TutorFlow is currently in an early development stage (hackathon/early-stage project). At this time, only the main branch is actively maintained and supported.
| Version | Supported |
|---|---|
| main | ✅ |
We take security vulnerabilities seriously. If you discover a security issue in TutorFlow, please report it responsibly.
Please do NOT create a public GitHub issue for security vulnerabilities, as this could expose sensitive information.
Instead, please use one of the following methods:
- GitHub Security Advisories (Recommended)
- Go to the repository's Security tab
- Click on "Report a vulnerability"
- Create a private security advisory
- This ensures that the issue is handled confidentially
When reporting a vulnerability, please provide:
- A clear description of the vulnerability
- Steps to reproduce the issue
- The affected version/commit (if applicable)
- Potential impact and severity assessment
- Any suggested fixes or mitigations (if available)
- Your environment details (OS, Python version, Django version, etc.)
- Acknowledgment: We will acknowledge receipt of your report within 48 hours
- Initial Assessment: We will provide an initial assessment within 5 business days
- Resolution: We aim to address security issues promptly, with critical vulnerabilities prioritized
- Disclosure: Security issues will be disclosed responsibly after a fix is available, following coordinated disclosure practices
All security-related information will be handled confidentially and will not be shared publicly until a fix is available.
- PII Sanitizer: Prompts are sanitized before AI usage (full name, address, email, phone, tax_id, dob, medical_info →
[REDACTED]). - Mock Mode Default:
MOCK_LLM=1is enabled for demos and CI to avoid external LLM traffic unless explicitly disabled. - No Prompt Logs in Demo Mode: User prompts/responses are not persisted when running with mock mode.
- Sanitized Context Only: Lesson plans are generated from sanitized context to reduce exposure of personal data.
- Opt-in for Production: Live AI calls require setting
LLM_API_KEYand disablingMOCK_LLMexplicitly. - Hackathon Demos: During hackathon demos (e.g., CodeCraze Hackathon), TutorFlow runs in
MOCK_LLMmode with synthetic data only; no real student data or real LLM calls are used.
When using TutorFlow:
- Never commit sensitive data (API keys, passwords, real student data) to the repository
- Use environment variables for configuration (see
docs/DEPLOYMENT.md) - Keep dependencies updated (Dependabot is configured to help with this)
- Follow deployment guidelines in
docs/DEPLOYMENT.mdfor production environments - Use strong passwords and enable authentication features as appropriate
- No hardcoded secrets:
SECRET_KEY, database credentials, and LLM keys are expected from environment variables; defaults in the repo are demo-only - Secure defaults: In Produktion niemals mit
DEBUG=Trueoder leerenALLOWED_HOSTSbetreiben; aktivieren Sie bei BedarfSECURE_SSL_REDIRECT,SESSION_COOKIE_SECURE,CSRF_COOKIE_SECUREvia ENV.
We appreciate the security research community's efforts to help keep TutorFlow secure. Responsible disclosure helps protect all users of the application.