innercontext/backend
Piotr Oleszczyk ada5f2a93b fix(llm): disable Gemini thinking to prevent MAX_TOKENS on structured output
Gemini 2.5 Flash (gemini-flash-latest) enables thinking by default.
Thinking tokens count toward max_output_tokens, leaving ~150 tokens for
actual JSON output and causing MAX_TOKENS truncation. Disable thinking
centrally in call_gemini via ThinkingConfig(thinking_budget=0).

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
2026-03-01 20:12:31 +01:00
..
alembic fix(llm): log and handle non-STOP finish_reason from Gemini 2026-03-01 20:08:22 +01:00
innercontext fix(llm): disable Gemini thinking to prevent MAX_TOKENS on structured output 2026-03-01 20:12:31 +01:00
tests feat(mcp): add FastMCP server with 14 tools for LLM agent access 2026-02-28 17:59:11 +01:00
.env.example fix: load .env via python-dotenv; SQLite default for local dev 2026-02-26 20:51:13 +01:00
.python-version Initial commit: backend API, data models, and test suite 2026-02-26 15:10:24 +01:00
alembic.ini feat(backend): add Alembic migrations 2026-02-28 20:14:57 +01:00
db.py Initial commit: backend API, data models, and test suite 2026-02-26 15:10:24 +01:00
main.py feat(routines): add minoxidil beard/mustache option to routine suggestions 2026-03-01 19:46:07 +01:00
pyproject.toml feat(backend): add Alembic migrations 2026-02-28 20:14:57 +01:00
README.md Initial commit: backend API, data models, and test suite 2026-02-26 15:10:24 +01:00
skincare.yaml Initial commit: backend API, data models, and test suite 2026-02-26 15:10:24 +01:00
uv.lock feat(backend): add Alembic migrations 2026-02-28 20:14:57 +01:00

See the root README for setup and usage instructions.