Skip to content

docs(CLAUDE.md): refresh LLM provider routing notes for new src/llm/ architecture#6

Merged
offendingcommit merged 1 commit intomainfrom
docs/refresh-llm-routing-notes
May 4, 2026
Merged

docs(CLAUDE.md): refresh LLM provider routing notes for new src/llm/ architecture#6
offendingcommit merged 1 commit intomainfrom
docs/refresh-llm-routing-notes

Conversation

@offendingcommit
Copy link
Copy Markdown
Owner

The April-16 gotchas section in CLAUDE.md was stale after the upstream sync (#4 + #5):

  • Legacy cf / custom provider tags removed — replaced by ModelTransport literal (anthropic/openai/gemini) and per-component <COMPONENT>_MODEL_CONFIG__* env vars
  • "deriver/summary must stay on cf" rule no longer applies — native GeminiBackend (src/llm/backends/gemini.py) honors response_format=json_schema
  • thoughtSignature multi-iteration workaround obsolete — preserved natively in src/llm/history_adapters.py:77-78 + src/llm/executor.py:43-44
  • LM Studio section: env var names switched from LLM_OPENAI_COMPATIBLE_* globals to MODEL_CONFIG__OVERRIDES__BASE_URL / __API_KEY per component

Adds a note that CF Gateway integration is now app-level (cf-aig-authorization auto-injected by src/llm/registry.py and src/embedding_client.py based on base_url pattern matching) rather than deployment-level URL routing.

Ollama Cloud structured-output limitation kept — that's still a real upstream constraint, just rephrased for the new transport model.

Test plan

  • Documentation-only change; no runtime impact
  • All file paths referenced in the new section verified to exist (src/llm/registry.py, src/llm/backends/gemini.py, src/llm/history_adapters.py, src/llm/executor.py, src/llm/backends/openai.py)

…architecture

The April-16 gotchas section was stale after the upstream sync (PR #4):

- Legacy 'cf' / 'custom' provider tags removed — replaced by ModelTransport
  literal (anthropic/openai/gemini) and per-component MODEL_CONFIG__* env vars
- 'deriver/summary must stay on cf' rule no longer applies — native gemini
  backend (src/llm/backends/gemini.py) honors response_format=json_schema
- thoughtSignature multi-iteration workaround obsolete — preserved natively
  in src/llm/history_adapters.py + src/llm/executor.py
- LM Studio section: env var names switched from LLM_OPENAI_COMPATIBLE_*
  globals to MODEL_CONFIG__OVERRIDES__BASE_URL / __API_KEY per component

Adds a note that CF Gateway integration is now app-level
(cf-aig-authorization auto-injected by src/llm/registry.py and
src/embedding_client.py based on base_url pattern matching) rather than
deployment-level URL routing.

Ollama Cloud structured-output limitation kept — that's still a real
upstream constraint, just rephrased for the new transport model.
@offendingcommit offendingcommit merged commit 9b74886 into main May 4, 2026
1 check failed
@offendingcommit offendingcommit deleted the docs/refresh-llm-routing-notes branch May 4, 2026 03:18
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant