Design Trade-offs: Determinism vs LLM Routing #1
jharjadi
announced in
Announcements
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Design Trade-offs: Determinism vs LLM Routing
One of the core architectural decisions in this project is how to balance deterministic routing with LLM-assisted classification.
The Trade-off
Pure Deterministic Routing
Pros:
Cons:
Pure LLM Routing
Pros:
Cons:
Our Hybrid Approach
This project uses a two-tier system:
See
core/intent_classifier.pyfor implementation.Why This Works
Open Questions
Should we add a confidence threshold?
Currently, if deterministic matching fails, we always call the LLM. Should we have a "confidence score" that triggers human review for ambiguous cases?
Should routing be cached?
Similar queries could reuse previous routing decisions. Trade-off: memory vs cost.
Should we support multi-intent queries?
Currently, we route to a single agent. Should we support queries that need multiple agents?
Your Thoughts?
What trade-offs have you encountered in agent routing? How would you approach this differently?
Related Documentation:
Beta Was this translation helpful? Give feedback.
All reactions