Token Usage #971
Replies: 2 comments
-
|
Had the same experience with multi-agent setups — V3's orchestration layer adds significant overhead because each agent pass includes the full conversation context plus coordination prompts. A few things that helped me get token usage under control: 1. Instrument first, optimize second If you're using the Node.js SDK directly, burn0 can give you per-request cost breakdowns with a single import — it intercepts HTTP calls and logs exactly what each agent step costs. Helped me identify that one summarization step was re-sending the entire research context (~80k tokens) when it only needed the conclusions. 2. Context window management
4. Check for retry loops The weekly limit issue suggests your usage jumped 3-4x, which lines up with what I've seen when moving to more sophisticated orchestration patterns. The tokens-per-task metric is the one to watch. |
Beta Was this translation helpful? Give feedback.
-
|
We investigated this exact problem. Here's where your tokens are going:
The intelligence layer is the biggest culprit: it builds a 100 MB graph from 5,706 entries (only ~20 unique, rest are duplicates), runs PageRank, then injects the same entry 5 times per message. Ironically, the README claims '30-50% token reduction' — but that metric is fabricated ( Quick fixes to reduce token burn:
Full token cost analysis: Independent Audit |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Since using V3 my token usage has increased significantly. I am burning thour my weekly limit in a two days. Anyone experiencing this also? Any Ideas how improve that? At this state I have to stop using V3.
Beta Was this translation helpful? Give feedback.
All reactions