| ▲ | empressplay 12 hours ago | |
That output is there for a reason. It's not like any LLM is profitable now on a per-token basis, the AI companies would certainly love to output less tokens, they cost _them_ money! The entire hypothesis for doing this is somewhat dubious. | ||
| ▲ | verdverm 8 hours ago | parent [-] | |
Why building / using a custom agent stack and paying per-token (not subscription) is more efficient and cost effective. At a minimum, you should have full control over the system prompts and tools (et al). | ||