| ▲ | AlphaWeaver 2 hours ago | |||||||
Do you think there's a path where you can pregenerate popular paths of dialogue to avoid LLM inference costs for every player? And possibly pair it with a lightweight local LLM to slightly adapt the responses? While still shelling out to a larger model when users go "off the rails"? | ||||||||
| ▲ | themanmaran 2 hours ago | parent [-] | |||||||
Not the founder, but having run conversational agents at decent scale, I don't think the cost actually matters much early on. It's almost always better to pay more for the smarter model, than to potentially give a worse player experience. If they had 1M+ players there would certainly be room to optimize, but starting out you'd certainly spend more trying engineer the model switcher than you would save in token costs. | ||||||||
| ||||||||