| ▲ | siim 9 hours ago | |
Curious what made you think the backend uses LLMs for content generation? To clarify: 1. transcription is local VOSK speech-to-text via WebSocket 2. live transcript post-processing has optional Gemini Flash-lite turned on which tries to fix obvious transcription mistakes, nothing else. The real fix here is more accurate transcriber. 3. backend: TypeGraphQL + MongoDB + Redis The anti-AI stance isn't "zero AI anywhere", it's about requiring human input. AI-generated audio is either too bad or too perfect. Real recorded voice has human imperfections. | ||