▲ | icelancer 3 months ago | |||||||||||||
Nice use of an LLM - we use Groq 70b models for this in our pipelines at work. (After using WhisperX ASR on meeting files and such) One of the better reasons to use Cerebras/Groq that I've found so you can return huge amounts of clean text back fast for processing in other ways. | ||||||||||||||
▲ | ldenoue 3 months ago | parent [-] | |||||||||||||
Although Gemini accepts very long input context, I found that sending more than 512 or so words at a time to the LLM for "cleaning up the text" yields hallucinations. That's why I chunk the raw transcript into 512-word chunks. Are you saying it works with 70B models on Groq? Mixtral, Llama? Other? | ||||||||||||||
|