▲ | amarcheschi a day ago | |||||||
Results don't seem that bad for 9b https://huggingface.co/blog/eurollm-team/eurollm-9b | ||||||||
▲ | KronisLV a day ago | parent | next [-] | |||||||
I've been running it with Ollama, it's actually pretty good for working with text in Latvian (and other EU languages). I'd be hard pressed to find another model of a similar size that's good at it, for example: https://huggingface.co/spaces/openGPT-X/european-llm-leaderb... This won't be relevant to most people here, but it's cool to see even the smaller languages getting some love, instead of getting garbage outputs from Qwen (some versions of which are otherwise pretty good for programming) and anything below Llama 70B, or maybe looking at Gemma as a middle ground. | ||||||||
▲ | belter a day ago | parent | prev [-] | |||||||
"...EuroLLM-9B was trained on approximately 4 trillion tokens, using 400 Nvidia H100 GPUs on the MareNostrum5 supercomputer..." | ||||||||
|