| ▲ | embedding-shape 4 hours ago | |||||||||||||||||||||||||
Right, and my point is that if you use 80% Brazilian Portuguese during base model training + 20% European Portuguese as post-training, you pretty much get exactly that, except with a ton more of available training data. | ||||||||||||||||||||||||||
| ▲ | KK7NIL 4 hours ago | parent [-] | |||||||||||||||||||||||||
What's your evidence for that? And if the first 80% doesn't bias the language after post-training (which I think is what you're claiming) why not go for English or a mixture of languages, which is essentially what they did by starting with EuroLLM? | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||