AMÁLIA and the future of European Portuguese LLMs

duarteocarmo.com

134 points by johnbarron 4 days ago


pu_pe - 16 hours ago

I'm not sure the direction should be to finetune a small local model for each country or language. These models are already not particularly great at information retrieval, so I doubt anyone would use them for questions like the author suggests (ie who was the president between X and Y). Similarly, they are a little too lightweight to be used for translations too.

If the budget is indeed so modest (5.5 million euros!), I would focus completely on preparing datasets and making sure all open cultural artifacts that we can find are well documented in them. That way every model, private or open, that gets trained in the future could better represent the culture and language of your country.

mariopt - 13 hours ago

This model is a waste of Public Funds.

There is no public website to use it, be it free or paid, the dataset is not public, the code is not public (The github URL in the article returns 404 ), the claimed model intelligence is so low that is pretty much useless at 32K context and massively inferior to GPT‑4o.

As per tradition in Portugal, some people managed to get 5.5 Million to produce nothing and no one is asking questions.

You want a better idea? Just fine tune the open source Kimi 2.6 with an open source Portuguese dataset, the cost would be under a million and we would be getting something useful.

It would be really nice to know what happened to 5.5 Millions whilst not being able to even provide a functional website to use the model.

alexaholic - 11 hours ago

The Amália model is not yet publicly available. Until it's ready, one can fool around with Anália at https://analia.pt

swiftcoder - 16 hours ago

It is definitely an interesting problem, because Portugal is a small enough country that the actual total corpus of available texts in (non-Brazilian) Portuguese is potentially problematic.

drivebyhooting - 11 hours ago

I’ve noticed that ChatGPT is noticeably dumber in languages other than English. It even will confidently repeat common but wrong superstitions from the target language as if they were fact.

mt_ - 13 hours ago

5 million for a llama-2 finetune, how is that impressive?

r2ob - 12 hours ago

"This model is a waste of Public Funds". There is no "public funds", this is a waste of money from the tax payers.

algoth1 - 16 hours ago

Wouldnt it be easier to fine tune a model to convert the Brazilian Portuguese corpus into European Portuguese and then use that corpus?

simianwords - 15 hours ago

Domain specific models will never be a thing. You don't get generalised intelligence with that.

https://simianwords.bearblog.dev/why-domain-specific-llms-wo...

hartator - 16 hours ago

What a waste of time and money.

Trying to force a LLM into a specific language makes you missed out on most of the world knowledge.