The Ollama library currently has zero natively trained Brazilian Portuguese models. Adding these models would significantly expand Ollama's reach in Latin America's largest AI market.
# [Feature Request] Add Brazilian Portuguese (PT-BR) native LLM models to the Ollama Library ## Summary The Ollama official library currently has **zero natively trained Brazilian Portuguese models**. With 260M+ Portuguese speakers worldwide (215M+ in Brazil alone) and a growing ecosystem of open-source PT-BR LLMs, adding these models would significantly expand Ollama's reach in Latin America's largest AI market. ## The Problem While models like Llama 3.2/3.3/4, Gemma, and Qwen include Portuguese as a "supported language," they are **not optimized** for Brazilian Portuguese. This leads to: - **Poor tokenizer efficiency**: Models using English-centric tokenizers (e.g., Llama 2's BPE) encode Portuguese text very inefficiently — often 2-3x more tokens than necessary for the same content ([research from Tucano paper](https://arxiv.org/html/2411.07854v1)). - **Cultural blindspots**: Mainstream models struggle with Brazilian-specific concepts (PIX payment system, CPF/CNPJ, ENEM, vestib