Developer of the LLaMA series of open-source large language models.
Meta
Latest Meta AI model, at ~400B parameters and trained on 40T tokens. Intended to push open-model performance to near-GPT-4 level. Released under a new Llama 4 community license. Early indications show significant gains in reasoning and multilingual understanding.
Third-generation Llama model, reportedly 405B parameters with 15.6T tokens of training. Introduced a massive context window (128k tokens) and expanded multilingual capability. At release, it was the largest openly available model to date.
70B-parameter successor to LLaMA, trained on 2 trillion tokens and fine-tuned for helpful dialogue. Released openly under a custom license allowing commercial use with some restrictions. Strong performance rivaling closed models like ChatGPT on many benchmarks.
Large Language Model Meta AI, released in sizes up to 65B parameters for research. Achieved strong performance by training on 1.4T tokens; weights were leaked, spurring a wave of fine-tuned variants.
120B-parameter model aimed at scientific text and knowledge integration. Notorious for generating plausible-sounding but incorrect scientific content; public demo was retracted.
Open Pretrained Transformer, a reproduction of GPT-3 (175B) made openly available to researchers. Notably released with training logs and a non-commercial license.