LLM.rocks 🤘

Welcome

Welcome to llm.rocks.

I'm a software engineer who finds LLMs (large language models) and their potential in the (near) future like magic. I don't work with LLMs, so this is my site to learn/educate about them.

I also keep this page up to date with LLM news/articles.

Biggest LLMs

namerelease datecompany# paramssize
BERT2018Google340 million3.3 billion words
GPT-22019OpenAI1.5 billion40 GB, 10 billion tokens
GPT-32020OpenAI175 billion500 billion tokens
GBP-Neo2021EleutherAI2.7 billion800 GB
GPT-J2021EleutherAI6 billion800 GB
Megatron-Turing NLG2021Nvidia/Microsoft500 billion300 billion tokens
Ernie 3.0 Titan2021Baidu260 billion4tb
Claude2021Anthropic25 billion400 billion tokens
GLaM (Generalist Language Model)2021Google1.2 trillion1.6 trillion tokens
Gopher2021DeepMind280 billion300 billion tokens
LaMDA (Language Models for Dialog Applications)2022Google137 billion1.5 trillion words, 160 billion tokens
GPT-NeoX2022EleutherAI20 billion800 gb
Chinchilla2022DeepMind70 billion1.4 trillion
PaLM (Pathways Language Model)2022Google540 billion768 billion tokens
OPT (Open pretrained transformer)2022Meta175 billion180 billion tokens
YaLM 100B2022Yandex100 billion1.7TB
Minerva2022Google540 billion38.5 tokens (from math content/academic papers)
BLOOM2022Hugging Face175 billion350 billion tokens/1.6TB
Galactica2022Meta120 billion106 billion tokens
AlexaTM (Teacher Models)2022Amazon20 billion1.3 trillion
LLaMA (large language model meta AI)2023Meta65 billion1.4 trillion
GPT-42023OpenAI
Cerebras-GPT2023Cerebras13 billion
Falcon2023Technology Innocation Institute40 billion1 Trillion tokens (1TB)
BloombergGPT2023Bloomberg50 billion360 billion tokens from Bloomberg data + 345 billion tokens from general datasets
PanGu2023Huawei1 trillion329 billion token

© 2019 llm.rocks • About/contact