-
2026-02-05
Why NVIDIA builds their own open models | Nemotron w/ Bryan Catanzaro
-
2026-01-28
"All in on building open models in the U.S."
-
2026-01-10
8 Graphs Telling Today's Story of Open Models
-
2025-12-24
2025 Hot or Not: Can Western AI Labs Catch Up? (Technical vs. Business Hurdles)
-
2025-12-19
Winners and Losers of Open Models in 2025
-
2025-12-19
GPT OSS: Why Itโs Brilliant, But Brittle | The Future of OpenAIโs Models
-
2025-12-11
My talk at the Curve '25
-
2025-12-11
How We Built a Leading Reasoning Model (Olmo 3)
-
2025-11-15
Some models are like one-way doors (that we cannot delete) and present risk
-
2025-11-14
Approximately 80% of American companies are using Chinese models like Qwen
-
2025-11-14
Big tech companies are going strong while the rest of the economy is being left behind
-
2025-11-12
They Built an AGI Lab in 8 Months
-
2025-11-11
China is dominating the Open Model space
-
2025-10-16
The State of Open Models
-
2025-08-08
GPT-5's simultaneous excitement and disappointment
-
2025-08-08
GPT 5 Release Live Reaction
-
2025-08-04
The ATOM Project: American Truly Open Models
-
2025-07-29
Ross Taylor, Ex-Llama reasoning lead, on Chinese open models, scaling RL, & the next 6 months in AI
-
2025-06-18
The art of training a good (reasoning) language model
-
2025-03-12
Self-play for Self-driving and where Scaling Reinforcement Learning is Heading with Eugene Vinitsky
-
2025-02-14
An Unexpected Reinforcement Learning Renaissance
-
2025-01-23
OLMo leads on the secrets of training language models (w Dirk Groeneveld, Kyle Lo, & Luca Soldaini)
-
2025-01-08
How language model post-training is done today
-
2025-01-03
Quick recap on the state of language model reasoning
-
2024-12-10
How to have impact in an AI job
-
2024-12-07
Google's AI Infrastructure Advantage -- Past and Present
-
2024-12-06
The Bitter Lesson: What people get wrong
-
2024-12-05
Finbarr Timbers on the Future of Reinforcement Learning
-
2024-11-07
Tim Dettmers on Open-source AI, LMs, SWE Bench, Agents, Quantization, & Optimization
-
2024-11-05
OpenAI's obsession with the "low key research preview" idea
-
2024-11-01
3 Pillars of OpenAI's Research Culture
-
2024-10-31
Why the Meta RayBans can be your favorite tech gadget
-
2024-10-31
Andrew Carr on Pushing the Boundaries of Generative AI (Beyond Text)
-
2024-10-31
How to effectively use OpenAI's New Voice Mode
-
2024-10-17
Arvind Narayanan on making sense of AI hype
-
2024-10-11
Should LLMs hold syntactic or semantic information?
-
2024-10-10
Andrew Trask on changing how LLMs store and access information
-
2024-10-10
Why do AI's struggle with reading clocks?
-
2024-10-09
Building the internet for non public information
-
2024-10-05
How Sebastian Raschka saves time with AI assistants #chatgpt #languagemodels
-
2024-10-01
OpenAI's o1, RL, and how it impacts prompting
-
2024-10-01
Prompting ChatGPT to "Think Silently"
-
2024-09-30
Riley Goodside on OpenAI's o1 and the frontier of prompting LLMs
-
2024-09-30
The government's love-hate relationship with open-source software (and AI)
-
2024-09-30
Can a language model discover new science?
-
2024-09-29
How LLMs work through reasoning
-
2024-09-29
Why compute thresholds don't work for AI policy
-
2024-09-28
Llama 3.2 vision vs. Molmo: Open multimodal models
-
2024-09-05
OpenAI's Strawberry and Spending More on Inference
-
2024-09-04
OLMoE and training better foundation models
-
2024-08-28
Defining open-source AI and open data
-
2024-08-08
Interviewing Ross Taylor on LLM reasoning, Llama post training, and Galactica
-
2024-08-01
Interviewing Sebastian Raschka on the state of open LLMs, Llama 3.1, and AI education
-
2024-07-24
Llama 3.1 405b and Meta's "open source" AI strategy
-
2024-07-04
Switching to Claude 3.5 from ChatGPT
-
2024-06-28
Interviewing Dean Ball on AI policy: SB-1047, Llama 405B, scaling laws, China, and more
-
2024-06-22
Frontiers in LLM synthetic data
-
2024-06-13
How Apple Intelligence Works
-
2024-06-05
A realistic path to robotic foundation models
-
2024-05-30
There is no LLM data wall