Sitemap - 2025 - Interconnects

What people get wrong about the leading Chinese open models: Adoption and censorship

Sycophancy and the art of the model

State of play of AI progress (and related brakes on an intelligence explosion)

Qwen 3: The new open standard

Transparency and (shifting) priority stacks

The latest open artifacts (#9): RLHF book draft, where the open reasoning race is going, and unsung heroes of open LM work

OpenAI's o3: Over-optimization is back and weirder than ever

OpenAI's GPT-4.1 and separating the API from ChatGPT

Llama 4: Did Meta just push the panic button?

RL backlog: OpenAI's many RLs, clarifying distillation, and latent reasoning

Recent reasoning research: GRPO tweaks, base model RL, and data curation

GPT-4o's images and lessons from native input-output multimodality

Gemini 2.5 Pro and Google's second chance with AI

The latest open artifacts (#8): The return of ~30B models, side effects of OpenAI's proposed DeepSeek ban, and yet another reasoning roundup

Managing frontier model training organizations (or teams)

Gemma 3, OLMo 2 32B, and the growing potential of open-source AI

Interviewing Eugene Vinitsky on self-play for self-driving and what else people do with RL

Elicitation, the simplest way to understand post-training

Where inference-time scaling pushes the market for AI companies

GPT-4.5: "Not a frontier model"?

Character training: Understanding and crafting a language model's personality

Claude 3.7 thonks and what's next for inference-time scaling

The latest open artifacts (#7): Alpaca era of reasoning models, China's continued dominance, and tons of multimodal advancements

Grok 3 and an accelerating AI roadmap

An unexpected RL Renaissance

Deep Research, information vs. insight, and the nature of science

Making the U.S. the home for open-source AI

Why reasoning models will generalize

The latest open artifacts (#6): Reasoning models, China's lead in open-source, and a growing multimodal space

Interviewing OLMo 2 leads: Open secrets of training language models

DeepSeek R1's recipe to replicate o1 and the future of reasoning LMs

Let me use my local LMs on Meta Ray-Bans

DeepSeek V3 and the actual cost of training frontier AI models

(Voiceover) DeepSeek V3 and the actual cost of training frontier AI models

The state of post-training in 2025

Quick recap on the state of reasoning