Sitemap - 2024 - Interconnects

OpenAI chases Her

OpenAI’s Model (behavior) Spec, RLHF transparency, personalization questions

ChatBotArena: The peoples’ LLM evaluation, the future of evaluation, the incentives of evaluation, and gpt2chatbot

How RLHF works, part 2: A thin line between useful and lobotomized

Phi 3 and Arctic: Outlier LMs are hints

AGI is what you want it to be

Llama 3: Scaling open LLMs to AGI

Stop "reinventing" everything to solve alignment

The end of the “best open LLM”

We disagree on what open-source AI should mean

DBRX: The new best open model and Databricks’ ML strategy

Evaluations: Trust, performance, and price (bonus, announcing RewardBench)

Model commoditization and product moats

The koan of an open-source LLM

Interviewing Louis Castricato of Synth Labs and Eleuther AI on RLHF, Gemini Drama, DPO, founding Carper AI, preference data, reward models, and everything in between

How to cultivate a high-signal AI feed

Google ships it: Gemma open LLMs and Gemini backlash

10 Sora and Gemini 1.5 follow-ups: code-base in context, deepfakes, pixel-peeping, inference costs, and more

OpenAI’s Sora for video, Gemini 1.5's infinite context, and a secret Mistral model

Why reward models are key for alignment

Alignment-as-a-service: Scale AI vs. the new guys

Open Language Models (OLMos) and the LLM landscape

Model merging lessons in The Waifu Research Department

Local LLMs, some facts some fiction

Multimodal blogging: My AI tools to expand your audience

RLHF learning resources in 2024

Multimodal LM roundup: Unified IO 2, inputs and outputs, Gemini, LLaVA-RLHF, and RLHF questions

Where 2024’s “open GPT4” can’t match OpenAI’s

It's 2024 and they just want to learn