Title: A Gentle Introduction to LLM Reasoning


Abstract: 
Large language models are built from three deceptively simple ideas: a parametrized function trained to predict the next word, human feedback that steers those same parameters toward preferred behavior, and a fictional frame made real by intercepting the right tokens. Yet when scaled, these models appear to reason, plan, and solve complex problems. This talk unpacks how modern LLMs reason and how we make them reason better. We trace the path from chain-of-thought prompting through inference-time scaling techniques (self-consistency and self-refinement) to reinforcement learning with verifiable rewards (GRPO) and knowledge distillation. A 0.6B-parameter model can be pushed from 15% to over 50% accuracy on competition math through a combination of these techniques. We discuss the key trade-offs between compute cost, accuracy, and training stability that define this rapidly evolving field.

Dates

March 11, 2026

Abstract submission deadline

March 18, 2026

Paper submission deadline

April 22, 2026

Author notification

June 10-12, 2026

Netys Conference

Proceedings

Partners & Sponsors (TBA)