More The TWIML AI Podcast episodes

The Evolution of Reasoning in Small Language Models with Yejin Choi thumbnail

The Evolution of Reasoning in Small Language Models with Yejin Choi

Published 29 Jan 2026

Duration: 3981

Language models can struggle to generate diverse responses, despite adjustments, and improving their performance requires addressing data quality and training techniques.

Episode Description

Today, we're joined by Yejin Choi, professor and senior fellow at Stanford University in the Computer Science Department and the Institute for Human-C...

Overview

The podcast examines how both small and large language models tend to generate similar, homogeneous outputs in response to open-ended prompts, even when temperature settings are modified to encourage diversity. It investigates research efforts aimed at improving small language models (SLMs) by refining their reasoning abilities through methods like better data curation, synthetic data generation, and hybrid model designs. The discussion points out the limitations of relying on internet-based training data and stresses the value of high-quality, specialized content created by humans to train SLMs more effectively.

The podcast also reviews techniques such as imitation learning, reinforcement learning with verification, and data filtering to enhance model performance and output diversity. Additionally, it raises broader concerns about the effects of AI on human creativity and thought, as well as the potential for AI to contribute to the homogenization of online content. Emphasis is placed on the importance of making AI more accessible, exploring diverse alignment approaches, and developing systems that are more data-efficient and ethically responsible.

Recent Episodes of The TWIML AI Podcast

26 Mar 2026 The Race to Production-Grade Diffusion LLMs with Stefano Ermon

The text traces generative models' evolution from early image generation to diffusion models' stability, highlights Mercury II's advancements in speed and efficiency, and addresses ongoing challenges in scalability, multimodal integration, and future research in controllability and cross-modal unification.

More The TWIML AI Podcast episodes