New Paradigm: AI Research Summaries

New Paradigm: AI Research Summaries

This podcast provides audio summaries of new Artificial Intelligence research papers. These summaries are AI generated, but every effort has been made by the creators of this podcast to ensure they are of the highest quality. As AI systems are prone to hallucinations, our recommendation is to always seek out the original source material. These summaries are only intended to provide an overview of the subjects, but hopefully convey useful insights to spark further interest in AI related matters.

Episodes

February 23, 2025 8 mins
This episode analyzes the study "Competitive Programming with Large Reasoning Models," conducted by researchers from OpenAI, DeepSeek-R1, and Kimi k1.5. The research investigates the application of reinforcement learning to enhance the performance of large language models in competitive programming scenarios, such as the International Olympiad in Informatics (IOI) and platforms like CodeForces. It compares general-purpose models, i...
Mark as Played
This episode analyzes the study "ZebraLogic: On the Scaling Limits of LLMs for Logical Reasoning," conducted by Bill Yuchen Lin, Ronan Le Bras, Kyle Richardson, Ashish Sabharwal, Radha Poovendran, Peter Clark, and Yejin Choi from the University of Washington, the Allen Institute for AI, and Stanford University. The research examines the capabilities of large language models (LLMs) in handling complex logical reasoning tasks by intr...
Mark as Played
This episode analyzes "s1: Simple test-time scaling," a research study conducted by Niklas Muennighoff, Zitong Yang, Weijia Shi, Xiang Lisa Li, Li Fei-Fei, Hannaneh Hajishirzi, Luke Zettlemoyer, Percy Liang, Emmanuel Candès, and Tatsunori Hashimoto from Stanford University, the University of Washington in Seattle, the Allen Institute for AI, and Contextual AI. The research investigates an innovative approach to enhancing language m...
Mark as Played
This episode analyzes "AI Tools in Society: Impacts on Cognitive Offloading and the Future of Critical Thinking," a study conducted by Michael Gerlich at the Center for Strategic Corporate Foresight and Sustainability, SBS Swiss Business School. The research examines how the use of artificial intelligence tools influences critical thinking skills by introducing the concept of cognitive offloading—relying on external tools to perfor...
Mark as Played
This episode analyzes the "Multimodal Visualization-of-Thought" (MVoT) study conducted by Chengzu Li, Wenshan Wu, Huanyu Zhang, Yan Xia, Shaoguang Mao, Li Dong, Ivan Vulić, and Furu Wei from Microsoft Research, the University of Cambridge, and the Chinese Academy of Sciences. The discussion delves into MVoT's innovative approach to enhancing the reasoning capabilities of Multimodal Large Language Models (MLLMs) by integrating visua...
Mark as Played
This episode analyzes the research paper titled "Increased Compute Efficiency and the Diffusion of AI Capabilities," authored by Konstantin Pilz, Lennart Heim, and Nicholas Brown from Georgetown University, the Centre for the Governance of AI, and RAND, published on February 13, 2024. It examines the rapid growth in computational resources used to train advanced artificial intelligence models and explores how improvements in hardwa...
Mark as Played
This episode analyzes the research paper "Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs," authored by Yue Wang and colleagues from Tencent AI Lab, Soochow University, and Shanghai Jiao Tong University. The study investigates the phenomenon of "underthinking" in large language models similar to OpenAI's o1, highlighting their tendency to frequently switch between lines of thought without thoroughly exploring ...
Mark as Played
This episode analyzes the study "On the Overthinking of o1-Like Models" conducted by researchers Xingyu Chen, Jiahao Xu, Tian Liang, Zhiwei He, Jianhui Pang, Dian Yu, Linfeng Song, Qiuzhi Liu, Mengfei Zhou, Zhuosheng Zhang, Rui Wang, Zhaopeng Tu, Haitao Mi, and Dong Yu from Tencent AI Lab and Shanghai Jiao Tong University. The research investigates the efficiency of o1-like language models, such as OpenAI's o1, Qwen, and DeepSeek, ...
Mark as Played
This episode analyzes the research paper titled "In-Context Learning of Representations," authored by Core Francisco Park, Andrew Lee, Ekdeep Singh Lubana, Yongyi Yang, Maya Okawa, Kento Nishi, Martin Wattenberg, and Hidenori Tanaka from Harvard University, NTT Research Inc., and the University of Michigan. The discussion delves into how large language models, specifically Llama3.1-8B, adapt their internal representations of concep...
Mark as Played
This episode analyzes the research paper titled "Agent Laboratory: Using LLM Agents as Research Assistants," authored by Samuel Schmidgall, Yusheng Su, Ze Wang, Ximeng Sun, Jialian Wu, Xiaodong Yu, Jiang Liu, Zicheng Liu, and Emad Barsoum from AMD and Johns Hopkins University. The discussion delves into how the Agent Laboratory framework leverages Large Language Models (LLMs) to enhance the scientific research process by automating...
Mark as Played
This episode analyzes the research paper "Evolving Deeper LLM Thinking" by Kuang-Huei Lee, Ian Fischer, Yueh-Hua Wu, Dave Marwood, Shumeet Baluja, Dale Schuurmans, and Xinyun Chen from Google DeepMind, UC San Diego, and the University of Alberta. It explores the innovative Mind Evolution approach, which employs evolutionary search strategies to enhance the problem-solving abilities of large language models (LLMs) without the need f...
Mark as Played
This episode analyzes the study "Predicting Human Brain States with Transformer" conducted by Yifei Sun, Mariano Cabezas, Jiah Lee, Chenyu Wang, Wei Zhang, Fernando Calamante, and Jinglei Lv from the University of Sydney, Macquarie University, and Augusta University. The discussion explores how transformer models, originally developed for natural language processing, are utilized to predict future brain states using functional magn...
Mark as Played
This episode analyzes "DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning," a study conducted by Daya Guo and colleagues at DeepSeek-AI, published on January 22, 2025. The discussion focuses on how the researchers utilized reinforcement learning to enhance the reasoning abilities of large language models (LLMs), introducing models such as DeepSeek-R1-Zero and DeepSeek-R1. It examines the models' impr...
Mark as Played
This episode analyzes the study "Titans: Learning to Memorize at Test Time" by Ali Behrouz, Peilin Zhong, and Vahab Mirrokni from Google Research. It examines the researchers' innovative approach to enhancing artificial intelligence models' memory capabilities, addressing the limitations of traditional recurrent neural networks and Transformer models. The discussion highlights the introduction of a neural long-term memory module an...
Mark as Played
This episode analyzes the research paper titled **"Search-o1: Agentic Search-Enhanced Large Reasoning Models,"** authored by Xiaoxi Li, Guanting Dong, Jiajie Jin, Yuyao Zhang, Yujia Zhou, Yutao Zhu, Peitian Zhang, and Zhicheng Dou from Renmin University of China and Tsinghua University, published on January 9, 2025. The discussion focuses on the Search-o1 framework, which enhances large reasoning models by incorporating an agentic ...
Mark as Played
This episode analyzes the research paper "TRANSFORMER2: SELF-ADAPTIVE LLM S" by Qi Sun, Edoardo Cetin, and Yujin Tang from Sakana AI and the Institute of Science Tokyo, published on January 14, 2025. It explores the development of Transformer2, a self-adaptive large language model designed to dynamically adjust its behavior in real time without requiring additional training or human intervention. The analysis delves into the novel ...
Mark as Played
This episode analyzes "OASIS: OpenAgent Social Interaction Simulations with One Million Agents," a research initiative conducted by a diverse team from institutions including the Shanghai Artificial Intelligence Laboratory, Oxford, and the Max Planck Institute. The discussion explores the development of OASIS, a scalable and generalizable social media simulator designed to model interactions among up to one million agents. By integ...
Mark as Played
This episode analyzes Rafid Mahmood's paper, "Pricing and Competition for Generative AI," authored by Mahmood from NVIDIA and the University of Ottawa, and published on November 4, 2024. It delves into the complexities of pricing strategies for generative artificial intelligence models, examining how companies determine optimal pricing based on model performance and competitive market dynamics. The discussion introduces key concept...
Mark as Played
This episode analyzes the research paper "**Compact Language Models via Pruning and Knowledge Distillation**" authored by Saurav Muralidharan, Sharath Turuvekere Sreenivas, Raviraj Joshi, Marcin Chochowski, Mostofa Patwary, Mohammad Shoeybi, Bryan Catanzaro, Jan Kautz, and Pavlo Molchanov from **NVIDIA**, published on November 4, 2024. It explores NVIDIA's strategies for reducing the size of large language models by implementing st...
Mark as Played
This episode analyzes the "Phi-4 Technical Report," published on December 12, 2024, by a team of researchers from Microsoft Research, including Marah Abdin, Jyoti Aneja, Harkirat Behl, Stéphane Bubeck, and others. The discussion delves into the Phi-4 language model's architecture, which comprises 14 billion parameters, and its innovative training approach that emphasizes data quality and the strategic use of synthetic data. It expl...
Mark as Played

Popular Podcasts

    If you've ever wanted to know about champagne, satanism, the Stonewall Uprising, chaos theory, LSD, El Nino, true crime and Rosa Parks, then look no further. Josh and Chuck have you covered.

    Dateline NBC

    Current and classic episodes, featuring compelling true-crime mysteries, powerful documentaries and in-depth investigations. Follow now to get the latest episodes of Dateline NBC completely free, or subscribe to Dateline Premium for ad-free listening and exclusive bonus content: DatelinePremium.com

    New Heights with Jason & Travis Kelce

    Football’s funniest family duo — Jason Kelce of the Philadelphia Eagles and Travis Kelce of the Kansas City Chiefs — team up to provide next-level access to life in the league as it unfolds. The two brothers and Super Bowl champions drop weekly insights about the weekly slate of games and share their INSIDE perspectives on trending NFL news and sports headlines. They also endlessly rag on each other as brothers do, chat the latest in pop culture and welcome some very popular and well-known friends to chat with them. Check out new episodes every Wednesday. Follow New Heights on the Wondery App, YouTube or wherever you get your podcasts. You can listen to new episodes early and ad-free, and get exclusive content on Wondery+. Join Wondery+ in the Wondery App, Apple Podcasts or Spotify. And join our new membership for a unique fan experience by going to the New Heights YouTube channel now!

    The Clay Travis and Buck Sexton Show

    The Clay Travis and Buck Sexton Show. Clay Travis and Buck Sexton tackle the biggest stories in news, politics and current events with intelligence and humor. From the border crisis, to the madness of cancel culture and far-left missteps, Clay and Buck guide listeners through the latest headlines and hot topics with fun and entertaining conversations and opinions.

    The Bobby Bones Show

    Listen to 'The Bobby Bones Show' by downloading the daily full replay.

Advertise With Us
Music, radio and podcasts, all free. Listen online or download the iHeart App.

Connect

© 2025 iHeartMedia, Inc.