Improving Reasoning Abilities of Large Language Models through Imperfect Synthetic Data
- 2025-04-07 (Mon.), 10:30 AM
- Auditorium, B1F, Institute of Statistical Science;The tea reception will be held at 10:10.
- Online live streaming through Cisco Webex will be available.
- Dr. 許大山
- MediaTek
Abstract
In this presentation, we will discuss our recent publication, "RL-STaR: Theoretical Analysis of Reinforcement Learning Frameworks for Self-Taught Reasoner." Our findings indicate that synthetic data do not need to be completely accurate; errors in reasoning steps are acceptable as large language models (LLMs) can still learn the correct answers over time. Parallel to the theoretical treatment, we reflect on the use of synthetic data in LLM-based assistant for programming in general and for Register Transfer Level (RTL) design in particular.
Please click here for participating the talk online.
Download
Update:2025-03-31 14:39