Tag: model training

AI Machine Learning & Data Science Research

Achieving 8× Performance Gains with Reinforcement Learning on Synthetic Data in Large Language Models

In a new paper RL on Incorrect Synthetic Data Scales the Efficiency of LLM Math Reasoning by Eight-Fold, a research team provides insights into how synthetic data affects performance, suggesting that a specific schema can achieve consistent gains over using only positive data, achieving performance by 8× in synthetic data volume.

AI Machine Learning & Data Science Research

DeepMind & Toulouse U Contribute Composable Function Preserving Transformations to Boost Transformer Training

In a new paper Composable Function-preserving Expansions for Transformer Architectures, a research team from Google DeepMind and University of Toulouse introduces parameter expansion transformations for transformer-based neural networks while preserving functionality, enabling the expansion of the capability of the model as needed.