Tag: large language model

AI Machine Learning & Data Science Research

From Token to Conceptual: Meta introduces Large Concept Models in Multilingual AI

A research team at Meta introduces the Large Concept Model (LCM), a novel architecture that processes input at a higher semantic level. This shift allows the LCM to achieve remarkable zero-shot generalization across languages, outperforming existing LLMs of comparable size.

AI Machine Learning & Data Science Research

Achieving 8× Performance Gains with Reinforcement Learning on Synthetic Data in Large Language Models

In a new paper RL on Incorrect Synthetic Data Scales the Efficiency of LLM Math Reasoning by Eight-Fold, a research team provides insights into how synthetic data affects performance, suggesting that a specific schema can achieve consistent gains over using only positive data, achieving performance by 8× in synthetic data volume.