Tag: ML

AI Machine Learning & Data Science Research

Breaking LLMs’ Limits: Upstage AI’s SOLAR 10.7B Shines Bright with Simple Scaling Magic

In a new paper SOLAR 10.7B: Scaling Large Language Models with Simple yet Effective Depth Up-Scaling, a Upstage AI research team introduces depth up-scaling (DUS), which emerges as an efficient and uncomplicated technique for amplifying LLMs, surpassing existing open-source state-of-the-art LLMs, such as Llama 2 and Mistral 7B.

AI Machine Learning & Data Science Nature Language Tech Research

A Robot Chemist Driven by GPT-4 Made Its Debut in Nature: Autonomously Designs Reactions and Performs Complex Experiments

In a new paper Autonomous chemical research with large language models, a research team from Carnegie Mellon University and Emerald Cloud Lab introduces an innovative LLMs-Powered system named Coscientist, which autonomously designs, plans, and executes complex scientific experiments, marking a significant leap forward in the integration of laboratory automation technologies with powerful language models.

AI Machine Learning & Data Science Research

Microsoft’s TaskWeaver: Empowering Intelligent Conversational Agents for Handling Domain-Specific Complex Tasks

A Microsoft research team introduces TaskWeaver, a cutting-edge, code-first framework designed to empower LLM-powered autonomous agents. TaskWeaver offers a potent and flexible platform for constructing intelligent conversational agents capable of handling complex tasks and seamlessly adapting to domain-specific scenarios.

AI Machine Learning & Data Science Research

Spatial-Temporal Innovation: STLVQE Redefines Real-Time Video Enhancement for an Unmatched Viewing Experience

A paper titled “Online Video Quality Enhancement with Spatial-Temporal Look-up Tables” introduces a novel method, STLVQE. This research, conducted by a team from Tongji University and Microsoft Research Asia, pioneers the exploration of the online video quality enhancement problem and presents the first method achieving real-time processing speed.

AI Machine Learning & Data Science Research

DeepMind’s DiLoCo Revolutionizes Language Model Training with 500× Less Communication

In a new paper DiLoCo: Distributed Low-Communication Training of Language Models, a Google DeepMind research team presents Distributed Low-Communication (DiLoCo). DiLoCo employs a distributed optimization algorithm that facilitates the training of language models on islands of poorly connected devices, surpassing the performance of fully synchronous models while reducing communication by 500 times.

AI Machine Learning & Data Science Research

Microsoft Orca 2’s Triumph: Comparable or Superior Performance to Models 5-10x Its Size in Mastering Reasoning Tasks

Microsoft has recently unveiled Orca 2 in a new paper titled “Orca 2: Teaching Small Language Models How to Reason.” to explore how enhanced training signals can augment the reasoning abilities of smaller language models. Notably, Orca 2 surpasses models of similar size, achieving performance levels comparable to or better than models 5-10 times larger.

AI Computer Vision & Graphics Machine Learning & Data Science Research

Adobe & ANU’s LRM Reconstructs Models For Single Image to 3D in 5s

In a new paper LRM: Large Reconstruction Model for Single Image to 3D, a research team from Adobe Research and Australian National Univerisity introduces an innovative Large Reconstruction Model (LRM). This groundbreaking model has the remarkable ability to predict a 3D model of an object from a single input image in a mere 5 seconds.

AI Machine Learning & Data Science Research

Google’s E3 TTS Provides Effortless Approach to High-Quality Audio Synthesis Through Diffusion Models

In a new paper E3 TTS: Easy End-to-End Diffusion-based Text to Speech, a Google research team proposes Easy End-to-End Diffusion-based Text to Speech. This streamlined and efficient text-to-speech model hinges solely on diffusion to preserve temporal structure, allowing it to accept plain text as input and generate audio waveforms directly.

AI Machine Learning & Data Science Research

Apple Repurposes Large Language Models for Reinforcement Learning challenges in Embodied AI

An Apple research team presents Large LAnguage model Reinforcement Learning Policy (LLaRP). LLaRP effectively repurposes LLMs for Reinforcement Learning (RL) challenges within the realm of Embodied Artificial Intelligence (AI), achieving a remarkable 1.7 times higher success rate compared to other established baselines and zero-shot LLM applications.

AI Machine Learning & Data Science Research

Microsoft Azure’s Idea2Img: Enabling Automatic Image Design and Generation with Enhanced Image Quality

A Microsoft Azure AI research team introduces “Idea2Img” in their paper, “Idea2Img: Iterative Self-Refinement with GPT-4V(ision) for Automatic Image Design and Generation.”, which leverages the capabilities of GPT-4V(ision) to revolutionize the process of automatic image design and generation with enhanced image quality.

AI Machine Learning & Data Science Research

Microsoft’s DeepSpeed-VisualChat: Breaking Boundaries in Multi-Modal Language Models

In a new paper DeepSpeed-VisualChat: Multi-Round Multi-Image Interleave Chat via Multi-Modal Causal Attention, a research team from DeepSpeed of Microsoft presents the DeepSpeed-VisualChat framework, which is designed to optimize LLMs by incorporating multi-modal capabilities, demonstrating superior scalability, even up to a 70 billion parameter model size.

AI Machine Learning & Data Science Research

NNAISENSE’s New Class of Generative Model: Bayesian Flow Networks Break Barriers in Handing Discrete Data

A NNAISENSE research team introduces a novel class of generative models known as Bayesian Flow Networks (BFNs). These BFNs combine the power of Bayesian inference with neural networks in an iterative modeling process, enabling successful application to continuous, discretized, and discrete data while maintaining competitive performance.

AI Machine Learning & Data Science Research

Standford U’s MAPTree: Redefining Decision Trees – Precision, Speed, and Efficiency Unleashed

In a new paper MAPTree: Beating “Optimal” Decision Trees with Bayesian Decision Trees, a Stanford University research team introduces MAPTree, an algorithm that confidently uncovers the maximum a posteriori tree within Bayesian Classification and Regression Trees (BCART) posterior, achieving strong performance with significantly leaner and faster trees.