Microsoft researchers have released technical details of an AI system that combines both approaches. The new Multi-Task Deep Neural Network (MT-DNN) is a natural language processing (NLP) model that outperforms Google BERT in nine of eleven benchmark NLP tasks.
In 2017 Google introduced Federated Learning (FL), “a specific category of distributed machine learning approaches which trains machine learning models using decentralized data residing on end devices such as mobile phones.” A new Google paper has now proposed a scalable production system for federated learning to enable increasing workload and output through the addition of resources such as compute, storage, bandwidth, etc.
The San Francisco-based AI non-profit however has raised eyebrows in the research community with its unusual decision to not release the language model’s code and training dataset. In a statement sent to Synced, OpenAI explained the choice was made to prevent malicious use: “it’s clear that the ability to generate synthetic text that is conditioned on specific subjects has the potential for significant abuse.”
Uber has unveiled Ludwig, a new TensorFlow-based toolkit that enables users to train and test deep learning models without writing any code. The toolkit will help non-experts understand models and accelerate their iterative development by simplifying the prototyping process and data processing.
In December Synced reported on a hyperrealistic face generator developed by US chip giant NVIDIA. The GAN-based model performs so well that most people can’t distinguish the faces it generates from real photos. This week NVIDIA announced that it is open-sourcing the nifty tool, which it has dubbed “StyleGAN”.
The internet loves those little looping action images we call GIFs. They can tell a short visual story in a small file size that has high portability. The visual quality of GIFs is however usually low compared to the videos they were sourced from. If you are sick of fuzzy, low resolution GIFs, then researchers from Stony Brook University, UCLA, and Megvii Research have just the thing for you: “the first learning-based method for enhancing the visual quality of GIFs in the wild.”
To make ML-based solutions available for a wider variety of deployment scenarios, Waymo’s autonomous driving team has collaborated with Google AI Brain Team researchers on a system that automates the creation of high quality and low latency neural networks on existing AutoML architectures.
A cooperative research group from Google, Stanford, and Johns Hopkins has proposed “Auto-DeepLab,” a new method which utilizes hierarchical Neural Architecture Search (NAS) for semantic image segmentation. The project team includes top AI researchers Director of the Stanford Vision Lab Fei-Fei Li; and UCLA Center for Cognition, Vision, and Learning Director Alan Yuille.
The proliferation of social media in our daily lives has profoundly changed the way we work and play with others. It has also created an entirely new job: thousands of people worldwide now work for Google, Facebook and Twitter “Community Operations Teams.” Whenever a user flags content as offensive, it’s sent to these guys for review.
Researchers using enhanced super-resolution technology are giving classic video games of the past incredible, texture-rich visual makeovers. The team has released ‘remastered’ versions of Return to Castle Wolfenstein, Doom, The Elder Scrolls III: Morrowind, and most recently — a visually enhanced version of 2001 third-person shooter game Max Payne.
The world’s largest and most exciting technology show, CES officially kicks off this morning in Las Vegas, USA. With an abundance of products to present, tech giants such as Nvidia, Qualcomm, Samsung, Intel started revving up their engines two days ago, hosting press conferences to showcase their “New Year’s Resolutions.”
Researchers at the Korea Advanced Institute of Science and Technology and Pohang University of Science and Technology have introduced a machine learning algorithm system, InstaGAN, which can perform multiple instance-aware image-to-image translation tasks — such as replacing sheep in photos with giraffes — on multiple image datasets.
Tsinghua Natural Language Processing Group (THUNLP) has published a great reading list for any budding AI researchers whose New Year’s resolution is to study machine translation. The list compiles the most influential machine translation papers from the past 30 years, spotlighting the 10 most important contributions to the development of machine translation.
Alibaba Cloud recently announced that it has open sourced Mars — its tensor-based framework for large-scale data computation — on Github. Mars can be regarded as “a parallel and distributed NumPy.” Mars can tile a large tensor into small chunks and describe the inner computation with a directed graph, enabling the running of parallel computation on a wide range of distributed environments, from a single machine to a cluster comprising thousands of machines.
Reinforcement learning (RL) has been making spectacular achievements, e.g., Atari games, AlphaGo, AlphaGo Zero, AlphaZero, DeepStack, Libratus, OpenAI Five, Dactyl, DeepMimic, Catch The Flag, learning to dress, data center cooling, chemical syntheses, drug design, etc. See more RL applications.
Text-based CAPTCHA remain one of the most visible and commonly used mechanisms for website security. As a sort of online gatekeeper that distinguishes between humans and bots, the little solvable image fields have critical commercial applications in blocking automatic spam and preventing e-transfer fraud; and can also stop bots from spreading fraudulent information, etc.
Synthesizing peptides — the chains of amino acids that conduct various functions within cells — has long been a research area of interest for scientists and engineers. There has however been little success thus far, as existing methods for synthesizing peptides have been prohibitively expensive and time-consuming.
Insilico Medicine, a drug discovery startup located in Johns Hopkins University, has introduced MOSES (Molecular Sets), a platform which can be used to compare model accuracy in molecular generation. MOSES provides a standardized benchmarking dataset, a set of open-sourced models with unified implementation, and evaluation metrics.
The digital painting tool GANpaint has gone viral on social media. The product of a team of high-profile researchers from MIT, IBM, Google, and the Chinese University of Hong Kong, GAPpaint allows anyone — even those with little knowledge of digital painting or photoshop — to “paint” incredibly complex and detailed photorealistic scenes.
In 2016 Google’s DeepMind stunned the world when their Go computer AlphaGo secured a historic victory over Korean grandmaster Lee Sedol. Yesterday the UK’s top AI team delivered their latest “wow moment” as their AI system AlphaFold topped the Critical Assessment of Structure Prediction (CASP) competition.