BERTLang Helps Researchers Choose Between BERT Models
Researchers from Bocconi University have prepared an online overview of the commonalities and differences between language-specific BERT models and mBERT.
AI Technology & Industry Review
Researchers from Bocconi University have prepared an online overview of the commonalities and differences between language-specific BERT models and mBERT.
A new study suggests that VSR models could perform even better if they used additional available visual information.
The earliest evidence of China’s recorded history is found in the Shang dynasty (~1600 to 1046 BC), and this hasContinue Reading
The model outperforms existing methods in image manipulation and offers researchers a possible solution to the scarcity of paired datasets.
Researchers proposed a new training scheme that targets this bias by controlling and exposing textural information slowly through the training process.
Researchers from the Berkeley Artificial Intelligence Research (BAIR) Lab at UC Berkeley explored the effect of Transformer model size on training and inference efficiency.
Researchers proposed an automatic structured pruning framework, AutoCompress, which adopts the 2018 ADMM-based weight pruning algorithm and outperforms previous automatic model compression methods while maintaining high accuracy.
A new study leverages an established AI-based drug discovery pipeline to produce molecular structures as part of the widening fight against the 2019-nCoV outbreak.
Researchers propose a flexible GNN benchmarking framework that can also accommodate the needs of researchers to add new datasets and models.
UC Berkeley and Adobe Research have introduced a “universal” detector that can distinguish real images from generated images regardless of what architectures and/or datasets were used for training.
Proposed by researchers from the Rutgers University and Samsung AI Center in the UK, CookGAN uses an attention-based ingredients-image association model to condition a generative neural network tasked with synthesizing meal images.
The KaoKore dataset includes 5552 RGB image files drawn from the 2018 Collection of Facial Expressions dataset of cropped face images from Japanese artworks.
Researchers propose a novel model compression approach to effectively compress BERT by progressive module replacing.
The crowdsourcing produced 111.25 hours of video from 54 non-expert demonstrators to build “one of the largest, richest, and most diverse robot manipulation datasets ever collected using human creativity and dexterity.”
A Google-led research team has introduced a new method for optimizing neural network parameters that is faster than all common first-order methods on complex problems.
Fast and accurate diagnosis is critical on the front line, and now an AI-powered diagnostic assessment system is helping Hubei medical teams do just that.
In an attempt to equip the TF-IDF-based retriever with a state-of-the-art neural reading comprehension model, researchers introduced a new graph-based recurrent retrieval approach.
The proposed system is capable of searching the continental United States at 1 -meter pixel resolution, corresponding to approximately 2 billion images, in around 0.1 seconds.
MonoLayout, a practical deep neural architecture that takes just a single image of a road scene as input and outputs an amodal scene layout in bird’s-eye view.
In a bid to raise awareness of the threats posed by climate change, the Mila team recently published a paper that uses GANs to generate images of how climate events may impact our environments — with a particular focus on floods.
Researchers have proposed a novel self-adversarial learning (SAL) paradigm for improving GANs’ performance in text generation.
Bayesian inference meanwhile leverages Bayes’ theorem to update the probability of a hypothesis as additional data becomes available. How can Bayesian inference benefit deep learning models?
DeepMind announced yesterday the release of Haiku and RLax — new JAX libraries designed for neural networks and reinforcement learning respectively.
Researchers from Italy’s University of Pisa present a clear and engaging tutorial on the main concepts and building blocks involved in neural architectures for graphs.
Researchers have proposed a novel generator network specialized on the illustrations in children’s books.
Researchers have proposed a simple but powerful “SimCLR” framework for contrastive learning of visual representations.
A recent Google Brain paper looks into Google’s hugely successful transformer network — BERT — and how it represents linguistic information internally.
The tool enables researchers to try, compare, and evaluate models to decide which work best on their datasets or for their research purposes.
Google teamed up with researchers from Synthesis AI and Columbia University to introduce a deep learning approach called ClearGrasp as a first step to teaching machines how to “see” transparent materials.
Researchers from Google Brain and Carnegie Mellon University have released models trained with a semi-supervised learning method called “Noisy Student” that achieve 88.4 percent top-1 accuracy on ImageNet.
Researchers have introduced the first unsupervised learning approach for identifying interpretable semantic directions in the latent space of generative adversarial network (GAN) models.
Deep learning models are getting larger and larger to meet the demand for better and better performance. Meanwhile, the timeContinue Reading
Researchers introduced semantic region-adaptive normalization (SEAN), a simple but effective building block for conditional Generative Adversarial Networks (cGAN).
The Godfathers of AI and 2018 ACM Turing Award winners Geoffrey Hinton, Yann LeCun, and Yoshua Bengio shared a stage in New York on Sunday night at an event organized by the Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI 2020).
The crucial step now is to develop matching vaccines and drugs to uproot its existence, and China’s big tech companies have stepped up to help.
Batchboost is a simple technique to accelerate ML model training by adaptively feeding mini-batches with artificial samples which are created by mixing two examples from the previous step – in favor of pairing those that produce the difficult one.
Leading scientific publication Nature announced it is launching a trial starting this week that will give authors of newly published papers the option of appending contents of the discussions they’ve had with and reports they’ve received from their reviewers.
In an effort to enrich resources for multispeaker singing-voice synthesis, a team of researchers from the University of Tokyo has developed a Japanese multispeaker singing-voice corpus.
Researchers proposed a “radioactive data” technique for subtly marking images in a dataset to help researchers later determine whether they were used to train a particular model.
In a new paper, researchers from the University of Toronto, Vector Institute, and University of Wisconsin-Madison propose SISA training, a new framework that helps models “unlearn” information by reducing number of updates that need to be computed when data points are removed.