In an unprecedented “call for collaboration,” a group of 22 respected AI experts that includes Andrew Ng, Yoshua Bengio, and Demis Hassabis have published a paper exploring how machine learning (ML) could help deal with climate change by reducing greenhouse gases (GHG).
Many cat owners have awoken to the sound of frenzied scrambling across the floor as their furry feline friend chases the mouse it snuck into the house in the middle of the night. Or looked up to find their pet staring down with those adorable big round eyes at the dead little fledgling it has deposited beside your pillow as a “present.“
In a joint effort with the Perimeter Institute for Theoretical Physics and Alphabet (Google) X, Google AI researchers recently announced a new open source library, TensorNetwork, which can greatly improve the efficiency of tensor calculations for tensor network algorithms.
A group of researchers from Tencent Technology, the Chinese University of Hong Kong, and Nankai University recently combined two commonly used techniques — Batch Normalization (BatchNorm) and Dropout — into an Independent Component (IC) layer inserted before each weight layer to make inputs more independent*.
Berkeley Artificial Intelligence Research (BAIR) has introduced a new reinforcement learning (RL) method, Stochastic Optimal Control with Latent Representations (SOLAR), which can help robots quickly learn tasks such as stacking blocks or pushing objects from visual inputs.
Automated machine learning (AutoML) is a hot topic in artificial intelligence. Researchers from German digital and software company USU Software AG and the University of Stuttgart recently published a review paper summarizing the latest academic and industrial developments in AutoML.
Designing accurate and efficient CNNs for mobile devices is challenging due to the large design space and expensive computational methods. Although many mobile CNNs are available for developers to train and deploy to mobile devices, existing CNN architecture may not be able to achieve the best results for some tasks on mobile devices.
A group of Google researchers led by Quoc Le — the AI expert behind Google Neural Machine Translation and AutoML — have published a paper proposing attention augmentation. In experiment results, the novel two-dimensional relative self-attention mechanismfor image classification delivers “consistent improvements in image classification.”
With its improved productivity and accuracy and more personalized experience, AI is revolutionizing medical imaging. According to Signify Research, the world market for AI in medical imaging — comprising software for automated detection, quantification, decision support, and diagnosis — will reach US$2 billion by 2023.
Advanced machine learning techniques and the widespread deployment of surveillance cameras have dramatically improved the efficiency and accuracy of human detection systems in airports, train stations, and other sensitive public places. Is this the end of anonymity?
Thanks to the CUDA architecture  developed by NVIDIA, developers can exploit GPUs’ parallel computing power to perform general computation without extra efforts. Our objective is to evaluate the performance achieved by TensorFlow, PyTorch, and MXNet on Titan RTX.
Researchers from Facebook, the National University of Singapore, and the Qihoo 360 AI Institute have jointly proposed OctConv (Octave Convolution), a promising new alternative to traditional convolution operations. Akin to a “compressor” for Convolutional Neural Networks (CNN), the OctConv method saves computational resources while boosting effectiveness.