The Eyes Have It: 20 Million Images Make TEyeD World’s Largest Human Eye Dataset
A new study by University Tübingen introduces the world’s largest unified eye dataset with over 20 million human eye images captured using head-mounted eye trackers.
AI Technology & Industry Review
A new study by University Tübingen introduces the world’s largest unified eye dataset with over 20 million human eye images captured using head-mounted eye trackers.
Researchers from Drexel University, Northeastern University and IT University Copenhagen explore how humans interact with AI in such contexts, with a focus on computer games.
In a new paper, researchers from Max Planck Institute for Informatics and Facebook Reality Labs propose an end-to-end trainable method that enables re-rendering of humans from one single image.
Recent AI research on speech separation has explored ways to associate lip motions in videos with audio, but this approach suffers when speakers’ lips are occluded, which they often are in busy multi-speaker environments.
VOGUE, an AI-powered optimization method that deforms garments according to a given body shape while preserving pattern and material details to deliver state-of-the-art photorealistic, high-resolution try-on images.
Researchers propose a novel framework and hierarchical predictive model that learns to identify what is predictable from unlabelled video.
The novel approach tackles dynamic 3D human-body synthesis from a sparse set of camera views, bettering existing methods on key metrics by significant margins.
Researchers proposed PGDrive, a driving simulator designed to evaluate and improve end-to-end driving agents’ generalization abilities.
Synced has invited Mr. Sheldon Fernandez to share his insights about the current development and future trends of artificial intelligence.
As part of our year-end series, Synced highlights 10 AI-powered art projects that inspired and entertained us in 2020.
Researchers combine the effectiveness of the inductive bias in CNNs with the expressivity of transformers to model and synthesize high resolution images.
In the new paper Canonical Capsules: Unsupervised Capsules in Canonical Pose, Turing Award Honoree Dr. Geoffrey Hinton and a team of researchers propose an architecture for unsupervised learning with 3D point clouds based on capsules.
University of Notre Dame and Facebook AI research propose Img2pose, real-time 6DoF 3D face pose estimation without face detection or landmark localization.
“Depix” is a new AI-powered tool that can easily undo pixelization to enable recovery of the information therein.
CSAIL researchers propose a framework for image reconstruction tasks using the state-of-the-art generative model StyleGAN2.
The new AI-powered Multi-Ingredient Pizza Generator (MPG) can deliver all these mouth-watering pies and many more.
NVIDIA blog introduced company’s latest NeurIPS presentation: applying a novel neural network training technique, adaptive discriminator augmentation, to the popular NVIDIA StyleGAN2 model.
The approach dramatically reduces bandwidth requirements by sending only a keypoint representation [of faces] and reconstructing the source video on the receiver side with the help of generative adversarial networks (GANs) to synthesize the talking heads.
A Princeton student designed a GAN framework for Chinese landscape painting generation that is so effective most humans can’t distinguish its works from the real thing.
Facebook’s new model enables free-viewpoint rendering of dynamic scenes in a single video.
Researchers from the City University of Hong Kong and SenseTime propose a lightweight matting objective decomposition network (MODNet) that can smoothly process real-time human matting from a single input image with diverse and dynamic backgrounds.
University of Alberta recently proposed U^2-Net, a novel deep network architecture that achieves very competitive performance in salient object detection.
A new study by South China University of Technology and Tencent WeChat AI researchers is the latest fruitful attempt to utilize transformer architectures in object detection.
A new AI-powered image synthesis framework makes “learning” to moonwalk or drop Blackpink dance moves a snap.
Stylized Neural Painter is a novel automatic image-to-painting translation method that generates vivid and realistic artworks in controllable styles
“Our research provides enriched AR user experiences by enabling a more fine-grained visual recognition feature in AR, which is desirable in a wide range of application scenarios including technical support,” IBM researchers say.
Google AI yesterday released its Objectron dataset — a collection of short, object-centric video clips capturing a large set of common objects from different angles.
Researchers introduced a modular primitive that uses existing, highly optimized hardware graphics pipelines to deliver performance superior to previous differentiable rendering systems.
Now, just in time for costume season, another indie developer has taken facial image transfer tech to the opposite end of the cuteness spectrum, building a zombie generator.
ICLR 2021 submission proposes LambdaNetworks, a transformer-specific method that reduces costs of modeling long-range interactions for CV and other applications.
Google AI recently launched the open-source browser-based toolset “rǝ,” which was created to enable the exploration of city transitions from 1800 to 2000 virtually in a three-dimensional view.
Pinkney and Adler NeurIPS 2020 workshop paper enables realistic image generation in domains such as animation and ukiyo-e with creative control on the output.
ICLR 2021 submitted paper proposes efficient VAEs that outperform PixelCNN-based autoregressive models in log-likelihood on natural image benchmarks.
ICLR 2021 paper An Image Is Worth 16×16 Words: Transformers for Image Recognition at Scale suggests Transformers can outperform top CNNs on CV at scale.
NeurIPS 2020 released its list of accepted papers this week with Google, Stanford, and MIT as the top affiliations.
Google AI researchers developed a sign language detection model for video conferencing applications that can perform real-time identification of a person signing as an active speaker.
Google AI has announced a new audiovisual speech enhancement feature in YouTube Stories (iOS) that enables creators to make better selfie videos by automatically enhancing their voices and reducing noise.
Novel model uses a quality estimator and evolutionary optimization to search the latent space of GANs trained on limited datasets.
Imaginaire, a universal PyTorch library designed for various GAN-based tasks and methods.
Chinese researchers propose a novel regression framework in pursuit of “fast, accurate and stable 3D dense face alignment simultaneously.”







































