NVIDIA Research is set to present more than 50 papers at the Computer Vision and Pattern Recognition (CVPR) conference in Seattle, from June 17-21, 2024, highlighting significant advancements in visual generative AI. The research covers potential applications across creative industries, autonomous vehicle development, healthcare, and robotics, according to NVIDIA Blog.
Generative AI for Diverse Applications
Among the notable projects, two papers focusing on the training dynamics of diffusion models and high-definition maps for autonomous vehicles are finalists for CVPR’s Best Paper Awards. NVIDIA also secured the CVPR Autonomous Grand Challenge’s End-to-End Driving at Scale track, showcasing comprehensive self-driving models that outperformed over 450 entries globally, earning the CVPR Innovation Award.
NVIDIA’s research includes a text-to-image model easily customizable for specific objects or characters, a new model for object pose estimation, techniques to edit neural radiance fields (NeRFs), and a visual language model capable of understanding memes. These innovations aim to empower creators, accelerate autonomous robot training, and assist healthcare professionals in processing radiology reports.
“Artificial intelligence, and generative AI in particular, represents a pivotal technological advancement,” said Jan Kautz, vice president of learning and perception research at NVIDIA. “At CVPR, NVIDIA Research is sharing how we’re pushing the boundaries of what’s possible — from powerful image generation models that could supercharge professional creators to autonomous driving software that could help enable next-generation self-driving cars.”
JeDi: Simplifying Custom Image Generation
One of the standout papers, JeDi, proposes a technique allowing users to personalize diffusion model outputs using reference images within seconds, outperforming existing fine-tuning methods. This innovation, developed in collaboration with Johns Hopkins University, Toyota Technological Institute at Chicago, and NVIDIA, could benefit creators needing specific character depictions or product visuals.
FoundationPose and NeRFDeformer
FoundationPose, another research highlight, is a foundation model for object pose estimation and tracking. It can be applied to new objects without fine-tuning, using reference images or 3D representations to track objects in 3D across videos, even in challenging conditions. This model could enhance industrial applications and augmented reality.
NeRFDeformer, developed with the University of Illinois Urbana-Champaign, simplifies transforming NeRFs with a single RGB-D image, streamlining the process of updating 3D scenes captured as 2D images.
VILA: Advancing Visual Language Models
In collaboration with the Massachusetts Institute of Technology, NVIDIA introduced VILA, a family of visual language models that outperforms prior models in answering questions about images. VILA’s pretraining process enhances world knowledge, in-context learning, and reasoning across multiple images, making it a powerful tool for various applications.
Generative AI in Autonomous Driving and Smart Cities
NVIDIA’s contributions to autonomous vehicle research at CVPR include a dozen papers focusing on this area. Additionally, NVIDIA provided the largest-ever indoor synthetic dataset to the AI City Challenge, aiding the development of smart city solutions and industrial automation. These datasets were generated using NVIDIA Omniverse, a platform enabling developers to build Universal Scene Description (OpenUSD)-based applications and workflows.
NVIDIA Research, with hundreds of scientists and engineers worldwide, continues to push the boundaries in AI, computer graphics, computer vision, self-driving cars, and robotics. Learn more about their groundbreaking work at CVPR 2024 on the NVIDIA Blog.
Image source: Shutterstock
Credit: Source link