Capra Robotics Lands 116M Boost As Autonomous Mobile Robotics Pioneer Expands Globally
Capra Robotics Seals $11.6M Funding to Accelerate Global Expansion and Robust Partnership Network …
23. December 2024
Autonomous vehicle developers are on the cusp of a revolution in their quest for more accurate and diverse data to improve autonomous driving systems. Helm.ai has unveiled GenSim-2, its cutting-edge generative AI model designed to enrich autonomous driving datasets with advanced video editing capabilities.
GenSim-2 boasts dynamic weather and illumination adjustments, object appearance modifications, and consistent multi-camera support. These features empower automakers to generate diverse, highly realistic video data tailored to specific requirements, a significant leap forward in the development of robust autonomous driving systems.
Built on Helm.ai’s proprietary Deep Teaching methodology and deep neural networks, GenSim-2 expands on the capabilities of its predecessor, GenSim-1. By leveraging generative AI, development teams can modify weather conditions such as rain, fog, snow, glare, and time of day in video data, creating a more realistic environment for autonomous vehicles to learn from.
In addition to enhancing realism, GenSim-2 enables customization and adjustments of object appearances, including road surfaces, vehicles, pedestrians, buildings, vegetation, and other road objects. These transformations can be applied consistently across multi-camera perspectives, ensuring self-consistency throughout the dataset.
“The ability to manipulate video data at this level of control and realism is a game-changer for autonomous driving development,” said Vladislav Voroninski, Helm.ai’s CEO and founder. “GenSim-2 equips automakers with unparalleled tools for generating high-fidelity labeled data for training and validation, bridging the gap between simulation and real-world conditions to accelerate development timelines and reduce costs.”
This technology addresses industry challenges by offering an alternative to resource-intensive traditional data collection methods. By generating scenario-specific video data, GenSim-2 supports a wide range of applications in autonomous driving, from developing and validating software across diverse geographies to resolving rare and challenging corner cases.
GenSim-2 aligns with Helm.ai’s broader portfolio of generative AI-based products, including VidGen-2 and WorldGen-1. VidGen-2 generates predictive video sequences with realistic appearances and dynamic scene modeling, offering double the resolution of its predecessor and improved realism at 30 frames per second. WorldGen-1 is a generative AI foundation model that can simulate the entire autonomous vehicle stack, generating driving scenes across multiple sensor modalities and perspectives.
With GenSim-2, Helm.ai aims to revolutionize the autonomous driving industry by providing automakers with unparalleled tools for generating high-fidelity labeled data. By bridging the gap between simulation and real-world conditions, the company seeks to accelerate development timelines and reduce costs, ultimately paving the way for safer and more efficient autonomous vehicles on our roads.