Monday, December 23, 2024

Helm.ai Announces DNN Foundation Models for Intent Prediction and Path Planning

Related stories

Doc.com Expands AI developments to Revolutionize Healthcare Access

Doc.com, a pioneering healthcare technology company, proudly announces the development...

Amesite Announces AI-Powered NurseMagic™ Growth in Marketing Reach to Key Markets

Amesite Inc., creator of the AI-powered NurseMagic™ app, announces...

Quantiphi Joins AWS Generative AI Partner Innovation Alliance

Quantiphi, an AI-first digital engineering company, has been named...
spot_imgspot_img

Helm.ai, provider of next-generation AI software for autonomous driving and automation of robotics, announced DNN (Deep Neural Network)-based foundation models for behavioral prediction and decision-making as part of the company’s AI software stack for high-end ADAS L2/L3 and L4 autonomous driving.

The company has trained DNN foundation models to make predictions about the behavior of vehicles and pedestrians in complex urban scenarios, as well as to predict the path an autonomous vehicle would take in those situations, which are critical ingredients of the decision-making capabilities for self-driving cars. Helm.ai leveraged its industry-validated surround view full scene semantic segmentation and 3D detection system as the core representation to enable training intent prediction and path planning capabilities. Additionally, the foundation models are trained using the company’s proprietary Deep Teaching technology to achieve broad predictive capability in a scalable way.

Helm.ai’s technology learns directly from real driving data and uses the company’s highly accurate and temporally stable perception system to capture information about complex behaviors of vehicles and pedestrians and the surrounding driving environment, leading to DNNs that automatically learn subtle yet important aspects of urban driving. The foundation models powering Helm.ai’s intent and path prediction gather input from a series of observed images and generate predicted video sequences that represent the most likely possible outcomes of what happens next. The models also provide a predicted path for the autonomous vehicle that is consistent with the intent prediction. Both the intent prediction and path prediction capabilities are essential for planning the safest optimal action by the autonomous vehicle.

Also Read: Codenotary Introduces Trustcenter 4.0 with New Machine Learning Guided Search Engine

Importantly, the Helm DNN foundation models for intent prediction and path planning are trained in the highly scalable Deep Teaching paradigm, enabling unsupervised learning about complex urban driving scenarios directly from real driving data. This approach circumvents cumbersome physics-based simulators and hand-coded rules, which are insufficient to capture the full complexity of driving in the real world. In particular, the Helm.ai development and validation pipeline, while optimized for high end ADAS L2/L3 mass production software, can also be directly applied to L4 fully autonomous applications. Moreover, the Helm.ai scalable AI approach readily generalizes to robotics domains beyond self-driving vehicles.

Helm.ai is building an AI-first approach to autonomous driving that is designed to seamlessly scale from high-end ADAS L2/L3 mass production programs all the way to large scale L4 deployments. The company’s software-only platform is hardware-agnostic and vision-first, addressing the critical perception problem for vision yet also incorporating sensor fusion between vision and radar/lidar as needed. The technology advancements announced today accelerate the value of Helm.ai’s software offering by paving the way for scalable development and validation of AI-based intent prediction and path planning software for autonomous vehicles.

SOURCE: PRNewswire

Subscribe

- Never miss a story with notifications


    Latest stories

    spot_img