Unified system architecture supporting
multiple automation levels and different hardware architectures

Utilizing model pruning, modular approach and a dual-track approach to flexibly adapt algorithm models
and ensure smooth transition between L4 and L2.

Learn more
  • Unified system architecture supporting multiple automation levels

    In software development, we leverage the technology stack, development platform and Tool chains built for our L2-L2+ solutions to significantly enhance R&D efficiency of L4 solutions while reducing development costs. AI models for both L2-L2+ and L4 solutions are built upon a unified foundational model, which is optimized and customized based on application scenarios and performance requirements through technologies such as model pruning, quantization and distillation to ensure consistency and efficiency across technology development.

    We have accumulated abundant AI technology and engineering experiences through extensive deliveries for L2-L2+ mass-production projects. The reliability and stability of our L2-L2+ solutions have been fully validated through mass-production, which provides a replicable engineering foundation for the development and deployment of our L4 systems.

    On the hardware side, by reusing sensors and domain controllers adopted in our L2-L2+ solutions, we not only ensure the reliability of hardware, but also reduce costs of our L4 solutions leveraging cost advantages of mass-production supply chain. In addition, our experience in algorithm optimization at the L2-L2+ level enables us to optimize sensor layout and chip selection for our L4 solutions. For instance, the sensor layout of our Robotaxi solution, ConnectOne, adopts fewer LiDARs and primarily utilizes Horizon Robotics’ Journey 6 processing hardware with mid-range computing power, which significantly improves the cost-efficiency of our L4 solutions.

    As we scale up the deployment of our L4 solutions, we can also accumulate a massive amount of high-quality data in real-world operating scenarios. Such high-quality data from L4 vehicles can be fed back to optimize our models for L2-L2+ solutions. After anonymization and cleansing, the ground-truth data eliminate biases inherent in human driving behaviors, which can accelerate our product iteration cycle and reduce data collection costs compared with traditional methods — data collection by human drivers.

  • Unified system architecture supporting different hardware architectures

    Leveraging our strong software-and-hardware co-development capabilities, we have developed a unified heterogeneous computing architecture adaptable to multiple SoC platforms to enable deep integration of AI algorithm models with various hardware (such as computing chips and sensors). Our system architecture adopts a unified platform interface and data protocols, and supports the rapid deployment of intelligent driving systems across different SoC platforms through an automated model conversion tool. We also unify and standardize external interfaces through a layered and decoupled system architecture. This allows us to effectively break the barriers between different hardware to ensure consistent, high performance of our systems across SoC platforms, enabling generalization and reusability of software and supporting rapid cross-platform deployment and validation of intelligent driving systems.

Advanced AI large model
capabilities delivering high-performance solutions

Learn more
Advantages of VLA/VLM

We have built up a VLA/VLM architecture through deep integration of end-to-end methodologies and large language models that enables advanced object recognition, accurate prediction of traffic participants’ behavioral intentions and comprehensive environment understanding. Our natural-language-driven trajectory generator strengthens the explainability of our AI systems, thereby enhancing trust from both users and regulators. Our systems achieve exceptional robustness leveraging multimodal fusion and redundant safety mechanism. Our system architecture can compensate for limitations of standalone end-to-end system in extreme conditions, reducing false detections, and optimizing our ability to address long-tail issues through an AI model with high generalization ability. We adopt the idea of developing light-weight AI models with enhanced reasoning efficiency. We compress model parameters by applying model pruning, quantization and distillation technologies. Based on our deep understanding of various SoC platforms, we have also implemented a multi-core parallel scheduling mechanism that maximizes the utilization of heterogeneous computing power under a CPU-NPU collaborative SoC architecture.

Evolving from multi-module collaboration toward
end-to-end large models for development and iterations

end-to-end · map-free · spot-to-spot

Self-developed, advanced end-to-end perception algorithms, combined with multi-sensor fusion and centimeter-level planning and control, and an automated data closed-loop, ensures the continuous evolution of autonomous driving capabilities for enhanced precision, stability, and intelligence.
CalmDrive 1.0

CalmDrive 1.0 forms a series of stable and reliable intelligent driving solutions by integrating multiple AI perception algorithm models, post-perception fusion models, and rule-based planning modules and control modules. It features clear functional modules and strong explainability, but requires rule design and fallback strategies for various scenarios. The interdependencies between different modules are complex, and the system’s ability to handle long-tail corner cases is limited. This was the mainstream technical approach at the early stage of the intelligent driving industry.

CalmDrive 2.0

CalmDrive 2.0 adopts a BEV-based pre-fusion perception algorithm to integrate visual and point-cloud data from different sensors, and combines rule-based planning strategy and model-based planner into one planning module to address cross-modal spatial matching challenges. The system enhances temporal modeling capability, reduces missed and false detections, and significantly improves adaptability and reliability under complex scenarios. It has become a mainstream technology architecture for mass-production projects in recent years.

CalmDrive 2.5

CalmDrive 2.5 adopts a two-stage end-to-end architecture composed of a perception module and a decision/planning module. The system leverages large models to power the perception module for precise environmental understanding, outputs the perception results to model-based planning module, and outputs control commands. It ensures seamless connections between perception, decision-making and planning, substantially simplifies system complexity, optimizes information flow, minimizes information loss between models, and improves overall system efficiency and accuracy.

CalmDrive 3.0

CalmDrive 3.0 is a one-stage end-to-end system integrating perception and planning. Leveraging a large model, the system starts from processing inputs of sensor data, and outputs control commands. This architecture further simplifies system structure, enhances efficiency of model execution and generalization capability, and improves system integration and operationality. In addition, we have further enhanced the system’s explainability through VLM/VLA architecture.

AI Planner

Multimodal fusion drives proactive planning and decision-making for human-like, scenario-adaptive driving.

Learn more

Smarter Decisions, Smoother Planning, Stronger Generalization

The system mimics the behavior of experienced human drivers leveraging reinforcement learning(RL) and contrastive imitation learning(CIL), delivering driving behaviors that are safe, smooth and efficient.The powerful decision model architecture boosts the autonomous driving system's cognitive intelligence and enhances performance in complex scenarios.

Comprehensiveness and temporal model
  • Multimodal Fusion: Fuse multi source data, provide 360-degree around-view perception.
  • Centimeter-Level Accuracy: Leverage localization and HD maps to enable precise matching.
  • Full-Range Driving Path Guidance: Provide a foundation for efficient and comfortable driving intent.
Temporal model and Foresight Capability
  • Dynamic Scenarios Understanding: Online modeling of continuous time-series data to interpret the dynamic evolution of scenes.
  • Prediction and Reasoning: Accurately predict the trajectories of dynamic objects and infer their behavioral intentions.
  • Coherence and Safety: Generate smooth, stable, and foresighted planning decisions to avoid jittery and short-sighted behaviors.
Multimodal Candidate Trajectory Generation
  • Multiple Reference Line Input: Process multiple reference lines simultaneously to comprehend diverse paths and lane options.
  • 12 Candidate Trajectories: Generate multiple driving strategies at once, including efficiency-oriented, safety-conservative, and balanced-robust behaviors.

Self-developed One-stop AI development platform CalmVolution
enables high R&D efficiency and superior adaptability to SoC platforms

Learn more
Integration of Multimodal AI Large Models

We have integrated multimodal AI large models into CalmVolution, which helps us to conduct analysis, breakdown and review of customer requirements during the R&D and delivery process, and to accelerate code generation and make code quality assessment. This approach also helps us sort out the traceability among requirements, development, and verification and validation, which further improves our systematic error-correction capability.

DataTurbo

Our data closed-loop subsystem, DataTurbo, was built on the philosophy of“full chain data closed-loop.”It has established a complete chain from data generation to value realization through deep coordination among three module—data collection, data management and data upload. It supports efficient processing of data—a core element of intelligent driving algorithm iteration and functional optimization.

CalmForge

Our model training and development platform, CalmForge, enables us to manage AI algorithm models at different stages throughout their lifecycle, including training, fine-tuning, evaluation and monitoring. With our optimization of training methodology, CalmForge enables independent training and rapid iteration of multiple models in parallel, which improves utilization of our AI training servers and reduces model training period by more than 30% compared to conventional approaches. In addition, the model conversion tool embedded in CalmForge automatically converts our AI models to adapt to various SoC platforms, which enhances our cross-platform compatibility while reducing R&D expenses.

CalmVergence

Our automated system integration platform, CalmVergence, provides essential support to the rapid deployment of our intelligent driving systems by seamlessly connecting algorithms, software and hardware. Our R&D team can easily integrate our AI algorithm models and software from our software library with various vehicle interfaces simply by ticking checkboxes, which enables us to maximize reuse of our R&D outputs across projects. The reuse rate of our software and AI algorithm models in different projects exceeds 70%. We are able to complete deployment of a full set of systems on a new SoC platform within one month for verification and validation purpose, while it takes less than six months to complete the full cycle from deployment, verification and validation, to the final delivery of a mass-production project for a new vehicle model on a previously-deployed SoC platform.

Real-time, high-precision localization
Cross-floor parking from slot-to-slot

Learn more

CalmCar SLAM System

  • Integrated Neural Network Architecture for Multi-dimensional, Multi-layered Map Information Extraction

    The neural network concurrently generates multi-dimensional data, including sparse feature point-cloud data, sparse semantic targets, and dense semantic targets. This integrated spatial map encompasses both ground and above-ground elements, addressing the issues of high computational redundancy and limited information density commonly found in traditional map extraction techniques.

  • Highly Robust Scene Relocalization Mechanism

    By combining rich semantic data with sparse point cloud information, the system achieves accurate scene matching and localization. It effectively overcomes challenges such as dynamic lighting changes, low-texture environments, and repetitive scene textures, enabling reliable initial localization within parking structures at various levels of a memory map.

  • Bidirectional Cruise Localization Hardware and Software Architecture

    Featuring a multi-camera redundancy design, the system overcomes localization challenges in obstructed environments. It supports single-pass mapping and bidirectional cruise localization, providing precise and efficient solutions for various parking and retrieval scenarios.

Convolutional Neural Network-Based Feature Extraction and Matching

神经网络特征示例

跨层泊车

Other Utility Toolchain Products Integrated
into the AI Development Platform

To maximize the efficiency and performance of our AI development platform, we have developed a series of specialized toolchain products.
These tools further enhance the platform’s advantages in data processing, model training, and post-delivery support. Set forth below are selected examples of the toolchain products.

Fast4D Automatic Data Labelling Tool

SuperPangoo Issue Data Management Platform