- Manufacturing
XPeng Quantum-Enhanced Perception for Autonomous Driving
XPeng Motors
XPeng Motors researched quantum ML for sensor fusion in autonomous driving, combining LiDAR, camera, and radar data using quantum convolutional neural networks and quantum kernel methods to explore efficiency improvements for 3D object detection.
- Key Outcome
- QCNN achieved 78% mAP on synthetic LiDAR object detection (vs 82% classical PointNet) with 40% fewer parameters; identified quantum advantage pathway for edge deployment on low-power autonomous driving chips.
The Problem
Autonomous driving perception requires fusing data from multiple sensor modalities (LiDAR point clouds, RGB cameras, and millimeter-wave radar) into a coherent 3D model of the environment. Classical deep learning approaches (PointNet, VoxelNet, BEVFusion) achieve strong accuracy but are computationally expensive, requiring dedicated AI accelerator chips that consume significant power. For XPeng’s XNGP full-stack autonomous driving system, edge deployment on the vehicle’s onboard computing unit imposes hard constraints on memory bandwidth and power draw.
XPeng’s AI research group explored whether quantum convolutional neural networks (QCNNs) and quantum kernel methods offer a parameter efficiency advantage: achieving comparable accuracy with fewer trainable parameters, which translates to smaller model footprints for edge chips. The research was conducted primarily in simulation, with selected circuits validated on IonQ Harmony hardware to characterize noise characteristics.
Quantum Feature Maps for Point Cloud Data
LiDAR point clouds are unordered sets of 3D coordinates. The first stage of the quantum pipeline converts local point neighborhoods (16-point patches) into fixed-size feature vectors via classical PointNet-style set abstraction, then encodes these 8-dimensional features into 3-qubit quantum states using angle encoding. A quantum feature map applies a sequence of Ry rotations and CNOT entangling layers, projecting the input into a high-dimensional Hilbert space for subsequent kernel or QCNN processing.
import pennylane as qml
import torch
import torch.nn as nn
import numpy as np
N_QUBITS = 6
dev = qml.device("default.qubit", wires=N_QUBITS)
def angle_encode(features, wires):
"""Encode normalized feature vector via Ry rotations."""
for i, wire in enumerate(wires):
qml.RY(features[i] * np.pi, wires=wire)
def entangling_layer(wires):
"""Ring of CNOT gates for entanglement."""
for i in range(len(wires)):
qml.CNOT(wires=[wires[i], wires[(i + 1) % len(wires)]])
@qml.qnode(dev, interface="torch")
def qcnn_layer(features, weights):
"""
QCNN layer: encode 6-dim point cloud features,
apply parameterized rotations + entanglement, measure.
"""
angle_encode(features, wires=range(N_QUBITS))
entangling_layer(wires=range(N_QUBITS))
# Parameterized rotation layer
for i in range(N_QUBITS):
qml.RZ(weights[i, 0], wires=i)
qml.RY(weights[i, 1], wires=i)
entangling_layer(wires=range(N_QUBITS))
# Second parameterized layer
for i in range(N_QUBITS):
qml.RZ(weights[i, 2], wires=i)
return [qml.expval(qml.PauliZ(i)) for i in range(N_QUBITS)]
class HybridQCNN(nn.Module):
"""Hybrid quantum-classical 3D object detector."""
def __init__(self, n_classes=3):
super().__init__()
# Classical point feature extraction (simplified PointNet-style)
self.point_encoder = nn.Sequential(
nn.Linear(3, 32), nn.ReLU(),
nn.Linear(32, 6)
)
# Quantum layer weights: N_QUBITS x 3 rotation angles
self.q_weights = nn.Parameter(
torch.randn(N_QUBITS, 3) * 0.1
)
# Classical classifier on quantum output
self.classifier = nn.Sequential(
nn.Linear(N_QUBITS, 16), nn.ReLU(),
nn.Linear(16, n_classes)
)
def forward(self, point_patch):
# point_patch: (batch, 16, 3) local neighborhood
pooled = self.point_encoder(point_patch).mean(dim=1) # (batch, 6)
q_out = torch.stack([
torch.stack(qcnn_layer(pooled[b], self.q_weights))
for b in range(pooled.shape[0])
]) # (batch, N_QUBITS)
return self.classifier(q_out)
# Parameter count comparison
model = HybridQCNN(n_classes=3)
q_params = sum(p.numel() for p in model.parameters())
print(f"Hybrid QCNN parameters: {q_params}")
# Classical PointNet equivalent: ~40,000 parameters
print(f"Parameter reduction vs PointNet: ~{(1 - q_params/40000)*100:.0f}%")
Accuracy vs Compute Trade-off
The QCNN was evaluated on a synthetic LiDAR dataset generated from CARLA simulation, containing three object classes: vehicles, pedestrians, and cyclists. The quantum model scored 78% mean average precision (mAP) across classes, compared to 82% for a classical PointNet with comparable training budget. The QCNN used 40% fewer trainable parameters, which is the primary advantage for edge deployment scenarios.
Latency on simulation was not competitive with classical inference; quantum circuit simulation is orders of magnitude slower than GPU-accelerated PointNet. The claimed edge advantage is theoretical: future quantum co-processors integrated with autonomous driving SoCs could execute these shallow circuits natively with sub-millisecond latency and milliwatt power. IonQ Harmony validation confirmed the circuits behave as simulated under low noise conditions, but gate fidelity degradation at scale (above 8 qubits) remains a barrier.
Quantum Advantage Pathway
XPeng’s conclusion is that the near-term advantage case for quantum autonomous driving lies not in current gate-based hardware but in the parameter efficiency argument: smaller models require less memory bandwidth, which is a real constraint on automotive-grade edge chips. The research team identified a specific deployment scenario where a quantum-inspired (classically simulated) model with QCNN-derived architecture could reduce inference chip area requirements by an estimated 25%, enabling higher sensor resolution within the same thermal envelope. This is the roadmap item that justifies continued quantum ML investment even before physical quantum hardware reaches automotive-grade reliability.