Why Multimodal Large Language Models (MLLM) is promise for Autonomous Driving? - Blockchain.News

Why Multimodal Large Language Models (MLLM) is promise for Autonomous Driving?

The integration of MLLMs in autonomous driving could revolutionize the global economy, with ARK's research suggesting a potential GDP increase of 20% over the next decade, driven by safety improvements, productivity gains, and a shift to electric vehicles.

  • Jan 12, 2024 07:23
Why Multimodal Large Language Models (MLLM) is promise for Autonomous Driving?

The integration of Multimodal Large Language Models (MLLMs) in autonomous driving is reshaping the landscape of vehicular technology and transportation. Recently, a paper "A Survey on Multimodal Large Language Models for Autonomous Driving" presents a comprehensive survey of recent advancements in MLLMs, particularly focusing on their application in autonomous driving systems.


MLLMs, which combine linguistic and visual information processing capabilities, are emerging as key enablers in the development of autonomous driving systems. These models enhance vehicle perception, decision-making, and human-vehicle interaction, leveraging large-scale data training on traffic scenes and regulations.

Development of Autonomous Driving

The journey towards autonomous driving has been marked by significant technological advancements. Early efforts in the late 20th century, like the Autonomous Land Vehicle project, laid the groundwork for current systems. The last two decades have seen improvements in sensor accuracy, computational power, and deep learning algorithms, driving advancements in autonomous driving systems.

The future of Autonomous Driving

A recent study by ARK Investment Management LLC highlights the transformative potential of autonomous vehicles, particularly autonomous taxis, on the global economy. ARK’s research forecasts a significant boost in global gross domestic product (GDP) due to the advent of autonomous vehicles, estimating an increase of approximately 20% over the next decade. This projection is based on various factors, including the potential for reduced accident rates and lowered transportation costs. The introduction of autonomous taxis, or robotaxis, is expected to have a profound impact on GDP. ARK estimates net GDP gains could approach $26 trillion by 2030. This is significant, amounting to about 26% of the current size of the US economy. ARK’s analysis indicates that autonomous taxis could be one of the most impactful technological innovations in history, potentially adding 2-3 percentage points to global GDP annually by 2030. This impact surpasses the combined contributions of the steam engine, robots, and IT to the economy. Consumers are likely to benefit from decreased transportation costs and increased purchasing power.

Role of MLLMs in Autonomous Driving

MLLMs are crucial in various aspects of autonomous driving:

Perception: MLLMs improve the interpretation of complex visual environments, translating visual data into text representations for enhanced understanding.

Planning and Control: MLLMs facilitate user-centric communication, allowing passengers to express their intentions in natural language. They also help in high-level decision-making for route planning and vehicle control.

Human-Vehicle Interaction: MLLMs advance personalized human-vehicle interaction, integrating voice commands and analyzing user preferences.

Challenges and Opportunities

Despite their potential, applying MLLMs in autonomous driving systems presents unique challenges, primarily due to the necessity of integrating inputs from diverse modalities like images, 3D point clouds, and HD maps. Addressing these challenges requires large-scale, diverse datasets and advancements in hardware and software technologies.



MLLMs hold significant promise for transforming autonomous driving, offering enhanced perception, planning, control, and interaction capabilities. Future research directions include developing robust datasets, improving hardware support for real-time processing, and advancing models for comprehensive environmental understanding and interaction.

Image source: Shutterstock
. . .