LMDrive Model Card

Model details

Model type: LMDrive is an end-to-end, closed-loop, language-based autonomous driving framework, which interacts with the dynamic environment via multi-modal multi-view sensor data and natural language instructions.

Model date: LMDrive-1.0 (based on LLaMA-7B) was trained in November 2023. The original LLaMA-7B also needs to be downloaded.

Paper or resources for more information:

Github: https://github.com/opendilab/LMDrive/README.md

Paper: https://arxiv.org/abs/2312.07488

Related weights for the vision encoder

https://huggingface.co/deepcs233/LMDrive-vision-encoder-r50-v1.0

Where to send questions or comments about the model:

https://github.com/opendilab/LMDrive/issues

Intended use

Primary intended uses:

The primary use of LMDrive is research on large multimodal models for autonomous driving.

Primary intended users:

The primary intended users of the model are researchers and hobbyists in computer vision, large multimodal model, autonomous driving, and artificial intelligence.

Training dataset

  • 64K instruction-sensor-control data clips collected in the CARLA simulator. dataset_webpage
    • where each clip includes one navigation instruction, several notice instructions, a sequence of multi-modal multi-view sensor data, and control signals. The duration of the clip spans from 2 to 20 seconds

Evaluation benchmark

LangAuto, LangAuto-short, LangAuto-tiny, LangAuto-notice

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Collection including OpenDILabCommunity/LMDrive-llama-7b-v1.0