[ECCV 2024] Official GitHub repository for "LingoQA: Visual Question Answering for Autonomous Driving", presenting the LingoQA benchmark, dataset and baseline model for autonomous driving Visual Question Answering (VQA).
[preprint][arxiv][huggingface]
In this repository you will find:
- A summary of the LingoQA dataset and evaluation metric
- An example of how to run the benchmark on your model predictions
- Details about how to download the datasets
- An example of how to run the novel evaluation metric, Lingo-Judge
To run the LingoQA benchmark on your predictions, simply install the requirements for the repository:
pip install -r ./requirements.txt
Export the predictions of your model to a predictions.csv
file with the columns question_id
, segment_id
and answer
corresponding to the evaluation dataset.
You should have 500 answers in the file. Subsequently, run the benchmark using the following command:
python ./benchmark/evaluate.py --predictions_path ./path_to_predictions/predictions.csv
The LingoQA training and evaluation datasets contain:
- Videos corresponding to driving scenarios
- Language annotations
Datset | Split | Videos | QA Pairs | QA Per Scenario | Link |
---|---|---|---|---|---|
Scenery | Train | 3.5k | 267.8k | ~10.9 | https://drive.google.com/drive/folders/1ivYF2AYHxDQkX5h7-vo7AUDNkKuQz_fL/scenery |
Action | Train | 24.5k | 152.5k | ~43.6 | https://drive.google.com/drive/folders/1ivYF2AYHxDQkX5h7-vo7AUDNkKuQz_fL/action |
Evaluation | Test | 100 | 1000 | 10 | https://drive.google.com/drive/folders/1ivYF2AYHxDQkX5h7-vo7AUDNkKuQz_fL/evaluation |
Lingo-Judge is an evaluation metric that aligns closely with human judgement on the LingoQA evaluation suite.
# Import necessary libraries
from transformers import pipeline
# Define the model name to be used in the pipeline
model_name = 'wayveai/Lingo-Judge'
# Define the question and its corresponding answer and prediction
question = "Are there any pedestrians crossing the road? If yes, how many?"
answer = "1"
prediction = "Yes, there is one"
# Initialize the pipeline with the specified model, device, and other parameters
pipe = pipeline("text-classification", model=model_name)
# Format the input string with the question, answer, and prediction
input = f"[CLS]\nQuestion: {question}\nAnswer: {answer}\nStudent: {prediction}"
# Pass the input through the pipeline to get the result
result = pipe(input)
# Print the result and score
score = result[0]['score']
print(score > 0.5, score)
If you find our work useful in your research, please consider citing:
@article{marcu2023lingoqa,
title={LingoQA: Visual Question Answering for Autonomous Driving},
author={Ana-Maria Marcu and Long Chen and Jan Hünermann and Alice Karnsund and Benoit Hanotte and Prajwal Chidananda and Saurabh Nair and Vijay Badrinarayanan and Alex Kendall and Jamie Shotton and Oleg Sinavski},
journal={arXiv preprint arXiv:2312.14115},
year={2023},
}