mwmathis commited on
Commit
abff0ca
·
verified ·
1 Parent(s): 19054f0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -0
README.md CHANGED
@@ -10,6 +10,8 @@ tags:
10
  - Video
11
  - MQA
12
  - multimodal
 
 
13
  metrics:
14
  - accuracy
15
  library_name: transformers
@@ -17,6 +19,24 @@ library_name: transformers
17
 
18
  # LLaVAction-0.5B
19
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  ## Model Summary
21
  The LLaVAction-0.5B model is trained on EPIC-KITCHENS-100-MQA, based on Qwen2 language model with a context window of 32K tokens.
22
 
 
10
  - Video
11
  - MQA
12
  - multimodal
13
+ - MLLMs
14
+ - LLaVAction
15
  metrics:
16
  - accuracy
17
  library_name: transformers
 
19
 
20
  # LLaVAction-0.5B
21
 
22
+ <div align="center">
23
+ <h2>LLaVAction: evaluating and training multi-modal large language models for action recognition
24
+ </h2>
25
+
26
+ [Shaokai Ye](https://yeshaokai.github.io/)<sup>1**</sup>&nbsp;
27
+ [Haozhe Qi](https://people.epfl.ch/haozhe.qi)<sup>1**</sup>&nbsp;
28
+
29
+ [Alexander Mathis](https://mathislab.org/)<sup>1</sup><sup>†</sup>&nbsp;
30
+ [Mackenzie Weygandt Mathis](https://www.mackenziemathislab.org/mackenziemathis)<sup>1</sup><sup>†</sup><sup>‡</sup>&nbsp;
31
+
32
+ <sup>1</sup> EPFL
33
+
34
+ <sup>**</sup> First authors <sup>†</sup> Senior Authors <sup>‡</sup> Corresponding Author
35
+
36
+ \[[arXiv Paper](https://www.arxiv.org/tbd)\] &nbsp; \[[Project Page](https://mmathislab.github.io/llavaction/)\] &nbsp; \[[Github Repo](https://github.com/AdaptiveMotorControlLab/LLaVAction)\] &nbsp;
37
+
38
+ </div>
39
+
40
  ## Model Summary
41
  The LLaVAction-0.5B model is trained on EPIC-KITCHENS-100-MQA, based on Qwen2 language model with a context window of 32K tokens.
42