nielsr HF staff ervin0307 commited on
Commit
61a33f7
·
verified ·
1 Parent(s): b3b624d

Update README.md (#13)

Browse files

- Update README.md (e8502659e3d5a90124a1294331de3b1bc8a9d5d4)


Co-authored-by: Ervin Castelino <[email protected]>

Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -16,7 +16,7 @@ Disclaimer: The team releasing LLaVa-NeXT-Video did not write a model card for t
16
  ## 📄 Model details
17
 
18
  **Model type:**
19
- LLaVA-Next-Video is an open-source chatbot trained by fine-tuning LLM on multimodal instruction-following data. The model is buit on top of LLaVa-NeXT by tuning on a mix of video and image data to achieves better video understanding capabilities. The videos were sampled uniformly to be 32 frames per clip.
20
  The model is a current SOTA among open-source models on [VideoMME bench](https://arxiv.org/abs/2405.21075).
21
  Base LLM: [lmsys/vicuna-7b-v1.5](https://huggingface.co/lmsys/vicuna-13b-v1.5)
22
 
 
16
  ## 📄 Model details
17
 
18
  **Model type:**
19
+ LLaVA-Next-Video is an open-source chatbot trained by fine-tuning LLM on multimodal instruction-following data. The model is buit on top of LLaVa-NeXT by tuning on a mix of video and image data to achieve better video understanding capabilities. The videos were sampled uniformly to be 32 frames per clip.
20
  The model is a current SOTA among open-source models on [VideoMME bench](https://arxiv.org/abs/2405.21075).
21
  Base LLM: [lmsys/vicuna-7b-v1.5](https://huggingface.co/lmsys/vicuna-13b-v1.5)
22