jiajunlong commited on
Commit
03bf8e8
·
verified ·
1 Parent(s): 481d547
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -3,13 +3,13 @@
3
  [![arXiv](https://img.shields.io/badge/Arxiv-2402.14289-b31b1b.svg?logo=arXiv)](https://arxiv.org/abs/2402.14289)[![Github](https://img.shields.io/badge/Github-Github-blue.svg)](https://github.com/TinyLLaVA/TinyLLaVA_Factory)[![Demo](https://img.shields.io/badge/Demo-Demo-red.svg)](http://8843843nmph5.vicp.fun/#/)
4
  TinyLLaVA has released a family of small-scale Large Multimodel Models(LMMs), ranging from 0.55B to 3.1B. Our best model, TinyLLaVA-Phi-2-SigLIP-3.1B, achieves better overall performance against existing 7B models such as LLaVA-1.5 and Qwen-VL.
5
  ### TinyLLaVA
6
- Here, we introduce TinyLLaVA-OpenELM-450M-SigLIP-0.55B, which is trained by the [TinyLLaVA Factory](https://github.com/TinyLLaVA/TinyLLaVA_Factory) codebase. For LLM and vision tower, we choose [OpenELM-450M-Instruct](apple/OpenELM-450M-Instruct) and [siglip-so400m-patch14-384](https://huggingface.co/google/siglip-so400m-patch14-384), respectively. The dataset used for training this model is the [ShareGPT4V](https://github.com/InternLM/InternLM-XComposer/blob/main/projects/ShareGPT4V/docs/Data.md) dataset.
7
 
8
  ### Usage
9
  Execute the following test code:
10
  ```python
11
  from transformers import AutoTokenizer, AutoModelForCausalLM
12
- hf_path = 'jiajunlong/TinyLLaVA-OpenELM-450M-SigLIP-0.55B'
13
  model = AutoModelForCausalLM.from_pretrained(hf_path, trust_remote_code=True)
14
  model.cuda()
15
  config = model.config
 
3
  [![arXiv](https://img.shields.io/badge/Arxiv-2402.14289-b31b1b.svg?logo=arXiv)](https://arxiv.org/abs/2402.14289)[![Github](https://img.shields.io/badge/Github-Github-blue.svg)](https://github.com/TinyLLaVA/TinyLLaVA_Factory)[![Demo](https://img.shields.io/badge/Demo-Demo-red.svg)](http://8843843nmph5.vicp.fun/#/)
4
  TinyLLaVA has released a family of small-scale Large Multimodel Models(LMMs), ranging from 0.55B to 3.1B. Our best model, TinyLLaVA-Phi-2-SigLIP-3.1B, achieves better overall performance against existing 7B models such as LLaVA-1.5 and Qwen-VL.
5
  ### TinyLLaVA
6
+ Here, we introduce TinyLLaVA-OpenELM-270M-SigLIP-0.55B, which is trained by the [TinyLLaVA Factory](https://github.com/TinyLLaVA/TinyLLaVA_Factory) codebase. For LLM and vision tower, we choose [OpenELM-270M-Instruct](apple/OpenELM-270M-Instruct) and [siglip-so400m-patch14-384](https://huggingface.co/google/siglip-so400m-patch14-384), respectively. The dataset used for training this model is the [ShareGPT4V](https://github.com/InternLM/InternLM-XComposer/blob/main/projects/ShareGPT4V/docs/Data.md) dataset.
7
 
8
  ### Usage
9
  Execute the following test code:
10
  ```python
11
  from transformers import AutoTokenizer, AutoModelForCausalLM
12
+ hf_path = 'jiajunlong/TinyLLaVA-OpenELM-270M-SigLIP-0.55B'
13
  model = AutoModelForCausalLM.from_pretrained(hf_path, trust_remote_code=True)
14
  model.cuda()
15
  config = model.config