giobin commited on
Commit
ef3ce55
·
1 Parent(s): c4c3f6c

Initial commit of LLaVA model and processor

Browse files
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7f035ba974d8ffa11b7d80d651b6549bf399407c53e5e5c651aa66f67028245
3
  size 4982327104
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b94abf308c1f0007659db2b1568447c0cc8f3766825d2a13148570a935e23dc
3
  size 4982327104
preprocessor_config.json CHANGED
@@ -1,74 +1,51 @@
1
  {
2
  "crop_size": {
3
- "height": 224,
4
- "width": 224
5
  },
6
  "do_center_crop": true,
7
  "do_convert_rgb": true,
8
  "do_normalize": true,
 
9
  "do_rescale": true,
10
- "do_resize": {
11
- "crop_size": {
12
- "height": 384,
13
- "width": 384
14
- },
15
- "do_center_crop": true,
16
- "do_convert_rgb": false,
17
- "do_normalize": true,
18
- "do_pad": true,
19
- "do_rescale": true,
20
- "do_resize": true,
21
- "image_grid_pinpoints": [
22
- [
23
- 768,
24
- 384
25
- ],
26
- [
27
- 384,
28
- 768
29
- ],
30
- [
31
- 768,
32
- 768
33
- ],
34
- [
35
- 384,
36
- 1152
37
- ],
38
- [
39
- 1152,
40
- 384
41
- ]
42
  ],
43
- "image_mean": [
44
- 0.5,
45
- 0.5,
46
- 0.5
47
  ],
48
- "image_std": [
49
- 0.5,
50
- 0.5,
51
- 0.5
52
  ],
53
- "size": {
54
- "shortest_edge": 384
55
- }
56
- },
 
 
 
 
 
57
  "image_mean": [
58
- 0.48145466,
59
- 0.4578275,
60
- 0.40821073
61
  ],
62
  "image_processor_type": "CLIPImageProcessor",
63
  "image_std": [
64
- 0.26862954,
65
- 0.26130258,
66
- 0.27577711
67
  ],
68
  "processor_class": "LlavaProcessor",
69
  "resample": 3,
70
  "rescale_factor": 0.00392156862745098,
71
  "size": {
72
- "shortest_edge": 224
73
  }
74
  }
 
1
  {
2
  "crop_size": {
3
+ "height": 384,
4
+ "width": 384
5
  },
6
  "do_center_crop": true,
7
  "do_convert_rgb": true,
8
  "do_normalize": true,
9
+ "do_pad": true,
10
  "do_rescale": true,
11
+ "do_resize": true,
12
+ "image_grid_pinpoints": [
13
+ [
14
+ 768,
15
+ 384
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  ],
17
+ [
18
+ 384,
19
+ 768
 
20
  ],
21
+ [
22
+ 768,
23
+ 768
 
24
  ],
25
+ [
26
+ 384,
27
+ 1152
28
+ ],
29
+ [
30
+ 1152,
31
+ 384
32
+ ]
33
+ ],
34
  "image_mean": [
35
+ 0.5,
36
+ 0.5,
37
+ 0.5
38
  ],
39
  "image_processor_type": "CLIPImageProcessor",
40
  "image_std": [
41
+ 0.5,
42
+ 0.5,
43
+ 0.5
44
  ],
45
  "processor_class": "LlavaProcessor",
46
  "resample": 3,
47
  "rescale_factor": 0.00392156862745098,
48
  "size": {
49
+ "shortest_edge": 384
50
  }
51
  }