droussis commited on
Commit
2515953
·
verified ·
1 Parent(s): 8f333f3

Improve things a bit

Browse files
Files changed (1) hide show
  1. README.md +7 -4
README.md CHANGED
@@ -19,7 +19,9 @@ Krikri is built on top of [Llama-3.1-8B](https://huggingface.co/meta-llama/Llama
19
  ![image/png](llama-krikri-image.jpg)
20
 
21
 
22
- # Base Model Information
 
 
23
 
24
  - Vocabulary extension of the Llama-3.1 tokenizer with Greek tokens
25
  - 128k context length (approximately 80,000 Greek words)
@@ -41,13 +43,14 @@ Krikri is built on top of [Llama-3.1-8B](https://huggingface.co/meta-llama/Llama
41
 
42
  Chosen subsets of the 91 billion corpus were upsampled resulting in a size of **110 billion tokens**.
43
 
44
- # Instruct Model Information
45
 
46
- 🚨 **More information of the post-training corpus and methdology coming soon.** 🚨
47
 
48
 
49
  # How to use
50
 
 
51
 
52
  ```python
53
  from transformers import AutoModelForCausalLM, AutoTokenizer
@@ -73,7 +76,7 @@ outputs = model.generate(input_prompt['input_ids'], max_new_tokens=256, do_sampl
73
  print(tokenizer.batch_decode(outputs)[0])
74
  ```
75
 
76
- # How to serve with OpenAI compatible server via vLLM
77
 
78
  ```bash
79
  vllm serve ilsp/Llama-Krikri-8B-Instruct \
 
19
  ![image/png](llama-krikri-image.jpg)
20
 
21
 
22
+ # Model Information
23
+
24
+ ## Base Model
25
 
26
  - Vocabulary extension of the Llama-3.1 tokenizer with Greek tokens
27
  - 128k context length (approximately 80,000 Greek words)
 
43
 
44
  Chosen subsets of the 91 billion corpus were upsampled resulting in a size of **110 billion tokens**.
45
 
46
+ ## Instruct Model
47
 
48
+ 🚨 **More information on the post-training corpus and methdology coming soon.** 🚨
49
 
50
 
51
  # How to use
52
 
53
+ ## With Transformers
54
 
55
  ```python
56
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
76
  print(tokenizer.batch_decode(outputs)[0])
77
  ```
78
 
79
+ ## With OpenAI compatible server via vLLM
80
 
81
  ```bash
82
  vllm serve ilsp/Llama-Krikri-8B-Instruct \