Update README.md
Browse files
README.md
CHANGED
@@ -9,21 +9,23 @@ Zamba-2-2.7B is a hybrid model between state-space models and transformers. It b
|
|
9 |
2.) Instead of a single shared attention block, we utilize two shared attention blocks which are interleaved in an ABAB pattern through the network.
|
10 |
3.) We apply a LoRA projector to each shared MLP block allowing the network to specialize the MLPs at each shared layer with a minimal increase in total parameter count.
|
11 |
|
12 |
-
Zamba was trained using next-token prediction. It uses the Mistral v0.1 tokenizer. Zamba2-2.7B was pre-trained on 3T tokens of text and code data sourced from open web-datasets. Subsequently in a second phase,
|
13 |
|
14 |
-
Note: this is a temporary HuggingFace implementation of
|
|
|
|
|
15 |
|
16 |
## Quick start
|
17 |
|
18 |
### Presequities
|
19 |
|
20 |
-
To download
|
21 |
1. `git clone https://github.com/Zyphra/transformers_zamba2.git`
|
22 |
2. `cd transformers_zamba2`
|
23 |
3. Install the repository: `pip install -e .`
|
24 |
|
25 |
|
26 |
-
You can run the model without using the optimized Mamba kernels, but it is **not** recommended as it will result in significantly higher latency.
|
27 |
|
28 |
To run on CPU, please specify `use_mamba_kernels=False` when loading the model using ``AutoModelForCausalLM.from_pretrained``.
|
29 |
|
@@ -44,13 +46,12 @@ outputs = model.generate(**input_ids, max_new_tokens=100)
|
|
44 |
print(tokenizer.decode(outputs[0]))
|
45 |
```
|
46 |
|
47 |
-
## Model Details
|
48 |
-
|
49 |
-
Zamba utilizes a unique hybrid SSM architecture. This architecture consists of a backbone of Mamba layers interspersed with a shared attention layer. This attention has shared weights to minimize the parameter cost of the model. We find that concatenating the original model embeddings to the input to this attention block improves performance, likely due to better maintenance of information across depth.
|
50 |
|
|
|
51 |
|
52 |
<center>
|
53 |
-
<img src="https://cdn-uploads.huggingface.co/production/uploads/65c05e75c084467acab2f84a/
|
54 |
</center>
|
55 |
|
56 |
|
|
|
9 |
2.) Instead of a single shared attention block, we utilize two shared attention blocks which are interleaved in an ABAB pattern through the network.
|
10 |
3.) We apply a LoRA projector to each shared MLP block allowing the network to specialize the MLPs at each shared layer with a minimal increase in total parameter count.
|
11 |
|
12 |
+
Zamba was trained using next-token prediction. It uses the Mistral v0.1 tokenizer. Zamba2-2.7B was pre-trained on 3T tokens of text and code data sourced from open web-datasets. Subsequently, in a second phase, Zamba2-2.7B was annealed on a mixture of 100B high-quality tokens.
|
13 |
|
14 |
+
Note: this is a temporary HuggingFace implementation of Zamba2-2.7B. It may not yet be fully compatible with all frameworks and tools intended to interface with HuggingFace models.
|
15 |
+
|
16 |
+
A pure Pytorch implementation of Zamba2-2.7B may be found [here](https://github.com/Zyphra/Zamba2).
|
17 |
|
18 |
## Quick start
|
19 |
|
20 |
### Presequities
|
21 |
|
22 |
+
To download Zamba2-2.7B, clone Zyphra's fork of transformers:
|
23 |
1. `git clone https://github.com/Zyphra/transformers_zamba2.git`
|
24 |
2. `cd transformers_zamba2`
|
25 |
3. Install the repository: `pip install -e .`
|
26 |
|
27 |
|
28 |
+
You can run the model without using the optimized Mamba kernels, but it is **not** recommended as it will result in significantly higher latency and memory usage.
|
29 |
|
30 |
To run on CPU, please specify `use_mamba_kernels=False` when loading the model using ``AutoModelForCausalLM.from_pretrained``.
|
31 |
|
|
|
46 |
print(tokenizer.decode(outputs[0]))
|
47 |
```
|
48 |
|
49 |
+
## Model Details
|
|
|
|
|
50 |
|
51 |
+
Zamba2-2.7B utilizes and extends our original Zamba hybrid SSM-attention architecture. The core Zamba architecture consists of a backbone of Mamba layers interleaved with a shared attention layer. This attention has shared weights to minimize the parameter cost of the model. We find that concatenating the original model embeddings to the input to this attention block improves performance, likely due to better maintenance of information across depth. The Zamba2 architecture also applies LoRA projection matrices to the shared MLP to gain some additional expressivity in each block and allow each shared block to specialize slightly to its own unique position while keeping the additional parameter overhead small.
|
52 |
|
53 |
<center>
|
54 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/65c05e75c084467acab2f84a/XrEIEBxd0fqIgh3LyArAV.png" width="300" alt="Zamba architecture">
|
55 |
</center>
|
56 |
|
57 |
|