Commit
·
7eb80ca
1
Parent(s):
8a502c1
Update README.md
Browse files
README.md
CHANGED
@@ -56,7 +56,8 @@ The BERT encoder is based on the following configuration:
|
|
56 |
## Training
|
57 |
This model was trained on a personal fork of [NeMo](http://github.com/NVIDIA/NeMo), specifically this [sbd](https://github.com/1-800-BAD-CODE/NeMo/tree/sbd) branch.
|
58 |
|
59 |
-
Model was trained on an A100 for
|
|
|
60 |
|
61 |
### Training Data
|
62 |
This model was trained on `OpenSubtitles`.
|
|
|
56 |
## Training
|
57 |
This model was trained on a personal fork of [NeMo](http://github.com/NVIDIA/NeMo), specifically this [sbd](https://github.com/1-800-BAD-CODE/NeMo/tree/sbd) branch.
|
58 |
|
59 |
+
Model was trained on an A100 for \~150k steps with a batch size of 256, with a $3 budget on the [Lambda cloud](https://cloud.lambdalabs.com/).
|
60 |
+
Model was allowed to converge with 25M training sentences (1M per language).
|
61 |
|
62 |
### Training Data
|
63 |
This model was trained on `OpenSubtitles`.
|