You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+3-1Lines changed: 3 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -33,6 +33,8 @@ Nanotron is a library for pretraining transformer models. It provides a simple a
33
33
34
34
📚 **Check out our [Ultrascale Playbook](https://huggingface.co/spaces/nanotron/ultrascale-playbook)** - A comprehensive guide to efficiently scale LLM training with Nanotron!
35
35
36
+
📝 **AI generated docs thanks to [DeepWiki](https://deepwiki.com/huggingface/nanotron)**
37
+
36
38
## Installation
37
39
38
40
To run the code in this project, first create a Python virtual environment using e.g. `uv`:
@@ -108,7 +110,7 @@ For detailed instructions on training your first model, check out our [Your Firs
Increase the value of `--tp` (tensor paralle) to accelerate generation with multiple GPUs and use a larger value of `--pp` (pipeline parallel) for very large models.
113
+
Increase the value of `--tp` (tensor parallel) to accelerate generation with multiple GPUs and use a larger value of `--pp` (pipeline parallel) for very large models.
112
114
113
115
### Debugging with VSCode
114
116
To debug with VSCode, add the following configuration to your `launch.json` file:
0 commit comments