We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 1981782 commit 68238b7Copy full SHA for 68238b7
README.md
@@ -143,13 +143,6 @@ For instance, if you want to work with larger contexts, you can expand the conte
143
llm = Llama(model_path="./models/7B/llama-model.gguf", n_ctx=2048)
144
```
145
146
-### Loading llama-2 70b
147
-
148
-Llama2 70b must set the `n_gqa` parameter (grouped-query attention factor) to 8 when loading:
149
150
-```python
151
-llm = Llama(model_path="./models/70B/llama-model.gguf", n_gqa=8)
152
-```
153
154
## Web Server
155
0 commit comments