You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+1-6Lines changed: 1 addition & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -256,6 +256,7 @@ List of command-line flags
256
256
| Flag | Description |
257
257
|-------------|-------------|
258
258
|`--tensorcores`| Use llama-cpp-python compiled with tensor cores support. This increases performance on RTX cards. NVIDIA only. |
259
+
|`--flash-attn`| Use flash-attention. |
259
260
|`--n_ctx N_CTX`| Size of the prompt context. |
260
261
|`--threads`| Number of threads to use. |
261
262
|`--threads-batch THREADS_BATCH`| Number of threads to use for batches/prompt processing. |
@@ -425,9 +426,3 @@ If you would like to contribute to the project, check out the [Contributing guid
425
426
## Acknowledgment
426
427
427
428
In August 2023, [Andreessen Horowitz](https://a16z.com/) (a16z) provided a generous grant to encourage and support my independent work on this project. I am **extremely** grateful for their trust and recognition.
428
-
429
-
## GitHub Sponsors
430
-
431
-
The following is a list of top-tier sponsors for this project here on GitHub:
432
-
433
-
* Be the first one! Visit https://github.com/sponsors/oobabooga/.
0 commit comments