-
Notifications
You must be signed in to change notification settings - Fork 11.9k
Added README.md for main with examples and explanations #1139
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
1a3d66d
to
e7a09bd
Compare
Just updated it to fix a link to the primary README. If you want to see it formatted, you can follow this link. Edit: And again, adding a link back to the front llama.cpp page in case someone finds their way to the main example from outside of github. |
2d94c77
to
9d72015
Compare
@mgroeber9110 is doing the same in #1131... |
Thanks. That's the problem with these large write ups. We probably both started days ago... Edit: I'm biased for sure, but mine feels more complete. If you want to commit @mgroeber9110's README.md first I can try to edit my information into his but they are structured different. I feel like I would probably just end up replacing everything. :( Still, it would get his |
I was trying to compare all the options to see if we could merge the two. I did find his has instruction for Mine is missing: I did have |
9d72015
to
b8cf6b6
Compare
@DannyDaemonic Sorry for the duplicate work - I should have probably made a note on the ticket that I intend to work on it, but your text definitely feels much more comprehensive, while I only took a quick stab at giving people a starting point. From my point of view, it probably makes more sense to merge yours first, and then I can perhaps transfer a few tidbits of information from mine (such as what exactly Just wondering: was it intentional to not mention |
Yeah, I feel bad that time was wasted either way. The I intended to put all the options in there. Even |
I added the missing options with the exception of On a bit of a tangent here, but I have had problems with Alpaca ending a response too soon, sometimes immediately, and never tried |
3784be4
to
befd875
Compare
|
||
These options help improve the performance and memory usage of the LLaMA models: | ||
|
||
- `-t N, --threads N`: Set the number of threads to use during computation. Using the correct number of threads can greatly improve performance. It is recommended to set this value to the number of CPU cores. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'll add physical CPU cores, and the number of performance cores in a chipset with efficiency/performance (E/P) cores
in PR #934.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm watching that pull request. My plan was to put another pull request through removing it from the ## common
section as soon as this is resolved.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I still think we should leave it in; even with a warning in the cmdline the user might miss it; it's good to have multiple paths to getting the information.
I've been slowly working on this. It contains a couple of examples and a longer explanation of the options a user might use.