-
Notifications
You must be signed in to change notification settings - Fork 712
Support multiple prompts in the runner #9817
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/9817
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 55678ad with merge base 753a88e ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
|
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
1e9a596 to
7d2dc2d
Compare
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
7d2dc2d to
24b6ca4
Compare
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
24b6ca4 to
e3ec4b5
Compare
|
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Summary: Pull Request resolved: pytorch#9817 As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
e3ec4b5 to
d86ecc8
Compare
|
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Summary: Pull Request resolved: pytorch#9817 As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
d86ecc8 to
b59f256
Compare
|
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Summary: Pull Request resolved: pytorch#9817 As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
b59f256 to
8a19c85
Compare
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
8a19c85 to
28292d0
Compare
|
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
28292d0 to
92ed635
Compare
|
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Summary: Pull Request resolved: pytorch#9817 As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
92ed635 to
40a42c6
Compare
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
40a42c6 to
43bd572
Compare
|
This pull request was exported from Phabricator. Differential Revision: D72276104 |
|
Just curious - will QNN be able to give response to multiple prompts? |
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
43bd572 to
55678ad
Compare
|
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Yeah, currently it's the same as a fresh prompt for each inference, but we will enable multi turn conversation and previous conversation will be part of the context |
Differential Revision: D72276104 Pull Request resolved: #9817
Summary:
As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command:
It will be hard to use any char as delimiter, so we use
--promptto explicitly mark a prompt and collect them together.Differential Revision: D72276104