-
Notifications
You must be signed in to change notification settings - Fork 537
Support multiple prompts in the runner #9817
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/9817
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 55678ad with merge base 753a88e ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
1e9a596
to
7d2dc2d
Compare
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
7d2dc2d
to
24b6ca4
Compare
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
24b6ca4
to
e3ec4b5
Compare
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Summary: Pull Request resolved: pytorch#9817 As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
e3ec4b5
to
d86ecc8
Compare
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Summary: Pull Request resolved: pytorch#9817 As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
d86ecc8
to
b59f256
Compare
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Summary: Pull Request resolved: pytorch#9817 As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
b59f256
to
8a19c85
Compare
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
8a19c85
to
28292d0
Compare
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
28292d0
to
92ed635
Compare
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Summary: Pull Request resolved: pytorch#9817 As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
92ed635
to
40a42c6
Compare
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
40a42c6
to
43bd572
Compare
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Just curious - will QNN be able to give response to multiple prompts? |
Summary: As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command: ``` ./qnn_llama3_2_runner --model_path hybrid_llama_qnn.pte --tokenizer_path tiktokenizer.bin --eval_mode 1 --prompt "Once upon a time" --prompt "girl named Lily." --prompt "her toys and her favorite toy was a big," --kv_updater "ShiftPointer" --logits_scale 0.1 --output_path output.txt --num_iters 1 ``` It will be hard to use any char as delimiter, so we use `--prompt` to explicitly mark a prompt and collect them together. Reviewed By: kirklandsign Differential Revision: D72276104
43bd572
to
55678ad
Compare
This pull request was exported from Phabricator. Differential Revision: D72276104 |
Yeah, currently it's the same as a fresh prompt for each inference, but we will enable multi turn conversation and previous conversation will be part of the context |
Differential Revision: D72276104 Pull Request resolved: #9817
Summary:
As a preparation for the multiturn conversation, we can provide multiple prompts and execute them in sequence. Example command:
It will be hard to use any char as delimiter, so we use
--prompt
to explicitly mark a prompt and collect them together.Differential Revision: D72276104