-
-
Notifications
You must be signed in to change notification settings - Fork 11.3k
Closed
Labels
performancePerformance-related issuesPerformance-related issues
Description
Proposal to improve performance
No response
Report of performance regression
I tested the accept length ( number of tokens per step) withtypical acceptance sampling. The accept length is even smaller than default reject sampling method.
Here is my experimental details:
- The dataset I used was mt_bench.
- Speculative decoding model's setup:
llama3.1 8b as target model and Qwama-0.5B-Instruct as a draft model (num of speculative tokens is 2)
llama3.1 8b as target model with MLP-speculator.
3 Temperature was set as 0.9
4posterior_thresholdandposterior_alphawere set as default values.
Do you have some experimental results on this? Or do I need to tune some parameters for typical acceptance sampling? Thanks a lot!
Misc discussion on performance
No response
Your current environment (if you think it is necessary)
The output of `python collect_env.py`
Before submitting a new issue...
- Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
Metadata
Metadata
Assignees
Labels
performancePerformance-related issuesPerformance-related issues