Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat(benchmark): tui based benchmarking tool #149

Merged
merged 15 commits into from
Mar 30, 2023
Merged

Conversation

OlivierDehaene
Copy link
Member

@OlivierDehaene OlivierDehaene commented Mar 30, 2023

image

cc: @njhill

@OlivierDehaene OlivierDehaene merged commit 610bb1f into main Mar 30, 2023
@OlivierDehaene OlivierDehaene deleted the feat/benchmark_tool branch March 30, 2023 13:26
@njhill
Copy link
Contributor

njhill commented Mar 31, 2023

Thanks @OlivierDehaene, looks great! I haven't had a chance to look at it closely yet but will soon.

@OlivierDehaene
Copy link
Member Author

Thanks!
I find very useful to compare modeling code or how the model behave with different number of shards. It is also useful to see what is the expected latency in worst case scenarios (maximum expected context size, maximum decode tokens) and pick the correct max batch size bellow a latency threshold.
However it is not perfect as the batch never contains any padding which make some models look better but it's a first version.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants