-
Notifications
You must be signed in to change notification settings - Fork 1
Issues: tenstorrent/tt-inference-server
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Setup tool to run in tt-metal dev env
enhancement
New feature or request
Model readiness suite
#123
opened Feb 25, 2025 by
tstescoTT
Model and vLLM test script
enhancement
New feature or request
Model readiness suite
#122
opened Feb 25, 2025 by
tstescoTT
5 tasks
Documentation for evals
documentation
Improvements or additions to documentation
Model readiness suite
#121
opened Feb 25, 2025 by
tstescoTT
LLM Eval run script
enhancement
New feature or request
Model readiness suite
#120
opened Feb 25, 2025 by
tstescoTT
Documentation for benchmarking
documentation
Improvements or additions to documentation
Model readiness suite
#119
opened Feb 25, 2025 by
tstescoTT
Benchmark performance run script
enhancement
New feature or request
Model readiness suite
#117
opened Feb 24, 2025 by
tstescoTT
remove model name registry from prompt_generation.py
bug
Something isn't working
#116
opened Feb 24, 2025 by
tstescoTT
run.sh script to codify the New feature or request
docker run
commands and do validation and error handling
enhancement
#115
opened Feb 24, 2025 by
tstescoTT
rename published tt-inference-server packages and remove old packages
#111
opened Feb 24, 2025 by
tstescoTT
pass vLLM KWArgs to run script
enhancement
New feature or request
#105
opened Feb 19, 2025 by
tstescoTT
Add evals support for 3.x meta Llama eval datasets in lm-evaluation-harness
enhancement
New feature or request
#81
opened Jan 28, 2025 by
tstescoTT
Verify device topology automatically
enhancement
New feature or request
#74
opened Jan 21, 2025 by
tstescoTT
MESH_DEVICE management for Llama 3.x implementations
enhancement
New feature or request
#73
opened Jan 21, 2025 by
tstescoTT
Missing
--max_prompt_length
argument running example_requests_client_alpaca_eval.py
#51
opened Dec 2, 2024 by
milank94
Provide example chat template usage
documentation
Improvements or additions to documentation
enhancement
New feature or request
#36
opened Nov 15, 2024 by
tstescoTT
Previous Next
ProTip!
What’s not been updated in a month: updated:<2025-01-28.