Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

tstesco/use-env-vars #96

Merged
merged 1 commit into from
Feb 5, 2025
Merged

tstesco/use-env-vars #96

merged 1 commit into from
Feb 5, 2025

Conversation

@tstescoTT tstescoTT merged commit f29afd3 into dev Feb 5, 2025
tstescoTT added a commit that referenced this pull request Feb 5, 2025
# change log

* Default handling of MESH_DEVICE for Llama 3.x models
* Modified setup script improvements:
    * Improved environment variable handling and persistence storage integration
    * Added IMPL_ID field (set to "tt-metal" for all current models)
    * Introduced MODEL_VERSION and MODEL_ID variables for better versioning
* Add image input support for image-text-to-text models in client scripts and tools
    * Added support for image input in trace capturing
    * Added new parameters for image width and height
    * Implemented handling of both text-only and image+text trace captures
* Rename client side scripts batch_size options to max_concurrent to indicate client side concurrent request limits
* Fixed the vLLM model registration logic. Added missing ModelRegistry.register_model call for TTLlamaForCausalLM for legacy implementation models
* Updated benchmark path handling to use $HOME environment variable instead of hardcoded /home/user path
* Add benchmark summary support handling for vllm benchmark script, add documentation example
* Added support for a new model "DeepSeek-R1-Distill-Llama-70B" in the model setup configurations
* use CACHE_ROOT and vllm_dir where possible, fix mock.vllm.openai.dockerfile (#96)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant