Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fixes for Llama 3.x support, adding image input tooling, tt-studio environment variable handling #83

Closed
wants to merge 4 commits into from

Conversation

tstescoTT
Copy link
Contributor

change log

  • Add benchmark summary support handling for vllm benchmark script, add documentation example
  • Rename client side scripts batch_size options to max_concurrent to indicate client side concurrent request limits
  • Stop-gap handling of MESH_DEVICE for Llama 3.x models
  • Add image input support for image-text-to-text models in client scripts and tools
    • Added support for image input in trace capturing
    • Added new parameters for image width and height
    • Implemented handling of both text-only and image+text trace captures
  • Modified setup script improvements:
    • Improved environment variable handling and persistence storage integration
    • Added IMPL_ID field (set to "tt-metal" for all current models)
    • Introduced MODEL_VERSION and MODEL_ID variables for better versioning
    • Enhanced HF token validation with character count check
  • Fixed the vLLM model registration logic:
    • Added missing ModelRegistry.register_model call for TTLlamaForCausalLM for legacy implementation models
  • Updated benchmark path handling to use $HOME environment variable instead of hardcoded /home/user path
  • Added support for a new model "DeepSeek-R1-Distill-Llama-70B" in the model setup configurations

…vironment variable handling (#82)

* Add benchmark summary support handling for vllm benchmark script, add documentation example
* Rename client side scripts batch_size options to max_concurrent to indicate client side concurrent request limits
* Stop-gap handling of MESH_DEVICE for Llama 3.x models
* Add image input support for image-text-to-text models in client scripts and tools
  - Added support for image input in trace capturing
  - Added new parameters for image width and height
  - Implemented handling of both text-only and image+text trace captures
* Modified setup script improvements:
  - Improved environment variable handling and persistence storage integration
  - Added IMPL_ID field (set to "tt-metal" for all current models)
  - Introduced MODEL_VERSION and MODEL_ID variables for better versioning
  - Enhanced HF token validation with character count check
* Fixed the vLLM model registration logic:
  - Added missing ModelRegistry.register_model call for TTLlamaForCausalLM for legacy implementation models
* Updated benchmark path handling to use $HOME environment variable instead of hardcoded /home/user path
* Added support for a new model "DeepSeek-R1-Distill-Llama-70B" in the model setup configurations
echo " llama-3-70b"
echo " llama-3-8b-instruct"
echo " llama-3-8b"
echo " DeepSeek-R1-Distill-Llama-70B"
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

i'm up for this, as it will align id with metal's expectations! (model_config.py, L137)

but we need we need to change the names in setup_model_environment as well then.
i've done that locally, will push soon. you can the usage chunk if you wish?

@@ -74,6 +75,7 @@ get_hf_env_vars() {
echo "HF_TOKEN environment variable is not set. Please set it before running the script."
read -r -s -p "Enter your HF_TOKEN: " input_hf_token
echo
echo "entered HF_TOKEN contains: ${#input_hf_token} characters, expected 37."
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

we can ping HF API to check if the token is valid
i have that in my change, will follow up!

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nice, can you open a separate PR on that or share the code you're using? I had a TODO to also check repo access for the target model and message to user on 404.

* fix casing

* clean up HF setup venv before repacking
* update refs to /home/container_app_user/ for container home usage
* update documentation
@tstescoTT
Copy link
Contributor Author

Moved this PR to #88 to use the proposed RC git workflow.

@tstescoTT tstescoTT closed this Feb 1, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants