Skip to content

Add quantization and tuning ops as part of model compile hash#180

Open
TedThemistokleous wants to merge 3 commits intorocm7.1_internal_testingfrom
add_quant_and_tune_flags_to_hash
Open

Add quantization and tuning ops as part of model compile hash#180
TedThemistokleous wants to merge 3 commits intorocm7.1_internal_testingfrom
add_quant_and_tune_flags_to_hash

Conversation

@TedThemistokleous
Copy link
Collaborator

Description

Add the quantization, tuning and memory limits as inputs to the final hashed output name fora model. This ensure that we're not reusing a different quantized, or tuned model from a previous session.

Motivation and Context

@TedThemistokleous TedThemistokleous self-assigned this Oct 3, 2025
@TedThemistokleous TedThemistokleous added the Bugfix Fix to a bug or reported issue label Oct 3, 2025
turns out we weren't passing exhaustive tune flags for the recompile in along with some other iflags like mem_limit
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Bugfix Fix to a bug or reported issue

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant