Detailed description of the requested feature
It would be great to have gemma 4 models in your published optimized models (i.e. https://huggingface.co/collections/nvidia/inference-optimized-checkpoints-with-model-optimizer).
Timeline
Soon, but not blocking.
Describe alternatives you've considered
Target hardware/use case
For me, a 5080 for local inference. But it would be broadly useful for anyone running gemma 4 models.
Detailed description of the requested feature
It would be great to have gemma 4 models in your published optimized models (i.e. https://huggingface.co/collections/nvidia/inference-optimized-checkpoints-with-model-optimizer).
Timeline
Soon, but not blocking.
Describe alternatives you've considered
Target hardware/use case
For me, a 5080 for local inference. But it would be broadly useful for anyone running gemma 4 models.