You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Can I perform the quantization using the "optimum" for my own executorch FP16 model, which needs to be quantized as q4fp16? Here, my model includes mamba blocks.
If yes, then can you guide with step by step process?