Inference speed is heavily dependent on memory read/write speed versus size. As long as you can fit the model in memory, what’ll determine functionality is the mem bandwidth.
This is not universally true although I see this phrase being repeated here too often. And it is especially not true with the small models. Small models are compute-bound.