Discussion about this post

User's avatar
VJAnand's avatar

@Nathan: "The second core axis by which local models are more likely to be optimized for latency is that companies like OpenAI are trying to take their best model and make it fast enough for real-time audio, while companies and hackers looking on-device can invert the question. It’s an existential question for frontier model providers to follow that path — the capital costs and growth dictate their hand" - Can you clarify who are you calling as frontier model providers? those serving from the cloud like OpenAI or from local devices where a localized version resides

VJ

Expand full comment
1 more comment...

No posts