They're probably 2-3 years from being able to run worthwhile LLM inference on device assuming they started a week after chatgpt dropped - and that's pushing the timelines. I'll be surprised if next year's pro can do it. I won't be surprised if the 2025 does. I'll be surprised if the 2026 doesn't.