This may be a case where humans do well on the test, but you can do very well on the test without doing anything the way a human would. The fact that GPTs aren’t very good at the test isn’t probably evidence that they’re not really very smart, but it doesn’t really mean that if we fix them to do very well on the test that they’ve gotten any smarter.
My thesis is that the calendar will play a significant role in the next OS UI paradigm, replacing the old grid of icons and overlapping windows. Everything will happen around a unified timeline, through which you will launch an app by time-blocking its use, check notifications by looking at the past, forecast battery life or weather by looking at the future, undo/redo actions (or view snapshots/backups) by time traveling, etc.
It is shocking how far behind our map of time (e.g., Google Calendar) is compared to our map of space (e.g., Google Maps). Every spatial feature has an obvious temporal equivalent that just isn't implemented. I want to be able to search, save, and review events. I want to schedule an itinerary of events. I want turn-by-turn navigation in time.
Of course, solving time is just the beginning. The real magic happens when you combine space and time. For this, we may need to wait for more AR/VR adoption, whose added dimension should make this realization obvious to most and significantly facilitate its implementation.
https://github.com/markwk/awesome-biomarkers#biomarker-track...
notes:
- photos/album search now includes video understanding, which imo seems very good from the first 2 examples they showed. includes scroll to exact time of the moment you describe.
- Mail and Notifications will show summaries instead of str[:x]
- Siri now knows iPhone, becomes the ultimate manual on how to use the increasingly complicated iOS 18. and can read your texts (!) to suggest actions with Personal Context Understanding (also it will try to advertise apple tv shows to you... i'm SURE it will be totally objective and aligned to your preferences amirite)
- new iphone 16 camera control button is PRIME real estate - notice how OpenAI/ChatGPT is now next to Google search, and both are secondary clicks to Apple's visual search, which comes first
- camera adds events to calendar!
- "all done on device" and on cloud (though craig doesnt say that haha)
overall i think insanely good ideas on ai + phone integrations.
In their demos, they use the action button to capture an ambient song for Shazam, the power button to capture a voice command for Siri, and the camera button for an image for Visual Intelligence. All 3 captures should be performed using the same button.
And screenshots still require pressing 2 buttons simultaneously. Unless you want to share your screen with Siri in which case it's the power button...
People are going to use this button as a voice recorder, and Apple will announce native support next year.
[1] https://miguelrochefort.com/blog/capture-button