Who Owns the Read Layer

Diving deeper into

Sam Hall, CEO of Wafer, on AI agent form factors

Interview
they want access to these types of insights they can't get at the application layer or through standard SDKs.
Analyzed 5 sources

The real prize is not taking actions inside apps, it is seeing enough cross app behavior to know which action matters before the user asks. Standard SDKs expose narrow, preapproved events inside one app at a time, while an OS level product can observe patterns across calendar, messages, navigation, ride sharing, and screen context, then rank the best next step from that combined picture.

  • A launcher can tell which app gets opened, and an assistant can call only the actions developers chose to expose. That is useful for simple commands, but it cannot compare Uber versus Lyft prices unless both apps deliberately hand over the same data in the same format.
  • Screen recording and accessibility tools go deeper because they can read what is actually on screen. That is why desktop products like Granola and Limitless moved below the meeting app itself, they want raw audio or screen context instead of waiting for Zoom or another platform to expose it.
  • Forking Android is the most extreme version because the OS sits above every sandbox. That lets a company combine signals from notifications, app usage, and prior actions into a user model, which is the foundation for proactive suggestions rather than reactive chatbot commands.

The next wave of AI products will compete on who owns the read layer of computing. Products that can see only app level APIs will remain command based helpers, while products with system level visibility will start to feel more like ranking engines that decide what to surface, when to surface it, and which app becomes just the execution backend.