Then again, with the Agentic Economy, who knows, there could be a currency-based consolidation. Not sure what that all means for the OS. Maybe look at the payment and finance systems first and work backwards. You do some work for someone, how are they going to pay you and how are you going to spend the money?
>>Please elaborate. How does this resonate with the average user who doesn't know anything about infosec?
Elaboration, with too much pop culture... ;-)
When you use cash, for example, you're using capabilities. You can hand out exactly $3.50 to the Loch Ness Monster[1], and no matter what, he's not going to be able to leverage that into taking out your entire bank balance, etc.
The current "ambient authority" system is like handing the Loch Ness Monster your wallet, and HOPING he only takes $3.50.
Another metaphor is power outlets, which limit how much of the power from the grid makes it to your device. The current system is much like the electric - i - cal, at the Douglass house in Green Acres.[2]
The point is, you can run any program you want, and give it only the files you want, and nothing else, by default in such a system. For the user, it really doesn't have to seem that different, they already use dialog boxes to select files to open and save things, they could use a "power box"[3] instead, which looks the same, except then the OS enforces their choices.
[1] https://www.quora.com/Why-does-the-Loch-Ness-monster-want-3-...
So we'll probably stick with what we've got until AI is truly empowered to change things, which we are probably a decade away from. At that point, it is far more likely that AI will be taking in full audio, video, and data from your environment, and will know you well enough that the mundane tasks will just happen, without need for any UX at all. Maybe a small device for you to tweak things and control non-standard tasks.
But again, that is a decade off, if not two. We're currently headed into the first downturn of the AI-driven world, when the hype dies, people really spell out the problems, platforms realize that most people don't want generative AI, and all of this quiets down, taking a back burner for 7-10 years while the research advances to move beyond today's problems and evolves into what people might actually want.
One of the main things I've gotten out of the whole OpenClaw/Moltbot/Clawdbot situation is that the general public has a dangerously low grasp on information security. There's usefulness to that type of assistant, but I have yet to see a compelling, general consumer take on it.
If you don't agree, take a step back and tell me how many people prefer navigating a terminal window using a keyboard instead of a graphic interface using a mouse.
The future belongs to a more frictionless, no keyboard, voice activated UI, IMHO.
Most people don't see innovation until it is materialized in front of them.
I guess, maybe because you don't know it any better(systems and device form factors), you're trying to correct an already dumbed down(for mass acceptance) interface paradigm, with one which is even more indirect and imprecise.
We are already seeing traditional coding evaporate overnight, let alone have people memorize commands and type it like we were in the 19th century.
I'm just thinking it's not as clear-cut as you make it to be, as the past shows, multiple times. For whichever, maybe technically unrelated reasons.
Also "use it or lose it" and "learned helplessness" comes to mind.
Plus the average user doesn't care about data sovereignty, what they care about is UX and dopamine.
How many users you know of that are concerned with data collection by big tech? How much does that account for percent wise?
In my opinion users won't choose a future where sovereignity matters. It is a future forced upon them, because they will simply pick whatever methods are still available for communicating.
We used to see this only in a few edge countries like NK and Cuba. Now it is becoming the norm on Russia, Iran, China, Japan and more recently even Europe if you follow the most recent developments. What was once an open garden is becoming quite fenced.
Furthermore I see nothing wrong with the desktop metaphor, it's just that we mostly only had a miserable magnifying glass, giving only a small viewport into a crammed childs toy, instead of real large high-resolution screens as can be had now, or sensible virtual desktops for more common sizes. To be expanded by "Metisse", an early 2.5D extension for FVWM, and later "User Interface Faćades". Maybe with some Zoomable UI sprinkled on top, like in https://eaglemode.sourceforge.net/ or whatever the clandestine weirdos from https://arcan-fe.com/ may come up with. (IF. EVER.)
For a handful of reasons (abusive and hostile actors being at the top) we focus elsewhere (https://www.divergent-desktop.org/blog/2026/01/26/a12web/ and https://arcan-fe.com/2025/01/27/sunsetting-cursed-terminal-e...).
AR/VR development in this space is a massive timesink for all the wrong reasons. Hardware vendors absolutely suck here. Everyone is openly or quietly dreaming of the vertically integrated 'app-store tax' being their real source of revenue rather than selling devices.
This means that if you don't want to fuzz around with half-baked proprietary SDKs that break more often than they do what they're supposed to, you get to sit around reverse engineering. As fun as that can be, it's much less so when that is not what you set out to be doing. Half my electronics 'donation boards bin' is discarded HMDs and input devices by now.
Even in the quirky missed opportunities like Tilt5 you have this situation.
Voice is natural, it is fluid, it conveys emotion, intent.
You cannot seriously be comparing metaverse immersion BS with voice commanded devices.
But one thing is seeing the content and another thing is interacting with the UI to accomplish an office task, for instance.
Since privacy will be an issue, "Shazam-like" filters will inhibit uncleared capture of voice.
So now the operating system will decide which recordings are "cleared" and which aren't? Fuck outta here with that nonsenseThink about it. Not everyone wants to be recorded as a bystander. Privacy will be an issue.
The technology for audio signature already exists and works fine.
It will be a matter of opt-in/opt-out from users, not an OS decision.