Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
Ask HN: The Next Big OS Leap
4 points by rafaelmdec 6 days ago | hide | past | favorite | 30 comments
After witnessing what is being said about the AI Botlers (like OpenClaw/Moltbot/Clawdbot), I believe UIs will start melting big time.

The point, click and type era is over.

Voice will take over as the primary interface.

UIs will be adaptive and enabled on demand.

There will be an AI agent layer on every single PC out there.

Since privacy will be an issue, "Shazam-like" filters will inhibit uncleared capture of voice.

Makes sense?





The next leap, I think will be when internet service providers are bought out by the hardware makers. Or vice versa? Everything tech will come in a bundle - and it will be global/multi-language and come with content. One ring to rule them all type stuff.

Then again, with the Agentic Economy, who knows, there could be a currency-based consolidation. Not sure what that all means for the OS. Maybe look at the payment and finance systems first and work backwards. You do some work for someone, how are they going to pay you and how are you going to spend the money?


Nope, that sounds like a small iteration on UX, not a revolution, so it is not worth the massive cultural change to make it happen. After all, despite what tech folk think, most people really dislike change.

So we'll probably stick with what we've got until AI is truly empowered to change things, which we are probably a decade away from. At that point, it is far more likely that AI will be taking in full audio, video, and data from your environment, and will know you well enough that the mundane tasks will just happen, without need for any UX at all. Maybe a small device for you to tweak things and control non-standard tasks.

But again, that is a decade off, if not two. We're currently headed into the first downturn of the AI-driven world, when the hype dies, people really spell out the problems, platforms realize that most people don't want generative AI, and all of this quiets down, taking a back burner for 7-10 years while the research advances to move beyond today's problems and evolves into what people might actually want.



Nope, because we already could have had that with VR/AR glasses, and while there are some (even impressive) options now, they aren't mainstream. Neither are the 'apps', nor the content interoperable, exchangable.

Furthermore I see nothing wrong with the desktop metaphor, it's just that we mostly only had a miserable magnifying glass, giving only a small viewport into a crammed childs toy, instead of real large high-resolution screens as can be had now, or sensible virtual desktops for more common sizes. To be expanded by "Metisse", an early 2.5D extension for FVWM, and later "User Interface Faćades". Maybe with some Zoomable UI sprinkled on top, like in https://eaglemode.sourceforge.net/ or whatever the clandestine weirdos from https://arcan-fe.com/ may come up with. (IF. EVER.)


There's been a closed version of https://arcan-fe.com/2021/04/12/introducing-pipeworld/ that was VR centric as well as wilder 'layouters' to https://arcan-fe.com/2018/03/29/safespaces-an-open-source-vr... that is still around in my piles here somewhere.

For a handful of reasons (abusive and hostile actors being at the top) we focus elsewhere (https://www.divergent-desktop.org/blog/2026/01/26/a12web/ and https://arcan-fe.com/2025/01/27/sunsetting-cursed-terminal-e...).

AR/VR development in this space is a massive timesink for all the wrong reasons. Hardware vendors absolutely suck here. Everyone is openly or quietly dreaming of the vertically integrated 'app-store tax' being their real source of revenue rather than selling devices.

This means that if you don't want to fuzz around with half-baked proprietary SDKs that break more often than they do what they're supposed to, you get to sit around reverse engineering. As fun as that can be, it's much less so when that is not what you set out to be doing. Half my electronics 'donation boards bin' is discarded HMDs and input devices by now.

Even in the quirky missed opportunities like Tilt5 you have this situation.


Who said people want to wear goggles? I mean, seriously, what on earth is Apple Vision Pro??

Voice is natural, it is fluid, it conveys emotion, intent.

You cannot seriously be comparing metaverse immersion BS with voice commanded devices.


There are other 'glasses' systems in industrial use, giving engineers/technicians plans overlaid into their augmented reality. Usually voice activated. So that exists. Less cumbersome than that Apple or Metaverse stuff. Because industrial users won't have any of that. (mostly)

Yeah. I see that working.

But one thing is seeing the content and another thing is interacting with the UI to accomplish an office task, for instance.


Does anyone actually ask for this? What problem is it solving other than following the hype?

One of the main things I've gotten out of the whole OpenClaw/Moltbot/Clawdbot situation is that the general public has a dangerously low grasp on information security. There's usefulness to that type of assistant, but I have yet to see a compelling, general consumer take on it.


I think that, for the first time in tech history, we have the tools to step away from ineffective app installs and menu cluttering and memorization and that is a rather big thing.

If you don't agree, take a step back and tell me how many people prefer navigating a terminal window using a keyboard instead of a graphic interface using a mouse.

The future belongs to a more frictionless, no keyboard, voice activated UI, IMHO.


Many professionals, not even necessarily in IT prefer the "green screen", because it enables them to do things faster with a few key-strokes, instead of having to click around in laggy menues.

I guess, maybe because you don't know it any better(systems and device form factors), you're trying to correct an already dumbed down(for mass acceptance) interface paradigm, with one which is even more indirect and imprecise.


Yup. Many like tens of thousands out of billions. Makes sense.

Trillions of flies eat shit. Makes sense?

You cannot be seriously thinking that the future lies in memorizing commands and typing words one stroke at a time on a keyboard.

We are already seeing traditional coding evaporate overnight, let alone have people memorize commands and type it like we were in the 19th century.


I can't tell how this is developing, and which parts will be adopted by the masses, if offered at all, and which wont, and how that will change what the few remaining professionals do.

I'm just thinking it's not as clear-cut as you make it to be, as the past shows, multiple times. For whichever, maybe technically unrelated reasons.

Also "use it or lose it" and "learned helplessness" comes to mind.


Tks for sharing your views on this

And, BTW, according to Henry Ford, if he listened to his customers, he would have gone after faster horses.

Most people don't see innovation until it is materialized in front of them.


The next big OS leap is a capabilities based security with a microkernel. The old model of assuming you wanted to share your authority with everything you run is unsustainable. It should have been a thing at least 20 years ago.

>>Please elaborate. How does this resonate with the average user who doesn't know anything about infosec?

Elaboration, with too much pop culture... ;-)

When you use cash, for example, you're using capabilities. You can hand out exactly $3.50 to the Loch Ness Monster[1], and no matter what, he's not going to be able to leverage that into taking out your entire bank balance, etc.

The current "ambient authority" system is like handing the Loch Ness Monster your wallet, and HOPING he only takes $3.50.

Another metaphor is power outlets, which limit how much of the power from the grid makes it to your device. The current system is much like the electric - i - cal, at the Douglass house in Green Acres.[2]

The point is, you can run any program you want, and give it only the files you want, and nothing else, by default in such a system. For the user, it really doesn't have to seem that different, they already use dialog boxes to select files to open and save things, they could use a "power box"[3] instead, which looks the same, except then the OS enforces their choices.

[1] https://www.quora.com/Why-does-the-Loch-Ness-monster-want-3-...

[2] https://youtu.be/EnGyq2JYrHk?si=c2iTB9BYxB0VwZ9u&t=184

[3] https://wiki.c2.com/?PowerBox


Please elaborate. How does this resonate with the average user who doesn't know anything about infosec.

Good stuff, tks

Or maybe the next big OS leap is decentralization along with data sovereignity. Each person being their own server without so many dependencies to clouds and huge processing/database power inside their own pockets.

I have difficulty to see that, as it requires proper packaging and distribution for mainstream adoption.

Plus the average user doesn't care about data sovereignty, what they care about is UX and dopamine.

How many users you know of that are concerned with data collection by big tech? How much does that account for percent wise?


From that perspective the operating system are a few apps on the browser since that is that users see and get their dopamine rush.

In my opinion users won't choose a future where sovereignity matters. It is a future forced upon them, because they will simply pick whatever methods are still available for communicating.

We used to see this only in a few edge countries like NK and Cuba. Now it is becoming the norm on Russia, Iran, China, Japan and more recently even Europe if you follow the most recent developments. What was once an open garden is becoming quite fenced.


Why does everyone think people want to talk to their computers? There are so many places where talking isn’t appropriate.

    Since privacy will be an issue, "Shazam-like" filters will inhibit uncleared capture of voice.
So now the operating system will decide which recordings are "cleared" and which aren't? Fuck outta here with that nonsense

Reminded me of the recent Q.ai acquisition by Apple. It's obvious that if you're using voice it should be as clear as possible.

I was going to suggest the next big leap will be some kind of "OmniLinux" that spreads across all devices, appliances and hardware that contains any kind of OS and enables interoperability, control and telemetry. Allows updating firmware from a central point, access control and power management. Will be used by humans first, then bots later. There might be some big retro movement to old world things as a result when people reject the idea of a "common dashboard" for the things they own. Might be useful for sharing and rentals though. Is a new OS needed for this though, why not some standards and protocols?

Had this vague thought that the OP is a bot. Does it matter?


OP is not a bot. I can tell you that for sure.

To your point, and I see that it is a recurring opinion around here.

I like the way you think, yet I find it difficult to see how one such movement would emerge.

Most likely from the Open Source community, as I do not see any incumbent intending to go in that direction, don't you think?


Generally you'd need a killer app/device designed from the standpoint of a vision and once the app/device spreads, so does the vision.

I had the idea of a trashcan that knows what's being thrown away, say with a handheld barcode scanner. The product can be a wifi connected barcode scanner. Each time you scan something, you'd say something like: need 3 more now, past expiry, didn't like, find alternative, order more in 2 weeks. And that barcode scanner would do the ordering through an Agent for replacements.

Not sure if it needs to be wifi connected, but needs web access somehow at some point.

On the other side, suppliers would pay to access and fulfill these orders. Eventually that would be done by Robotaxis and drones. There could be a screen attached to this barcode scanner too ( see this: https://news.ycombinator.com/item?id=46699782 )

So I guess the goal is to setup a site MyStuff. And aim to slot in OmniLinux in there. Also worth taking into account the recycling aspect.. trash is money.

Maybe you could even take the scanner shopping to discover and scan things with it... not sure how retailers would feel about that.

Could also be used to quickly and easily sell one's stuff, or tokenize it.

But yeah, this is just zooming in on a starting point. The other idea I had was community notice boards. Always important across the world, especially for launching a business. Maybe they could have little cameras in them to record new pinups. Sell to local councils.

So with MyStuff, I am not sure how that overlaps with IoT and Tokenization, some things would be assets, some just things.

Maybe get a really good website going and find a way to let people populate it with data.

I am attempting to move away from smartphones as well. Maybe there's a way to give away these scanners forA Social Network for AI Agents free or provide rebates if the data they collect is solid and/or leads to real transactions. Can also be used as a crypto wallet in face-to-face transactions. That'd be nice - and a way to limit the loss if the scanner is lost/stolen.

Good luck.


I'm not gonna say don't do it. By all means, go ahead and try it out.

My only advice is: don't get married to the solution.

Or, in wiser words, fall in love with the problem, not the solution (I think Ury Levine coined that one).

At the end of the day, either you'll have learned or succeeded.


I see it as a rather logical step with the advances in voice first AI wearables.

Think about it. Not everyone wants to be recorded as a bystander. Privacy will be an issue.

The technology for audio signature already exists and works fine.

It will be a matter of opt-in/opt-out from users, not an OS decision.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: