Hacker Newsnew | past | comments | ask | show | jobs | submit | ForHackernews's commentslogin

Cool. Fork it and maintain your own runtime, then. Why are you complaining about what the bun team does with their project?

Cause I'm sick of this amateur hour shit

Then show it wrong by making something people want to use more instead. If it's too difficult then the existing project must not be very amateur.

From what I'm reading, it's too late for Bun. I hear the whole dev stream is slop now. It was nice while it lasted, but that's not a foundation to build rock-solid stuff on top of. Not for me, not for them, not for anyone.

It's totally fair to use GPL code, it just means all the models built by Anthropic, OpenAI, etc. using GPL-licensed source are themselves bound by the GPL. Plus, any works created downstream using those AI tools.

We're on the verge of a golden age of software as soon as someone finds a court with courage.


Ah, you have much more faith in the legal system than I do. It's nice to dream, though.

Frustrating because the Slack snap is broken so every day you have to downgrade it and I guess you can't without connectivity.

This might be the incentive I need to finally purge snap.


Snap recently got much more polished.

I used to have to find a script to purge excess old snaps that would fill up my hard drive. Now Ubuntu only keeps two versions of each snap.

I was wondering why the script didn't have to ever clean more than one version, even when I took longer between running updates.


Just move to flatpak, much nicer to deal with

I got rid of both and my system is much better for it. The only thing I still use that is distributed in such a format is AppImage, and mainly because it has never given me trouble.

Both fail hard for so many things. If you need any sort of hardware acceleration, just use an rpm/deb.

In my testing I find the exact reverse. I much prefer snap to flatpak.

Snap is mostly limited to Ubuntu and has to run as a daemon.

Flatpak gives me cross-platform/cross-distro software directly/certified by the project or company that has additional security sandboxing and doesn't open up potential security issues.

I don't have to wait for a distro package, and yet there are no system integration concerns.

It also works great for atomic distros (SilverBlue, etc)


snaps work on anything that has systemd so I don't quite know where you got the idea is that it is mostly limited to ubuntu.

IDK just using Linux for like 30 years.

They're the only main distro I know of that uses it at all by default.

Red Hat, Fedora, Debian, Ubuntu, Arch, etc. Ubuntu is the only one.


> Ubuntu is the only one.

I have to say that I really do not understand the point of comments like this.

The reality of user-facing Linux is this:

Most of it is Android. It sells billions of units per year.

Of the rest, something like 90% is ChromeOS. That recruits 2-3 hundred million new hardware sales a year, which is roughly 10x as many users annually as the entire history of the Linux distro world. ChromeBooks sell more than desktop PCs and more than Macs by value although Macs cost at least 5x as much.

But of the remaining, oh let's be generous and say 10% of Linux users, most is Ubuntu.

Debian-based distros are 2/3 to 3/4 of all traceable measurable Linux usage, and 2/3 of Debian-family use is Ubuntu.

Red Hat and its fantastic valuation is like 1% of the Linux world.

So what you are saying is "only the Linux mainstream uses snap, and all the weird little hacker/hobbyist distros that have like ten users don't use it."


Designed For Chromenet Explorer 6.0


I'm not familiar with Ed Zitron but failing to call the top of a bubble doesn't mean you're wrong about it being a bubble. People who were calling out the housing bubble in the 2000s were "wrong" right up until they were right. e.g. from 2006 https://www.nytimes.com/2006/01/02/opinion/no-bubble-trouble...

My own feeling is that it is a bubble: AI models are the new virtual machines. They will become commodified and low-margin hosting providers will dominate the market. Investors in OpenAI/Anthropic will lose their shirts.


> They will become commodified and low-margin hosting providers will dominate the market.

I have my doubts about this. We have not seen a viable YouTube alternative because the underlying costs of handling video content are significant and YT has custom hardware and sophisticated software. When we look to the broader cloud market, hyperscalers dominate. We are likely seeing similar when it comes to Google's TPU and access to Nvidia's best offerings.

That being said, I did just pick up a DGX Spark and it runs qwen-3.6 sufficiently well to be a viable interactive coding assistant. Certainly more than enough for unattended agents.


YouTubes biggest moat the last 10 years is probably more that all the viewers and creators are already there. Any competitor has a huge disadvantage - creators are not interested in a place without viewers, and viewers not in a place without creators/content.

yeah, network effect is real, and you cannot get viewers without competitive video delivery, so perhaps the moat is more like having an ocean on both sides

Network effect means it will be a huge and risky undertaking, and one needs to solve the bootstrap problem. But the costs of video delivery means that one would have to burn serious cash in the meantime. So it works in tandem.

TikTok kinda did manage to make a dent though - I suspect it substitutes for YouTube in some cases (though not all).


Maybe, but if custom hardware and economies of scale are the determining factor, that favors Google (and Amazon/Microsoft), not OpenAI or Anthropic.

Definitely, Google Vertex Ai serves up other companies models better than they can themselves. The TPU is the bee's knees. I really hope Google makes a take on the DGX Spark

arguing that the reason youtube is succeeding is because of video hosting costs is hilariously misguided.

the content and creators are the only competitive advantage they have. there are MANY video hosting platforms out there but they just don't have the content to attract large audiences like youtube does. they have a strong early mover advantage


Theres no viable yt alternative because of network effects not the video hosting

Those same network effects dont exist (yet) on models


There is a bit of apples vs. oranges

It is unlikely that models will have network effect because (1) there is less of a two-sided marketplace and (2) people are already forming brand preferences. We also see significant convergence among the agent harnesses as well.

I'm currently building out an internal agentic orchestration platform for business and development and a requirement is to support multiple models and tools so people have an amount of choice.


Discoverability of other content and ad money. And then critical mass of viewers leading to sponsorships and other exploitative models of monetising outside Google.

Ads might be questionable model for lot of use cases. And network model only works for promotion but does not lock users in because content is only available in one place.


im not sure i understand your reply, but it sounds like you're agreeing with me that yts biggest advantage is the network effect?

Yes. And that won't work wit AI.

> AI models are the new virtual machines

Deepseek v4 flash is priced at 1/10 that of openai/anthropic. I can see a race to the bottom - or perhaps an android vs iphone split - where, the premium market is served by openai/anthropic and there is a long-tail of commodity vendors.


It's priced at 1/10, but deepseek is probably not profitable, also it's slow.

Even more interesting is the question if we would have a deepseek model without the US frontier models.

And then what's the value of the advantage that the frontier models have. It's definitely 100x more valuable to find zero days 3months earlier. Probably not in every domain but in enough domains having the smartest model is valuable.


False. Deepseek and other providers who host deepseek have no incentive to subsidise. They also price it similarly. So it is the true value.

iPhone is a consumer brand, and to some extent a fashion/status signalling choice. The market pressures in the B2B space are quite different, I expect lots of cheap good-enough models (Deepseek and others) will end up powering customer service chatbots and the like.

Who will pay 500x the price for a 1% better model? Quants and traders?


Much of the agentic intelligence is at the client. The llm backend is largely swappable. For instance, claude-code paired with any model performs well enough for many usecases. In fact, the real breakthrough is how an agent paired with an unreliable llm could perform well. Given this dynamic, I see llm tokens as the electrons or electricity, and agents as the toasters, and appliances using those electrons. If you extend this analogy, value will bubble up into the appliances which would each have consumer preferences. A token is a token no matter who produces it, just as an electron is, but I like my KitchenAid toaster, whats your preference ?

But Claude Code works about as well as Codex and roughly the same as Github Copilot. So what part of this pipeline is supposed to command premium pricing?

You might prefer your KitchenAid toaster but I'd wager you won't pay enough to support a trillion dollar valuation.


- If costs are going up: oh no AI is not sustainable! Bubble bursting!

- If AI costs are going down: oh no it’s getting commoditised, OpenAI bankrupt anytime soon TM

- If companies have moat and get bigger: oh no companies are getting powerful! It’s bad and we must oppose them because they can rug pull anytime and enshittify!!

What situation is something that you would be okay with? Because people seem to have a problem with any outcome.


Huh? What makes you think I'm objecting to anything?

I'm actually pretty happy if we have a competitive market for AI that maximizes consumer surplus. For a while there it looked like AI might remain in the hands of two or three corporate giants.

(I won't be buying the OpenAI IPO, that's all)


There has to be some time-based discount factor to calling a bubble.

I've corrected the typo now, but I almost let it stand as a testament to my humanity.

Yea no worries, just wanted others to know since when I typed the comment, the submission was quite fresh. It was my hope that it would make less people confused.

I didn't mind that there was a typo.


I assume most readers could decipher that typo, but I recognize your need to be pedantically correct as part of the grand tradition of HN.

They might be the most complex mass-produced commercial machines but the Large Hadron Collider has a plausible claim to the title of "world's most complex machine" https://www.guinnessworldrecords.com/world-records/103591-la...

agree

I think something like this wins that category:

https://en.wikipedia.org/wiki/Eastern_Interconnection


It's strange to count this as single machine. But if we go this way, the north Chinese interconnection and the continental Europe interconnection are bigger (https://en.wikipedia.org/wiki/Wide_area_synchronous_grid).

I suppose it's partly a semantic question that hinges on what you count as a single "machine" and what's a system or a network.

The fact that there’s more than one copy of TWINSCAN EXE 5k seems important

What makes it read-only? You told the AI "don't"?

Actually, it's deterministic -- our product doesn't move money, so when the user gives us access through Plaid, we're only getting read-level permissions. We actually don't even get full account numbers.

A company working with Plaid has to request separate product "scopes" through Plaid in order to be able to move money.


I'm not that familiar with Plaid, but if it works like Yodlee, users have to hand over their credentials so there's no real security, it's just that their scraper is designed to be look not touch.

Plaid has OAuth-based access for most of the big institutions now, but yes, for smaller institutions, they do scraping. Thankfully, Plaid's been around for a while now and has a good track record. It would be a non-starter to give your credentials to a small startup directly.

the question isn't whether the user is trusting Plaid with too much access, the question is whether Plaid is trusting these apps with too much.

It's read only until they get hacked like dozens of other ai start ups are...

I seriously hope this doesn't happen by the way but yea. This is not for me.


The awesome thing is, if it's something you're interested in, you can basically set this up yourself -- a local AI model, some browser automation, and it does mostly work! It's just a pain to keep running. But it's definitely doable.

I am the kind of geek who doesn't even want a second record of their finances around. My physical bank statements even have a procedure for how long they live until being shredded.

If I were interested in this though, yea, a local LLM would be the only way.


I respect that 100%. Thanks for sharing your thoughts!


https://www.yodlee.com/ is another provider

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: