Hacker Newsnew | past | comments | ask | show | jobs | submit | awesome_dude's commentslogin

Speaking as someone that has spent a large amount of time unemployed because I have a moral compass - let me know when you actually walk that talk.

For me I could only do it because I had "f*ck you" money gained through investments, other people are able to do it because of welfare systems, or even through friends and family.


>> If you are wealthy

Then.. you wouldn't be working...


Why is Elon still working then?

I'm not sure that posting deranged tweets at three in the morning _really_ qualifies as work.

At the risk of drawing moderation ire..

When does Elon work?


He works pretty hard to destabilize democracy

I saw the other day people complaining about AI slop being posted on this site by new accounts - which I agree is bad.

Someone suggested that people with 10k karma and/or 10 years subscription to this site should be able to do things (such as auto-ban) to those accounts.

The account that misrepresented your comment and thus acted in bad faith is one of those 10k+ accounts.

To me, this is a data point showing the fallacy of long term subscription and/or karma accrual as evidence of their quality/good faith abilities


I admit now after rereading that I did misrepresent what they said and I should have read their comment more closely and it was a knee jerk reaction and that its my fault.

This is one of the key "inefficiencies" of the private sector - there might be one winner at the end of the day providing the product that fills the market niche, but there was always multiple competitors giving it a go in the mean time.

A recent example, Mitchell Hashimoto was pointing out that he wasn't "first to market" with his product(s), he was (at least) SEVENTH


Almost tautologically it's not "inefficient" to do so, because free market economics has decided that all the attempts are mathematically worth it, for a high-margin low-marginal-cost product like software.

I'm a little lost as to why seven teams duplicating effort is more "efficient" in any sense of the word than one or two teams working iteratively toward the same goal.

If this were seven government funded teams solving the same problem, people would lose their minds over the 'waste' But when private companies do it, we call it efficient market competition. The duplication is the same - we just frame it differently.

Edit: fixed some typos caused by fat fingers on a phone keyboard


The benefit from having a 5% better product that hundreds of millions of people will use is worth the duplicated effort in the beginning. The numbers just make sense.

>If this were seven government funded teams solving the same problem

The problem here is "government funded" - the trials are not rationalized by free-market economics. That is, a 5% better product in the end would not be worth seven competing developments initially.


> The benefit from having a 5% better product that hundreds of millions of people will use is worth the duplicated effort in the beginning. The numbers just make sense.

This assumes that the duplicated effort arrives at a solution that is better than if it were done by a single team.

> >If this were seven government funded teams solving the same problem

> The problem here is "government funded" - the trials are not rationalized by free-market economics. That is, a 5% better product in the end would not be worth seven competing developments initially.

I think you're saying that 5% is worth it when the free market does it, but 5% gain isn't when the government does it?

I'm hoping you're not because that's impossible - the end result is precisely the same


> The duplication is the same

It is not. Seven teams all working under one leadership is quite different to seven leaderships each working with one team.

When different governments (e.g. USA and USSR), and thus different leaderships, are both trying to solve the same problem (e.g. travel to the moon), that too is considered efficient competition.


Oh, so seven /leaderships/ is what's made the difference?

If a government did this (e.g., seven independent agencies competing for a moon landing), people would call it "fragmented," "uncoordinated," and "bureaucratic infighting."


Seven independent government agencies are still an arm of the same leadership.

When complete organizational separation is introduced, the concerns you speak of go away. In the USA, the ARPA (you might recognize that name from the thing you're using right now) program regularly enables "seven" independent leaders to tackle a problem and this is widely considered a resounding success.


No real scotsmen

Remember, when it comes to government — at least a democratic one — the people complaining are also the leadership. Think about it from their perspective:

- If they do a good job with leadership, only one team will be necessary. Anything else is truly a waste.

- If they do a poor job with leadership, every team will fail. Any more than one is also truly a waste[1].

The latter is the most likely outcome, of course. Now, when you absolve yourself from the process then those points still apply, but now you have several leaders duking it out to see which one doesn't fail. But, for the same reasons, those leaders each only benefit from having one team.

[1] You could argue that all teams are truly a waste, but one team is necessary to show that leadership failed. That brings abstract value, even if it fails to deliver the intended value. You don't know until you try.


If all people complaining are the leadership - then so are all the customers (potential, or otherwise)

The repeated movement of the goalposts here is only evidence of the no real scotsmen strategy being employed.


> If all people complaining are the leadership - then so are all the customers

Not necessarily. Unless you think a global democratic government formed overnight?

> The repeated movement of the goalposts here

Whatever it is you are reading in other threads has no relevance to this one.


> Not necessarily. Unless you think a global democratic government formed overnight?

This is a distraction. Whether it's 300 million voters or 300 million iPhone users, both groups act as the ultimate arbiter of value. If a customer stops paying, the "leadership" of a company fails. If a voter stops voting for one party, or the other, the "leadership" of a state fails. The mechanical result on the "seven teams" is identical: the unsuccessful ones are defunded.

Further, this proves the detachment from reality you are bringing to the conversation - everybody in the private sector knows the golden rule - your customers ARE your employers

THEY dictate what they will pay for, and therefore what can be sold (unless you are a fan of monopolies forcing people to buy things they do not want to)

> Whatever it is you are reading in other threads has no relevance to this one.

Your dishonesty only highlights your bad faith, and as such we are done here.


> If a voter stops voting for one party, or the other, the "leadership" of a state fails.

Political parties in democracy are quite literally labor unions. The people in them do not independently lead the state, they are merely employees, hired by the leadership. You know, that's what you host elections for — to choose which employee you want to hire from the set of candidates who want the job. They may act as sub-leaders within the capacity of their job, but they are not the top leaders we are talking about. "Leadership" here was never intended to be about "middle managers".

That seems pretty obvious, but perhaps this confusion is the source of your misunderstandings?

> and as such we are done here.

Done with what? Thinking other threads are related this one? That is a good idea.


> So, ATMs did impact bank teller jobs by a significant amount. A third of them were made redundant.

That's not quite my read - the original says per branch there was a 1/3 reduction, but your comment appears to say 1/3 total redundancy.

There was, according to the original, a 40% increase in number of branches, meaning a net increase in tellers (my math might be off though)

edit:

100 branches → 140 branches = +40%

100 tellers/branch → 67 tellers/branch = -33%

140 × 67 = 9,380

100 × 100 = 10,000

net difference -620 or just over 6% (loss)


To be fair, almost every society portrays itself as the defender of whatever is right/good.

And, to be equally as fair, the only genuinely good guys are the ones that are too small to enforce their will upon others directly - small countries without arms who are forced to find other ways to engage with others in order to achieve whatever goals they have (resource acquisition)

The Americans have been extremely adept at dominating the discourse via non-government pathways (Hollywood)


I think that when some people talk about "AI" they have "AGI" in mind, and when others talk about "AI" they have "latest computer does the smarts" in mind.

I personally would prefer "AI" to be "AGI" but there's no point fighting the way people use language (see: every damned pedantic comment about English usage ever!! :-)


Agreed that people increasingly interpret AI to mean AGI, but the academic use of "Artificial Intelligence" has been mostly consistent since the famous 1950s Dartmouth workshop that coined the term. It's not just a recent phenomenon and AI has never really meant "broad human-equivalent intelligence". Fun quote from John McCarthy, who helped coined the term: "Artificial intelligence is not, by definition, simulation of human intelligence".

But beyond the pedanticness and authority appeals, I think keeping the term AI distinct from AGI is just useful so it can be an umbrella term for all the human-like smart-ish things computers do. And so its Wikipedia page doesn't have to be re-written.


AI didn't invent the terms, they were a part of the training data given to it.

The real tell is that you've not been in the group of people that use these terms frequently enough for you to think they're normal.

It's like the emdash alarmism, AI never invented emdash, nor did it invent using it frequently. Its training was full of examples, so many that AI picked up using it frequently.


> like the emdash alarmism, AI never invented emdash, nor did it invent using it frequently. Its training was full of examples, so many that AI picked up using it frequently

Look at my comment history. I emdash. But I adapted by removing the spaces around them—AI hasn’t similarly adapted.

Most comments on HN with emdashes aren’t slop. But if it starts getting into Wernicke word-salad territory and there are emdashes? With spaces? At that point, it’s fair to flag.


LOL

I'm laughing not at you but the ludicrousness of the times - I use endash heavily, have done for a minute, but now I see endash used by LLMs with no surrounding spaces.

I think that the "identify AI by some artefact" is just another game of whack-a-mole, and the better approach is to look at the quality of what's being presented.

I have argued before, and still feel strongly, that LLM/AI generated images/audio/text is causing a stronger inspection of what's being presented as fact, which is a healthy thing (how far that will go is yet to be determined, as per when the availability of Photoshop generated content exploded)


> now I see endash used by LLMs with no surrounding spaces

Goddamit. (Flippity floppity floop.)

> LLM/AI generated images/audio/text is causing a stronger inspection of what's being presented as fact, which is a healthy thing

If it is, I agree. What I think is actually happening is folks are skimming and then concluding on vibes. Unfortunately, that means “I don’t agree” gets lumped in with “this is slop.”


Is it that easy?

Machine Learning papers, for example, used to have a terrible reputation for being inconsistent and impossible to replicate.

That didn't make them (all) fraudulent, because that requires intent to deceive.


What do you think it is about machine learning that makes it hard to replicate? I'm an outsider to academic research, but it seems like computer based science would be uniquely easy - publish the code, publish the data, and let other people run it. Unless it's a matter of scale, or access to specific hardware.

A lot of things are easy if you ignore the incentive structure. E.g. a lot of papers will no longer be published if the data must be published. You’d lose all published research from ML labs. Many people like you would say “that’s perfectly okay; we don’t need them” but others prefer to be able to see papers like Language Models Are Few-Shot Learners https://arxiv.org/abs/2005.14165

So the answer is that we still want to see a lot of the papers we currently see because knowing the technique helps a lot. So it’s fine to lose replicability here for us. I’d rather have that paper than replicability through dataset openness.


But the lab must publish at least the general category of data, and if that doesn't replicate, then the model only works on a more specific category than they claim (e.g. only their dataset).

Even with the exact same dataset and architecture, ML results aren't perfectly replicable due to random weight initialisation, training data order, and non-deterministic GPU operations. I've trained identical networks on identical data and gotten different final weights and performance metrics.

This doesn't mean the model only works on that specific dataset - it means ML training is inherently stochastic. The question isn't 'can you get identical results' but 'can you get comparable performance on similar data distributions.


Then researchers should re-train their models a couple times, and if they can't get consistent results, figure out why. This doesn't even mean they must throw out the work: a paper "here's why our replications failed" followed by "here's how to eliminate the failure" or "here's why our study is wrong" is useful for future experiments and deserves publication.

As per my previous comment - we are discussing stochastic systems.

By definition, they involve variance that cannot be explained or eliminated through simple repetition. Demanding a 'deterministic' explanation for stochastic noise is a category error; it's like asking a meteorologist to explain why a specific raindrop fell an inch to the left during a storm replication.


Lack of will. That was one of the main results from the survey from Whitaker in 2020. Making your code reusable and easy to understand is significant work that had no direct benefits for a researcher's career. Particularly because research code grows wildly as researchers keep trying thungs.

Working on the next paper is seem as the better choice.

Moreover if your code is easy for others to run then you're likely to be hit with people wanting support, or even open yourself to the risk of someone finding errors in your code (the survey's result, not my own beliefs).

There are other issues, of course. Just running the code doesn't mean something is replicable. Science is replicated when studies are repeated independently by many teams.

There are many other failure modes SOTA-hacking, benchmarking, and lack of rigorous analysis of results, for example. And that's ignoring data leakage or other more silly mistakes (that still happen in published work! In work published in very good venues even)

Authors don't do much of anything to disabuse readers that they didn't simply get really look with their pseudorandom number generators during initialization, shuffling, etc. As long as it beats SOTA who cares if it is actually a meaningful improvement? Of course doing multiple runs with a decent bootstrap to get some estimation of the average behavior os often really expensive and really slow, and deadlines are always so tight. There is also the matter that the field converged on a experimentation methodology that isn't actually correct. Once you start reusing test sets your experiments stop being approximations of a random sampling process and you quickly find yourself outside of the grantees provided by statistical theory (this is a similar sort of mistake as the one scientists in other fields do when interpreting p-values). There be dragons out there and statistical demons might come to eat your heart or your network could converge to an implementation of nethack.

Scale also plays into that, of course, and use of private data as the other comment mentioned.

Ultimately Machine Learning research is just too competitive and moves too fast. There are tens of thousands (hundreds maybe?) of people all working on closely related problems, all rushing to publish their results before someone else published something that overlaps too much with their own work. Nobody is going to be as careful as they should, because they can't afford to. It's more profitable to carefully find the minimal publishable amount of work and do that, splitting a result into several small papers you can pump every few months. The first thing that tends to get sacrificed during that process is reliability.


It's the committee vs the dictator issue - a small driven individual (or group) can achieve a lot, but they can also turn into tyrants.

A committee forms when there's widespread disagreement on goals or priorities - representing stakeholders who can't agree. The cost is slower decisions and compromise solutions. The benefit is avoiding tyranny of a single vision that ignores real needs.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: