Human life includes a lot of adversarial training (lying relatives) and training in temporal logics, which would seem to be a somewhat different domain than purely linguistic computations (e.g. staying up late, feeling bad; working hard at a task for months, getting better at it; feeling physical skills, even editing Go with emacs, move from the conscious layer into the cerebrellar layer). I think attention is a poor mans "OODA" loop; cognitive science is learning that a primary function of the brain is predicting what will be going on with the body in the immediate future, and prepping for it; that's not a thing that LLMs are architecturally positioned to do. Maybe swarms of agents (although in my mind that's more of a way to deal with LLM poor performance with large context of instructions (as opposed to large context of data) than a way to have contending systems fighting to make a decision for the overall entity), but they still lack both the real-time computational aspect and the continuously tricky problem of other people telling partially correct information.
There's plenty of training data, for a human. The LLM architecture is not as efficient as the brain; perhaps we can overcome that with enough twitter posts from PhDs, and enough YouTubes of people answering "why" to their four year olds and college lectures, but that's kind of an experimental question.
Starting a network out in a contrained body and have it learn how to control that, with a social context of parents and siblings would be an interesting experiment, especially if you could give it an inherent temporality and a good similar-content-addressable persistent memory. Perhaps a bit terrifying experiment, but I guess the protocols for this would be air-gapped, not internet connected with a credit card.
Why would you think a system that can reason well in one domain could not reason well in other domains? Intelligence is a generic, on-the-fly programmable quality. And perhaps your coding is different from mine, but it includes a great deal of general reasoning, going from formal statements to informal understandings and back until I get a formalization that will solve the actual real world problem as constrained.
Email itself cannot be regarded as a reliable delivery method. That said, I host my own email service, have for decades, and often have problems sending to people. I am not running a product on it, and so my recipients usually will check in spam since they want my email. My family knows to txt if there is an email I need to read (that isn’t a mail hosting problem but I don’t really read email consistently). I also have a small web site where I can put family recipes and my resume and the odd file that is too large for email. And a mastodon instance, sync thing, dns, and an old fingerd I wrote in Lisp in 2008 when I was done being a stay at home dad and needed an industry job.
It is a great hobby, and a good way to keep aware of current trends in internet infrastructure. And, like riding a bicycle to commute, maximally free of red tape or external regulation.
The gas town discord has two people that are doing transformation of extremely legacy in house Java frameworks. Not reporting great success yet but also probably work that just wouldn’t be done otherwise.
If this was about a useful part of the economy rather than financialized credit access, I would care. Paying 10x to US residents to provide a better Amex experience really doesn’t incentivize anything that is long term useful. Now when materials scientists and skilled scientific technicians are fired from US based jobs, well that is a problem. The economy in the US has fallen prey to the demands of irreality based capital which is seeking, not profit, not good products and healthy trade, but seeking to extract capital and to erect legislative barriers to Adam Smith style competition (between innovative small businesses striving to find an innovative way to push product utility up a bit). If you can’t brag about the good your job is doing for people, life, or history, try again.
Credit cards alter the power dynamic between customer and merchant, especially in a world where so much commerce is remote. I appreciate protection against fraudulent merchant activity, as well as the ease of use relative to alternatives (who likes carrying cash or a checkbook?). Not having to pay for a month is nice but not the reason I use credit cards.
It isn't the managers it is the business. All those geniuses hired and over years and years no one came up with another business model but ads. I pay for ad-free YouTube and would happily pay for ad free search. As would many. Many people would like a google scale micropayments system that isn't ads. The failure to do this led directly to social media becoming customer devouring experiences rather than making good products people want.
Paradoxically, the people who pay for adfree experiences would be the most valuable targets for ads, so I suspect any pay for no ads arrangement will be temporary at best.
> All those geniuses hired and over years and years no one came up with another business model but ads
This isn't true, there were many other ideas. It's just that only KPI was how much money they can make, thus ads won. Companies don't have an axis of ethics or morality.
I think we are talking past each other. I'm saying there are proposed models which are not ads, but they don't maximize earnings. A silicon valley company will always choose earnings over anything else.
1. Allowing non-developers to provide very detailed specs for the tools they want or experiences they are imagining
2. Allowing developers to write code using frameworks/languages they only know a bit of and don't like; e.g. I use it to write D3 visualizations or PNG extracts from datastores all the time, without having to learn PNG API or modern javascript frameworks. I just have to know enough to look at the console.log / backtrace and figure out where the fix can be.
3. Analysing large code bases for specific questions (not as accurate on "give me an overall summary" type questions - that one weird thing next to 19 normal things doesn't stick in its craw as much as for a cranky human programmer.
It does seem to benefit cranking thru a list of smallish features/fixes rapidly, but even 4.5 or 4.6 seem to get stuck in weird dead ends rarely enough that I'm not expecting it, but often enough to be super annoying.
I've been playing around with Gas Town swarming a large scale Java migration project, and its been N declarations of victory and still mvn test isn't even compiling. (mvn build is ok, and the pom is updated to the new stack, so it's not nothing). (These are like 50/50 app code/test code repos).
Why do all of that when you can just keep a tight hold on an agent that is operating at the speed that you can think about what you're actually doing?
Again, if you're just looking to spend a lot of money on the party trick, don't let me yuck your yum. It just seems like doing things in a way that is almost guaranteed to lead to the outcomes that people love to complain aren't very good.
As someone getting excellent results on a huge (550k LoC) codebase only because I'm directing every feature, my bottleneck is always going to be the speed at which I can coherently describe what needs to be done + a reasonable amount of review to make sure that what happened is what I was looking for. This can only work because I explicitly go through a planning cycle before handing it to the agent.
I feel like if you consider understanding what your LLM is doing for you to be unacceptably slow and burdensome, then you deserve exactly what you're going to get out of this process.
There's plenty of training data, for a human. The LLM architecture is not as efficient as the brain; perhaps we can overcome that with enough twitter posts from PhDs, and enough YouTubes of people answering "why" to their four year olds and college lectures, but that's kind of an experimental question.
Starting a network out in a contrained body and have it learn how to control that, with a social context of parents and siblings would be an interesting experiment, especially if you could give it an inherent temporality and a good similar-content-addressable persistent memory. Perhaps a bit terrifying experiment, but I guess the protocols for this would be air-gapped, not internet connected with a credit card.
reply