As a non-American, I love what Chinese companies are doing. The progress they are showing and the fact they are sharing the weights of the models is great. I can't wait for the day when companies that "have no moat" like A. , Cursor or even OpenAI are left with a bunch of float matrices and hardware.
I understand people from the US will have an anti-Chinese reaction, but for us in the "third world" that can use both techs, the openess is always good.
I've mention before that we should have a look at Telegraph/telegram speak. There was a HUGE industry in word efficiency at that time. There are hundreds of books.
I even think an LLM trained to communicate using telegram style might even be faster and way cheaper.
Reminds me of the terminus agent/harness on the terminal-bench coding benchmark - they just send send keystrokes to a tmux session. They score pretty well.
I just read a tech "policy" document of one of the largest packaged food makers in Mexico. They explicitly say they ONLY use paid services/software to ensure there is liability and support.
There are A Lot of businesses thar are happy to burn cash for a false sense of security. They don't know better.
> They explicitly say they ONLY use paid services/software to ensure there is liability and support.
Hahahahaha. Yes, by all means, call IBM or Oracle and ask them to compensate you for downtime. Support, sure, of a sort. If you throw enough money at them, you'll get an answer from them in an arbitrarily short time frame.
I would love to see a "distributed LLM" system, where people can easily setup a system to perform a "piece" of a "mega model" inference or training. Kind of like SETI@home but for an open LLM (like https://github.com/evilsocket/cake but massive )
Ideally if you "participate" in the network, you would get "credits" to use it proportionally to how much GPU power you have provided to the network. Or if you can't, then buy credits (payment would be distributed as credits to other participants).
That way we could build huge LLMs that area really open and are not owned by any network.
I would LOVE to participate in building that as well.
This is exactly what I've been building — mycellm.ai. Seed GPU when idle, earn credits, spend on inference. BitTorrent economics for LLM compute. It's live now (my homelab serves the public chat). Would love to have you involved — github.com/mycellm/mycellm
reply