Hacker Newsnew | past | comments | ask | show | jobs | submit | MetalSnake's commentslogin

I talk to it in non-English. But have rules to have everything in code and documentation in english. Only speaking with me should use my native language. Why would that be a problem?

Because 90% of training data was in English and therefore the model perform best in this language.

In my experience these models work fine using another language, if it’s a widely spoken one. For example, sometimes I prompt in Spanish, just to practice. It doesn’t seem to affect the quality of code generation.

They literally just have to subtract the vector for the source language and add the vector for the target.

It’s the original use case for LLMs.


Thank you. +1. There are obviously differences and things getting lost or slightly misaligned in the latent space, and these do cause degradation in reasoning quality, but the decline is very small in high resource languages.

It’s just a subjective observation.

It just can’t be a case simply because how ML works. In short, the more diverse and high quality texts with reasoning reach examples were in the training set, the better model performs on a given language.

So unless Spanish subset had much more quality-dense examples, to make up for volume, there is no way the quality of reasoning in Spanish is on par with English.

I apologise for the rambling explanation, I sure someone with ML expertise here can it explain it better.


I saw a curious post recently that explored this idea, and showed that it isn’t really the case. The internal layers of the model aren’t really reasoning in English, or in any human language.

Translation in/out of human languages only happens at the edges of the model.

Internal layer activations for the same concept are similar regardless of language, while activations at the top/bottom layers diverge. Meanwhile the pattern is reversed for same language different content.


So we do at least agree on the fact that quality of human language <-> embeddings transition depends on how good target language is represented in the training dataset?

Even if it happens at the edge, on every conversation turn, I may assume non captured small subtleties of meaning over time can accumulate into significant error.


The agent may be translating to English behind the scenes. There are all kinds of hacks layered on top of the underlying models. But however exactly it works, it’s clearly not the case that these things only make use of Spanish training data when you prompt them in Spanish.

Lets take a language with orders of magnitude less training data, like Nahuatl. The following is what happens if I ask Claude Code in Nahuatl to give me an example of a simple program that I can execute inside a web browser (avoiding any direct keyword prompts like 'Javascript' or 'Hello World'). Interestingly, it responds in Spanish. I doubt that it has any relevant training data in Nahuatl. (Apologies if the Nahuatl is wrong – I just used Google translate.)

    ¿Uelis techmakas se neskayotl tlen se programa tlen amo ouij tlen ueliskia nijchiuas ipan se navegador web?

    -----

    Entiendo tu pregunta en náhuatl. Aquí un ejemplo sencillo — una calculadora en HTML puro que puedes abrir directo en el navegador, sin servidor:

    <!DOCTYPE html>
    <html>
    <head>
        <meta charset="UTF-8">
        <title>Calculadora</title>
        <style>
        body { font-family: sans-serif; display: flex; justify-content: center; padding: 40px; }
        .calc { border: 1px solid #ccc; padding: 20px; border-radius: 8px; width: 200px; }
        input { width: 100%; margin-bottom: 10px; font-size: 1.2em; text-align: right; }
        button { width: 22%; margin: 2px; padding: 8px; font-size: 1em; cursor: pointer; }
        .wide { width: 47%; }
        </style>
    </head>
    <body>
    <div class="calc">
        <input type="text" id="pantalla" readonly>
        <div>
        <button onclick="tecla('7')">7</button>
        <button onclick="tecla('8')">8</button>
        <button onclick="tecla('9')">9</button>
        <button onclick="tecla('/')">÷</button>
    
        <button onclick="tecla('4')">4</button>
        <button onclick="tecla('5')">5</button>
        <button onclick="tecla('6')">6</button>
        <button onclick="tecla('*')">×</button>
    
        <button onclick="tecla('1')">1</button>
        <button onclick="tecla('2')">2</button>
        <button onclick="tecla('3')">3</button>
        <button onclick="tecla('-')">−</button>
    
        <button onclick="tecla('0')">0</button>
        <button onclick="tecla('.')">.</button>
        <button onclick="tecla('+')">+</button>
        <button onclick="limpiar()">C</button>
    
        <button class="wide" onclick="calcular()">=</button>
        <button class="wide" onclick="borrar()">⌫</button>
        </div>
    </div>
    
    <script>
        const pantalla = document.getElementById('pantalla');
    
        function tecla(val) {
        pantalla.value += val;
        }
    
        function calcular() {
        try {
            pantalla.value = eval(pantalla.value);
        } catch {
            pantalla.value = 'Error';
        }
        }
    
        function limpiar() { pantalla.value = ''; }
        function borrar() { pantalla.value = pantalla.value.slice(0, -1); }
    </script>
    </body>
    </html>
    
    Guarda esto como calculadora.html y ábrelo en cualquier navegador — no necesita servidor ni dependencias. Es un buen punto de partida para aprender HTML,
    CSS y JavaScript.

> it’s clearly not the case that these things only make use of Spanish training data when you prompt them in Spanish.

It’s not! And I’ve never said that.

Anyways, I’m not even sure what we are arguing about, as it’s 100% fact that SOTA models perform better in English, the only interesting question here how much better, is it negligible or actually makes a difference in real world use-cases.


It’s negligible as far as I can tell. If the LLM can “speak” the language well then you can prompt it in that language and get more or less the same results as in English.

When you say Claude learned it. That's in the current context window it is able to do that, right? Or is there a more permanent way to make it learn something?

You ask it to read the docs or site and create a 'skill' out of it.

I'm surprised that Opus 4.5 is better than Opus 4.6 and Sonnet 4.6 is even better than Opus 4.5 (and 4.6). Shouldn't Opus 4.6 be the best of the Claude models?


I can’t really tell the difference between the two models for the things I do any more.


Why is it getting expensive when using the phone as a hotspot, but not when using the iPad or MacBook directly with cellular data?


I think some (most?) carriers in the US charge for hotspot traffic separately from direct access from the phone (by looking at packet's TTL, it's lower by 1)


Mmm the Freedom of US Enterprise strikes again. Such a stupid billing rule.


Even in that case, there are tethering apps that can bypass this issue.


That "cult" and "attention to details" was a long time ago, when Steve Jobs was CEO. But even then there were many bugs, like the .DS_Store files, which never got fixed. It got much worse when Tim Cook took over. Now I only use macOS because the next best thing is still worse, and also the Mac hardware became really good with the switch to ARM (after a long long dark time with x86 where the hardware got worse and worse with each iteration).


I wish .DS_Store files were automatically gitignored...

(though why should git fix such a stupid problem it didn't create...)


Pretty easy with a global gitignore


it's not my computer that produces these files... but yes, perhaps MacOS should ship with .DS_Store in a global gitignore by default (though I don't even know if it ships with git at all...)


It does not (unless you install XCode developer tools)


I played and finished Super Bit Boy. :D Reminded me of the game "Thomas Was Alone", that was nice. :)


Aren't most old game like that? Mostly because scrolling was hard back in the days.

The Oxyd games including Esprit and the FOSS version Enigma

The Atari ST game Thriller N.T. and maybe also Shocker and Shocker 2 but not sure on those as I didn't play them in a long time sadly.

Head over Heels

And probably many more, and more popular games, do that, but for some reason my mind comes up with those only.

And maybe even Zork counts? You only "see" (get a description of) one room at a time.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: