• Zak@lemmy.world
    link
    fedilink
    arrow-up
    8
    ·
    18 hours ago

    LLMs don’t understand things. They repeatedly predict the next token given previous tokens.

    I don’t think something without predictable patterns is likely to work as a language. A very complex grammar probably means the LLM will make grammatical errors more often, but that’s probably the most that can be done to make a language hard for LLMs. Other comments mention languages without much training data, but I don’t think that’s what you’re asking.

  • howrar@lemmy.ca
    link
    fedilink
    arrow-up
    18
    ·
    23 hours ago

    The difficulty in training an AI is dependent on data availability. So this is just a question of choosing a language that has the least amount of writing. You can trivially choose any language that doesn’t have a writing system at all and invent a writing system for it. But then you’d also run into the problem of learning the language yourself.

  • disregardable@lemmy.zip
    link
    fedilink
    arrow-up
    14
    ·
    23 hours ago

    The AI doesn’t understand anything in any language. It’s just putting words in order by their association.

  • SillyGooseQuacked@lemmy.world
    link
    fedilink
    arrow-up
    3
    ·
    edit-2
    20 hours ago

    Not to be a downer if you’re anti-AI, but you should know a functional, small, 1B parameter model only needs ~85GB of data if the training data set is high quality (the four-year old chinchilla paper set out the 20 to 1 optimization rule for ai training, so it may require even less today).

    That’s basically nothing. If a language has over ~130,000 books or an equivalent amount of writing (1,500 books is about a gig in plain ascii), a functional text-based ai model could be built that uses it.

    My understanding is there are next to zero languages in existence today that do not have this amount of quality text. Certainly, spoken languages that have no written word are not accessible like this, but most endangered languages with few speakers that have a historical written word could in theory have ai models built that effectively communicate in those languages.

    To give you an idea of what this means for less-written languages and a website revolving around them, look at worldcat (which does NOT have anywhere near most of the written text available entirely online for each language listed, it’s JUST a resource for libraries): https://www.oclc.org/en/worldcat/inside-worldcat.html

    But this gets even harder for a theoretical website used to avoid an LLM that can read it, because this is all assuming creating an ai model for language from scratch. That is not necessary today because of transfer learning.

    Major LLM models with over 100 diverse major languages can be fined-tuned on an insignificant amount of data (even 1GB could work in theory) and produce results like those of a 1B parameter model trained solely on one language. This is because the multi-lingual models developed cross-cultural vector-based understandings of Grammer.

    In truth, the only remaining major barriers for any language not understood by fine-tuning an ai model today are both (1) digitization and (2) character recognition. Digitization will vanish as an issue for basically every written language that has a unique script within the next ten years. Character recognition (and more specifically, the economic viability of building the character recognition) will be the only remaining issue.

    Ironically, in creating such a website, you will be creating more data for a future potential ai model to use in training. Especially if whatever you write makes the language of greater economic importance.

  • Axolotl@feddit.it
    link
    fedilink
    arrow-up
    4
    ·
    23 hours ago

    Probably a dialect or a dead language could be, since they have little to none written texts to train AI on

  • theneverfox@pawb.social
    link
    fedilink
    English
    arrow-up
    2
    ·
    20 hours ago

    You can literally train them on animal languages. They’re language models, their primary attribute is modeling a language

    And they are very good at it

  • tal@lemmy.today
    link
    fedilink
    English
    arrow-up
    3
    ·
    23 hours ago

    https://en.wikipedia.org/wiki/Linear_A

    Linear A is a writing system that was used by the Minoans of Crete from 1800 BC to 1450 BC. Linear A was the primary script used in palace and religious writings of the Minoan civilization. It evolved into Linear B, which was used by the Mycenaeans to write an early form of Greek. It was discovered by the archaeologist Sir Arthur Evans in 1900. No texts in Linear A have yet been deciphered.

  • Zwuzelmaus@feddit.org
    link
    fedilink
    arrow-up
    2
    ·
    edit-2
    23 hours ago

    All languages that are spoken, but not written.

    Current AI models are trained on written stuff only.

  • CombatWombatEsq@lemmy.world
    link
    fedilink
    arrow-up
    1
    ·
    20 hours ago

    I suppose you could create one: https://en.wikipedia.org/wiki/Constructed_language

    It would only stop LLMs until the corpus grew large enough for someone to train a model, but that would be quite a long time, given the uptake on previous conlangs.

    You can achieve much the same effect with slang, though — LLMs are trained on old data, so if every human starts referring to blindats all of a sudden, the LLMs won’t catch on until there is a critical mass of text that uses that slang.