• notfromhere
    link
    fedilink
    arrow-up
    1
    ·
    9 months ago

    I think this is because of how transformer style LLMs tokenizers work, the models don’t really understand the spelling of the words when asked like that. Karpathy has a great learning video on the GPT2 tokenizer that just came out.