I’ve been using airoboros-l2-70b for writing fiction, and while overall I’d describe the results as excellent and better than any llama1 model I’ve used, it doesn’t seem to be living up to the promise of 4k token sequence length.

Around 2500 tokens output quality degrades rapidly, and either starts repeating previous text verbatim, or becomes incoherent (grammar, punctuation and capitalization disappear, becomes salad of vaguely related words)

Any other experiences with llama2 and long context? Does the base model work better? Are other fine tunes behaving similarly? I’ll try myself eventually, but the 70b models are chunky downloads, and experimentation takes a while at 1 t/s.

(I’m using GGML Q4_K_M on kobold.cpp, with rope scaling off like you’re supposed to do with llama2)

  • creolestudios@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    1
    ·
    8 months ago

    Yes, the 4k context length of Llama2 is indeed real. Llama2 is a cutting-edge language model developed by OpenAI, and its impressive capability to understand and generate text with such a lengthy context is one of its remarkable features. If you’re interested in leveraging advanced AI models like Llama2 for chatbot development or other applications, you may consider reaching out to an AI chatbot development company for assistance in harnessing this technology effectively.