• brucethemoose@lemmy.world
    link
    fedilink
    English
    arrow-up
    6
    ·
    2 days ago

    Anyone tested it at high context yet? I find all Mistral models peter out after like 16K-24K tokes no matter what they advertise the context length as.