Poking around the network requests for ChatGPT, I’ve noticed the /backend-api/models response includes information for each model, including the maximum tokens.

For me:

  • GPT-3.5: 8191
  • GPT-4: 4095
  • GPT-4 with Code Interpreter: 8192
  • GPT-4 with Plugins: 8192

It seems to be accurate. I’ve had content that is too long for GPT-4, but is accepted by GPT-4 with Code Interpreter. The quality feels about the same, too.

Here’s the response I get from /backend-api/models, as a Plus subscriber:

{
    "models": [
        {
            "slug": "text-davinci-002-render-sha",
            "max_tokens": 8191,
            "title": "Default (GPT-3.5)",
            "description": "Our fastest model, great for most everyday tasks.",
            "tags": [
                "gpt3.5"
            ],
            "capabilities": {}
        },
        {
            "slug": "gpt-4",
            "max_tokens": 4095,
            "title": "GPT-4",
            "description": "Our most capable model, great for tasks that require creativity and advanced reasoning.",
            "tags": [
                "gpt4"
            ],
            "capabilities": {}
        },
        {
            "slug": "gpt-4-code-interpreter",
            "max_tokens": 8192,
            "title": "Code Interpreter",
            "description": "An experimental model that can solve tasks by generating Python code and executing it in a Jupyter notebook.\nYou can upload any kind of file, and ask model to analyse it, or produce a new file which you can download.",
            "tags": [
                "gpt4",
                "beta"
            ],
            "capabilities": {},
            "enabled_tools": [
                "tools2"
            ]
        },
        {
            "slug": "gpt-4-plugins",
            "max_tokens": 8192,
            "title": "Plugins",
            "description": "An experimental model that knows when and how to use plugins",
            "tags": [
                "gpt4",
                "beta"
            ],
            "capabilities": {},
            "enabled_tools": [
                "tools3"
            ]
        },
        {
            "slug": "text-davinci-002-render-sha-mobile",
            "max_tokens": 8191,
            "title": "Default (GPT-3.5) (Mobile)",
            "description": "Our fastest model, great for most everyday tasks.",
            "tags": [
                "mobile",
                "gpt3.5"
            ],
            "capabilities": {}
        },
        {
            "slug": "gpt-4-mobile",
            "max_tokens": 4095,
            "title": "GPT-4 (Mobile, V2)",
            "description": "Our most capable model, great for tasks that require creativity and advanced reasoning.",
            "tags": [
                "gpt4",
                "mobile"
            ],
            "capabilities": {}
        }
    ],
    "categories": [
        {
            "category": "gpt_3.5",
            "human_category_name": "GPT-3.5",
            "subscription_level": "free",
            "default_model": "text-davinci-002-render-sha",
            "code_interpreter_model": "text-davinci-002-render-sha-code-interpreter",
            "plugins_model": "text-davinci-002-render-sha-plugins"
        },
        {
            "category": "gpt_4",
            "human_category_name": "GPT-4",
            "subscription_level": "plus",
            "default_model": "gpt-4",
            "code_interpreter_model": "gpt-4-code-interpreter",
            "plugins_model": "gpt-4-plugins"
        }
    ]
}

Anyone seeing anything different? I haven’t really seen this compared anywhere.

  • Anony Moose@lemmy.ca
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    Tangentially, I think I read about a 32k token version of GPT4 that was supposed to roll out. Did anybody get access to that?