-
Notifications
You must be signed in to change notification settings - Fork 3.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
copilot: Correct o3-mini context length #24152
Conversation
This comment was marked as resolved.
This comment was marked as resolved.
This comment was marked as resolved.
This comment was marked as resolved.
This comment was marked as resolved.
This comment was marked as resolved.
Ah, I didn't find that, but appreciate the context. I haven't tested the API, if they are limiting it to that then obviously close the PR. Wish they publicly posted what their limits were regardless. |
That was for o1-mini, so the question is whether o3-mini has a higher token count and what that is. It looks like there is an API endpoint retrieve the information: #20362 (comment) It doesn't seem to work unauthenticated for me, so I'd need to figure out how to auth against it. |
Sadly my work account for copilot doesn't give us access to the new models right now, later tonight I can see if the free version of copilot lets you use o3-mini and if the context is higher. |
Can confirm 200k input and 100k output tokens, according to the
I have not personally tried to input 200k tokens, however. |
Thanks! |
Hm. I've just sniffed the request made by Copilot in VSCode, and I get this:
Namely, for
So indeed, it looks like only 20k tokens can be sent in the prompt. Assuming Copilot does even allow a full 200k to be sent (as in, assuming it's not restricting beyond the model's capabilities), I presume it's not directly via the chat context. I can see file contexts are being chunked and sent to a separate API endpoint, which I presume is the 'correct' way to use the entire available token window. I'll try look into it a bit more. I presume this change should be reverted though. My apologies the first time -- it seems the different methods of calling Copilot give different results and limits. |
Reverts #24152 See comment: #24152 (comment) Manually confirmed >20k generates error.
Reverts zed-industries#24152 See comment: zed-industries#24152 (comment) Manually confirmed >20k generates error.
Thanks @SirSilver and @itsaphel for getting to the bottom of this, it is unfortunate that copilot doesn't actually document this. It is an interesting distinction though, where you can potentially go up to 200k tokens but only send up to 20k at a time. I wonder if it considers cached tokens (things it's seen before like openai). |
It should be 200k (with 100k output). I can't find anything that puts it at 20k and the changeover in 2f82374 only changed the name from o1-mini to o3-mini
References:
Release Notes: