-
Notifications
You must be signed in to change notification settings - Fork 2.4k
feat: add GPT-OSS-120B model to SambaNova provider #8187
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
- Added GPT-OSS-120B to SambaNovaModelId type - Added model configuration with 64K context window - Added test coverage for the new model Fixes #8185
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I only added half the models requested. Classic me.
| | "Llama-4-Maverick-17B-128E-Instruct" | ||
| | "Llama-3.3-Swallow-70B-Instruct-v0.4" | ||
| | "Qwen3-32B" | ||
| | "GPT-OSS-120B" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Issue #8185 mentions that SambaNova has both "deepseek v3.1 and gpt oss 120 b" available. While we've added GPT-OSS-120B, we're missing DeepSeek V3.1.
Looking at other providers in the codebase, they have DeepSeek V3.1 defined:
- Fireworks has it as "deepseek-v3.1"
- Chutes has it as "deepseek-ai/DeepSeek-V3.1"
- Vertex has it as "deepseek-v3.1-maas"
Should we add "DeepSeek-V3.1" here to fully address the issue? The existing "DeepSeek-V3-0324" appears to be a date-versioned model (March 24), not the V3.1 mentioned in the issue.
| }, | ||
| "GPT-OSS-120B": { | ||
| maxTokens: 8192, | ||
| contextWindow: 65536, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Have these specifications (64K context window, /bin/sh.6/.2 pricing) been verified against SambaNova's actual API documentation? It would be good to confirm these values are accurate.
| expect(model.info).toEqual(sambaNovaModels[testModelId]) | ||
| expect(model.info.contextWindow).toBe(65536) | ||
| expect(model.info.description).toContain("GPT-OSS 120B") | ||
| }) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Good test coverage for GPT-OSS-120B! If we add DeepSeek V3.1, we should add a similar test for it as well.
|
solved in #8186 |
This PR attempts to address Issue #8185 by updating the SambaNova model list.
Changes
Notes
The issue also mentioned DeepSeek v3.1, but the codebase already contains several DeepSeek models:
These existing models may already satisfy the DeepSeek requirement. If a specific DeepSeek v3.1 model is needed, please let me know and I can add it.
Testing
Fixes #8185
Feedback and guidance are welcome!
Important
Add
GPT-OSS-120Bmodel to SambaNova provider with test coverage insambanova.tsandsambanova.spec.ts.GPT-OSS-120Btosambanova.tswith 64K context window.GPT-OSS-120Binsambanova.spec.tsto verify model retrieval and properties.This description was created by
for d9e48db. You can customize this summary. It will automatically update as commits are pushed.