112 lines
4.3 KiB
Markdown
112 lines
4.3 KiB
Markdown
# Known GitHub Issues for gpt-oss and vLLM
|
|
|
|
## Active Issues
|
|
|
|
### vLLM Repository
|
|
|
|
#### Issue #22519: Token Error with gpt-oss-20b Tool Calls
|
|
- **URL**: https://github.com/vllm-project/vllm/issues/22519
|
|
- **Error**: `Unexpected token 12606 while expecting start token 200006`
|
|
- **Status**: Open, To Triage
|
|
- **Model**: gpt-oss-20b
|
|
- **Symptoms**:
|
|
- Error occurs after model returns token 200012
|
|
- Token 12606 = "comment"
|
|
- Hypothesis: Model incorrectly splitting "commentary" into "comment" + "ary"
|
|
- **Workaround**: None currently documented
|
|
|
|
#### Issue #22515: Same Error, Fixed by Config Update
|
|
- **URL**: https://github.com/vllm-project/vllm/issues/22515
|
|
- **Error**: Same token parsing error
|
|
- **Status**: Open
|
|
- **Fix**: Update generation_config.json from HuggingFace
|
|
- Specific commit: 8b193b0ef83bd41b40eb71fee8f1432315e02a3e
|
|
- User andresC98 confirmed this resolved the issue
|
|
- **Version**: Reported in vLLM v0.10.2
|
|
|
|
#### Issue #22578: gpt-oss-120b Tool Call Support
|
|
- **URL**: https://github.com/vllm-project/vllm/issues/22578
|
|
- **Error**: Chat Completions endpoint tool_call not working
|
|
- **Status**: Open
|
|
- **Model**: gpt-oss-120b
|
|
- **Symptoms**: Tool calling doesn't work correctly via /v1/chat/completions
|
|
|
|
#### Issue #22337: Empty Tool Calls Array
|
|
- **URL**: https://github.com/vllm-project/vllm/issues/22337
|
|
- **Error**: tool_calls returning empty arrays
|
|
- **Status**: Open
|
|
- **Model**: gpt-oss-120b
|
|
- **Symptoms**: Content appears in wrong format, tool_calls=[]
|
|
|
|
#### Issue #23567: Unexpected Tokens in Message Header
|
|
- **URL**: https://github.com/vllm-project/vllm/issues/23567
|
|
- **Error**: `openai_harmony.HarmonyError: unexpected tokens remaining in message header`
|
|
- **Status**: Open
|
|
- **Symptoms**: Occurs in multi-turn conversations with gpt-oss-120b
|
|
- **Version**: vLLM v0.10.1 and v0.10.1.1
|
|
|
|
#### PR #24787: Tool Call Turn Tracking
|
|
- **URL**: https://github.com/vllm-project/vllm/pull/24787
|
|
- **Title**: Pass toolcall turn to kv cache manager
|
|
- **Status**: Merged (September 2025)
|
|
- **Description**: Adds toolcall_turn parameter for tracking turns in tool-calling conversations
|
|
- **Impact**: Enables better prefix cache statistics for tool calling
|
|
|
|
### HuggingFace Discussions
|
|
|
|
#### gpt-oss-20b Discussion #80: Tool Calling Configuration
|
|
- **URL**: https://huggingface.co/openai/gpt-oss-20b/discussions/80
|
|
- **Summary**: Community discussion about tool calling best practices
|
|
- **Key Findings**:
|
|
- Explicit tool listing in system prompt improves results
|
|
- Better results with tool_choice='required' or 'auto'
|
|
- Avoid requiring JSON response format
|
|
- Configuration and prompt engineering significantly impact tool calling behavior
|
|
|
|
#### gpt-oss-120b Discussion #69: Chat Template Spec Errors
|
|
- **URL**: https://huggingface.co/openai/gpt-oss-120b/discussions/69
|
|
- **Summary**: Errors in chat template compared to spec
|
|
- **Impact**: May affect tool calling format
|
|
|
|
### openai/harmony Repository
|
|
|
|
#### Issue #33: EOS Error While Waiting for Message Header
|
|
- **URL**: https://github.com/openai/harmony/issues/33
|
|
- **Error**: `HarmonyError: Unexpected EOS while waiting for message header to complete`
|
|
- **Status**: Open
|
|
- **Context**: Core Harmony parser issue affecting message parsing
|
|
|
|
## Error Pattern Summary
|
|
|
|
### Token Mismatch Errors
|
|
- **Pattern**: `Unexpected token X while expecting start token Y`
|
|
- **Root Cause**: Model generating text tokens instead of Harmony control tokens
|
|
- **Common Triggers**: Tool calling, multi-turn conversations
|
|
- **Primary Fix**: Update generation_config.json
|
|
|
|
### Streaming Errors
|
|
- **Pattern**: Parse failures during streaming responses
|
|
- **Root Cause**: Incompatibility between request format and vLLM token generation
|
|
- **Affected**: Both 20b and 120b models
|
|
|
|
### Tool Calling Failures
|
|
- **Pattern**: Empty tool_calls arrays or text descriptions instead of calls
|
|
- **Root Cause**: Configuration issues or outdated model files
|
|
- **Primary Fix**: Correct vLLM flags and update generation_config.json
|
|
|
|
## Version Compatibility
|
|
|
|
### vLLM Versions
|
|
- **v0.10.2**: Multiple token parsing errors reported
|
|
- **v0.10.1/v0.10.1.1**: Multi-turn conversation errors
|
|
- **Latest**: Check for fixes in newer releases
|
|
|
|
### Recommended Actions by Version
|
|
- **Pre-v0.11**: Update to latest, refresh model files
|
|
- **v0.11+**: Verify tool calling flags are set correctly
|
|
|
|
## Cross-References
|
|
|
|
- Model file updates: See model-updates.md
|
|
- Tool calling configuration: See tool-calling-setup.md
|