Initial commit
This commit is contained in:
111
skills/gpt-oss-troubleshooting/reference/known-issues.md
Normal file
111
skills/gpt-oss-troubleshooting/reference/known-issues.md
Normal file
@@ -0,0 +1,111 @@
|
||||
# Known GitHub Issues for gpt-oss and vLLM
|
||||
|
||||
## Active Issues
|
||||
|
||||
### vLLM Repository
|
||||
|
||||
#### Issue #22519: Token Error with gpt-oss-20b Tool Calls
|
||||
- **URL**: https://github.com/vllm-project/vllm/issues/22519
|
||||
- **Error**: `Unexpected token 12606 while expecting start token 200006`
|
||||
- **Status**: Open, To Triage
|
||||
- **Model**: gpt-oss-20b
|
||||
- **Symptoms**:
|
||||
- Error occurs after model returns token 200012
|
||||
- Token 12606 = "comment"
|
||||
- Hypothesis: Model incorrectly splitting "commentary" into "comment" + "ary"
|
||||
- **Workaround**: None currently documented
|
||||
|
||||
#### Issue #22515: Same Error, Fixed by Config Update
|
||||
- **URL**: https://github.com/vllm-project/vllm/issues/22515
|
||||
- **Error**: Same token parsing error
|
||||
- **Status**: Open
|
||||
- **Fix**: Update generation_config.json from HuggingFace
|
||||
- Specific commit: 8b193b0ef83bd41b40eb71fee8f1432315e02a3e
|
||||
- User andresC98 confirmed this resolved the issue
|
||||
- **Version**: Reported in vLLM v0.10.2
|
||||
|
||||
#### Issue #22578: gpt-oss-120b Tool Call Support
|
||||
- **URL**: https://github.com/vllm-project/vllm/issues/22578
|
||||
- **Error**: Chat Completions endpoint tool_call not working
|
||||
- **Status**: Open
|
||||
- **Model**: gpt-oss-120b
|
||||
- **Symptoms**: Tool calling doesn't work correctly via /v1/chat/completions
|
||||
|
||||
#### Issue #22337: Empty Tool Calls Array
|
||||
- **URL**: https://github.com/vllm-project/vllm/issues/22337
|
||||
- **Error**: tool_calls returning empty arrays
|
||||
- **Status**: Open
|
||||
- **Model**: gpt-oss-120b
|
||||
- **Symptoms**: Content appears in wrong format, tool_calls=[]
|
||||
|
||||
#### Issue #23567: Unexpected Tokens in Message Header
|
||||
- **URL**: https://github.com/vllm-project/vllm/issues/23567
|
||||
- **Error**: `openai_harmony.HarmonyError: unexpected tokens remaining in message header`
|
||||
- **Status**: Open
|
||||
- **Symptoms**: Occurs in multi-turn conversations with gpt-oss-120b
|
||||
- **Version**: vLLM v0.10.1 and v0.10.1.1
|
||||
|
||||
#### PR #24787: Tool Call Turn Tracking
|
||||
- **URL**: https://github.com/vllm-project/vllm/pull/24787
|
||||
- **Title**: Pass toolcall turn to kv cache manager
|
||||
- **Status**: Merged (September 2025)
|
||||
- **Description**: Adds toolcall_turn parameter for tracking turns in tool-calling conversations
|
||||
- **Impact**: Enables better prefix cache statistics for tool calling
|
||||
|
||||
### HuggingFace Discussions
|
||||
|
||||
#### gpt-oss-20b Discussion #80: Tool Calling Configuration
|
||||
- **URL**: https://huggingface.co/openai/gpt-oss-20b/discussions/80
|
||||
- **Summary**: Community discussion about tool calling best practices
|
||||
- **Key Findings**:
|
||||
- Explicit tool listing in system prompt improves results
|
||||
- Better results with tool_choice='required' or 'auto'
|
||||
- Avoid requiring JSON response format
|
||||
- Configuration and prompt engineering significantly impact tool calling behavior
|
||||
|
||||
#### gpt-oss-120b Discussion #69: Chat Template Spec Errors
|
||||
- **URL**: https://huggingface.co/openai/gpt-oss-120b/discussions/69
|
||||
- **Summary**: Errors in chat template compared to spec
|
||||
- **Impact**: May affect tool calling format
|
||||
|
||||
### openai/harmony Repository
|
||||
|
||||
#### Issue #33: EOS Error While Waiting for Message Header
|
||||
- **URL**: https://github.com/openai/harmony/issues/33
|
||||
- **Error**: `HarmonyError: Unexpected EOS while waiting for message header to complete`
|
||||
- **Status**: Open
|
||||
- **Context**: Core Harmony parser issue affecting message parsing
|
||||
|
||||
## Error Pattern Summary
|
||||
|
||||
### Token Mismatch Errors
|
||||
- **Pattern**: `Unexpected token X while expecting start token Y`
|
||||
- **Root Cause**: Model generating text tokens instead of Harmony control tokens
|
||||
- **Common Triggers**: Tool calling, multi-turn conversations
|
||||
- **Primary Fix**: Update generation_config.json
|
||||
|
||||
### Streaming Errors
|
||||
- **Pattern**: Parse failures during streaming responses
|
||||
- **Root Cause**: Incompatibility between request format and vLLM token generation
|
||||
- **Affected**: Both 20b and 120b models
|
||||
|
||||
### Tool Calling Failures
|
||||
- **Pattern**: Empty tool_calls arrays or text descriptions instead of calls
|
||||
- **Root Cause**: Configuration issues or outdated model files
|
||||
- **Primary Fix**: Correct vLLM flags and update generation_config.json
|
||||
|
||||
## Version Compatibility
|
||||
|
||||
### vLLM Versions
|
||||
- **v0.10.2**: Multiple token parsing errors reported
|
||||
- **v0.10.1/v0.10.1.1**: Multi-turn conversation errors
|
||||
- **Latest**: Check for fixes in newer releases
|
||||
|
||||
### Recommended Actions by Version
|
||||
- **Pre-v0.11**: Update to latest, refresh model files
|
||||
- **v0.11+**: Verify tool calling flags are set correctly
|
||||
|
||||
## Cross-References
|
||||
|
||||
- Model file updates: See model-updates.md
|
||||
- Tool calling configuration: See tool-calling-setup.md
|
||||
Reference in New Issue
Block a user