Skip to content

feat: add MiniMax as LLM provider#1568

Open
octo-patch wants to merge 1 commit intomicrosoft:mainfrom
octo-patch:feature/add-minimax-provider
Open

feat: add MiniMax as LLM provider#1568
octo-patch wants to merge 1 commit intomicrosoft:mainfrom
octo-patch:feature/add-minimax-provider

Conversation

@octo-patch
Copy link

@octo-patch octo-patch commented Mar 15, 2026

Summary

Add MiniMax as a new LLM provider for JARVIS/HuggingGPT, alongside the existing OpenAI and Azure OpenAI options. MiniMax provides an OpenAI-compatible chat completions API with models offering up to 204K context length.

Changes

  • Provider detection (awesome_chat.py): Added MiniMax to the provider priority chain (local > azure > minimax > openai) with proper API endpoint construction, key resolution (config file or MINIMAX_API_KEY env var), and Bearer token authentication
  • Temperature handling (awesome_chat.py): MiniMax requires temperature in (0.0, 1.0] — zero values are automatically adjusted to 0.01
  • Token management (get_token_ids.py): Added MiniMax-M2.5 and MiniMax-M2.5-highspeed with cl100k_base encoding and 204,800 max context length
  • Configuration template (config.minimax.yaml): New config file for MiniMax setup with all required fields
  • Documentation (README.md): Updated with MiniMax setup instructions and model options

Supported Models

Model Context Length
MiniMax-M2.5 204K
MiniMax-M2.5-highspeed 204K

Usage

cd hugginggpt
python awesome_chat.py --config configs/config.minimax.yaml --mode server

Test Plan

  • Python syntax validation passes for all modified files
  • YAML config validation passes
  • Manual testing with MiniMax API key

Add MiniMax (MiniMax-M2.5, MiniMax-M2.5-highspeed) as a new LLM provider
option alongside OpenAI and Azure OpenAI. MiniMax offers an
OpenAI-compatible API with up to 204K context length.

Changes:
- Add MiniMax provider detection and API endpoint construction in
  awesome_chat.py with priority: local > azure > minimax > openai
- Handle MiniMax temperature constraint (must be > 0) by adjusting
  zero values to 0.01
- Add MiniMax model encodings and context lengths in get_token_ids.py
- Create config.minimax.yaml template for MiniMax configuration
- Update README.md with MiniMax setup instructions
@octo-patch
Copy link
Author

@microsoft-github-policy-service agree

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant