Kilo Code is an open source AI coding agent for VS Code. It generates code from natural language, automates tasks, runs terminal commands, and supports inline autocomplete.
Prerequisites
Configuration
Step 1: Open Kilo Code Settings
- Open VS Code
- Click the Kilo Code icon in the sidebar
- Open the settings panel
Step 2: Select Provider
- In the API Provider dropdown, select OpenAI Compatible
- Enter the base URL:
https://api.infercom.ai/v1
- Enter your Infercom API key
- Enter the model name:
MiniMax-M2.5
Step 3: Model Configuration
In the Model Configuration section, set:
| Setting | Value |
|---|
| Context Window | 163840 |
| Max Output Tokens | 16384 |
| Image Support | Disabled |
Image Support must be disabled - MiniMax-M2.5 doesn’t support image input.
Step 4: Verify Connection
Start a conversation to verify the setup works.
Model
Use MiniMax-M2.5 - optimized for agentic coding with 160K context and 75.8% SWE-bench.
Usage
With Kilo Code configured:
- Click the Kilo Code icon in VS Code sidebar
- Select a mode: Architect, Coder, or Debugger
- Type your request in the chat
- Kilo Code will autonomously:
- Read relevant files
- Write or edit code
- Run terminal commands
- Iterate until the task is complete
Modes
| Mode | Best For |
|---|
| Architect | Planning, design decisions |
| Coder | Writing and editing code |
| Debugger | Finding and fixing bugs |
Example Tasks
- “Add error handling to the login function”
- “Write unit tests for the User class”
- “Refactor this file to use async/await”
Troubleshooting
Connection Failed
Verify your configuration:
- Base URL:
https://api.infercom.ai/v1
- API key is valid
- Model name is exact:
MiniMax-M2.5
Invalid API Key Error
Double-check your Infercom API key is entered correctly. Test it:
curl -s https://api.infercom.ai/v1/models \
-H "Authorization: Bearer your-infercom-api-key"
Model Not Found
Ensure the model name matches exactly (case-sensitive): MiniMax-M2.5
Slow Responses
MiniMax-M2.5 runs at 400+ tokens/sec. If responses seem slow:
- Check your network connection
- Large context (many files) increases processing time
- First request may be slower due to model loading
Next Steps