Kilo for Slack: AI Coding Agents Where Engineering Decisions Actually Happen
Published on 16.01.2026
Kilo for Slack: AI Coding Agents Where Engineering Decisions Actually Happen
TLDR: Kilo launched a Slack integration that can read conversation context, access GitHub repositories, implement fixes, and create PRs directly from Slack threads. It uses MiniMax M2.1 as the default model, available free for the first week.
Engineering teams don't make decisions in IDE sidebars. They make them in Slack. As startups have moved away from siloed workflows, Slack has become the central operating system where critical decisions happen and work gets done. Yet most AI coding tools still force developers to leave context behind.
The pattern is familiar: someone reports a bug in Slack, the team discusses potential causes, then you Alt+Tab to your IDE and explain the whole situation over again. You wait for the fix, push to GitHub, hop back to Slack to share the PR link. This happens dozens of times daily, and context switching is costly.
Kilo for Slack aims to eliminate this friction. Mention @Kilo in a thread, reference a bug discussion, and ask it to implement a fix. The bot reads the entire Slack thread, pulls your GitHub repo context, spins up a cloud agent, implements the fix, and pushes a PR. No copy-pasting, no context switching.
The technical differentiation is worth examining. Unlike Cursor's Slackbot, which is limited to a single repository and requires manual configuration, Kilo infers the relevant repository from conversation and works across multiple repos. Unlike Claude Code's Slack integrations designed for one-shot interactions, Kilo builds on existing threads and PRs, allowing conversations to evolve as work progresses.
The model choice is interesting. Kilo chose MiniMax M2.1 as the default—an open-weight model from a lab that just had a successful Hong Kong IPO. The argument is that 2025 was the year OSS models caught up to proprietary ones, reducing performance gaps from 8% to 1.7% on key benchmarks. As model quality converges, cost, flexibility, and control matter more.
For teams and architects, this represents a shift in how AI integrates into workflows. The tool goes where the conversation already is rather than requiring developers to bring context to the tool. The pricing is straightforward usage-based with 450+ model choices. Setup takes about two minutes: connect GitHub repos, add the Slack integration, start mentioning @Kilo.
The four main use cases: asking questions about your codebase, debugging issues on the fly with stack traces, implementing fixes from Slack discussions, and pushing PRs without context switching. Each is triggered by mentioning @Kilo in a thread with appropriate context.
The broader thesis is that AI coding shouldn't be trapped in a single interface. Kilo already exists in VS Code, JetBrains, and CLI. Slack is where teams discuss code before anyone opens an editor, so having AI present in that conversation removes friction. The tool fits into how engineers already work rather than demanding they change workflows.
Key takeaways:
- Kilo for Slack creates PRs and implements fixes directly from Slack conversation context
- Multi-repository support infers relevant repos from conversation—no manual configuration
- Uses MiniMax M2.1 open-weight model by default, free for the first week
- Continuous conversations build on threads and existing PRs, not one-shot interactions
- 450+ models available with usage-based pricing
Tradeoffs:
- Gain zero context switching but need robust GitHub integration and permissions setup
- Get AI where decisions happen but must trust the bot's repository inference accuracy
- Use open-weight models for cost savings but may sacrifice some edge-case performance vs. frontier models
Link: Announcing Kilo for Slack
This article was generated from newsletter content. For the original source and to subscribe, visit the links above.