About
Kibela Integration Server enables AI assistants to search, read, and manage notes in Kibela, a team knowledge management platform popular in Japan for documentation and internal wikis. Key features: - Search Kibela notes by keywords across your workspace - Retrieve full note content and comments by ID or path - Fetch your personal notes and recent activity - Create new notes directly from AI assistants - Update existing note content by note ID The integration uses Kibela's GraphQL API with team-based authentication via access tokens.
README
mcp-kibela 🗒️
[](https://smithery.ai/server/@kj455/mcp-kibela) [](https://www.npmjs.com/package/@kj455/mcp-kibela) [](https://opensource.org/licenses/MIT)
A Model Context Protocol (MCP) server implementation that enables AI assistants to search and reference Kibela content. This setup allows AI models like Claude to securely access information stored in Kibela.
Features 🚀
The mcp-kibela server provides the following features:
---
Prerequisites 📋
Before you begin, ensure you have:
Installation 🛠️
Usage with Cursor
{
"kibela": {
"command": "docker",
"args": [
"run",
"-i",
"--rm",
"-e",
"KIBELA_TEAM",
"-e",
"KIBELA_TOKEN",
"ghcr.io/kj455/mcp-kibela:latest"
],
"env": {
"KIBELA_TEAM": "your-team-name from https://[team-name].kibe.la",
"KIBELA_TOKEN": "your-token"
}
}
}
Usage with VSCode
{
"mcp": {
"inputs": [
{
"type": "promptString",
"id": "kibela_team",
"description": "Kibela team name",
"password": false
},
{
"type": "promptString",
"id": "kibela_token",
"description": "Kibela token",
"password": true
},
],
"servers": {
"kibela": {
"command": "docker",
"args": [
"run",
"-i",
"--rm",
"-e",
"KIBELA_TEAM",
"-e",
"KIBELA_TOKEN",
"ghcr.io/kj455/mcp-kibela:latest"
],
"env": {
"KIBELA_TEAM": "${input:kibela_team}",
"KIBELA_TOKEN": "${input:kibela_token}"
}
}
}
}
}
Usage with Claude Desktop
{
"mcpServers": {
"mcp-kibela": {
"command": "docker",
"args": [
"run",
"-i",
"--rm",
"-e",
"KIBELA_TEAM",
"-e",
"KIBELA_TOKEN",
"ghcr.io/kj455/mcp-kibela:latest"
],
"env": {
"KIBELA_TEAM": "your-team-name from https://[team-name].kibe.la",
"KIBELA_TOKEN": "your-token"
}
}
}
}
Using Smithery
npx -y @smithery/cli install @kj455/mcp-kibela --client claude
Environment Variables
The following environment variables are required:
KIBELA_TEAM: Your Kibela team name (required). You can find it from the URL of your Kibela team page. e.g. https://[team-name].kibe.laKIBELA_TOKEN: Your Kibela API token (required)Contributing
Any contributions are welcome!
Development
1. Use npm run build:watch to build the project in watch mode.
npm run build:watch
2. Use npx @modelcontextprotocol/inspector to inspect the MCP server.
npx @modelcontextprotocol/inspector node /path/to/mcp-kibela/dist/index.js
License 📄
MIT
Related MCP Servers
AI Research Assistant
hamid-vakilzadeh
AI Research Assistant provides comprehensive access to millions of academic papers through the Semantic Scholar and arXiv databases. This MCP server enables AI coding assistants to perform intelligent literature searches, citation network analysis, and paper content extraction without requiring an API key. Key features include: - Advanced paper search with multi-filter support by year ranges, citation thresholds, field of study, and publication type - Title matching with confidence scoring for finding specific papers - Batch operations supporting up to 500 papers per request - Citation analysis and network exploration for understanding research relationships - Full-text PDF extraction from arXiv and Wiley open-access content (Wiley TDM token required for institutional access) - Rate limits of 100 requests per 5 minutes with options to request higher limits through Semantic Scholar
Linkup
LinkupPlatform
Linkup is a real-time web search and content extraction service that enables AI assistants to search the web and retrieve information from trusted sources. It provides source-backed answers with citations, making it ideal for fact-checking, news gathering, and research tasks. Key features of Linkup: - Real-time web search using natural language queries to find current information, news, and data - Page fetching to extract and read content from any webpage URL - Search depth modes: Standard for direct-answer queries and Deep for complex research across multiple sources - Source-backed results with citations and context from relevant, trustworthy websites - JavaScript rendering support for accessing dynamic content on JavaScript-heavy pages
Math-MCP
EthanHenrickson
Math-MCP is a computation server that enables Large Language Models (LLMs) to perform accurate numerical calculations through the Model Context Protocol. It provides precise mathematical operations via a simple API to overcome LLM limitations in arithmetic and statistical reasoning. Key features of Math-MCP: - Basic arithmetic operations: addition, subtraction, multiplication, division, modulo, and bulk summation - Statistical analysis functions: mean, median, mode, minimum, and maximum calculations - Rounding utilities: floor, ceiling, and nearest integer rounding - Trigonometric functions: sine, cosine, tangent, and their inverses with degrees and radians conversion support