Skip to content

evolsb/ask-intercom

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

90 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Ask-Intercom

Your always-on customer insights engine

Transform customer conversations into product decisions. Start with Intercom analysis today, evolve into your complete customer intelligence platform tomorrow. Ask natural language questions like "show me issues from last week" and get structured insights with customer details.

🚀 Quick Start

🐳 Docker (One Command - Recommended) ✅

# Clone and setup
git clone https://github.com/your-username/ask-intercom
cd ask-intercom
cp .env.example .env
# Edit .env with your API keys

# Run with Docker
docker-compose up

# Access at http://localhost:8000

🌐 Development (Local)

# Start backend and frontend
env -i HOME="$HOME" PATH="$PATH" ~/.local/bin/poetry run uvicorn src.web.main:app --port 8000 --reload &
cd frontend && npm run dev &

# Open http://localhost:5173

💻 CLI Usage

# Setup environment  
~/.local/bin/poetry install
cp .env.example .env
# Edit .env with your API keys

# Ask questions
env -i HOME="$HOME" PATH="$PATH" ~/.local/bin/poetry run python -m src.cli "What are the top customer complaints this month?"

✨ Features

  • 🌐 Web interface with real-time progress tracking
  • 🤖 AI-powered analysis using OpenAI GPT-4
  • Natural language queries ("show me issues from last week")
  • 🎯 Structured insights with customer details and priorities
  • 🔗 Direct links to Intercom conversations
  • ⚙️ Optional conversation limits (user-controlled via Settings)
  • 💰 Cost tracking and optimization
  • 🚀 MCP Architecture with FastIntercom for 400x speedup
  • 🔄 Graceful fallbacks (MCP → FastIntercom → Local → REST)

📋 Requirements

  • Python 3.13+ and Poetry
  • Node.js and npm (for web interface)
  • Intercom access token
  • OpenAI API key

⚡ MCP Configuration (Optional Performance Boost)

Enable MCP for 400x faster cached queries:

# Add to your .env file
ENABLE_MCP=true
MCP_BACKEND=fastintercom  # or 'official', 'local', 'auto'

Backends:

  • fastintercom - High-performance caching with SQLite (400x speedup)
  • official - Standard Intercom MCP server
  • local - Local development MCP server
  • auto - Automatically choose best available backend

📚 Documentation

Core Documentation

MCP Architecture

🛠️ Development

See CLAUDE.md for Claude Code specific guidance and docs/ for full documentation.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •