Thanks to visit codestin.com
Credit goes to github.com

Skip to content

qcrao/copilot

Repository files navigation

Roam Copilot

Watch the video 👆 Click to watch the demo video

🚀 Transform your Roam Research experience with AI-powered insights

Roam Copilot brings the power of advanced AI directly into your Roam workspace. Get instant, context-aware assistance for brainstorming, analysis, and idea connections without ever leaving your graph.

Key Highlights:

  • 🧠 Context-Aware Intelligence - Automatically understands your current page content
  • 🏠 Privacy-First with Local AI - Full support for local models via Ollama
  • Multi-Provider Support - OpenAI, Anthropic, Grok, DeepSeek, Qwen, Gemini, Groq, XAI, Ollama, and custom models
  • 🎯 User-friendly Interface - Resizable window on the right side of your main workspace

✨ What Makes Roam Copilot Special

🎯 Smart Context Understanding

  • Automatically reads your current page content
  • Uses visible blocks, including sidebar notes, as conversation context
  • No manual copy-pasting required

🏠 Privacy & Control

  • Local AI Models - Run powerful models privately via Ollama. Your data stays where you want it
  • Multiple Providers - OpenAI, Anthropic, Grok, or custom endpoints

💬 Intuitive Chat Experience

  • User-friendly Interface - Draggable, resizable, and minimizable
  • Conversation History - Never lose your AI conversations
  • Quick Templates - Pre-built prompts for common workflows
  • Command Palette - Keyboard shortcuts for power users
  • Smart Input Features - @ to search pages, drag & drop blocks, / for templates

📝 Powerful Templates

  • Official Templates - Pre-built prompts for common workflows

    • Insights
    • Writing assistance
  • Custom Templates - Create your own templates

    • Define custom prompts
    • Add descriptions and icons
    • Organize by categories

🚀 Quick Start

Installation from Roam Depot

  1. Install from Roam Depot - Search for "Roam Copilot" in your Roam Research extensions
  2. Configure Your AI Provider - Add API keys for OpenAI, Anthropic, or Grok in settings
  3. Start Chatting - Click the lightbulb icon in the bottom-right corner

Prerequisites

Using Local Models (Ollama)

To use Roam Copilot with a local Ollama instance, you must configure Ollama to accept requests from Roam Research. This is a critical security step required by Ollama. Here is a detailed guide to get it working:

1. Check if Ollama is Running

Open your terminal and run the following command. You should see a response from the Ollama server.

curl http://localhost:11434/

2. List Your Installed Models

Verify which models you have available locally.

ollama list
# or via the API
curl http://localhost:11434/api/tags

3. Set the Allowed Origin for Roam Research

This is the most important step. You need to tell Ollama to accept connections from https://roamresearch.com.

One-Click Setup Script (Recommended)

We provide scripts that automatically set up, test, and clean up CORS configuration. You can either download them from the repository or use our hosted versions:

Download and run the setup script:

# Option 1: Download from public URL
curl -fsSL https://public.qcrao.com/rr-copilot/setup_ollama_cors.sh | bash

This script will:

  • Add the environment variable to your shell configuration file (.zshrc, .bash_profile, etc.)
  • Create a persistent launchd configuration on macOS
  • Restart Ollama automatically
  • Verify the CORS configuration

To test if CORS is working correctly:

# Download and run the test script
curl -fsSL https://public.qcrao.com/rr-copilot/test_ollama_cors.sh | bash

To remove the CORS configuration later:

# Download and run the cleanup script
curl -fsSL https://public.qcrao.com/rr-copilot/cleanup_ollama_cors.sh | bash

💬 How to Use

Getting Started

  • 💡 Click the lightbulb icon in the bottom-right corner, or
  • ⌨️ Use Command Palette (Cmd+P / Ctrl+P) → Search "Roam Copilot"

Chat Interface

  • 🔄 Auto-Context - Your current page content is automatically included
  • 📚 Chat History - Access previous conversations from the sidebar
  • ➕ New Conversations - Start fresh chats anytime
  • 🔽 Minimize/Expand - Keep it out of the way when not needed

Settings

You can configure the AI models and other settings from the settings panel.

Contributing

Contributions are welcome! Please feel free to submit a pull request or open an issue.

License

This project is licensed under the MIT License.

About

Roam Research Copilot

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Sponsor this project

 

Packages

No packages published