Advanced AI Assistant Chrome Extension with Ollama Integration
A powerful Chrome extension that brings the full capabilities of Ollama directly to your browser sidebar. Features intelligent model switching, advanced tool calling, vision support, and seamless web integration.
π Live Demo: Install the extension to experience the full feature set with your local Ollama models!
- Multi-Model Support: Seamlessly switch between text, vision, and thinking models
- Auto-Configuration: Intelligent tool call management based on model capabilities
- Thinking Models: Support for reasoning models like DeepSeek-R1, Qwen2.5-Coder
- Vision Models: Process images with LLaVA, Qwen2-VL, and other multimodal models
- Web Search: Real-time search with Serper API integration
- Page Context: Extract and analyze current webpage content
- Screenshot Analysis: Capture and analyze screenshots with vision models
- Smart Disable: Automatically disables tools for vision models to prevent conflicts
- Streaming Responses: Real-time response generation with typing indicators
- Rich Media Support: Drag & drop images, paste from clipboard
- Context Awareness: Optional webpage context integration
- Message History: Persistent conversation storage
- Error Prevention: Smart duplicate error handling
- Model Detection: Automatic capability detection (tools, vision, thinking)
- Auto-Management: Smart tool call enabling/disabling based on model type
- User Override: Full manual control when needed
- Settings Sync: Real-time synchronization across extension components
- Chrome Browser (Version 88+)
- Ollama installed and running locally
- At least one Ollama model pulled (e.g.,
ollama pull qwen2.5:7b)
-
Download Extension
# Download the latest release wget https://github.com/xsploit/SideLLAMA/releases/latest/download/SideLlama-v2.0.0-final-complete.zip -
Install in Chrome
- Open Chrome and navigate to
chrome://extensions/ - Enable "Developer mode" (top right toggle)
- Click "Load unpacked"
- Select the extracted SideLlama folder
- Open Chrome and navigate to
-
Configure Ollama
- Ensure Ollama is running:
ollama serve - Default endpoint:
http://localhost:11434 - Pull some models:
ollama pull qwen2.5:7b
- Ensure Ollama is running:
-
Open SideLlama
- Click the SideLlama icon in your toolbar
- Or use keyboard shortcut:
Ctrl+Shift+O
- Open the sidebar panel
- Select your preferred model from the dropdown
- Type your message and press Enter
- Enjoy streaming responses!
- Switch to a vision model (e.g.,
llava,qwen2-vl:7b) - Drag & drop an image or paste from clipboard
- Add your question about the image
- Send for analysis
- Enable web search in settings
- Add your Serper API key (optional, for enhanced search)
- Use search directly in chat or via context menu
- AI will analyze search results automatically
- Select a thinking model (e.g.,
deepseek-r1,qwen2.5-coder) - Enable "Thinking Mode" in settings
- Ask complex questions requiring reasoning
- Watch the model's thought process unfold
Access via extension options or the settings button in the sidebar.
- Ollama URL: Local Ollama endpoint (default:
http://localhost:11434) - Default Model: Your preferred model for new conversations
- System Prompt: Custom instructions for the AI assistant
- Enable Tool Calls: Allow AI to use web search and other tools
- Auto-Manage Tool Calls: Automatically enable/disable based on model type
- Enable Thinking Mode: Show reasoning process for thinking models
- Streaming: Real-time response generation (recommended)
- Context Length: Maximum conversation context (default: 128k)
- Temperature: Response creativity (0.0-1.0)
- Keep Alive: Model memory duration
- Performance Stats: Show response timing and token usage
Ctrl+Shift+O: Open SideLlama panelCtrl+Shift+S: Summarize current pageCtrl+Shift+E: Explain selected textEnter: Send messageShift+Enter: New line in message
SideLlama/
βββ manifest.json # Extension manifest (V3)
βββ service-worker.js # Background service worker
βββ sidepanel.js # Main UI logic
βββ sidepanel.html # Sidebar interface
βββ settings.js # Settings management
βββ settings.html # Settings page
βββ content-script.js # Page content extraction
βββ shared-utils.js # Shared utility functions
βββ model-utils.js # Model capability detection
βββ icons/ # Extension icons
# Clone the repository
git clone https://github.com/xsploit/SideLLAMA.git
cd SideLLAMA
# Install dependencies (if any)
npm install
# Load in Chrome for development
# Open chrome://extensions/, enable Developer mode, click "Load unpacked"- Fork the repository
- Create a feature branch:
git checkout -b feature/amazing-feature - Commit changes:
git commit -m 'Add amazing feature' - Push to branch:
git push origin feature/amazing-feature - Open a Pull Request
qwen2.5:7b- General purpose, tool callingllama3.1:8b- Meta's latest, tool supportmistral-nemo- Mistral's efficient modelphi3.5- Microsoft's compact model
llava:7b- Leading vision-language modelqwen2-vl:7b- Alibaba's vision modelllama3.2-vision:11b- Meta's multimodal modelmoondream- Lightweight vision model
deepseek-r1:7b- Advanced reasoning modelqwen2.5-coder- Code-focused reasoningo1-preview- OpenAI-style reasoning (if available)
Models that support function calling:
qwen2.5:*- Full tool supportllama3.1:*- Function callingfirefunction-v2- Specialized for toolscommand-r- Cohere's tool model
- Local Processing: All AI processing happens locally via Ollama
- No Data Collection: Extension doesn't collect or transmit personal data
- Optional Web Search: External API calls only when explicitly enabled
- Secure Communication: All local requests use localhost/127.0.0.1
- Content Script Isolation: Minimal content script with limited permissions
"Unable to connect to Ollama"
- Ensure Ollama is running:
ollama serve - Check if port 11434 is accessible
- Verify firewall settings
"Model does not support tools" Error
- Enable "Auto-Manage Tool Calls" in settings
- Or manually disable tool calls for vision models
Images not displaying properly
- Ensure you're using a vision-capable model
- Check image format (JPEG, PNG supported)
- Try smaller image sizes if processing is slow
Streaming responses stop working
- Refresh the extension (disable/enable in chrome://extensions/)
- Check Ollama server status
- Restart Ollama service
- Use smaller models for faster responses (7B vs 70B)
- Enable "Keep Alive" for frequently used models
- Adjust context length based on your needs
- Use tool calling only when necessary
- β Smart model auto-configuration
- β Thinking mode toggle for reasoning models
- β Enhanced error handling and duplicate prevention
- β Code deduplication and performance improvements
- β Vision model tool call auto-disable
- β Shared utility system
- Basic Ollama integration
- Sidebar chat interface
- Model switching
- Basic tool calling
- Issues: GitHub Issues
- Discussions: GitHub Discussions
- Documentation: Wiki
- Releases: Latest Releases
This project is licensed under the MIT License - see the LICENSE file for details.
- Ollama for the amazing local AI platform
- LLaVA for pioneering vision-language models
- Qwen Team for excellent multimodal models
- Chrome Extensions team for Manifest V3 framework
Built with β€οΈ for the open-source AI community
SideLlama - Bringing the power of local AI to your browser, one conversation at a time.
