Skip to content

A collection of guidance for terminal-based AI coding tools (Qwen Code, Aider,Copilot CLI, and Gemini CLI) and API provider integration instructions.

BNLNPPS/terminal-ai-toolkit

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

9 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

πŸ› οΈ Terminal AI Toolkit

Supercharge your terminal with AI-powered tools and scripts

Awesome Scripts


πŸ“‘ Table of Contents


🌟 Overview

This repository is your comprehensive guide to getting the most out of AI tools in your terminal. It contains curated scripts, expert tips, and detailed guides for terminal-based AI development.

πŸ’‘ Pro Tip: This is a companion to the awesome-terminals-ai listβ€”your one-stop resource for terminal AI tools!


πŸ“œ Scripts

Useful scripts to enhance your AI terminal workflow:

Script Description Guide
πŸ“Š copilot-usage.sh Check your GitHub Copilot usage and quota Copilot CLI Guide
πŸ€– run-claude-copilot.sh Run Claude Code with GitHub Copilot models See below ⬇️

πŸ†“ Free API Providers

πŸ’Ž Gemini API

Access powerful Google Gemini models with generous free tier limits:

Feature Free Tier
πŸš€ Model Gemini 2.5 Pro
⚑ Rate Limit 5 requests/minute
πŸ“… Daily Limit 100 requests/day

πŸ™ GitHub Models

GitHub provides two types of AI model access for developers:

πŸ€– GitHub Copilot Models

Overview:

  • 🌐 Endpoint: https://api.githubcopilot.com
  • πŸ“– Documentation: Supported Models
  • ⚑ Rate Limits: 300 premium requests/month (Copilot Pro)

List available models:

curl -L \
  -H "Accept: application/vnd.github+json" \
  -H "Authorization: Bearer ${OAUTH_TOKEN}" \
  https://api.githubcopilot.com/models | jq -r '.data[].id'
πŸ›’ GitHub Market Models

Overview:

  • 🌐 Endpoint: https://models.github.ai/inference
  • πŸ” Browse: GitHub Marketplace Models
  • πŸ“Š Rate Limits: 4k input tokens, 4k output tokens per request

List available models:

curl -L \
  -H "Accept: application/vnd.github+json" \
  -H "Authorization: Bearer ${OAUTH_TOKEN}" \
  -H "X-GitHub-Api-Version: 2022-11-28" \
  https://models.github.ai/catalog/models | jq -r '.[].id'

πŸ”€ OpenRouter

OpenRouter provides unified API access to multiple AI modelsβ€”try different models using one API to find your best fit!

πŸ†“ Free Models Available

Model Link
GPT OSS 20B Try it
Qwen3 Coder Try it
GLM 4.5 Air Try it
Kimi K2 Try it
DeepSeek Chat v3.1 Try it

Setup: πŸ”‘ Generate API Key

πŸ’‘ Rate Limits:

  • With 10+ credits purchased: 1,000 requests/day
  • Otherwise: 50 requests/day

⚑ Groq

Groq offers high-speed inference with free tier access.

πŸ†“ Free Models

Available models from Rate Limits documentation:

  • openai/gpt-oss-120b
  • openai/gpt-oss-20b
  • qwen/qwen3-32b
  • moonshotai/kimi-k2-instruct-0905

Setup: πŸ”‘ Generate API Key

πŸš€ NVIDIA Build

NVIDIA Build provides free API access to a wide selection of AI models optimized on NVIDIA infrastructure.

πŸ†“ Example Models Available

Model Full Model Name Link
Qwen3 Next 80B qwen/qwen3-next-80b-a3b-instruct Try it
Qwen3 Coder 480B qwen/qwen3-coder-480b-a35b-instruct Try it
GPT-OSS 120B openai/gpt-oss-120b Try it
Kimi K2 Instruct moonshotai/kimi-k2-instruct-0905 Try it
DeepSeek V3.1 deepseek-ai/deepseek-v3_1 Try it

Setup:

πŸ’‘ Note: Use the full model name (with namespace) when making API requests.


πŸ’» Local Model Providers

πŸ¦™ Ollama

Ollama - Lightweight framework for running LLMs locally via command line.

Key Features:

  • ⚑ Simple CLI interface
  • 🌐 RESTful API
  • 🐳 Docker-like model management
  • πŸ€– Popular models: LLaMA, Gemma, DeepSeek
  • πŸ”Œ OpenAI-compatible API
  • πŸ–₯️ Cross-platform support

πŸ“Š Ollama Model Performance

Model Sizes:

Model Size
gpt-oss:120b 65 GB
gpt-oss:20b 13 GB
qwen3:8b 5.2 GB
qwen3:30b 18 GB

Performance Benchmark (tokens/second):

Machine gpt-oss:120b gpt-oss:20b qwen3:8b qwen3:30b
πŸ–₯️ Windows PC (Intel i9) - 15 t/s 12 t/s 22 t/s
πŸ’» MacBook Pro (M3 Max) - 70 t/s 57 t/s 74 t/s
πŸ–₯️ Linux Server (Dual RTX 4090) 36 t/s 156 t/s 140 t/s 163 t/s
πŸ“‹ Machine Specifications
  • Windows PC (Intel i9):

    • CPU: Intel i9-12900
    • GPU: Intel UHD Graphics 770 (2 GB)
    • RAM: 64 GB
  • MacBook Pro (M3 Max):

    • Apple M3 Max with 64 GB RAM
  • Linux Server (Dual RTX 4090):

    • CPU: Xeon(R) w7-3445 (40 CPUs)
    • GPU: 2 Γ— Nvidia RTX 4090
    • RAM: 128 GB

πŸ–₯️ LM Studio

LM Studio - User-friendly desktop GUI for running local LLMs with no technical setup required.

Key Features:

  • πŸ›οΈ Model marketplace
  • 🌐 OpenAI-compatible API server
  • πŸ’¬ Chat interface
  • πŸ“¦ GGUF model support
  • πŸ’° Free for personal & commercial use

πŸ”€ API Proxies

Most AI tools support OpenAI-compatible APIs. For tools requiring Anthropic-compatible APIs, these solutions provide compatibility:

πŸ”„ Claude Code Router

Claude Code Router - Routes Claude Code requests to different models with request customization.

πŸ“¦ Installation (Linux/macOS)
# Install Claude Code CLI (prerequisite)
npm install -g @anthropic-ai/claude-code

# Install Claude Code Router
npm install -g @musistudio/claude-code-router
βš™οΈ Configuration Examples

Create ~/.claude-code-router/config.json with your preferred providers:

{
  "LOG": true,
  "API_TIMEOUT_MS": 600000,
  "Providers": [
    {
      "name": "gemini",
      "api_base_url": "https://generativelanguage.googleapis.com/v1beta/models/",
      "api_key": "$GEMINI_API_KEY",
      "models": ["gemini-2.5-flash", "gemini-2.5-pro"],
      "transformer": { "use": ["gemini"] }
    },
    {
      "name": "openrouter",
      "api_base_url": "https://openrouter.ai/api/v1/chat/completions",
      "api_key": "$OPENROUTER_API_KEY",
      "models": ["google/gemini-2.5-pro-preview", "anthropic/claude-sonnet-4"],
      "transformer": { "use": ["openrouter"] }
    },
    {
      "name": "grok",
      "api_base_url": "https://api.x.ai/v1/chat/completions",
      "api_key": "$GROK_API_KEY",
      "models": ["grok-beta"]
    },
    {
      "name": "github-copilot",
      "api_base_url": "https://api.githubcopilot.com/chat/completions",
      "api_key": "$GITHUB_TOKEN",
      "models": ["gpt-4o", "claude-3-7-sonnet", "o1-preview"]
    },
    {
      "name": "github-marketplace",
      "api_base_url": "https://models.github.ai/inference/chat/completions",
      "api_key": "$GITHUB_TOKEN",
      "models": ["openai/gpt-4o", "openai/o1-preview", "xai/grok-3"]
    },
    {
      "name": "ollama",
      "api_base_url": "http://localhost:11434/v1/chat/completions",
      "api_key": "ollama",
      "models": ["qwen3:30b", "gpt-oss:20b", "llama3.2:latest"]
    }
  ],
  "Router": {
    "default": "gemini,gemini-2.5-flash",
    "background": "ollama,qwen3:30b",
    "longContext": "openrouter,google/gemini-2.5-pro-preview"
  }
}
πŸ’» Usage Commands
# Start Claude Code with router
ccr code

# Use UI mode for configuration
ccr ui

# Restart after config changes
ccr restart

# Switch models dynamically in Claude Code
/model ollama,llama3.2:latest

⚠️ Known Issue: The proxy for Ollama models does not work properly with Claude Code.

πŸŒ‰ Copilot API Bridge

copilot-api - Converts GitHub Copilot into OpenAI/Anthropic API compatible server for use with Claude Code.

πŸš€ Deployment Example (Singularity/CVMFS)

The copilot-api tool is available in the modern-linuxtools Singularity image on CVMFS:

# Setup the environment
$ source /cvmfs/atlas.sdcc.bnl.gov/users/yesw/singularity/alma9-x86/modern-linuxtools/setupMe.sh

# Start the API wrapper
$ copilot-api start -c
[...]
  ➜ Listening on: http://130.199.48.146:4141/

# In another terminal, use with Aider
$ export ANTHROPIC_BASE_URL=http://130.199.48.146:4141 && aider --no-git --anthropic-api-key dummy --model anthropic/claude-sonnet-4

# Or use with Claude Code CLI (also included in modern-linuxtools)
$ export ANTHROPIC_BASE_URL=http://130.199.48.146:4141 && claude-code

πŸ“Œ Important Notes:

  • Use your own URL in ANTHROPIC_BASE_URL and remove trailing /
  • Enable X11 forwarding when SSH-ing: ssh -X username@hostname
  • All GitHub Copilot models (excluding Market models) become accessible

πŸ€– Automated Setup with run-claude-copilot.sh

For a streamlined experience, this script automates the entire setup process for using Claude Code with GitHub Copilot models.

✨ Key Features:

Feature Description
πŸ“¦ Auto Dependency Management Installs nvm, npm, copilot-api, and claude-code
⚑ Simplified Usage Single command to start fully configured Claude session
πŸ”„ Model Selection Specify which Copilot model to use
πŸ› οΈ Utility Functions Check usage, list models, update packages
πŸ”— Transparent Args Forwards arguments directly to claude command

πŸ’» Usage Examples:

# Run Claude with default settings
./scripts/run-claude-copilot.sh

# List available Copilot models
./scripts/run-claude-copilot.sh --list-models

# Check your Copilot API usage
./scripts/run-claude-copilot.sh --check-usage

# Run Claude with a specific model and pass a prompt
./scripts/run-claude-copilot.sh --model claude-sonnet-4 -- -p "Explain quantum computing"

# Get help on the script's options
./scripts/run-claude-copilot.sh --help

# Get help on Claude's own options
./scripts/run-claude-copilot.sh -- --help

πŸ“š Detailed Tool Guides

Comprehensive documentation for each AI terminal tool:

Tool Description Guide
🀝 Aider AI pair programming in your terminal Read Guide
πŸ€– GitHub Copilot CLI Copilot coding agent directly in your terminal Read Guide
πŸ’Ž Gemini CLI Google's Gemini in your terminal Read Guide
πŸš€ Qwen Code Qwen3-Coder models in your terminal Read Guide

πŸ–₯️ AI-Enhanced Terminals

⚑ Warp Terminal

AI-first terminal that integrates intelligent agents directly into the command line.

✨ Key Features:

πŸ’¬ Natural Language Commands

Generate commands with # trigger

πŸ€– Real-time AI

Autosuggestions and error detection

🎀 Voice Commands

Multi-agent parallel workflows

🏒 Enterprise Ready

SAML SSO, BYOL, zero data retention

πŸ“Š Usage Limits:

  • πŸ†“ Free tier: 150 requests/month
  • πŸ’Ž Paid plans available for higher usage

πŸ“¦ Installation:

brew install --cask warp    # macOS
winget install Warp.Warp    # Windows

# Linux - Multiple package formats available
# See: https://www.warp.dev/blog/warp-for-linux
# Packages include: .deb (apt), .rpm (yum/dnf/zypper), Snap, Flatpak, AppImage, and AUR

🌊 Wave Terminal

Open-source terminal that brings graphical capabilities into the command line.

✨ Key Features:

πŸ–ΌοΈ Inline Previews

Images, markdown, CSV, video files

πŸ“ VSCode-like Editor

Integrated editor for remote files

🌐 Built-in Browser

Web browser and SSH connection manager

πŸ“Š Custom Widgets

Dashboard creation capabilities

πŸ–₯️ Cross-platform

Local data storage for privacy

πŸ€– AI Integration:

  • βœ… Built-in AI assistance for command suggestions
  • βš™οΈ Configurable AI models via "Add AI preset..."
  • πŸ¦™ Support for Ollama and other local models
  • 🎯 Context-aware recommendations

πŸ“¦ Installation:

Download from waveterm.dev/download

Available as: Snap, AppImage, .deb, .rpm, and Windows installers


Made with ❀️ by the Community

⭐ Star on GitHub | πŸ› Report Issues | πŸ’‘ Contribute

Supercharge your terminal workflow! πŸš€

About

A collection of guidance for terminal-based AI coding tools (Qwen Code, Aider,Copilot CLI, and Gemini CLI) and API provider integration instructions.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages