A CLI tool that uses the Qodo-Embed-1-1.5B embedding model to analyze code, query knowledge, and understand code repositories.
⚠️ ALPHA RELEASE: This tool is currently in alpha status and not production-ready. Use at your own risk.
- Code Embedding: Generate embeddings for code files using pre-trained models
- Query Current Knowledge: Ask questions about embedded code or general info the model knows
- Directory Analysis: Provide a directory, embed all code files, and ask questions about them
- Memory-Efficient Processing: Options to control memory usage for different hardware profiles
This development branch contains work-in-progress features that are not yet ready for release. Current development focuses on:
- Improving memory management
- Adding support for more programming languages
- Enhancing query capabilities
- Optimizing performance on different hardware platforms
# Create virtual env
python3 -m venv venv
source venv/bin/activate
# Install core dependencies first
pip install torch
pip install click transformers faiss-cpu numpy accelerate
# Install package locally
pip install -e .
-
Add a Directory:
# Standard mode (will try to use Qodo-Embed-1-1.5B if possible) code_grasp add /path/to/code/dir # Lightweight mode (good for memory-constrained systems) code_grasp add --lightweight /path/to/code/dir # Control batch size for better memory management code_grasp add --batch-size 2 /path/to/code/dir
-
Ask About Current Knowledge:
# Standard mode code_grasp ask "How do I implement a binary search?" # Lightweight mode code_grasp ask --lightweight "How do I implement a binary search?"
-
Ask About a Directory:
# Standard mode code_grasp ask-dir /path/to/code/dir "What functions handle authentication?" # Lightweight mode with smaller batch size code_grasp ask-dir --lightweight --batch-size 2 /path/to/code/dir "What functions handle authentication?"
-
Display Information:
code_grasp info # Lightweight mode (recommended for Apple Silicon) code_grasp info --lightweight
- Use standard mode for best embedding quality
- Increase batch size for faster processing:
--batch-size 8
or higher - On NVIDIA GPUs, CUDA acceleration is used automatically
- For very large codebases, consider processing in chunks of 5,000-10,000 files
- Standard mode should work with default batch size
- If experiencing memory issues, reduce batch size:
--batch-size 2
- Process larger codebases in smaller chunks
- Use lightweight mode:
--lightweight
- Use minimal batch size:
--batch-size 1
- Process smaller directories at a time (few hundred files max)
- MPS acceleration is used automatically for improved performance
- The
info
command automatically uses lightweight mode on Apple Silicon to prevent segmentation faults - 8GB RAM models: Use
--lightweight
mode for all commands with larger codebases - 16GB+ RAM models: Standard mode should work well with default settings for most commands
- For optimal performance on 8GB models, use
--batch-size 2
with--lightweight
- CUDA acceleration is used automatically
- For GTX 1650-1660 class GPUs: Use standard settings but consider
--batch-size 2
for larger files - For RTX 3000+ series: Can increase batch size to
8
or higher for better performance - For GPUs with 8GB+ VRAM: Can use standard mode with large batch sizes
- Tested on macOS Ventura+ with Apple Silicon
- Intel Macs should work with CPU acceleration
- MPS acceleration automatically used on M1/M2/M3
- Works on major distributions (Ubuntu, Debian, CentOS, etc.)
- CUDA support for NVIDIA GPUs
- ROCm support for AMD GPUs may work but is untested
- Should work on Windows 10/11 with proper Python environment
- CUDA support for NVIDIA GPUs
- Installation via WSL (Windows Subsystem for Linux) is recommended for best compatibility
- Python (.py)
- C++ (.cpp, .hpp, .h, .c)
- C# (.cs)
- Go (.go)
- Java (.java)
- Javascript (.js, .jsx)
- PHP (.php)
- Ruby (.rb)
- Typescript (.ts, .tsx)
- HTML (.html)
- CSS (.css)
Code Grasp offers several memory optimization features:
- Lightweight Mode: Use
--lightweight
to force using a smaller, more memory-efficient model - Batch Processing: Control the number of files processed at once with
--batch-size
- Automatic Fallback: If the large model fails to load, automatically falls back to a smaller model
- Half-Precision: Uses float16 precision for the large model to reduce memory usage
- Large Codebases: May struggle with repositories containing tens of thousands of files
- Memory Usage: The Qodo-Embed model requires significant RAM/VRAM
- Token Limits: Individual files larger than 4,096 tokens will be truncated
- Embedding Quality: Lightweight mode provides lower quality embeddings than standard mode
- Non-Text Files: Binary files and non-supported text formats are skipped
- File Encodings: Assumes UTF-8 encoding for all files
If you experience segmentation faults, try these solutions:
- Use lightweight mode:
code_grasp --lightweight ...
(particularly important for theinfo
command on Apple Silicon) - Reduce batch size:
code_grasp --batch-size 1 ...
- Process smaller directories at a time
- Ensure you have at least 8GB of free RAM
Note: The info
command now automatically uses lightweight mode on Apple Silicon to prevent segmentation faults.
For better performance:
- On M1/M2 Macs, the tool automatically uses MPS acceleration
- On systems with NVIDIA GPUs, CUDA will be used automatically
- Increase batch size for faster processing if you have enough memory
If you encounter problems installing dependencies:
- Ensure you have the latest pip:
pip install --upgrade pip
- On Windows, you may need Microsoft Visual C++ Build Tools
- For CUDA support, ensure you have compatible CUDA drivers installed
- For Apple Silicon, ensure you have Xcode command-line tools installed
Please see CONTRIBUTING.md for guidelines on how to contribute to this project.
This project uses a dual licensing model:
-
Code Grasp Code: The Code Grasp tool itself is licensed under the MIT License, allowing for free use, modification, and distribution.
-
Qodo-Embed-1-1.5B Model: The embedding model used by this tool (Qodo-Embed-1-1.5B) is licensed under the Qodo Open RAIL++-M License, which:
- Allows commercial and non-commercial use
- Includes use-based restrictions to prevent harmful applications
- Requires derivatives to retain these restrictions
- Was released by QodoAI on February 19, 2025
When using Code Grasp, you must comply with both licenses - MIT for the code and Qodo Open RAIL++-M for the model. If you redistribute or create derivative works, you must include both license terms.
For the full Qodo Open RAIL++-M License text, please refer to QodoAI's official website.