NodeTool includes a guided setup that handles dependencies automatically.
Quick Start
- Download NodeTool from nodetool.ai
- Run the installer
- Choose where to install (default is fine)
- Wait for setup to complete (~5-10 minutes)
- Start building workflows!
System Requirements
Basics
| Component | Need |
|---|---|
| OS | Windows 10+, macOS 11+, Linux (Ubuntu 20.04+) |
| RAM | 8 GB minimum, 16 GB recommended |
| Storage | 20 GB free (SSD recommended) |
| Internet | For setup and cloud AI |
For Local AI
Running models locally gives you privacy and offline use, but needs more resources:
| Hardware | Can Do |
|---|---|
| NVIDIA GPU (8+ GB VRAM) | All local AI including image generation |
| Apple Silicon (M1/M2/M3) | Excellent performance via MLX |
| CPU only | Works, slower |
No GPU? Use cloud providers (OpenAI, Anthropic) instead. Add API key in Settings.
What Different Tasks Need
| GPU Tier | Recommended Setup | Best Local Experience (Optimized) |
|---|---|---|
| Entry (8 GB) | RTX 4060 / 5060 | Flux.1 Schnell (Nunchaku), Qwen-Image-Lightning, 8B LLMs (Llama 3/4). |
| Mid (12–16 GB) | RTX 4070 Ti / 5070 | Qwen-Image-Edit (4-bit), Flux.1 Dev (Nunchaku), 32B Reasoning LLMs (DeepSeek R1 Distill). |
| Pro (24–32 GB) | RTX 3090 / 4090 / 5090 | Full Qwen-Image 2512, Wan2.1 Video (720p), 70B LLMs (Llama 3.3/4 Q4). |
| Ultra (48 GB+) | Dual 5090s / Mac Ultra | DeepSeek-V3 (Full Local), 4K Video Gen (LTX-2), LoRA training in minutes. |
In 2026, Apple hardware is particularly strong for local AI because of Unified Memory Architecture (UMA). Unlike Windows PCs where you are limited by the VRAM on your graphics card, a Mac can use a large portion of its total RAM for AI models.
With the M4 chip family and the MLX framework, Macs are now competitive with NVIDIA for “compute-heavy” tasks like Flux and Qwen-Image.
2026 Apple Silicon AI Capability Table
| Chip Model | Min. RAM | Ideal RAM | Can Handle (MLX Optimized) |
|---|---|---|---|
| M4 (Base) | 16 GB | 32 GB | 8B LLMs (Llama 4), Flux.1 Schnell (8-bit), Sana 4K images. |
| M4 Pro | 24 GB | 64 GB | Qwen-Image-Edit, 32B Reasoning models (DeepSeek R1), Flux.1 Dev. |
| M4 Max | 48 GB | 128 GB | 70B Flagship LLMs, Full-precision Flux, 720p Video (Wan2.1). |
| M2/M3 Ultra | 128 GB | 512 GB | DeepSeek-V3 (671B), 4K Video workflows, massive Batch-processing. |
Specific Task Guide for Mac (2026)
1. Image Generation (MLX / MFLUX)
Apple users should use MLX-based tools (like mflux) rather than standard PyTorch for a 3x speed boost.
- Flux.1 Dev: Requires at least 32GB RAM to run smoothly at 8-bit.
- Qwen-Image-Edit: Now natively supported via MLX. On an M4 Max, it can perform complex “Multi-Image” edits in under 10 seconds.
- Sana (4K): Runs exceptionally well on the base M4 because of its low parameter count but high resolution output.
2. Language Models (LLMs)
The rule of thumb for Mac: Your Model Size (GB) + 4GB (System) < Total RAM.
- Llama 3.3/4 (70B) @ Q4: Needs ~42 GB. Runs great on a 64GB M4 Pro/Max.
- DeepSeek-V3 (MoE): This massive model requires at least 128GB RAM (Ultra chips) even when heavily quantized.
3. Video Generation
- Wan2.1 (Small): Can run on M4 Pro (48GB).
- CogVideoX: Best on M4 Max due to high memory bandwidth requirements ( GB/s).
Platform-Specific Instructions
Windows
- Download the
.exeinstaller from nodetool.ai - Run the installer – Windows Defender may ask for permission (click “Run anyway”)
- Approve any firewall prompts so NodeTool can run its local server
- NVIDIA users: Ensure you have recent GPU drivers installed for best performance
macOS
- Download the
.dmgfile from nodetool.ai - Open the DMG and drag NodeTool to Applications
- First launch: Right-click and choose “Open” (required for unsigned apps)
- Apple Silicon: NodeTool automatically uses MLX for optimized local AI
Linux
- Download the AppImage or
.debpackage from nodetool.ai - AppImage: Make executable with
chmod +xand run directly - Debian/Ubuntu: Install with
sudo dpkg -i nodetool.deb - NVIDIA users: Ensure CUDA drivers are installed for GPU acceleration
What Gets Installed
NodeTool automatically sets up everything it needs to run. Here’s what happens behind the scenes:
Core Components
- Python Environment – Self-contained Python installation (doesn’t affect your system Python)
- AI Engines – Tools for running AI models locally:
- Ollama – For language models
- llama.cpp – Optimized inference (GPU-accelerated where available)
- Dependencies – All required libraries and packages
Why 20 GB?
NodeTool itself is small, but AI models can be large:
| Component | Typical Size |
|---|---|
| NodeTool + Python environment | 2-4 GB |
| GPT-OSS (recommended LLM) | ~4 GB |
| Flux (image generation) | ~12 GB |
| Total with recommended models | ~20 GB |
You can install fewer models to save space, or use cloud providers instead.
Step-by-Step Installation
1. Download NodeTool
Visit nodetool.ai and click the download button for your operating system.
2. Run the Installer
Launch the downloaded file. NodeTool’s setup wizard will guide you through the process.
3. Choose Installation Location
You’ll be asked where to install NodeTool’s environment:
- Default location – Recommended for most users
- Custom folder – Choose any location with enough free space
Tip: Use an SSD for faster AI model loading and workflow execution.
Select Optional Packages
Choose additional features:
- Cloud AI Services – OpenAI, Anthropic, Google integrations
- Document Processing – PDF extraction, OCR
- Audio/Video Tools – Media processing nodes
Additional packages can be installed later from Settings → Packages.
5. Wait for Download
NodeTool downloads and sets up all components. Typically 5-10 minutes depending on internet connection.
6. Launch NodeTool
Once installation completes, NodeTool opens automatically. You’re ready to start building!
After Installation
First Launch
- Firewall prompts: Approve any requests – NodeTool runs a local server that needs network access
- Model Manager: Open Models → Model Manager to download AI models
- Templates: Check the Dashboard for ready-to-use workflow templates
Sign In (Optional)
- Sign in with Supabase: Sync workflows across devices
- Localhost Mode: Keep everything local and private
Install AI Models
To run workflows locally, install some AI models:
- Go to Models → Model Manager
- Install GPT-OSS for text generation
- Install Flux for image generation
- Or skip and use cloud providers with your API keys
Troubleshooting Installation
Common Installation Issues
Installation takes too long
- Large models take time to download
- Check internet connection
- Try pausing/resuming or restart the installer
Not enough disk space
- Free up space or choose a different installation location
- Use cloud providers instead of local models
GPU not detected
- Update GPU drivers
- On Windows, ensure CUDA is installed for NVIDIA GPUs
- See CUDA Troubleshooting
Can’t connect to server
- Approve firewall prompts
- Restart NodeTool
- Check if antivirus is blocking the connection
CUDA and NVIDIA Driver Issues
NodeTool uses CUDA for GPU acceleration on NVIDIA cards. If you’re having GPU issues:
Check Your CUDA Version
Open a terminal/command prompt and run:
nvidia-smi
You should see your GPU model and driver version. NodeTool requires:
- CUDA 11.8 or CUDA 12.x (12.1+ recommended)
- Driver version 525.60+ for CUDA 12.x
Common CUDA Problems
“CUDA out of memory”
- Close other GPU-intensive applications (browsers, games, other AI tools)
- Use smaller/quantized models (see Hardware Requirements)
- Reduce batch sizes in workflow settings
- Check if another process is using GPU:
nvidia-smishows GPU memory usage
“No CUDA-capable device detected”
- Verify your GPU is NVIDIA and supports CUDA (GTX 900 series or newer)
- Update NVIDIA drivers from nvidia.com/drivers
- Reinstall CUDA Toolkit if needed: developer.nvidia.com/cuda-downloads
“CUDA version mismatch” or “cuDNN errors”
- Multiple CUDA versions can conflict. Check installed versions:
# Windows nvcc --version where nvcc # Linux/macOS nvcc --version which nvcc - If multiple versions exist, ensure your PATH points to the correct one
- NodeTool’s bundled environment usually handles this, but system conflicts can occur
“torch.cuda.is_available() returns False”
- Your PyTorch installation may not have CUDA support
- NodeTool includes its own PyTorch; if using custom Python, install the CUDA version:
pip install torch --index-url https://download.pytorch.org/whl/cu121
Windows-Specific CUDA Issues
- Visual C++ Redistributable: Install from Microsoft
- Windows Defender: May quarantine CUDA files. Add NodeTool folder to exclusions
- Path length: Install NodeTool in a short path (e.g.,
C:\NodeTool) to avoid Windows path limits
Antivirus and Firewall Issues
Security software can interfere with NodeTool’s local server and AI model execution.
Symptoms
- NodeTool installs but won’t start
- “Connection refused” errors
- Models download but won’t load
- Slow performance despite adequate hardware
Solutions by Antivirus
Windows Defender
- Open Windows Security → Virus & threat protection
- Click “Manage settings” under Virus & threat protection settings
- Scroll to “Exclusions” and click “Add or remove exclusions”
- Add these folders:
- NodeTool installation directory
%USERPROFILE%\.nodetool%USERPROFILE%\.cache\huggingface
Norton, McAfee, Bitdefender, etc.
- Add NodeTool to your antivirus’s trusted/excluded programs list
- Temporarily disable real-time scanning during installation
- Some AV software blocks Python processes – whitelist
python.exein NodeTool’s folder
Firewall Configuration
NodeTool runs a local server (default port 7777). Allow it through your firewall:
Windows Firewall
- Open Windows Firewall → “Allow an app through firewall”
- Click “Change settings” then “Allow another app”
- Browse to NodeTool’s executable and add it
- Ensure both Private and Public are checked
macOS Firewall
- System Preferences → Security & Privacy → Firewall
- Click “Firewall Options”
- Add NodeTool and set to “Allow incoming connections”
Linux (ufw)
sudo ufw allow 7777/tcp
Python Environment Issues
NodeTool includes its own Python environment, but system Python can sometimes conflict.
“Python not found” or “Module not found”
- NodeTool uses a bundled Python – this error usually means installation incomplete
- Try reinstalling NodeTool, ensuring the installer completes fully
- Check that you’re launching NodeTool from the correct location
Conflicting Python Environments
If you have Anaconda, Miniconda, or other Python distributions:
- Don’t activate conda before running NodeTool – it uses its own Python
- If issues persist, temporarily rename or move your conda installation to test
- Check your PATH doesn’t override NodeTool’s Python
Virtual Environment Issues (for developers)
If running NodeTool from source:
# Create fresh environment
python -m venv .venv
source .venv/bin/activate # Linux/macOS
.venv\Scripts\activate # Windows
# Install dependencies
pip install -e .
Platform-Specific Troubleshooting
Windows
“Missing DLL” errors
- Install Visual C++ Redistributable (x64): Download
- Restart after installation
“Access denied” during installation
- Run installer as Administrator
- Install to a user-writable location (not
C:\Program Files) - Disable controlled folder access temporarily
Long path errors
- Enable long paths in Windows (requires admin):
New-ItemProperty -Path "HKLM:\SYSTEM\CurrentControlSet\Control\FileSystem" -Name "LongPathsEnabled" -Value 1 -PropertyType DWORD -Force - Or install NodeTool in a short path like
C:\NT
macOS
“App is damaged” or “unidentified developer”
- Right-click the app and select “Open” (bypasses Gatekeeper once)
- Or: System Preferences → Security & Privacy → “Open Anyway”
- If still blocked:
xattr -cr /Applications/NodeTool.app
Rosetta 2 (Intel apps on Apple Silicon)
- NodeTool is native Apple Silicon – no Rosetta needed
- If you installed the wrong version, delete and reinstall the ARM version
Permissions
- Grant Full Disk Access if accessing files outside standard locations
- Grant accessibility permissions if prompted
Linux
AppImage won’t run
chmod +x NodeTool-*.AppImage
./NodeTool-*.AppImage
Missing libraries
# Ubuntu/Debian
sudo apt update
sudo apt install libfuse2 libgl1 libglib2.0-0
# Fedora
sudo dnf install fuse-libs mesa-libGL glib2
GPU not detected (NVIDIA)
# Check driver installation
nvidia-smi
# Install NVIDIA drivers if needed (Ubuntu)
sudo ubuntu-drivers autoinstall
# Install CUDA toolkit
sudo apt install nvidia-cuda-toolkit
Resetting NodeTool
If all else fails, try a clean reinstall:
- Uninstall NodeTool (see Uninstalling below)
- Delete configuration folders:
- Windows:
%USERPROFILE%\.nodetool - macOS:
~/.nodetooland~/Library/Application Support/NodeTool - Linux:
~/.nodetooland~/.config/nodetool
- Windows:
- Delete model caches (optional, saves redownloading):
~/.cache/huggingface~/.ollama
- Reinstall from nodetool.ai
Getting Help
If you’re still stuck:
- Discord Community – Ask questions and get help from users
- GitHub Issues – Report bugs with system details
- Troubleshooting Guide – For workflow and runtime issues (not installation)
Uninstalling
Windows
Use Add/Remove Programs in Windows Settings
macOS
Drag NodeTool from Applications to Trash, then remove ~/Library/Application Support/NodeTool
Linux
Remove the AppImage or use sudo dpkg -r nodetool for Debian packages
Next Steps
Ready to build your first workflow? See the Getting Started guide.