NodeTool includes a guided setup that handles dependencies automatically.


Quick Start

  1. Download NodeTool from nodetool.ai
  2. Run the installer
  3. Choose where to install (default is fine)
  4. Wait for setup to complete (~5-10 minutes)
  5. Start building workflows!

System Requirements

Basics

Component Need
OS Windows 10+, macOS 11+, Linux (Ubuntu 20.04+)
RAM 8 GB minimum, 16 GB recommended
Storage 20 GB free (SSD recommended)
Internet For setup and cloud AI

For Local AI

Running models locally gives you privacy and offline use, but needs more resources:

Hardware Can Do
NVIDIA GPU (8+ GB VRAM) All local AI including image generation
Apple Silicon (M1/M2/M3) Excellent performance via MLX
CPU only Works, slower

No GPU? Use cloud providers (OpenAI, Anthropic) instead. Add API key in Settings.

What Different Tasks Need

GPU Tier Recommended Setup Best Local Experience (Optimized)
Entry (8 GB) RTX 4060 / 5060 Flux.1 Schnell (Nunchaku), Qwen-Image-Lightning, 8B LLMs (Llama 3/4).
Mid (12–16 GB) RTX 4070 Ti / 5070 Qwen-Image-Edit (4-bit), Flux.1 Dev (Nunchaku), 32B Reasoning LLMs (DeepSeek R1 Distill).
Pro (24–32 GB) RTX 3090 / 4090 / 5090 Full Qwen-Image 2512, Wan2.1 Video (720p), 70B LLMs (Llama 3.3/4 Q4).
Ultra (48 GB+) Dual 5090s / Mac Ultra DeepSeek-V3 (Full Local), 4K Video Gen (LTX-2), LoRA training in minutes.

In 2026, Apple hardware is particularly strong for local AI because of Unified Memory Architecture (UMA). Unlike Windows PCs where you are limited by the VRAM on your graphics card, a Mac can use a large portion of its total RAM for AI models.

With the M4 chip family and the MLX framework, Macs are now competitive with NVIDIA for “compute-heavy” tasks like Flux and Qwen-Image.

2026 Apple Silicon AI Capability Table

Chip Model Min. RAM Ideal RAM Can Handle (MLX Optimized)
M4 (Base) 16 GB 32 GB 8B LLMs (Llama 4), Flux.1 Schnell (8-bit), Sana 4K images.
M4 Pro 24 GB 64 GB Qwen-Image-Edit, 32B Reasoning models (DeepSeek R1), Flux.1 Dev.
M4 Max 48 GB 128 GB 70B Flagship LLMs, Full-precision Flux, 720p Video (Wan2.1).
M2/M3 Ultra 128 GB 512 GB DeepSeek-V3 (671B), 4K Video workflows, massive Batch-processing.

Specific Task Guide for Mac (2026)

1. Image Generation (MLX / MFLUX)

Apple users should use MLX-based tools (like mflux) rather than standard PyTorch for a 3x speed boost.

  • Flux.1 Dev: Requires at least 32GB RAM to run smoothly at 8-bit.
  • Qwen-Image-Edit: Now natively supported via MLX. On an M4 Max, it can perform complex “Multi-Image” edits in under 10 seconds.
  • Sana (4K): Runs exceptionally well on the base M4 because of its low parameter count but high resolution output.

2. Language Models (LLMs)

The rule of thumb for Mac: Your Model Size (GB) + 4GB (System) < Total RAM.

  • Llama 3.3/4 (70B) @ Q4: Needs ~42 GB. Runs great on a 64GB M4 Pro/Max.
  • DeepSeek-V3 (MoE): This massive model requires at least 128GB RAM (Ultra chips) even when heavily quantized.

3. Video Generation

  • Wan2.1 (Small): Can run on M4 Pro (48GB).
  • CogVideoX: Best on M4 Max due to high memory bandwidth requirements ( GB/s).

Platform-Specific Instructions

Windows

  1. Download the .exe installer from nodetool.ai
  2. Run the installer – Windows Defender may ask for permission (click “Run anyway”)
  3. Approve any firewall prompts so NodeTool can run its local server
  4. NVIDIA users: Ensure you have recent GPU drivers installed for best performance

macOS

  1. Download the .dmg file from nodetool.ai
  2. Open the DMG and drag NodeTool to Applications
  3. First launch: Right-click and choose “Open” (required for unsigned apps)
  4. Apple Silicon: NodeTool automatically uses MLX for optimized local AI

Linux

  1. Download the AppImage or .deb package from nodetool.ai
  2. AppImage: Make executable with chmod +x and run directly
  3. Debian/Ubuntu: Install with sudo dpkg -i nodetool.deb
  4. NVIDIA users: Ensure CUDA drivers are installed for GPU acceleration

What Gets Installed

NodeTool automatically sets up everything it needs to run. Here’s what happens behind the scenes:

Core Components

  • Python Environment – Self-contained Python installation (doesn’t affect your system Python)
  • AI Engines – Tools for running AI models locally:
    • Ollama – For language models
    • llama.cpp – Optimized inference (GPU-accelerated where available)
  • Dependencies – All required libraries and packages

Why 20 GB?

NodeTool itself is small, but AI models can be large:

Component Typical Size
NodeTool + Python environment 2-4 GB
GPT-OSS (recommended LLM) ~4 GB
Flux (image generation) ~12 GB
Total with recommended models ~20 GB

You can install fewer models to save space, or use cloud providers instead.


Step-by-Step Installation

1. Download NodeTool

Visit nodetool.ai and click the download button for your operating system.

2. Run the Installer

Launch the downloaded file. NodeTool’s setup wizard will guide you through the process.

3. Choose Installation Location

You’ll be asked where to install NodeTool’s environment:

  • Default location – Recommended for most users
  • Custom folder – Choose any location with enough free space

Tip: Use an SSD for faster AI model loading and workflow execution.

Select Optional Packages

Choose additional features:

  • Cloud AI Services – OpenAI, Anthropic, Google integrations
  • Document Processing – PDF extraction, OCR
  • Audio/Video Tools – Media processing nodes

Additional packages can be installed later from Settings → Packages.

5. Wait for Download

NodeTool downloads and sets up all components. Typically 5-10 minutes depending on internet connection.

6. Launch NodeTool

Once installation completes, NodeTool opens automatically. You’re ready to start building!


After Installation

First Launch

  1. Firewall prompts: Approve any requests – NodeTool runs a local server that needs network access
  2. Model Manager: Open Models → Model Manager to download AI models
  3. Templates: Check the Dashboard for ready-to-use workflow templates

Sign In (Optional)

  • Sign in with Supabase: Sync workflows across devices
  • Localhost Mode: Keep everything local and private

Install AI Models

To run workflows locally, install some AI models:

  1. Go to Models → Model Manager
  2. Install GPT-OSS for text generation
  3. Install Flux for image generation
  4. Or skip and use cloud providers with your API keys

Troubleshooting Installation

Common Installation Issues

Installation takes too long

  • Large models take time to download
  • Check internet connection
  • Try pausing/resuming or restart the installer

Not enough disk space

  • Free up space or choose a different installation location
  • Use cloud providers instead of local models

GPU not detected

Can’t connect to server

  • Approve firewall prompts
  • Restart NodeTool
  • Check if antivirus is blocking the connection

CUDA and NVIDIA Driver Issues

NodeTool uses CUDA for GPU acceleration on NVIDIA cards. If you’re having GPU issues:

Check Your CUDA Version

Open a terminal/command prompt and run:

nvidia-smi

You should see your GPU model and driver version. NodeTool requires:

  • CUDA 11.8 or CUDA 12.x (12.1+ recommended)
  • Driver version 525.60+ for CUDA 12.x

Common CUDA Problems

“CUDA out of memory”

  • Close other GPU-intensive applications (browsers, games, other AI tools)
  • Use smaller/quantized models (see Hardware Requirements)
  • Reduce batch sizes in workflow settings
  • Check if another process is using GPU: nvidia-smi shows GPU memory usage

“No CUDA-capable device detected”

  1. Verify your GPU is NVIDIA and supports CUDA (GTX 900 series or newer)
  2. Update NVIDIA drivers from nvidia.com/drivers
  3. Reinstall CUDA Toolkit if needed: developer.nvidia.com/cuda-downloads

“CUDA version mismatch” or “cuDNN errors”

  • Multiple CUDA versions can conflict. Check installed versions:
    # Windows
    nvcc --version
    where nvcc
      
    # Linux/macOS
    nvcc --version
    which nvcc
    
  • If multiple versions exist, ensure your PATH points to the correct one
  • NodeTool’s bundled environment usually handles this, but system conflicts can occur

“torch.cuda.is_available() returns False”

  • Your PyTorch installation may not have CUDA support
  • NodeTool includes its own PyTorch; if using custom Python, install the CUDA version:
    pip install torch --index-url https://download.pytorch.org/whl/cu121
    

Windows-Specific CUDA Issues

  • Visual C++ Redistributable: Install from Microsoft
  • Windows Defender: May quarantine CUDA files. Add NodeTool folder to exclusions
  • Path length: Install NodeTool in a short path (e.g., C:\NodeTool) to avoid Windows path limits

Antivirus and Firewall Issues

Security software can interfere with NodeTool’s local server and AI model execution.

Symptoms

  • NodeTool installs but won’t start
  • “Connection refused” errors
  • Models download but won’t load
  • Slow performance despite adequate hardware

Solutions by Antivirus

Windows Defender

  1. Open Windows Security → Virus & threat protection
  2. Click “Manage settings” under Virus & threat protection settings
  3. Scroll to “Exclusions” and click “Add or remove exclusions”
  4. Add these folders:
    • NodeTool installation directory
    • %USERPROFILE%\.nodetool
    • %USERPROFILE%\.cache\huggingface

Norton, McAfee, Bitdefender, etc.

  • Add NodeTool to your antivirus’s trusted/excluded programs list
  • Temporarily disable real-time scanning during installation
  • Some AV software blocks Python processes – whitelist python.exe in NodeTool’s folder

Firewall Configuration

NodeTool runs a local server (default port 7777). Allow it through your firewall:

Windows Firewall

  1. Open Windows Firewall → “Allow an app through firewall”
  2. Click “Change settings” then “Allow another app”
  3. Browse to NodeTool’s executable and add it
  4. Ensure both Private and Public are checked

macOS Firewall

  1. System Preferences → Security & Privacy → Firewall
  2. Click “Firewall Options”
  3. Add NodeTool and set to “Allow incoming connections”

Linux (ufw)

sudo ufw allow 7777/tcp

Python Environment Issues

NodeTool includes its own Python environment, but system Python can sometimes conflict.

“Python not found” or “Module not found”

  • NodeTool uses a bundled Python – this error usually means installation incomplete
  • Try reinstalling NodeTool, ensuring the installer completes fully
  • Check that you’re launching NodeTool from the correct location

Conflicting Python Environments

If you have Anaconda, Miniconda, or other Python distributions:

  • Don’t activate conda before running NodeTool – it uses its own Python
  • If issues persist, temporarily rename or move your conda installation to test
  • Check your PATH doesn’t override NodeTool’s Python

Virtual Environment Issues (for developers)

If running NodeTool from source:

# Create fresh environment
python -m venv .venv
source .venv/bin/activate  # Linux/macOS
.venv\Scripts\activate     # Windows

# Install dependencies
pip install -e .

Platform-Specific Troubleshooting

Windows

“Missing DLL” errors

  • Install Visual C++ Redistributable (x64): Download
  • Restart after installation

“Access denied” during installation

  • Run installer as Administrator
  • Install to a user-writable location (not C:\Program Files)
  • Disable controlled folder access temporarily

Long path errors

  • Enable long paths in Windows (requires admin):
    New-ItemProperty -Path "HKLM:\SYSTEM\CurrentControlSet\Control\FileSystem" -Name "LongPathsEnabled" -Value 1 -PropertyType DWORD -Force
    
  • Or install NodeTool in a short path like C:\NT

macOS

“App is damaged” or “unidentified developer”

  1. Right-click the app and select “Open” (bypasses Gatekeeper once)
  2. Or: System Preferences → Security & Privacy → “Open Anyway”
  3. If still blocked: xattr -cr /Applications/NodeTool.app

Rosetta 2 (Intel apps on Apple Silicon)

  • NodeTool is native Apple Silicon – no Rosetta needed
  • If you installed the wrong version, delete and reinstall the ARM version

Permissions

  • Grant Full Disk Access if accessing files outside standard locations
  • Grant accessibility permissions if prompted

Linux

AppImage won’t run

chmod +x NodeTool-*.AppImage
./NodeTool-*.AppImage

Missing libraries

# Ubuntu/Debian
sudo apt update
sudo apt install libfuse2 libgl1 libglib2.0-0

# Fedora
sudo dnf install fuse-libs mesa-libGL glib2

GPU not detected (NVIDIA)

# Check driver installation
nvidia-smi

# Install NVIDIA drivers if needed (Ubuntu)
sudo ubuntu-drivers autoinstall

# Install CUDA toolkit
sudo apt install nvidia-cuda-toolkit

Resetting NodeTool

If all else fails, try a clean reinstall:

  1. Uninstall NodeTool (see Uninstalling below)
  2. Delete configuration folders:
    • Windows: %USERPROFILE%\.nodetool
    • macOS: ~/.nodetool and ~/Library/Application Support/NodeTool
    • Linux: ~/.nodetool and ~/.config/nodetool
  3. Delete model caches (optional, saves redownloading):
    • ~/.cache/huggingface
    • ~/.ollama
  4. Reinstall from nodetool.ai

Getting Help

If you’re still stuck:


Uninstalling

Windows

Use Add/Remove Programs in Windows Settings

macOS

Drag NodeTool from Applications to Trash, then remove ~/Library/Application Support/NodeTool

Linux

Remove the AppImage or use sudo dpkg -r nodetool for Debian packages


Next Steps

Ready to build your first workflow? See the Getting Started guide.