Getting Started
Get FileMind up and running in under five minutes. This guide covers installation, first-run setup, and your first library scan.
System Requirements
| Minimum | Recommended | |
|---|---|---|
| OS | Windows 10 / macOS 12 | Windows 11 / macOS 14 |
| RAM | 8 GB | 16 GB (for larger AI models) |
| Disk | 2 GB free | 5 GB+ (AI model + embeddings) |
| GPU | Not required | Any CUDA/Metal GPU speeds up AI |
Installation
Windows
- Download
FileMind-setup.exefrom the download page - Run the installer and follow the prompts
- FileMind appears in your Start menu and system tray
macOS
- Download
FileMind.dmgfrom the download page - Open the DMG and drag FileMind to Applications
- Launch from Applications — macOS may ask you to confirm since it's not from the App Store
First Launch
On first launch, FileMind runs the Setup Wizard automatically. The wizard installs Ollama (the local AI runtime), downloads an AI model matched to your hardware, and verifies everything works with a test query. The whole process takes about 2-5 minutes depending on your internet speed.
Your First Scan
- Click Add Library Folder in the toolbar (or drag a folder onto the window)
- Select the folder containing your PDFs — FileMind scans recursively through subfolders
- Watch the scan progress: FileMind extracts text, detects metadata, and generates embeddings
- When the scan completes, rename proposals appear in the Rename Queue
FileMind processes files in the background. You can keep working while it scans — the UI updates in real time as papers are indexed.
What Happens During a Scan
For each PDF, FileMind:
- Extracts text — native extraction for digital PDFs, OCR (via Tesseract) for scanned documents
- Scores text quality — determines if OCR is needed based on character quality
- Detects identifiers — looks for DOI, arXiv ID, ISBN, and other identifiers
- Fetches metadata — queries CrossRef, arXiv API, and other sources using detected identifiers
- Parses heuristically — extracts title, authors, year from document structure when APIs don't help
- Falls back to LLM — uses the local AI model only when deterministic methods fail
- Generates embeddings — creates vector representations for semantic search and related-paper discovery
- Creates a rename proposal — builds a filename from extracted metadata with a confidence score