Indexing
Select files or folders, and FrameQuery handles the rest. Frames and audio are extracted on your device and sent for analysis, nothing else. Your originals never leave your machine, and the extracted data is discarded the moment analysis completes.
Select
Files or folders, local or cloud
Extract
Frames and audio on-device
Analyse
Scenes, objects, speech, faces
Index
Stored locally and searchable
Select
Files or folders, local or cloud
Extract
Frames and audio on-device
Analyse
Scenes, objects, speech, faces
Index
Stored locally and searchable
Paid plans include hours of video each month: 10 hrs of video/mo on Starter, 50 hrs of video/mo on Pro, 300 hrs of video/mo on Max. Hours count the duration of your footage, not how long indexing takes. Most videos finish in 5-10 minutes. See pricing.
What Gets Extracted
Video and audio are analysed in parallel. Scene detection, object recognition, and speech-to-text with speaker separation all run simultaneously, so every clip is automatically tagged and indexing stays fast.
People, vehicles, animals, props, text overlays, and hundreds more, identified frame-by-frame and stored as searchable metadata. Search for “red car” or “laptop” and jump straight to the frame.
Learn about visual searchNatural-language summaries of each scene including setting, action, mood, and composition. Also extracts shot type (wide, close-up, medium), shot angle, and dominant color, so you can search by how a shot looks, not just what's in it.
Learn about scene detectionFull speech-to-text with word-level timestamps and automatic speaker diarization. Each segment is tagged with a speaker so you can search by who said something, not just what was said.
Learn about transcription searchNo transcoding required. Point FrameQuery at your media as-is and it handles the rest. Professional camera RAW formats use native SDKs with GPU acceleration; everything else goes through FFmpeg.
On-Device Recognition
Tag a person once and FrameQuery finds every appearance across your entire library, by what they look like and what they sound like. Both models run 100% on your machine. Your biometric data never leaves your device, and embeddings and labels are never included when you share an index.
Sarah
3F 1V
James
5F 2V
Right-click or drag onto a person to assign
Right-click or drag onto a person to assign
All biometric data processed and stored on-device only
Runs entirely on your machine on footage you've already analysed, so the recognition itself adds no extra cost. Optional and consent-gated, and FrameQuery will backfill any videos already in your library when you turn it on.
Search
Once indexed, your data is stored locally using Tantivy (a Rust port of Lucene). Search runs entirely on your machine: no internet required, no API calls, no per-query cost. Results are ranked with BM25 relevance scoring and field-based boosting across transcripts, scene descriptions, objects, and filenames.
Keyword search is only half of it. Semantic search lets you describe what you're looking for in plain language. Try “sunset over a mountain,” “person celebrating after a goal,” or “empty boardroom at night,” and FrameQuery surfaces clips that match the meaning, not just the words. It works across the visual content itself via CLIP embeddings, and across the text of your transcripts, scene descriptions, and detected objects. So “telling a story about a road trip” finds dialogue about travel, “camping gear” surfaces clips tagged with tents and backpacks, and “quiet morning light” picks up scenes described as calm and golden hour. All of it runs locally.
3 results · 12ms
A001_C003.mov
VisualB_ROLL_LAKE.mp4
TranscriptINTERVIEW_EXT.mov
Object@SarahFilter by person, face or voice. Use @"full name" for multi-word names.
"quarterly goals"Exact phrase match. Finds the precise moment someone said this.
-interviewExclude a term. Combine with other queries to narrow results.
codec:prores res:4kMetadata filters: codec, resolution, camera, FPS, ISO, lens, and more.
Also filter by time range, dominant color, match type (transcript, object, scene), date range, source folder, project, or named index. Combine any of these with free-text search.
NLE Export
Export search results or selections as timeline-ready files for Final Cut Pro, Premiere Pro, DaVinci Resolve, Avid, and more. Free on every plan.
Learn about NLE exportExplore Features
Transcription Search
Search every word spoken across your video library.
Face Recognition
Find anyone by face or voice across your footage.
Visual Search
Describe what you need in plain language.
NLE Export
FCPXML, Premiere XML, EDL, LosslessCut.
Audio Sync
Timecode and waveform matching for dual-system audio.
Scene Detection
Scene boundaries, take organization, and grouping.
RAW Format Support
R3D, BRAW, ARRI, ProRes RAW with native SDKs.
Collaboration & Review
Timecoded comments, annotations, and approvals.
Sharing & Backup
Share indexes and back up to local, Drive, or NAS.