Workflows
How to Search Video by Content Instead of Filename
A001_C003.MOV tells you nothing. Neither does DJI_0047.MP4. Content-based search means typing what you remember and finding the exact moment. Here is how the shift from filename to content retrieval works.
Open your footage folder. What do you see?
A001_C003.MOV. A001_C004.MOV. A001_C005.MOV. DJI_0047.MP4. GH010089.MP4. BRAW_001.braw. Maybe a hundred more. Maybe a thousand.
You know what you are looking for. The moment in the interview where John talks about the budget overrun. The wide shot of the lobby with the morning light. The drone pull-back over the new building. You can picture it. But nothing in front of you helps you find it.
This is the filename problem. It affects every editor, every project, and every footage library. And it is the reason finding a clip often takes longer than editing it.
A001_C012_0814KN.R3D
personLena detected at 04:10, 21:44, 38:02
C0034_sunset_harbor.MP4
sceneGolden hour establishing shot, harbor with boats
DOC_Interview_EP02.mp4
transcript"...quarterly goals and marketing strategy across all channels..."
The filename problem
Camera manufacturers assign filenames based on internal logic: card slot, reel number, clip counter. A RED camera produces files like A001_C003_0214K7.R3D. A DJI drone produces DJI_0047.MP4. A GoPro produces GH010089.MP4. A Blackmagic camera produces BRAW_001.braw.
None of these names encode anything about what is in the clip. They are identifiers, not descriptions.
Some teams try to fix this by renaming files. Interview_John_Day2_TakeA.mov. But renaming introduces its own problems. It is tedious, error-prone, and only captures the broadest context. "Interview_John_Day2" tells you who and when, not what he said. And renaming raw cinema files risks breaking sidecar metadata links.
Even descriptive filenames are shallow. The best filename you can write is still just a few words of context attached to a container that might hold 45 minutes of footage. You are labeling the box, not indexing the contents.
The folder structure workaround
When filenames fail, teams build folder hierarchies. Project > Shoot Date > Camera > Card. Or Project > Subject > Setup. Or some hybrid that seemed logical at the time.
Folder structures get you to the neighborhood. They answer "which camera card was rolling during John's interview on Day 2." They do not answer "which clip on that card contains the moment where John discusses the budget" or "did we also capture a reaction shot during that answer."
Folder organization also assumes discipline. Every person who touches the media needs to follow the convention. One collaborator dumps files into the wrong folder, one card gets copied without following the naming scheme, and the system degrades. The larger the team and the longer the project, the more likely this happens.
The manual logging detour
The thoroughgoing solution is manual clip logging. A human watches every clip, writes descriptions, notes timecodes, and creates a searchable spreadsheet or database. This produces excellent metadata.
It also takes two to three times the footage duration. A ten-hour shoot requires 20 to 30 hours of logging. That is time and money most projects do not have. So most footage goes unlogged, and finding clips reverts to memory, thumbnails, and scrubbing.
The shift to content-based search
Content-based search takes a fundamentally different approach. Instead of searching labels attached to files, you search the actual content inside them.
You type "John discusses budget" and the search engine checks the transcript of every processed video for those words, or words with similar meaning. It finds the clip, the timestamp, and gives you a direct link to the moment.
You type "wide shot of lobby with morning light" and the search engine checks AI-generated scene descriptions for visual matches. It finds the establishing shot regardless of whether anyone labeled it, organized it, or even remembers shooting it.
The filename becomes irrelevant. The folder structure becomes irrelevant. What matters is what is in the footage, and the search engine knows what is in the footage because it analyzed it.
What content-based search actually searches
A comprehensive content-based search system indexes four layers of information from each video:
Transcript. Every word spoken, timestamped and attributed to specific speakers. This is what makes "John discusses budget" work. The search matches against the actual transcript, not a filename or tag.
Scene descriptions. AI-generated natural language descriptions of what each shot looks like. "Wide shot of a modern office lobby with floor-to-ceiling windows, morning sunlight streaming in." This is what makes "wide shot of lobby with morning light" work.
Objects. Specific items detected in each frame: furniture, vehicles, equipment, products, signage. This lets you search for "laptop" or "forklift" and find every clip where one appears on screen.
People. Faces and voices recognized and clustered across your library. Search for a specific person and find every clip they appear in, across all cameras and all shoots.
Together, these four layers mean you can find footage by what someone said, what the scene looks like, what objects are visible, or who is in the shot. Usually some combination of all four.
Practical examples of the shift
Here is what the shift from filename search to content search looks like in practice:
Finding a soundbite. Filename approach: open each interview clip, scrub to find the right section, listen for the quote. Content approach: type the quote (or your best memory of it), get a timestamped result, click to jump there.
Finding B-roll. Filename approach: open the B-roll folder, scrub through each clip, hope you recognize the shot. Content approach: type "aerial shot of construction site" or "close-up of hands assembling product" and get every matching clip.
Finding a person. Filename approach: check which interview clips are labeled with the person's name, then check every other clip manually. Content approach: search the person's name and get every clip where their face appears or their voice is heard, across the entire library.
Finding a reaction shot. Filename approach: impossible unless someone labeled it. Content approach: type "person reacting" or search for the person's face in clips timestamped around the relevant moment.
In each case, the content-based approach is faster and more complete. It does not depend on someone having organized the footage correctly. It does not depend on memory. It works even for footage shot years ago by people who are no longer on the team.
How FrameQuery handles the shift
FrameQuery processes your footage through a cloud-based analysis pipeline that generates transcripts, scene descriptions, and object labels automatically. Face and voice recognition run locally on your machine. The results are stored in a local search index powered by Tantivy, a Rust search engine, with both BM25 text matching and MiniLM semantic embeddings.
The semantic layer is what makes imprecise queries work. You do not need to guess the exact words in the scene description. Searching "golden hour exterior" can match a scene described as "outdoor shot at sunset with warm orange light" because the embedding model understands the relationship between those phrases.
Processing takes roughly five minutes per hour of footage. Your original files never leave your machine. The index supports 50+ formats natively, including R3D, BRAW, ProRes, MXF, ARRIRAW, and CinemaDNG, so you do not need to transcode anything before indexing.
When you find what you need, export your selections as FCPXML, EDL, Premiere XML, or LosslessCut CSV and drop them straight into your timeline.
The filename still has a job
Content-based search does not mean filenames and folders stop mattering. Good organization is still valuable for project management, archiving, and backup. Camera-original filenames preserve provenance and should generally not be changed.
What changes is that filenames are no longer your primary tool for finding footage. They become administrative identifiers, which is what they were designed to be. The job of finding what is inside the footage shifts to a system that actually looked inside the footage.
Join the waitlist to stop searching by filename and start searching by content when FrameQuery launches.