-
-
Notifications
You must be signed in to change notification settings - Fork 618
Description
Describe the feature
Problem
Currently, PictoPy doesn't provide any special search feature.
There is no way to search using natural, conversational queries like
"show me photos from the beach" or "find pictures with my friends
from last year".
Proposed Solution
Implement a natural language search feature powered by AI that allows
users to query their photo library conversationally. This would involve:
- Generating text embeddings for each photo's metadata (tags, detected
objects, faces, dates, locations) - Storing embeddings in a vector index (e.g. FAISS)
- Accepting natural language queries from the user
- Converting queries to embeddings and retrieving the most semantically
relevant photos
Additional AI Features (can be separate issues)
- Duplicate photo detection — identify near-identical images using
embedding similarity and allow bulk removal - Auto-generated albums — automatically group photos by event,
location, or date using clustering
Why This Fits PictoPy
PictoPy already uses YOLO and FaceNet to extract rich metadata from
photos. This feature would make that metadata actually queryable in a
natural way, significantly improving the user experience without
compromising the privacy-first approach since all processing stays local.
Tech Stack Suggestion
- Python, FAISS, Sentence Transformers (all-MiniLM-L6-v2 for lightweight
local embeddings) - No external API calls — fully offline and privacy-preserving
Additional Notes
I'm interested in working on this feature and have prior experience
building RAG pipelines (FAISS + embeddings). Happy to discuss the
approach before starting implementation.
Add ScreenShots
Record
- I agree to follow this project's Code of Conduct
- I want to work on this issue