forked from npna/CoreMLPlayer
-
Notifications
You must be signed in to change notification settings - Fork 0
Open
Description
Feature: Real-time Object Detection with Meta Glasses
Integrate CoreMLPlayer with the Meta Glasses Relay for real-time CV on live video streams.
The Vision
Meta Glasses → iPhone → Mac (CoreMLPlayer) → Display with overlays
Use Cases
- Hands-free AR - Wear glasses, see live detection on Mac display
- Accessibility - Real-time scene description/object localization for blind/low-vision
- Professional Tools - Technician/doctor inspector with live CV feedback
- Research/Testing - Test models in real environments, collect training data
Technical Approach
- Use MacReceiver to get frames from Meta glasses
- Pass frames to CoreMLPlayer's detection pipeline
- Render detection boxes/labels over live video
- Add model selection (YOLO, SAM2, etc.)
- Optional: recording, annotation, audio feedback
Existing Infrastructure
- ✅ Meta glasses relay (mac-relay-yolo branch)
- ✅ CoreMLPlayer detection pipeline
- ✅ Video rendering system
- ✅ Model loading/switching
Main Challenges
- Latency - Multi-hop adds delay (glasses→iPhone→Mac)
- Performance - Real-time detection needs to be fast
- UI - How to overlay detection results on live video feed
Related
- Meta glasses relay: https://github.com/ebowwa/meta-wearables-dat-ios/tree/wip/mac-relay-yolo
- SAM2 support discussion: feat: Add SAM2 (Segment Anything Model) support #3
Metadata
Metadata
Assignees
Labels
No labels