-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Help wanted : How to get position data from the video #1
Comments
Hello,
Then you can run:
In |
Hi! |
Is the query mask given, as in, you know which object you are tracking? If not, you can use something like Segment Anything to turn a point query (for example from eye tracking) into an object mask and then run TCOW from there. If the query mask is given as input however, and you simply want to correspond that with a sequence of eye tracking coordinates, then my recommendation would be to write an additional script that looks at the video outputs and measures the strength of the predictions at those coordinates [x, y]. You can easily load video frames as numpy arrays in Python and then the ROI could be defined as a thresholded output (for example, for each pixel, whether the confidence is >0.5 for either the target/occluder/container channel, which map to green/red/blue in the video respectively). If you want the ROI to be a rectangle instead of an arbitrary shape you could also do some light image processing to calculate the encompassing rectangle of the activated pixels. For completeness, my script exports multiple output video files, make sure you load the one that directly stores the segmentation against a black background (I forgot the exact file names right now), and not overlaid on top of the original video. Hope this helps and please let me know if not! ;) |
Thank you for your advice! turned out we found some other way to get the position data in matlab, thanks so much for your help! |
Hi! I'm an undergraduate cognitive science student using tcow to track my experiment videos. I'm interested in extracting data for everything inside the contour of my target objects, not just their center positions. Is there a way to access this data from the model's output? Thank you!
The text was updated successfully, but these errors were encountered: