-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Roadmap of MMAction2 #19
Comments
I strongly recommend adding the support for dataset FineGym99 with video dataset_type, it would be more convenient for users to validate the ideas for fine-grained action recognition or localization tasks. Hoping this would come true in a not so long future! |
it will be nice if mmaction2 could support ava dataset and spatio-temporal action detection models. |
it will be nice if mmaction2 can give some pretrained backbone models for user,, such as ResNet3dSlowFast and so on. |
Yes it is in the plan. |
There are already lots of pretrained models in the model zoo. |
It will be better if the model can output in video format such as mp4. I have tired the demo.py, it feedbacks text. |
It now supports to output video format and gif format in |
@dreamerlin could you pls sort out all feature requests in one grand post here, so that we can easily track the status? 🏃 |
Introducing Multi-Grid or mixed precision training strategy would be helpful for faster prototype iteration. |
In the action localization task,you provided the code to get the AUC metric for action proposal evaluation. |
It can be used to recognize real-time videos with webcamera or something else? thanks |
There are many trained models in Model Zoo, while all of them are just used to test the performance of the proposed works. Do you plan to make them available for backbone pre-training? Say I may want to use the i3d pre-trained on kinetics-400 as the pre-trained backbone of my own model. It seems that we don't have much choice of pre-trained backbones except a Resnet50 on ImageNet. |
To use the pre-trained model for the whole network, the new config adds the link of pre-trained models in the load_from. See Tutorial 1: Finetuning Models # Use Pre-Trained Model and example. And to use backbone pre-training, you can change |
Wow! Fantastic! I think you can mention this feature somewhere in case others, like me, may don't know that they directly use pre-trained weights of the whole model for the backbone. |
Could you please support X3D |
Here is the X3D config files. https://github.com/open-mmlab/mmaction2/tree/master/configs/recognition/x3d |
Could you please add Video Action/Activity Temporal Segmentation models? |
Also, could you please add Video models on MovieNet data set? |
Hi, I'm struggling to train a model using a dataset structured like the AVA dataset. Does anyone have a config file that they have used for this type of dataset that they would be willing to share? There is a code to create an ava dataset, but I haven't been able to find any config files. Otherwise, is there a different framework I can train where I have bounding boxes in the training data? |
Recently I learned about action localization/detection/segmentation(They seem to be the same thing ), it seems that it can generate a file like caption, i found it very interesting and practical. I will be very apreciate it if mmaction2 could have the action localization demo and more docs about it, thanks ! |
Very happy to have spatio-temporal action detection model today... Two related features could be very helpful:
|
Do you have a plan to add flow models for TSN and I3D? |
How about adding some models for temporal action segmentation? |
Thanks for the great repo! Do you have plans adding S3D and S3D-G from https://arxiv.org/abs/1712.04851? They achieve better performance than the I3D model while runs much faster. Here is a reproduced implementation of the S3D model: https://github.com/kylemin/S3D. And for S3D-G model https://github.com/antoine77340/S3D_HowTo100M/blob/master/s3dg.py, https://github.com/tensorflow/models/blob/master/research/slim/nets/s3dg.py |
Thanks in advance for this great unceasing progressing repo. Recently, I saw that on ava-kinetics challenge, the new method 'Actor-Context-Actor Relation Network for Spatio-Temporal Action Localization' has a very good performance and take the lead of nearly 6 percent to the second place in the competition 2020. And I think is a good candidate to enrich the area of spatio temporal action localization in mmaction2. Will you consider to include this network? |
Could you please add the algorithm proposed in the paper of AVA dataset [1]. It is helpful for comparing experiment for spatio-temporal action localization when using AVA dataset. The model is consist of Faster-Rcnn and I3D. Reference: |
Is there any plan or current work for multi modal action classification? |
Maybe MMAction2 could support some of the models and datasets from PytorchVideo |
Do you plan to support Action Genome dataset and model ? |
Add output predictions as JSON in I have implemented this but need to polish it so that it's clean and similar to the rest of the codebase here. Will do a PR in the future. |
Hope to have Kinetics-TPS FineAction MultiSports data set support, pre-training model, training and web video inference demo |
I hope to use posec3d to realize bone-based spatiotemporal motion detection |
Add demo scripts for temporal action detection models Was mentioned in #746 any progress? |
Hope to have web_cam demo for posec3d in near future. |
Do you plan to add a new model to spatio-temporal action detection? The ACRN (Actor Centric Relation Network) is great. However, ACAR adopts the previous work and builds on it with better results. |
I would appreciate it if you could add ViViT.Because I feel there are few transformer-based method for action recognition in MMaction2. |
Hope to have two-stream dataset support , such as in slow-fast : rgb for slow-pathway and optical flow for fast-pathway. |
怎么用自己的数据集跑bsn模型呢,有没有具体的步骤,我需要做些什么,谢谢 |
We keep this issue open to collect feature requests from users and hear your voice. Our monthly release plan is also available here.
You can either:
The text was updated successfully, but these errors were encountered: