Coder Social home page Coder Social logo

bsn-boundary-sensitive-network's People

Contributors

wzmsltw avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

bsn-boundary-sensitive-network's Issues

Baidu Yun files are broken, how to re-generate the files?

I download the tsn_feature files for many times from the link you provided: https://pan.baidu.com/s/19GI3_-uZbd_XynUO6g-8YQ (i also tried the google drive download link, but will be abruptly ended every time) but every time when i unzip the BaiduYun files,

$ unzip -d ../BMN-Boundary-Matching-Network/data/activitynet_feature_cuhk/ csv_mean_100.zip

i got hundreds of bad zipfile offset errors in my terminal like:

Archive:  csv_mean_100.zip
error: End-of-centdir-64 signature not where expected (prepended bytes?)
  (attempting to process anyway)
warning [csv_mean_100.zip]:  zipfile claims to be last disk of a multi-part archive;
  attempting to process anyway, assuming all parts have been concatenated
  together in order.  Expect "errors" and warnings...true multi-part support
  doesn't exist yet (coming soon).
warning [csv_mean_100.zip]:  5242880000 extra bytes at beginning or within zipfile
  (attempting to process anyway)
file #1:  bad zipfile offset (local header sig):  5242880004
  (attempting to re-compensate)
...
file #3998:  bad zipfile offset (local header sig):  5293978810
file #3999:  bad zipfile offset (local header sig):  5294255897
file #4000:  bad zipfile offset (local header sig):  5294532787
file #4001:  bad zipfile offset (lseek):  5294809088
file #4002:  bad zipfile offset (lseek):  5295079424
file #4003:  bad zipfile offset (lseek):  5295357952
file #4004:  bad zipfile offset (lseek):  5295636480
file #4005:  bad zipfile offset (lseek):  5295915008
file #4006:  bad zipfile offset (lseek):  5296185344
file #4007:  bad zipfile offset (lseek):  5296463872
file #4008:  bad zipfile offset (lseek):  5296742400
file #4009:  bad zipfile offset (lseek):  5297004544
file #4010:  bad zipfile offset (lseek):  5297283072
file #4011:  bad zipfile offset (lseek):  5297553408
file #4012:  bad zipfile offset (lseek):  5297831936
file #4013:  bad zipfile offset (lseek):  5298110464

...

Also, when i run BMN program in pytorch(https://github.com/JJBOY/BMN-Boundary-Matching-Network), i will find some files are missing:

like:

FileNotFoundError: [Errno 2] File b'./data/activitynet_feature_cuhk/csv_mean_100/v_sx_npA4wRrw.csv' does not exist: b'./data/activitynet_feature_cuhk/csv_mean_100/v_sx_npA4wRrw.csv'

Could you please tell me how to re-generate the tsn feature files? So i can make them by myself...

same issue at: JJBOY/BMN-Boundary-Matching-Network#11

About action detection with proposals

Hello, thank you for your contribution.
In the paper, On ActivityNet-1.3, we use top-1 video-level class generated by classification model [44] for all proposals in a video. ([44] means UTS at ActivityNet 2016). But I can't find the source code.
So my question is how do you actually do action detection with proposals?
Thank you very much!

Could we get features for THUMOS'14?

Hi Tianwei,

Thanks for your great work! Could we also have extracted features of THUMOS'14? It would be great if the rescaled and original features for THUMOS'14 are provided. Thank you so much!

PEM - sampling the high low and medium

There are some inconsistencies between the paper and the implementation when it comes to sampling data for predicting the iou in the PEM. According to the paper you sample with a 1 to 2 ratio between negative and positives. according to the code, you measure according to the ratio between positive and negatives and according to the ratio between medium and positive. can you add more information about that?
Thanks,
Ilan

Apply for video

Hi,

This is nice work, I would like to apply this for video, How this applies to video?

Could I get original TSN features?

Hello,

I'm trying to experiment with this model using original features.
But I'm in trouble of acquiring video features from "CUHK & ETH & SIAT Solution to ActivityNet Challenge 2016" model.
I want to get the original features used in your paper. (2048-dim I think, and whose length is not rescaled to 100)
Will you able to provide the 2048-dim original two stream features not rescaled?

Thank you.

Thumos14 features required

Hello,

Can you please also provide the rescaled features for Thumos14 similar to ActivityNet dataset?

Could we get C3D features for THUMOS'14?

Hi Tianwei,
Thanks for your great work! Can you provide the C3D features for THUMOS14?It would be great if the original features for THUMOS'14 are provided. Thanks!

what is AR@AN results of rescaled feature on THUMOS14?

Hi, Tianwei Lin, Thank you for your contribution! I have do some experiments on thumos14 based on your training code for ActivityNet! During feature encoding, I rescale the feature sequence to 100, but I got poor results that AR@100 is about 22%. I notice that you do not rescale feature on Thumos14.
Have you do experiments based on rescaled feature on Thumos14 and what about the results? Does the operation of rescaling have so much impact on the results?
Thank you very much!

About BSN improvement B

Hi,
Mentioned in your ActivityNet report improvement B that you adopt TSN and P3D pretrained on Kinetics-400 dataset for video feature extraction. I wonder whether the model fine-tuned on the ActivityNet dataset. If not, are you using the feature of the last pooling layer?

Feature extraction for THUMOS14 is strange

According to the "Implementation Details" part in Section 4 of the original paper, you use the model pre-trained on the training set of ActivityNet-1.3 as the features extractor. And I don't make anything wrong, your BMN uses the output of the last layer as the feature. That's why you get 400-dimensions length features (200 classes, 2 streams).

But the output of the last layer should represent the predicted class scores of the input frame of 200 ActivityNet actions. This kind of feature should be meaningless for THUMOS14 because half of the action classes in THUMOS14 don't belong to ActivityNet. How can you detect an unseen type of action using the classification scores of 200 irrelative actions?

urllib2.URLError: <urlopen error [Errno 110] Connection timed out>

Thanks for your great work.
I want to get the evaluation results. So I followed the readme in Evaluation and use the command python get_classification_performance.py data/activity_net.v1-3.min.json sample_classification_prediction.json. But I got the error urllib2.URLError: <urlopen error [Errno 110] Connection timed out>. Then I checked the url: http://ec2-52-11-11-89.us-west-2.compute.amazonaws.com/challenge16/api.py?action=get_blocked. Ant it is not accessible. Could you please help me with the probelm ?
Thank you !

csv_mean_100/v_nt4Ag91306U.csv does not exist

Hi,

I got this problem of data/activitynet_feature_cuhk/csv_mean_100/v_nt4Ag91306U.csv does not exist. It seems that some of the files are missing.Can you please tell me how to solve this problem?

About THUMOS14 annotations

Hi tianwei!
Thanks a lot about BSN code release.
I tried to train BSN on THUMOS14, do u have annotations of THUMOS14 used in BSN training procedure?

how to unpack the original features?

@wzmsltw Lin, thanks for your repo.

I have downloaded the original features you provided, but how to unpack and use the features, any ideas?

i use command cat original_spatial_csv.tar.gz.0* > original_spatial_csv.tar.gz, and then tar -zxvf original_spatial_csv.tar.gz, but error occurred

tar: Unrecognized archive format
tar: Error exit delayed from previous errors.

How to classify generated proposals?

As described in your paper, you perform action detection by classifying proposals using Untrimmed Net or SCNN. I am trying to do this with UntrimmendNet but don't know how.
Can you provide more details? Do I need to re-train UntrimmedNet based on BSN generated proposals or the "shots" described in the UntrimmedNet paper? And how about testing? What are the inputs and outputs?

Abount feature extraction on THUMOS14?

Thank you for your great help!
I have another question. About THUMOS14, may I ask what is the way you take to extract features?
The FC layer output of anet is 200 classes. Did you modify the FC layer and then train on THUMOS14?

How to get the results of AR@50, AR@100, AR@200...

Hi Will,
Could you please tell me how to control the different numbers of proposals precisely, e.g. 50, 100, 200?
I know it maybe by setting the threshold, but I have no idea how to do it and make the number of proposals are excatly the amount you want.
Thanks!

Visualize Actionness probability Sequence

Hi, thanks for the great work.

For my project, I would like to visualize the the probability actionness score for input video. How can I visualize that sequence? Any help will be appreciated.

Thank you!

Question about the thumos14 features

Hi, Tianwei Lin
Thanks for your great sharing.
When I use the thumos14 features you shared, I find the row number of rgb and flow features of the same videos are different,why so and how can I do?

HOW about THUMOS2014 result?

this repo will get ActivityNet v1.3 result according to your README,but in your paper,we can also see THUMOS2014 result, so could you please release some pre-trained model on THUMOS2014 dataset?
I believe the model in ./models/TEM or PEM is pre-trained model on ActivityNet ?

IOError

Hi~
When i run the code python TEM_train.py, i meet a error about feature,like this:
IOError:File ./data/activitynet_feature_chunk/csv_mean_100/v_QOlsCBRmfWY.csv does not exit
I check the zip_csv_mean_100.zip and cannot find this file...
What's wrong with it? Is it the feature missing?

how to get video_info.csv

Excusme ,could you tell me how you get video_info.csv and anet_anno_action.csv files ?
Is the two files you created from activity-net.v1-3.min.json ? Then could you please send the script to me?
Thank you very much!!!
My email is [email protected]

Issue regarding rescaling the feature length

Hi,

(1) You mentioned that "For convenience of training and testing, we rescale the feature length of all videos to same length 100", if I understand correctly, this means that you only extract 100 snippets per video clip? Also would you mind elaborating on how the video resizing is done?

(2) Also there are many improvements for the winning solution compared to BSN baseline, will you consider releasing the code for the winning solution?

Thank you

ERROR about activitynet dataset

Hi, thanks for your great contribution!
However I have two questions while I am downloading the dataset activitynet-1.3 as follows:

  1. It shows that some of the youtube accountis terminated so the video is not available either. 1
  2. Due to some unknown reasons, the video is unable to extract, also the webpage cannot be reached.
    2

Now I have downloaded just 18449 videos, and I notice that there are 19228 csv files instead of original 19994. So we could not download them all, right?

Request classification results

Thank you for your contribution! I've learned a lot from your code. The caffe files about anet_16 and UntrimmedNet can't compiled on my computer. Could you please send me the classification results of anet and UntrimmedNet on ActivityNet and THUMOS data sets respectively?
email : [email protected]

Inconsistency in ground truth region generation

Hi, I think there is some inconsistency between the code and your paper in terms of how ground truth regions are been generated. In the TEM_load_data.py file, you have:

gt_lens=gt_xmaxs-gt_xmins
gt_len_small=np.maximum(tgap,0.1*gt_lens)
gt_start_bboxs=np.stack((gt_xmins-gt_len_small/2,gt_xmins+gt_len_small/2),axis=1)
gt_end_bboxs=np.stack((gt_xmaxs-gt_len_small/2,gt_xmaxs+gt_len_small/2),axis=1)

I believe that this correspond to 3.4 part of the paper where you describe TEM training as:

For ground truth action instance φ g = (t s , t e ) in Ψ ω , we denote its region
as action region r ... as r = [t s − d g /10, t s + d g /10]
and r = [t e − d g /10, t e + d g /10] separately, where d g = t e − t s ...

So it seems that in your paper the start and end regions are generated by +/- one-tenth of the ground truth duration. But in your code it is +/- duration/20 (divide by 10 and then divide by 2 again).

It will great if you can help clarify this part. Thank you.

Thumos annotation files

Hi! Thanks a lot for the code! I am referring to this model as well as BMN model for my own work and I need to test my model on the Thumos dataset too. Can you please share the annotation files for the THUMOS dataset? I would really appreciate it if you do!
Thanks!

About the TSN feature downloading

Hi~ I've downloaded the tsn feature from Baidu Yun, but I found the total number of .csv files in the csv_mean_100 directory is 3992 after unzip. I want to know where are the video features for all the videos?

Training code for THUMOS14

Thank you for your contribution.Limited by the computer, I could only do experiments on the small data set THUMOS14.However, there was something I did not understand about the training process of THUMOS14, which caused me to fail to achieve the effect in the paper.For example, the setting of the TEM training set in THUMOS14 (the sliding window size is 100, so how to regularize time for each group) and the extraction of BSP features.
Could you please send the code to me? E-mail: 349249680 @qq.com
Thank you very much!

Sharing proposals of THUMOS14?

Thanks for sharing your work on github!
Is it possible to share the THUMOS14 proposal results? We do not need the code or model, just a json file with the proposals (video-id, start, end, score) would be fine. We want to compare our proposal methods with this and do some analysis.
Many thanks.

question about action detection performance on activity-net

Hello, Tianwei~ I tried to reimplement your action detection experiments on activity-net, but only get a poor result. Here is my results using different iou thresh on NMS:

IOU thresh [email protected] [email protected] [email protected] average
0.3 0.450 0.274 0.078 0.281
0.4 0.455 0.280 0.078 0.287
0.5 0.450 0.288 0.078 0.292
0.6 0.437 0.298 0.079 0.293
0.7 0.422 0.304 0.080 0.291

I found average number of proposals is far smaller than 100. Does it matter?

I have checked my code several times, but can not find any mistake. Can you help me? Or share your code which generates the detection result. Many thanks.

How to train this on a custom dataset ?

I've been looking at this for the past few days and I can't seem to figure out how I can use this to train on any other dataset than Thumos or ActivityNet. How would you train this in the end-to-end fashion proposed in the article please ?

400-d vectors in csv_mean_100.csv file

Hi, as you mentioned in the paper, RGB and optical flow features are extracted by two-stream network.
So I am wondering what is the meaning of 400-d vector in .csv file?
Are they ordered? For example, the first 200-d means RGB feature and the rest means optical?

question about poolData

Hi, Tianwei Lin
Thanks for your BSN codes sharing.
From paper and the function poolData code in this project, I noticed that there is another operation before you pass TSN features to BSN network.
That operation does interpolation from var-length TSN feature to fixed-length processed feature.

I have read your BSN paper and ActivityNet 2017 paper, but can't find much instruction about that.
Can you give me some reference for it? Any paper or forum discussion will be appreciated.

Another question, as BSN network accepted fix-length feature, how about extract fixed number of frames (and optical flow) for every single video (comparison to extract 1 snippet every 16 frames)?
I think the poolData function, after the TSN feature extraction stage, is the reason why we need to use a pre-trained TSN network. otherwise, we can train the feature extraction network (RGB and Flow) together with the BSN (and BMN ) network online, which may achieve better results.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.