Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ActivityNet 1.2 feature #1

Open
zhenyingfang opened this issue Feb 28, 2019 · 10 comments
Open

ActivityNet 1.2 feature #1

zhenyingfang opened this issue Feb 28, 2019 · 10 comments

Comments

@zhenyingfang
Copy link

Hi, will you release the feature of ActivityNet 1.2?

@Rheelt
Copy link

Rheelt commented Mar 29, 2019

@sujoyp @zhenyingfang Hi! Recently, I re-implemented the code of W-TALC on ActivityNet v1.2. I obtain the classification mAP of 94% on ActivityNet v1.2. But the detection performance is bad. Why is that? Is it parameter setting?

@sujoyp
Copy link
Owner

sujoyp commented Mar 30, 2019

@Rheelt We smoothened the output of our model before applying the threshold for detection in ActivityNet1.2. Good to hear that you obtained better results than what we reported in the paper.

@zhenyingfang you can find the features in the link under the Data section.

@Rheelt
Copy link

Rheelt commented Mar 31, 2019

@sujoyp Thanks for your reply. I appreciate your help very much. I downloaded the features of ActivityNet v1.2. I used the following Settings to get [email protected]=16.62.
def smooth(v): l = min(351, len(v)); l = l - (1-l%2) if len(v) <= 3: return v return savgol_filter(v, l, 1) threshold = np.max(tmp) - (np.max(tmp) - np.min(tmp))*0.5
But But when I change the threshold to threshold = np.max(tmp) - (np.max(tmp) - np.min(tmp))*0.7, I got [email protected] = 33.99. I think it's because the action instances in the ActivityNet v1.2 dataset occupy most of the video and each video contains only 1.5 action instances. So turning the threshold down can get better results. Is there any skill in parameter setting to get the results in the paper?

@sujoyp
Copy link
Owner

sujoyp commented Apr 1, 2019

@Rheelt we actually used the savgol_filter, which is already there in the smooth() function of detectionMAP.py

@sunsiyang2015
Copy link

hi, when I train on ActivityNet v1.2. the I got mIOU is nan.
how to solve the problem?

@sunsiyang2015
Copy link

hi, when I train on ActivityNet v1.2. the I got mIOU is nan.
how to solve the problem?

@Rheelt we actually used the savgol_filter, which is already there in the smooth() function of detectionMAP.py

hi, when I train on ActivityNet v1.2. the I got mIOU is nan.
how to solve the problem?

@Rheelt
Copy link

Rheelt commented Jul 29, 2019

@sunsiyang2015 First, check whether detection weight is right. Then, check evaluation code

@sunsiyang2015
Copy link

@sunsiyang2015 First, check whether detection weight is right. Then, check evaluation code

I use the reference evaluation code as same as detectionMAP.py

@sunsiyang2015
Copy link

@sunsiyang2015 First, check whether detection weight is right. Then, check evaluation code

@sujoyp Thanks for your reply. I appreciate your help very much. I downloaded the features of ActivityNet v1.2. I used the following Settings to get [email protected]=16.62.
def smooth(v): l = min(351, len(v)); l = l - (1-l%2) if len(v) <= 3: return v return savgol_filter(v, l, 1) threshold = np.max(tmp) - (np.max(tmp) - np.min(tmp))*0.5
But But when I change the threshold to threshold = np.max(tmp) - (np.max(tmp) - np.min(tmp))*0.7, I got [email protected] = 33.99. I think it's because the action instances in the ActivityNet v1.2 dataset occupy most of the video and each video contains only 1.5 action instances. So turning the threshold down can get better results. Is there any skill in parameter setting to get the results in the paper?

@sujoyp Thanks for your reply. I appreciate your help very much. I downloaded the features of ActivityNet v1.2. I used the following Settings to get [email protected]=16.62.
def smooth(v): l = min(351, len(v)); l = l - (1-l%2) if len(v) <= 3: return v return savgol_filter(v, l, 1) threshold = np.max(tmp) - (np.max(tmp) - np.min(tmp))*0.5
But But when I change the threshold to threshold = np.max(tmp) - (np.max(tmp) - np.min(tmp))*0.7, I got [email protected] = 33.99. I think it's because the action instances in the ActivityNet v1.2 dataset occupy most of the video and each video contains only 1.5 action instances. So turning the threshold down can get better results. Is there any skill in parameter setting to get the results in the paper?

I use the reference evaluation code as same as detectionMAP.py, but the result is nan.

@GGQ1996
Copy link

GGQ1996 commented Aug 17, 2019

@sujoyp Hi ! Thanks for your excellent work. Can you share the parameter (window length) of savgol_filter?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants