Skip to content

Commit a1fab49

Browse files
author
YapengTian
authored
Update README.md
1 parent b25238d commit a1fab49

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

README.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ python supervised_main.py --model_name AV_att --train
3333

3434

3535
### Weakly-supervised audio-visual event localization
36-
We add some videos without audio-visual events into training data. Therefore, the labels of these videos are background. Processed visual features can be found in [visual_feature_noisy.h5](https://drive.google.com/file/d/1I3OtOHJ8G1-v5G2dHIGCfevHQPn-QyLh/view?usp=sharing).
36+
We add some videos without audio-visual events into training data. Therefore, the labels of these videos are background. Processed visual features can be found in [visual_feature_noisy.h5](https://drive.google.com/file/d/1I3OtOHJ8G1-v5G2dHIGCfevHQPn-QyLh/view?usp=sharing). Put the feature into data folder.
3737

3838
Testing:
3939

@@ -44,6 +44,7 @@ Training:
4444
python weak_supervised_main.py --train
4545

4646
### Cross-modality localization
47+
For this task, we developed a cross-modal matching network. Here, we used visual feature vectors via global average pooling, and you can find [here](https://drive.google.com/file/d/1l-c8Kpr5SZ37h-NpL7o9u8YXBNVlX_Si/view?usp=sharing). Please put the feature into data folder.
4748

4849

4950
### Citation

0 commit comments

Comments
 (0)