Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to get bert_ft.h5 for myself dataset #3

Open
wangbq18 opened this issue Jun 20, 2021 · 9 comments
Open

How to get bert_ft.h5 for myself dataset #3

wangbq18 opened this issue Jun 20, 2021 · 9 comments

Comments

@wangbq18
Copy link

wangbq18 commented Jun 20, 2021

Hello, I want to ask you another question, how to get bert_ft.h5 for myself dataset? how to encode qusetion and answer with bert? Are they separate or together ? thanks!

@doc-doc
Copy link
Owner

doc-doc commented Jun 20, 2021

Hi, please refer to the link given in readme & our paper. Answers are appended behind the corresponding question for multi-choice QA.

@wangbq18
Copy link
Author

Hi, please refer to the link given in readme & our paper. Answers are appended behind the corresponding question for multi-choice QA.

OK, I see. a another question, how to get motion feature with a shape as (16, 2048). With code provided by [HCRN], the motion feature shape is (8, 2048) with 8 clips, Dose that mean I should set clips=16? And your paper said the best performance is
from using ResNet as an appearance feature along with I3D ResNeXt as a motion feature (Res+I3D), How to get I3D feature. Can you share the code?

@doc-doc
Copy link
Owner

doc-doc commented Jun 21, 2021

Hi, we use I3D with ResNeXt as backbone to capture motion info. The code can also be found in HCRN. The number of sampled clips depends on your dataset, usually ranges from 8~32..

@wangbq18
Copy link
Author

Hi, we use I3D with ResNeXt as backbone to capture motion info. The code can also be found in HCRN. The number of sampled clips depends on your dataset, usually ranges from 8~32..

Thanks a lot, I have solved the problem above. There is no HCRN model implementation base on bert, I try to implementation, but When I repalce glove with bert, It doesn't convergence. Can you share the code?

@doc-doc
Copy link
Owner

doc-doc commented Jun 25, 2021

You need to finetune BERT on your own dataset, and then extract token representations for sentences. Afterwards, you can use the extracted BERT features to replace the GloVe embedding layer in HCRN. You can learn from NExT-QA (this repo.) on how to replace GloVe with BERT features. We are not going to release this part of work so far..

@doc-doc
Copy link
Owner

doc-doc commented Nov 17, 2021

Hi, we use I3D with ResNeXt as backbone to capture motion info. The code can also be found in HCRN. The number of sampled clips depends on your dataset, usually ranges from 8~32..

Thanks a lot, I have solved the problem above. There is no HCRN model implementation base on bert, I try to implementation, but When I repalce glove with bert, It doesn't convergence. Can you share the code?

Hi, HCRN-BERT implementation is available here.

@doc-doc
Copy link
Owner

doc-doc commented Aug 1, 2022

Hi, we have released the edited code for fintuning BERT on NExT-QA here. You can also fine-tune other datasets by using the code.

@PolarisHsu
Copy link

Hi, we have released the edited code for fintuning BERT on NExT-QA here. You can also fine-tune other datasets by using the code.

Hi, this link has expired, can you provide it again?

@doc-doc
Copy link
Owner

doc-doc commented Apr 30, 2023

Yes. Please download it via this link.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants