-
Notifications
You must be signed in to change notification settings - Fork 24
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
VTLN across genders #6
Comments
Hi! |
I think computing one VTLN coefficient per audio sample may take too long (for example if we have 100h training data)? Another issue is inference time. If we have 1 sample during test time, warp factor will be 1.0. |
I wanna note my observations here. I seperated dataset into 2 speaker as I said and run mfcc_vtln.py. When I limit duration with 10min and 20min, warp factors become 1.0 for each speaker. I am incresing duration limit. |
Hi!
I am trying to extract mfcc+vtln features for a dataset, but I couldn't get the pipeline exactly. When I run mfcc_vtln.py, I get warp factors for each speakers. So we can use these warps for normalization of dataset. However if we have no speaker ids in dataset, but have female/male labels, can we do vtln across genders? I mean, in this setup, we have only 2 speaker_id which are female and male. Do you think if this will work?
The text was updated successfully, but these errors were encountered: