You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In the Hubert model convolutional positional encoding - add a support for batch norm instead of weight norm as added in Fairseq at - facebookresearch/fairseq@4db2649
This will give the opportunity to support other textlesslib/fairseq models and specifically the newer, better HuBERT model - mhubert-base-25hz which uses this and is currently unsupported.
This model is frequently used for training speech language models, and would benefit the community as well as myself in a project I am working on.
Your contribution
I can create a PR to implement this, but would love some guidance @ylacombe
The text was updated successfully, but these errors were encountered:
Would you like to open a PR to correct this ?
You'd have to add the possibility to use batch norm in the configuration_hubert.py, propagate the change to the modeling file and the converting file, and finally add an integration test. How does it sound?
Feature request
Motivation
Your contribution
I can create a PR to implement this, but would love some guidance @ylacombe
The text was updated successfully, but these errors were encountered: