Giter VIP home page Giter VIP logo

Comments (7)

pengyizhou avatar pengyizhou commented on August 26, 2024

Is it wrong to train an i-vector or x-vector model using language-id instead of speaker-id in lid? Because if we see language id as speaker id in baseline system, we get a better results than using the original speaker id to train the model.

from asv-subtools.

Snowdar avatar Snowdar commented on August 26, 2024

from asv-subtools.

pengyizhou avatar pengyizhou commented on August 26, 2024

Thanks a lot!

from asv-subtools.

ASR2020Guru avatar ASR2020Guru commented on August 26, 2024

Hi @Snowdar,

Thanks for this killing toolkit. I have a followup question:

For the OLR2020 baseline and LID task, the training dataset includes multiple spks(each spk has multiple utterances and one language label). And the test dataset includes multiple spks which not been seen in the training dataset before.

From my understanding based on your answer, I see that the difference between i-vector/-vector SRE task and i-vector/-vector LID task is use either spk-id or languge-id as label.
For the LID task, it is true that we should use language-id as the label. For example, if we do an LID task on such dataset which has 6 different languages, each language has 1000 spks, each spk has only one utterance. It is easy to explain that trained model successfully learned the language-related feature not the spk-related.

But if we do an LID task on such dataset which has 6 languages, each language has 30 spks, each spk has 300 utterances. If we use language-id as label here, 24 spk for train and 6 for test. How do we know the trained model can successfully predict the language label because it learned the language-related features or spk-related features?

I am confused about this part.
Thanks in advance

from asv-subtools.

Snowdar avatar Snowdar commented on August 26, 2024

from asv-subtools.

ASR2020Guru avatar ASR2020Guru commented on August 26, 2024

Hi @Snowdar,

Thanks for your reply.

For the training loss and validation loss in the training process, how we split the training dataset in the training part and validation part. Do we split the training dataset by different spk? In such case, the spks in the validation part do not appeared in the training part.
Or we shuffled the whole training dataset first, and then i.e., use the first 90% portion as the training part and the rest as the validation part? In such case, the spks in the validation part could appeared in the training part as well.

Thanks

from asv-subtools.

Snowdar avatar Snowdar commented on August 26, 2024

from asv-subtools.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.