Automatic Speaker Verification Challenge - Overview
Welcome to the Automatic Speaker Verification (ASV) Challenge! This competition focuses on developing and evaluating speaker verification systems that can accurately determine whether two speech utterances belong to the same speaker.
Speaker verification is a biometric authentication technology that verifies a person's claimed identity based on their voice characteristics. It answers the question: "Is this the person they claim to be?"
In this challenge, participants will develop systems that:
This challenge uses the Vietnam-Celeb dataset, which is a large-scale spontaneous dataset gathered under noisy environments, with over 87,000 utterances from 1,000 Vietnamese speakers of many professions, covering 3 main Vietnamese dialects.
Key features of the dataset:
Systems will be evaluated using the Equal Error Rate (EER) metric, which is the operating point where the False Acceptance Rate (FAR) equals the False Rejection Rate (FRR). Lower EER values indicate better performance.
For questions or support, please use the competition forum or contact the organizers at [email protected], cc-ing [email protected], [email protected].
Automatic Speaker Verification Challenge - Evaluation
The primary evaluation metric for this challenge is the Equal Error Rate (EER), expressed as a percentage (%). EER is the operating point where the False Acceptance Rate (FAR) equals the False Rejection Rate (FRR).
In speaker verification:
The EER is the point where FAR = FRR. Lower EER values indicate better performance. A perfect system would have an EER of 0%, while a random classifier would have an EER around 50%.
We use the following implementation to calculate EER: https://github.com/wenet-e2e/wespeaker/blob/310a15850895b54e20845e107b54c9a275d39a2d/wespeaker/utils/score_metrics.py#L79
Your submission should be a ZIP file, containing a text file named predictions.txt
, each line is the similarity score for one trial pair. Please note that the order of the lines should exactly match the order of the trials in the provided test list. For example, if the first line of the test list is "audio1.wav audio2.wav", the first line of your predictions.txt should be the similarity score for that pair.
The evaluation process works as follows:
For example, if your system produces these similarity scores:
0.75 0.23 0.92
And the ground truth is:
test_001 enroll_005 1 # Same speaker test_002 enroll_008 0 # Different speakers test_003 enroll_003 1 # Same speaker
Then your system would be performing well because it assigned higher scores to same-speaker pairs and lower scores to different-speaker pairs.
Participants will be ranked based on their EER score, with lower values being better. In case of ties, earlier submissions will be ranked higher.
Automatic Speaker Verification Challenge - Terms and Conditions
The challenge organizers reserve the right to modify these terms and conditions at any time. Participants will be notified of any changes. The decisions of the challenge organizers regarding any aspect of the competition are final.
For questions or clarifications regarding these terms, please contact the challenge organizers at [email protected], cc-ing [email protected], [email protected].
Q: Why is my result showing 50% EER?
A: By default, if the function to calculate EER fails to run, it will return the result of 50% EER (which is similar to random guessing). You can check the output log and error log of your submission in the submission window to debug. Please make sure that:
- The score matches the order of the pairs provided in the test list
- The file name is exactly "predictions.txt", with an "s". The text file is zipped in an archive with any name
- Make sure the score is float, not string, and not encapsulated with any quotes.
With that being said, if your result is somewhat even worse than 50%. Congratulations! Your score is bug-free, yet hopeless result for now. Try to improve that!
Q: What dataset can I use for training and development?
A: You can use Vietnam-Celeb-T (provided, downloadable) as the training set with any other datasets you can find. However, you have to disclose the dataset publicly on our Forum.
For development, you can use Vietnam-Celeb-E and Vietnam-Celeb-H provided in the Vietnam-Celeb dataset. The public and private test sets in the challenge are completely different and never published before.
Start: April 1, 2025, midnight
Start: Sept. 1, 2025, midnight
Dec. 31, 2025, midnight
You must be logged in to participate in competitions.
Sign In# | Username | Score |
---|---|---|
1 | vuhl | 3.880 |