Spaces:
Running
title: phone_errors
tags:
- evaluate
- metric
description: >-
Error rates in terms of distance between articulatory phonological features
can help understand differences between strings in the International Phonetic
Alphabet (IPA) in a linguistically motivated way. This is useful when
evaluating speech recognition or orthographic to IPA conversion tasks.
sdk: gradio
sdk_version: 3.19.1
app_file: app.py
pinned: false
Metric Card for Phone Errors
Metric Description
Error rates in terms of distance between articulatory phonological features can help understand differences between strings in the International Phonetic Alphabet (IPA) in a linguistically motivated way. This is useful when evaluating speech recognition or orthographic to IPA conversion tasks. These are Levenshtein distances for comparing strings where the smallest unit of measurement is based on phones or articulatory phonological features, rather than Unicode characters.
How to Use
import evaluate
phone_errors = evaluate.load("ginic/phone_errors")
phone_errors.compute(predictions=["bob", "ði"], references=["pop", "ðə"])
Inputs
- predictions (
list
ofstr
): Transcriptions to score. - references (
list
ofstr
) : Reference strings serving as ground truth. - feature_model (
str
): Set which panphon.distance.Distance feature parsing model is used, choose from"strict"
,"permissive"
,"segment"
. Defaults to"segment"
. - is_normalize_pfer (
bool
): Set toTrue
to normalize PFER by the largest number of phones in the prediction, reference pair. Defaults toFalse
. When this is used PFER will no longer obey the triangle inequality.
Output Values
The computation returns a dictionary with the following key and values:
- phone_error_rates (
list
offloat
): Phone error rate (PER) gives edit distance in terms of phones for each prediction-reference pair, rather than Unicode characters, since phones can consist of multiple characters. It is normalized by the number of phones of the reference string. The result with have the same length as the input prediction and reference lists. - mean_phone_error_rate (
float
): Overall mean of PER. - phone_feature_error_rates (
list
offloat
): Phone feature error rate (PFER) is Levenshtein distance between strings where distance between individual phones is computed using Hamming distance between phonetic features for each prediction-reference pair. By default it is a metric that obeys the triangle equality, but can also be normalized by number of phones. - mean_phone_feature_error_rate (
float
): Overall mean of PFER. - feature_error_rates (
list
offloat
): Feature error rate (FER) is the edit distance in terms of articulatory features normalized by the number of phones in the reference, computed for each prediction-reference pair. - mean_feature_error_rate (
float
): Overall mean of FER.
Values from Popular Papers
Universal Automatic Phonetic Transcription into the International Phonetic Alphabet (Taguchi et al.) reported an overall PER of 0.21 and PFER of 0.057 on supervised phonetic transcription of in-domain languages, a PER of 0.632 and PFER of 0.213 on zero-shot phonetic transcription of languages not seen in training data. On the zero-shot languages they also reported inter-annotator scores between human annotators as PER 0.533 and PFER 0.196.
Examples
Simplest use case to compute phone error rates between two IPA strings:
>>> phone_errors.compute(predictions=["bob", "ði", "spin"], references=["pop", "ðə", "spʰin"])
{'phone_error_rates': [0.6666666666666666, 0.5, 0.25], 'mean_phone_error_rate': 0.47222222222222215,
'phone_feature_error_rates': [0.08333333333333333, 0.125, 0.041666666666666664], 'mean_phone_feature_error_rate': 0.08333333333333333,
'feature_error_rates': [0.027777777777777776, 0.0625, 0.30208333333333337], 'mean_feature_error_rate': 0.13078703703703706}
Normalize phone feature error rate by the length of the reference string:
>>> phone_errors.compute(predictions=["bob", "ði"], references=["pop", "ðə"], is_normalize_pfer=True)
{'phone_error_rates': [0.6666666666666666, 0.5], 'mean_phone_error_rate': 0.5833333333333333,
'phone_feature_error_rates': [0.027777777777777776, 0.0625], 'mean_phone_feature_error_rate': 0.04513888888888889,
'feature_error_rates': [0.027777777777777776, 0.0625], 'mean_feature_error_rate': 0.04513888888888889}
Error rates may be greater than 1.0 if the reference string is shorter than the prediction string:
>>> phone_errors.compute(predictions=["bob"], references=["po"])
{'phone_error_rates': [1.0], 'mean_phone_error_rate': 1.0,
'phone_feature_error_rates': [1.0416666666666667], 'mean_phone_feature_error_rate': 1.0416666666666667,
'feature_error_rates': [0.020833333333333332], 'mean_feature_error_rate': 0.020833333333333332}
Empty reference strings will cause an ValueError, you should handle them separately:
>>> phone_errors.compute(predictions=["bob"], references=[""])
Traceback (most recent call last):
...
raise ValueError("one or more references are empty strings")
ValueError: one or more references are empty strings
Limitations and Bias
- Phone error rate and feature error rate can be greater than 1.0 if the reference string is shorter than the prediction string.
- Since these are error rates, not edit distances, the reference strings cannot be empty.
Citation
@inproceedings{Mortensen-et-al:2016,
author = {David R. Mortensen and
Patrick Littell and
Akash Bharadwaj and
Kartik Goyal and
Chris Dyer and
Lori S. Levin},
title = {PanPhon: {A} Resource for Mapping {IPA} Segments to Articulatory Feature Vectors},
booktitle = {Proceedings of {COLING} 2016, the 26th International Conference on Computational Linguistics: Technical Papers},
pages = {3475--3484},
publisher = {{ACL}},
year = {2016}
}
Further References
- PER and PFER are used as evaluation metrics in Universal Automatic Phonetic Transcription into the International Phonetic Alphabet (Taguchi et al.)
- Pierce Darragh's blog post Introduction to Phonology, Part 3: Phonetic Features gives an overview of phonetic features for speech sounds.
- panphon Github repository