Preprint
Article

This version is not peer-reviewed.

Comparative Analysis of Audio Features for Unsupervised Speaker Change Detection

A peer-reviewed article of this preprint also exists.

Submitted:

14 October 2024

Posted:

16 October 2024

You are already at the latest version

Abstract
This study analyzes how various audio features influence speaker change detection in different unsupervised approaches. Ten different audio features, including MFCC, mel-spectrogram, spectral centroid, and pitch, etc. were chosen for analysis. Two distinct unsupervised methods were selected for comparing the effectiveness of these feature: the Bayesian information criterion with Gaussian mixture model (BIC-GMM), a model-based approach, and the Kullback-Leibler divergence with Gaussian (KL-GMM), a metric-based approach. In order to gain a prior insight into which features have a significant impact on the SCD problem, we calculate statistic for two probabilities: the probability of speaker change given there is a feature change, and vice versa. To experimentally check the prior insights, a set of experiments carried out to analyze the influence of various audio feature for SCD. The results showed that in both methods, MFCC demonstrates a robust performance, and ranks at the top. Zero crossing rate, chroma and spectral contrast gives a higher performance in BIC-GMM. Mel-spectrogram consistently ranks at the bottom, suggesting it has the least impact on performance across both methods.
Keywords: 
;  ;  

1. Introduction

Speaker change detection is an important task in speech processing, vital for applications like speaker diarization [1,2], automated transcription [3], and audio indexing [4]. The goal of speaker change detection is to identify timestamps in an audio stream where a speaker’s segment ends and a new one begins, enabling accurate segmentation of speech data. With the rapid growth of audio data in various domains, from meetings and lectures to podcasts and media archives, the demand for accurate and efficient speaker change point detection has significantly increased. This surge in data volume necessitates robust methods capable of handling diverse and challenging scenarios. A fast and simple approach to SCD is therefore essential to process this vast array of audio efficiently, ensuring that these applications can operate effectively even in the face of increasing data complexity. Given the complex nature of audio signals, which often include varying pitch, tone, and amplitude, the task becomes even more challenging when multiple voices overlap, background noise is present.
Speaker change detection has been studied for more than a decade, and unsupervised methods for detecting speaker change points using metrics have become representative. In these methods, speaker change points are detected by comparing features from two consecutive audio frames. For instance, in a approach of Bayesian information criterion (BIC) [5], audio features are extracted from audio frames, then the distribution of the speech feature is assumed to follow mixture Gaussian distribution. The speaker change points are identified by evaluating two hypotheses from these distribution of consecutive frames: i) the first hypothesis assumes that the left and right segments of speech originate from the same speaker, ii) while the second hypothesis posits that the two segments come from different speakers.
Numerous criteria for the metric-based approach have been proposed to assess the likelihood of the two hypotheses, such as the Kullback Leibler (KL) distance [6], generalized likelihood ratio (GLR) [7,8]. Among these criteria, KL has been the most widely used method, along with its numerous variants [9,10]. Recent developments in audio feature extraction, such as i-vector[11] and deep learning-based approaches [12,13], have changed the SCD method to uniform segmentation, as varying segment lengths introduced additional variability into speaker representation and deteriorated the fidelity of the speaker representations.
Despite advancements in feature extraction of audio data and modeling techniques for SCD, there’s still a gap in understanding how various audio features influence different methods in SCD, particularly in unsupervised settings. Audio features such as mel-frequency cepstral coefficients (MFCC), pitch, and mel-spectrogram, etc, and they contain different aspects of speech, but their effectiveness in identifying speaker changes is not clearly studied. Without a thorough comparative analysis, it’s challenging to pinpoint which features are most effective for unsupervised SCD.
In this paper, an empirical comparative study using various audio features with different methods for unsupervised SCD is presented. The goal of study is to analyze how various audio features influence the different method in SCD. In order to conduct an comparative analysis, various features were analyzed such as MFCC, mel-spectrogram, spectral centroid, spectral bandwidth, spectral contrast, spectral rolloff, RMS, pitch, zero crossing rate. These audio features were evaluated for two methods: BIC-GMM and KL divergence with GMM, the former approach is no need threshold optimization, but the latter does.
A set of experiments were carried out for comparing the combinations of method across features :i) optimizing thresholds with Bayesian optimization for KL-GMM, ii) reporting two naive systems as baselines for class imbalance, iii) evaluating feature-method combinations on test data, and iv) averaging results to identify the most suitable features for SCD. v) comparing the relation between maximum probability of speaker change given feature change with results of the method.
Experimental results showed that in both methods, MFCC demonstrates a robust performance, and ranks at the top. Zero crossing rate, chroma and spectral contrast gives a higher performance in BIC-GMM. Mel-spectrogram consistently ranks at the bottom, suggesting it has the least impact on performance across both methods.

2. Related Work

In general, machine learning approaches [14] designed for speaker change point detection contains two stages: i) audio feature extraction, and ii) change point detection stage. Many approaches with these stages have been studied for speaker change point detection in recent decades, and examined different audio features with different detection method.
In the first stage, a number of audio features have been applied for SCD [15,16,17]. The paper [15] describes the evaluation of different feature extraction techniques and processing methods in the context of speaker diarization. Their experimental results showed that perceptual linear predictive (PLP) features, particularly when excluding energy components, often outperform MFCC.
The author presented [17] an approach to speech classification, and unsupervised speaker segmentation using linear spectral pair (LSP) divergence analysis. In first step, they classify the audio into four categories: speech, music, environment sound and silence. The zero-crossing rate, short-time energy, spectrum flux were used for characterizing different audio signals. If classify it a speech segment, the process will be involved to detection of speaker transitions. The method includes incremental speaker modeling and adaptive threshold setting, enabling effective unsupervised segmentation. Experimental results demonstrate the algorithm’s effectiveness, achieving an overall recall of 89.89% and a precision of 83.66%. Recent development of features for SCD is to extract speaker-specific information, like i-vector [18,19] and feature vectors [20] from neural networks. d-vector [21] was successfully applied to SCD as feature and yielding a excellent result. The latest trends is to use end-to-end manner[22,23], to extract deep speaker embedding as features.
For the second stage of speaker change detection, the approaches can be divided into three groups: i) metric-based and ii) energy or model-based approach, iii) hybrid method. Metric-based methods for SCD involve measuring the differences between consecutive audio frames as they move through the signal. Distance between these frames are computed and if the value is larger than a predefined threshold, a change point is detected. Frequently used distance measures include the Bayesian information criterion [5,24], the generalized likelihood ratio [7], the Kullback-Leibler divergence [6]. Model-based approaches are usually supervised, train on a labeled audio dataset for detecting speaker changes. Early methods include the GMM [25], and hidden Markov models (HMM) [26]. Latter, deep learning models like DNN [27], CNN [28] and bi-direction LSTMs [29] were applied.

3. Methodology

In order to explore the effectivenness of various audio features and different distance metrics for the SCD. We use unsupervised metric-based approach, which computes the distance between two consecutive audio features. If the distance value is larger than a predefined threshold, then it is considered as speaker change point.

3.1. Task Definition

Given an audio signal S, the task of SCD is to find a set of change points for the segments, denoted as Y = [ y 1 , , y m ] , where y i R r , r is the total number of frames in each window, and m is the total number of segments.
Let X i = { x i } i = 1 n , x i R d represent the sequence of feature vectors of segment y i , where x i is the features and n is the sample size, d is dimension of feature vector.
Generally, speaker change point could be detected by calculating a distance or divergence between feature distributions of two adjacent audio segments and it can be represented as:
  • a first segment of X i feature vector, the sequence of feature vectors from time t i to t i + 1 .
  • a second segment X i + 1 , the sequence of feature vectors from time t i + 1 to t i + 2 .
A speaker change is detected at time t i + 1 if the distance between these two segments exceeds a threshold τ :
d X i , X i + 1 > τ .
Where d is a distance function, and τ is a threshold to detect changes in speaker.

3.2. Change Point Detection

In order to examine various features in a different settings, in this study, two methods are used for speaker change detection: Bayesian information criterion with Gaussian mixture model (BIC-GMM) and KL divergence with Gaussian mixture model (KL-GMM) method. BIC-GMM, which is based on model selection, and KL-GMM, which is metric-based.

3.2.1. BIC-GMM Method

This method uses the Bayesian information criterion to compare consecutive audio segments X i and X i + 1 . It evaluates the following two hypotheses:
  • H 0 : both segments are derived from a Gaussian { X i , X i + 1 } N ( μ , Σ ) distribution, in other words, both segments belong to the same speaker.
  • H 1 : both segments are derived from two distinct Gaussians X i N ( μ 1 , Σ 1 ) and X i + 1 N ( μ 2 , Σ 2 ) . The segments belong to different speakers.
With these hypotheses in mind, a BIC difference Δ B I C is calculated as:
Δ B I C = ( N 1 + N 2 ) log | Σ | N 1 log | Σ 1 | N 2 log | Σ 2 | λ P
P = 1 2 d + 1 2 d ( d + 1 ) log ( N )
where P is penalty term and N = N 1 + N 2 are the sample sizes of the segments, and d is the dimension of the feature vector, and λ is a penalty weight.
The larger value of Δ B I C i , the less similar the two segments are, indicating a greater distance between them. Instead of setting a threshold to it, we compare two Δ B I C i 1 value, if Δ B I C i > Δ B I C i 1 , a change point is detected; otherwise, it is not.

3.2.2. KL Divergence

In this method, audio segments are modeled as Gaussian distributions. The KL divergence between two segments X i N ( μ 1 , Σ 1 ) and X i + 1 N ( μ 2 , Σ 2 ) is computed as:
D K L ( X i X i + 1 ) = 1 2 tr ( Σ 2 1 Σ 1 ) + ( μ 2 μ 1 ) T Σ 2 1 ( μ 2 μ 1 ) d + log | Σ 2 | | Σ 1 |
If this difference is larger than the given threshold, it indicates a change point; otherwise, it does not. This approach requires specifically choosing the threshold value, as it is sensitive to this value. An inappropriate threshold may lead to either missed detections or false alarms, affecting the accuracy of change point detection.

3.3. Audio Features

As known, audio features play a vital role in the process of identifying change points. In order to analyze the different features, we choose different categories of feature based on their relevance to different aspects of the audio signal, can be summarized as follows:
  • Mel-frequency cepstral coefficients (MFCC) - captures the timbral and phonetic content, which is useful for distinguishing different speakers.
  • Mel-Spectrogram - similar to the spectrogram but focused on the mel scale, which is more aligned with human hearing.
  • Chroma - captures the harmonic content and pitch class, which can indicate tonal changes that may differentiate speakers.
  • Spectral centroid - represents the "center of mass" of the spectrum, which often relates to the brightness of a sound.
  • Spectral bandwidth - measures the spread of the spectrum, providing information about timbre and energy distribution.
  • Spectral contrast - captures the difference between peaks and valleys in a spectrum, highlighting spectral variation which may signal speaker changes.
  • Spectral rolloff - indicates the frequency below which a certain percentage of the total spectral energy is contained, useful for tracking energy shifts.
  • Root mean square energy (RMS) - represents the power of the audio signal, which can help distinguish speakers based on loudness and energy pattern.
  • Pitch - provides fundamental frequency information that can help differentiate speakers based on vocal pitch.
  • Zero crossing rate - indicates the rate of signal sign changes, which can be useful for detecting changes in speech quality and distinguishing between different speakers.

3.3.1. Feature Changes vs. Speaker Changes

In the above we briefly describe each feature, in order to analyze those features how they useful to speaker change detection, the meetings audio from AMI corpus [30] is used in the following statistical counting. This section outlines the process of detecting changes in audio features and analyzing their co-occurrence with speaker changes.
By identifying points where specific features change (FC), we investigate how these changes align with speaker changes (SC) to inform feature selection for speaker change detection. For each audio feature, values are extracted over time, and points of significant change are identified by analyzing the differences between consecutive frames. This method accommodates both one-dimensional and two-dimensional feature data: i) one-dimensional features (such as pitch) are straightforward; absolute differences between consecutive values are calculated to capture changes over time. ii) two-dimensional features (such as MFCCs) involve multiple components per time frame. In this case, differences across dimensions are aggregated using the mean, which reflects overall changes in the feature values. Using the mean helps capture general trends across all components of a feature, making it suitable for analyzing broad patterns in audio data. Thresholds are applied to these feature differences to mark significant change points. By adjusting the threshold, it is possible to control the sensitivity of change detection for each feature.
After determining the change times for each feature, co-occurrence analysis is conducted to assess how often these feature changes align with speaker changes. A defined time window, set by the tolerance parameter, is used to determine if two events are co-occurred. A tolerance was set to 1 seconds, any feature change within this window of a speaker change is marked as a co-occurrence.
In the Figure 1, we plot thresholds against the probabilities P ( F C | S C ) and P ( S C | F C ) , which represent the likelihood of feature changes given speaker changes, and vice versa, across various audio features.
Across all graphs, P ( F C | S C ) generally starts at higher values at lower thresholds, indicating greater sensitivity to minor feature changes when a speaker change occurs. As thresholds increase, P ( F C | S C ) gradually declines, indicating a reduction in sensitivity as the system filters out smaller changes. Conversely, P ( S C | F C ) tends to rise as the threshold increases, indicating improved specificity. However, it does not reach unity (1) for any feature, meaning that even at high thresholds, the system does not perfectly align feature changes with speaker changes. However, no feature reaches unity (1) for P ( S C | F C ) , suggesting that even at high thresholds, the perfect alignment between feature changes and speaker changes is hard to be observed.
Crossing points where P ( S C | F C ) and P ( F C | S C ) intersect are can be observed. These points represent the thresholds at which sensitivity and specificity are balanced, marking where the two probabilities are equal.
Table 1 provides the crossing points and the maximum observed P ( S C | F C ) values for each feature. Crossing points where P ( F C | S C ) and P ( S C | F C ) intersect, and represent the thresholds at which sensitivity and specificity are balanced, marking where the two probabilities are equal. P ( S C | F C ) value reflects the peak specificity that each feature achieves at a particular threshold, giving insights into how closely feature changes align with speaker changes. The highest values of P ( S C | F C ) are observed for MFCC (0.534), chroma (0.515), pitch (0.40), and all the spectral related features have over 0.39 values. Other features, such as zero crossing rate and RMS energy, have lower maximum values around 0.32, indicating weaker alignment with speaker changes.

4. Experiments

In order to explore the effectiveness of different features and various distance metrics for SCD, we designed a set of experiments and report them in the following order:
i) the threshold is a hyperparameter in metric-based SCD that determines the sensitivity of the detection process. This experiment seeks to effectively identify the optimal τ threshold for different feature and distance metric combinations on a subset of the validation set. ii) considering a situation of imbalanced problem, the dataset may contains many negative samples, while only a small portion are positive samples. Under these conditions, a naive system that only outputs negative sample labels (in this case, 0) would achieve high accuracy but would fail to correctly identify positive samples. iii) evaluate each combination of audio feature with different methods on the test data, and report the evaluation results. iv) averaging the results for each method across all features to identify the most suitable features to SCD. v) compare the empirically obtained results with the statistically obtained insights, and identifying the robust features by visualizing the comparison between the maximum probability of P(SCFC) and a combined score.
Robust features were clearly identified by visualizing the comparison between the maximum probability of P(SCFC) and the methods’ results.
To effectively compare the metric-based approach, we designed an experiment using two naive systems as a baselines for comparisons.

4.1. Dataset and Evaluation Metric

Experiments were conducted using the AMI meeting corpus [30], a audio dataset commonly used for speech and audio processing research. The AMI corpus provided a diverse set of audio recordings that allowed us to evaluate the performance of these distance metrics under various conditions and scenarios. The test data includes 16 meeting’s audios with a total duration of around 9 hours, and an average duration of 33.38 minutes per meeting. The full data statistics are given in Table 2. It can be observed that the change points, overlaps, total number of speakers, and audio length are provided. For this statistic, 40% of the change points are overlapped.
For evaluation, the following metrics were chosen for the comparative analysis. False alarm rate (FAR) indicates the frequency of incorrect detections of speaker changes. It is calculated as the proportion of false positives (FP) among all non-change points:
FAR = False positives ( FP ) False positives ( FP ) + True negatives ( TN )
Missed detection rate (MDR) measures how often the system fails to detect an actual speaker change. It is the proportion of false negatives (FN) among all true change points:
MDR = False negatives ( FN ) True positives ( TP ) + False negatives ( FN )
Performance Score = 1 FAR + MDR 2
Performance score reflects both types of errors (false alarm rate and missed detection rate), this score is only used for threshold optimization.

4.2. Experimental Setup

To conduct a fair and comparable experiments, we define the following points: i) choose 16 k Hz as the main sample rate of audios; ii) set parameter of hop length to 512 and n_fft set to 2048 for extracting the audio feature from raw signals. Each meeting was segmented to short 1s audio, before extracting the features.
Since the true labels of speaker change points are given as timestamps on the raw audio signal, such as beginning time, we match this beginning times for obtain the labels for each segment. The test set was used solely for final evaluation, guaranteeing that the chosen thresholds are fair, as they are not influenced by direct tuning on the test data. Validation set was used for threshold optimization for KL divergence with GMM. For simplicity, in the experiments, we refer to the method of KL-divergence with GMM as KL-GMM. BIC-GMM models are simplified to BIC.
Table 3. Optimal thresholds of KL divergence method for each feature after optimized on validation set with Bayesian optimization.
Table 3. Optimal thresholds of KL divergence method for each feature after optimized on validation set with Bayesian optimization.
Feature Optimal threshold Performance score
pitch 5.999179 0.528707
mfcc 4.331410 0.519567
chroma 9.998953 0.505080
mel_spectrogram 0.604765 0.502307
zero_crossing_rate 0.988847 0.606065
spectral_centroid 0.441515 0.534054
spectral_bandwidth 0.763678 0.547675
spectral_contrast 4.665370 0.498247
spectral_rolloff 0.539069 0.534808
rms 0.259262 0.514598

4.3. Results

In the threshold optimization process for SCD, Bayesian optimization was applied to determine the optimal threshold for each combination of audio feature for KL-GMM , with the aim of maximizing the performance score on the validation set. The optimization process is configured with the following parameters:
  • Threshold parameter is explored within the range of 0 to 10, allowing the optimizer to identify both highly sensitive and more robust threshold values across various feature and metric combinations.
  • Initialization points set to 5, this parameter specifies the number of random points the optimizer initially samples to gain a broad understanding of the threshold’s impact on the combined metric as peformance score.
  • Iterations set to 25, this parameter specifies the number of optimization steps following the initial exploration.
The process begins with a set number of random samples, after which the optimizer uses these observations to inform subsequent iterations. B focusing on the most promising thresholds, Bayesian optimization quickly converges on the values that maximize the performance score. It is shown in Figure 2, the fluctuation is process of optimization, and with the blue cure, the convergence curve with performance score is plotted. The final optimized threshold values are list in Table 3.
Two naive systems considered as baselines for comparison, they are i) naive 0 : it is only output 0 (negative) label. ii) naive 1 : the system only output 1 (positive) label. Table 4 lists the FAR and MDR for the naive systems, which are shown here for comparison purposes with the KL-GMM and BIC-GMM.
Table A1 to Table A10 lists the detailed results of two approaches across different features: chroma, mel-spectrogram, MFCC, pitch, RMS, spectral bandwidth, spectral centroid, spectral contrast, spectral rolloff, and zero crossing rate for different meeting records.
First of all, compared to the naive systems, both KL-GMM and BIC-GMM that processing the audio features, doing a real detection of speaker changes.
They reduce the errors found in naive systems by analyzing the audio, rather than relying on fixed assumptions.
It can be seen from Figure 2 (e, d), although the threshold of KL-GMM with mel-spectrogram (Table A2) feature and RMS feature (Table A5) was optimized for the validation set well, but on the test data, it produces extremely high FAR values (close to 1.0) but very low MDR values, indicating that it detects almost all change points, but at the cost of many false alarms. In this case, the performance of the KL-GMM is similar to the naive 1 system due to improper threshold values. KL-GMM with the pitch (Table A4) feature gives a high MDR value and a low FAR. In this situation, it behaves similarly to the naive 0 system, but slightly better.
However, KL-GMM as a sensitive method to threshold, with feature MFCC (Table A3), and zero cross rate feature (Table A10) shows a balanced result with FAR and MDR across different meeting test sets. This reflects the fact that optimizing a threshold value for KL divergence is challenging and may not always work well across different audio data. Variations in audio signals, such as slight environmental changes or differences in speaker characteristics, can lead to significant changes in feature values, causing the method to fail in detecting speaker changes. BIC-GMM appears less influenced by variability of different feature value, it achieves more stable performance across different meetings. It also maintains more balanced FAR and MDR rates across all features.
Table 5 reports the average results of meetings for different features. It can be observed that the lowest MDR value of 44.99% was achieved by BIC-GMM with the MFCC feature, with a corresponding FAR of 48.37%. KL-GMM with the MFCC feature gave an MDR of 46.71% and an FAR of 54.28%, which is considered a good performance compared to other features. These results indicate that MFCC features is good fit to SCD problem when using KL and BIC techniques. Additionally, the conditional probability of speaker change given a feature change, P ( S C | F C ) M F C C , is 53.40% for the MFCC feature (see Table 1). This value aligns with the experimental result for both BIC-GMM and KL-GMM with MFCC feature. The conditional probability for the chroma feature, P ( S C | F C ) c h r o m a , is 51.49%, showing a similar pattern.
When we calculate the standard deviation for KL divergence and BIC-GMM model, it can be observed that, for KL divergence (0.29 for FAR, 0.28 for MDR), its std values are much higher than BIC-GMM’s (0,011 for FAR, 0.029 for MDR). This indicates that the BIC-GMM model has more consistent performance across different instances, with lower variability in both FAR and MDR. In contrast, the higher standard deviation values for KL divergence suggest that its performance fluctuates more depending on the specific audio data and conditions.
Figure 3 presents the relation between maximum probability of speaker change given feature change Max P ( S C | F C ) with the FAR and MDR results for BIC-GMM and KL-GMM. To compare different features, we need to compare both MDR and FAR value in a balanced way. For example, in BIC-GMM with feature RMS (see sub-figure (a), (b)), its FAR value is lowest among others, but it has high MDR.
It can be observed that for MFCC in BIC-GMM (see sub-figure (a), (b)), the highest probability of P ( S C | F C ) corresponds to the lowest MDR and FAR values. It shows mel-spectrogram has the highest value for both MDR and FAR scores in BIC-GMM (see sub-figure (a), (b)) It may indicate it is not suits to the SCD compared to others.
To clearly demonstrate the effectiveness of different features on two distinct methods, we use a combined score of MDR and FAR by approximating the F1 score:
Approx . F 1 2 × ( 1 FAR ) × ( 1 MDR ) ( 1 FAR ) + ( 1 MDR )
Figure 4 compares the performance of various audio features across two speaker change detection methods: KL-GMM and BIC-GMM, using approx. F1 scores, and shows the relation with Max P ( S C | F C ) . In both methods, MFCC achieving the highest approx. F1 score, which indicates effective detection with a balanced reduction in both false alarms and missed detection. Correspondingly, its Max P ( S C | F C ) value is also the highest.
BIC-GMM method demonstrates more consistent approx. F1 scores across features, with RMS, spectral centroid, spectral bandwidth, and chroma all performing reasonably well. Zero crossing rate ranks highest for KL-GMM, with an approx. F1 and a relatively low Max P ( S C | F C ) value. However, this feature did not outperform others in BIC-GMM (sub-figure (b)), except for mel-spectrogram feature that consistently underperforms across both methods, although it shows slightly improved performance in BIC-GMM.

5. Conclusion

In this paper, an empirical comparative study using various audio features with different methods for unsupervised SCD is presented. A various of audio features were evaluated for two methods: BIC-GMM and KL divergence with GMM, the former approach is no need threshold optimization, but the latter does. Various features were experimented such as MFCC, mel-spectrogram, spectral centroid, spectral bandwidth, spectral contrast, spectral rolloff, RMS, pitch, zero crossing rate. In order to gain insight into which features have a significant impact on the SCD problem, two conditional distributions from a corpus, such as the probability of feature change given a speaker change P ( F C | S C ) and the probability of speaker change given feature change were calculated. Crossing points between these two probabilities and the maximum P ( S C | F C ) for each features were calculated. The crossing points represent the thresholds at which sensitivity and specificity are balanced, marking where the two probabilities are equal. The maximum P ( S C | F C ) giving insights how closely feature changes align with speaker changes.
Different experiments were carried out for comparing the combinations of method across features: i) the threshold was optimized on the validation set for KL-GMM approach using Bayesian optimization process. ii) results for two naive systems were reported as baseline considering the situation of unbalanced class of SCD. iii) evaluate each combination of audio feature with different methods on the test data, and report the evaluation results. iv) averaging the results for each method across all features to identify the most suitable features to SCD.
For evaluation, false alarm rate and missed detection rate were calculated. Experimental results can be summarized as follows:
i) For different features, KL-GMM approache using Bayesian optimization process to find optimal threshold value on the validation set and the optimal value with the optimization process was reported.
ii) Comparing with two naive systems, both KL-GMM and BIC-GMM that processing the audio features, doing a real detection of speaker changes.
iii) KL-GMM peformance with some feature combination similar to two naive systems, like giving extreme high value to FAR or low value to MDR, vice versa. Although, the thresholds are well optimized on the validation set, it still fails to handle the variability of new audio dataset. KL-GMM as a sensitive mehtod to threshold, but with features like MFCC and zero cross rate shows a better performance. BIC-GMM showed a robust performance on handling the variability of different feature values compared it with KL-GMM, and it achieves more stable performance across different meetings.
iv) Average results indicate that both BIC-GMM and KL-GMM perform effectively on the SCD when using MFCC features. This suggests that MFCC features are well-suited for SCD with these techniques. With the highest P ( S C | F C ) value, MFCC, also has the lowest MDR and FAR, confirming the expectations derived from the statistics.
v) The relation between Max P ( S C | F C ) vs approx. F1 score showed that MFCC consistently ranks high across both methods, confirming its effectiveness. Zero crossing rate, chroma and spectral contrast gives a higher performance in BIC-GMM. Mel-spectrogram consistently ranks at the bottom, suggesting it has the least impact on performance across both methods.

Author Contributions

Conceptualization, A.T. and G.T.; methodology, A.T., G.T.; software, A.T., G.T.; validation, A.T., G.T.; formal analysis, A.T, G.T.; investigation, G.T, A.T; resources, A.T.,G.T.; data curation, A.T.; writing—original draft preparation, A.T., G.T.; writing—review and editing, A.T.,G.T.; visualization, A.T., G.T.; supervision, A.T.,G.T.; project administration, A.T., A.K.,R.M.; funding acquisition, G.T., B.Z. All authors have read and agreed to the published version of the manuscript.

Funding

The research was funded by the Committee of Science of Ministry of Science and Higher Education of the Republic of Kazakhstan under the grant number AP19676744.

Data Availability Statement

The data presented in this study are publicly available. The corresponding URLs to the used datasets are provided in the following link. [the AMI Corpus] [accessed on August 01 2023]

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

The following tables report the detailed results of two approaches across various features: chroma, mel-spectrogram, MFCC, pitch, RMS, spectral bandwidth, spectral centroid, spectral contrast, spectral rolloff, and zero crossing rate for different meeting records.
Table A1. Speaker change detection results of KL-GMM and BIC-GMM for chroma.
Table A1. Speaker change detection results of KL-GMM and BIC-GMM for chroma.
Meeting id FAR (KL) MDR (KL) FAR (BIC) MDR (BIC)
EN2002a 0.3081 0.6658 0.4827 0.4496
EN2002b 0.3066 0.7148 0.4951 0.5190
EN2002c 0.2777 0.7174 0.4841 0.4943
EN2002d 0.2776 0.7220 0.4872 0.5007
ES2004a 0.2385 0.6996 0.4939 0.4798
ES2004b 0.2792 0.6868 0.4958 0.4698
ES2004c 0.3121 0.6401 0.4824 0.4547
ES2004d 0.3184 0.6529 0.4912 0.4314
IS1009a 0.3549 0.6474 0.4984 0.4842
IS1009b 0.3046 0.6629 0.4889 0.4600
IS1009c 0.3539 0.6751 0.4903 0.4230
IS1009d 0.3714 0.6183 0.4967 0.4379
TS3003a 0.2170 0.7629 0.4745 0.5015
TS3003b 0.2347 0.7639 0.4884 0.4587
TS3003c 0.2260 0.7374 0.5015 0.4667
TS3003d 0.2422 0.7384 0.5070 0.5139
Table A2. Speaker change detection results of KL-GMM and BIC-GMM for mel-spectrogram.
Table A2. Speaker change detection results of KL-GMM and BIC-GMM for mel-spectrogram.
Meeting id FAR (KL) MDR (KL) FAR (BIC) MDR (BIC)
EN2002a 0.9935 0.0080 0.5173 0.5106
EN2002b 0.9910 0.0114 0.5049 0.5114
EN2002c 0.9913 0.0046 0.5082 0.5034
EN2002d 0.9926 0.0041 0.4973 0.5035
ES2004a 0.9927 0.0090 0.4939 0.5157
ES2004b 0.9932 0.0067 0.5016 0.4944
ES2004c 0.9930 0.0086 0.5089 0.5388
ES2004d 0.9935 0.0066 0.4828 0.4926
IS1009a 0.9968 0.0105 0.5032 0.5000
IS1009b 0.9934 0.0057 0.5111 0.5200
IS1009c 0.9882 0.0056 0.5000 0.4958
IS1009d 0.9960 0.0023 0.4973 0.5012
TS3003a 0.9921 0.0030 0.4991 0.5046
TS3003b 0.9929 0.0019 0.5235 0.5374
TS3003c 0.9908 0.0020 0.5073 0.5556
TS3003d 0.9930 0.0093 0.4979 0.5206
Table A3. Speaker change detection results of KL-GMM and BIC-GMM for mfcc.
Table A3. Speaker change detection results of KL-GMM and BIC-GMM for mfcc.
Meeting id FAR (KL) MDR (KL) FAR (BIC) MDR (BIC)
EN2002a 0.5144 0.5146 0.4697 0.4788
EN2002b 0.5352 0.5190 0.4820 0.4544
EN2002c 0.5140 0.5172 0.4855 0.4554
EN2002d 0.4778 0.5173 0.4818 0.4703
ES2004a 0.5896 0.4260 0.4976 0.4395
ES2004b 0.5432 0.4609 0.4694 0.4340
ES2004c 0.5495 0.4828 0.4900 0.4547
ES2004d 0.5445 0.4479 0.4698 0.4380
IS1009a 0.5394 0.4684 0.4748 0.4421
IS1009b 0.4973 0.4686 0.4955 0.3971
IS1009c 0.5381 0.5042 0.4958 0.4146
IS1009d 0.5272 0.4496 0.4861 0.4543
TS3003a 0.5650 0.4347 0.4754 0.4802
TS3003b 0.6025 0.3877 0.4920 0.4626
TS3003c 0.5795 0.4465 0.4903 0.4626
TS3003d 0.5671 0.4276 0.4828 0.4595
Table A4. Speaker change detection results of KL-GMM and BIC-GMM for pitch.
Table A4. Speaker change detection results of KL-GMM and BIC-GMM for pitch.
Meeting id FAR (KL) MDR (KL) FAR (BIC) MDR (BIC)
EN2002a 0.1429 0.8302 0.5079 0.4841
EN2002b 0.1295 0.8574 0.4918 0.4962
EN2002c 0.1254 0.8467 0.4851 0.4771
EN2002d 0.1294 0.8520 0.5013 0.4965
ES2004a 0.1768 0.8027 0.5048 0.4664
ES2004b 0.1475 0.8367 0.4942 0.5168
ES2004c 0.1303 0.8793 0.5073 0.5172
ES2004d 0.1365 0.8430 0.4867 0.4595
IS1009a 0.2192 0.7842 0.4621 0.5211
IS1009b 0.1349 0.8543 0.4985 0.4886
IS1009c 0.1247 0.8683 0.5007 0.4874
IS1009d 0.1379 0.8712 0.4881 0.4637
TS3003a 0.1397 0.8632 0.5158 0.5106
TS3003b 0.1783 0.8560 0.4979 0.4837
TS3003c 0.1591 0.8869 0.4908 0.4828
TS3003d 0.1396 0.8805 0.4823 0.4794
Table A5. Speaker change detection results of KL-GMM and BIC-GMM for rms.
Table A5. Speaker change detection results of KL-GMM and BIC-GMM for rms.
Meeting id FAR (KL) MDR (KL) FAR (BIC) MDR (BIC)
EN2002a 0.8824 0.0955 0.4834 0.5199
EN2002b 0.8820 0.1464 0.4795 0.4715
EN2002c 0.8722 0.1247 0.4802 0.4966
EN2002d 0.8753 0.1494 0.4656 0.4799
ES2004a 0.8862 0.1076 0.4976 0.4933
ES2004b 0.8825 0.1432 0.4842 0.4810
ES2004c 0.8945 0.1466 0.4808 0.4569
ES2004d 0.8785 0.1223 0.5062 0.4893
IS1009a 0.8438 0.1842 0.4558 0.4368
IS1009b 0.8423 0.1514 0.4792 0.4486
IS1009c 0.8511 0.1905 0.4834 0.4118
IS1009d 0.8727 0.1475 0.4814 0.4965
TS3003a 0.9060 0.1064 0.4763 0.4711
TS3003b 0.9156 0.0806 0.4831 0.4722
TS3003c 0.9093 0.0869 0.4869 0.4505
TS3003d 0.8996 0.0969 0.4828 0.4768
Table A6. Speaker change detection results of KL-GMM and BIC-GMM for spectral bandwidth.
Table A6. Speaker change detection results of KL-GMM and BIC-GMM for spectral bandwidth.
Meeting id FAR (KL) MDR (KL) FAR (BIC) MDR (BIC)
EN2002a 0.6429 0.3820 0.4935 0.4867
EN2002b 0.6336 0.3878 0.4877 0.4658
EN2002c 0.6249 0.3913 0.5034 0.4966
EN2002d 0.5957 0.3817 0.5054 0.4620
ES2004a 0.6320 0.3453 0.4782 0.4619
ES2004b 0.6128 0.3893 0.4984 0.4474
ES2004c 0.6247 0.3405 0.4943 0.4741
ES2004d 0.6277 0.3570 0.4867 0.4810
IS1009a 0.4858 0.4368 0.4842 0.4211
IS1009b 0.5250 0.4600 0.5099 0.5229
IS1009c 0.5111 0.4594 0.4910 0.4650
IS1009d 0.5351 0.4895 0.4934 0.4660
TS3003a 0.5633 0.4134 0.4956 0.4620
TS3003b 0.6352 0.2821 0.4979 0.4683
TS3003c 0.6261 0.3475 0.4850 0.4485
TS3003d 0.6412 0.3267 0.4968 0.4914
Table A7. Speaker change detection results of KL-GMM and BIC-GMM for spectral centroid.
Table A7. Speaker change detection results of KL-GMM and BIC-GMM for spectral centroid.
Meeting id FAR (KL) MDR (KL) FAR (BIC) MDR (BIC)
EN2002a 0.7799 0.1976 0.5159 0.5186
EN2002b 0.7713 0.2053 0.4738 0.4354
EN2002c 0.7686 0.2151 0.4860 0.4886
EN2002d 0.7588 0.2282 0.4933 0.4772
ES2004a 0.7712 0.2915 0.4673 0.4798
ES2004b 0.7571 0.2282 0.4805 0.4765
ES2004c 0.7772 0.2069 0.4922 0.4483
ES2004d 0.7921 0.2264 0.4737 0.4314
IS1009a 0.6893 0.2474 0.4874 0.4211
IS1009b 0.7225 0.2400 0.4937 0.4857
IS1009c 0.7244 0.2745 0.4931 0.4762
IS1009d 0.7341 0.2881 0.4934 0.4848
TS3003a 0.7645 0.2219 0.4859 0.4650
TS3003b 0.7950 0.1919 0.4795 0.4434
TS3003c 0.7706 0.2263 0.4961 0.4606
TS3003d 0.7927 0.1899 0.5000 0.5113
Table A8. Speaker change detection results of KL-GMM and BIC-GMM for spectral contrast.
Table A8. Speaker change detection results of KL-GMM and BIC-GMM for spectral contrast.
Meeting id FAR (KL) MDR (KL) FAR (BIC) MDR (BIC)
EN2002a 0.0866 0.9005 0.4820 0.4788
EN2002b 0.0885 0.8954 0.4754 0.4734
EN2002c 0.0858 0.9005 0.4908 0.4863
EN2002d 0.0876 0.9225 0.4704 0.4716
ES2004a 0.1453 0.8565 0.4879 0.4484
ES2004b 0.1312 0.8591 0.4879 0.4765
ES2004c 0.1195 0.9030 0.4803 0.4224
ES2004d 0.1001 0.8959 0.4886 0.4727
IS1009a 0.0710 0.9105 0.4984 0.4842
IS1009b 0.0662 0.9457 0.4913 0.4543
IS1009c 0.0817 0.9440 0.4875 0.3950
IS1009d 0.1061 0.8829 0.4768 0.4543
TS3003a 0.0826 0.9301 0.4956 0.5167
TS3003b 0.1176 0.8752 0.4896 0.4645
TS3003c 0.0946 0.8747 0.4791 0.4545
TS3003d 0.0961 0.9177 0.4860 0.4914
Table A9. Speaker change detection results of KL-GMM and BIC-GMM for spectral rolloff.
Table A9. Speaker change detection results of KL-GMM and BIC-GMM for spectral rolloff.
Meeting id FAR (KL) MDR (KL) FAR (BIC) MDR (BIC)
EN2002a 0.7937 0.2202 0.4921 0.5066
EN2002b 0.8008 0.2072 0.4918 0.4221
EN2002c 0.7927 0.1922 0.4860 0.4783
EN2002d 0.7877 0.2172 0.4892 0.4924
ES2004a 0.8160 0.2332 0.4831 0.5112
ES2004b 0.7998 0.1879 0.4852 0.4564
ES2004c 0.7923 0.2414 0.4932 0.4483
ES2004d 0.8038 0.1950 0.4873 0.4793
IS1009a 0.6972 0.3632 0.4826 0.4526
IS1009b 0.6930 0.2886 0.4865 0.4914
IS1009c 0.6884 0.2633 0.4979 0.4930
IS1009d 0.6989 0.2693 0.4887 0.4660
TS3003a 0.7522 0.2340 0.4921 0.5106
TS3003b 0.7914 0.2226 0.4843 0.4491
TS3003c 0.7648 0.1879 0.4893 0.4788
TS3003d 0.7948 0.1859 0.4941 0.5020
Table A10. Speaker change detection results of KL-GMM and BIC-GMM for zero crossing rate.
Table A10. Speaker change detection results of KL-GMM and BIC-GMM for zero crossing rate.
Meeting id FAR (KL) MDR (KL) FAR (BIC) MDR (BIC)
EN2002a 0.5173 0.4748 0.5022 0.5159
EN2002b 0.5369 0.4411 0.4893 0.4544
EN2002c 0.5405 0.4714 0.4981 0.4714
EN2002d 0.5195 0.4965 0.4879 0.4730
ES2004a 0.5508 0.4395 0.4927 0.4843
ES2004b 0.5300 0.4586 0.4989 0.5034
ES2004c 0.4927 0.4978 0.4878 0.4504
ES2004d 0.5465 0.4562 0.4899 0.4893
IS1009a 0.4259 0.5316 0.5016 0.5053
IS1009b 0.4792 0.5143 0.5063 0.5257
IS1009c 0.4986 0.5042 0.5125 0.5490
IS1009d 0.4609 0.4403 0.4940 0.4801
TS3003a 0.4692 0.4772 0.5009 0.4894
TS3003b 0.5068 0.5317 0.4955 0.5048
TS3003c 0.4918 0.5091 0.4937 0.4929
TS3003d 0.5306 0.4794 0.4979 0.4954

References

  1. Anguera, X.; Bozonnet, S.; Evans, N.; Fredouille, C.; Friedland, G.; Vinyals, O. Speaker Diarization: A Review of Recent Research. IEEE Transactions on Audio, Speech, and Language Processing 2012, 20, 356–370. [Google Scholar] [CrossRef]
  2. Tranter, S.; Reynolds, D.A. An overview of automatic speaker diarization systems. IEEE Transactions on Audio, Speech, and Language Processing 2006, 14, 1557–1565. [Google Scholar] [CrossRef]
  3. Tranter, S.E.; Evermann, G.; Kim, D.Y.; Woodland, P.C. An Investigation into the Interactions between Speaker Diarisation Systems and Automatic Speech Transcription B Accuracy of Cts Forced Alignments 44.
  4. Aronowitz, H.; Burshtein, D.; Amir, A. Speaker Indexing in Audio Archives Using Gaussian Mixture Scoring Simulation. Machine Learning for Multimodal Interaction; Bengio, S., Bourlard, H., Eds.; Springer Berlin Heidelberg: Berlin, Heidelberg, 2005; pp. 243–252. [Google Scholar]
  5. Chen, S.; Gopalakrishnan, P.S.; Watson, I.T.J. Speaker, Environment and Channel Change Detection and Clustering via the Bayesian Information Criterion. 1998.
  6. Siegler, M.A.; Jain, U.; Raj, B.; Stern, R.M. Automatic Segmentation, Classification and Clustering of Broadcast News Audio. 1997.
  7. Gish, H.; Siu, M.H.; Rohlicek, R. Segregation of speakers for speech recognition and speaker identification. [Proceedings] ICASSP 91: 1991 International Conference on Acoustics, Speech, and Signal Processing, 1991, pp. 873–876 vol.2. [CrossRef]
  8. Bonastre, J.F.; Delacourt, P.; Fredouille, C.; Merlin, T.; Wellekens, C. A speaker tracking system based on speaker turn detection for NIST evaluation. Proceedings of the Acoustics, Speech, and Signal Processing, 2000. on IEEE International Conference - Volume 02; IEEE Computer Society: USA, 2000; ICASSP’00, p. II1177–II1180. [Google Scholar] [CrossRef]
  9. Malegaonkar, A.; Ariyaeeinia, A.; Sivakumaran, P.; Fortuna, J. Unsupervised speaker change detection using probabilistic pattern matching. IEEE Signal Processing Letters 2006, 13, 509–512. [Google Scholar] [CrossRef]
  10. Ajmera, J.; McCowan, I.; Bourlard, H. Robust speaker change detection. IEEE Signal Processing Letters 2004, 11, 649–651. [Google Scholar] [CrossRef]
  11. Dehak, N.; Kenny, P.J.; Dehak, R.; Dumouchel, P.; Ouellet, P. Front-End Factor Analysis for Speaker Verification. IEEE Transactions on Audio, Speech, and Language Processing 2011, 19, 788–798. [Google Scholar] [CrossRef]
  12. Snyder, D.; Garcia-Romero, D.; Sell, G.; Povey, D.; Khudanpur, S. X-Vectors: Robust DNN Embeddings for Speaker Recognition. 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2018, pp. 5329–5333.
  13. Variani, E.; Lei, X.; McDermott, E.; López-Moreno, I.; Gonzalez-Dominguez, J. Deep neural networks for small footprint text-dependent speaker verification. 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2014, pp. 4052–4056.
  14. Mukhamediev, R.I.; Popova, Y.; Kuchin, Y.; Zaitseva, E.; Kalimoldayev, A.; Symagulov, A.; Levashenko, V.; Abdoldina, F.; Gopejenko, V.; Yakunin, K.; Muhamedijeva, E.; Yelis, M. Review of Artificial Intelligence and Machine Learning Technologies: Classification, Restrictions, Opportunities and Challenges. Mathematics 2022, 10. [Google Scholar] [CrossRef]
  15. Sinha, R.; Tranter, S.E.; Gales, M.J.F.; Woodland, P.C. The Cambridge University March 2005 speaker diarisation system. Proc. Interspeech 2005, 2005, pp. 2437–2440. [Google Scholar] [CrossRef]
  16. Wang, D.; Lu, L.; Zhang, H.J. Speech segmentation without speech recognition. 2003 International Conference on Multimedia and Expo. ICME ’03. Proceedings (Cat. No.03TH8698), 2003, Vol. 1, pp. I–405. [CrossRef]
  17. Lu, L.; Zhang, H.J.; Jiang, H. Content analysis for audio classification and segmentation. IEEE Transactions on Speech and Audio Processing 2002, 10, 504–516. [Google Scholar] [CrossRef]
  18. Desplanques, B.; Demuynck, K.; Martens, J.P. Factor analysis for speaker segmentation and improved speaker diarization. Interspeech, 2015.
  19. Neri, L.V.; Pinheiro, H.N.; Tsang, I.R.; da, C. Cavalcanti, G.D.; Adami, A.G. Speaker segmentation using i-vector in meetings domain. 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2017, pp. 5455–5459. [CrossRef]
  20. Sarkar, A.; Dasgupta, S.; Naskar, S.K.; Bandyopadhyay, S. Says Who? Deep Learning Models for Joint Speech Recognition, Segmentation and Diarization. 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2018, pp. 5229–5233.
  21. Wang, R.; Gu, M.; Li, L.; Xu, M.; Zheng, T.F. Speaker segmentation using deep speaker vectors for fast speaker change scenarios. 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2017, pp. 5420–5424. [CrossRef]
  22. Zhang, A.; Wang, Q.; Zhu, Z.; Paisley, J.W.; Wang, C. Fully Supervised Speaker Diarization. ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2018, pp. 6301–6305.
  23. Jati, A.; Georgiou, P. Speaker2Vec: Unsupervised Learning and Adaptation of a Speaker Manifold Using Deep Neural Networks with an Evaluation on Speaker Segmentation. Proc. Interspeech 2017, 2017, pp. 3567–3571. [Google Scholar] [CrossRef]
  24. Sivakumaran, P.; Fortuna, J.; Ariyaeeinia, A.M. On the use of the Bayesian information criterion in multiple speaker detection. Interspeech, 2001.
  25. Malegaonkar, A.S.; Ariyaeeinia, A.M.; Sivakumaran, P. Efficient Speaker Change Detection Using Adapted Gaussian Mixture Models. IEEE Transactions on Audio, Speech, and Language Processing 2007, 15, 1859–1869. [Google Scholar] [CrossRef]
  26. Meignier, S.; Bonastre, J.F.; Igounet, S. E-HMM approach for learning and adapting sound models for speaker indexing. ISCA, A Speaker Odyssey, The Speaker Recognition Workshop;, 2001; p. 5.
  27. Gupta, V. Speaker change point detection using deep neural nets. 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2015, pp. 4420–4424. [CrossRef]
  28. Hrúz, M. CONVOLUTIONAL NEURAL NETWORK FOR SPEAKER CHANGE DETECTION IN. 2017.
  29. Hrúz, M.; Hlaváč, M. LSTM Neural Network for Speaker Change Detection in Telephone Conversations. Speech and Computer; Karpov, A., Jokisch, O., Potapova, R., Eds.; Springer International Publishing: Cham, 2018; pp. 226–233. [Google Scholar]
  30. Carletta, J.; Ashby, S.; Bourban, S.; Flynn, M.; Guillemot, M.; Hain, T.; Kadlec, J.; Karaiskos, V.; Kraaij, W.; Kronenthal, M.; Lathoud, G.; Lincoln, M.; Lisowska, A.; Post, W.; Reece, S.; Renals, S.; Summers-Stay, D.; Wellner, P. The AMI Meeting Corpus: A Pre-announcement. Machine Learning for Multimodal Interaction. Springer, Berlin, Heidelberg, 2005, pp. 25–29.
Figure 1. Probability distributions of speaker change (SC) and feature change (FC) across corresponding threshold values for various audio features.
Figure 1. Probability distributions of speaker change (SC) and feature change (FC) across corresponding threshold values for various audio features.
Preprints 121190 g001
Figure 2. Threshold optimization with Bayesian approach for KL divergence on validation set.
Figure 2. Threshold optimization with Bayesian approach for KL divergence on validation set.
Preprints 121190 g002
Figure 3. Relation between Max P ( S C | F C ) with results of BIC-GMM (the lower value is good).
Figure 3. Relation between Max P ( S C | F C ) with results of BIC-GMM (the lower value is good).
Preprints 121190 g003
Figure 4. Relation between Max P ( S C | F C ) vs. Approx . F 1 score (the higher value is good).
Figure 4. Relation between Max P ( S C | F C ) vs. Approx . F 1 score (the higher value is good).
Preprints 121190 g004
Table 1. Crossing points and the maximum values of P ( S C | F C ) for various audio features.
Table 1. Crossing points and the maximum values of P ( S C | F C ) for various audio features.
Feature Crossing Points Max P ( S C | F C )
Pitch 0.2869 0.4008
MFCC 0.3153 0.5340
Zero crossing rate 0.2194 0.2931
RMS energy 0.2879 0.3196
Mel spectrogram 0.3614 0.3944
Chroma 0.4171 0.5149
Spectral centroid 0.3593 0.3941
Spectral bandwidth 0.3844 0.4521
Spectral contrast 0.4090 0.4746
Spectral rolloff 0.3332 0.4611
Table 2. Dataset statistics: total change points, overlaps, speakers, and audio length.
Table 2. Dataset statistics: total change points, overlaps, speakers, and audio length.
Meeting ID #Change Points #Overlaps % Overlaps #Speakers Length (min)
EN2002a 1366 678 49.63 4 43.90
EN2002b 954 444 46.55 4 33.52
EN2002c 1468 681 46.39 3 56.46
EN2002d 1331 662 49.74 4 46.02
ES2004a 360 145 40.28 4 15.71
ES2004b 687 257 37.41 4 37.48
ES2004c 722 262 36.29 4 37.71
ES2004d 893 351 39.31 4 34.02
IS1009a 290 122 42.07 4 11.91
IS1009b 602 251 41.70 4 33.24
IS1009c 494 142 28.74 4 26.51
IS1009d 688 245 35.61 4 29.27
TS3003a 424 95 22.41 4 17.54
TS3003b 658 157 23.86 4 30.63
TS3003c 629 160 25.44 4 31.83
TS3003d 1061 345 32.52 4 35.16
Total 13127 5253 40.02 - 548.70
Average 821.31 328.19 37.57 3.94 33.38
Table 4. Results of naive systems for comparison.
Table 4. Results of naive systems for comparison.
FAR MDR
naive 0 0 1
naive 1 1 0
Table 5. Average SCD results for KL-GMM and BIC-GMM across features.
Table 5. Average SCD results for KL-GMM and BIC-GMM across features.
Feature FAR (KL) MDR (KL) FAR (BIC) MDR (BIC)
chroma 0.2889 0.6941 0.4911 0.4716
mel-spectrogram 0.9927 0.0062 0.5034 0.5128
mfcc 0.5428 0.4671 0.4837 0.4499
pitch 0.1470 0.8508 0.4947 0.4894
rms 0.8809 0.1300 0.4817 0.4720
spectral bandwidth 0.5948 0.3869 0.4938 0.4700
spectral centroid 0.7606 0.2299 0.4882 0.4690
spectral contrast 0.0975 0.9009 0.4855 0.4653
spectral rolloff 0.7667 0.2318 0.4890 0.4774
zero crossing rate 0.5061 0.4827 0.4968 0.4928
Average 0.5578 0.4380 0.4908 0.4770
Standard Deviation 0.2919 0.2899 0.0110 0.0292
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated