-
CiteScore
-
Impact Factor
Volume 2, Issue 3, Chinese Journal of Information Fusion
Volume 2, Issue 3, 2025
Submit Manuscript Edit a Special Issue
Chinese Journal of Information Fusion, Volume 2, Issue 3, 2025: 253-274

Open Access | Research Article | 26 September 2025
MgEL: Quantum Entanglement-Inspired Evidence Fusion for Learning with Noisy Labels
1 Key Laboratory of Measurement and Control of CSE, School of Automation, Southeast University, Nanjing 210018, China
2 Faculty of Robot Science and Engineering, Northeastern University, Shenyang 110167, China
3 Southeast University Shenzhen Research Institute, Shenzhen 518063, China
* Corresponding Author: Xinde Li, [email protected]
Received: 21 April 2025, Accepted: 23 July 2025, Published: 26 September 2025  
Abstract
With the rise of data engineering-driven automatic annotation strategies, deep learning has demonstrated remarkable performance and strong competitiveness in intelligent fault diagnosis. However, the inherent limitations of automatic annotators inevitably introduce noisy labels, which in turn hinder the generalization and accuracy of diagnostic models. Although numerous Learning with Noisy Labels (LNL) methods attempt to alleviate the impact of label noise through sample selection or label correction, most rely heavily on model predictions to guide training. This self-reinforcing mechanism frequently leads to confirmation bias, especially under high-noise conditions, thereby limiting their effectiveness. To address these challenges while preserving the full data utility, this paper proposes a novel approach termed the Multi-granularity Evidence Labels (MgEL), inspired by the principles of quantum entanglement and collapse. In MgEL, we perform feature-space fusion between entangled sub-distributions to construct a superposition state, from which two auxiliary labels are derived: a pseudo-label obtained by selecting the class with the maximum amplitude and a collapsed label sampled probabilistically according to the class-wise amplitude distribution. The collapsed label represents an uncertainty-aware observation, while the pseudo-label represents the most confident class estimation. These are then fused with the original annotation to form multi-granularity evidence labels. This approach allows MgEL to suppress confirmation bias and improve robustness under noisy supervision. Extensive experiments validate the effectiveness and reliability of MgEL, particularly in high-noise scenarios (e.g., noise intensity $\eta \ge 80\%$), underscoring its potential for practical deployment in low-cost, data-driven intelligent fault diagnosis systems.

Keywords
learning with noisy labels
multi-granularity information fusion
fault diagnosis
deep learning
classification of time series signals

1. Introduction

With the rapid development of big data technologies, deep learning has made remarkable strides in both theoretical research and practical applications of intelligent fault diagnosis [5, 3]. However, the success of these models hinges on the availability of large-scale, high-quality annotated data, which often entails substantial manual labeling costs [4]. This requirement severely limits the scalability and industrial deployment of deep learning-based diagnostic models, especially in scenarios with constrained annotation resources [1].

To address this bottleneck, cost-effective alternatives, such as crowdsourcing [6] and data-engine-driven automatic labeling [7], have been widely adopted. While these strategies significantly reduce annotation efforts, they inevitably introduce noisy labels—instances [8] where the annotated label deviates from the true class (𝒴=k𝒴^). Such label noise distorts the data distribution, misaligns decision boundaries [9], and ultimately undermines both the accuracy and generalization of diagnostic models [10].

To enhance model robustness under noisy supervision, Learning with Noisy Labels (LNL) [11] has emerged as a prominent research direction. Two mainstream strategies have been extensively explored: sample separation [17] and label correction [12]. Sample separation aims to iteratively identify clean samples via model predictions and use only those for further training [13], while label correction refines noisy labels by integrating observed labels with predictive cues (e.g., latent features [14] or logits [15]). Despite their empirical success, both approaches suffer from a key limitation: they are heavily reliant on model predictions, which are themselves corrupted by noise [16], leading to the accumulation of confirmation bias during self-guided training.

Figs/Confirm.pdf
Figure 1  Confirmation bias in LNL: Model predictions, inherently corrupted by label noise, are repeatedly used to guide training, leading to the accumulation of bias Δ and degraded generalization.

Essentially, these models are guided by their own decisions in noisy environments. Erroneous predictions may be repeatedly reinforced, eventually forming a biased representation of the data distribution (see Figure 1), which motivates the central question of this work:

How can we effectively leverage the full information of noisy datasets while mitigating confirmation bias caused by inaccurate predictions during self-guided training, in order to improve the generalization of diagnostic models under severe noise?

To address this challenge, we draw inspiration from quantum entanglement [18] and collapse, proposing a novel framework called "Multi-granularity Evidence Labels (MgEL)". MgEL constructs a set of multi-granularity labels by integrating three label sources: the original annotation labels 𝒴, dynamically generated pseudo-labels 𝒴, and probabilistic collapsed labels 𝒴¯ derived from the feature space. This process of combining two sub-distributional structures simulates the observation-collapse process of a quantum-entangled system, enabling more robust supervision in the presence of label noise.

Specifically, MgEL conceptualizes a feature cluster 𝒞 containing conflicting labels as a superposition state φ=kck|ϕk. The complex amplitudes ck are transformed into real-valued probabilities by borrowing the Born rule—P(k)=ck2=λk—to enable probabilistic reasoning. Beyond this, class-wise evidential values are introduced to quantify the "belief mass" that a sample belongs to each latent class (or basis state). These evidential values are not a direct conversion from the Born rule but rather are derived by modeling the internal distribution within each feature cluster under uncertainty, providing a structured representation of decision confidence.

Unlike conventional classification schemes that treat each sample as a single independent observation, MgEL constructs a pseudo-multi-source observation framework in the feature space, which better reflects the multi-view nature of real-world noisy data. The fusion of the original label yi, pseudo-label yi, collapsed label y¯i, and evidential i={ek}k=1K simulates repeated observations across different measurement bases. This design suppresses overreliance on single predictions and alleviates the accumulation of confirmation bias, thereby improving robustness during training.

Table 1 Notation summary table.
Symbol Description
xiC×L Input sample with C channels and L time points
yi{1,,K} Annotated label for xi
zid Latent feature of xi extracted by the backbone network
𝒵A,𝒵B Randomly partitioned feature subspaces (entangled sets)
𝒮cos(zi,zj) Exponential cosine similarity between features zi and zj
𝒞i Feature cluster centered at zi
φiK Superposition state encoding class affiliation amplitudes
λik[0,1] Amplitude associated with class k for sample i
|φkK Canonical basis vector for class k
i={ek}k=1K Evidence vector obtained from pseudo-source fusion
y~iK Multi-granularity fused label used for training
(θb,θc;) Diagnostic model with backbone θb and classifier θc
η~[0,1] Estimated noise intensity for dynamic cluster adjustment
A,B Hilbert spaces of entangled subsystems A and B
piK Model prediction vector for input xi
CE(pi,y~i) Cross-entropy loss between prediction pi and label y~i
D={(xi,yi)}i=1N Training dataset containing N noisy samples
yi Pseudo-label derived via superposition collapse
y¯i Collapsed label via multinomial sampling from φi
yi~ Final fused label
yi Clean class-conditional manifold for ground-truth label yi
η Feature extractor perturbed by label noise rate η
ξi Noise-induced deviation of zi from its clean manifold
Tkl Probability of class-k label being flipped to class-l

We conduct extensive experiments on three fault diagnosis datasets to validate the effectiveness of MgEL, which demonstrate that MgEL outperforms other methods, particularly under scenarios with severe label noise (i.e., noise intensity η80%), significantly improving diagnostic accuracy and reducing sensitivity to annotation quality. These findings suggest that MgEL has the potential to reduce annotation costs and enable the reliable deployment of intelligent diagnostic models in practical industrial settings.

The key contributions of this work are summarized as follows:

  1. Theory: We establish a pseudo-multi-source observation modeling method in the feature space, extending the theoretical foundation of decision-level information fusion.

  2. Methodology: We propose a robust multi-granularity label construction strategy by introducing class-wise evidential values under uncertainty, which enhances the model's tolerance to noisy supervision.

  3. Empirical validation: We conduct extensive experiments across three real-world fault diagnosis datasets with varying noise intensities. MgEL consistently outperforms baselines, especially under severe corruption (e.g., η80%), demonstrating its practical feasibility for robust learning under noisy supervision—though not yet deployed in full industrial pipelines.

2. Preliminary

2.1 Nomenclature

To improve clarity and reduce potential ambiguity, Table 1 provides a formal summary of all key notations used throughout the paper.

2.2 Problem formulation (Fault diagnosis with noisy labels)

Fault diagnosis with noisy labels is typically formulated as a multi-class classification task under annotation uncertainty [4]. Let the training dataset be 𝒟={𝒳,𝒴}={(xi,yi)}i=1N, where each xiC×L represents a multivariate time-series signal with C channels and L sampling points, and yi[1,K] is the corresponding fault label drawn from K predefined categories. These samples typically come from industrial systems, including mechanical, electrical, and structural components [21], where condition monitoring sensors collect time-series signals for predictive maintenance and fault detection [19].

In practice, noisy labels are prevalent due to multiple factors [20], including ambiguous or overlapping fault manifestations, inconsistencies among domain experts, and, more importantly, the limited reliability of automated annotation systems [23], which may mislabel large volumes of data due to heuristic rules or insufficient context [22]. Let yi denote the latent true label of xi. If yiyi, the sample is considered mislabeled. However, the noise distribution or transition matrix T(𝒴𝒴) is typically unknown [24]. Depending on whether the noise occurs randomly or in a class-dependent fashion, it is commonly categorized as symmetric (uniform) or asymmetric (class-dependent)—the latter often reflecting realistic diagnostic confusion among fault types with similar signal patterns [13].

The learning objective is to construct a diagnostic model (θb,θc;), comprising a backbone θb and a classifier θc, that maintains robust generalization to clean labels despite being trained on corrupted data, which can be formalized as:

argminθb,θc{𝔼xi𝒳((θb,θc;xi)y^i)}.

2.3 Related works on LNL

Deep neural networks (DNNs) tend to overfit when trained on noisy labels, leading to poor generalization and unreliable predictions [26]. To address this, numerous strategies have been proposed to enhance model robustness under label noise [11], which are broadly categorized into three research areas: robust loss functions [27], sample separation [28], and label correction [29].

Robust loss functions reduce the negative impact of mislabeled samples during training. For example, SCE [30] includes a symmetric regularization term to reduce the dominance of noisy labels, while NLS [31] uses label smoothing to lower label confidence. However, these approaches often rely on strong assumptions about the noise distribution, such as the need for access to or accurate estimation of the label transition matrix T(𝒴𝒴), which is difficult to obtain in practical scenarios [32].

A second class of methods focuses on sample separation, aiming to distinguish clean from noisy samples based on training dynamics. Many early works exploit loss-based heuristics, assuming that samples with lower loss are more likely to be correctly labeled. Representative methods include JoCoR [33], which trains peer networks on mutually selected small-loss samples to refine the dataset. More recent approaches, such as DISC [25], introduce memory-based dynamic thresholding to categorize training data into clean, hard, and correctable subsets. However, these methods depend fundamentally on the model's own predictions to estimate sample reliability. As a result, incorrect assessments during early training stages may be reinforced across epochs—this phenomenon is known as confirmation bias (see Figure 1). Even dynamic sample selection strategies cannot fully overcome this limitation, as the thresholds are still updated based on past model behavior [34].

To mitigate data underutilization caused by sample removal, label correction approaches that generate pseudo-labels from model predictions to supervise training have emerged as a promising alternative [35]. For instance, SED [32] uses a mean-teacher framework to stabilize label updates. Despite their effectiveness, these methods also suffer from confirmation bias: initial incorrect predictions can propagate and become increasingly difficult to reverse during subsequent training [16].

This shared limitation in both sample separation and label correction methods arises from their "self-guided" nature—they rely solely on the model's internal signals, making them vulnerable to early-stage errors that propagate unchecked. To overcome this, inspired by principles of quantum mechanics, MgEL emulates the repeated observation-collapse mechanism of entangled quantum systems, introducing controlled uncertainty throughout the training process. By periodically re-evaluating fused labels with multi-granularity representations, MgEL disrupts confirmation bias and prevents training from being overly influenced by prior incorrect predictions. In this way, MgEL fully leverages the information contained in the original dataset, even when labels are partially corrupted, and improves the robustness of the corrected supervision signals.

2.4 Differentiation from previous works

This work extends our earlier research on learning with noisy labels, specifically MgCF [40] and MgL [41], both of which leverage feature clustering for robust label correction. Although MgCF, MgL, and the proposed MgEL share a unified design philosophy: transforming noisy labels into structured supervision via latent-space modeling, they differ significantly in terms of label granularity representation, fusion strategy, belief assignment, and computational scalability. To clarify these distinctions and avoid any confusion regarding academic overlap, we provide a comparative analysis, also visually summarized in Figure 2.

Figs/Previous.pdf
Figure 2  Comparative illustration of the MgCF [40], MgL [41], and the proposed MgEL, which highlights the core distinctions among the three methods in terms of computational scalability, granularity expressiveness, uncertainty modeling, and the belief assignment strategy adopted during label fusion.

First, regarding label granularity semantics, all three frameworks aim to suppress confirmation bias by constructing multi-granularity supervisory signals. In this context, MgCF introduces a dual-granularity labeling scheme: fine-grained labels are formed when annotation and pseudo-labels agree, indicating high class certainty, whereas coarse-grained labels are used when disagreement occurs, signaling ambiguity. MgL builds on this structure by incorporating an additional label source—the collapsed label sampled from the superposition state φi—which allows the construction of medium-grained labels when only partial agreement exists among the three sources (annotation, pseudo, collapsed). In contrast, MgEL reverts to a two-level granularity structure as in MgCF but complements it with a downstream decision deferral mechanism that handles fully inconsistent labels by reverting to the original superposition state φi. While this does not introduce an explicit third granularity level, it implicitly absorbs semantically hesitant cases through adaptive rejection, thereby preserving the interpretive flexibility offered by multi-granularity labeling.

In terms of label fusion mechanisms, both MgCF and MgL directly use the fused multi-granularity label 𝒴~ as the supervisory signal in training. MgL distinguishes itself by explicitly incorporating collapsed labels 𝒴¯ into the fusion rule alongside annotations and pseudo-labels, thus enabling medium-grained representations. MgEL, in contrast, restricts the fusion to annotation and pseudo-labels but introduces a Coherence Mechanism (Sec. 3.5) that evaluates whether any two of the three available labels (annotation, pseudo, collapsed) are consistent. If no consistency is found, the fused label 𝒴~ is rejected and replaced with φ_i, effectively reverting supervision to a probabilistic representation. This mechanism guards against overconfident but unreliable updates and introduces an adaptive rejection pathway not present in either MgCF or MgL.

The belief assignment strategies adopted by the three methods also exhibit fundamental differences. In both MgCF and MgL, the amplitude probabilities derived from the superposition state φi—which reflects the relative class distribution within each feature cluster—are directly treated as empirical belief masses. These unregularized values are used to assign fusion weights between annotations and pseudo-labels, implicitly treating intra-cluster frequencies as reliable class evidence. In contrast, MgEL introduces a belief regularization mechanism. Rather than using the raw amplitudes in φi, MgEL calibrates the belief assignment by incorporating a global noise estimate η~, constructing an evidential vector i={ek}k=1K that reflects class-wise reliability under dataset-level uncertainty. This transformation from empirical frequencies to noise-aware belief masses enhances the robustness of label fusion, particularly in high-noise scenarios, and aligns with the principles of uncertainty modeling in evidential reasoning frameworks.

Finally, in terms of computational scalability, MgCF and MgL both compute global similarity matrices of size 𝒪(N2) to construct feature clusters across the full dataset, which incurs substantial memory and runtime costs. MgEL, inspired by the partitioned structure of entangled quantum systems, proposes an entangled subspace design: the latent feature space is randomly split into two disjoint subsets, with each sample querying only across the opposite subspace. This reduces similarity computation to 𝒪(N2/4) and introduces randomization effects similar to dropout. This modification not only improves training efficiency but also enhances cluster robustness by avoiding deterministic, potentially biased global modeling.

In summary, MgEL consolidates and extends the prior frameworks by integrating evidential trust modeling, a coherence-driven rejection mechanism, and a more scalable partitioned clustering strategy. These contributions allow it to generalize effectively across high-noise environments while maintaining theoretical and algorithmic distinctions from both MgCF and MgL.

2.5 Entangled Quantum-inspired foundations

In quantum mechanics, the observation-collapse mechanism explains how the measurement of one particle in an entangled system causes the entire wavefunction to collapse instantaneously into a corresponding eigenstate [36]. This collapse not only determines the measured particle's state but also instantaneously defines the state of its entangled counterpart, reflecting a non-local correlation [37]. This principle embodies measurement-induced state transitions and decision outcomes under uncertainty [38], providing rich inspiration for information fusion in noisy, uncertain environments.

Formally, we consider a bipartite quantum system composed of two subsystems A and B, with a joint entangled state expressed as:

|Ψ=k=1Kλk|akA|bkB,

where |Ψ denotes the overall quantum state in the composite Hilbert space AB, with A and B denoting the Hilbert spaces of subsystems A and B, respectively. The vectors |akA and |bkB are orthonormal eigenstates of the respective subsystems, and λk is the complex amplitude associated with each joint basis pair, normalized such that k=1K|λk|2=1. The index k{1,,K} enumerates the possible entangled basis components, forming the mathematical foundation of our analogy.

Upon measuring subsystem A and obtaining the outcome |akA, the entire system collapses into the product state |akA|bkB, with a probability of |λk|2. This structure implies that the state of B is instantaneously determined by observing A, without direct interaction with B. Such non-local inference allows one subsystem to act as an informational proxy for the other—a foundational feature of quantum entanglement.

In MgEL, this collapse mechanism serves as a conceptual inspiration, rather than a physical simulation. We draw an analogy between the measurement-induced resolution of uncertainty and the process of integrating multiple label sources into a consistent supervisory signal. Concretely, three label sources are constructed for each sample: (I) the original annotation, (II) a pseudo-label corresponding to the maximum-amplitude component in the superposition state, and (III) a collapsed label probabilistically sampled based on the amplitude distribution. These label sources represent diverse, imperfect observations of the same latent feature representation.

To further emulate the informational interdependence observed in entangled systems, we partition the latent feature space 𝒵 into two disjoint subspaces, 𝒵A and 𝒵B, and construct feature clusters through cross-subspace querying—samples in 𝒵A are clustered based on their similarity to samples in 𝒵B, and vice versa. This design allows feature clusters derived from one subspace to act as interpretive surrogates for evaluating the class membership of samples in the other. In doing so, each subspace imposes a structural constraint on its counterpart, akin to the inference structure in bipartite entanglement. Additionally, since each sample observes only a subset of the latent space per epoch, this formulation introduces Dropout-like stochasticity, reducing memory complexity and enhancing generalization.

When partial agreement arises among the constructed label sources, it is interpreted as a consistency-triggered decision signal, and the fused label is then adopted for training. Conversely, when all sources disagree, no definitive decision is made; the model retains the superposition-based representation, deferring commitment due to measurement incoherence.We emphasize that all quantum-theoretic terminology is used metaphorically to guide the modeling of uncertainty, structural supervision, and decision consistency. No physical entanglement or non-local interaction is simulated or implied.

2.6 Manifold perturbation under label noise: Modeling and implications

To theoretically examine whether extreme label noise (e.g., η80%) induces structural shifts in the latent space, we first formalize the notion of label-induced manifold perturbation. We posit that such shifts do not alter the marginal distribution P(𝒳) but instead distort the conditional representation P(𝒵|𝒴) learned by the model. The following assumption summarizes our conclusion:

.

(Noise-induced manifold perturbation) Under label noise with a corruption rate of η, the latent representation zi=η(xi) deviates from its ideal, noise-free class manifold yi with an expected squared perturbation:
𝔼[Dist(zi,yi)2]ηTr(ΣT),
where ΣT is the covariance of the perturbation process governed by the label transition matrix T. This distortion causes a representation-level structural shift that scales with noise intensity.

We now justify Assumption 1 through formal modeling and analysis.

Let (xi,yi) denote a clean training sample, and yi the observed (potentially corrupted) label. Due to the incorrect supervision, the model learns a perturbed representation:

zi=η(xi)=zi+ξi,

where ξi denotes the feature-level perturbation induced by the label error. Since this perturbation accumulates over multiple training steps via gradient descent, we express it as:

ξi=t=1TΔzi(t)(yi).

We consider the label corruption process governed by a class transition matrix T={Tkl} with Tkl=P(yi=lyi=k). The induced perturbation distribution is thus:

ξi𝒫η=(1η)δ0+η𝒬(T),

where δ0 denotes the no-perturbation case (correct labels), and 𝒬(T) is a distribution over perturbations generated by incorrect labels sampled according to T.

This mixture formulation in Eq. 5 provides a unified framework for modeling both symmetric and asymmetric label noise. In the symmetric case, where Tkl=1K1 for all kl, the perturbation distribution 𝒬(T) approximates an isotropic Gaussian, i.e., 𝒩(0,σ2I). In contrast, under asymmetric noise where T is sparse and class-dependent, 𝒬(T) becomes a mixture of Gaussians with non-zero means, as defined in Eq. 6.

𝒬(T)=lkTkl𝒩(μkl,Σkl).

We define the manifold perturbation error (MPE) as follows:

MPEi:=minzyiziz2=ξi,

and hence,

𝔼[MPEi2]=η𝔼ξi𝒬(T)[ξi2]=ηTr(ΣT),

where ΣT is the effective covariance structure aggregated over the perturbation distribution. This result confirms that as η increases, the average deviation from the class-specific latent manifold grows linearly, reflecting a progressive distortion in P(𝒵|𝒴).

To quantify the global structural deformation, we define the manifold distortion index (MDI) as:

MDI(η):=1Ni=1NMPEiηTr(ΣT).

The derivation above substantiates the hypothesis in Assumption 1 by showing how high-intensity label noise induces representation-level perturbations that scale with both the noise rate η and the structural properties of the label transition matrix T. In particular, the manifold distortion index MDI(η) provides a global quantitative measure of such perturbations, confirming that noisy supervision leads to a pseudo distribution shift that manifests not in the input space but in the conditional representation space P(𝒵|𝒴). This shift disrupts the geometric coherence of latent class manifolds and ultimately impairs generalization performance.

These findings motivate the core design of MgEL: to suppress the emergence and propagation of noise-induced structural drift by fundamentally rethinking how label supervision is incorporated during training. Rather than relying solely on potentially corrupted labels or unstable model predictions, MgEL introduces controlled structural uncertainty and diversified supervisory signals to counteract the convergence toward biased representations. This is achieved not by architectural overhauls or post hoc corrections, but by embedding uncertainty-aware regularization into the label construction process itself. By doing so, MgEL aims to retain the semantic integrity of the class manifolds even under extreme noise, thereby preserving the model's capacity for generalization.

3. Methodology

3.1 Design rationale of MgEL

Figs/TOC.pdf
Figure 3  Conceptual workflow: The MgEL establishes a quantum-inspired label fusion mechanism by simulating the observation-collapse process inherent to entangled systems. Given a noisy dataset 𝒟={(xi,yi)}i=1N, features 𝒵 are extracted via a backbone encoder (θb;), and the dataset is randomly split to mimic entangled subsystems. For each sample, a feature cluster 𝒞i is constructed from its most similar counterparts in the other subset and abstracted as a superposition state φi. From this, a pseudo-label yi, a collapsed label yi¯, and an evidence i are derived. These are fused with the original annotation yi to form a multi-granularity label y~i. A Coherence mechanism assesses label consistency to decide whether to reintegrate yi~ into φi for mitigating confirmation bias. The final evidence labels 𝒴~ is used to supervise training, improving generalization under severe noise.

DNNs with strong generalization capabilities tend to induce class-discriminative clustering structures in the latent feature space [16], where the similarity between two sample embeddings reflects the likelihood that they share the same true class. Empirical observations confirm [40] that the closer two representations zi,zj𝒵 are, the more likely yi=yj.

.

(Label consistency) For xi,xj𝒳, if their feature similarity 𝒮(zi,zj) approaches the self-similarity limit, i.e., 𝒮(zi,zj)𝒮(zi,zi)=𝒮(zj,zj), the probability of label consistency increases:
P(y^i=y^j)1𝒮(zi,zj)𝒮(zi,zi)=𝒮(zj,zj)

Motivated by Assumption 2, MgEL (see Figure 3) draws inspiration from the observation-collapse process in quantum entangled systems. To simulate such a system, the latent feature space 𝒵 is randomly divided into two disjoint subsets, 𝒵𝒜 and 𝒵. For a sample xi𝒵𝒜, a feature cluster 𝒞i𝒵 is formed by retrieving its top-n most similar neighbors, each associated with the observed annotation yj. These cluster members are treated as independent observations of xi under different measurement conditions, with their label distribution abstracted as a probabilistic superposition over possible class outcomes.

Through these local observations, MgEL simulates multi-view fusion from pseudo sources by leveraging the structured observations of these clusters (see Figure 4). Specifically, for each sample xi, a pseudo-label yi𝒴 is derived from the distribution of labels in its feature cluster. This pseudo-label is then fused with the original annotation yi𝒴 to create an evidence-aware multi-granularity label y~i𝒴~. Based on the principles of Dempster-Shafer theory (DST) [39], we interpret the label distribution in the feature cluster as a class-supporting evidence mass function. The fusion process assigns different weights to the pseudo and observed labels, reflecting their respective levels of credibility under uncertainty.

Figs/Approximating.pdf
Figure 4 Illustration of multi-view fusion from pseudo sources in MgEL: This method simulates multi-view observations by leveraging structured feature clusters, where each cluster 𝒞i is approximated by a set of proxy samples. The belief bj of each pseudo-source zi(j) is assigned based on its similarity 𝒮(zi,zi(j)) to the cluster center zi, simulating multiple independent observations. These pseudo sources are fused by aggregating the evidence from belief-weighted samples, effectively simulating multi-source information fusion.

To further enhance interpretability and control, MgEL categorizes the fused label y~i into two levels of semantic granularity based on the consistency between the pseudo-label yi and the annotation yi. When the two labels are identical, the fused result is interpreted as a fine-grained confident label, reflecting strong agreement across sources and indicating high certainty in the class assignment. In contrast, if the pseudo and original labels disagree, the resulting label is considered a coarse-grained hesitant label, capturing ambiguity in the sample's class attribution and preserving the possibility of it belonging to multiple candidate classes.

Since this correction is based on model-generated latent representations and predictions, it risks amplifying confirmation bias during self-guided training. To mitigate this, MgEL introduces a coherence mechanism that assesses the consistency among multiple label sources (pseudo, original, collapsed). When inconsistency is high, the fused label y~i is reverted to its superposition state φi, preserving uncertainty and deferring final decisions until further optimization.

Overall, MgEL reinterprets noisy or conflicting labels as structured evidence, offering a principled framework to improve robustness by aligning quantum-inspired modeling with uncertainty-aware label correction.

3.2 Constructing feature clusters via entangled partitioning

Consider a noisy dataset 𝒟={(xi,yi)}i=1N with latent representations 𝒵={zi}i=1N extracted using a backbone network (θb;).

Drawing upon quantum collapse dynamics and superposition principles, we establish a pseudo multi-source label fusion framework where each latent representation zi models an analogous quantum state. Central to this approach is the entanglement-emulating partition: 𝒵 undergoes random bisection into disjoint subsets 𝒵A and 𝒵B of equal cardinality. This bipartite configuration achieves two objectives: reducing computational complexity of similarity tensors from 𝒪(N2) to 𝒪(N2/4) and implementing stochastic structural dropout.

Furthermore, each subset acts as an independent measurement apparatus, simulating the observational asymmetry inherent in quantum systems. The bipartite design is a deliberate computational compromise. While multi-partite partitioning (n>2) incurs prohibitive 𝒪((N/n)n) complexity and generates intractable similarity tensors, the current formulation maintains theoretical fidelity while ensuring computational tractability.

For each sample xi, neighborhood retrieval is performed exclusively from the complementary subset. These neighbors {zi(j)}j=1n are not merely correlated instances but simulate independent measurements of the underlying quantum state. Each neighbor thus represents an eigen-label state |k (class k collapse) under different measurement contexts. This quantum interpretation underpins the formulation of the superposed label state:

φi=k=1Kλik|k

where φiK encodes amplitude-based label uncertainty, with coefficients λik[0,1] quantifying class evidence strength (see Sec. 3.3). Neighborhood retrieval uses exponentially transformed cosine similarity to circumvent the curse of dimensionality.

Based on Assumption 2, for xi𝒳, with latent representation zi=(θb;xi), we identify its n most similar feature vectors from the entangled subset to form a feature cluster 𝒞i. These neighbors {zi(j)}j=1n are ranked such that

𝒮(zi,zi(i+1))>𝒮(zi,zi(j)) if j1

To quantify sample similarity, we use an exponential cosine similarity function (see Eq. 8). This decision is motivated by the well-known phenomenon of distance concentration in high-dimensional spaces, which makes traditional distance metrics, such as Euclidean distance, ineffective. Specifically, for two normalized vectors zi,zjn, the squared Euclidean distance is defined as:

d2(zi,zj)=zizj2=zi2+zj22zi,zj

Assuming zi=zj=1, we have:

d2(zi,zj)=22cos(zi,zj),

where zi,zj is the angle between zi and zj. As dimensionality n, most angles zi,zjπ2, and hence cos(zi,zj)0, implying:

d2(zi,zj)2,

which leads to the so-called concentration of measure, where almost all pairwise distances converge to a constant. Formally,

(|d(zi,zj)μd|<ε)1asn,

making it nearly impossible to discriminate between samples based on Euclidean distance.

To avoid this, we use the cosine similarity:

𝒮cos(zi,zj)=cos(zi,zj)=zi×zjzizj,

which measures angular similarity and is less sensitive to magnitude and dimensionality. Still, in high dimensions, even cosine similarities between neighbors may vary only slightly. To accentuate such differences, we introduce the exponential transformation:

𝒮(zi,zj)=ecos(zi,zj)=exp(zi×zjzizj)

This maps cosine similarity from [1,1] to [e1,e1], i.e. 𝒮(zi,zj)[e1,e1][0.37,2.72], nonlinearly amplifying differences between close neighbors and enabling sharper separation of structurally similar samples in latent space.

Based on this similarity, the feature cluster 𝒞i is formally defined as:

𝒞i={{zj𝒮(zi,zj)𝒮(zi,zi(n)),zj𝒵A}if zi𝒵B,{zj𝒮(zi,zj)𝒮(zi,zi(n)),zj𝒵B}otherwise.

The cluster size n is dynamically adapted based on the estimated noise intensity η~ (see Sec. 3.4) from the previous training epoch:

n=max(η~×Initialize(n),Truncation(n))

In this work, Initialize(n)=27 and Truncation(n)=23. This allows the framework to flexibly adjust the neighborhood scope: under high noise, more neighbors are included to stabilize superposition formation; under low noise, smaller clusters avoid introducing irrelevant variance.

3.3 Simulating pseudo-source observations via cluster fusion

Cluster fusion simulates the integration of multiple pseudo-observations by leveraging the feature and label distributions within the feature cluster 𝒞i (as constructed in Eq. 9). This process is analogous to quantum state tomography [42] in quantum mechanics, where the quantum state is reconstructed by sampling probability distributions across different measurement bases, determining the amplitude probabilities of the eigenstates in a superposition.

In MgEL, the superposition state φi constructed for each sample xi can be interpreted as the probability distribution of xi's class membership across all potential categories. The associated belief mass i={ek}k=1K encodes the basic belief assignment for each class, which is used to fuse the pseudo-label and the original annotation during label correction, considering the current dataset's noise level.

To improve the robustness and accuracy of the superposition state φi, MgEL adopts the evidence combination concept from DST. Each sample in the feature cluster 𝒞i is treated as an independent source of evidence, and a cluster-based fusion process based on similarity is designed:

{φi=k=1Kλik|ϕk,λik=zj𝒞i𝒮(zi,zj)𝟙(yj=k)zj𝒞i𝒮(zi,zj),|ϕk=OneHot(k),k[1,K].

where K represents the total number of classes, and 𝟙(yj=k) is an indicator function that equals 1 when yj=k, and 0 otherwise.

Using this similarity-based fusion method, MgEL combines the feature distribution and label frequency information within the feature cluster 𝒞i to form the superposition state φi. This enables MgEL to robustly model the class membership probability distribution for sample xi, even in noisy and uncertain environments.

However, considering the potential misleading effects of noisy labels, there is a risk that the feature mapping during training may inaccurately represent the data, leading to imprecise class probabilities in φi. If these probabilities were directly used as fusion belief masses for label correction, they could amplify errors and introduce confirmation bias.

To address this, MgEL does not use φi's class probabilities directly as belief masses in the fusion of annotations and pseudo-labels. Instead, after considering the estimated noise intensity η~, MgEL adjusts the evidence i for each class membership distribution:

ek=zj𝒞i𝒮(zi,zj)𝟙(yj=k)zj𝒞i𝒮(zi,zj)+K×η~.

3.4 Multi-granularity labels construction with evidence awareness

After obtaining the superposition state φi and class evidence i through feature cluster fusion, MgEL generates pseudo-labels yi using the maximum posterior decision rule (commonly applied in multi-class classification tasks, e.g. [1]). The amplitude probability αyi associated with the pseudo-label represents the frequency or intra-cluster label consistency of the dominant class within the feature cluster. This is used to estimate the noise intensity of the current dataset by calculating the mean label consistency α𝒴 across all feature clusters:

{yi=argmax(φi),αyi=|yi×φi=max(φi),η~=i=1NαyiN.

At this stage, both the pseudo-label and the original annotation label serve as independent evidence sources for making decisions about sample xi. MgEL then fuses these two sources to correct the noisy labels, based on their respective confidence levels. Specifically, if the pseudo-label and original annotation labels agree, it is highly likely that the original annotation is not noisy and can be used for training. In contrast, when the two labels disagree, the pseudo-label is more likely to be accurate. The confidence in the decision-making process is captured by the evidence vector i, which is derived from the pseudo-source fusion.

This allows the original annotation label to be corrected toward the pseudo-label, with a weighted confidence:

yi~=(|yi×i)yi+(1|yi×i)yi

The fused multi-granularity labels 𝒴~ integrate both fine-grained and coarse-grained information, reflecting the model's certainty in the predicted class. When the pseudo-label and the original annotation agree, the label corresponds to a fine-grained confident label, offering a more precise supervision signal. When the labels disagree, the output represents a coarse-grained hesitant label, with the decision uncertain between two potential classes, and the confidence distributed between them according to the evidence mass i. This distinction allows the model to dynamically adjust its decision-making process based on the consistency of the label sources.

3.5 Coherence mechanism triggered by consistency of labels

To mitigate the confirmation bias that may arise from directly using pseudo-labels to correct original annotations, MgEL employs a resampling strategy on the superposition state φi, simulating a random collapse process. This generates a collapsed label yi¯, where the sampling probability for each class corresponds to the amplitude probability of the corresponding eigenstate in φi.

Subsequently, MgEL introduces a Coherence mechanism triggered by label consistency. Before performing supervised training, the fused labels 𝒴~ go through an additional processing step. In this step, the system accepts a fused label only when at least two of the following labels are consistent: the original annotation yi, the pseudo-label yi, and the collapsed label yi¯. If none of these labels are consistent, the system retains the superposition state φi, deferring the final label commitment until more evidence is provided in subsequent training iterations.

The update rule for the fused label is formally defined as follows:

yi~={φi, if |yi+|yi+|yi¯=3,yi~, otherwise.

In this rule, represents the norm of the vector formed by the sum of the three labels, and the condition ensures that the fused label yi~ is accepted when at least two of the three labels are consistent.

Through this process, MgEL combines the statistical consistency of the pseudo-labels, the randomness introduced by the collapsed labels, and the original annotation labels to create more robust labels 𝒴~. This approach reduces the confirmation bias typically associated with prediction-based label correction and improves the model's generalization ability, especially in high-noise environments.

3.6  Overview

The Alg. 1 summarizes the key steps of MgEL, integrating feature clustering, pseudo-label generation, and evidence-based label fusion for label correction, simulating quantum-inspired mechanisms and adapting them for noisy label correction.

Algorithm 1  MgEL Label Correction Process
  •  Noisy dataset 𝒟={𝒳,𝒴}={(xi,yi)}i=1N, model (θb,θc;), estimated noise intensity η~

  • (θb,θc;)# Trained diagnostic model

  •  Initialize model f(θb,θc;), η~=1.0

  • for epoch in range(Total Epochs) do

  •   𝒵=(θb;𝒳)# Extract latent features

  •   𝒵𝒜,𝒵=Split(𝒵)# Partition dataset into two subsets

  •   𝒞𝒜={𝒞i}i=1N2η~,𝒵, 𝒞={𝒞i}i=1N2η~,𝒵𝒜# Construct feature clusters via Eq. 9

  •   φ=𝒜{φi}i=1N2𝒞𝒜, φ={φi}i=1N2𝒞# Compute superposition state via Eq. 11

  •   {i}i=1Nφ𝒜φ# Get class evidence via Eq. 12

  •   η~,𝒴φ𝒜φ# Estimated noise intensity and compute pseudo-labels via Eq. 13

  •   𝒴~{i}i=1N,𝒴,𝒴# Construct fused labels via Eq. 14

  •   𝒴¯φ𝒜φ# Generate collapsed labels via multinomial sampling

  •   𝒴¯φ𝒜φ,𝒴¯,𝒴,𝒴# Apply Coherence Mechanism for final multi-granularity labels via Eq. 15

  •   for each sample (xi,y~i) in 𝒟~ do

  •    pi=f(θb,θc;xi)# Compute prediction

  •    (θb,θc)Update(θb,θc,(pi,y~i))# Update model parameters via fused label y~i

  •   end for

  • end for

Considering that MgEL uses exponential cosine similarity (Eq. 8) to measure sample similarity in the latent feature space, we have aligned the classifier architecture accordingly. Instead of adding a fully connected layer to the backbone network for classification, we employ a cosine classifier to directly classify the features, which ensures better alignment with the feature space's geometry and optimization direction, as cosine similarity is more suitable for high-dimensional feature manifolds.

It is important to emphasize that the proposed MgEL framework operates exclusively during the training phase. It acts as a robust label correction strategy by adjusting supervisory signals based on latent-space evidence fusion. Consequently, the inference pipeline remains entirely unaltered—both structurally and computationally. MgEL introduces no additional latency, memory, or computational overhead at deployment time. The final diagnostic model inherits its real-time performance characteristics solely from the underlying backbone architecture, making MgEL fully compatible with industrial runtime constraints.

4.  Experimental verification and discussion

4.1 Experimental settings

To thoroughly evaluate the performance of MgEL, we conducted a series of experiments on three benchmark datasets for bearing fault diagnosis: Single [44], Multiple [43], and Damage [40]. These datasets consist of vibration signals collected directly from mechanical systems during operation, covering various working conditions, rotational speeds, and load conditions. These datasets were selected to represent a wide range of fault scenarios and complexities, ensuring that MgEL can handle diverse real-world conditions effectively.

Table 2 The detailed information about all datasets.
Parameter Detailed information about three datasets
Single Multiple Damage
No. of categories 6+1=7 (6+1)2=49 4×6+1=25
Fault location Inner, Outer, Ball, Inner and Outer, Inner and Ball, Outer and Ball
Damage degree 0.3 mm 0.3 mm 0.2, 0.4, 0.6, and 0.8 mm
[1770,1775] rpm [1770,1775] rpm
Motor speed [1443,1478] rpm [2366,2370] rpm [2366,2370] rpm
[2959,2962] rpm [2959,2962] rpm
Motor load 0.0, 0.1, 0.2 and 0.3 NM 0.0 and 0.3 NM 0.0 and 0.3 NM
Sampling frequency 12 kHz 12 kHz 16 kHz
No. of tra. samples 4032 46452 34100
No. of val. samples 924 10290 7600
No. of tes. samples 4032 46452 34100

Table 2 summarizes the details of each dataset, including fault categories, operating conditions, and the number of samples. Notably, the datasets were chosen without addressing class imbalance, ensuring that the number of training and testing samples for each class is roughly equal. Each sample consists of 2048 continuous data points, with no overlap between any two samples, ensuring a clean and consistent evaluation framework.

With the goal of conducting a comprehensive evaluation of MgEL under various noise environments, two types of noise were introduced: class-dependent asymmetric noise and class-independent symmetric noise. Asymmetric noise was introduced by swapping labels between similar classes at five different intensity levels (η=25%,30%,35%,40%,45%). For example, in the Damage dataset, severity labels within the same fault category were swapped, such as replacing the "0.2 mm" label with "0.4 mm" or swapping "0.8 mm" and "0.6 mm". This type of noise simulates realistic scenarios in which label errors occur within similar categories. Symmetric noise was introduced by randomly replacing the original labels with labels from other categories, with five different noise intensities applied (η=50%,60%,70%,80%,90%).

To ensure a fair comparison and reproducibility, we adopt MgNet [43], an open-source fault diagnosis architecture released alongside the Multiple dataset and well-suited for time-series analysis in industrial scenarios, as the backbone for evaluating the performance of MgEL. All models were trained using the AdamW optimizer, with momentum coefficients β1=0.9, β2=0.999, and weight decay 102. The learning rate follows a CosineAnnealing schedule, with the initial value scaled as 3×103×batch size512, ensuring consistency across different training scales. Training is performed for 100 epochs, including a 5-epoch warm-up stage, during which only standard supervised learning is applied, without any additional label correction operations.

All experiments were conducted on an NVIDIA A100 GPU using PyTorch version 2.1.1+cu118. To eliminate the influence of random initialization and stochastic variation, each experiment was repeated at least ten times, with the detailed evaluation results and corresponding analyses presented in the following subsections.

4.2 Case I: Effectiveness of MgEL

Table 3 Comparison of effectiveness between MgEL and other methods for LNL.
Datasets Methods Accuracy(%) under asymmetric noise Accuracy(%) under symmetric noise Mean
η=25% η=30% η=35% η=40% η=45% η=50% η=60% η=70% η=80% η=90% (%)
Single Baseline 87.93±5.51 83.16±4.79 75.88±1.97 69.42±3.47 59.26±0.10 68.12±4.96 58.04±6.44 45.60±5.87 29.04±9.35 5.37±3.79 58.18
SL (2019) 95.80±0.42 86.35±3.15 85.61±7.28 72.60±3.34 61.06±1.38 81.66±6.66 73.14±8.38 58.99±1.20 25.31±5.43 3.38±1.35 64.39
JoCoR (2020) 88.51±0.17 89.71±7.14 81.80±8.32 73.87±4.63 60.44±0.81 79.37±2.00 63.38±5.80 53.36±4.41 34.43±1.30 7.42±1.62 63.23
JNPL (2021) 64.66±9.98 54.91±3.24 54.22±4.37 36.95±7.29 43.46±4.35 28.80±7.35 29.21±5.07 27.98±1.28 21.73±9.13 8.86±4.39 37.08
NCR (2022) 97.54±1.54 90.09±3.63 86.09±6.91 81.73±2.12 75.11±6.49 84.25±1.31 76.59±5.17 68.97±7.78 49.63±2.68 5.03±0.38 71.50
ALFs (2023) 89.16±7.78 88.08±7.80 83.60±2.93 69.47±1.07 59.24±5.62 90.68±0.47 86.87±2.08 73.94±2.46 53.65±2.96 5.11±2.26 69.98
LSL (2024) 96.24±3.03 95.28±4.32 94.73±3.82 82.56±6.03 70.14±9.81 95.33±3.98 93.34±0.57 82.30±9.58 64.83±3.16 6.73±3.39 78.15
ANNE (2025) 91.10±7.84 89.32±8.27 83.79±1.22 75.15±5.49 69.50±7.78 98.74±1.21 97.43±1.57 95.47±0.57 73.17±8.33 4.44±1.13 77.81
MgEF (Ours) 99.49±0.38 99.19±0.35 99.09±0.63 96.06±2.58 80.89±4.69 99.55±0.17 98.92±0.77 98.25±0.41 80.46±3.19 23.02±2.53 87.49
Multiple Baseline 75.07±3.78 69.21±3.16 64.11±3.13 55.04±0.27 51.49±0.92 53.00±2.29 48.36±1.99 36.36±3.22 23.03±0.44 7.39±0.69 48.31
SL (2019) 80.40±1.85 75.52±0.43 71.80±2.82 53.57±1.73 49.67±0.84 72.56±1.22 62.51±2.54 58.29±2.29 34.38±2.62 9.34±0.53 56.80
JoCoR (2020) 88.47±1.14 88.70±0.68 87.31±1.53 72.50±4.18 64.88±1.23 74.13±3.67 66.38±0.98 49.36±2.57 31.64±1.03 9.97±1.43 63.33
JNPL (2021) 63.59±0.24 63.45±0.19 61.66±1.21 60.12±1.02 55.93±1.67 50.45±0.33 41.54±2.07 33.78±1.27 23.24±1.28 9.57±1.37 46.33
NCR (2022) 90.25±0.57 89.34±0.26 86.66±1.72 82.49±0.85 69.89±1.48 66.50±3.83 55.60±2.50 43.01±1.53 26.52±3.07 9.07±1.69 61.93
ALFs (2023) 77.04±1.50 73.96±1.75 68.02±0.74 64.00±3.36 54.59±1.65 65.27±2.76 52.35±0.61 32.91±0.69 22.92±0.73 8.40±1.39 51.95
LSL (2024) 64.06±1.57 62.43±4.91 66.19±1.33 59.88±1.83 55.32±0.07 47.67±0.83 42.58±2.20 34.91±1.08 24.05±1.66 8.72±0.99 46.58
ANNE (2025) 65.55±2.49 67.69±2.57 63.46±1.44 63.60±1.34 57.80±1.04 52.88±2.58 51.15±4.12 39.90±4.36 23.58±1.72 11.10±0.51 49.67
MgEF (Ours) 89.33±0.52 89.16±0.87 87.70±1.15 86.20±0.51 75.65±1.50 86.20±1.28 83.49±1.16 81.45±0.52 75.46±1.80 43.18±5.46 79.78
Damage Baseline 91.18±0.17 86.31±1.18 82.38±1.10 67.28±3.45 54.57±2.16 88.78±0.48 84.66±1.02 76.72±0.30 55.24±1.62 20.18±3.84 70.73
SL (2019) 93.83±0.42 90.53±2.13 88.04±1.00 75.70±4.60 57.45±0.45 94.55±0.17 92.51±0.19 87.88±1.48 76.11±1.50 20.86±1.74 77.74
JoCoR (2020) 93.84±0.19 89.22±1.72 85.28±0.81 70.00±3.85 56.66±1.52 91.78±0.64 88.94±1.25 83.64±1.32 62.57±1.67 22.41±2.13 74.43
JNPL (2021) 90.18±0.05 86.10±0.51 81.11±0.99 69.62±0.76 57.32±1.66 87.56±0.93 84.95±0.34 76.17±2.60 52.13±4.56 24.30±2.23 70.94
NCR (2022) 94.72±0.64 92.00±0.30 86.24±1.44 75.71±2.07 61.99±1.99 91.75±0.77 89.44±2.55 83.49±0.74 61.45±0.75 18.20±1.09 75.50
ALFs (2023) 96.93±0.62 94.96±1.41 90.11±2.67 78.45±2.18 59.09±0.83 96.03±0.62 94.01±1.71 87.32±2.06 62.74±0.85 19.77±1.66 77.94
LSL (2024) 90.09±0.84 86.56±0.44 81.67±0.99 70.83±0.72 57.55±1.07 86.76±1.06 83.61±1.56 75.46±1.08 54.41±4.17 20.28±1.20 70.72
ANNE (2025) 96.63±0.18 96.50±0.23 92.75±0.31 87.08±0.45 64.41±6.91 97.36±0.65 96.99±0.26 90.41±0.41 83.42±3.21 33.66±8.18 83.92
MgEF (Ours) 97.48±0.42 97.19±0.45 96.70±0.87 93.25±0.65 75.03±1.01 97.31±0.22 96.47±0.34 95.81±0.27 94.31±0.35 75.07±2.35 91.86

To evaluate the effectiveness of the proposed MgEL framework, we conducted comprehensive comparisons with seven state-of-the-art LNL approaches: SL (2019) [30], JoCoR (2020) [33], JNPL (2021) [45], NCR (2022) [46], ALFs (2023) [27], LSL (2024) [16], and ANNE (2025) [47], across three datasets with varying types and intensities of label noise. We visualize the performance comparison in Figure 5 and provide the detailed results in Table 3, where bold values represent the best accuracy achieved by other methods under each noise setting, facilitating an intuitive comparison with MgEL.

Figs/Effectiveness.pdf
Figure 5  Comparison of MgEL and other methods under various label noise environments. Where the shaded area indicates the accuracy fluctuation range of each method, and the dashed baseline represents the performance of MgNet trained with standard supervised learning, without any LNL strategy.

Overall, MgEL consistently outperforms all competing approaches across different datasets and noise scenarios, demonstrating superior robustness and generalization even in the presence of severe label corruption. On the Single dataset, MgEL achieves a mean accuracy of 87.49%, surpassing the strongest baseline (LSL, 78.15%) by 9.34%. This advantage is amplified further under high symmetric noise (η=90%), where MgEL maintains a robust 23.02% accuracy compared to 7.42% for JoCoR. A similar pattern is observed on the Multiple dataset, where MgEL yields an average accuracy of 79.78%, outperforming the second-best baseline (JoCoR) by 16.45%. On the more challenging Damage dataset, MgEL delivers the highest performance (91.86%), surpassing ANNE (83.92%) by 7.94%.

These consistent improvements across datasets and noise conditions are attributed to several technical innovations in MgEL. One key reason is that, instead of relying on heuristic sample selection or loss-based reweighting, MgEL constructs a quantum-inspired superposition state φi for each sample by aggregating local observations from entangled feature clusters. This formulation captures class attribution uncertainty with greater fidelity. Additionally, MgEL fuses the pseudo-label and original annotation via an evidence-aware belief mass i, derived from the distribution of cluster-level statistics. This enables adaptive weighting of conflicting information sources based on their consistency. Moreover, the Coherence mechanism safeguards against early confirmation bias by retaining the superposition state whenever the pseudo, annotation, and collapsed labels fail to reach consensus, postponing hard commitments until sufficient evidence accumulates in later training iterations.

The performance advantage of MgEL becomes especially prominent under high-noise symmetric settings, indicating its tolerance against purely stochastic label perturbations that often mislead traditional LNL strategies. Moreover, its stability across diverse datasets highlights its generalizability and scalability for real-world fault diagnosis tasks.

4.3 Case II: Ablation for MgEL

To further investigate the source of MgEL's effectiveness, we conduct an ablation study by selectively removing each of its three core components: Entanglement, Evidence, and Coherence. The results are summarized in Table 4 and visualized in Figure 6.

Table 4 The results of the ablation on MgEL.
Methods Entanglement Evidence Coherence Average accuracy(%)
Single Multiple Damage
Baseline 58.18 48.31 70.43
A1 85.64 76.89 90.22
A2 84.61 75.58 89.12
A3 82.09 72.95 87.74
A4 79.41 70.16 86.09
A5 81.62 72.33 87.67
A6 83.88 74.34 88.78
MgEL 87.49 79.78 91.86

Taken together, Table 4 and Figure 6 show that the full MgEL framework, which integrates all three components, achieves the highest accuracy across all datasets. Removing any component consistently leads to a decline in performance, indicating that the effectiveness of MgEL stems from the complementary roles of entanglement-based pseudo-source modeling, evidence-aware label fusion, and coherence-guided filtering.

Figs/Ablation.pdf
Figure 6  Performance comparison of MgEL and its ablated variants on three datasets, highlighting the individual contribution of Entanglement, Evidence, and Coherence modules, where dashed lines represent the accuracy of baseline.

The effect of removing Entanglement is evident when comparing configurations A1 and A4 to the full model. A1, which disables Entanglement while retaining Evidence and Coherence, shows a performance decline (e.g., 76.89% vs. 79.78% on Multiple). The degradation becomes more pronounced in A4, where only Entanglement is active, leading to further reductions in accuracy (e.g., 70.16%). These trends suggest that Entanglement is critical for constructing reliable pseudo-source representations. Its role in partitioning the feature space introduces stochastic diversity in cluster formation, capturing the distributional uncertainty of class attribution through superposition states, which is essential for robust label inference.

Figs/tSNE.pdf
Figure 7  t-SNE visualization of latent feature distributions for test and training samples across the three datasets. The first column shows test samples from the baseline model trained without noise handling. The second and third columns show test and training samples learned under the MgEL framework. Compared to the baseline, MgEL induces more compact and well-separated class clusters, even under high noise.

The role of Evidence is assessed by examining A2 and A5, both of which remove this component. A2 maintains Entanglement and Coherence but excludes Evidence, resulting in performance drops across all datasets (e.g., 75.58% vs. 79.78% on Multiple). A5, retaining only Entanglement, performs even worse. These observations indicate that while structural diversity is necessary, accurate fusion of annotations and pseudo-labels requires confidence calibration. The belief mass i, computed from local similarity-weighted label distributions and adjusted for noise intensity, serves this purpose. Its removal forces reliance on raw class probabilities, weakening the model's ability to distinguish reliable label sources. Thus, Evidence enhances label consistency by integrating supervision signals based on their reliability within the cluster context.

The effect of Coherence is reflected in A3 and A6. A3 excludes Coherence while preserving Entanglement and Evidence, leading to a modest performance reduction (e.g., 72.95% vs. 79.78% on Multiple), while A6—disabling both Coherence and Entanglement—shows the lowest performance among all variants. These results confirm that Coherence plays a significant role in label stability, especially under ambiguous or conflicting supervision. The mechanism selectively defers label commitments by reverting to the superposition state when the pseudo, annotation, and collapsed labels do not agree. Therefore, Coherence mitigates confirmation bias and stabilizes predictions by withholding unreliable updates in the presence of low label consensus.

Collectively, the ablation results demonstrate that MgEL's tolerance to noisy labels comes from the interplay of three complementary mechanisms: entanglement introduces structured diversity for uncertainty modeling, evidence enables reliability-aware label fusion via belief-based weighting, and coherence mitigates confirmation bias by deferring low-consensus decisions. Their unified integration equips MgEL with robust generalization capabilities across diverse and corrupted learning environments.

4.4 Case III: Visualization analysis

To empirically support Assumption 1, we visualize the t-SNE embeddings of both training and test samples from the three datasets under 90% symmetric noise, as shown in Figure 7. The resulting projections reveal significant differences in the topological organization of feature manifolds, providing intuitive evidence of noise-induced structural perturbations and highlighting MgEL's ability to counteract these distortions through more coherent class representations.

Compared to the baseline, the latent features learned by MgEL show significantly improved intra-class compactness and inter-class separability. On the Single dataset, clear and well-separated clusters emerge even for complex fault types, while the baseline features remain entangled with indistinct boundaries. The Multiple dataset, which introduces greater domain and fault variability, further demonstrates MgEL's robustness: despite the increased difficulty, MgEL preserves coherent topological structures for most classes, while the baseline embeddings collapse into ambiguous, overlapping manifolds.

Notably, the effect is most pronounced on the Damage dataset, which consists of 49 fine-grained fault categories and extreme label corruption (η=90%). In this challenging setting, the baseline model fails to preserve class-discriminative geometry, leading to severe semantic drift. In contrast, MgEL successfully induces structured manifolds with meaningful class-wise alignment, consistent with the suppression of structural perturbations predicted by the theoretical model.

Furthermore, training embeddings offer additional insight into MgEL's manifold-regularizing behavior. Although residual intra-cluster variance persists due to corrupted supervision, the global alignment between training and test distributions is significantly improved under MgEL. This suggests that the framework filters noise at the label level and stabilizes representation learning by regularizing optimization trajectories.

The observed improvements are attributed to MgEL's evidence-aware label construction process, which incorporates uncertainty through probabilistic superposition and belief-based fusion mechanisms. These mechanisms effectively mitigate the propagation of biased gradients and restore the semantic coherence of the latent space.

Taken together, the t-SNE results corroborate the theoretical analysis in Assumption 1, showing that extreme label noise induces a representation-level structural shift. MgEL mitigates this phenomenon by preserving class-consistent manifolds in the latent space, improving both robustness and generalization.

5. Conclusion

Motivated by the challenge of confirmation bias in LNL, this work proposes MgEL, a quantum-inspired framework that introduces a novel multi-granularity evidence labeling mechanism by simulating the observation–collapse behavior of entangled systems. Through pseudo-source construction, evidence-aware fusion, and coherence-guided filtering, MgEL effectively integrates annotations, pseudo-labels, and collapsed labels to create robust supervision signals. Extensive experiments across three fault diagnosis benchmarks confirm that MgEL outperforms existing methods in most scenarios, particularly under severe symmetric noise, while consistently improving the quality of latent representations. MgEL provides a theoretically grounded and practically scalable solution for trustworthy fault diagnosis in data-driven industrial systems. Beyond its empirical success, MgEL introduces a principled information fusion strategy that integrates multiple uncertain label sources, offering new insights into uncertainty modeling and decision-level fusion in noisy environments.

Despite its promising performance, MgEL has several limitations. The stochastic nature of random partitioning may cause instability in small or highly imbalanced datasets, and reliance on similarity-based neighborhood construction can be sensitive to feature distortions during early training. Future research could explore more robust entanglement schemes, incorporate adaptive clustering, or integrate causal and temporal priors into the label-fusion process. Moreover, extending the proposed framework to semi-supervised, active, or federated learning settings holds promise for advancing both the theory and practice of information fusion under uncertainty.


Data Availability Statement
Data will be made available on request.

Funding
This work was supported in part by the National Natural Science Foundation of China under Grant 62233003 and Grant 62073072; in part by the Key Projects of Key R&D Program of Jiangsu Province under Grant BE2020006 and Grant BE2020006-1; in part by the Shenzhen Science and Technology Program under Grant JCYJ20210324132202005 and Grant JCYJ20220818101206014.

Conflicts of Interest
The authors declare no conflicts of interest.

Ethical Approval and Consent to Participate
Not applicable.

References
  1. Dunkin, F., Li, X., Hu, C., Wu, G., Li, H., Lu, X., & Zhang, Z. (2024). Like draws to like: A Multi-granularity Ball-Intra Fusion approach for fault diagnosis models to resists misleading by noisy labels. Advanced Engineering Informatics, 60, 102425.
    [CrossRef]   [Google Scholar]
  2. Li, X., Dunkin, F., & Dezert, J. (2024). Multi-source information fusion: Progress and future. Chinese Journal of Aeronautics, 37(7), 24–58.
    [CrossRef]   [Google Scholar]
  3. Zheng, X., Nie, J., He, Z., & Gao, M. (2025). Specific Task-Guided Collaborative Domain Generalization Network for Intelligent Fault Diagnosis under Unseen Conditions. IEEE Internet of Things Journal.
    [CrossRef]   [Google Scholar]
  4. Dunkin, F., Li, X., Li, H., Wu, G., Hu, C., & Ge, S. S. (2025). MgCNL: A Sample Separation Approach via Multi-Granularity Balls for Fault Diagnosis With the Interference of Noisy Labels. IEEE Transactions on Automation Science and Engineering, 22, 7748–7761.
    [CrossRef]   [Google Scholar]
  5. Hu, C., Zhang, Z., Li, C., Leng, M., Wang, Z., Wan, X., & Chen, C. (2025). A state of the art in digital twin for intelligent fault diagnosis. Advanced Engineering Informatics, 63, 102963.
    [CrossRef]   [Google Scholar]
  6. Zhang, W., Jiang, L., & Li, C. (2025). ELDP: Enhanced Label Distribution Propagation for Crowdsourcing. IEEE Transactions on Pattern Analysis and Machine Intelligence, 47(3), 1850–1862.
    [CrossRef]   [Google Scholar]
  7. Yang, H., Wang, L., Pan, Y., & Chen, J.-J. (2025). A Teacher-Student Framework Leveraging Large Vision Model for Data Pre-Annotation and YOLO for Tunnel Lining Multiple Defects Instance Segmentation. Journal of Industrial Information Integration, 100790.
    [CrossRef]   [Google Scholar]
  8. Zhang, Y., Chen, Y., Fang, C., Wang, Q., Wu, J., & Xin, J. (2025). Learning from open-set noisy labels based on multi-prototype modeling. Pattern Recognition, 157, 110902.
    [CrossRef]   [Google Scholar]
  9. Bian, Z., Chang, Q., Wang, J., Pedrycz, W., & Pal, N. R. (2024). Takagi–sugeno–kang fuzzy systems for high-dimensional multilabel classification. IEEE Transactions on Fuzzy Systems, 32(6), 3790-3804.
    [CrossRef]   [Google Scholar]
  10. Sun, Y., Song, H., Guo, L., Gao, H., & Cao, A. (2025). A transfer learning method: Universal domain adaptation with noisy samples for bearing fault diagnosis. Advanced Engineering Informatics, 65, 103243.
    [CrossRef]   [Google Scholar]
  11. Song, H., Kim, M., Park, D., Shin, Y., & Lee, J.-G. (2023). Learning From Noisy Labels With Deep Neural Networks: A Survey. IEEE Transactions on Neural Networks and Learning Systems, 34(11), 8135–8153.
    [CrossRef]   [Google Scholar]
  12. Liu, Y., Zhong, Y., Ma, A., Zhao, J., & Zhang, L. (2023). Cross-resolution national-scale land-cover mapping based on noisy label learning: A case study of China. International Journal of Applied Earth Observation and Geoinformation, 118, 103265.
    [CrossRef]   [Google Scholar]
  13. Zhang, J., Song, B., Wang, H., Han, B., Liu, T., Liu, L., & Sugiyama, M. (2024). BadLabel: A Robust Perspective on Evaluating and Enhancing Label-Noise Learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 46(6), 4398–4409.
    [CrossRef]   [Google Scholar]
  14. Chen, M., Zhao, Y., He, B., Han, Z., Huang, J., Wu, B., & Yao, J. (2024). Learning with noisy labels over imbalanced subpopulations. IEEE Transactions on Neural Networks and Learning Systems, 36(4), 6544-6555.
    [CrossRef]   [Google Scholar]
  15. Lu, Y., & He, W. (2024). Mitigating Noisy Supervision Using Synthetic Samples with Soft Labels. arXiv preprint arXiv:2406.16966.
    [Google Scholar]
  16. Kim, N.-R., Lee, J.-S., & Lee, J.-H. (2024). Learning with Structural Labels for Learning with Noisy Labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 27610–27620).
    [CrossRef]   [Google Scholar]
  17. Kim, S., Lee, D., Kang, S., Chae, S., Jang, S., & Yu, H. (2024, June). Learning Discriminative Dynamics with Label Corruption for Noisy Label Detection. In 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 22477-22487). IEEE.
    [CrossRef]   [Google Scholar]
  18. ATLAS Collaboration. (2024). Observation of quantum entanglement with top quarks at the ATLAS detector. Nature, 633(8030), 542.
    [CrossRef]   [Google Scholar]
  19. Schwarz, A., Rahal, J. R., Sahelices, B., Barroso-García, V., Weis, R., & Duque Anton, S. (2024). Data augmentation in predictive maintenance applicable to hydrogen combustion engines: a review. Artificial Intelligence Review, 58(1), 32.
    [CrossRef]   [Google Scholar]
  20. Mo, Z., Zhang, Z., Miao, Q., & Tsui, K.-L. (2025). Extended Invariant Risk Minimization for Machine Fault Diagnosis With Label Noise and Data Shift. IEEE Transactions on Neural Networks and Learning Systems, 36(8), 15476-15489.
    [CrossRef]   [Google Scholar]
  21. Yu, W., Dillon, T., Mostafa, F., Rahayu, W., & Liu, Y. (2020). A Global Manufacturing Big Data Ecosystem for Fault Detection in Predictive Maintenance. IEEE Transactions on Industrial Informatics, 16(1), 183–192.
    [CrossRef]   [Google Scholar]
  22. Wang, X., Wang, S., Liang, Y., & Lei, Z. (2025). Decisive vector guided column annotation. Pattern Recognition, 158, 110958.
    [CrossRef]   [Google Scholar]
  23. Nguyen, T., Ibrahim, S., & Fu, X. (2024). Noisy Label Learning with Instance-Dependent Outliers: Identifiability via Crowd Wisdom. Advances in Neural Information Processing Systems, 37, 97261–97298.
    [Google Scholar]
  24. Lin, Y., Yao, Y., & Liu, T. (2024). Learning the latent causal structure for modeling label noise. Advances in Neural Information Processing Systems, 37, 120549–120577.
    [Google Scholar]
  25. Li, Y., Han, H., Shan, S., & Chen, X. (2023). DISC: Learning From Noisy Labels via Dynamic Instance-Specific Selection and Correction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 24070–24079).
    [CrossRef]   [Google Scholar]
  26. Li, F., Li, K., Tian, J., & Zhou, J. (2024). Regroup Median Loss for Combating Label Noise. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 38, No. 12, pp. 13474–13482).
    [CrossRef]   [Google Scholar]
  27. Zhou, X., Liu, X., Zhai, D., Jiang, J., & Ji, X. (2023). Asymmetric Loss Functions for Noise-Tolerant Learning: Theory and Applications. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(7), 8094–8109.
    [CrossRef]   [Google Scholar]
  28. Fang, C., Cheng, L., Mao, Y., Zhang, D., Fang, Y., Li, G., Qi, H., & Jiao, L. (2024). Separating Noisy Samples From Tail Classes for Long-Tailed Image Classification With Label Noise. IEEE Transactions on Neural Networks and Learning Systems, 35(11), 16036–16048.
    [CrossRef]   [Google Scholar]
  29. Xu, G., Yi, L., Xu, P., Li, J., Pu, R., Shui, C., Ling, C., McLeod, A. I., & Wang, B. (2025). Unraveling the Mysteries of Label Noise in Source-Free Domain Adaptation: Theory and Practice. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1–17.
    [CrossRef]   [Google Scholar]
  30. Wang, Y., Ma, X., Chen, Z., Luo, Y., Yi, J., & Bailey, J. (2019). Symmetric cross entropy for robust learning with noisy labels. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 322–330).
    [CrossRef]   [Google Scholar]
  31. Wei, J., Liu, H., Liu, T., Niu, G., Sugiyama, M., & Liu, Y. (2021). To smooth or not? when label smoothing meets noisy labels. arXiv preprint arXiv:2106.04149.
    [Google Scholar]
  32. Sheng, M., Sun, Z., Chen, T., Pang, S., Wang, Y., & Yao, Y. (2025). Foster Adaptivity and Balance in Learning with Noisy Labels. In European Conference on Computer Vision (pp. 217–235).
    [CrossRef]   [Google Scholar]
  33. Wei, H., Feng, L., Chen, X., & An, B. (2020). Combating noisy labels by agreement: A joint training method with co-regularization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 13726–13735).
    [CrossRef]   [Google Scholar]
  34. Kim, D., Ryoo, K., Cho, H., & Kim, S. (2025). SplitNet: learnable clean-noisy label splitting for learning with noisy labels. International Journal of Computer Vision, 133(2), 549–566.
    [CrossRef]   [Google Scholar]
  35. Zhang, R., Cao, Z., Huang, Y., Yang, S., Xu, L., & Xu, M. (2025). Visible-Infrared Person Re-identification with Real-world Label Noise. IEEE Transactions on Circuits and Systems for Video Technology, 1–1.
    [CrossRef]   [Google Scholar]
  36. Bender, C. M., & Hook, D. W. (2024). $\mathcal{PT$-symmetric quantum mechanics. Reviews of Modern Physics, 96(4), 045002.
    [CrossRef]   [Google Scholar]
  37. Bose, S., Fuentes, I., Geraci, A. A., Khan, S. M., Qvarfort, S., Rademacher, M., Rashid, M., Toroš, M., Ulbricht, H., & Wanjura, C. C. (2025). Massive quantum systems as interfaces of quantum mechanics and gravity. Reviews of Modern Physics, 97(1), 015003.
    [CrossRef]   [Google Scholar]
  38. Liang, X.-B., Li, B., & Fei, S.-M. (2024). Signifying quantum uncertainty relations by optimal observable sets and the tightest uncertainty constants. Science China Physics, Mechanics & Astronomy, 67(9), 290311.
    [CrossRef]   [Google Scholar]
  39. Zhang, X., Wang, C., Zhou, W., Xu, J., & Han, T. (2024). Trustworthy diagnostics with out-of-distribution detection: A novel max-consistency and min-similarity guided deep ensembles for uncertainty estimation. IEEE Internet of Things Journal, 11(13), 23055-23067.
    [CrossRef]   [Google Scholar]
  40. Ruan, H., Wang, Y., Qin, Y., & Tang, B. (2021, October). An enhanced intelligent fault diagnosis method to combat label noise. In 2021 International Conference on Sensing, Measurement & Data Analytics in the era of Artificial Intelligence (ICSMD) (pp. 1-6). IEEE.
    [CrossRef]   [Google Scholar]
  41. Li, M., He, S., Chen, J., Feng, Y., & Xie, J. (2025). Label-smoothing dynamic decoupling augmented network for intelligent fault diagnosis under imbalanced data distribution with noisy labels. Measurement, 118664.
    [CrossRef]   [Google Scholar]
  42. Hu, C. K., Wei, C., Liu, C., Che, L., Zhou, Y., Xie, G., ... & Yu, D. (2024). Experimental sample-efficient quantum state tomography via parallel measurements. Physical Review Letters, 133(16), 160801.
    [CrossRef]   [Google Scholar]
  43. Deng, J., Liu, H., Fang, H., Shao, S., Wang, D., Hou, Y., Chen, D., & Tang, M. (2023). MgNet: A fault diagnosis approach for multi-bearing system based on auxiliary bearing and multi-granularity information fusion. Mechanical Systems and Signal Processing, 193, 110253.
    [CrossRef]   [Google Scholar]
  44. Fang, H., Deng, J., Chen, D., Jiang, W., Shao, S., Tang, M., & Liu, J. (2023). You can get smaller: A lightweight self-activation convolution unit modified by transformer for fault diagnosis. Advanced Engineering Informatics, 55, 101890.
    [CrossRef]   [Google Scholar]
  45. Kim, Y., Yun, J., Shon, H., & Kim, J. (2021). Joint Negative and Positive Learning for Noisy Labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 9442–9451).
    [CrossRef]   [Google Scholar]
  46. Iscen, A., Valmadre, J., Arnab, A., & Schmid, C. (2022). Learning With Neighbor Consistency for Noisy Labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 4672–4681).
    [CrossRef]   [Google Scholar]
  47. Cordeiro, F. R., & Carneiro, G. (2025). ANNE: Adaptive Nearest Neighbours and Eigenvector-based sample selection for robust learning with noisy labels. Pattern Recognition, 159, 111132.
    [CrossRef]   [Google Scholar]

Cite This Article
APA Style
Dunkin, F., & Li, X. (2025). MgEL: Quantum Entanglement-Inspired Evidence Fusion for Learning with Noisy Labels. Chinese Journal of Information Fusion, 2(3), 253–274. https://doi.org/10.62762/CJIF.2025.151851

Article Metrics
Citations:

Crossref

0

Scopus

0

Web of Science

0
Article Access Statistics:
Views: 82
PDF Downloads: 13

Publisher's Note
ICCK stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions
CC BY Copyright © 2025 by the Author(s). Published by Institute of Central Computation and Knowledge. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/), which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made.
Chinese Journal of Information Fusion

Chinese Journal of Information Fusion

ISSN: 2998-3371 (Online) | ISSN: 2998-3363 (Print)

Email: [email protected]

Portico

Portico

All published articles are preserved here permanently:
https://www.portico.org/publishers/icck/