Preprint
Article

This version is not peer-reviewed.

Critical Observability Enforcement in Discrete-event Systems Using Differential Privacy

A peer-reviewed article of this preprint also exists.

Submitted:

11 November 2024

Posted:

12 November 2024

You are already at the latest version

Abstract
In the context of discrete-event systems (DESs), critical states usually refer to a system configuration of interest, describing certain important system properties, e.g., fault diagnosability, state/language opacity, and state/event concealment. Technically, a DES is critically observable if an intruder can always unambiguously infer, by observing the system output, whether the plant is currently in a predefined set of critical states or the current state set is disjoint with the critical states. In this paper, given a partially observable DES modeled with a finite-state automaton that is not critically observable, we focus on how to make it critically observable, which is achieved by proposing a novel enforcement mechanism based on differential privacy (DP). Specifically, we consider two observations, where one cannot determine whether a system is currently in the predefined critical states (i.e., the observation violating the critical observability), while the other is randomly generated by the system. When these two observations are processed separately by the differential privacy mechanism (DPM), the system generates an output, exposed to the intruder, that is randomly modified such that its probability approximates the two observations. In other words, the intruder cannot determine the original input of a system by observing its output. In this way, even if the utilized DPM is published to the intruder, he/she is unable to identify whether critical observability is violated.
Keywords: 
;  ;  

1. Introduction

Over the past decade, state estimation in discrete-event systems (DESs) (without a full sensor deployment) has received unprecedented interest thanks to the extensive deployment of cyber-physical systems that are mathematically characterized by discrete-event systems. Within this problem, several core challenges, such as opacity analysis [1,2,3], diagnosability assessment [4,5], detectability verification [6,7], and controller design optimization [8], are deeply dependent on the ability to accurately estimate the system states by its partially observed nature. Particularly, critical observability, as an important concern in the direction of state estimation, is the focus of our exploration in this research. The concept of critical observability, formulated in [9], is widely applied to networked systems, medical & health systems, semiconductor manufacturing, etc. A system is considered to be critically observable if, through the observation of a system output, it is possible to infer without ambiguity whether or not the current system states fully fall into a predefined set of critical states that can characterize several system properties of interest. Petri nets and automata are the main tools for the modeling and control of DESs with potential mutual transformation [10].
There are two commonly employed techniques for checking the critical observability of a DES that is modeled with finite state automata: regular language [11] and observer-based approaches [12,13,14]. In [11], the authors utilize the operations of regular languages to develop a computational procedure to address the observability verification problem, executable in polynomial time. In [12], k-step critical observability is defined and an observer called a k-extended detector is presented to verify k-step critical observability. The critical observability of large-scale finite-state automaton networks is considered in [13] and decentralized critical observers are designed. Lai et al. propose a novel method for constructing observers for polynomially-ambiguous max-plus automata to check their critical observability, which extends earlier work on unambiguous max-plus automata [14].
When a system is verified to be non-critically observable, a significant problem is how to make it critically observable. This research reports a mechanism that, borrowing from DP [15], aims to consolidate/enforce the critical observability for a non-critically observable system. Further, the mechanism ensures that the critical states of the system cannot be detected by an intruder, i.e., achieving critical state protection/concealment.
Differential privacy (DP) is a data processing technique within the field of cryptography, first proposed by Dwork [15]. Building on a solid mathematical foundation, DP guarantees that the statistical characteristics exhibited by the overall dataset will almost remain unchanged when a single entity (a user, a record, or any form of individual data point) is added to or removed from the dataset. In essence, the inclusion/exclusion of a single-record barely alters the computational outcomes of the dataset. Consequently, DP can withstand a wide range of privacy attacks, making it extremely difficult for attackers to accurately infer individual data by observing the calculation results.
The privacy protection mechanism due to DP is primarily achieved by introducing noise to obfuscate the original data during data processing, thereby concealing personal characteristics and sensitive information. The Laplace mechanism [16], which is particularly suited for protecting numerical data, and the exponential mechanism [17], designed for handling non-numerical data, are two of the most common and effective methods for injecting noise in order to implement DP.
As shown in the literature, DP has been widely applied in various fields, including big data analytics, federated learning, recommendation systems, and trajectory data publishing. In [18], DP is combined with the Haar Wavelet technique to protect the privacy of big data in body sensor networks. The authors developed a tree-based structure to analyze the data, reducing errors and enabling long-range queries. In [19], the authors concentrated on three aspects of DP protection for adolescents’ ultrasound examination health datasets: DP protection for basic statistical output perturbations, publication of DP marginal histograms and synthesized data, and machine learning DP learning algorithms. The problem of location and traffic information leakage in vehicular network environments is addressed in [20] by integrating local DP and federated learning. Specifically, the authors introduce four local DPMs, including the three-outputs and optimal piecewise mechanism, to perturb gradients from vehicles, ensuring privacy while maintaining model accuracy and minimizing communication overhead. In [21], Ren et al. employ the data from various social network platforms to generate personalized recommendations, where, to ensure the privacy of the users, DP is applied to the user attributes and the connections present in the social network graph. Presented in [22] and [23] is a privacy protection mechanism for symbolic systems whose traces (symbol sequences) can be protected. Based on this framework, the authors further develop an innovative exponential mechanism. In addition, DP finds its extensive applications in other domains such as logistics [24] and time series trajectories [25]. We believe that more research will be exposed as this kind of privacy protection paradigm prevails corresponding to the increasingly demanding requirements of system security.
Although DP is effective in protecting sensitive information within various information-intensive systems, its potential applications in DESs have yet to be fully explored. This paper primarily focuses on the problem how to utilize DP to enforce critical observability in DESs modeled with finite-state automata. Furthermore, we explore how DP is employed to guarantee that the critical states are not discovered by an intruder. The research follows the schematic illustrated in Figure 1. Specifically, given a system A n d modeled with a nondeterministic finite-state automaton, a set of observable events Σ o , a set of critical states X s , and a set of non-critical states X n s , we assume that the system is not critically observable with respect to (w.r.t.) the predefined set of observable events Σ o and critical states X s . Using the standard determinization procedure presented in [26], we construct a critical state estimator O b s ( A n d ) w.r.t. observable events. The system’s behavior can be described as a scenario that the event sequence s generated by A n d is transmitted to the critical state estimator O b s ( A n d ) , which then produces an observation ρ .
By estimating the state set that is consistent with an observation ρ , denoted by C ( ρ ) , we draw a crucial conclusion. If this set is a subset of the non-critical states, i.e., C ( ρ ) X n s , the observation ρ is regarded as a safe observation. If the set is a subset of the critical states, i.e., C ( ρ ) X s , the observation ρ is considered an unsafe observation. In both cases, the critical states can be revealed/detected by an intruder. Otherwise, the set C ( ρ ) is neither a subset of the critical states nor a subset of the non-critical states, i.e, C ( ρ ) X s C ( ρ ) X n s . In this situation, the observation ρ is also deemed unsafe since the critical observability of the system is violated. In other words, when ρ is acquired by an intruder, he/she can determine that the system is not critically observable. To prevent unsafe observations from being captured, a DP module is introduced as a protection shield. The function of this module is that, when an unsafe observation ρ is received, the module outputs a protected observation ρ with a probability similar to that of another random observation, thereby obscuring the intruder such that it is difficult to trace back to the original unsafe observation. If the observation ρ is safe, the system directly outputs the undecorated ρ .
We further analyze how DP works. For an unsafe observation ρ , it is first mapped into a state sequence ω , which is a state string interleaved with ρ in O b s ( A n d ) , where O b s ( A n d ) is the observer of the plant A n d . Then, ω is sent to the DP module. We assume that the operational logic (running principals) of the DP module is public to an intruder, i.e., the intruder knows the probability distribution of the module generating all possible outputs ρ w.r.t. an input observation ρ . However, when the two state sequences respectively corresponding to an unsafe observation ρ and a random observation τ are input to the module, the protected output observation ρ is highly similar in probability. This design guarantees that even if an intruder captures ρ , it is difficult to distinguish whether it originates from ρ or τ , thereby effectively protecting unsafe observations that violate critical observability (i.e., C ( ρ ) X s C ( ρ ) X n s ) or reveal critical states (i.e., C ( ρ ) X s ). Under the aforementioned lines, the main contributions of this work are summarized as follows.
(1) A novel framework based on DP for the enforcement of critical observability for DESs is proposed. Given a non-critically observable system, we define an unsafe observation by computing the set of states consistent with it, i.e., an observational system output. In this paper, the designed method aims to confuse the unsafe observations with the randomly generated observations from a probabilistic viewpoint, such that a non-critically observable system satisfies critical observability. Moreover, this method guarantees that the critical states of the system cannot be detected by an intruder.
(2) To expand the application field of DP theory, we propose a new concept, called “state sequence DP” and apply it to finite state automata.
(3) Thanks to the concept of state sequence DP, we design an optimal DPM that ensures the optimal protection of unsafe observations.
The preliminaries of DESs and DP are presented in Section 2. The definition of critical observability is recalled in Section 3. In Section 4, a novel DPM specialized for the enforcement of critical observability of a system is designed. We discuss a realistic case of enforcing critical observability in Section 5 to show the practical applicability of the proposed approach. Finally, concluding remarks are given in Section 6.

2. Preliminaries

The basics of notations and notions necessary for the development of this research are presented, including languages and finite-state machines (deterministic finite automata–DFA and non-deterministic finite automata–NFA), as they are appropriate for the modeling of a DES. A reader is referred to [26]. Note that an NFA is equivalent to a DFA from the lens of languages that they can generate, i.e, they have the same modeling power.

2.1. Languages and Automata

Let E be a finite set of event labels a DES and E * be a set of the finite strings defined over E (the Kleene closure of set E), consisting of the empty string ε . Given a string σ E * , its length is the number of symbols contained in it, denoted by | σ | . In particular, we have | ε | = 0 . Given two strings σ 1 and σ 2 , the string σ = σ 1 σ 2 is said to be the concatenation of σ 1 and σ 2 .
A DFA is a quadruple A d = ( X d , E , δ d , x 0 ) , where X d is the finite set of states, E is, as previously stated, the collection of event labels, δ d : X d × E X d is the partially defined deterministic transition function, and x 0 X d is the initial state. In order to facilitate the development of this research, one readily extends the transition function, as again denoted by δ d in the case of no confusion, recursively defined from X d × E X d to X d × E * X d , where δ d ( x , s e ) = δ d ( δ d ( x , s ) , e ) for s E * and e E (specifically, we have δ d ( x , ε ) = x for all x X d ). Let k N with N = { 0 , 1 , 2 , 3 , } be a non-negative integer. We use E k to denote the set of event sequences with length k, i.e., E k = { s E * | s | = k } . Analogously, we write X d k to represent the set of all state sequences with length k. The language generated by A d is defined as L ( A d ) = { s E * δ d ( x 0 , s ) is defined } . The set of events that are enabled (feasible) at a given state x X d is defined by E ( x ) = { e E δ d ( x , e ) is defined } . The number of enabled events at state x is thus | E ( x ) | .
Given a DFA A d = ( X d , E , δ d , x 0 ) , a sequence ξ (interleaving of states and events) starting from state x 0 and ending at a state in X d , i.e., ξ = x 0 e 1 x 1 e 2 e k x k is said to be a production of A d with e i E ( x i 1 ) ( i = 1 , 2 , , k ). We write Ξ to represent the set of all productions ξ in the DFA A d . Given a production ξ = x 0 e 1 x 1 e 2 e k x k of A d , the state sequence x 0 x 1 x k associated with ξ is called a production state sequence, denoted by F ( ξ ) = x 0 x 1 x k . Analogously, the event sequence e 1 e 2 e k associated with ξ is said to be a production event sequence, is denoted by T ( ξ ) = e 1 e 2 e k . Let γ = x 0 x 1 x k with k 1 be a non-empty state sequence. We write T ( γ ) = { s E * { ε } ξ Ξ : F ( ξ ) = γ & s = T ( ξ ) } to represent the set of event sequences consistent with γ . Given a non-empty event sequence s = e 1 e k with k 1 , its consistent state sequence set is defined as F ( s ) = { γ X d * { ε } ξ Ξ : T ( ξ ) = s & γ = F ( ξ ) } . Note that X d * is the set of all finite strings of elements in X d .
Example 1.
Consider a DFA A d = ( X d , E , δ d , x 0 ) shown in Figure 2, where X d = { x 0 , x 1 , x 2 , x 3 , x 4 } and E = { α , β , λ } . We have δ d ( x 0 , β ) = x 1 , δ d ( x 0 , λ ) = x 2 , δ d ( x 0 , α ) = x 3 , E ( x 0 ) = β , λ , α , and E ( x 0 ) = 3 . Given a production ξ = x 0 α x 3 β x 4 , the production state sequence and the production event sequence associated with ξ are F ( ξ ) = x 0 x 3 x 4 and T ( ξ ) = α β , respectively. Let us consider an event sequence λ α in A d . The set of state sequences consistent with λ α is F ( λ α ) = { x 0 x 2 x 2 } . □
An NFA is a quadruple A n d = ( X n d , E , δ n d , X 0 ) , where X n d is the finite set of states, E is the same as before explained, δ n d : X n d × E 2 X n d is the (partial) transition function, and X 0 X n d is the set of initial states. The transition function δ n d is similarly extended to δ n d : X n d × E * 2 X n d by defining δ n d ( x , ε ) = { x } for all x X n d and δ n d ( x , s e ) = x δ n d ( x , s ) δ n d ( x , e ) , where s E * and e E . The language generated by A n d from state x X n d is formally defined as L ( A n d , x ) = { s E * δ n d ( x , s ) } . Given a set of states X X n d , we define L ( A n d , X ) = x X L ( A n d , x ) . The language of A n d is L ( A n d ) = L ( A n d , X 0 ) .
Given an automaton, E is partitioned into two disjoint sets: E = E o ˙ E u o , where E o ( E u o ) is the set of observable (unobservable) events. The natural projection P : E * E o * is defined by
P ( ε ) = ε and P ( s e ) = P ( s ) e , if e E o ; P ( s ) , if e E u o ,
where s E * and e E . Function P can be extended to P : 2 E * 2 E o * , i.e., given Z E * , P ( Z ) = { z E o * z Z : z = P ( z ) } .
Given an NFA A n d = ( X n d , E o ˙ E u o , δ n d , X 0 ) and a state set X X n d , the set of states unobservably reached from X is defined as U R ( X ) = { x X n d x X , s E u o * : x δ n d ( x , s ) } . We use the notation N e ( X , e ) to represent the set of states reached by the occurrence of an observable event e E o from the states in X, i.e., N e ( X , e ) = { x X n d x X : x δ n d ( x , e ) } . The observer of A n d , denoted as O b s ( A n d ) = ( X o b s , E o , δ o b s , x 0 , o b s ) , can be computed following a standard procedure outlined in [26]. More specifically, X o b s 2 X n d is the state set, E o is the set of observable events, δ o b s : X o b s × E o X o b s is the (partial) transition function (for all x o b s X o b s and for all e E o , δ o b s ( x o b s , e ) = U R ( N e ( x o b s , e ) ) ), and x 0 , o b s = U R ( X 0 ) is the initial state.
Due to the nature of partial observation of a considered plant, we can only observe the observable event labels, while unobservable events can be directly observed. However, different event sequences may have the same observational results. Given an NFA A n d , a string ρ P ( L ( A n d ) ) is said to be an observation. The state estimation w.r.t. ρ can be expressed as C ( ρ ) = { x X n d x 0 X 0 , s L ( A n d ) : P ( s ) = ρ & x δ n d ( x 0 , s ) } . By definition, it comes C ( ρ ) = δ o b s ( x 0 , o b s , ρ ) . Note that an observer is a DFA where all events are observable.

2.2. Differential Privacy

Differential privacy is an effective vehicle for privacy protection by providing a mechanism (or a function) to protect sensitive personal information from being disclosed during statistical analysis or data dissemination. Broadly speaking, whenever we use a dataset as the input to this mechanism, the addition or removal of any single record from the dataset will not affect the data query results. This means that even if the dataset is slightly modified, it is difficult or impossible for external intruders to obtain any specific privacy information [15].
Definition 1
(Differential privacy  [27]). Let ϵ be a positive real number and M be a randomized mechanism (function) that takes a dataset as input. Let I m ( M ) denote the image of M . The mechanism M is said to provide ϵ -differential privacy if for any two datasets D 1 and D 2 that differ on a single element, and for any O I m ( M ) , it holds
P M ( D 1 ) O exp ( ϵ ) P [ M ( D 2 ) O ] ,
where the value of M at a dataset D 1 or D 2 is contained in the sample space, i.e., I m ( M ) , with a probability decided by the randomness used in the mechanism, and P M ( D 1 ) O is the probability of M ( D 1 ) O representing the possibility that the output of M at D 1 belongs to O. ♢
In Definition 1, the performance of DP is evaluated by the value of ϵ . Specifically, a smaller ϵ indicates a more subtle distinction between the probabilities of events M ( D 1 ) O and M ( D 2 ) O , signifying a lower likelihood for an intruder to distinguish between the two datasets. Conversely, a larger value of ϵ implies a diminished level of protection for users’ private information.
Remark 1.
According to Definition 1, two datasets, namely D 1 and D 2 , are required to differ in only one element. However, in this study, we extend the concept of DP to the field of DESs modeled with a DFA A d = ( X d , E , δ d , x 0 ) (the formal definition of DP within the DFAs framework, called state sequence DP, will be presented in Section 4). In the definition of DP tailored for DESs, we define M ω : X d k X d k as a mechanism specifically designed for a set of state sequences. The input and output of M ω are two state sequences of equal length. The degree of similarity between these sequences is assessed based on their Hamming distance, as defined in Section 4.

3. Critical Observability for Automata

In this section, we recall the formalization of critical observability verification problems. Given an NFA A n d = ( X n d , E , δ n d , X 0 ) that is partially observable, i.e., E = E o ˙ E u o , in practice, certain important physical states of the NFA are called critical states, represented by a subset of states X s X n d . By acquiring a random sequence of observable events, if an intruder can always unambiguously infer whether the system is currently in a predefined set of critical states, A n d is said to be critically observable.
Definition 2
(Critical observability [13]). Given an NFA A n d = ( X n d , E , δ n d , X 0 ) , a set of observable events E o E , and a set of critical states X s X n d , the system A n d is said to be critically observable w.r.t. E o and X s , if the following predicate holds
ρ P ( L ( A n d ) ) : C ( ρ ) X s C ( ρ ) X n d X s .
A critically observable system requires that, for any observation ρ P ( L ( A n d ) ) , the state set consistent with the observation should be either a subset of critical states or a subset of non-critical states. To verify whether an NFA A n d = ( X n d , E , δ n d , X 0 ) is critically observable, one can construct a critical state estimator of A n d , i.e., the observer O b s ( A n d ) = ( X o b s , E o , δ o b s , x 0 , o b s ) of A n d , which stores all possible state estimations [26]. Then, a system A n d is critically observable w.r.t. X s and E o , if for any state x X o b s , either x X s or x X n d X s holds.
Let A n d = ( X n d , E , δ n d , X 0 ) be an NFA that is not critically observable, E o E be a set of observable events, X s X n d be a set of critical states, and O b s ( A n d ) = ( X o b s , E o , δ o b s , x 0 , o b s ) be the critical state estimator of A n d . The set of critical observability violative states is denoted as X N = { x X o b s x X s x X n d X s } .
Example 2.
Consider the NFA A n d = { X n d , E , δ n d , X 0 } shown in Figure 3, where X n d = { x 0 , x 1 , x 2 , x 3 , x 4 , x 5 } is the set of states, E = E o ˙ E u o = { α , β , θ , λ } is the set of events with E o = { α , β , λ } and E u o = { θ } , and X 0 = { x 0 } is the set of initial states. The critical state estimator O b s ( A n d ) = ( X o b s , E o , δ o b s , x 0 , o b s ) of A n d is portrayed in Figure 4. We assume that the set of critical states is X s = { x 3 , x 4 } . In Figure 4, by considering an observation ρ = β α λ , the system A n d is not critically observable w.r.t. E o and X s , thanks to C ( β α λ ) = δ o b s ( x 0 , o b s , β α λ ) = x o 4 = { x 3 , x 4 , x 5 } , i.e, there exist a state x o 4 X o b s such that x o 4 X s and x o 4 X n d X s .
However, the NAF A n d is critically observable if the critical state is X s = { x 2 } , since intruders can always determine whether the critical state is reached by observing the system’s behavior. That is, in O b s ( A n d ) , for all x X o b s , x X s or x X n d X s holds. □
Consider a system A n d = ( X n d , E , δ n d , X 0 ) that is not critically observable w.r.t. the set of observable events E o E and the set of critical states X s X n d . Let O b s ( A n d ) be the critical state estimator of A n d . Given a production ξ = x 0 e 1 x 1 e 2 e k x k of O b s ( A n d ) with the observation ρ = e 1 e 2 e k , if the set of states consistent with ρ is a subset of critical observability violative states or a subset of critical states, i.e., C ( ρ ) X N or C ( ρ ) X s , the production is said to be unsafe. In this case, the corresponding sequences F ( ξ ) = x 0 x 1 x k and T ( ξ ) = ρ = e 1 e 2 e k are defined as an unsafe production state sequence and an unsafe production event sequence (or an unsafe observation), respectively. Otherwise, the set of states consistent with ρ is included in X n d X s , i.e., C ( ρ ) X n d X s , the production is identified as safe. Its corresponding production event sequence T ( ξ ) and production state sequence F ( ξ ) are also considered safe. Once an unsafe observation is exposed to an intruder, he/she can determine that the system is not critically observable or the system’s critical states are revealed. In the following, we focus on how to make a non-critically observable system satisfy critical observability. Moreover, we guarantee that the critical states of the system cannot be revealed by an intruder.

4. Critical State Concealment Based on Differential Privacy

In this section, we aim to design a DPM to enforce critical observability for DESs modeled with DFAs, and to ensure that the predefined critical states cannot be detected to intruders. Specifically, we first define the Hamming distance of two state sequences in a DFA to describe the difference between them. Then, to prevent critical states or critical observability violative states from being captured by an intruder, we define a novel DPM for DFAs, called state sequence DP. Finally, an algorithm that provides a probability distribution for the generation of each state in an output state sequence is proposed and the DPM is designed.

4.1. Problem Statement

In O b s ( A n d ) , when an unsafe observation ρ is generated, we can design a DPM associated with ρ to effectively conceal the unsafe observation. In particular, given an observation ρ , we derive its corresponding production state sequence ω 1 F ( ρ ) . By inputting ω 1 and a random production state sequence ω 2 , such that the Hamming distance between ω 1 and ω 2 is restricted to a bound l N , into the designed mechanism, a protected production state sequence ω 3 is output with approximate probability, where the lengths of ω 1 , ω 2 , and ω 3 are equal (i.e., ω 1 = ω 2 = ω 3 ). Since both ω 1 and ω 2 can lead to the generation of ω 3 with probabilities that are sufficiently close, it becomes impossible for an intruder to determine whether ω 1 or ω 2 is generated by the estimator O b s ( A n d ) . In other words, from the observed behavior contained in T ( ω 3 ) , the intruder cannot infer its corresponding input behavior, and thereby the unsafe observation is not identifiable. Before discussing the use of DP to protect unsafe observations from being acquired, we first provide some definitions.
Definition 3
(Hamming distance). Given two production state sequences ω 1 = x 0 x 1 x k and ω 2 = x 0 x 1 x k generated by a critical state estimator O b s ( A n d ) with the same length ( ω 1 = ω 2 ), the Hamming distance between ω 1 and ω 2 is defined as J ( ω 1 , ω 2 ) = | { j ω 1 ( j ) ω 2 ( j ) } | , where ω i ( j ) represents the j-th entry of production state sequence ω i , i = 1 , 2 , and j = 1 , 2 , , | ω 1 | . ♢
Let O b s ( A n d ) = ( X o b s , E o , δ o b s , x 0 , o b s ) be a critical state estimator of A n d , k N + = { 1 , 2 , } be a positive integer representing the length of a production state sequence, and l N be a parameter. We define W k , l as the set of all pairs of production state sequences ( ω 1 , ω 2 ) with ω 1 = ω 2 = k and the Hamming distance J ( ω 1 , ω 2 ) between them is less than or equal to l, i.e.,
W k , l = { ( ω 1 , ω 2 ) X o b s k × X o b s k J ( ω 1 , ω 2 ) l } .
Considering Definition 1 and the properties of DES models, we now define state sequence DP in a DFA.
Definition 4
(State sequence differential privacy). Given an NFA A n d , let O b s ( A n d ) = ( X o b s , E o , δ o b s , x 0 , o b s ) be its critical state estimator and X o b s k be the set of state sequences with length k in O b s ( A n d ) . Let ϵ 0 be a real number. A mechanism M ω : X o b s k X o b s k is said to be of state sequence DP w.r.t. W k , l if for all production state sequence pairs ( ω 1 , ω 2 ) W k , l and for all ω 3 X o b s k , it holds
P [ M ω ( ω 1 ) = ω 3 ] exp ( ϵ ) P [ M ω ( ω 2 ) = ω 3 ] ,
where P [ M ω ( ω i ) = ω 3 ] represents the probability of M ω ( ω i ) = ω 3 with i { 1 , 2 } .
According to Definition 4, a mechanism that guarantees DP of state sequences w.r.t. W k , l implies that for all pairs of production state sequences ( ω 1 , ω 2 ) W k , l with ω 1 , ω 2 X o b s k , and for all possible output production state sequences ω 3 X o b s k , the probability of ω 1 being decorated as ω 3 and that of ω 2 being decorated as ω 3 is close enough such that an intruder cannot identify which is the real input to M ω .
To evaluate the performance of the mechanism, we introduce the concept of the security degree of a state sequence. Given an NFA A n d = ( X n d , E , δ n d , X 0 ) , a set of observable events E o E , a set of critical states X s X n d , and the critical state estimator O b s ( A n d ) = ( X o b s , E o , δ o b s , x 0 , o b s ) of A n d , let ω = x 0 x 1 x k X o b s k + 1 be a state sequence in O b s ( A n d ) . If the production state sequence ω is safe, the security degree of ω is defined as d e g ( ω ) = | x k | , where x k = C ( ρ ) with ρ T ( ω ) . Otherwise, the sequence ω is unsafe, whose security degree is defined as d e g ( ω ) = 0 . In other words, the state sequence ω consistent with an observation ρ that violates critical observability or reveals the critical states (i.e., C ( ρ ) X N or C ( ρ ) X s ) has the lowest security level. Note that for all ρ T ( ω ) , C ( ρ ) is unique. Given a production state sequence ω in O b s ( A n d ) , a larger value d e g ( ω ) means that the set of critical observability violative states X N or the set of critical states X s is less likely to be detected by an intruder.
Consider a critical state estimator O b s ( A n d ) of a system A n d and a parameter k N + . The maximum security degree of production state sequences ω = x 0 x 1 x k X o b s k + 1 contained in O b s ( ( A n d ) is defined as d e g m a x ( O b s ( A n d ) , k + 1 ) = m a x { g N ω X o b s k + 1 : d e g ( ω ) = g } 1.
Example 3.
Consider the NFA A n d = ( X n d , E , δ n d , X 0 ) and a set of critical states X s = { x 3 } in Example 2. Let ξ = x 0 , o b s α x o 1 λ x o 3 β x o 1 be a production of O b s ( A n d ) portrayed in Figure 4 with F ( ξ ) = ω = x 0 , o b s x o 1 x o 3 x o 1 . The security degree of ω is 2, i.e., d e g ( ω ) = | { x 1 , x 4 } | = 2 . Given another production ξ = x 0 , o b s β x o 2 λ x o 4 , the production state sequence F ( ξ ) = ω = x 0 , o b s x o 2 x o 4 is unsafe and ω has the lowest security degree d e g ( ω ) = 0 thanks to x o 4 X N . Given a non-negative integer k = 4 representing the length of a state sequence, we have d e g m a x ( O b s ( A n d ) , 4 ) = 2 . □
In this paper, we focus on a mechanism that provides the highest degree of protection for an unsafe production state sequence. In simpler terms, the mechanism is capable of confusing unsafe state sequences with safe ones that have the maximum security degree. In particular, the designed mechanism w.r.t. W k + 1 , l ensures that, for an unsafe production state sequence ω 1 = x 0 x 1 x k , there exists a production state sequence ω 2 = x 0 x 1 x k with maximum degree d e g m a x ( O b s ( A n d ) , k + 1 ) such that ( ω 1 , ω 2 ) W k + 1 , l and P [ M ω ( ω 1 ) = ω 3 ] exp ( ϵ ) P [ M ω ( ω 2 ) = ω 3 ] for all ω 3 X o b s k + 1 .
The distance bound l is another parameter that significantly impacts the performance of the mechanism M ω . Given an unsafe state sequence ω 1 , it is crucial to select a value l that is as small as possible while guaranteeing that the mechanism M ω can maximally confuse ω 1 with ω 2 . In other words, the state sequence ω 2 , which serves as a suitable alternative for ω 1 , should exhibit the utmost similarity to ω 1 .
Definition 5.
Let A n d = ( X n d , E , δ n d , X 0 ) be an NFA that is not critically observable, X s X n d be a set of critical states, E o E be a set of observable events, O b s ( A n d ) = ( X o b s , E o , δ o b s , x 0 , o b s ) be a critical state estimator of A n d , and M ω : X o b s k + 1 X o b s k + 1 be a mechanism that provides state sequence DP w.r.t. W k + 1 , l . Given an unsafe state sequence ω 1 , the mechanism M ω is optimal w.r.t. ω 1 if it satisfies the following statements:
  • There exists a production state sequence ω 2 in O b s ( A n d ) with d e g ( ω 2 ) = d e g m a x ( O b s ( A n d ) , k + 1 ) such that ( ω 1 , ω 2 ) W k + 1 , l holds;
  • For all ω 3 X o b s k + 1 , the inequality P [ M ω ( ω 1 ) = ω 3 ] exp ( ϵ ) P [ M ω ( ω 2 ) = ω 3 ] holds;
  • For all ω 2 in O b s ( A n d ) with d e g ( ω 2 ) = d e g m a x ( O b s ( A n d ) , k + 1 ) , there does not exist l < l ( l , l N ) such that J ( ω 1 , ω 2 ) = l . ♢
According to Definition 5, given an unsafe ω 1 , condition 1) ensures that there exists at least one state sequence ω 2 with the maximum security degree such that ( ω 1 , ω 2 ) belongs to the set W k + 1 , l . Condition 2) indicates that an optimal mechanism provides (guarantees) the state sequence DP w.r.t. W k + 1 , l . Condition 3) implies that the parameter l with l = J ( ω 1 , ω 2 ) must be minimal for all state sequences ω 2 that satisfy d e g ( ω 2 ) = d e g m a x ( O b s ( A n d ) , k + 1 ) . In the following, we will detail the design of an optimal mechanism M ω that satisfies the definition of state sequence DP.

4.2. State Sequence Differential Privacy

Based on Definitions 4 and 5, we first design a probability allocation strategy for the optimal mechanism M ω in a critical state estimator O b s ( A n d ) = ( X o b s , E o , δ o b s , x 0 , o b s ) of system A n d . Given an unsafe input production state sequence ω 1 = x 0 x 1 x k X o b s k + 1 , the aim is to generate a random state sequence ω 3 = x 0 x 1 x k X o b s k + 1 with a certain probability from M ω such that P [ M ω ( ω 1 ) = ω 3 ] . Specifically, the strategy to be defined enables assigning a probability P [ x i ] to each state included in the output production state sequence ω 3 = x 0 x 1 x k X o b s k + 1 , where i = 0 , , k . Then, by multiplying the probabilities P [ x i ] of all output states x i X o b s ( i = 0 , 1 , , k ), we can obtain the probability of generating ω 3 , i.e., P [ M ω ( ω 1 ) = ω 3 ] = P [ x 0 ] × P [ x 1 ] × × P [ x k ] . In terms of the strategy, we explore the conditions for ensuring the DP of state sequences and further design an algorithm to synthesize this optimal mechanism M ω .
Given a critical state estimator O b s ( A n d ) = ( X o b s , E o , δ o b s , x 0 , o b s ) , let M ω : X o b s k + 1 X o b s k + 1 be a mechanism that takes a production state sequence ω 1 = x 0 x 1 x k X o b s k + 1 as input and ω 3 = x 0 x 1 x k X o b s k + 1 as output. Before elaborating on the probability allocation strategy to a state in ω 3 , we introduce an indicator function denoted as ν ( x i 1 , x i ) for two states x i 1 , x i X o b s , defined as follows:
ν ( x i 1 , x i ) = 1 , if e E o : δ o b s ( x i 1 , e ) = x i ; 0 , otherwise ,
where i = 1 , 2 , , k . If x i included in ω 3 satisfies x i = x i , it is considered a correct state. On the other hand, if x i x i , the state x i is referred to as an incorrect state.
Now, we describe a probability allocation strategy that assigns a probability P [ x i ] to each state x i in the state sequence ω 3 = x 0 x 1 x k X o b s k + 1 , where i { 0 , , k } . Given an input production state sequence ω 1 = x 0 x 1 x k X o b s k + 1 , let X ω 1 be the set that collects all the states in ω 1 , i.e., X ω 1 = { x 0 , x 1 , , x k } . The strategy for ω 3 associated with ω 1 is a (partial) function φ : X o b s × X ω 1 × X o b s [ 0 , 1 ] that outputs a probability of generating state x i X o b s in relation to two states x i 1 X o b s and x i X ω 1 , where ( x i 1 , x i , x i ) X o b s × X ω 1 × X o b s . Particularly, for the initial state, we define x 0 = x 0 = x 0 , o b s , and it is assigned a probability value P [ x 0 ] = 1 . Then, for all i { 1 , , k } , if x i is a correct state, it is allocated a correct probability φ ( x i 1 , x i , x i ) = P [ x i ] = η ( x i ) (the value of η ( x i ) is examined in detail in Theorem 1). On the other hand, if x i is not a correct state, let
φ ( ( x i 1 , x i , x i ) = P [ x i ] = 1 η ( x i ) ν ( x i 1 , x i ) E ( x i 1 ) ν ( x i 1 , x i ) .
Algorithm 1: A probability Allocation Strategy φ .
Preprints 139195 i001
Then, we design an algorithm (namely Algorithm 1) to compute the probability allocation strategy φ for a random output ω 3 = x 0 x 1 x k associated with an input state sequence ω 1 = x 0 x 1 x k . The algorithm takes as input a critical state estimator, an input state sequence, probabilities of correct states, and an indicator function. It iterates through each state in the input sequence and calculates the probability of generating each possible output state based on the given probabilities of correct states and the indicator function. In Algorithm 1, for all i = 1 , , k , if i = 1 , we have x 0 = x 0 = x 0 , o b s . For each possible state x i X o b s , it checks whether x i = x i and whether the indicator function ν ( x i 1 , x i ) = 1 . If both conditions hold, the correct state x i is allocated a correct probability η ( x i ) . If x i x i and the indicator function ν ( x i 1 , x i ) = 1 , the state x i is assigned an incorrect probability
1 η ( x i ) ν ( x i 1 , x i ) E ( x i 1 ) ν ( x i 1 , x i ) .
Otherwise, if ν ( x i 1 , x i ) = 0 , φ ( x i 1 , x i , x i ) = 0 is defined. If 1 < i k , for all x i 1 X o b s , the codes in lines 4–10 are repeated. The computational complexity of Algorithm 1 is O ( k × X o b s ) .
Example 4.
Consider a state sequence ω 1 = x 0 x 1 x 2 = x 0 , o b s x o 2 x o 4 in the critical state estimator in Figure 4 and feed into a mechanism M ω . Let us elaborate on how to calculate the probability allocation strategy φ for a randomly generated output state sequence ω 3 = x 0 x 1 x 2 with x 0 = x 0 = x 0 , o b s (all possible output state sequences ω 3 with ω 3 = 3 are illustrated in Figure 5).
If x 1 = x 1 = x o 2 , according to the strategy φ, the state x 1 is correct and is assigned a correct probability η ( x 1 ) with φ ( x 0 , x 1 , x 1 ) = φ ( x 0 , o b s , x o 2 , x o 2 ) = η ( x 1 ) (the value of η ( x 1 ) is discussed in Theorem 1); if x 1 = x o 1 with x o 1 x o 2 , by the line 8 of Algorithm 1, we allocate an incorrect probability to state x o 1 , i.e., φ ( x 0 , x 1 , x 1 ) = φ ( x 0 , o b s , x o 2 , x o 1 ) = 1 η ( x 1 ) × 1 2 1 = 1 η ( x 1 ) .
At the state x 1 = x o 1 , since the feasible state x 2 = x o 3 is not correct, i.e., x 2 x 2 and ν ( x 1 , x 2 ) = ν ( x o 1 , x o 3 ) = 1 , we have φ ( x 1 , x 2 , x 2 ) = φ ( x o 1 , x o 4 , x o 3 ) = 1 η ( x 1 ) × 0 1 0 = 1 . Analogously, at the state x 1 = x o 2 , we have φ ( x o 2 , x o 4 , x o 4 ) = η ( x 2 ) and φ ( x o 2 , x o 4 , x o 2 ) = φ ( x o 2 , x o 4 , x o 3 ) = 1 η ( x 2 ) × 1 3 1 = 1 η ( x 2 ) 2 . □
By utilizing the probability allocation strategy φ , we are able to synthesize a mechanism that generates a production state sequence assigned with a probability. In other words, for all states contained in ω 3 = x 0 x 1 x k , the probability of generating ω 3 is given by i = 1 k φ ( x i 1 , x i , x i ) with P [ x 0 ] = P [ x 0 , o b s ] = 1 . Now, we demonstrate the condition of η ( x i ) for a mechanism M ω to provide state sequence DP w.r.t. W k + 1 , l , based on Algorithm 1.
Theorem 1.
Given a critical state estimator O b s ( A n d ) = ( X o b s , E o , δ o b s , x 0 , o b s ) of a system A n d that is not critically observable, let ϵ > 0 be a real number, l N be a parameter, k N + be the length of a state sequence, M ω : X o b s k + 1 X o b s k + 1 be a mechanism whose probability allocation strategy is computed by Algorithm 1, and x i ( 0 i k ) be a correct state contained in an output state sequence ω 3 = x 0 x 1 x k X o b s k + 1 of M ω . The mechanism M ω guarantees state sequence DP w.r.t. W k + 1 , l if the probabilities of correct states η ( x i ) satisfy the following condition:
η ( x i ) = 1 ( | E ( x i 1 ) | 1 ) exp ( ϵ l ) + 1 .
Proof. Let ω 1 = x 0 x 1 x k and ω 2 = x 0 x 1 x k be two production state sequences generated by O b s ( A n d ) such that ( ω 1 , ω 2 ) W k + 1 , l . We consider M ω ( ω 1 ) = M ω ( ω 2 ) = ω 3 with ω 3 = x 0 x 1 x k X o b s k + 1 (the initial states in ω 1 , ω 2 , and ω 3 are the same, i.e., x 0 = x 0 = x 0 = x 0 , o b s ). By Definition 4, we need to prove the following inequality:
exp ( ϵ ) P [ M ω ( ω 1 ) = ω 3 ] P [ M ω ( ω 2 ) = ω 3 ] exp ( ϵ ) .
To begin with, we show P [ M ω ( ω 1 ) = ω 3 ] P [ M ω ( ω 2 ) = ω 3 ] exp ( ϵ ) . Using Algorithm 1, we have
P [ M ω ( ω 1 ) = ω 3 ] P [ M ω ( ω 2 ) = ω 3 ] = η ( x 0 ) × i = 1 k φ ( x i 1 , x i , x i ) η ( x 0 ) × j = 1 k φ ( x j 1 , x j , x j ) = r R φ ( x r 1 , x r , x r ) φ ( x r 1 , x r , x r ) ,
where R = { r N + x r x r } and | R | l .
Note that when x i is an incorrect state, the values of φ ( x i 1 , x i , x i ) depend on ν ( x i 1 , x i ) , i.e.,
φ ( x i 1 , x i , x i ) = 1 η ( x i ) | E ( x i 1 ) | 1 if ν ( x i 1 , x i ) = 1 ; 1 | E ( x i 1 ) | otherwise .
We now claim η ( x i ) 1 | E ( x i 1 ) | 1 η ( x i ) | E ( x i 1 ) | 1 .
1) First, we prove η ( x i ) 1 | E ( x i 1 ) | . One has
η ( x i ) 1 | E ( x i 1 ) | = 1 ( | E ( x i 1 ) | 1 ) exp ( ϵ l ) + 1 1 | E ( x i 1 ) | = ( | E ( x i 1 ) | 1 ) ( 1 exp ( ϵ l ) ) ( ( | E ( x i 1 ) | 1 ) exp ( ϵ l ) + 1 ) | E ( x i 1 ) | 0 .
2) Let us prove the second inequality 1 | E ( x i 1 ) | 1 η ( x i ) | E ( x i 1 ) | 1 .
1 | E ( x i 1 ) | 1 η ( x i ) | E ( x i 1 ) | 1 = | E ( x i 1 ) | 1 | E ( x i 1 ) | + | E ( x i 1 ) | η ( x i ) | E ( x i 1 ) | · ( | E ( x i 1 ) | 1 ) = | E ( x i 1 ) | η ( x i ) 1 | E ( x i 1 ) | · ( | E ( x i 1 ) | 1 ) = ( | E ( x i 1 ) | 1 ) · ( 1 exp ( ϵ l ) ) ( | E ( x i 1 ) | 1 ) · exp ( ϵ l ) + 1 | E ( x i 1 ) | · ( | E ( x i 1 ) | 1 ) 0 .
Then we conclude that
P [ M ω ( ω 1 ) = ω 3 ] P [ M ω ( ω 2 ) = ω 3 ] = r R φ ( x r 1 , x r , x r ) φ ( x r 1 , x r , x r ) r R η ( x r ) 1 η ( x r ) | E ( x r 1 ) | 1 = r R exp ( ϵ l ) = i = 1 | R | exp ( ϵ l ) exp ( ϵ ) .
The constraint r R φ ( x r 1 , x r , x r ) φ ( x r 1 , x r , x r ) exp ( ϵ ) can be proved similarly, which completes the proof.
By Definition 4 and Theorem 1, Algorithm 2 computes an optimal mechanism for an unsafe input state sequence ω 1 , which aims to prevent the leakage of critical observability violative states or critical states in a non-critically observable system. [2]We use a number in square brackets after a tuple to represent the entries of the tuple.
[3]A number in square brackets after a set of tuples represents the set contained by the entries of the tuple set. Here, B [ 2 ] = { b 1 [ 2 ] , , b m [ 2 ] } with b i B ( i = 1 , , m ) .
Algorithm 2: Construction of a mechanism M ω .
Preprints 139195 i002
Algorithm 2 first initializes two empty sets B and G for storing the triplet elements b B and g G . Then, it iterates through all the productions ξ = x 0 e 1 x 1 e 2 e k x k with F ( ξ ) = ω 2 . In lines 3–6, for each such sequence ω 2 , the algorithm executes three key computations: assigning ω 2 , the degree d e g ( ω 2 ) of ω 2 , and the Hamming distance J ( ω 1 , ω 2 ) to b [ 1 ] , b [ 2 ] , and b [ 3 ] , respectively. In line 7, the maximum value in set B [ 2 ] is computed and is stored in the variable s. Next, for each triplet b in the set B, check whether b [ 2 ] is equal to the maximum number s. If the condition is true, then add this element b to the set G. In line 11, the value of the parameter h is obtained by calculating the minimum number in set G [ 3 ] . For all input state sequences ω = x 0 x 1 x k X o b s k + 1 and for all ω 3 = x 0 x 1 x k X o b s k + 1 , in line 15, we compute the correct probability η ( x i ) for each output state x i . Finally, in lines 16–17, Algorithm 2 synthesizes the mechanism M ω by invoking the probability allocation strategy φ of Algorithm 1.
The computational complexity of Algorithm 2 is mainly decided by the synthesis of the mechanism M ω in lines 12–17. In lines 12–13, we enumerate all the input and output sequences, resulting in complexity O ( | X o b s | 2 k ) . Furthermore, in lines 14–16, we compute η ( x i ) and utilize Algorithm 1 to calculate φ . This step has complexity O ( k + k × | X o b s | ) . Taking into account the above statements, the complexity can be expressed as O ( k × | X o b s | 2 k × ( 1 + | X o b s | ) ) , i.e., O ( k × | X o b s | 3 k ) .
Theorem 2.
Consider a critical state estimator O b s ( A n d ) = ( X o b s , E o , δ o b s , x 0 , o b s ) of a system A n d that is not critically observable. Let ω 1 = x 0 x 1 x k X o b s k + 1 be an unsafe production state sequence in O b s ( A n d ) . The mechanism M ω computed by Algorithm 2 guarantees DP of state sequences w.r.t. W k + 1 , h , in which h N is a parameter calculated by Algorithm 2. Moreover, the mechanism is optimal w.r.t. an unsafe production state sequence ω 1 .
Proof. First, we prove that the mechanism M ω designed in Algorithm 2 provides DP of state sequences w.r.t. W k + 1 , h . Specifically, in lines 12–17, for all input state sequences ω and any possible output state sequence ω 3 , we have P [ M ω ( ω ) = ω 3 ] = i = 1 k φ ( x i 1 , x i , x i ) . Besides, we accurately compute the correct state probability η ( x i ) based on Equation (2). According to Theorem 1, the mechanism M ω provides DP of state sequences w.r.t. W k + 1 , h .
Then we show that the mechanism M ω is optimal w.r.t. the unsafe state sequence ω 1 . In lines 2–6, for all productions ξ , the security degrees of their corresponding state sequences F ( ξ ) = ω 2 are calculated and these values are stored in the set B [ 2 ] . Then the algorithm calculates the Hamming distance between ω 1 and ω 2 , which reflects the degree of difference between them. In lines 7–10 of the algorithm, elements b are assigned to G such that b [ 2 ] = s = d e g ( ω 2 ) = d e g m a x ( O b s ( A n d ) , k + 1 ) . In line 11, for all g [ 3 ] G [ 3 ] , the minimum value h in the set G [ 3 ] is obtained, i.e., compute the minimum value h of the Hamming distance between ω 1 and ω 2 . To this end, there exists a state sequence ω 2 with d e g ( ω 2 ) = d e g m a x ( O b s ( A n d ) , k + 1 ) such that ( ω 1 , ω 2 ) W k + 1 , h and there does not exist h < h such that J ( ω 1 , ω 2 ) = h for all ω 2 with the maximum degree. According to Definition 5, the mechanism M ω is optimal.
In Algorithm 2, we design an optimal mechanism M ω w.r.t. an unsafe production state sequence ω 1 . When the mechanism M ω receives ω 1 as input, a protected state sequence ω 3 is probabilistically output. In addition, the mechanism guarantees that there exists at least one state sequence ω 2 with the maximum security degree such that when ω 2 is fed, it outputs ω 3 with an approximate probability. Then, when an intruder obtains an observation ρ T ( ω 3 ) derived from ω 3 , he/she is unable to infer whether ω 1 or ω 2 is generated in the critical state estimator O b s ( A n d ) . Consequently, the disclosure of critical states or critical observability violative states becomes unattainable for the intruder. Furthermore, in order to maximize the obfuscation effect between ω 1 and ω 2 , we especially emphasize the optimization of the parameter h in Algorithm 2, i.e., the Hamming distance between ω 1 and ω 2 is minimal.
Example 5.
We consider the NFA A n d = ( X n d , E , δ n d , X 0 ) shown in Figure 3. Based on the results in Example 2 by assuming that the observable event set is E o = { α , β , λ } and the secret is X s = { x 3 } , the system A n d is not critically observable. Suppose that an unsafe observation ρ = β α λ is generated, in which the state estimate C ( ρ ) = x o 4 is neither a subset of the critical states nor a subset of the non-critical states. The unsafe state sequence ω 1 = x 0 , o b s x o 2 x o 2 x o 4 F ( ρ ) corresponding to ρ is obtained in the critical state estimator shown in Figure 4. By lines 2–6 of Algorithm 2, we calculate the set B illustrated in Table 1 (the contents of columns 1, 2 and 3 record B [ 1 ] , B [ 2 ] and B [ 3 ] , respectively). Then, we compute the maximum degree of the production state sequences in B [ 1 ] , i.e., s = m a x { b 1 [ 2 ] , b 2 [ 2 ] , , b 8 [ 2 ] } = 2 and choose the minimum parameter h = 3 .
We take the input sequence ω 1 = x 0 , o b s x o 2 x o 2 x o 4 and a random output ω 3 = x 0 , o b s x o 1 x o 3 x o 5 for an example. The probabilities of the correct states are
η ( x i ) = 1 ( | E ( x i 1 ) | 1 ) exp ( ϵ 3 ) + 1
with i = { 1 , 2 , 3 } , where in general ϵ [ 0.01 , 10 ] . Here, we consider ϵ = 0.1 .
By Equation (3), we have
η ( x 0 ) = η ( x 0 , o b s ) = 1 , η ( x 1 ) = η ( x o 2 ) 0.5083 , η ( x 2 ) = η ( x o 2 ) 0.3408 , and η ( x 3 ) = η ( x o 4 ) 0.3408 .
The probability for generating ω 3 is
P [ M ω ( ω 1 ) = ω 3 ] = P [ x 0 , o b s ] × P [ x o 1 ] × P [ x o 3 ] × P [ x o 5 ] = η ( x 0 ) × 1 η ( x 1 ) 2 1 × 1 0 × η ( x 2 ) 1 0 × 1 0 × η ( x 3 ) 2 0 1 × ( 1 0.5083 ) × 1 × 1 2 0.2458 .
Analogously, for another random input state sequence (e.g., ω 2 = x 0 , o b s x o 1 x o 3 x o 1 with the maximum security degree), we have probabilities for the correct states
η ( x 0 ) = η ( x 0 , o b s ) = 1 , η ( x 1 ) = η ( x o 1 ) 0.5083 , η ( x 2 ) = η ( x o 3 ) = 1 , and η ( x 3 ) = η ( x o 1 ) 0.5083 .
When considering ω 2 as an input of M ω , the probability of generating ω 3 is
P [ M ω ( ω 2 ) = ω 3 ] = P [ x 0 , o b s ] × P [ x o 1 ] × P [ x o 3 ] × P [ x o 5 ] = η ( x 0 ) × η ( x 1 ) × η ( x 2 ) × 1 η ( x 3 ) 2 1 1 × 0.5038 × 1 × ( 1 0.5038 ) 0.2499 ,
and one has
P [ M ω ( ω 1 ) = ω 3 ] exp ( ϵ ) P [ M ω ( ω 2 ) = ω 3 ] .
In other words, if the observation ρ = β α λ is directly sent, the critical observability violative state x o 4 can be detected by an intruder due to C ( ρ ) = { x 3 , x 4 , x 5 } X s and C ( ρ ) X n d X s . However, by inputting ω 1 = x 0 , o b s x o 2 x o 2 x o 4 F ( ρ ) into the designed optimal DPM, a modified state sequence ω 3 = x 0 , o b s x o 1 x o 3 x o 5 is output. Just by acquiring the sequence ω 3 , it is difficult for an intruder to backtrack or infer the original sequence ω 1 , since there exists another random sequence ω 2 = x 0 , o b s x o 1 x o 3 x o 1 which can also generate ω 3 with an approximate probability. □

5. Case Study

In this section, we consider a realistic case of enforcing critical observability in air traffic management (ATM) systems. In recent years, with the development of ATM systems, their network architecture has gradually shifted from closed, physically isolated networks to more interconnected and open cyber-physical systems. This evolution has made data exchange and information sharing efficient and convenient, but it has also created more possibilities for potential network attacks. Frequent and diverse types of network attacks pose a serious risk to the safety and stability of aviation. In this context, to strengthen security defenses by enhancing the system’s awareness and response capability to potential threats, we propose an innovative solution: unsafe observations protection based on DP.
An ATM system is illustrated in Figure 6, detailing the communication process between the airport control tower and the pilot during the various stages of flight. On the one hand, the control tower sends a series of instructions to the pilot, such as pre-flight/stop-flight instructions, engine start instructions, etc. These instructions guide the pilot’s actions and determine that the aircraft is airworthy and all parts of the system are operating normally. On the other hand, the pilot’s role is to wait for and follow instructions from the control tower, performing the corresponding operations. If there are any faults or abnormal situations during the flight, the pilot must immediately report to the tower. This real-time feedback mechanism allows the control tower to take action quickly to resolve issues that may threaten flight safety. The exchange of information between the control tower and the pilot runs throughout the entire flight process and jointly affects the aircraft’s maneuvers (represented by the downward arrows in Figure 6). A complete aerial flight can be divided into five stages: taxiing, climbing, cruising, descending, and landing. Each stage requires precise coordination between the pilot and the control tower. Based on Figure 6, we model an ATM system as an NFA (shown in Figure 7) by using a single aircraft as a study case and define a series of important physical states as critical states. The meanings of states and events in the NFA are listed in Table 2. We further assume that an intruder launches an attack as soon as he/she captures the critical states or critical observability violative states. To prevent the system A n d from being corrupted, we introduce a DPM, such that the unsafe states are not accurately captured by an intruder. From a practical standpoint, this research establishes the groundwork for subsequent security defense of cyber-physical systems and has significant implications for safeguarding air transportation safety and maintaining airspace order.
In Figure 7, the pilot initially executes boarding and waits for instructions from the airport control tower (state x 0 ). When the pilot receives a pre-flight instruction sent by the tower (event a), the state is transferred from x 0 to x 1 , at which point the pilot executes a pre-flight check. For any error found by the pilot during the check, an fault is reported (event d) to the tower, and the state x 3 is reached. In state x 2 , the aircraft is waiting for the engine to start. If the engine fails to start (event e) due to safety issues or other conditions, the state remains in x 2 ; otherwise, the state transitions to x 4 , where the pilot communicates with the tower in the hope of obtaining a taxi clearance. The taxiing process is represented by states x 6 , x 7 , and x 8 , which correspond to different runways. The event h, obtaining taxi clearance, triggers the transition from x 4 to one of these taxing states. Once the taxi clearance is not granted (event n), the state transitions from x 4 to x 13 or x 5 .
The event i, sending a climbing request, triggers the transition from one of the taxiing states ( x 6 , x 7 , or x 8 ) to state x 9 . If the climb request is granted (event j), the aircraft takes off and leaves the ground (state x 10 ). In addition, fault reports (event d) can lead to a transition from x 10 to x 3 , allowing the aircraft to report any troubles encountered during the flight. However, if the climb request is rejected (event l), the state is switched from x 9 to x 12 , in which case the flight is stopped and the aircraft returns to its origin. In exceptional cases, the aircraft may be started without authorization, resulting in a move from x 9 to x 13 , as triggered by event l. In x 13 , the pilot sends an anomaly report to the tower (event m), where the aircraft refuses to execute the flight mission and x 12 or x 5 is reached.
Example 6.
Let us consider the NFA A n d = ( X n d , E , δ n d , X 0 ) portrayed in Figure 7, where E = E o ˙ E u o is the set of events with E o = { a , b , c , d , e , f , g , h , i , j , k , l , n , o , p } and E u o = { m } , and X 0 = { x 0 } is the set of initial states. We assume that the set of critical states is X s = { x 5 , x 12 } . The critical state estimator O b s ( A n d ) = { X o b s , E o , δ o b s , x 0 , o b s } of A n d is shown in Figure 8, where the components of all states are listed in Table 3.
Given an observation ρ = a c f n , we conclude that the critical observability is violated thanks to C ( ρ ) = δ o b s ( x 0 , o b s , a c f n ) = x o 7 = { x 5 , x 12 , x 13 } X s and C ( ρ ) X n d X s . Therefore, ρ is an unsafe observation and ω 1 = x 0 , o b s x o 1 x o 2 x o 4 x o 7 F ( ρ ) is an unsafe production state sequence.
We now show how the unsafe production state sequence is protected by designed optimal mechanisms M ω that guarantees DP. Consider ω 1 as an input of M ω . By Algorithm 2, for all ω 2 X o b s 5 , the security degrees d e g ( ω 2 ) and Hamming distances J ( ω 1 , ω 2 ) are listed in Table 4. Then we compute the maximum degree s = 3 and for all ω 2 with d e g ( ω 2 ) = d e g m a x ( O b s ( A n d ) , 5 ) = 3 , we select the minimum parameter h = 1 . For all ( ω 1 , ω 2 ) W 5 , 1 , and for all ω 3 X o b s 5 , it holds that
P M ω ( ω 1 ) = ω 3 exp ( ϵ ) P [ M ω ( ω 2 ) = ω 3 ] ,
where in general ϵ [ 0.01 , 10 ] . Here, we consider ϵ = 0.1 .
When observing ρ = a c f n , the state that violates critical observability is detectable by an intruder. However, by applying the designed optimal mechanism that outputs a randomly modified production state sequence ω 3 , the unsafe sequence ω 1 F ( ρ ) cannot be ensured, since another sequence ω 2 with ( ω 1 , ω 2 ) W 5 , 1 can generate ω 3 with a similar probability. In this way, unsafe observations in the ATM system cannot be accurately captured. Thereby, the intruder is unable to launch an attack to the ATM system A n d . □

6. Conclusions

This work proposes a differential privacy (DP) mechanism for the enforcement of critical observability in DESs modeled with finite state automata. Meanwhile, the mechanism protects the predefined critical states from being captured by an intruder. We first recall the formalization of critical observability verification problems. Then, a mechanism called state sequence DP is introduced to protect critical states or critical observability violative states such that it cannot be detected by an intruder. Furthermore, in order to most effectively protect unsafe production state sequences, we define an optimal mechanism and design an algorithm for synthesizing the optimal mechanism. Finally, the problem of enforcing critical observability is explored in an ATM system, where model construction of the ATM system and protection implementation of unsafe observations using DP are illustrated in detail.
From a practical perspective, this research has potential applications in various cyber-physical systems. In the future, we will consider applying the differential privacy mechanism to the verification and enforcement of other properties such as k-step and infinite-step opacity. Moreover, it is worth exploring the critical observability problem when a system is subject to sensor attacks, i.e., sensor readings can altered by an external attacker.

References

  1. Lin, F. Opacity of discrete event systems and its applications. Automatica 2011, 47, 496–503. [Google Scholar] [CrossRef]
  2. Yang, J.; Deng, W.; Qiu, D.; Jiang, C. Opacity of networked discrete event systems. Inf. Sci. 2021, 543, 328–344. [Google Scholar] [CrossRef]
  3. Deng, W.; Qiu, D.; Yang, J. Opacity measures of fuzzy discrete Event Systems. IEEE Trans. Fuzzy Syst. 2021, 29, 2612–2622. [Google Scholar] [CrossRef]
  4. Cao, L.; Shu, S.; Lin, F.; Chen, Q.; Liu, C. Weak diagnosability of discrete-event systems. IEEE Trans. Control. Netw. Syst. 2022, 9, 184–196. [Google Scholar] [CrossRef]
  5. Liu, F.; Qiu, D. Safe diagnosability of stochastic discrete event systems. IEEE Trans. Autom. Control. 2008, 53, 1291–1296. [Google Scholar] [CrossRef]
  6. Balun, J.; Masopust, T. On verification of D-detectability for discrete event systems. Automatica 2021, 133. [Google Scholar] [CrossRef]
  7. Zhu, H.; Liu, G.; Yu, Z.; Li, Z. Detectability in discrete event systems using unbounded Petri nets. Mathematics 2023, 11. [Google Scholar] [CrossRef]
  8. Li, Y.; Wang, M.; Jones, A. An approach for the design of supervisory controller of discrete event systems. 3rd World Congress on Intelligent Control and Automation, 2000, pp. 2341–2346.
  9. De Santis, E.; Di Benedetto, M.D.; Di Gennaro, S.; D’Innocenzo, A.; Pola, G. Critical observability of a class of hybrid systems and application to air traffic management. In Stochastic Hybrid Systems: Theory and Safety Critical Applications; 2006; pp. 141–170.
  10. Zhu, G.; Yin, L.; Li, Y.; Li, Z.; Wu, N. Identification of labeled Petri nets from finite automata. Inf. Sci. 2024, 667, 120448. [Google Scholar] [CrossRef]
  11. Di Benedetto, M.D.; Di Gennaro, S.; D’Innocenzo, A. Discrete state observability of hybrid systems. Int. J. Robust Nonlinear Control 2009, 19, 1564–1580. [Google Scholar] [CrossRef]
  12. Tong, Y.; Ma, Z. Verification of k-step and definite critical observability in discrete-event systems. IEEE Trans. Autom. Control 2023, 68, 4305–4312. [Google Scholar] [CrossRef]
  13. Pola, G.; De Santis, E.; Di Benedetto, M.D.; Pezzuti, D. Design of decentralized critical observers for networks of finite state machines: A formal method approach. Automatica 2017, 86, 174–182. [Google Scholar] [CrossRef]
  14. Lai, A.; Lahaye, S.; Komenda, J. Observer construction for polynomially ambiguous max-plus automata. IEEE Trans. Autom. Control 2022, 67, 1582–1588. [Google Scholar] [CrossRef]
  15. Dwork, C. Differential privacy. International Colloquium on Automata, Languages, and Programming. Springer, 2006, pp. 1–12.
  16. Dwork, C.; McSherry, F.; Nissim, K.; Smith, A. Calibrating noise to sensitivity in private data analysis. 3rd Theory of Cryptography Conference. Springer, 2006, pp. 265–284.
  17. McSherry, F.; Talwar, K. Mechanism design via differential privacy. 48th Annual IEEE Symposium on Foundations of Computer Science. IEEE, 2007, pp. 94–103.
  18. Lin, C.; Wang, P.; Song, H.; Zhou, Y.; Liu, Q.; Wu, G. A differential privacy protection scheme for sensitive big data in body sensor networks. Ann. Telecommun. 2016, 71, 465–475. [Google Scholar] [CrossRef]
  19. Li, W.; Wang, S.; Wang, H.; Lu, Y. Child health dataset publishing and mining based on differential privacy preservation. Mathematics 2024, 12. [Google Scholar] [CrossRef]
  20. Zhao, Y.; Zhao, J.; Yang, M.; Wang, T.; Wang, N.; Lyu, L.; Niyato, D.; Lam, K.Y. Local differential privacy-based federated learning for internet of things. IEEE Internet Things J. 2021, 8, 8836–8853. [Google Scholar] [CrossRef]
  21. Ren, J.; Jiang, L.; Peng, H.; Lyu, L.; Liu, Z.; Chen, C.; Wu, J.; Bai, X.; Yu, P.S. Cross-network social user embedding with hybrid differential privacy guarantees. 31st ACM International Conference on Information and Knowledge Management, 2022, pp. 1685–1695.
  22. Jones, A.; Leahy, K.; Hale, M. Towards differential privacy for symbolic systems. 2019 American Control Conference, 2019, pp. 372–377.
  23. Chen, B.; Leahy, K.; Jones, A.; Hale, M. Differential privacy for symbolic systems with application to Markov Chains. Automatica 2023, 152. [Google Scholar] [CrossRef]
  24. Kang, H.; Zhang, S.; Jia, Q. A method for time-series location data publication based on differential privacy. Wuhan University Journal of Natural Sciences 2019. [Google Scholar] [CrossRef]
  25. Hua, J.; Gao, Y.; Zhong, S. Differentially private publication of general time-serial trajectory data. 2015 IEEE Conference on Computer Communications, 2015.
  26. Cassandras, C.G.; Lafortune, S. Introduction to Discrete Event Systems; Springer, 2009.
  27. Dwork, C.; Roth, A. The algorithmic foundations of differential privacy. Foundations and Trends in Theoretical Computer Science 2014, 9, 211–406. [Google Scholar] [CrossRef]
1
The notation m a x (or m i n ) applied to a set represents the maximum (or minimum) element contained in the set.
Figure 1. Framework of critical state concealment.
Figure 1. Framework of critical state concealment.
Preprints 139195 g001
Figure 2. DFA A d .
Figure 2. DFA A d .
Preprints 139195 g002
Figure 3. NFA A n d .
Figure 3. NFA A n d .
Preprints 139195 g003
Figure 4. Critical state estimator O b s ( A n d ) of A n d .
Figure 4. Critical state estimator O b s ( A n d ) of A n d .
Preprints 139195 g004
Figure 5. Probability allocation strategy for all possible output production state sequences ω 3 = x 0 x 1 x 2 X o b s 3 .
Figure 5. Probability allocation strategy for all possible output production state sequences ω 3 = x 0 x 1 x 2 X o b s 3 .
Preprints 139195 g005
Figure 6. Air traffic management systems.
Figure 6. Air traffic management systems.
Preprints 139195 g006
Figure 7. An NFA model A n d of an air traffic management system.
Figure 7. An NFA model A n d of an air traffic management system.
Preprints 139195 g007
Figure 8. The critical state estimator O b s ( A n d ) of the air traffic management system A n d .
Figure 8. The critical state estimator O b s ( A n d ) of the air traffic management system A n d .
Preprints 139195 g008
Table 1. Components of the set B in Example 5.
Table 1. Components of the set B in Example 5.
States Sequence Security Hamming
Degree Distance
x 0 , o b s x o 2 x o 2 x o 2 0 1
x 0 , o b s x o 2 x o 4 x o 3 1 2
x 0 , o b s x o 2 x o 2 x o 3 1 1
x 0 , o b s x o 2 x o 3 x o 5 1 2
x 0 , o b s x o 2 x o 4 x o 4 0 1
x 0 , o b s x o 2 x o 4 x o 2 0 2
x 0 , o b s x o 1 x o 3 x o 1 2 3
x 0 , o b s x o 1 x o 3 x o 5 1 3
Table 2. Meanings of states and events of the NFA portrayed in Figure 7.
Table 2. Meanings of states and events of the NFA portrayed in Figure 7.
state Meanings of states event Meanings of events
x 0 Pilot boards an aircraft and waits for instructions a Send a pre-flight instruction
x 1 Pilot executes a pre-flight check b Incomplete pre-flight check
x 2 Waiting for the engine to start c Complete pre-flight check
x 3 Fault reports are received d Send a fault report
x 4 Waiting for taxi clearance e Reject engine start
x 5 Rejected take off f Allow engine start
x 6 Aircraft taxiing to runway A 1 g Send stop-flight instructions
x 7 Aircraft taxiing to runway A 2 h Obtain taxi clearance
x 8 Aircraft taxiing to runway A 3 i Send a climb request
x 9 Aircraft waiting for authorization at the runway j Climb request granted
x 10 Aircraft take off k Send a cruise instruction
x 11 Aircraft cruise l Climb request rejected
x 12 Stop flight and return to origin m Send an anomaly report
x 13 Aircraft started without authorization n Taxi clearance not granted
x 14 Aircraft descent o Send a descent instruction
x 15 Aircraft landing p Send a landing instruction
Table 3. Components of the states in the estimator shown in Figure 8.
Table 3. Components of the states in the estimator shown in Figure 8.
state state estimation state state estimation
x 0 , o b s { x 0 } x o 7 { x 5 , x 12 , x 13 }
x o 1 { x 1 } x o 8 { x 9 }
x o 2 { x 2 } x o 9 { x 10 }
x o 3 { x 3 } x o 10 { x 11 }
x o 4 { x 4 } x o 11 { x 14 }
x o 5 { x 5 } x o 12 { x 15 }
x o 6 { x 6 , x 7 , x 8 }
Table 4. All production state sequences with length 5 in O b s ( A n d ) .
Table 4. All production state sequences with length 5 in O b s ( A n d ) .
States Sequence Security Hamming
Degree Distance
x 0 , o b s x o 1 x o 1 x o 1 x o 1 1 3
x 0 , o b s x o 1 x o 1 x o 1 x o 2 1 3
x 0 , o b s x o 1 x o 1 x o 1 x o 3 1 3
x 0 , o b s x o 1 x o 1 x o 3 x o 5 0 3
x 0 , o b s x o 1 x o 1 x o 2 x o 2 1 3
x 0 , o b s x o 1 x o 1 x o 2 x o 4 1 3
x 0 , o b s x o 1 x o 2 x o 2 x o 2 1 2
x 0 , o b s x o 1 x o 2 x o 2 x o 4 1 2
x 0 , o b s x o 1 x o 2 x o 4 x o 6 3 1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated