Preprint
Article

This version is not peer-reviewed.

An Algebraic Method for Constructing Bases in Binary Linear Codes for Information Dispersal Algorithms

A peer-reviewed version of this preprint was published in:
Mathematics 2026, 14(7), 1097. https://doi.org/10.3390/math14071097

Submitted:

13 February 2026

Posted:

14 February 2026

You are already at the latest version

Abstract
The algebraic analysis of linear code parameters reveals deep connections with cryptographic constructions, including the information dispersal algorithms (IDAs) and secret-sharing schemes. In this work, we propose an algebraic method for constructing bases of binary linear codes from subsets of codewords selected according to their generalized Hamming weights (GHWs). The approach employs a degree-compatible monomial ordering on the polynomial ring F2[x1, . . . , xn] and imposes the conditions d1(C) = 1 and dk (C) = n. Under these assumptions, we prove the existence of a generator matrix containing an invertible k × k submatrix, which guarantees correct information reconstruction. This structural property enables the direct application of binary linear codes to information dispersal and recovery mechanisms without the need for larger finite fields. We validate the proposed framework through algebraic proofs and an explicit example illustrating both the dispersal and recovery procedures. These results provide a theoretical foundation for the design of information dispersal schemes relying exclusively on binary linear codes.
Keywords: 
;  ;  ;  ;  

1. Introduction

The study of linear codes provides a mathematical foundation for secure communication, error detection and correction, and information protection. Furthermore, it underpins modern cryptographic and distributed systems, such as distributed storage, secure multiparty computation, secret-sharing schemes, and information dispersal algorithms (IDAs), along with their associated information recovery algorithms (IRAs) [1,2,8,15,16]. IDAs, introduced by Rabin [14], disperse data into n pieces, allowing reconstruction from any t < n pieces to ensure reliability and fault tolerance. Rabin’s construction employs matrices over algebraic structures such as polynomial rings, Euclidean rings, or finite fields [10], for example, Cauchy or Vandermonde matrices with invertible submatrices that enable reconstruction[19].
Binary linear codes are seldom used in IDAs due to generator matrices lacking invertible k × k submatrices, unlike those over R or C . However, parameters like minimum distance and generalized Hamming weights (GHWs) can indicate such properties. Binary codes, being fundamental, offer rigid structures ideal for studying GHWs, bases, and invertible submatrices. We focus on algebraic conditions for binary codes to support IDAs. We show that codes with d 1 ( C ) = 1 and d k ( C ) = n enable IDAs via an invertible k × k submatrix in the generator matrix, ensuring Rabin’s IDA validity. The method uses a degree compatible monomial order on F 2 [ x 1 , , x n ] to select codewords matching GHWs, forming a basis for binary IDAs without larger-field overhead. Sections 4.2 and 4.3 of [3] motivated us to consider dispersal matrices given by binary linear codewords with special characteristics, as well as those in [2], which allowed us to study the secret-sharing schemes and, therefore, the information dispersal algorithms in a more current way.
The paper is organized as follows. Section 2 reviews the necessary background on linear codes, GHWs, monomial orders, and IDAs. Section 3 presents the main algebraic construction, including the selection of a basis guided by GHWs and the definition of the IDA and IRA procedures. Section 4 illustrates the proposed framework through an example of a binary linear code that satisfies the required conditions. Section 5 presents the conclusions and outlines directions for future research. Finally, Appendix Section 1.1 details the method proposed by Johnsen and Verdure [6] for calculating the GHWs of a linear code.

1.1. Literature Review

This work intersects linear coding theory, algebraic methods, and IDAs. The relevant literature is summarized as follows:
1.
The algebraic foundations of linear codes, including generator matrices, parity-check matrices, and structural parameters, are classical and well established; see, for instance, [12,15,16]. These works provide the basic framework for understanding linear codes as vector spaces over finite fields and for analyzing their fundamental parameters.
2.
Generalized Hamming weights (GHWs), introduced by Wei [18], extend the notion of minimum distance and have become central invariants in the structural analysis of linear codes. Subsequent developments connected GHWs with algebraic and combinatorial objects, particularly through the study of matroids and Stanley-Reisner ideals. In this direction, Johnsen and Verdure [6,7] established a deep relationship between GHWs and graded free resolutions, while García-Marco et al. [5] further explored these ideas in the binary case. These results highlight the role of GHWs as tools for understanding the internal structure of linear codes beyond error-correcting capabilities.
3.
From the algebraic point of view, polynomial rings and monomial orders play an important role in the study of linear codes and their associated ideals; see, e.g., [11]. Degree-compatible monomial orders provide a natural way to organize monomials according to their weight, making them suitable for linking algebraic structures with combinatorial properties of codes.
4.
Information dispersal algorithms (IDA) were introduced by Rabin [1,2,14] as a method for distributing information among multiple participants in a fault-tolerant manner. Classical constructions of IDA usually use matrices from large finite fields, such as Vandermonde or Cauchy matrices, to ensure the existence of invertible submatrices [19]. While these approaches are effective, they inherently depend on non-binary fields.

1.2. Contributions

This work presents an algebraic method for constructing bases of binary linear codes with structural properties suitable for information dispersal. The contributions of this paper can be summarized as follows.
  • An algebraic method is introduced for selecting a basis of an [ n , k ] binary linear code by combining GHWs with a degree-compatible monomial order on the polynomial ring F 2 [ x 1 , , x n ] . The resulting basis reflects the hierarchy of supports determined by the GHWs of the code.
  • Under the structural assumptions d 1 ( C ) = 1 and d k ( C ) = n , we prove that the basis obtained through this construction yields a generator matrix containing at least one invertible k × k submatrix. This result establishes a direct link between GHWs and the existence of invertible substructures in binary generator matrices.
  • We provide a systematic procedure for identifying an invertible k × k submatrix from the generator matrix associated with the constructed basis. The procedure is derived from the algebraic properties of the selected codewords and does not rely on probabilistic or numerical arguments.
  • Based on the above structural results, we formalize an information dispersal and reconstruction scheme defined entirely over the binary field. The correctness of the reconstruction follows from the algebraic properties of the constructed basis and the imposed GHWs conditions.
This work highlights new ideas on GHWs that depart from the traditional perspective of error or security analysis. Instead, they are used as structural invariants that provide information about the possible supports of subcodes and, consequently, about the existence of codewords with controlled support properties. These features will be exploited in Section 2 to guide the selection of a suitable basis for the code.

2. Preliminaries

This section establishes the fundamentals of linear coding theory, specifically generator matrices, parity-check matrices, and Generalized Hamming Weights (GHWs), adopting the standard definitions and notation found in [12,15,16,18]. Furthermore, we detail the use of polynomial rings and monomial orders as discussed in [11]. In this context, these algebraic concepts serve as a structural tool to impose a strict hierarchy on codeword coordinates, facilitating the algebraic analysis of supports and GHWs. Finally, we formally define the Information Dispersal Algorithm (IDA), based on the framework presented in [1].

2.1. Linear Codes and Information Dispersal Algorithm

Let q = p r be a positive integer, where p is a prime number and r is a positive integer. The finite field with q elements is denoted by F q . Let n , k , d be positive integers such that n k and d 1 . An [ n , k , d ] linear codeC is defined as a subspace of the vector space F q n .
A matrix G Mat k × n ( F q ) is called a generator matrix of C if its rows form a basis of C. Similarly, a matrix H Mat ( n k ) × n ( F q ) is said to be a parity-check matrix of C if its null space coincides with C, that is, x H T = 0 for all x C , where 0 denotes the zero vector in F q n k . The code C can be generated by multiplying any vector x F q k by the generator matrix G, that is, by computing the product x G .
Given two nonzero vectors x = ( x 1 , , x n ) and y = ( y 1 , , y n ) in F q n , the support of x is defined as supp ( x ) = i x i 0 . The Hamming distance between x and y is given by
d ( x , y ) = i x i y i .
The Hamming weight of a vector x is defined as w ( x ) = d ( x , 0 ¯ ) , where 0 ¯ denotes the zero vector. The minimum distance of a linear code C is then defined as
d = min w ( x ) x C , x 0 ¯ .
When the minimum distance of a linear code C is known, the code is described as an [ n , k , d ] linear code. The elements of C are called codewords, and the parameters n, k, and d are referred to as the basic parameters of the code, representing:
1.
n, the length of the codewords;
2.
k, the dimension of C as a vector space over F q ;
3.
d, the minimum distance of C.
The support of a subset  D C is defined as
supp ( D ) = i c D such that c i 0 .
In the case of a binary linear code, the support of a subset D C coincides with the set of coordinate positions for which at least one element of D has a nonzero entry.
Definition 2.1
([5]). The h-thgeneralized Hamming weightof a linear code C is defined as
d h ( C ) = min supp ( E ) E D h ( C ) ,
where D h ( C ) denotes the collection of all h–dimensional linear subspaces of C, for h 1 , , k .
It is worth noting that when h = 1 , the generalized Hamming weight coincides with the minimum distance of the code, that is, d 1 ( C ) = d . Hence, Definition 2.1 generalizes the classical notion of minimum distance.
The following theorem establishes fundamental properties of the generalized Hamming weights.
Theorem 2.1
([18]). Let C be an [ n , k , d ] linear code. Then:
1.
1 d 1 ( C ) < d 2 ( C ) < < d k ( C ) n ;
2.
(Generalized Singleton Bound) d h ( C ) n k + h .
The following classical results from linear algebra and commutative algebra can be found in standard references (see, for instance, [20]).
Corollary 2.1
([20]). Let V be a finite–dimensional vector space of dimension n. Any set of n linearly independent vectors in V forms a basis of V.
Theorem 2.2
([15]). Let A Mat n × n ( K ) . Then A is invertible if and only if det ( A ) 0 .
Some fundamental results on polynomial rings are presented in [5]. A degree–compatible monomial order is introduced in [11].
Definition 2.2
([11]). Let X be a collection of n vector variables X 1 , , X n , where each X i decomposes into q 1 components x i , 1 , , x i , q 1 for i [ n ] . A monomial in X is an expression of the form
X u = X 1 u 1 X n u n = i = 1 n j = 1 q 1 x i , j u i , j ,
where u Z 0 n ( q 1 ) .
Definition 2.3.
The polynomial ring K [ X ] is defined as the set of all polynomials in the variables X with coefficients in the field K .
The total degree of a monomial X u is given by deg ( X u ) = i = 1 n j = 1 q 1 u i , j . In particular, when u = ( 0 , , 0 ) , the monomial reduces to X u = 1 .
Let X be a set. A partial order on X is a binary relation ≤ that is reflexive, antisymmetric, and transitive. A partial order is said to be a total order if, for any a , b X , either a b or b a hold.
Let K be a field and let S = K [ x 1 , , x n ] . Denote by Mon ( S ) the set of all monomials in S.
Definition 2.4.
Amonomial orderon S is a total order ≤ on Mon ( S ) satisfying:
1.
1 u for all u Mon ( S ) ;
2.
if u < v and w Mon ( S ) , then u w < v w .
Definition 2.5
([11]). A monomial order ≺ on K [ x 1 , , x n ] is said to bedegree compatibleif
deg ( X a ) < deg ( X b ) X a X b
for all monomials X a , X b .
An example of a degree–compatible monomial order on F 2 [ x 1 , , x n ] is the graded lexicographic order. It is defined by X a < deglex X b if and only if:
1.
i = 1 n a i < i = 1 n b i ; or
2.
i = 1 n a i = i = 1 n b i and a j < b j for j = min i a i b i .
The following definitions, adapted from [1], formally describe the structure of an information dispersal algorithm.
Definition 2.6
([1]). Let P = P 1 , P 2 , , P n be a set of n participants, and let t be a positive integer called thethreshold. Aninformation dispersal algorithm(IDA) consists of a pair of algorithms ( IDA , IRA ) defined as follows:
IDA ( M , P , t ) ( X 1 , X 2 , , X n ) , where:
  • M denotes the plaintext message;
  • P is the set of participants;
  • t is the reconstruction threshold.
The algorithm produces a collection of shares ( X 1 , X 2 , , X n ) , where each share X i is distributed to the participant P i .
IRA ( { X i : P i A } , A ) :
  • the input consists of a subset A P and the corresponding shares;
  • if | A | t , the algorithm reconstructs M;
  • otherwise, it produces an error symbol (e.g.,ERROR).
The scheme satisfies the following correctness condition:
  • (Decoding) For every subset A P with | A | t ,
    IRA ( { X i : P i A } , A ) = M .
In classical IDA, correctness is ensured by the existence of invertible submatrices. In contrast, our binary variant relies on GHWs and degree compatible monomial orders.

3. An Algebraic Framework for Code-Based IDA/IRA Schemes

Let n be a positive integer and consider the notation [ n ] = { 1 , , n } . Let C be an [ n , k ] binary linear code. We define the set M of codewords of C that form generating sets such that the cardinality of their support equals d i ( C ) , as follows:
M = m C m i 1 , , m i j C d i ( C ) = s u p p m , m i 1 , , m i j ,
where i { 1 , , k } , i j { 1 , , k 1 } , and i = i j if i 1 and i k .
Observe that for i = 1 , we have m C d 1 ( C ) = s u p p ( m ) M , and for i = k , m C m 1 , , m k 1 C d k ( C ) = s u p p m , m 1 , , m k 1 M .
Let ≺ be a graded monomial order compatible with the total degree in F 2 [ x 1 , , x n ] . We define the minimal elements in M as follows:
1.
m 1 = min ( M ) ,
2.
m i = min m M | d i ( C ) = s u p p m , m i 1 , , m i j , where i { 2 , , k } , i j { 1 , , k 1 } , and i = i j if i 1 and i k .
Proposition 3.1.
Let C be an [ n , k ] -binary linear code. Then the set O = { m 1 , , m k } is a basis of C.
Proof. 
We prove this by induction on the index i [ k ] .
For i = 1 , the codeword m 1 satisfies d 1 ( C ) = s u p p ( m 1 ) . Clearly, m 1 0 , and since α m 1 = 0 implies α = 0 , the vector m 1 is linearly independent.
For illustrative purposes, consider the case i = 2 . The codewords m 1 and m 2 satisfy d 2 ( C ) = s u p p ( m 1 , m 2 ) . Moreover, m 1 m 2 because d 1 < d 2 , so there is at least one position j where m 2 has a 1 and m 1 has a 0. Let m 1 = ( b 1 , , b j 1 , 0 , b j + 1 , , b n ) and m 2 = ( b 1 , , b j 1 , 1 , b j + 1 , , b n ) , where b i , b i { 0 , 1 } . Then,
α 1 m 1 + α 2 m 2 = α 1 ( b 1 , , b j 1 , 0 , b j + 1 , , b n ) + α 2 ( b 1 , , b j 1 , 1 , b j + 1 , , b n ) = ( α 1 b 1 + α 2 b 1 , , α 1 b j 1 + α 2 b j 1 , α 2 , α 1 b j + 1 + α 2 b j + 1 , , α 1 b n + α 2 b n ) = ( 0 , , 0 ) .
From this, α 2 = 0 . Thus,
α 1 m 1 + α 2 m 2 = α 1 ( b 1 , , b j 1 , 0 , b j + 1 , , b n ) = α 1 m 1 ,
and since { m 1 } is linearly independent, α 1 = 0 . Therefore, α 1 = α 2 = 0 , so { m 1 , m 2 } is linearly independent.
Assume that { m 1 , , m k 1 } is linearly independent in C. Now consider { m 1 , , m k } and suppose i = 1 k α i m i = 0 . Note that m i m j for i j since d k ( C ) = s u p p ( m 1 , , m k ) and d 1 ( C ) < d 2 ( C ) < < d k ( C ) . This implies that m k has at least one 1 in positions r 1 , , r j where the others have 0s for some j [ t ] .
Fix position r 1 . The equation becomes
i = 1 k α i m i = i = 1 n α i b i , , α r 1 b r 1 , , i = 1 n α i b i = ( 0 , , 0 ) ,
where b i { 0 , 1 } . From the r 1 -th coordinate, α r 1 b r 1 = 0 and since b r 1 = 1 , α r 1 = 0 . Thus, the system reduces to i = 1 k 1 α i m i = 0 . By the induction hypothesis, { m 1 , , m k 1 } is linearly independent, so α i = 0 for i [ k 1 ] . Hence, α 1 = = α k = 0 , and O = { m 1 , , m k } is linearly independent.
By Corollary 2.1, O C is a basis of C.    □
We restate Proposition 3.1 as the following algorithm:
Algorithm 3.1.
Algorithm for computing a basis of a binary linear code.  
Input: An [ n , k ] -binary linear code C and a degree compatible monomial order in the polynomial ring F 2 [ x 1 , , x n ] .  
Output: A basis of C. Operations: A basis of C is obtained as follows:
1.
Compute the codewords in the subset M of C defined by
M = m C m i 1 , , m i j C d i ( C ) = s u p p m , m i 1 , , m i j ,
where i { 1 , , k } , i j { 1 , , k 1 } , and i = i j whenever i 1 and i k .
2.
Order the elements of M according to the order ≺ for obtain the sequence of codewords { m 1 , , m k } .
3.
Output: The ordered set of codewords O = { m 1 , , m k } .
Algorithm 3.2.
Algorithm for IDA using a basis of a binary linear code.
Input: An [ n , k ] -binary linear code C satisfying d 1 ( C ) = 1 , d k ( C ) = n , and a plaintext message M.
Output: A dispersal vector and an n × k matrix and the threshold t.
Operations: The dispersal vector and the matrix A of size n × k are obtained as follows:
1.
Apply Algorithm 3.1 to the [ n , k ] -binary linear code C to obtain a basis O = { m i } i = 1 k of C.
2.
Convert the plaintext M into its binary representation, yielding a vector (or a set of vectors) F F 2 k .
3.
Define A = [ m 1 , m 2 , , m k ] . Multiply F (on the right) by A to obtain the dispersal vector d, i.e., A · F = d .
4.
Define t = n k .
5.
Output: The dispersal vector d, the matrix A of size n × k and the threshold t.
Algorithm 3.3.
Algorithm for the recovery of information from a given vector and an associated dispersion matrix, referred to as the Information Recovery Algorithm (IRA).
Input: A vector b of length k and a matrix A M a t n × k ( F 2 ) .
Output: The dispersion vector d.
Operations: The dispersion vector d is obtained from b and A as follows.
1.
Consider A = [ m 1 , m 2 , , m k ] and denote by r i , i [ n ] , the rows of A. If s u p p ( r i ) = s u p p ( r j ) , remove the row r j . Repeat this step until all remaining rows satisfy r i r j , or until n k deletions are performed. If A contains no repeated rows, proceed to the next step.
2.
For each i [ n ] , if a row r i = 0 in A, remove that row.
3.
Suppose steps 1 and 2 result in t and s row deletions respectively, with p , s < n k . Identify rows r i of A such that | s u p p ( r i ) | | s u p p ( r j ) | for all j [ n ( p + s ) ] , and remove such rows. Repeat until a total of q additional deletions is performed so that p + s + q = n k .
4.
After deleting n k rows, rename the resulting column vectors of A as C 1 , C 2 , , C k , and define the matrix B = [ C 1 , C 2 , , C k ] .
5.
Solve the linear system B d = b for d, i.e., compute d = B 1 b .
6.
Output: The dispersion vector d.
Proposition 3.2.
Let C be an [ n , k ] -binary linear code such that d 1 ( C ) = 1 and d k ( C ) = n . Then Algorithms 3.2 and 3.3 are valid.
Proof. 
Consider a basis of C given by Algorithm 3.1, namely O = { m i } i = 1 k . Since d 1 = 1 , the codeword m 1 has exactly one nonzero coordinate in the i-th position; that is, m 1 = ( 0 , , 0 , 1 i , 0 , , 0 ) . Note that the i-th row in matrix A is not repeated because m 1 has a single 1 in position i. Therefore, this row is not omitted. Thus, the vector C 1 of matrix B preserves the unique 1 of m 1 while removing n k zero entries, so C 1 = ( 0 , , 0 , 1 i , 0 , , 0 ) .
Furthermore, Algorithm 3.3 omits zero rows and repeated rows in matrix A, so the rows of B satisfy r i r j if i j and r i 0 for i , j [ k ] .
Now, suppose i = 1 k α i r i = 0 , which gives the system of equations:
α 1 b 11 + α 2 b 12 + + α k b 1 k = 0 α 1 b 21 + α 2 b 22 + + α k b 2 k = 0 α 1 b k 1 + α 2 b k 2 + + α k b k k = 0 ( * )
This can be rewritten as α 1 C 1 + α 2 C 2 + + α k C k = 0 . Note that C 1 = ( 0 , , 1 i , , 0 ) . Applying Gaussian elimination to the system,
0 b 12 b 1 k 0 1 b i 2 b i k 0 0 b k 2 b k k 0 1 b i 2 b i k 0 0 b 22 b 2 k 0 0 b k 2 b k k 0
Since m i 0 and s u p p ( C i ) s u p p ( m i ) , some b i j in C i are nonzero. Step 2 of Algorithm 3.3 omits zero rows, and step 3 removes rows with larger supports, leaving at least the row in B where b i j = 1 in C i and b i j = 0 in C i 1 .
For each column C i , apply row swaps to place a 1 in position b i j , then eliminate below it. Repeating this k times yields:
1 b i 2 b i k 0 0 1 b 2 j 0 0 0 1 0
Thus, system ( * ) is equivalent to
α 1 + α 2 b 12 + + α k b 1 k = 0 α 2 + + α k b 2 j = 0 α k = 0
From the last equation, α k = 0 . From the penultimate, α k 1 + α k b k 1 , k = 0 , so α k 1 = 0 . Back substituting yields α k = α k 1 = = α 1 = 0 . Thus, the rows r i of B are linearly independent over F 2 k .
By Corollary 2.1, B is a basis of F 2 k . Since B has full rank k, it is invertible.    □

4. Analysis and Results

This section provides a concrete example to substantiate the algebraic framework established in Section 3. The example illustrates how basis selection, governed by GHWs, results in a generator matrix suitable for the Information Dispersal Algorithm (IDA). The primary objective is to confirm the theoretical consistency of our approach and elucidate the operational procedures, rather than to evaluate empirical performance. The GHWs are calculated following the method of Johnsen and Verdure [6]. For reference, the pseudocode for determining the GHWs of an [ n , k ] -binary linear code is included in Appendix Section 1.1.
Example 4.1.
Let C be a [ 10 , 7 ] -binary linear code with generator matrix
G = 1 1 0 0 0 0 1 1 0 1 1 1 0 1 0 0 1 0 0 1 1 1 0 0 1 1 1 1 0 0 0 0 1 1 0 1 0 1 0 0 0 1 0 0 0 1 1 0 1 1 0 1 0 1 0 0 1 1 1 1 0 0 1 0 0 1 1 0 1 1 .
Using the algorithm described in Appendix 1, the generalized Hamming weights of C are computed as
d 1 ( C ) = 1 , d 2 ( C ) = 2 , d 3 ( C ) = 3 , d 4 ( C ) = 5 , d 5 ( C ) = 7 , d 6 ( C ) = 9 , d 7 ( C ) = 10 .
For each i [ k ] , we determine some of the elements of the set M by selecting the keywords that satisfy the defining condition of this set, namely:
For i = 1 , ( 0000010000 ) , ( 0010000000 ) , ( 0100000000 ) M For i = 2 , ( 0000010000 ) , ( 0010000000 ) , ( 0100000000 ) M For i = 3 , ( 0000010000 ) , ( 0010000000 ) , ( 0100000000 ) , ( 0001000100 ) , ( 0010010000 ) , ( 0000100001 ) , ( 0110000000 ) , ( 0100010000 ) , M For i = 4 , ( 0010010000 ) , ( 0000100001 ) , ( 0001000100 ) , ( 0110000000 ) , ( 0100010000 ) , ( 0000001011 ) , M For i = 5 , ( 0001000100 ) , ( 0000001011 ) , ( 0000110001 ) , M For i = 6 , ( 0000001011 ) , ( 0000110001 ) , ( 0010100001 ) , M For i = 7 , ( 1000000110 ) , ( 0000101010 ) , ( 1110000110 ) , M .
We consider the lexicographic graded order ≺, and using SageMath we calculate the codewords m i C , which are:
O = m 1 = ( 0000010000 ) , m 2 = ( 0010000000 ) , m 3 = ( 0100000000 ) , m 4 = ( 0000100001 ) m 5 = ( 0001000100 ) , m 6 = ( 0000001011 ) , m 7 = ( 1000000110 )
Furthermore, by the definition of the set M we see that:
d 1 ( C ) = | s u p p ( m 1 ) | = 1 d 2 ( C ) = | s u p p ( m 1 , m 2 ) | = 3 d 7 ( C ) = | s u p p ( m 1 , m 2 , m 3 , m 4 , m 5 , m 6 , m 7 ) | = 10
By proposition 3.1, we see that the set O forms a basis of the code C.
We then apply the IDA mechanism proposed by Rabin [14] to the set O as follows:
Let M = “Quantum cryptography is secure" is the plain-text message. We convert the message M into ASCII code, and subsequently into its binary representation,
M = 01010001 01110101 01100001 01101110 01110100 01110101 01101101 00100000 01100011 01110010 01111001 01110000 01110100 01101111 01100111 01110010 01100001 01110000 01101000 01111001 00100000 01101001 01110011 00100000 01110011 01100101 01100011 01110101 01110010 01100101
Let P = P 1 , P 2 , P 3 , P 4 , P 5 , P 6 , P 7 , P 8 , P 9 , P 10 be a set of 10 participants and let the threshold be defined as t = n k = 3 . We consider the matrix formed by the codewords m i ,
A = m 1 , m 2 , m 3 , m 4 , m 5 , m 6 , m 7 = 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 1 0 0 0 0 1 0 1 0 0 0 0 0 1 0 0 0 0 1 0 1 1
The message M is partitioned into blocks of length 7, and each block is multiplied by A. For illustration, only the first three resulting vectors are shown:
D = A M = ( d 1 , d 2 , d 3 , d 4 , d 5 , ) = 0 1 0 0 1 0 1 0 1 1 1 0 1 0 1 0 1 1 1 0 1 1 1 0 0 0 1 0 0 1 0 1 0 1 0 0 0 1 1 1 0 0 0 1 1 1 0 1 1 0 10 × | M | / 7
where · denotes the ceiling function, and if the length of the final block is less than 7, it is padded with zeros.
Each component of the resulting vectors is distributed among the participants. We assume that participants P 7 , P 8 , and P 10 are unavailable. Let P represent the set of participants used for reconstruction:
P = { P 1 , P 2 , P 3 , P 4 , P 5 , P 6 , P 9 } , satisfying | P | = 7 t .
Applying the Information Recovery Algorithm (IRA), the components corresponding to the unavailable participants are removed from each dispersal vector, yielding the vectors f i . To recover the original message, we construct a submatrix B of the dispersal matrix A by deleting the rows corresponding to the unavailable participants.
By Proposition 3.2, the resulting matrix B F 2 7 × 7 is invertible. Its inverse, computed usingSageMath, is explicitly given by
B 1 = 0 0 0 0 0 1 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 1 0 0 0 0 0 0 .
Consequently, the dispersal vectors d i are recovered by solving the linear system
B M = f i , then , M = B 1 f i , so ,
D = 0 0 0 0 0 1 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 1 1 0 0 0 0 0 0 0 1 0 0 1 0 1 0 1 1 1 0 1 0 1 0 1 1 1 0 1 1 1 0 0 0 1 0 0 1 0 0 0 1 1 = 0 1 0 0 1 0 1 0 1 1 1 0 1 0 1 0 1 1 1 0 1 1 1 0 0 0 1 0 0 1 0 1 0 1 0 0 0 1 1 1 0 0 0 1 1 1 0 1 1 0
where i indexes the set of message blocks. This result shows that the original message can be retrieved even if up to t = n k components are lost in each dispersal vector.

5. Conclusions

The contributions of this work are fundamentally theoretical. We have demonstrated that Generalized Hamming Weights (GHWs) serve as effective algebraic tools for constructing bases of binary linear codes suitable for information dispersal. These results establish sufficient algebraic conditions to guaranty the existence of generator matrices with invertible submatrices over F 2 , thereby enabling the application of the Information Dispersal Algorithm (IDA) and the Information Recovery Algorithm (IRA) within a strictly binary framework
This study suggests several avenues for future research, particularly regarding performance and generalized applicability.
Question 5.1.
How does the efficiency of the proposed binary IDA benchmark against classical implementations over finite fields like G F ( 2 8 ) and G F ( 2 16 ) , as discussed in [17]?
Question 5.2.
In which parameter regimes ( n , k ) does the binary IDA/IRA framework offer superior performance or distinct trade-offs compared to traditional approaches over larger fields?
Furthermore, relaxing the Generalized Hamming Weight conditions presents a compelling direction. It is natural to inquire whether the strict constraints used here can be loosened to accommodate a wider class of binary linear codes.
Question 5.3.
What are the admissible ranges for d 1 ( C ) and d k ( C ) (with d 1 ( C ) > 1 and d k ( C ) n ) that still permit valid IDA-based dispersal and recovery?
Addressing these questions would deepen the understanding of the interplay between GHWs, algebraic structures, and information dispersal, potentially paving the way for new families of binary codes suitable for distributed storage and secure data dissemination

Author Contributions

Conceptualization, O.C.-M. and L.P.-L.; methodology, R.M.-J. and O.C.-M.; software, O.C.-M. and L.P.-L; validation, O.C.-M., R.M.-J., R. V.-M. and L.P.-L.; formal analysis, O.C.-M. and L.P.-L.; investigation, O.C.-M. and L.P.-L.; resources, all authors; writing—original draft preparation, O.C.-M., R.V.-M. and L.P.-L.; writing—review and editing, O.C.-M., R.M.-J., R.V.-M., and L.P.-L.; visualization, O.C.-M., R.M.-J, R.V.-M., and L.P.-L.; supervision, L.P.-L.; project administration, L.P.-L.; funding acquisition, R.V.-M., and L.P.-L. All authors have read and agreed to the published version of the manuscript.

Funding

Instituto Politécnico Nacional, [Grant numbers SIP-20250150 (R. Vázquez-Medina)

Acknowledgments

The authors acknowledge the Área de Redes y Telecomunicaciones of the Universidad Autónoma Metropolitana and the Secretaría de Ciencia, Humanidades, Tecnología e Innovación, Mexico, for their support under Grant No. CVU-1082416 (O. Casimiro-Muñoz).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A Algorithm for Computing the Generalized Hamming Weights of an [n,k]-Binary Linear Code

In this appendix, we present an algorithm for computing the generalized Hamming weights (GHWs) of an [ n , k ] –binary linear code. The procedure is based on the algebraic-combinatorial approach introduced by Johnsen and Verdure [6], which relates the GHWs of a linear code to homological invariants of a monomial ideal associated with a matroid. The algorithm is implemented using SageMath and exploits the correspondence between the parity-check matrix of the code, the associated binary matroid, and the graded Betti numbers arising from a minimal free resolution of the corresponding circuit ideal.
Algorithm A1:Computation of the GHWs of a binary linear code
Require:
Integers n , k
Ensure:
Generalized Hamming weights of an [ n , k ] binary linear code
1:
Input the parameters n and k
2:
Construct a binary linear code C F 2 n with parameters [ n , k ]
3:
Compute a generator matrix G and a parity-check matrix H of C
4:
Construct the binary matroid M represented by the matrix H
5:
Determine the set of circuits C of the matroid M
6:
Define the polynomial ring R = F 2 [ x 1 , , x n ]
7:
Construct the circuit ideal I C R generated by the monomials corresponding to the circuits of M
8:
Compute a minimal graded free resolution of I C
9:
Extract the graded Betti numbers and store them in a matrix M C
10:
Initialize an empty list G H W
11:
for each homological degree i in M C  do
12:
   for each internal degree j in M C  do
13:
     if  M C i , j 0  then
14:
        Append i + j to G H W
15:
        break
16:
     end if
17:
   end for
18:
end for
return G H W

References

  1. Auleriano, I.; Chen, A.; D’olivera, R. Optimal Computational Secret Sharing. IEEE International Symposium on Information Theory (ISIT) 2025, 1–10. [Google Scholar]
  2. Beimel, A. Secret-Sharing Schemes for General Access Structures: An Introduction, 1st ed.; University of the Negev, Israel, 2025; pp. 10–97. [Google Scholar]
  3. Dertil, R.; Eren, S. Secret Sharing Schemes Based on Linear Codes over F2RS. Asian J. Math Appl. 2025, 9. [Google Scholar]
  4. Deyanira, M. Evaluación de rendimiento del Algoritmo de Dispersión de Información sobre los campos finitos GF(28) y GF(216). Master of Science in Information Technology, Autonomous Metropolitan University, CDMX, 2019. [Google Scholar]
  5. García-Marco, I.; Márquez-Corbella, I.; Martínez-Moro, E.; Pitones, Y. Free Resolutions and Generalized Hamming Weights of Binary Linear Codes. Mathematics 2022, 10, 2079. [Google Scholar] [CrossRef]
  6. Johnsen, T.; Verdure, H. Hamming weights and Betti numbers of Stanley-Reisner rings associated to matroids. Appl. Algebra Eng. Commun. Comput. 2013, 24, 73–93. [Google Scholar] [CrossRef]
  7. Johnsen, T.; Verdure, H. Stanley-Reisner resolution of constant weight linear codes. Designs, Codes and Cryptography 2012, 72. [Google Scholar] [CrossRef]
  8. Ling, M.; Xing, C. Coding Theory: A First Course, 1st ed.; Cambridge University Press, USA, 2004; pp. 39–66. [Google Scholar]
  9. Li, Mingqiang. On the Confidentiality of Information Dispersal Algorithms and Their Erasure Codes. ArXiv 2012, abs/1206.4123. [Google Scholar]
  10. Lin, S. J.; Chung, W. H. An Efficient (n,k) Information Dispersal Algorithm Based on Fermat Number Transforms. IEEE 2013, 8, 1371–1383. [Google Scholar]
  11. Márquez-Corbella, I.; Martínez-Moro, E.; Suárez-Canedo, E. On the ideal associated to a linear code. Adv. Math. Commun. 2016, 10, 229–254. [Google Scholar] [CrossRef]
  12. Pless, V.; Huffman, W. C. Fundamentals of Error-Correcting Codes, 1st ed.; New York, USA, 2003; pp. 1–52. [Google Scholar]
  13. Qian, Q.; Yu, ZT.; Zhang, R.; Hung, CH. A multi-layer information dispersal based encryption algorithm andits application for access control. Sustainable Computing: Informatics and Systems 2018, 20, 76–87. [Google Scholar]
  14. Rabin, M. O. Efficient dispersal of information for security, load balancing, and fault tolerance. J. ACM 1989, 36, 335–348. [Google Scholar] [CrossRef]
  15. Ron, M. Introduction to Coding Theory, 2nd ed.; Cambridge University Press: UK, 2006; pp. 26–47. [Google Scholar]
  16. Van Lint, J.; Van der Geer, G. Introduction to Coding Theory and Algebraic Geometry, 1st ed.; Boston, Berlin, USA, Germany, 1998; pp. 11–28. [Google Scholar]
  17. Velázquez, B.; Marcelín, R. Estudio sobre el desempeño del Algoritmo de Dispersión de Información. CIAI 2018. [Google Scholar]
  18. Wei, V. K. Generalized Hamming weights for linear codes. IEEE Trans. Inf. Theory 1991, 37, 1412–1418. [Google Scholar] [CrossRef]
  19. Wijayanto, A.; Harjito, B. Reduce Rounding Off Errors in Information Dispersal Algorithm. IC3INA 2019, 36–40. [Google Scholar]
  20. Zaldivar, F. Introducción al álgebra lineal, 1st ed.; Universidad Nacional Autónoma de México: México, 2019; pp. 1–48. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated