[IEEE 2006 IEEE 12th Digital Signal Processing Workshop & 4th IEEE Signal Processing Education...

5

Click here to load reader

Transcript of [IEEE 2006 IEEE 12th Digital Signal Processing Workshop & 4th IEEE Signal Processing Education...

Page 1: [IEEE 2006 IEEE 12th Digital Signal Processing Workshop & 4th IEEE Signal Processing Education Workshop - Teton National Park, WY, USA (2006.9.24-2006.9.24)] 2006 IEEE 12th Digital

ERROR DETECTION WITH REAL-NUMBER CODES BASED ON RANDOM MATRICES

Vieira, Jose M. N., Santos, Dorabella M. S., Ferreira, Paulo J. S. G.

Departamento de Electronica, Telecomunicacoes e Informatica / IEETAUniversidade de Aveiro

Aveiro Portugal

ABSTRACT

Some well-known real-number codes are DFT codes. Sincethese codes are cyclic, they can be used to correct erasures(errors at known positions) and detect errors, using the loca-tor polynomial via the syndrome, with efficient algorithms.The stability of such codes are, however, very poor for bursterror patterns. In such conditions, the stability of the sys-tem of equations to be solved is very poor. This amplifies therounding errors inherent to the real number field. In order toimprove the stability of real-number error-correcting codes,other types of coding matrices were considered, namely ran-dom orthogonal matrices. These type of codes have provento be very stable, when compared to DFT codes. However,the problem of detecting errors (when the positions of theseerrors are not known) with random codes was not addressed.Such codes do not possess any specific structure which couldbe exploited to create an efficient algorithm. In this paper, wepresent an efficient method to locate errors with codes basedon random orthogonal matrices.

Index Terms— Error correction, random matrices, realnumber codes, sparse solutions

1. INTRODUCTION

Conventional error correcting codes are defined over finitefields. However, real-number codes can also be employedwhen floating point issues become important. A finite fieldcan be embedded in a real or complex field and several au-thors studied the properties of real-number codes [1, 2, 3].

Some such codes include the DFT and DCT codes. Theseparticular codes are cyclic, and, hence, can be used to detectand correct errors using the so-called locator polynomial viathe syndrome, with efficient algorithms. Unfortunately, dueto their structure, the stability of these codes is very poor forburst error patterns. Shannon showed in his seminal paper[4] that most of the codes can achieve channel capacity. Thestructure imposed on the codes, is only intended for practicalalgorithms to code and decode.

Surprisingly enough, if the matrix involved does not haveany particular structure, the code obtained performs quite wellindependently of the error pattern. These codes which are

based on random matrices are dubbed random codes. Theyhave been shown to have a good stability in correcting era-sures (errors in known positions) [5]. However, the problemof correcting errors (detecting and correcting errors) was notaddressed by the authors. This problem presents itself as acombinatorial problem, but that can be tackled as a linear pro-gramming problem, under certain conditions.

2. CODES AND STRUCTURE

In the particular case of DFT codes, to circumvent the poorstability of the code when facing burst erasure patterns, a par-allel concatenated code (PCC) is used, which consists of atwo-channel structure, where the first channel employs a DFTcode and the second channel employs additionally an inter-leaver (Figure 1).

Fig. 1. Two channel coder with K < N . The interleaver isdenoted by P .

These two-channel codes outperform their single-channelcounterparts, specially for occurrences of bursty losses. Theircondition number can be several orders of magnitude lowerthan the single channel DFT codes, which makes the two-channel structure very stable, even when bursts of losses oc-cur. The improvement in stability has to do with the random-ness introduced by the interleaver [6, 7], although it has notyet been possible to demonstrate this mathematically.

5261-4244-0535-1/06/$20.00/©2006 IEEE

Page 2: [IEEE 2006 IEEE 12th Digital Signal Processing Workshop & 4th IEEE Signal Processing Education Workshop - Teton National Park, WY, USA (2006.9.24-2006.9.24)] 2006 IEEE 12th Digital

3. RANDOM CODES

However, stable single channel real-number codes even underbursty erasure patterns have been proposed [5]. These codesare based on random gaussian matrices and, hence, have nospecific structure.

In this case, the original signal m ∈ RK is coded as

c = Gm (1)

with c ∈ RN where G is an N × K submatrix of an N × N

gaussian matrix (K < N ). Defining the set of the index ofthe lost samples (erasures) by J ⊂ {1, . . . , N} with at mostN −K samples, and J as its complement, then we can defineG(J) as the matrix with the known rows of G. To get themessage m from the known samples of c it suffices to solvethe system of equations (1)with the known rows of G

m = G(J)+c(J),

where G(J)+ stands for the pseudo-inverse of G(J). It hasbeen shown in [5] that with high probability, such codes arevery stable even under bursty losses.

As a comparison between random codes and a structuredreal-number code, Figure 2 shows the condition number of aDFT code, its PCC counterpart and a random code, with N =256 (128 in each channel for the PCC) and K = 21, as thenumber of bursty losses increases. It can be seen that for therandom code, the condition number never exceeds 10, whilefor the DFT code the condition number explodes quickly. ThePCC, in turn, is much more stable than the DFT code (by sixorders of magnitude for many losses), but the random codeoutperforms the two-channel code.

The connection between randomness and stability is againhighlighted: the more structure, the worse the performance.A question arises, however: since it is not possible to takeadvantage of the structure of random matrices, can detectionof errors be carried out with these new codes?

4. ERROR CORRECTION WITH RANDOM CODES

In the previous section we presented a method to recover theerror amplitudes solving an overdetermined system of equa-tions minimizing the �2 norm. In this section we show how toconstruct a real number code in order to correct errors in thereceived vector, finding their positions and amplitudes. Con-sider the random matrix F ∈ R

N×N with elements followinga normal distribution and the following partition

F =[

GN×K

... HN×(N−K)

],

where G ∈ RN×N is the coding matrix and H ∈ R

N×(N−K)

the parity check matrix. The matrix F can be orthogonalized

80 100 120 140 160 180 20010

0

102

104

106

108

1010

1012

1014

1016

Condition number

Total number of consecutive erasures |J|

conditio

n n

um

ber

random codeDFT codePCC code

Fig. 2. Log plot of the condition number versus the totalnumber of consecutive missing samples for the DFT code(N = 256), the two-channel counterpart (N = 128 and |J |/2missing samples per channel) and a random code (N = 256),with K=21.

by a QR factorization using a Gram-Smith orthogonalizationand in this case we have

HT G = 0.

Consider a vector m ∈ RK and the coding operation

c = Gm,

that results in the codeword c ∈ RN where N > K . Consider

that the vector c is corrupted in L samples with index givenby the set J ⊂ {1, . . . , N} and we can write the corruptedvector y as

y = c + e,

with e a sparse error vector with L samples different fromzero at positions given by J . If we multiply the vector y byHT we have

s = HT y = HT (c + e) = HT Gm + HT e = HT e.

The vector s is known as the syndrome [8] and because thecolumns of H are orthogonal to the columns of G, the syn-drome is a function of only the error vector e. From the re-ceived signal y it is possible to calculate the syndrome s, andthe error vector e can be recovered by solving the underdeter-mined system of equations

HT e = s. (2)

527

Page 3: [IEEE 2006 IEEE 12th Digital Signal Processing Workshop & 4th IEEE Signal Processing Education Workshop - Teton National Park, WY, USA (2006.9.24-2006.9.24)] 2006 IEEE 12th Digital

This system has many solutions and we want the sparsest one.This is a combinatorial problem very hard to solve and can bestated as

(P0) min ‖e‖0 subject to HT e = s

with ‖e‖0 = # {i ∈ 1, . . . , N : ei �= 0} the �0-norm (pseudo-norm). This problem has a unique solution if the vector e issufficiently sparse. The folowing theorem states this property.

Theorem 1 If ‖e‖0 = L, then e is the unique solution of (P0)if and only if

L <K(HT ) + 1

2.

The function K(HT ) is the ambiguity index of HT , andis defined as the largest number such that any set of K(HT )columns of HT is linearly independent. This bound is wellknown in the coding theory as the Singleton bound [9]. IfH ∈ RN×(N−K), then K(HT ) ≤ N − K and we have L <(N − K + 1)/2. Note that to construct a code capable ofcorrecting t errors we have to add two new samples to themessage vector m or have N − K = 2t.

The system of equations of (2) can be expanded to

� � �

h1 h2 · · · hN

� � �

e1

e2

...eN

= s

and it is easy to see that the syndrome s can be written as alinear combination of the vectors hi

s =∑i∈J

hiei.

4.1. Obtaining sparse solutions with �1 norm

Donoho and Elad [10, 11], showed that in certain conditionsthe (P0) can be solved by the more tractable problem

(P1) min ‖e‖1 subject to HT e = s

with ‖e‖1 =∑

i |ei| the �1-norm. This is a linear program-ming problem and can be solved using the simplex or an in-terior point algorithm [12, 13, 14]. The following theoremstates in which conditions the two problems are equivalent[10, 11].

Theorem 2 If ‖e‖0 = L and

L <1 + 1/M(HT )

2= ebp,

then (P0) is equivalent to (P1) and solving the latter problemleads to the same sparse solution obtained by solving (P0).

The mutual incoherence M(HT ) = maxi�=j

∣∣hTi hj

∣∣ (as-sume HT normalized with ‖hi‖2 = 1) gives a measure of thedistribution of the frame vectors of HT . If M(HT ) is high itmeans that certain vectors of HT are almost collinear. Theebp stands for equivalence break point and gives the numberof errors that is guaranteed the code to correct. Note that ifthe L < ebp the equivalence between (P0) and (P1) is guaran-teed, but if L ≥ ebp the equivalence may be broken. It is wellknown [15] that if we assume K(HT ) = N −K (not withoutsignificance, since we are dealing with random matrices) thatthe following holds for the mutual incoherence

M(HT ) ≥√

K

(N − K)(N − 1)= Mopt.

M(HT ) measures how spread-out the columns of HT are anddue to the Schwartz inequality we have 0 ≤ M(HT ) ≤ 1.Note that for a square matrix A we have M(A) = 0 if A hasorthogonal columns.

4.2. Grassmannian Frames

A code with a parity check matrix H that attains an optimalmutual incoherence M(HT ), is expected to achieve good re-sults. Such matrices exist and are known as Grassmannianframes [16, 17]. This type of frames has applications in cod-ing with real numbers and they guarantee that the mutual in-coherence is maximum M(HT ) = Mopt and the vectors hi

are distributed in an optimal way in order to maximize the rel-ative angle defined by

∣∣hTi hj

∣∣ with i �= j and ‖hi‖2 = 1. Forthis type of frames the equivalence break point is optimized

ebpopt =1 + 1/Mopt

2,

and it would be expected that they perform much better thanthe random case. However, that is not true, and simulationresults showed that in most cases the random frames performbetter. Note that the mutual incoherence does not give allthe information about the frame subspace. We also have toconsider the angle between each vector hi and the subspacesformed by all the possible sets of two and more vectors hj .

Assuming normalized vectors hi, we can calculate theGram matrix of HT which has all the possible dot products∣∣hT

i hj

∣∣ with i �= j. The mutual incoherence gives only theworst case, and we can get more insight on the distributionof the angles if we evaluate the histogram of the off diagonalelements of the Gram matrix. In the figure 3 we can comparethe histograms for a random generated matrix and a Grass-mannian matrix. Notice the normal distribution of the frameangles for the random matrix and the maximum dot productof 0.25 for the Grassmannian matrix.

In order to validate the error correction with random ma-trices we have implemented error correction algorithms thatgenerates a syndrome for a positive valued error pattern and

528

Page 4: [IEEE 2006 IEEE 12th Digital Signal Processing Workshop & 4th IEEE Signal Processing Education Workshop - Teton National Park, WY, USA (2006.9.24-2006.9.24)] 2006 IEEE 12th Digital

gets the sparse solution by solving (P1) using a linear pro-gramming method of interior point. For each number of er-rors we tested 105 error patterns and counted the number cor-rectly recovered. In the figure 4 we can see that for this sim-ulation the random code performed slightly better than theGrassmannian code. We can also see that both methods cor-rect most of the error patterns well above the ebp threshold.

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 10

5000

10000

15000

|<hi,h

j>| Grassmanian

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 10

100

200

300

400

500

600

|<hi,h

j>| Random

Fig. 3. Histograms of all possible angles between any twoframe vectors. The top plot is for the Grassmannian matrixand the bottom one for the random one.

One of the problems in solving (P1) directly with a lin-ear programming algorithm, is that all the errors should bepositive. To circumvent this limitation, we define for any er-ror vector e the two variables e+ and e− as e+

i = max{ei, 0}and e−i = max{−ei, 0} from which e can be recovered as e =e+−e−. Solving the following problem with e = [e+ e−]T ,

min ‖e‖1 subject to[HT −HT

]e = s and e ≥ 0,

we can correct positive and negative error amplitudes.

5. CONCLUSIONS

A method for correcting errors in real-number codes was pro-posed. Instead of solving the hard combinatorial problem itwas shown that it is possible to obtain the same solution bysolving a linear programming problem. We also showed thatthe optimal Grassmannian matrices do not always performbetter than a simple random code. This observation makesclear that the mutual incoherence is not enough to character-ize the code capacity. In Figure 4 we can see that the ebp(◦) measured for the random code is much more conservativethan the ebp (�) for the Grassmannian code. We also pro-posed a different measure to characterize a code, but it is notpractical because of the combinatorics.

0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 150

10

20

30

40

50

60

70

80

90

100

Number of errors

%

N= 128 N−K= 16

GrassRandom

Fig. 4. Percentage of corrected error patterns for several num-ber of errors (horizontal axis). Unexpectedly, the randomcode outperformed the Grassmannian code. ◦ and � repre-sents the ebp of the random and Grassmannian codes respec-tively.

6. REFERENCES

[1] T. G. Marshall Jr., “Methods for error correction withdigital signal processors,” in Proceedings 25th MidwestSymposium on Circuits and Systems. IEEE, 1982, pp.1–5.

[2] T. G. Marshall Jr., “Coding of real-number sequencesfor error correction: A digital signal processing prob-lem,” IEEE Journal on Selected Areas of Communica-tion, vol. 2, no. 2, pp. 381–391, 1984.

[3] Gagan Rath and Christine Guillemot, “Subspace basederror and erasure correction with dft codes for wirelesschannels,” IEEE Transactions on Signal Processing,vol. 52, no. 11, pp. 3241– 3252, 2004.

[4] C. E. Shannon, “A mathematical theory of communi-cation,” Bell Systems Technical Journal, vol. 27, pp.379–423, 1948.

[5] Zizhong Chen and Jack Dongarra, “Numerically sta-ble real-number codes based on random matrices,” inITW2004, San Antonio, Texas, 2004, IEEE, pp. 24–29.

[6] Paulo J. S. G. Ferreira and Jose M. N. Vieira, “Stabledft codes and frames,” IEEE Signal Processing Letters,vol. 10, no. 2, pp. 50–53, 2003.

[7] Jose Vieira, “Stability analysis of non-recursive par-allel concatenated real number codes,” in 2nd Sig-nal Processing Education Workshop, Callaway Gardens,Pine Mountain, Georgia, USA, 2002, IEEE.

529

Page 5: [IEEE 2006 IEEE 12th Digital Signal Processing Workshop & 4th IEEE Signal Processing Education Workshop - Teton National Park, WY, USA (2006.9.24-2006.9.24)] 2006 IEEE 12th Digital

[8] Richard E. Blahut, Theory and Practice of Error ControlCodes, Addison-Wesley, NY, 1983.

[9] Richard E. Blahut, Algebraic Codes for Data Transmis-sion, Cambridge University Press, Cambridge, 2002.

[10] David Donoho and Xiaoming Huo, “Uncertainty prin-ciples and ideal atomic decomposition,” IEEE Transac-tions on Information Theory, vol. 47, no. 7, pp. 2845–2862, 2001.

[11] Michael Elad and Alfred M. Bruckstein, “A generalizeduncertainty principle and sparse representation in pairsof bases,” IEEE Transactions on Information Theory,vol. 48, no. 9, pp. 2558 –2567, 2002.

[12] Margaret H. Wright, “The interior-point revolution inconstrained optimization,” Tech. Rep. 98, Bell Labora-tories.

[13] Stephen Boyd and Lieven Vandenberghe, Convex Opti-mization, Cambridge University Press, Cambridge, UK,2004.

[14] Edwin K. P. Chong and Stanislaw H. Zak, An Introduc-tion to Optimization, John Wiley & Sons. Inc., USA,1996.

[15] L. R. Welch, “Lower bounds on the maximum cross cor-relation of signals,” IEEE Transactions on InformationTheory, vol. 20, no. 3, pp. 397–399, 1974.

[16] Thomas Strohmer and Robert W. Heath Jr., “Grass-mannian frames with applications to coding and com-munication,” Applied and Computational HarmonicAnalysis, vol. 14, no. 3, pp. 257–275, 2003.

[17] Hardin Ronald H. Conway, John H. and Neil J. A.Sloane, “Packing lines, planes, etc.: Packings in grass-mannian spaces,” Experimental Mathmatics, vol. 5, no.2, pp. 139–159, 1996.

530