CTT

871

description

coding theory

Transcript of CTT

Page 1: CTT
Page 2: CTT

Course Outline:

Lectures 1-15

Basics of Block Codes

(2)

Mathematical Preliminaries:

Linear Codes

(3)

Bounds on Code Mathematical Preliminaries:

groups, rings and fields

(3)

Mathematical Preliminaries:

vector spaces

(3)

Bounds on Code

Size

(2)

Standard Array

Decoding

(2)

Page 3: CTT

Convolutional

Codes

(5)

Generalized

Finite Fields

(4)

Course Outline:

Lectures 16-42

Generalized

Distributive Law

(8)

Low-Density Parity-

Check Codes

(7)

Cyclic Codes (3)

The End!The End!The End!The End!

Page 4: CTT
Page 5: CTT
Page 6: CTT
Page 7: CTT
Page 8: CTT
Page 9: CTT
Page 10: CTT
Page 11: CTT
Page 12: CTT
Page 13: CTT
Page 14: CTT
Page 15: CTT
Page 16: CTT
Page 17: CTT
Page 18: CTT

REFERENCES – ERROR-CORRECTING CODES

P. VIJAY KUMAR

1. Shu Lin, Daniel J.Costello, Error Control Coding - Second Edition,(often used as a coursetextbook) Pearson Education Inc. Pearson Prentice Hall, 2004.

2. W. Cary Huffman, and Vera Pless, Fundamentals of Error-Correcting Codes,(block codingemphasis) Cambridge University Press, 2010.

3. Ron M.Roth, Introduction to Coding Theory,(block coding emphasis) Cambridge UniversityPress, 2006.

4. Tom Richardson and Ruediger Urbanke, Modern Coding Theory,(emphasis on LDPC andrelated codes) Cambridge University Press, 2008.

5. F.J.Macwilliams and N.J.A.Sloane, The Theory of Error - Correcting Codes,(older but clas-sical book on block coding) North-Holland Mathematical Library, 1977.

6. Stephen B. Wicker, Error Control Systems for Digital Communication and Storage,(blockcoding emphasis) Prentice Hall Englewood Cliffs, NJ 07632, 1995.

7. Keith Chugg, Achilleas Anastasopoulos and Xiaopeng Chen, Iterative Detection:Adaptivity,Complexity Reduction, and Applications,(iterative decoding emphasis) Kluwer Academic Pub-lishers, 2001.

8. Rolf Johannesson and Kamil Sh. Zigangirov, Fundamentals of Convolutional Coding, (con-volutional coding only)IEEE Press, 1999.

9. Andrew J. Viterbi and Jim K. Omura, Principles of Digital Communication and Coding,(nicediscussion on on convolutional coding) McGraw-Hill Book Company, 2009.

10. William E. Ryan and Shu Lin, Channel Codes: Classical and Modern,(recent book treatingboth block and LDPC codes) Cambridge University Press, 2009.

11. Chris Heegard and Stephen B. Wicker, Turbo Coding,(early book on turbo codes) KluwerAcademic Publishers, 2010.

12. T. R. N. Rao and Eiji Fujiwara, Error-Control Coding for Computer Systems, Prentice Hallseries in computer engineering, Jan 1989.

13. Hideki Imai,Essentials of Error-Control Coding Techniques , Academic Press, 1990.

14. Ezio Biglieri , Dariush Divsalar , Peter J. McLane and Marvin K. Simon, Introduction toTrellis-Coded Modulation With Applications,(trellis-coded modulation) Macmillan Publishs-ing Company, 1991.

1

Page 19: CTT

15. V.S.Pless and W.C.Huffman, Handbook of Coding Theory : Volumes I & II,(two volumehandbook on coding theory: block and convolutional codes, allied topics) North Holland,1998.

16. Richard E. Blahut, Algebraic Codes for Data Transmission,Cambridge University Press, 2011.

17. Richard E. Blahut, Theory and Practice of Error Control Codes,Addison-Wesley PublishingCompany, 1983.

18. George C. Clark Jr. and J. Bibb Cain, Error-Correction Coding for Digital Communica-tions,(an older book, but makes for easy reading) Plenum Press, New York, June 1981.

19. Irving S. Reed and Xuemin Chen, Error-Control Coding for Data Networks , Kluwer AcademicPublishers, 1999.

20. R. J. McEliece, The Theory of Information and Coding,Cambridge University Press, 2004.

21. Elwyn R. Berlekamp, Algebraic Coding Theory Revised 1984 Edition,Aegean Park Press, 1984.

22. Stephen B. Wicker and Vijay K. Bhargava, Reed-Solomon Codes and Their Applications,IEEEPress, 1999.

23. Scott A. Vanstone and Paul C. van Oorschot, An Introduction to Error Correcting Codes withApplications,Kluwer Academic Publishers, 1989.

24. J. van Lint and G. van der Geer, Introduction to Coding Theory and Algebraic Geome-try,Birkhauser Verlag, 1989.

25. V.D.Goppa, Mathematics and Its Applications V.D.Goppa Geometry and Codes,Kluwer Aca-demic Publishers, 1988.

26. Richard B. Wells, Applied Coding and Information Theory for Engineers,Prentice Hall Infor-mation and System Sciences Series, 1998.

27. Peter Sweeney, Error Control Coding: From Theory to Practice , John Wiley and Sons, Ltd,2002.

28. Shu Lin,Tadao Kasami,Toru Fujiwara and Marc Fossorier, Trellises and Trellis-Based Decod-ing Algorithms for Linear Block Codes,Kluwer Academic Publishers, 1998.

29. H. Stichtenoth , Algebraic Function Fields and Codes,Springer-Verlag Berlin Heidelberg, 2010.

Book Chapter

30. P. V. Kumar, M. Win, H-F. Lu, C. Georghiades, “Error-Control Coding Techniques andApplications,” invited chapter in the handbook, Optical Fiber Telecommunications IV, editedby Ivan P. Kaminow and Tingye Li, Spring 2002.

2

Page 20: CTT

Journal Articles

31. Srinivas M. Aji and Robert J. McEliece, “Generalized Distributive Law” IEEE Trans. Inform.Theory, March 2000.

32. T. J. Richardson, A. Shokrollahi, and R. Urbanke, Design of capacity-approaching low-densityparity-check codes, IEEE Trans. Inform. Theory, vol. 47, Feb. 2001.

33. A. J. Viterbi,“Convolutional codes and their performance in communication systems,” IEEETransactions on Communications Technology, , October 1971.

3

Page 21: CTT
Page 22: CTT
Page 23: CTT
Page 24: CTT
Page 25: CTT
Page 26: CTT
Page 27: CTT
Page 28: CTT
Page 29: CTT
Page 30: CTT
Page 31: CTT
Page 32: CTT
Page 33: CTT
Page 34: CTT
Page 35: CTT
Page 36: CTT
Page 37: CTT
Page 38: CTT
Page 39: CTT
Page 40: CTT
Page 41: CTT
Page 42: CTT
Page 43: CTT
Page 44: CTT
Page 45: CTT
Page 46: CTT
Page 47: CTT
Page 48: CTT
Page 49: CTT
Page 50: CTT
Page 51: CTT
Page 52: CTT
Page 53: CTT
Page 54: CTT
Page 55: CTT
Page 56: CTT
Page 57: CTT
Page 58: CTT
Page 59: CTT
Page 60: CTT
Page 61: CTT
Page 62: CTT
Page 63: CTT
Page 64: CTT
Page 65: CTT
Page 66: CTT
Page 67: CTT
Page 68: CTT
Page 69: CTT
Page 70: CTT
Page 71: CTT

Summary of Lecture 5:

· Equivalence classes defined via cosets:○ Proof that it is an equivalence relation○ The nature of the equivalence class E_b=Hb ○ Examples:

□ integers modulo 6 and even subset□ Even parity check code

○ Elements in different cosets can be placed in 1-1 correspondence

· Rings and Fields ○ Axioms of a ring○ Ring with identity○ Commutative ring○ Integral domain ○ Division ring ○ Examples: where do we place them ?

Page 72: CTT

Lecture 6: Vector Spaces, Linear Independence and Basis

·Rings and Fields ○ Examples: where do we place them ?

Vector Spaces○ Axioms○ Examples○ Derived properties

· Subspaces○ Definition○ Example 1: plane in R^3○ Test for a subspace○ Further examples: repetition code and spc code

· Definition of a linear code○ Show how the test applies to the Hamming code (nullspace

of a matrix) ○ Point out that as far as subsets of F_2^n are concerned,

Page 73: CTT
Page 74: CTT
Page 75: CTT
Page 76: CTT
Page 77: CTT
Page 78: CTT
Page 79: CTT
Page 80: CTT
Page 81: CTT
Page 82: CTT
Page 83: CTT
Page 84: CTT
Page 85: CTT
Page 86: CTT
Page 87: CTT
Page 88: CTT
Page 89: CTT
Page 90: CTT
Page 91: CTT
Page 92: CTT
Page 93: CTT
Page 94: CTT
Page 95: CTT
Page 96: CTT
Page 97: CTT
Page 98: CTT
Page 99: CTT
Page 100: CTT
Page 101: CTT
Page 102: CTT
Page 103: CTT
Page 104: CTT
Page 105: CTT
Page 106: CTT
Page 107: CTT
Page 108: CTT
Page 109: CTT
Page 110: CTT
Page 111: CTT
Page 112: CTT
Page 113: CTT
Page 114: CTT
Page 115: CTT
Page 116: CTT
Page 117: CTT
Page 118: CTT
Page 119: CTT
Page 120: CTT
Page 121: CTT
Page 122: CTT
Page 123: CTT
Page 124: CTT
Page 125: CTT
Page 126: CTT
Page 127: CTT
Page 128: CTT
Page 129: CTT
Page 130: CTT
Page 131: CTT
Page 132: CTT
Page 133: CTT
Page 134: CTT
Page 135: CTT
Page 136: CTT
Page 137: CTT
Page 138: CTT
Page 139: CTT
Page 140: CTT
Page 141: CTT
Page 142: CTT
Page 143: CTT
Page 144: CTT
Page 145: CTT
Page 146: CTT
Page 147: CTT
Page 148: CTT
Page 149: CTT
Page 150: CTT
Page 151: CTT
Page 152: CTT
Page 153: CTT
Page 154: CTT
Page 155: CTT
Page 156: CTT
Page 157: CTT
Page 158: CTT
Page 159: CTT
Page 160: CTT
Page 161: CTT
Page 162: CTT
Page 163: CTT
Page 164: CTT
Page 165: CTT
Page 166: CTT
Page 167: CTT
Page 168: CTT
Page 169: CTT
Page 170: CTT
Page 171: CTT
Page 172: CTT
Page 173: CTT
Page 174: CTT
Page 175: CTT
Page 176: CTT
Page 177: CTT
Page 178: CTT
Page 179: CTT
Page 180: CTT
Page 181: CTT
Page 182: CTT
Page 183: CTT
Page 184: CTT
Page 185: CTT
Page 186: CTT
Page 187: CTT
Page 188: CTT
Page 189: CTT
Page 190: CTT
Page 191: CTT
Page 192: CTT
Page 193: CTT
Page 194: CTT
Page 195: CTT
Page 196: CTT
Page 197: CTT
Page 198: CTT
Page 199: CTT
Page 200: CTT
Page 201: CTT
Page 202: CTT
Page 203: CTT
Page 204: CTT
Page 205: CTT
Page 206: CTT
Page 207: CTT
Page 208: CTT
Page 209: CTT
Page 210: CTT
Page 211: CTT
Page 212: CTT
Page 213: CTT
Page 214: CTT
Page 215: CTT
Page 216: CTT
Page 217: CTT
Page 218: CTT
Page 219: CTT
Page 220: CTT
Page 221: CTT
Page 222: CTT
Page 223: CTT
Page 224: CTT
Page 225: CTT
Page 226: CTT
Page 227: CTT
Page 228: CTT
Page 229: CTT
Page 230: CTT
Page 231: CTT
Page 232: CTT
Page 233: CTT
Page 234: CTT
Page 235: CTT
Page 236: CTT
Page 237: CTT
Page 238: CTT
Page 239: CTT
Page 240: CTT
Page 241: CTT
Page 242: CTT
Page 243: CTT
Page 244: CTT
Page 245: CTT
Page 246: CTT
Page 247: CTT
Page 248: CTT
Page 249: CTT
Page 250: CTT
Page 251: CTT
Page 252: CTT
Page 253: CTT
Page 254: CTT
Page 255: CTT
Page 256: CTT
Page 257: CTT
Page 258: CTT
Page 259: CTT
Page 260: CTT
Page 261: CTT
Page 262: CTT
Page 263: CTT
Page 264: CTT
Page 265: CTT
Page 266: CTT
Page 267: CTT
Page 268: CTT
Page 269: CTT
Page 270: CTT
Page 271: CTT
Page 272: CTT
Page 273: CTT
Page 274: CTT
Page 275: CTT
Page 276: CTT
Page 277: CTT
Page 278: CTT
Page 279: CTT
Page 280: CTT
Page 281: CTT
Page 282: CTT
Page 283: CTT
Page 284: CTT
Page 285: CTT
Page 286: CTT
Page 287: CTT
Page 288: CTT
Page 289: CTT
Page 290: CTT
Page 291: CTT
Page 292: CTT
Page 293: CTT
Page 294: CTT
Page 295: CTT
Page 296: CTT
Page 297: CTT
Page 298: CTT
Page 299: CTT
Page 300: CTT
Page 301: CTT
Page 302: CTT
Page 303: CTT
Page 304: CTT
Page 305: CTT
Page 306: CTT
Page 307: CTT
Page 308: CTT
Page 309: CTT
Page 310: CTT
Page 311: CTT
Page 312: CTT
Page 313: CTT
Page 314: CTT
Page 315: CTT
Page 316: CTT
Page 317: CTT
Page 318: CTT
Page 319: CTT
Page 320: CTT
Page 321: CTT
Page 322: CTT
Page 323: CTT
Page 324: CTT
Page 325: CTT
Page 326: CTT
Page 327: CTT
Page 328: CTT
Page 329: CTT
Page 330: CTT
Page 331: CTT
Page 332: CTT
Page 333: CTT
Page 334: CTT
Page 335: CTT
Page 336: CTT
Page 337: CTT
Page 338: CTT
Page 339: CTT
Page 340: CTT
Page 341: CTT
Page 342: CTT
Page 343: CTT
Page 344: CTT
Page 345: CTT
Page 346: CTT
Page 347: CTT
Page 348: CTT
Page 349: CTT
Page 350: CTT
Page 351: CTT
Page 352: CTT
Page 353: CTT
Page 354: CTT
Page 355: CTT
Page 356: CTT
Page 357: CTT
Page 358: CTT
Page 359: CTT
Page 360: CTT
Page 361: CTT
Page 362: CTT
Page 363: CTT
Page 364: CTT
Page 365: CTT
Page 366: CTT
Page 367: CTT
Page 368: CTT
Page 369: CTT
Page 370: CTT
Page 371: CTT
Page 372: CTT
Page 373: CTT
Page 374: CTT
Page 375: CTT
Page 376: CTT
Page 377: CTT
Page 378: CTT
Page 379: CTT
Page 380: CTT
Page 381: CTT
Page 382: CTT
Page 383: CTT
Page 384: CTT
Page 385: CTT
Page 386: CTT
Page 387: CTT
Page 388: CTT
Page 389: CTT
Page 390: CTT
Page 391: CTT
Page 392: CTT
Page 393: CTT
Page 394: CTT
Page 395: CTT
Page 396: CTT
Page 397: CTT
Page 398: CTT
Page 399: CTT
Page 400: CTT
Page 401: CTT
Page 402: CTT
Page 403: CTT
Page 404: CTT
Page 405: CTT
Page 406: CTT
Page 407: CTT
Page 408: CTT
Page 409: CTT
Page 410: CTT
Page 411: CTT
Page 412: CTT
Page 413: CTT
Page 414: CTT
Page 415: CTT
Page 416: CTT
Page 417: CTT
Page 418: CTT
Page 419: CTT
Page 420: CTT
Page 421: CTT
Page 422: CTT
Page 423: CTT
Page 424: CTT
Page 425: CTT
Page 426: CTT
Page 427: CTT
Page 428: CTT
Page 429: CTT
Page 430: CTT
Page 431: CTT
Page 432: CTT
Page 433: CTT
Page 434: CTT
Page 435: CTT
Page 436: CTT
Page 437: CTT
Page 438: CTT
Page 439: CTT
Page 440: CTT
Page 441: CTT
Page 442: CTT
Page 443: CTT
Page 444: CTT
Page 445: CTT
Page 446: CTT
Page 447: CTT
Page 448: CTT
Page 449: CTT
Page 450: CTT
Page 451: CTT
Page 452: CTT
Page 453: CTT
Page 454: CTT
Page 455: CTT
Page 456: CTT
Page 457: CTT
Page 458: CTT
Page 459: CTT
Page 460: CTT
Page 461: CTT
Page 462: CTT
Page 463: CTT
Page 464: CTT
Page 465: CTT
Page 466: CTT
Page 467: CTT
Page 468: CTT
Page 469: CTT
Page 470: CTT
Page 471: CTT
Page 472: CTT
Page 473: CTT
Page 474: CTT
Page 475: CTT
Page 476: CTT
Page 477: CTT
Page 478: CTT
Page 479: CTT
Page 480: CTT
Page 481: CTT
Page 482: CTT
Page 483: CTT
Page 484: CTT
Page 485: CTT
Page 486: CTT
Page 487: CTT
Page 488: CTT
Page 489: CTT
Page 490: CTT
Page 491: CTT
Page 492: CTT
Page 493: CTT
Page 494: CTT
Page 495: CTT
Page 496: CTT
Page 497: CTT
Page 498: CTT
Page 499: CTT
Page 500: CTT
Page 501: CTT
Page 502: CTT
Page 503: CTT
Page 504: CTT
Page 505: CTT
Page 506: CTT
Page 507: CTT
Page 508: CTT
Page 509: CTT
Page 510: CTT
Page 511: CTT
Page 512: CTT
Page 513: CTT
Page 514: CTT
Page 515: CTT
Page 516: CTT
Page 517: CTT
Page 518: CTT
Page 519: CTT
Page 520: CTT
Page 521: CTT
Page 522: CTT
Page 523: CTT
Page 524: CTT
Page 525: CTT
Page 526: CTT
Page 527: CTT
Page 528: CTT
Page 529: CTT
Page 530: CTT
Page 531: CTT
Page 532: CTT
Page 533: CTT
Page 534: CTT
Page 535: CTT
Page 536: CTT
Page 537: CTT
Page 538: CTT
Page 539: CTT
Page 540: CTT
Page 541: CTT
Page 542: CTT
Page 543: CTT
Page 544: CTT
Page 545: CTT
Page 546: CTT
Page 547: CTT
Page 548: CTT
Page 549: CTT
Page 550: CTT
Page 551: CTT
Page 552: CTT
Page 553: CTT
Page 554: CTT
Page 555: CTT
Page 556: CTT
Page 557: CTT
Page 558: CTT
Page 559: CTT
Page 560: CTT
Page 561: CTT
Page 562: CTT
Page 563: CTT
Page 564: CTT
Page 565: CTT
Page 566: CTT
Page 567: CTT
Page 568: CTT
Page 569: CTT
Page 570: CTT
Page 571: CTT
Page 572: CTT
Page 573: CTT
Page 574: CTT
Page 575: CTT
Page 576: CTT
Page 577: CTT
Page 578: CTT
Page 579: CTT
Page 580: CTT
Page 581: CTT
Page 582: CTT
Page 583: CTT
Page 584: CTT
Page 585: CTT
Page 586: CTT
Page 587: CTT
Page 588: CTT
Page 589: CTT
Page 590: CTT
Page 591: CTT
Page 592: CTT
Page 593: CTT
Page 594: CTT
Page 595: CTT
Page 596: CTT
Page 597: CTT
Page 598: CTT
Page 599: CTT
Page 600: CTT
Page 601: CTT
Page 602: CTT
Page 603: CTT
Page 604: CTT
Page 605: CTT
Page 606: CTT
Page 607: CTT
Page 608: CTT
Page 609: CTT
Page 610: CTT
Page 611: CTT
Page 612: CTT
Page 613: CTT
Page 614: CTT
Page 615: CTT
Page 616: CTT
Page 617: CTT
Page 618: CTT
Page 619: CTT
Page 620: CTT
Page 621: CTT
Page 622: CTT
Page 623: CTT
Page 624: CTT
Page 625: CTT
Page 626: CTT
Page 627: CTT
Page 628: CTT
Page 629: CTT
Page 630: CTT
Page 631: CTT
Page 632: CTT
Page 633: CTT
Page 634: CTT
Page 635: CTT
Page 636: CTT
Page 637: CTT
Page 638: CTT
Page 639: CTT
Page 640: CTT
Page 641: CTT
Page 642: CTT
Page 643: CTT
Page 644: CTT
Page 645: CTT
Page 646: CTT
Page 647: CTT
Page 648: CTT
Page 649: CTT
Page 650: CTT
Page 651: CTT
Page 652: CTT
Page 653: CTT
Page 654: CTT
Page 655: CTT
Page 656: CTT
Page 657: CTT
Page 658: CTT
Page 659: CTT
Page 660: CTT
Page 661: CTT
Page 662: CTT
Page 663: CTT
Page 664: CTT
Page 665: CTT
Page 666: CTT
Page 667: CTT
Page 668: CTT
Page 669: CTT
Page 670: CTT
Page 671: CTT
Page 672: CTT
Page 673: CTT
Page 674: CTT
Page 675: CTT
Page 676: CTT
Page 677: CTT
Page 678: CTT
Page 679: CTT
Page 680: CTT
Page 681: CTT
Page 682: CTT
Page 683: CTT
Page 684: CTT
Page 685: CTT
Page 686: CTT
Page 687: CTT
Page 688: CTT
Page 689: CTT
Page 690: CTT
Page 691: CTT
Page 692: CTT
Page 693: CTT
Page 694: CTT
Page 695: CTT
Page 696: CTT
Page 697: CTT
Page 698: CTT
Page 699: CTT
Page 700: CTT
Page 701: CTT
Page 702: CTT
Page 703: CTT
Page 704: CTT
Page 705: CTT
Page 706: CTT
Page 707: CTT
Page 708: CTT
Page 709: CTT
Page 710: CTT
Page 711: CTT
Page 712: CTT
Page 713: CTT
Page 714: CTT
Page 715: CTT
Page 716: CTT
Page 717: CTT
Page 718: CTT
Page 719: CTT
Page 720: CTT
Page 721: CTT
Page 722: CTT
Page 723: CTT
Page 724: CTT
Page 725: CTT
Page 726: CTT
Page 727: CTT
Page 728: CTT
Page 729: CTT
Page 730: CTT
Page 731: CTT
Page 732: CTT
Page 733: CTT
Page 734: CTT
Page 735: CTT
Page 736: CTT
Page 737: CTT
Page 738: CTT
Page 739: CTT
Page 740: CTT
Page 741: CTT
Page 742: CTT
Page 743: CTT
Page 744: CTT
Page 745: CTT
Page 746: CTT
Page 747: CTT
Page 748: CTT
Page 749: CTT
Page 750: CTT
Page 751: CTT
Page 752: CTT
Page 753: CTT
Page 754: CTT
Page 755: CTT
Page 756: CTT
Page 757: CTT
Page 758: CTT
Page 759: CTT
Page 760: CTT
Page 761: CTT
Page 762: CTT
Page 763: CTT
Page 764: CTT
Page 765: CTT
Page 766: CTT
Page 767: CTT
Page 768: CTT
Page 769: CTT
Page 770: CTT
Page 771: CTT
Page 772: CTT
Page 773: CTT
Page 774: CTT
Page 775: CTT
Page 776: CTT
Page 777: CTT
Page 778: CTT
Page 779: CTT
Page 780: CTT
Page 781: CTT
Page 782: CTT
Page 783: CTT
Page 784: CTT
Page 785: CTT
Page 786: CTT
Page 787: CTT
Page 788: CTT
Page 789: CTT
Page 790: CTT
Page 791: CTT
Page 792: CTT
Page 793: CTT
Page 794: CTT
Page 795: CTT
Page 796: CTT
Page 797: CTT
Page 798: CTT
Page 799: CTT
Page 800: CTT
Page 801: CTT
Page 802: CTT
Page 803: CTT
Page 804: CTT
Page 805: CTT
Page 806: CTT
Page 807: CTT
Page 808: CTT
Page 809: CTT
Page 810: CTT
Page 811: CTT
Page 812: CTT
Page 813: CTT
Page 814: CTT
Page 815: CTT
Page 816: CTT
Page 817: CTT
Page 818: CTT
Page 819: CTT
Page 820: CTT
Page 821: CTT
Page 822: CTT
Page 823: CTT
Page 824: CTT
Page 825: CTT
Page 826: CTT
Page 827: CTT
Page 828: CTT
Page 829: CTT
Page 830: CTT
Page 831: CTT
Page 832: CTT
Page 833: CTT
Page 834: CTT
Page 835: CTT
Page 836: CTT
Page 837: CTT
Page 838: CTT
Page 839: CTT
Page 840: CTT
Page 841: CTT
Page 842: CTT
Page 843: CTT
Page 844: CTT
Page 845: CTT
Page 846: CTT
Page 847: CTT
Page 848: CTT
Page 849: CTT
Page 850: CTT
Page 851: CTT
Page 852: CTT
Page 853: CTT
Page 854: CTT
Page 855: CTT
Page 856: CTT
Page 857: CTT

NPTEL: Error-Correcting CodesExercises

P. Vijay KumarIndian Institute of Science

Basics of Block Codes

1. What is the smallest possible minimum distance of a block code of length n that cancorrect 2 errors and detect 5 errors ? If used only for error-detection, what is themaximum number of errors that the code can detect ?

2. A ternary code C is a code whose symbol alphabet is the set {0, 1, 2}, i.e., C is a subsetof {0, 1, 2}n. Even in {0, 1, 2}n, the definitions of Hamming weight and Hammingdistance remain as in the binary case. In the binary case, a code is a (td, tc) code iff

dmin ≥ td + tc + 1. (1)

Is this also true in the ternary case ? (The definition of a (td, tc) code remains as inthe binary case.) Justify your answer.

Mathematical Preliminaries

3. Prove that if G is an Abelian group under the operation + and H is a finite subset ofG, then H is a subgroup of G if and only if

a+ b ∈ H whenever a ∈ H and b ∈ H.

Hint: Associativity and commutativity carry over to any subset. If a ∈ H consider thelist {la | 0 ≤ l, l an integer } (la is a the sum of l copies of a). There are only a finitenumber of distinct elements in this list as H is finite. Thus l1a = l2a for some distinctintegers l1, l2. From this you should be able to conclude the existence of the identityelement and of the inverse.

4. Consider the groupG = Z4

2 = { all binary 4 tuples }

1

Page 858: CTT

Let H be the subgroup

H = {[0 0 0 0], [1 1 0 0], [1 1 1 1], [0 0 1 1].

Define two elements a, b ∈ G of G to be equivalent, written, a ≡ b if a + b ∈ H. Forexample,

[1 0 0 0] ≡ [0 1 0 0],

since[1 0 0 0] + [0 1 0 0] = [1 1 0 0] ∈ H.

This equivalence relation allows G to be partitioned into 4 subsets of size 4 calledequivalence classes where all the 4 elements within a subset are equivalent. Identifythe four equivalence classes and the elements that they contain.

5. Provide as best an algebraic characterization as you can (i.e., specify if it is a group, aring, a field, a vector space, etc) of the set

S = {2∑

i=2

aizi | ai ∈ F2}

endowed with the addition operation

2∑

i=0

aizi +

2∑

i=0

bizi =

2∑

i=0

cizi

where ci = ai + bi (mod 2) and with multiplication given by

2∑

i=0

aizi

2∑

i=0

bizi =

2∑

i,j=0

aibizi⊕j

where i⊕ j := i+ j (mod 3).

6. If G is a non-Abelian group having subgroup H, and if for g1, g2 ∈ G, we define g1 ∼ g2

iff g−12 g1 ∈ H, does this represent an equivalence relation. Justify your answer.

7. Identify coset representatives for the cosets of the subgroup

H = {a0 + a2x2 | a0, a2 ∈ F2}

of the group

G = {3∑

i=0

aixi | ai ∈ F2}.

2

Page 859: CTT

Linear Codes

8. Consider decoding of the length n = 7 repetition code using two methods:

(a) using bounded distance decoding (BDD) with dmin = 7

(b) maximum likelihood decoding (MLD) assuming that the channel is a binary sym-metric channel having crossover probability ε < 1

2

Is there a difference between the two methods when applied to this code ? Explainyour answer.

9. Write down a parity-check (p.c.) matrix for the binary linear code C whose generatormatrix is given by

G =

1 1 1 0 01 1 0 1 01 0 0 0 1

.

10. Identify a [4, 2] linear code C such that C⊥ is C itself, i.e., such that the code and itsdual are one and the same. Such codes are called self-dual codes.

11. For the purposes of this problem, let us define a linear [n, k] block code C to be system-atic if there exists a generator matrix for the code such that when the code is encodedusing that generator matrix, the first k code symbols (c0, c1, · · · , ck−1) are preciselythe k message symbols, (m0, m1, · · · , mk−1).

Under this definition, is the linear block code C having generator matrix

G =

1 1 0 1 1 1 10 1 1 0 1 1 01 0 1 0 0 1 11 0 1 1 1 0 1

a systematic code ?

Justify your answer in a few words. Show all your working.

12. Consider the linear block code C having generator matrix given by

G =

1 0 0 1 1 10 1 1 0 1 10 1 0 0 0 0

.

(a) What is the minimum distance of C ? Explain how you obtained your answer.

(b) What is the minimum distance of the dual code C⊥ ? Again, explain how youobtained your answer.

3

Page 860: CTT

13. Determine the minimum distance dmin of the [7, 3] linear block code C having parity-check matrix

H =

1 0 0 0 1 1 10 1 0 0 1 1 00 0 1 0 0 1 10 0 0 1 1 0 1

.

Explain your reasoning and show all your working.

14. Show that the repetition code and the parity-check code are the only possible MDScodes of length n = 7. (Hint: Start by attempting to construct an [n, k] MDS code byattempting to build up a parity-check matrix H for the code, one column at a time.Keep in mind that the parity-check matrix has n− k rows and it is required that anyn− k columns of H be linearly independent. ) Note: The same proof carries over toany length n. However, you are only required to do the case n = 7.

15. Use cosets of the linear block code C having parameters [5, 2] and generator matrix:

G =

[0 1 1 1 01 0 1 0 1

].

to partition the set F 52 , i.e., identify all cosets of C.

16. A linear block code C is used to accomplish error-correction over a Binary SymmetricChannel (BSC) with cross-over probability ε. The standard array is used to carry outmaximum-likelihood decoding (MLD) of the code. Then the probability of codeworderror Pwe can be determined

(a) just from knowing the weight distribution of the code

(b) just from knowing the list of all coset leaders,

(c) only if both the weight distribution of the code and the list of coset leaders isknown

(d) only if the entire standard array is provided.

Identify the most appropriate answer(s).

17. Use the Hamming bound to determine an upper limit to the size of a binary block codeof length n = 15 and minimum distance dmin = 7.

18. Consider the linear block code of length 5 and dimension 2 with the following generatormatrix

G =

[1 0 1 1 00 1 1 0 1

].

4

Page 861: CTT

(a) Choose as coset leaders the zero vector, all 5-tuples with weight 1 and {00011, 10001}.Construct the standard array together with syndromes for complete decoding.(The first row in this table should list the codewords and the first column tothe left should contain all the coset leaders. The last column should list thecorresponding syndromes. )

(b) Given that the received vector

r = [1 1 1 0 1]t,

what is the decoded codeword ? What is the residual error ? How many messagebits are in error ? You may assume that the transmitted codeword is the all-zerocodeword.

Repeat for the caser = [1 1 0 0 0]t.

Again, you may assume that the transmitted codeword is the all-zero codeword.

(c) When the code is used only for the purposes for correcting error, what is theprobability Pwe of codeword error when the crossover probability of the BSC isε = 10−4 ?

(d) On the same channel as in the previous problem, what is the probability of un-detected error if the code is used solely for the purposes of error detection ?

(e) With systematic encoding, the codeword corresponding to the message vector[m0, m1], is given by [m0, m1, p0, p1, p2]. On the same channel as in theprevious problem, what is the probability that the first message symbol will bein error if the code is used solely for the purpose of error correction ?

(f) On the same channel as in the previous problem, what is the probability thatof the two message bits [m0, m1], only the second message bit m1 is decodedincorrectly ?

19. Let C be an [n, k, d] linear code having (k × n) generator matrix G. Prove that anycollection of n−d+1 columns selected from the n columns of G is a linearly independentset.

20. The covering radius of a linear [n, k] code C is the smallest integer ρ such that for anyx ∈ Fn2 , there exists a codeword c ∈ C such that

dH(c, x) ≤ ρ.

(a) How would you determine ρ from a standard array decoding table of the code ?

(b) How would you determine ρ from inspection of the parity-check matrix H of thecode ?

(c) What is the covering radius of the [7, 4, 3] Hamming code ?

21. Derive the analogue of the Hamming bound as it applied to ternary codes, i.e., to codeshaving the ternary alphabet {0, 1, 2}.

5

Page 862: CTT

Convolutional Codes

22. In the field of of formal power series F2[[D]], find the first 7 terms in the power-seriesexpansion of

(a) 11+D2

(b) D1+D+D2

(c) D2

1+D2+D5

23. Determine whether the convolutional codes encoded using the G(D) below given beloware catastrophic. If so, find an infinite weight input sequence that generates a codewordof finite weight.

(a) G(D) = [1 +D +D3, 1 +D +D2, 1 +D2 +D3].

(b) G(D) = [1 +D3, 1 +D +D2 +D4, 1 +D2 +D3 +D4].

Hint: The irreducible polynomials of degree ≤ 3 over GF (2) are listed below:

degree 1: D, 1 +Ddegree 2: 1 +D +D2

degree 3: 1 +D +D3, 1 +D2 +D3 .

24. Consider the rate 1/2 convolutional code with

G(D) = [1 +D +D2 1 +D2].

(a) Draw a complete trellis diagram up to node level 6 (beginning at node level 0).Label all branches with code symbols.

(b) Use the trellis to determine the free distance dfree of the code.

(c) If the received sequence (across a BSC) is

r = (01 00 01 00 00 00 . . . ...)

find (the information sequences associated to) all survivors at node level 6.

(d) If the received sequence (across an AWGN channel) is

r = (4 − 1 − 3 2 6 − 5 2 4 5 3 5 5 . . . ...)

find (the information sequences associated to) all survivors at node level 6.

25. Consider a rate 1/3 convolutional code with

G(D) = [1 +D 1 +D2 1 +D +D2].

(a) Draw the state diagram for the encoder.

6

Page 863: CTT

(b) Compute the generating function AF (L = 1, D, I)

(c) Use this generating functon to determine the free distance dfree of the code.

26. Will the choice of generator matrix,

G(D) =[

1 +D +D2 +D3, 1 +D2 +D3 +D5, 1 +D4],

cause the associated convolutional code C to exhibit catastrophic error propagation ?

Explain fully your answer.

27. This question pertain to convolutional codes of rate 1n, with m memory elements in the

encoder that are required to encode a given set of N message symbols {ui}N−1i=0 that

are i.i.d, and equally likely to be 0 or 1. Under the conventional encoding of messagesymbols using a terminated convolutional code, the convolutional encoder is forced tobegin and end at the all-zero state. In encoding using a “tail-biting” however, the onlyrestriction that is placed is that the encoder is required to begin and end at the samestate, but this state could be any of the possible encoder states. Derive an expressionfor the exact rate of the convolutional code when operated in tail-biting fashion. [Hint:How many code symbols does the tail-biting convolutional encoder need to transmit ?]

The Generalized Distributive Law

28. Consider the “min-star” semi-ring ((−∞,∞], min∗, +) in which the min∗ operationis given by:

∗min(x, y) := −ln(e−x + e−y).

(a) Identify the identity element under the min-star operation

(b) Verify that the distributive law holds.

29. Consider the single parity-check code of length 3 having parity check matrix H =[1 1 1]. Thus vT = [v1, v2, v3] is a codeword if and only if Hv = 0. In acertain instance, when communicating over a binary symmetric channel (BSC) havingcrossover probability ε < 0.5, the received vector was found to be

yT = [0 1 0].

Use the GDL to carry out ML code-symbol decoding of this code. Show all intermediatesteps

(a) the formulation as an MPF problem

(b) the junction tree

(c) the message-passing schedule and the messages passed

7

Page 864: CTT

(d) the result of decoding

30. Consider the rate 12

convolutional code having polynomial generator matrix

G(D) = [1, 1 +D].

On a certain transmission, two message symbols u0 and u1 were encoded using thecode and then transmitted across a binary symmetric channel (BSC) having crossoverprobability ε < 1

2. The corresponding received symbols were

yT = [01 01].

Use the GDL and the junction tree shown below to carry out maximum aposteriori(MAP) decoding of ONLY the message bit u1. (Note that the encoder is NOT returnedto the all-zero state, i.e., there are no tail bits inserted into the message stream).

Show ALL intermediate steps and all your working clearly.

S0

U0, S0, S1 U1, S1U0, S0

U0 U1

Figure 1: Junction Tree associated to the convolutional code.

31. Write down the distributive law as it applies to the semi-rings numbered 5,6,8,9 inTable I of the “GDL paper” (the paper by Aji and McEliece).

32. Read Example 2.2 of the “GDL paper” and determine the savings in computationbetween using the brute force approach to computing the 8 transform coefficientsF (x1, x2, x3) and the approach that makes intelligent use of the distributive law.

33. Consider the rate 1/2 convolutional code with

G(D) = [1 +D +D2 1 +D2].

If the received sequence (across a discrete memoryless AWGN channel) is

r = (4 − 1 − 3 2 6 − 5),

8

Page 865: CTT

use the GDL algorithm to implement minimum probability of bit error decoding of themessage bits u0, u1, u2. Show all your working including the graphs that you use andthe message passing schedule.

34. Consider decoding the [7, 4, 2] linear block code for the case when the received vectoracross a binary symmetric channel with crossover probability ε << 1, is the vector

y = [1 0 0 0 0 0 0 ]T .

Use the GDL to make decisions based on maximizing the aposteriori probabilities

p(ui/y)

of the code symbols ui, i = 1, 2, 3, 4, 5, 6, 7 .

35. Verify that the min∗ sum semi-ring is in fact a semi-ring, starting from the defintionof the min∗ operation:

min∗(x, y) = min{x, y} − ln(1 + e−|y−x|) = −ln(e−x + e−y).

Identify the underlying set and the identity element under the min∗ operation.

36. Set up a schedule for computing the objective function at vertex W for Example 2.4of the GDL paper. Draw the corresponding message trellis.

37. Write down the distributive law as it applies to the semi-rings numbered 9,10 in TableI of the “GDL paper” (by Aji and McEliece).

38. Set up an efficient message-passing schedule for computing the objective function atvertex W for Example 2.4 of the GDL paper, i.e., identify the sequence in which youwould pass messages.

39. In the computation:

β(x3) =∑

x1, x2, x4, x5

f(x1)g(x2)h(x1, x2, x3)p(x3, x4)q(x3, x5),

all the variables xi, i = 1, 2, 3, 4, 5 take on values from an alphabet A of size | A |= q. Ifyou were to reorganize this expression to minimize the number of operations (additionsand multiplications), how would you do it and how many operations would you endup needing ?

40. Consider the problem of computing

F (x1) =9∑

x2=0

9∑

x3=0

9∑

x4=0

f(x2, x3, x4)g(x3, x4)h(x1, x2, x4).

The functions f(·), g(·),h(·), are all real-valued functions.

9

Page 866: CTT

(a) It is desired to pose this problem as a marginalize a product function problem.Identify the corresponding universal set, the corresponding local domains and thelocal and global kernels.

(b) Organize if possible these local domains into a junction tree. Make clear all yourworking.

41. Consider maximum-likelihood code-symbol decoding of the binary block code havingparity check matrix H given by

H =

1 1 0 1 0 0 00 0 1 1 0 1 00 0 0 1 1 0 1

.

Thus all codewords c = [c1, c2, c3, c4, c5, c6, c7]T , satisfy Hc = 0. The decoding problem

when posed as an marginalize-a-product function problem, leads to the graph shown inFig. 2. Use the distributive law in conjunction with this graph, to efficiently computethe probabilities Pr(c4 = 0/r) and Pr(c4 = 1/r) where the received vector r is givenby r = [1111111]T . You may assume that the channel is a binary symmetric channel(BSC) having crossover probability 0 < ε << 1.

A suggested approximation In your computations, you will run into expressionsof the form

aiεi + ai+1ε

i+1 + ai+2εi+2 + · · · + ai+kε

i+k,

where i ≥ 0, k ≥ 0 and the ai are integers ≤ 10. It is suggested that whenever youencounter such an expression, you make the approximation

aiεi + ai+1ε

i+1 + ai+2εi+2 + · · · + ai+kε

i+k ≈ aiεi.

42. Consider the joint probability function

p({ui}3i=0, {si}4i=0, {yi}3i=0) = p(s0)3∏

i=0

p(ui)p(si+1/si, ui)p(yi/si, ui)

associated with a convolutional code. As in class the {ui ∈ {0, 1}} represent the binarymessage symbols, the {si} is the state sequence and {yi} are the received symbols.Consider the problem of maximum-likelihood code-symbol decoding of this code, i.e.,of computing p(uk/{yi}3i=0), 0 ≤ k ≤ 3.

(a) Present this as an MPF problem,

(b) organize the local domains into a junction tree

(c) show that message passing can be organized into a forward wave and a backwardwave and that the forward wave is in essence, a sequence of matrix multiplications

Note: It is NOT necessary to do anything beyond what is asked above!

10

Page 867: CTT

1

3

6

2

4

5 346

124

7

457

A

B C

Figure 13: Example of junction tree.

corresponding to the local domains {1, 2, 4}, {3, 4, 6}, and {4, 5, 7}, respectively. Fig. 13shows the local domains organized into a junction tree. We are interested here in computingthe objective functions at all vertices Vi, i = 1, . . . , 7. With regard to the message schedule,we will adopt an inward-outward schedule in which the outlying vertices send their messagesinward first. Once the innermost vertex V4 has acquired knowledge of all the local kernels,the outward phase of message passing begins. The validity of this message schedule can beverified with the aid of a message trellis if desired. The messages passed are as follows:

Phase 1

µ1,A(S1) = α1(x1) ⇒ µ1,A

:=

[µ 1,A(0)µ1,A(1)

]=

[1θ

]= µ

6,B= µ

7,C. (119)

Similarly,

µ2,A

=

[θ1

]= µ

3,B= µ

5,C. (120)

Phase 2µA,4(x4) =

x1,x2

µ1,A(x1)µ2,A(x2)χA (x1, x2, x4) . (121)

Since χA(x1 + x2 + x4) = 0 for x4 6= x1 + x2, we have

µA,4(0) =∑

x1+x2=0

µ1,A(x1)µ2,A(x2) = µ1,A(0)µ2,A(0) + µ1,A(1)µ2,A(1) = θ + θ = 2θ. (122)

Similarly, µA,4(1) = 1+θ2, which implies µA,4

=

[2θ

1 + θ2

]. Also, µ

C,4=

[2θ

1 + θ2

]= µ

B,4.

Phase 3µ4,A(x4) = µB,4(x4)µC,4(x4)α4(x4). (123)

55

Figure 2: Junction Tree associated to the [7, 4, 2] code..

LDPC Codes

43. In the density evolution analysis of (dv, dc)-regular LDPC codes, where the goal is todetermine the evolution of the density of number of incorrect messages passed betweenvariable nodes and check nodes, it is customary to assume that the all-1 codeword istransmitted. What are the assumptions on the channel and the processing carried outat the variable and check node under which this assumption is valid ? Explain youranswer while making clear any notation that you introduce.

44. Derive from first principles, the transformation of densities that takes place during aniteration at a check node.

45. Consider density evolution associated to Gallager Decoding Algorithm A applied toan LDPC code C. Thus the channel is a BSC with cross-over probability ε << 1 andall messages passed are either 1 or −1. You may assume that the neighborhood ofevery node in the Tanner graph of C is tree-like to depth 8. What is the probabilityp

(1)−1that at the end of iteration 1, the message passed from a variable node to check

node will be in error. Notation is as in class. Following an initial round of messagepassing, from the variable nodes to check nodes, based only on channel inputs, eachsubsequent iteration is composed of two rounds of message passing: from check nodeto variable node followed by from variable node back to check node. Show all your

11

Page 868: CTT

working clearly. You may use the fact that ε << 1 to simplify calculations. Hence aε2

for integer constants a < 100 (say) may safely be ignored in comparison with ε, etc.

46. Consider the variation of belief propagation decoding of binary LPDC codes in which,in place of beliefs, the messages passed correspond to log-likelihood ratios (as discussedin class).

(a) Identify (it is not necessary to derive them) the variable and check node maps

ψ(0)v (l0), ψ

(l)v (l0, l1, l2, . . . , ldv−1), ψ

(l)c (l1, l2, . . . , ldc−1).

(b) Do these maps satisfy the variable-node and check-node symmetry conditionswhich (along with the channel symmetry condition) permit us to conclude thatthe number of incorrect messages passed is the same regardless of the transmittedcodeword ? Make clear your reasoning

47. Is the computational tree associated with variable node 11 (at the top of the graphand incorrectly labelled as node 10 :-) ) in the Tanner graph in Fig. 3 of a certainLDPC code a junction tree ? If so, identify the associated MPF problem along withthe local domains and the local kernels. What is the objective function being computedif messages are passed as indicated by the arrows ?

Figure 3: Computational tree associated to node 10 in Tanner Graph.

48. (a) In the context of the performance analysis of LDPC codes, state (in terms ofthe notation introduced in class), the variable and check-node symmetry assump-tions that go into showing that the probability of passing an incorrect message isindependent of the transmitted codeword.

12

Page 869: CTT

(b) Show clearly that the check-node symmetry condition holds when LDPC codesare decoding using belief propagation with log-likelihood ratios (LLR) in place ofbeliefs.

Finite Fields & Cyclic Codes

49. Use the Euclidean division algorithm (EDA) to determine the gcd of 6711 and 831.Express the gcd as a linear combination u ∗ 6711 + v ∗ 831 of 6711 and 831.

50. Find the inverse of 7 modulo 13 using the EDA.

51. Identify all primitive elements of the finite fields of size 7 and 13 (the finite field of size13 is the set of all integers modulo 13).

52. Over GF (2), compute if possible, the inverse of (1 + x) modulo (1 + x+ x2 + x3 + x4).

53. Let α be a primitive element of GF (64). Identify all the elements in all the subfieldsof GF (64) in terms of α.

54. Use the irreducible polynomial (irreducible over GF (5)) x2 +x+2 to construct a finitefield of 25 elements. If α denotes a root of x2 + x+ 2, then α is known to be primitivein GF (25). Set up an add-1 table for GF (25). Identify the 5-cyclotomic cosets modulo24. Find the minimal polynomials of all elements in the field. Compute the product ofall the minimal polynomials (each distinct polynomial is taken just once) including theminimal polynomial x of the zero element. Which powers of α constitute the subfieldGF (5) of GF (25) ?

55. Identify the 3-cyclotomic cosets modulo 26 as well as the 2-cyclotomic cosets modulo19.

56. Let α be a primitive element of GF (26). Identify all the correct answers below with a√

• α + α4 ∈ GF (4)

• α + α8 ∈ GF (8)

• none of the above

57. The polynomials over GF (2) given below are all irreducible. Identify with a√

, allthose having the property that all of their zeros are contained in GF (256).

• x2 + x+ 1

• x3 + x2 + 1

• x4 + x3 + 1

13

Page 870: CTT

• x5 + x2 + 1

• x6 + x+ 1

• x8 + x6 + x5 + x4 + 1

58. Identify the smallest finite field of characteristic 2 that contains a primitive 17-th rootof unity.

59. In the notation used in class with regard to finite field Fourier transforms, let q = 2,N = 15 and α be a primitive element of F16 satisfying α4 + α + 1 = 0. Let

(s(t), t = 0, 1, 2, . . . , 14) = 000110000101101 .

Compute the Fourier transform s(λ) of s(t). Compute also the Fourier transform ofs(t) + s(t+ 2).

60. In the notation used in class with regard to finite field Fourier transforms, let q = 2,N = 15 and α be a primitive element of F16 satisfying α4 + α + 1 = 0. Determinethe basic sequence b(t) of “frequency” λ = 6. Determine the Fourier transform of thesequence c(t) given by c(t) = b(2t), 0 ≤ t ≤ 14.

61. Consider the binary (i.e., q = 2) cyclic code of length N = 15 consisting of all binarycodewords (c(t), 0 ≤ t ≤ 14) satisfying

C = {c(t) | c(λ) = 0, λ = 0, 7, 14, 13, 11}.

Transforms are computed using a primitive element of GF (16) satisfying α4+α+1 = 0.Find a codeword B(t) ∈ C such that every codeword c(t) in C can be expressed as alinear combination of cyclic shifts of B(t), i.e., can be expressed in the form

c(t) =14∑

t=0

u(τ)B(t− τ)

where u(τ) ∈ {0, 1},∀τ .

62. Determine the number of binary sequences {at} of period N = 15 that satisfy thecondition

aλ ∈ {0, 1}, all λ, 0 ≤ λ ≤ 14.

63. Let q = 2 and N = 23. What is the order m of q (mod N) ? Let α be a primitiveN -th root of unity lying in GF (2m). Determine the dimension of the binary q = 2cyclic code of length N = 23 all of whose codewords c(t) satisfy

c(λ) = 0, λ = 1 .

64. Why are there no interesting linear, cyclic binary codes of length N = 19 ?

14

Page 871: CTT

65. Design a single-error correcting, double-error detecting binary linear, cyclic code oflength 21. Naturally you would like to have dimension k as large as possible.

66. Identify the null spectrum of a Reed-Solomon (RS) code over GF (9) code of lengthN = 8 and designed distance dmin = 6.

67. How many distinct binary cyclic codes of length 41 are there ? (Include in your count,the cyclic code corresponding to the set of all binary 41-tuples as well as the cycliccode consisting of just the all-zero codeword).

68. Consider the binary cyclic code C of lengthN = 15 with null spectrum {1, 2, 3, 4, 5, 6, 8, 9, 10, 12}.You may assume that transforms are computed using primitive element α ∈ GF (16)satisfying α4 + α + 1 = 0.

Does the all-one codeword (1, 1, . . . 1, 1) belong to the cyclic code C ? Explain youranswer.

69. Use the irreducible polynomial (irreducible over F3) x2 + x + 2 to construct a finite

field of 9 elements. If α denotes a root of x2 + x+ 2, then α is known to be primitivein F9.

(a) Set up an add-1 table for F9.

(b) Identify the 3-cyclotomic cosets modulo 8.

(c) Find the minimal polynomials of all elements in the field.

70. Use the Mobius inversion formulae to determine the number of irreducible polynomialsof degree 12 over the binary field F2 .

71. If α, β in F16 have orders a, b, then is it always true that αβ has order = lcm(a, b) ?Justify your answer.

72. (a) How many binary cyclic codes of length 23 are there ?

(b) Design a double-error-correcting cyclic code of length 23 and identify its dimen-sion.

15