1 / 49

CIS 6930 Powerline Communications Error Handling

(c) 2013 Richard Newman. CIS 6930 Powerline Communications Error Handling. Error Handling - Outline. Analysis Forward Error Correction Copy codes Block codes Convolutional codes Scrambling Concatenated codes Turbo-codes Low Density Parity Check (LDPC) Backward Error Correction (BEC)

royal
Télécharger la présentation

CIS 6930 Powerline Communications Error Handling

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. (c) 2013 Richard Newman CIS 6930Powerline CommunicationsError Handling

  2. Error Handling - Outline • Analysis • Forward Error Correction • Copy codes • Block codes • Convolutional codes • Scrambling • Concatenated codes • Turbo-codes • Low Density Parity Check (LDPC) • Backward Error Correction (BEC) • Stop-and-Wait • ARQ • Conclusions

  3. Analysis • Measures • Nature of errors detected/corrected • Efficiency • Efficiency: U = Uf Up Ue * • Uf = framing efficiency = D/(D+H) ** where D = Data length and H = Header length • Up = protocol utilization • Depends on protocol, propagation delay, transmission time • Ue = error caused utilization efficiency = 1-P where P = frame error rate

  4. Caveats • Efficiency formula does not show • PHY layer efficiency • Variable data rate effects • PHY Efficiency * • Must consider time spent to deliver frame • Include required gaps/spaces, preamble • Also includes PHY layer FEC • Variable Data Rate Effects ** • Frame header/delimiter may be sent at lower modulation rate • Esp. for wireless and PLC

  5. Error Correction Strategies • Forward Error Correction • Include sufficient redundancy in transmitted units that errors can be corrected • Simplifies sender protocol – used in PHY • Backward Error Correction • Include sufficient redundancy in transmitted units that errors can be detected • Retransmit damaged units • More efficient* – used in MAC and above • Limitations • Hamming Distance of code limits capabilities • Always possible to “fool” receiver * Efficiency depends – if have to resend large frame for single error, maybe not so much...

  6. General ECC Considerations • Systematic vs. non-systematic • Systematic = data bits appear in coded stream • Non-systematic = no data bits identifiable • Hamming Distance • H(x,y) = number of bits where x and y differ • Code C = {x1, x2, ..., xN} set of valid codewords • d = H(C) = min{H(x,y) | x and y are distinct codewords in C} • Maximum detection ability = d-1 • Maximum correction ability = (d-1)/2

  7. Error Correction/Detection Theory • Code C is subset of n-bit binary words • Received word R = T + E, for transmitted word T and error E • H(C) determines maximum detection/correction capability • Max detection is d-1 • Max correction is floor (d-1)/2 • If correcting, then won't detect! • Can reduce correction to increase detection of uncorrectable errors, but correction + detection < d • Efficiency • Detection can be done with fixed # parity bits • Larger block has less overhead, but is more prone to errors • Correction requires parity bits scale with info bits • Economy of scale still exists – may be non-linear

  8. Forward Error Correction • Block vs. continuous • Block = set number of information symbols encoded into set number of code symbols • Internal fragmentation • Need for delimitation • Continuous = stream of information symbols encoded into stream of code symbols • Memory/constraint length – must “fill the pipeline” • Linearity • Sum of two code words is a code word • Concatenation • Combine two codes (inner and outer) to increase correction capabilities

  9. Forward Error Correction • Efficiency = code rate • Rate = k/n for (n,k) code • k = “information bits” • n = total bits • t = n-k = redundant bits • With continuous codes, need to account for “tail” - the number of bits in the memory

  10. Block Codes • Copy codes • LRC • Hamming codes • BCH • Reed-Solomon • LDPC

  11. Block Codes • Copy Codes • Simplest code, still useful at times • Copy data bits r times to encode • Use received copies to “vote” for input value • Can survive a burst error if scrambled • LRC – Longitudinal Redundancy Check • Information bits arranged in p-1 by q-1 matrix • Each row has parity bit at the end • Each column has parity bit at the bottom • n = pq, k = (p-1)(q-1), r = p+q-1 • Detects single bit errors

  12. LRC Example 1 0 1 1 0 1 1 0 1 0 1 1 = information bits 1 0 1 1 _ 1 0 1 1 1 0 1 1 0 _ -> 0 1 1 0 0 1 0 1 1 _ 1 0 1 1 1 _ _ _ _ _ 01100 <- LRC ^ VRC 1 0 1 1 1 0 1 1 0 0 1 0 1 1 1 0 1 1 0 0 = code word

  13. LRC Example 1 0 1 1 1 0 1 1 0 0 1 0 1 1 1 0 1 1 0 0 = sent 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 = error 1 0 1 1 1 0 1 0 0 0 1 0 1 1 1 0 1 1 0 0 = received 1 0 1 1 1 0 1 0 0 0 X 1 0 1 1 1 01100 X errors in LRC and VRC locate bit error

  14. Hamming Codes • Hamming Codes • “perfect” 1-bit error correction • 2n-1 bits per code word • n parity bits, remainder systematic information bits • Parity bit i is in position 2i • Parity bit i checks even parity of bits in positions with ith bit of location non-zero

  15. Hamming Code Example n = 4, length N = 16-1 = 15 bits, k = 11, r = 4 information bits = 10110110101 f e d c b a 9 8 7 6 5 4 3 2 1 bit positions 1 0 1 1 0 1 1 _ 0 1 0 _ 1 _ _ info bits in posn ---------------->1 parity bit 3 --------- ------->0 parity bit 2 ----- ----- ----- -->0 parity bit 1 -- -- -- -- -- -- -- >0 parity bit 0 1 0 1 1 0 1 1 1 0 1 0 0 1 00 code word

  16. Hamming Code Example f e d c b a 9 8 7 6 5 4 3 2 1 bit positions 1 0 1 1 0 1 1 1 0 1 0 0 1 0 0 code word 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 error 1 0 1 1 0 0 1 1 0 1 0 0 1 0 0 received word ---------------->0 parity bit 3 X --------- ------->0 parity bit 2 ----- ----- ----- -->1 parity bit 1 X -- -- -- -- -- -- -- >0 parity bit 0 • Syndrome = 1010 = a = location of error • Bit error => invert received bit to correct it

  17. BCH Codes • Bose, Chaudhuri, Hocquenghem ca. 1960 • Easy to decode • Syndrome decoding • Easy to implement in hardware • Cyclic code • Polynomial code over a finite field • Reed-Solomon codes a special form of BCH

  18. Reed-Solomon Codes • Based on oversampled polynomial • Redundant samples allow optimal polynomial to be recovered if most samples are good • Symbols greater than 1 bit • m-bit symbols => limit block size to 2m-1 symbols • 2 parity symbols needed to correct each error • One to locate position (or none) • One to specify correction to symbol • 1 parity symbol needed to correct each erasure • Handles small bursts • Popular • DVDs, CDs, Blu-Ray, DSL, WiMax, DVB, ATSC, Raid-6

  19. Low Density Parity Check Codes • Linear code • Capacity approaching code • Can get near to Shannon limit in symmetric, memoryless channel • Uses iterative belief propagation • Defined by sparse parity check matrix • No encumbering patents • Used in DVB-S2 digital TV, ITU-T G.hn, 10GBase-T

  20. Low Density Parity Check Codes Forney Factor Graph representation (fragment) Inputs Variable nodes Constraint nodes Probability values associated with variable nodes used with constraints to update value and confidence in variable nodes iteratively

  21. Convolutional Codes • May be systematic or not • Shift register for information bits • Each output bit has one or more taps into shift register • Tapped values are XORed to produce output • Outputs are sent round robin • May “puncture” output to increase coding rate • May “scramble” input to spread errors out • May “puncture” (drop out bits) to increase rate

  22. + Convolutional Codes out in 1 0 0 1 1 0 1 -> = info bits 10001 1 1 0 1 0 1 1 0 1 0 0 1 1 -> = output --------- tail Initialize shift register with 0’s, then shift in one bit at a time, then read one bit from each output

  23. Convolutional Codes Constraint length m (i.e., amount of memory or size of shift register) m = n+1, where n = degree of G(x)

  24. Decoding Convolutional Codes • Maximum Likelihood Decoding • Viterbi Algorithm • “Trellis” decoding • Dynamic programming • Number of states = 2m, m=constraint length • State = contents of shift regisiter • Cost = HD for transition based on received bits http://www.cambridge.org/resources/0521882672/7934_kaeslin_dynpro_new.pdf

  25. Shift Register Trellis Graph

  26. Trellis edges depend on code Start state? Start state defined by code specification Usually all 0s

  27. Trellis Decoding of Convolutional Codes Illegal transitions? Only two legal transitions – 0 or 1 shift in G(x) will determine arc labels depending on state and input shifted in

  28. Trellis Decoding of Convolutional Codes Illegal transitions follow same edges, but cost according to errorse.g., if receive 01 or 10 in state 000 then 000 or 100 at cost 1

  29. Trellis Decoding – our code Illegal transitions follow same edges, but cost according to errorse.g., if receive 01 or 10 in state 000 then 000 or 100 at cost 1

  30. Trellis Decoding – our example On each pair of received bits, update minimum cost for each state, preserving path(s) that produce that cost – 4 bits later can correct

  31. Decoding Convolutional Codes • Maximum Likelihood Decoding • At any point in received stream, output the stream of input bits most likely to have produced that output (i.e., fewest errors) • Viterbi Algorithm • “Trellis” keeps track of state of memory • For each possible state, track least cost to get there • Accumulate costs – cost of state B is minimum over all previous states A of cost of A plus cost to transition to B from A given current input • If multiple ways to reach a state, take cheapest one • Only have to maintain costs for current state set, then update based on received bits

  32. Scrambling • Convolutional codes correct well when errors are sparse • Tend to have problems with burst errors • Scramble bits after encoding, before decoding • Concatenated codes – allow errors/resynch • Scrambling • Shuffle order of bits on the way out/in • Interleaver depth = memory required to shuffle • E.g., fill block in row order, read out column order

  33. Concatenated Codes • Can increase correcting power by combining two codes – inner and outer • Apply outer first when encoding, last decoding • Apply inner last encoding, first decoding • RSV – Reed-Solomon outer, Viterbi inner • Very popular • Convolutional inner code corrects sparse errors • R-S outer code corrects bursts • (n,k) and (n’,k’) code produce (nn’,kk’) code

  34. Concatenated Codes Scrambler Outer Coder Inner Coder Sender Channel Noise Outer Coder Inner Coder Scrambler Receiver

  35. Turbo Codes • Essentially concatenating two convolutional codes (may be the same code) • One code operates on straight input • Other code operates on delayed and interleaved input • Decoding involves iteration between the two codes • Can approach Shannon Limit • Patents held by French Telecom

  36. Recall Error Correction Strategies • Forward Error Correction • Include sufficient redundancy in transmitted units that errors can be corrected • Simplifies sender protocol – used in PHY • Backward Error Correction • Include sufficient redundancy in transmitted units that errors can be detected • Retransmit damaged units • “More efficient” – used in MAC and above • Limitations • Hamming Distance of code limits capabilities • Always possible to “fool” receiver • Efficiency of BEC depends on size of frame & FER

  37. Backward Error Correction (BEC) • Received data cannot be corrected • Include checksum/redundancy check to detect errors (detection is cheap) • Retransmit frames that have errors • How does sender know which to resend? • ACK – OK, don’t resend • NAK – Received damaged frame • No response – time out and resend • ACKs • Cumulative vs. individual vs. SACK

  38. BEC Requirements • Sender • Must add error detection code to frames • Must store frames sent until ACKed • Must know which frames receiver got correctly • Must know when to resend frame • Must have timer for ARQ • Receiver • Must check error detection code on frames • Must be able to distinguish duplicate frames • Must be able to signal sender on state • May buffer out-of-order frames

  39. Acknowledgements • ACK types • Individual ACK – just the SN indicated • Cumulative ACK – indicates next expected unit • Beneficial when ACKs are lost - redundancy • SACK – indicates multiple good/back units • Sequence Number (SN) per unit • Units may be frames, bytes, cells, etc. • SNs eventually wrap around • Need to avoid confusion – send/receive window • Larger SN = more framing overhead

  40. Error Detection • Parity • ARC – Arithmetic Redundancy Check • Checksum adds values (with carry) • Drop overflow • May rotate addends • CRC – Cyclic Redundancy Check • Linear code, generated by polynomial • Easy to implement in feedback shift register • Cryptographic Hashes – MD5, SHA, HMAC • Useful for secure integrity checks

  41. Stop-and-Wait ARQ • One unit sent at a time • Wait for response (ACK/NAK) • Units have sequencing information (1 bit) • Accept received units in strict sequence • Gaps are not allowed • Duplicates are discarded • One-bit acknowledgement (ACK) • Resend unit on time-out or NAK

  42. Stop-and-Wait ARQ L/R L = Unit length (bits) R = Transmission rate (bps) T = Propagation delay (sec) Let a = T/(L/R) = propagation delay in frames Cycle time = Tc = 1+2a Data Tx time = 1 Utilization due to protocol = Up = 1/(1+2a) 2T 1 2a

  43. Stop-and-Wait ARQ Receiver needs way to distinguish duplicate frame from new frame, since ACK may be lost.... Frame 1 Frame 1 X Resent Frame 1 Frame 2

  44. Go-Back-N ARQ • Transmit units according to send window, k • Units have sequencing information • Accept received units in strict sequence • Gaps are not allowed • Duplicates, out of order units are discarded • Cumulative acknowledgement (ACK) • Resend un-acked units • Includes those received after a lost unit

  45. GBN ARQ k Cycle time = Tc = 1+2a Data Tx time = k Up = k/(1+2a) if no errors, and k <= 1+2a If errors at probability P, Expect to resend k units per error, so if k <= 1+2a, Up = k/(2a+1)(1-P-kP) U = k(1-P)/(2a+1)(1-P-kP) 1 2a 2a

  46. Selective Repeat ARQ • Transmit units according to transmit window • Units have sequencing information • Store received units according to sequencing information • Gaps are allowed • Duplicates are discarded • Selective acknowledgement (SACK) • Resend un-acked units

  47. Selective Repeat ARQ • Efficiency is best possible • Only redundancy is sequence info and error detection • Never resend units unnecessarily • If P = prob(unit is lost), k = Tx window, then Utilization U = (1-P)k/(2a+1) for k <= 1+2a (U = (1-P) if k > 1+2a)

  48. Observations w.r.t. PLC • Need a variety of techniques to deal with PL channel impairments • Modulation and coding used to get near channel capacity means there will be errors • Channel changes means there will be too many errors sometimes • If raw error rate too low, then get closer!!! • Longer frames mean greater error probability • Want to correction unit size to ensure success

  49. Observations w.r.t. PLC • Different coding levels needed for different PPDU/MPDU parts • Frame delimiter needs to be heard by all, so must be very robust • Payload can be adapted to highest rate path can handle • Coding method needs to deal with impulse noise • Want to be selective about what must be resent • Cross-layer design issues

More Related