fec.go 7.5 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311
  1. package kcp
  2. import (
  3. "encoding/binary"
  4. "sync/atomic"
  5. "github.com/klauspost/reedsolomon"
  6. )
  7. const (
  8. fecHeaderSize = 6
  9. fecHeaderSizePlus2 = fecHeaderSize + 2 // plus 2B data size
  10. typeData = 0xf1
  11. typeFEC = 0xf2
  12. )
  13. type (
  14. // fecPacket is a decoded FEC packet
  15. fecPacket struct {
  16. seqid uint32
  17. flag uint16
  18. data []byte
  19. }
  20. // fecDecoder for decoding incoming packets
  21. fecDecoder struct {
  22. rxlimit int // queue size limit
  23. dataShards int
  24. parityShards int
  25. shardSize int
  26. rx []fecPacket // ordered receive queue
  27. // caches
  28. decodeCache [][]byte
  29. flagCache []bool
  30. // zeros
  31. zeros []byte
  32. // RS decoder
  33. codec reedsolomon.Encoder
  34. }
  35. )
  36. func newFECDecoder(rxlimit, dataShards, parityShards int) *fecDecoder {
  37. if dataShards <= 0 || parityShards <= 0 {
  38. return nil
  39. }
  40. if rxlimit < dataShards+parityShards {
  41. return nil
  42. }
  43. dec := new(fecDecoder)
  44. dec.rxlimit = rxlimit
  45. dec.dataShards = dataShards
  46. dec.parityShards = parityShards
  47. dec.shardSize = dataShards + parityShards
  48. codec, err := reedsolomon.New(dataShards, parityShards)
  49. if err != nil {
  50. return nil
  51. }
  52. dec.codec = codec
  53. dec.decodeCache = make([][]byte, dec.shardSize)
  54. dec.flagCache = make([]bool, dec.shardSize)
  55. dec.zeros = make([]byte, mtuLimit)
  56. return dec
  57. }
  58. // decodeBytes a fec packet
  59. func (dec *fecDecoder) decodeBytes(data []byte) fecPacket {
  60. var pkt fecPacket
  61. pkt.seqid = binary.LittleEndian.Uint32(data)
  62. pkt.flag = binary.LittleEndian.Uint16(data[4:])
  63. // allocate memory & copy
  64. buf := xmitBuf.Get().([]byte)[:len(data)-6]
  65. copy(buf, data[6:])
  66. pkt.data = buf
  67. return pkt
  68. }
  69. // decode a fec packet
  70. func (dec *fecDecoder) decode(pkt fecPacket) (recovered [][]byte) {
  71. // insertion
  72. n := len(dec.rx) - 1
  73. insertIdx := 0
  74. for i := n; i >= 0; i-- {
  75. if pkt.seqid == dec.rx[i].seqid { // de-duplicate
  76. xmitBuf.Put(pkt.data)
  77. return nil
  78. } else if _itimediff(pkt.seqid, dec.rx[i].seqid) > 0 { // insertion
  79. insertIdx = i + 1
  80. break
  81. }
  82. }
  83. // insert into ordered rx queue
  84. if insertIdx == n+1 {
  85. dec.rx = append(dec.rx, pkt)
  86. } else {
  87. dec.rx = append(dec.rx, fecPacket{})
  88. copy(dec.rx[insertIdx+1:], dec.rx[insertIdx:]) // shift right
  89. dec.rx[insertIdx] = pkt
  90. }
  91. // shard range for current packet
  92. shardBegin := pkt.seqid - pkt.seqid%uint32(dec.shardSize)
  93. shardEnd := shardBegin + uint32(dec.shardSize) - 1
  94. // max search range in ordered queue for current shard
  95. searchBegin := insertIdx - int(pkt.seqid%uint32(dec.shardSize))
  96. if searchBegin < 0 {
  97. searchBegin = 0
  98. }
  99. searchEnd := searchBegin + dec.shardSize - 1
  100. if searchEnd >= len(dec.rx) {
  101. searchEnd = len(dec.rx) - 1
  102. }
  103. // re-construct datashards
  104. if searchEnd-searchBegin+1 >= dec.dataShards {
  105. var numshard, numDataShard, first, maxlen int
  106. // zero caches
  107. shards := dec.decodeCache
  108. shardsflag := dec.flagCache
  109. for k := range dec.decodeCache {
  110. shards[k] = nil
  111. shardsflag[k] = false
  112. }
  113. // shard assembly
  114. for i := searchBegin; i <= searchEnd; i++ {
  115. seqid := dec.rx[i].seqid
  116. if _itimediff(seqid, shardEnd) > 0 {
  117. break
  118. } else if _itimediff(seqid, shardBegin) >= 0 {
  119. shards[seqid%uint32(dec.shardSize)] = dec.rx[i].data
  120. shardsflag[seqid%uint32(dec.shardSize)] = true
  121. numshard++
  122. if dec.rx[i].flag == typeData {
  123. numDataShard++
  124. }
  125. if numshard == 1 {
  126. first = i
  127. }
  128. if len(dec.rx[i].data) > maxlen {
  129. maxlen = len(dec.rx[i].data)
  130. }
  131. }
  132. }
  133. if numDataShard == dec.dataShards {
  134. // case 1: no loss on data shards
  135. dec.rx = dec.freeRange(first, numshard, dec.rx)
  136. } else if numshard >= dec.dataShards {
  137. // case 2: loss on data shards, but it's recoverable from parity shards
  138. for k := range shards {
  139. if shards[k] != nil {
  140. dlen := len(shards[k])
  141. shards[k] = shards[k][:maxlen]
  142. copy(shards[k][dlen:], dec.zeros)
  143. }
  144. }
  145. if err := dec.codec.ReconstructData(shards); err == nil {
  146. for k := range shards[:dec.dataShards] {
  147. if !shardsflag[k] {
  148. recovered = append(recovered, shards[k])
  149. }
  150. }
  151. }
  152. dec.rx = dec.freeRange(first, numshard, dec.rx)
  153. }
  154. }
  155. // keep rxlimit
  156. if len(dec.rx) > dec.rxlimit {
  157. if dec.rx[0].flag == typeData { // track the unrecoverable data
  158. atomic.AddUint64(&DefaultSnmp.FECShortShards, 1)
  159. }
  160. dec.rx = dec.freeRange(0, 1, dec.rx)
  161. }
  162. return
  163. }
  164. // free a range of fecPacket, and zero for GC recycling
  165. func (dec *fecDecoder) freeRange(first, n int, q []fecPacket) []fecPacket {
  166. for i := first; i < first+n; i++ { // recycle buffer
  167. xmitBuf.Put(q[i].data)
  168. }
  169. copy(q[first:], q[first+n:])
  170. for i := 0; i < n; i++ { // dereference data
  171. q[len(q)-1-i].data = nil
  172. }
  173. return q[:len(q)-n]
  174. }
  175. type (
  176. // fecEncoder for encoding outgoing packets
  177. fecEncoder struct {
  178. dataShards int
  179. parityShards int
  180. shardSize int
  181. paws uint32 // Protect Against Wrapped Sequence numbers
  182. next uint32 // next seqid
  183. shardCount int // count the number of datashards collected
  184. maxSize int // track maximum data length in datashard
  185. headerOffset int // FEC header offset
  186. payloadOffset int // FEC payload offset
  187. // caches
  188. shardCache [][]byte
  189. encodeCache [][]byte
  190. // zeros
  191. zeros []byte
  192. // RS encoder
  193. codec reedsolomon.Encoder
  194. }
  195. )
  196. func newFECEncoder(dataShards, parityShards, offset int) *fecEncoder {
  197. if dataShards <= 0 || parityShards <= 0 {
  198. return nil
  199. }
  200. enc := new(fecEncoder)
  201. enc.dataShards = dataShards
  202. enc.parityShards = parityShards
  203. enc.shardSize = dataShards + parityShards
  204. enc.paws = (0xffffffff/uint32(enc.shardSize) - 1) * uint32(enc.shardSize)
  205. enc.headerOffset = offset
  206. enc.payloadOffset = enc.headerOffset + fecHeaderSize
  207. codec, err := reedsolomon.New(dataShards, parityShards)
  208. if err != nil {
  209. return nil
  210. }
  211. enc.codec = codec
  212. // caches
  213. enc.encodeCache = make([][]byte, enc.shardSize)
  214. enc.shardCache = make([][]byte, enc.shardSize)
  215. for k := range enc.shardCache {
  216. enc.shardCache[k] = make([]byte, mtuLimit)
  217. }
  218. enc.zeros = make([]byte, mtuLimit)
  219. return enc
  220. }
  221. // encodes the packet, outputs parity shards if we have collected quorum datashards
  222. // notice: the contents of 'ps' will be re-written in successive calling
  223. func (enc *fecEncoder) encode(b []byte) (ps [][]byte) {
  224. enc.markData(b[enc.headerOffset:])
  225. binary.LittleEndian.PutUint16(b[enc.payloadOffset:], uint16(len(b[enc.payloadOffset:])))
  226. // copy data to fec datashards
  227. sz := len(b)
  228. enc.shardCache[enc.shardCount] = enc.shardCache[enc.shardCount][:sz]
  229. copy(enc.shardCache[enc.shardCount], b)
  230. enc.shardCount++
  231. // track max datashard length
  232. if sz > enc.maxSize {
  233. enc.maxSize = sz
  234. }
  235. // Generation of Reed-Solomon Erasure Code
  236. if enc.shardCount == enc.dataShards {
  237. // fill '0' into the tail of each datashard
  238. for i := 0; i < enc.dataShards; i++ {
  239. shard := enc.shardCache[i]
  240. slen := len(shard)
  241. copy(shard[slen:enc.maxSize], enc.zeros)
  242. }
  243. // construct equal-sized slice with stripped header
  244. cache := enc.encodeCache
  245. for k := range cache {
  246. cache[k] = enc.shardCache[k][enc.payloadOffset:enc.maxSize]
  247. }
  248. // encoding
  249. if err := enc.codec.Encode(cache); err == nil {
  250. ps = enc.shardCache[enc.dataShards:]
  251. for k := range ps {
  252. enc.markFEC(ps[k][enc.headerOffset:])
  253. ps[k] = ps[k][:enc.maxSize]
  254. }
  255. }
  256. // counters resetting
  257. enc.shardCount = 0
  258. enc.maxSize = 0
  259. }
  260. return
  261. }
  262. func (enc *fecEncoder) markData(data []byte) {
  263. binary.LittleEndian.PutUint32(data, enc.next)
  264. binary.LittleEndian.PutUint16(data[4:], typeData)
  265. enc.next++
  266. }
  267. func (enc *fecEncoder) markFEC(data []byte) {
  268. binary.LittleEndian.PutUint32(data, enc.next)
  269. binary.LittleEndian.PutUint16(data[4:], typeFEC)
  270. enc.next = (enc.next + 1) % enc.paws
  271. }