Fast Erasure Coding for Data Storage: A Comprehensive Study of the Acceleration Techniques
This paper integrates various optimization for ECs as a computation train. The procedure is: use a bitmatrix to produce the computation schedule (XOR level vectorization, XOR reduction and caching). Results also suggest that vectorizing XOR is a better choice than directly vectorizing finite field operations.
Normalization of parity coding matrix to make it more suitable for computation.
Reusing parity computation to reduce overall computation. Introduced by Plank
Maximum caldinality unweighted/weighted matching algorithm
Caching optimization
Vectorized XOR
Performance of individual techniques are compared. It shows that V-XOR appears to be able to provide the most significant performance improvement, with avg 130.04%. But it’s completely different computation chain at all. The remaining techs have improvements ranging from 4.81% to 36.63% individually.
Different 8 combinations of stratigies under optimized BM are compared. Total number of XORs for each combination of acceleration methods with different (n, k, w) are compared.
N/A