raptorq/src/pi_solver.rs

1221 lines
43 KiB
Rust
Raw Normal View History

use crate::arraymap::{BoolArrayMap, UndirectedGraph};
use crate::arraymap::{U16ArrayMap, U32VecMap};
use crate::matrix::BinaryMatrix;
2019-03-23 01:25:07 +00:00
use crate::octet::Octet;
use crate::octet_matrix::DenseOctetMatrix;
2019-03-23 01:25:07 +00:00
use crate::symbol::Symbol;
use crate::systematic_constants::num_hdpc_symbols;
use crate::systematic_constants::num_intermediate_symbols;
use crate::systematic_constants::num_ldpc_symbols;
use crate::systematic_constants::num_pi_symbols;
use crate::util::get_both_indices;
use serde::{Deserialize, Serialize};
2020-01-13 00:56:45 +00:00
use std::mem::size_of;
2019-02-19 01:02:54 +00:00
#[derive(Clone, Debug, Serialize, Deserialize, PartialEq, PartialOrd, Eq, Ord, Hash)]
enum SymbolOps {
AddAssign {
dest: usize,
src: usize,
},
MulAssign {
dest: usize,
scalar: Octet,
},
FMA {
dest: usize,
src: usize,
scalar: Octet,
},
}
#[derive(Clone, Debug, Serialize, Deserialize, PartialEq, PartialOrd, Eq, Ord, Hash)]
2019-02-19 01:02:54 +00:00
struct FirstPhaseRowSelectionStats {
original_degree: U16ArrayMap,
ones_per_row: U16ArrayMap,
ones_histogram: U32VecMap,
2019-02-19 01:02:54 +00:00
start_col: usize,
end_col: usize,
2019-03-28 21:13:19 +00:00
start_row: usize,
rows_with_single_one: Vec<usize>,
2020-01-05 23:51:39 +00:00
// Scratch data struct that is reused across calls because it's expensive to construct
scratch_adjacent_nodes: UndirectedGraph,
2019-02-19 01:02:54 +00:00
}
impl FirstPhaseRowSelectionStats {
#[inline(never)]
#[allow(non_snake_case)]
pub fn new<T: BinaryMatrix>(matrix: &T, end_col: usize) -> FirstPhaseRowSelectionStats {
2019-02-19 01:02:54 +00:00
let mut result = FirstPhaseRowSelectionStats {
original_degree: U16ArrayMap::new(0, 0),
ones_per_row: U16ArrayMap::new(0, matrix.height()),
ones_histogram: U32VecMap::new(0),
2019-02-19 01:02:54 +00:00
start_col: 0,
end_col,
2019-03-28 21:13:19 +00:00
start_row: 0,
rows_with_single_one: vec![],
scratch_adjacent_nodes: UndirectedGraph::new(0, end_col),
2019-02-19 01:02:54 +00:00
};
for row in 0..matrix.height() {
let ones = matrix.count_ones(row, 0, end_col);
result.ones_per_row.insert(row, ones as u16);
result.ones_histogram.increment(ones);
if ones == 1 {
result.rows_with_single_one.push(row);
2019-04-09 02:08:09 +00:00
}
2019-02-19 01:02:54 +00:00
}
// Original degree is the degree of each row before processing begins
result.original_degree = result.ones_per_row.clone();
2019-02-19 01:02:54 +00:00
result
}
#[allow(dead_code)]
pub fn size_in_bytes(&self) -> usize {
let mut bytes = size_of::<Self>();
bytes += self.original_degree.size_in_bytes();
bytes += self.ones_per_row.size_in_bytes();
bytes += self.ones_histogram.size_in_bytes();
bytes += self.scratch_adjacent_nodes.size_in_bytes();
bytes
}
2019-02-19 01:02:54 +00:00
pub fn swap_rows(&mut self, i: usize, j: usize) {
self.ones_per_row.swap(i, j);
self.original_degree.swap(i, j);
for row in self.rows_with_single_one.iter_mut() {
2019-04-09 02:08:09 +00:00
if *row == i {
*row = j;
} else if *row == j {
2019-04-09 02:08:09 +00:00
*row = i;
}
}
2019-02-19 01:02:54 +00:00
}
// Recompute all stored statistics for the given row
pub fn recompute_row<T: BinaryMatrix>(&mut self, row: usize, matrix: &T) {
let ones = matrix.count_ones(row, self.start_col, self.end_col);
self.rows_with_single_one.retain(|x| *x != row);
if ones == 1 {
self.rows_with_single_one.push(row);
2019-04-09 02:08:09 +00:00
}
self.ones_histogram
.decrement(self.ones_per_row.get(row) as usize);
self.ones_histogram.increment(ones);
self.ones_per_row.insert(row, ones as u16);
2019-02-19 01:02:54 +00:00
}
pub fn eliminate_leading_value(&mut self, row: usize, value: &Octet) {
debug_assert_ne!(*value, Octet::zero());
debug_assert_eq!(*value, Octet::one());
self.ones_per_row.decrement(row);
let ones = self.ones_per_row.get(row);
if ones == 0 {
self.rows_with_single_one.retain(|x| *x != row);
} else if ones == 1 {
self.rows_with_single_one.push(row);
2019-02-19 01:02:54 +00:00
}
self.ones_histogram.decrement((ones + 1) as usize);
self.ones_histogram.increment(ones as usize);
2019-02-19 01:02:54 +00:00
}
// Set the valid columns, and recalculate statistics
// All values in column "start_col - 1" in rows start_row..end_row must be zero
2019-02-19 01:02:54 +00:00
#[inline(never)]
pub fn resize<T: BinaryMatrix>(
2019-03-28 21:13:19 +00:00
&mut self,
start_row: usize,
end_row: usize,
start_col: usize,
end_col: usize,
2019-04-06 04:18:25 +00:00
matrix: &T,
2019-03-28 21:13:19 +00:00
) {
2019-02-19 01:02:54 +00:00
// Only shrinking is supported
assert!(end_col <= self.end_col);
assert_eq!(self.start_row, start_row - 1);
assert_eq!(self.start_col, start_col - 1);
2019-02-19 01:02:54 +00:00
self.ones_histogram
.decrement(self.ones_per_row.get(self.start_row) as usize);
self.rows_with_single_one.retain(|x| *x != start_row - 1);
2019-02-19 01:02:54 +00:00
for col in end_col..self.end_col {
for row in matrix.get_col_index_iter(col, start_row, end_row) {
2019-02-19 01:02:54 +00:00
if matrix.get(row, col) == Octet::one() {
self.ones_per_row.decrement(row);
let ones = self.ones_per_row.get(row);
if ones == 0 {
self.rows_with_single_one.retain(|x| *x != row);
} else if ones == 1 {
self.rows_with_single_one.push(row);
2019-04-09 02:08:09 +00:00
}
self.ones_histogram.decrement((ones + 1) as usize);
self.ones_histogram.increment(ones as usize);
2019-02-19 01:02:54 +00:00
}
}
}
self.start_col = start_col;
self.end_col = end_col;
self.start_row = start_row;
}
#[inline(never)]
fn first_phase_graph_substep_build_adjacency<T: BinaryMatrix>(
2020-01-05 23:51:39 +00:00
&mut self,
2019-12-23 19:18:45 +00:00
rows_with_two_ones: &[usize],
2019-04-06 04:18:25 +00:00
matrix: &T,
2020-01-05 23:51:39 +00:00
) {
self.scratch_adjacent_nodes
.reset(self.start_col, self.end_col);
2019-02-19 01:02:54 +00:00
for row in rows_with_two_ones.iter() {
let mut ones = [0; 2];
let mut found = 0;
for (col, value) in matrix.get_row_iter(*row, self.start_col, self.end_col) {
2019-02-19 01:02:54 +00:00
// "The following graph defined by the structure of V is used in determining which
// row of A is chosen. The columns that intersect V are the nodes in the graph,
// and the rows that have exactly 2 nonzero entries in V and are not HDPC rows
// are the edges of the graph that connect the two columns (nodes) in the positions
// of the two ones."
// This part of the matrix is over GF(2), so "nonzero entries" is equivalent to "ones"
if value == Octet::one() {
2019-02-19 01:02:54 +00:00
ones[found] = col;
found += 1;
}
if found == 2 {
break;
}
}
2019-03-22 03:31:54 +00:00
assert_eq!(found, 2);
self.scratch_adjacent_nodes.add_edge(ones[0], ones[1]);
2019-02-19 01:02:54 +00:00
}
}
#[inline(never)]
fn first_phase_graph_substep<T: BinaryMatrix>(
2020-01-05 23:51:39 +00:00
&mut self,
2019-03-28 21:13:19 +00:00
start_row: usize,
end_row: usize,
2019-12-23 19:18:45 +00:00
rows_with_two_ones: &[usize],
2019-04-06 04:18:25 +00:00
matrix: &T,
2019-03-28 21:13:19 +00:00
) -> usize {
2020-01-05 23:51:39 +00:00
self.first_phase_graph_substep_build_adjacency(rows_with_two_ones, matrix);
let mut visited = BoolArrayMap::new(start_row, end_row);
let mut examplar_largest_component_node = None;
let mut largest_component_size = 0;
let mut node_queue = Vec::with_capacity(10);
2020-01-05 23:51:39 +00:00
for key in self.scratch_adjacent_nodes.keys() {
let mut component_size = 0;
// We can choose any edge (row) that connects this col to another in the graph
let mut examplar_node = None;
// Pick arbitrary node (column) to start
node_queue.clear();
node_queue.push(key);
while !node_queue.is_empty() {
let node = node_queue.pop().unwrap();
if visited.get(node) {
continue;
}
visited.insert(node, true);
component_size += 1;
for next_node in self.scratch_adjacent_nodes.get_adjacent_nodes(node) {
node_queue.push(*next_node);
examplar_node = Some(node);
2019-02-19 01:02:54 +00:00
}
}
2019-03-23 20:10:54 +00:00
if component_size > largest_component_size {
examplar_largest_component_node = examplar_node;
largest_component_size = component_size;
2019-02-19 01:02:54 +00:00
}
}
let node = examplar_largest_component_node.unwrap();
for row in matrix.get_col_index_iter(node, start_row, end_row) {
if matrix.get(row, node) == Octet::one() && self.ones_per_row.get(row) == 2 {
return row;
}
}
unreachable!();
2019-02-19 01:02:54 +00:00
}
#[inline(never)]
2019-03-28 21:13:19 +00:00
fn first_phase_original_degree_substep(
&self,
start_row: usize,
end_row: usize,
r: usize,
) -> usize {
// There's no need for special handling of HDPC rows, since Errata 2 guarantees we won't
// select any, and they're excluded in the first_phase solver
let mut chosen = None;
let mut chosen_original_degree = std::u16::MAX;
2019-04-09 02:08:09 +00:00
// Fast path for r=1, since this is super common
if r == 1 {
assert_ne!(0, self.rows_with_single_one.len());
for &row in self.rows_with_single_one.iter() {
let ones = self.ones_per_row.get(row);
2019-04-09 02:08:09 +00:00
let row_original_degree = self.original_degree.get(row);
if ones as usize == r && row_original_degree < chosen_original_degree {
chosen = Some(row);
chosen_original_degree = row_original_degree;
2019-04-09 02:08:09 +00:00
}
}
} else {
2019-04-09 02:08:09 +00:00
for row in start_row..end_row {
let ones = self.ones_per_row.get(row);
2019-04-09 02:08:09 +00:00
let row_original_degree = self.original_degree.get(row);
if ones as usize == r && row_original_degree < chosen_original_degree {
chosen = Some(row);
chosen_original_degree = row_original_degree;
2019-02-19 01:02:54 +00:00
}
}
}
return chosen.unwrap();
2019-02-19 01:02:54 +00:00
}
// Verify there there are no non-HPDC rows with exactly two non-zero entries, greater than one
#[inline(never)]
#[cfg(debug_assertions)]
2019-03-28 21:13:19 +00:00
fn first_phase_graph_substep_verify(
&self,
start_row: usize,
end_row: usize,
2019-12-23 19:18:45 +00:00
rows_with_two_ones: &[usize],
2019-03-28 21:13:19 +00:00
) {
2019-02-19 01:02:54 +00:00
for row in start_row..end_row {
if self.ones_per_row.get(row) == 2 {
assert!(rows_with_two_ones.contains(&row));
2019-02-19 01:02:54 +00:00
}
}
}
// Helper method for decoder phase 1
// selects from [start_row, end_row) reading [start_col, end_col)
// Returns (the chosen row, and "r" number of non-zero values the row has)
pub fn first_phase_selection<T: BinaryMatrix>(
2020-01-05 23:51:39 +00:00
&mut self,
2019-03-28 21:13:19 +00:00
start_row: usize,
end_row: usize,
2019-04-06 04:18:25 +00:00
matrix: &T,
) -> (Option<usize>, Option<usize>) {
2019-02-19 01:02:54 +00:00
let mut r = None;
2019-12-23 19:18:45 +00:00
for i in 1..=(self.end_col - self.start_col) {
if self.ones_histogram.get(i) > 0 {
2019-02-19 01:02:54 +00:00
r = Some(i);
break;
}
}
if r == None {
return (None, None);
2019-02-19 01:02:54 +00:00
}
if r.unwrap() == 2 {
2020-01-13 07:34:31 +00:00
// TODO: optimize to not allocate this Vec
2019-02-19 01:02:54 +00:00
let mut rows_with_two_ones = vec![];
for row in start_row..end_row {
let ones = self.ones_per_row.get(row);
if ones == 2 {
2019-02-19 01:02:54 +00:00
rows_with_two_ones.push(row);
}
}
// Paragraph starting "If r = 2 and there is no row with exactly 2 ones in V" can
// be ignored due to Errata 8.
assert!(!rows_with_two_ones.is_empty());
2019-02-19 01:02:54 +00:00
// See paragraph starting "If r = 2 and there is a row with exactly 2 ones in V..."
#[cfg(debug_assertions)]
self.first_phase_graph_substep_verify(start_row, end_row, &rows_with_two_ones);
let row =
self.first_phase_graph_substep(start_row, end_row, &rows_with_two_ones, matrix);
return (Some(row), r);
2019-03-28 21:13:19 +00:00
} else {
let row = self.first_phase_original_degree_substep(start_row, end_row, r.unwrap());
return (Some(row), r);
2019-02-19 01:02:54 +00:00
}
}
}
// See section 5.4.2.1
#[allow(non_snake_case)]
#[derive(Clone, Debug, Serialize, Deserialize, PartialEq, PartialOrd, Eq, Ord, Hash)]
pub struct IntermediateSymbolDecoder<T: BinaryMatrix> {
2019-04-06 04:18:25 +00:00
A: T,
// If present, these are treated as replacing the last rows of A
// Errata 3 guarantees that these do not need to be included in X
A_hdpc_rows: Option<DenseOctetMatrix>,
2019-04-06 04:18:25 +00:00
X: T,
2019-02-19 01:02:54 +00:00
D: Vec<Symbol>,
c: Vec<usize>,
d: Vec<usize>,
i: usize,
u: usize,
L: usize,
// Operations on D are deferred to the end of the codec to improve cache hits
deferred_D_ops: Vec<SymbolOps>,
2019-02-19 01:02:54 +00:00
num_source_symbols: u32,
debug_symbol_mul_ops: u32,
debug_symbol_add_ops: u32,
debug_symbol_mul_ops_by_phase: Vec<u32>,
2019-03-28 21:13:19 +00:00
debug_symbol_add_ops_by_phase: Vec<u32>,
2019-02-19 01:02:54 +00:00
}
#[allow(non_snake_case)]
impl<T: BinaryMatrix> IntermediateSymbolDecoder<T> {
2019-03-28 21:13:19 +00:00
pub fn new(
2019-04-06 04:18:25 +00:00
matrix: T,
hdpc_rows: DenseOctetMatrix,
2019-03-28 21:13:19 +00:00
symbols: Vec<Symbol>,
num_source_symbols: u32,
2019-04-06 04:18:25 +00:00
) -> IntermediateSymbolDecoder<T> {
2019-02-19 01:02:54 +00:00
assert!(matrix.width() <= symbols.len());
assert_eq!(matrix.height(), symbols.len());
let mut c = Vec::with_capacity(matrix.width());
let mut d = Vec::with_capacity(symbols.len());
for i in 0..matrix.width() {
c.push(i);
}
for i in 0..symbols.len() {
d.push(i);
}
let intermediate_symbols = num_intermediate_symbols(num_source_symbols) as usize;
let num_rows = matrix.height();
let pi_symbols = num_pi_symbols(num_source_symbols) as usize;
let mut A = matrix.clone();
A.enable_column_acccess_acceleration();
let mut X = matrix;
// Drop the PI symbols, since they will never be accessed in X. X will be resized to
// i-by-i in the second phase.
X.resize(X.height(), X.width() - pi_symbols);
let mut temp = IntermediateSymbolDecoder {
A,
A_hdpc_rows: None,
X,
2019-02-19 01:02:54 +00:00
D: symbols,
c,
d,
i: 0,
u: pi_symbols,
L: intermediate_symbols,
deferred_D_ops: Vec::with_capacity(70 * intermediate_symbols),
2019-02-19 01:02:54 +00:00
num_source_symbols,
debug_symbol_mul_ops: 0,
debug_symbol_add_ops: 0,
debug_symbol_mul_ops_by_phase: vec![0; 5],
2019-03-28 21:13:19 +00:00
debug_symbol_add_ops_by_phase: vec![0; 5],
};
// Swap the HDPC rows, so that they're the last in the matrix
let S = num_ldpc_symbols(num_source_symbols) as usize;
let H = num_hdpc_symbols(num_source_symbols) as usize;
// See section 5.3.3.4.2, Figure 5.
for i in 0..H {
temp.swap_rows(S + i, num_rows - H + i);
temp.X.swap_rows(S + i, num_rows - H + i);
2019-02-19 01:02:54 +00:00
}
temp.A_hdpc_rows = Some(hdpc_rows);
temp
2019-02-19 01:02:54 +00:00
}
#[inline(never)]
fn apply_deferred_symbol_ops(&mut self) {
for op in self.deferred_D_ops.drain(..) {
match op {
SymbolOps::AddAssign { dest, src } => {
let (dest, temp) = get_both_indices(&mut self.D, dest, src);
*dest += temp;
}
SymbolOps::MulAssign { dest, scalar } => {
self.D[dest].mulassign_scalar(&scalar);
}
SymbolOps::FMA { dest, src, scalar } => {
let (dest, temp) = get_both_indices(&mut self.D, dest, src);
dest.fused_addassign_mul_scalar(&temp, &scalar);
}
}
}
}
2019-02-19 01:02:54 +00:00
// Returns true iff all elements in A between [start_row, end_row)
// and [start_column, end_column) are zero
#[cfg(debug_assertions)]
2019-03-28 21:13:19 +00:00
fn all_zeroes(
&self,
start_row: usize,
end_row: usize,
start_column: usize,
end_column: usize,
) -> bool {
2019-02-19 01:02:54 +00:00
for row in start_row..end_row {
for column in start_column..end_column {
if self.get_A_value(row, column) != Octet::zero() {
2019-02-19 01:02:54 +00:00
return false;
}
}
}
return true;
}
#[cfg(debug_assertions)]
fn get_A_value(&self, row: usize, col: usize) -> Octet {
if let Some(ref hdpc) = self.A_hdpc_rows {
if row >= self.A.height() - hdpc.height() {
return hdpc.get(row - (self.A.height() - hdpc.height()), col);
}
}
return self.A.get(row, col);
}
2019-02-19 01:02:54 +00:00
// Performs the column swapping substep of first phase, after the row has been chosen
#[inline(never)]
fn first_phase_swap_columns_substep(&mut self, r: usize) {
let mut swapped_columns = 0;
// Fast path when r == 1, since this is very common
if r == 1 {
// self.i will never reference an HDPC row, so can ignore self.A_hdpc_rows
// because of Errata 2.
for (col, value) in self
.A
.get_row_iter(self.i, self.i, self.A.width() - self.u)
.clone()
{
if value != Octet::zero() {
// No need to swap the first i rows, as they are all zero (see submatrix above V)
self.swap_columns(self.i, col, self.i);
// Also apply to X
self.X.swap_columns(self.i, col, 0);
2019-04-14 23:51:38 +00:00
swapped_columns += 1;
break;
2019-02-19 01:02:54 +00:00
}
}
} else {
2019-04-14 23:51:38 +00:00
for col in self.i..(self.A.width() - self.u) {
// self.i will never reference an HDPC row, so can ignore self.A_hdpc_rows
// because of Errata 2.
if self.A.get(self.i, col) != Octet::zero() {
let mut dest;
if swapped_columns == 0 {
dest = self.i;
} else {
dest = self.A.width() - self.u - swapped_columns;
// Some of the right most columns may already contain non-zeros
while self.A.get(self.i, dest) != Octet::zero() {
dest -= 1;
swapped_columns += 1;
}
}
2019-04-14 23:51:38 +00:00
if swapped_columns == r {
break;
}
// No need to swap the first i rows, as they are all zero (see submatrix above V)
self.swap_columns(dest, col, self.i);
// Also apply to X
self.X.swap_columns(dest, col, 0);
swapped_columns += 1;
if swapped_columns == r {
break;
}
2019-02-19 01:02:54 +00:00
}
}
}
2019-04-14 23:51:38 +00:00
assert_eq!(r, swapped_columns);
2019-02-19 01:02:54 +00:00
}
// First phase (section 5.4.2.2)
#[allow(non_snake_case)]
#[inline(never)]
fn first_phase(&mut self) -> bool {
// First phase (section 5.4.2.2)
// ----------> i u <--------
// | +-----------+-----------------+---------+
// | | | | |
// | | I | All Zeros | |
// v | | | |
// i +-----------+-----------------+ U |
// | | | |
// | | | |
// | All Zeros | V | |
// | | | |
// | | | |
// +-----------+-----------------+---------+
// Figure 6: Submatrices of A in the First Phase
let num_hdpc_rows = self.A_hdpc_rows.as_ref().unwrap().height();
let mut selection_helper =
FirstPhaseRowSelectionStats::new(&self.A, self.A.width() - self.u);
2019-02-19 01:02:54 +00:00
while self.i + self.u < self.L {
// Calculate r
// "Let r be the minimum integer such that at least one row of A has
// exactly r nonzeros in V."
2020-01-11 22:27:29 +00:00
// Exclude the HDPC rows, since Errata 2 guarantees they won't be chosen.
let (chosen_row, r) = selection_helper.first_phase_selection(
self.i,
self.A.height() - num_hdpc_rows,
&self.A,
);
2019-02-19 01:02:54 +00:00
if r == None {
return false;
}
let r = r.unwrap();
let chosen_row = chosen_row.unwrap();
2019-04-14 23:51:38 +00:00
assert!(chosen_row >= self.i);
2019-02-19 01:02:54 +00:00
// See paragraph beginning: "After the row is chosen in this step..."
// Reorder rows
let temp = self.i;
self.swap_rows(temp, chosen_row);
self.X.swap_rows(temp, chosen_row);
selection_helper.swap_rows(temp, chosen_row);
// Reorder columns
self.first_phase_swap_columns_substep(r);
// Zero out leading value in following rows
let temp = self.i;
// self.i will never reference an HDPC row, so can ignore self.A_hdpc_rows
// because of Errata 2.
2019-04-10 04:30:25 +00:00
let temp_value = self.A.get(temp, temp);
2019-04-12 00:51:18 +00:00
// Cloning the iterator is safe here, because we don't re-read any of the rows that
// we add to
for row in self
.A
.get_col_index_iter(temp, self.i + 1, self.A.height() - num_hdpc_rows)
.clone()
{
2019-02-19 01:02:54 +00:00
let leading_value = self.A.get(row, temp);
if leading_value != Octet::zero() {
// Addition is equivalent to subtraction
2019-04-10 04:30:25 +00:00
let beta = &leading_value / &temp_value;
2019-02-19 01:02:54 +00:00
self.fma_rows(temp, row, beta);
if r == 1 {
// Hot path for r == 1, since it's very common due to maximum connected
// component selection, and recompute_row() is expensive
selection_helper.eliminate_leading_value(row, &leading_value);
2019-03-28 21:13:19 +00:00
} else {
2019-02-19 01:02:54 +00:00
selection_helper.recompute_row(row, &self.A);
}
}
}
for i in 0..(r - 1) {
self.A
.hint_column_dense_and_frozen(self.A.width() - self.u - 1 - i);
}
// apply to hdpc rows as well, which are stored separately
let pi_octets = self
.A
.get_sub_row_as_octets(temp, self.A.width() - (self.u + r - 1));
for row in 0..num_hdpc_rows {
let leading_value = self.A_hdpc_rows.as_ref().unwrap().get(row, temp);
if leading_value != Octet::zero() {
// Addition is equivalent to subtraction
let beta = &leading_value / &temp_value;
self.fma_rows_with_pi(
temp,
row + (self.A.height() - num_hdpc_rows),
beta,
// self.i is the only non-PI column which can have a nonzero,
// since all the rest were column swapped into the PI submatrix.
Some(temp),
Some(&pi_octets),
);
// It's safe to skip updating the selection helper, since it will never
// select an HDPC row
}
}
2019-02-19 01:02:54 +00:00
self.i += 1;
self.u += r - 1;
2019-03-28 21:13:19 +00:00
selection_helper.resize(
self.i,
self.A.height() - self.A_hdpc_rows.as_ref().unwrap().height(),
2019-03-28 21:13:19 +00:00
self.i,
self.A.width() - self.u,
&self.A,
);
2019-02-19 01:02:54 +00:00
#[cfg(debug_assertions)]
self.first_phase_verify();
}
self.record_symbol_ops(0);
return true;
}
// See section 5.4.2.2. Verifies the two all-zeros submatrices and the identity submatrix
#[inline(never)]
#[cfg(debug_assertions)]
fn first_phase_verify(&self) {
for row in 0..self.i {
for col in 0..self.i {
if row == col {
assert_eq!(Octet::one(), self.A.get(row, col));
2019-03-28 21:13:19 +00:00
} else {
2019-02-19 01:02:54 +00:00
assert_eq!(Octet::zero(), self.A.get(row, col));
}
}
}
assert!(self.all_zeroes(0, self.i, self.i, self.A.width() - self.u));
assert!(self.all_zeroes(self.i, self.A.height(), 0, self.i));
}
// Second phase (section 5.4.2.3)
#[allow(non_snake_case)]
#[inline(never)]
fn second_phase(&mut self) -> bool {
#[cfg(debug_assertions)]
self.second_phase_verify();
self.X.resize(self.i, self.i);
// Convert U_lower to row echelon form
let temp = self.i;
let size = self.u;
// HDPC rows can be removed, since they can't have been selected for U_upper
let hdpc_rows = self.A_hdpc_rows.take().unwrap();
if let Some(submatrix) = self.record_reduce_to_row_echelon(hdpc_rows, temp, temp, size) {
// Perform backwards elimination
self.backwards_elimination(submatrix, temp, temp, size);
} else {
2019-02-19 01:02:54 +00:00
return false;
}
self.A.resize(self.L, self.L);
self.record_symbol_ops(1);
return true;
}
// Verifies that X is lower triangular. See section 5.4.2.3
#[inline(never)]
#[cfg(debug_assertions)]
fn second_phase_verify(&self) {
for row in 0..self.i {
for col in (row + 1)..self.i {
assert_eq!(Octet::zero(), self.X.get(row, col));
}
}
}
// Third phase (section 5.4.2.4)
#[allow(non_snake_case)]
#[inline(never)]
fn third_phase(&mut self) {
#[cfg(debug_assertions)]
self.third_phase_verify();
// A[0..i][..] = X * A[0..i][..]
self.A.mul_assign_submatrix(&self.X, self.i);
// Now apply the same operations to D.
// Note that X is lower triangular, so the row must be processed last to first
for row in (0..self.i).rev() {
if self.X.get(row, row) != Octet::one() {
self.debug_symbol_mul_ops += 1;
self.deferred_D_ops.push(SymbolOps::MulAssign {
dest: self.d[row],
scalar: self.X.get(row, row),
});
2019-02-19 01:02:54 +00:00
}
for (col, value) in self.X.get_row_iter(row, 0, row) {
if value == Octet::zero() {
2019-02-19 01:02:54 +00:00
continue;
}
if value == Octet::one() {
2019-02-19 01:02:54 +00:00
self.debug_symbol_add_ops += 1;
self.deferred_D_ops.push(SymbolOps::AddAssign {
dest: self.d[row],
src: self.d[col],
});
2019-03-28 21:13:19 +00:00
} else {
2019-02-19 01:02:54 +00:00
self.debug_symbol_mul_ops += 1;
self.debug_symbol_add_ops += 1;
self.deferred_D_ops.push(SymbolOps::FMA {
dest: self.d[row],
src: self.d[col],
scalar: self.X.get(row, col),
});
2019-02-19 01:02:54 +00:00
}
}
}
self.record_symbol_ops(2);
#[cfg(debug_assertions)]
self.third_phase_verify_end();
}
#[inline(never)]
#[cfg(debug_assertions)]
fn third_phase_verify(&self) {
for row in 0..self.A.height() {
for col in 0..self.A.width() {
if row < self.i && col >= self.A.width() - self.u {
// element is in U_upper, which can have arbitrary values at this point
continue;
}
// The rest of A should be identity matrix
if row == col {
assert_eq!(Octet::one(), self.A.get(row, col));
2019-03-28 21:13:19 +00:00
} else {
2019-02-19 01:02:54 +00:00
assert_eq!(Octet::zero(), self.A.get(row, col));
}
}
}
}
#[inline(never)]
#[cfg(debug_assertions)]
fn third_phase_verify_end(&self) {
for row in 0..self.i {
for col in 0..self.i {
assert_eq!(self.X.get(row, col), self.A.get(row, col));
}
}
}
// Fourth phase (section 5.4.2.5)
#[allow(non_snake_case)]
#[inline(never)]
fn fourth_phase(&mut self) {
for i in 0..self.i {
// TODO: optimize for sparse
2019-02-19 01:02:54 +00:00
for j in 0..self.u {
let b = self.A.get(i, j + self.i);
if b != Octet::zero() {
let temp = self.i;
self.fma_rows(temp + j, i, b);
}
}
}
self.record_symbol_ops(3);
#[cfg(debug_assertions)]
self.fourth_phase_verify();
}
#[inline(never)]
#[cfg(debug_assertions)]
fn fourth_phase_verify(&self) {
// ---------> i u <------
// | +-----------+--------+
// | |\ | |
// | | \ Zeros | Zeros |
// v | \ | |
// i | X \ | |
// u +---------- +--------+
// ^ | | |
// | | All Zeros | I |
// | | | |
// +-----------+--------+
// Same assertion about X being equal to the upper left of A
#[cfg(debug_assertions)]
self.third_phase_verify_end();
assert!(self.all_zeroes(0, self.i, self.A.width() - self.u, self.A.width()));
assert!(self.all_zeroes(self.A.height() - self.u, self.A.height(), 0, self.i));
for row in (self.A.height() - self.u)..self.A.height() {
for col in (self.A.width() - self.u)..self.A.width() {
if row == col {
assert_eq!(Octet::one(), self.A.get(row, col));
2019-03-28 21:13:19 +00:00
} else {
2019-02-19 01:02:54 +00:00
assert_eq!(Octet::zero(), self.A.get(row, col));
}
}
}
}
// Fifth phase (section 5.4.2.6)
#[allow(non_snake_case)]
#[inline(never)]
fn fifth_phase(&mut self) {
// "For j from 1 to i". Note that A is 1-indexed in the spec, and ranges are inclusive,
// this is means [1, i], which is equal to [0, i)
for j in 0..self.i as usize {
2020-01-11 22:39:14 +00:00
// Skip normalizing the diagonal, since there can't be non-binary values due to
// Errata 7
2019-02-19 01:02:54 +00:00
// "For l from 1 to j-1". This means the lower triangular columns, not including the
// diagonal, which is [0, j)
for (l, _) in self.A.get_row_iter(j, 0, j).clone() {
let temp = self.A.get(j, l);
2019-02-19 01:02:54 +00:00
if temp != Octet::zero() {
self.fma_rows(l, j, temp);
}
}
}
self.record_symbol_ops(4);
#[cfg(debug_assertions)]
self.fifth_phase_verify();
}
#[inline(never)]
#[cfg(debug_assertions)]
fn fifth_phase_verify(&self) {
assert_eq!(self.L, self.A.height());
for row in 0..self.A.height() {
assert_eq!(self.L, self.A.width());
for col in 0..self.A.width() {
if row == col {
assert_eq!(Octet::one(), self.A.get(row, col));
2019-03-28 21:13:19 +00:00
} else {
2019-02-19 01:02:54 +00:00
assert_eq!(Octet::zero(), self.A.get(row, col));
}
}
}
}
fn record_symbol_ops(&mut self, phase: usize) {
self.debug_symbol_add_ops_by_phase[phase] = self.debug_symbol_add_ops;
self.debug_symbol_mul_ops_by_phase[phase] = self.debug_symbol_mul_ops;
for i in 0..phase {
self.debug_symbol_add_ops_by_phase[phase] -= self.debug_symbol_add_ops_by_phase[i];
self.debug_symbol_mul_ops_by_phase[phase] -= self.debug_symbol_mul_ops_by_phase[i];
}
}
// Reduces the size x size submatrix, starting at row_offset and col_offset as the upper left
// corner, to row echelon form.
// Returns the reduced submatrix, which should be written back into this submatrix of A.
// The state of this submatrix in A is undefined, after calling this function.
2019-02-19 01:02:54 +00:00
#[inline(never)]
// TODO: optimize for sparse
fn record_reduce_to_row_echelon(
&mut self,
hdpc_rows: DenseOctetMatrix,
row_offset: usize,
col_offset: usize,
size: usize,
) -> Option<DenseOctetMatrix> {
// Copy U_lower into a new matrix and merge it with the HDPC rows
let mut submatrix = DenseOctetMatrix::new(self.A.height() - row_offset, size, 0);
let first_hdpc_row = self.A.height() - hdpc_rows.height();
for row in row_offset..self.A.height() {
for col in col_offset..(col_offset + size) {
let value = if row < first_hdpc_row {
self.A.get(row, col)
} else {
hdpc_rows.get(row - first_hdpc_row, col)
};
submatrix.set(row - row_offset, col - col_offset, value);
}
}
2019-02-19 01:02:54 +00:00
for i in 0..size {
// Swap a row with leading coefficient i into place
for j in i..submatrix.height() {
if submatrix.get(j, i) != Octet::zero() {
submatrix.swap_rows(i, j);
// Record the swap, in addition to swapping in the working submatrix
// TODO: optimize to not perform op on A
self.swap_rows(row_offset + i, j + row_offset);
2019-02-19 01:02:54 +00:00
break;
}
}
if submatrix.get(i, i) == Octet::zero() {
2019-02-19 01:02:54 +00:00
// If all following rows are zero in this column, then matrix is singular
return None;
2019-02-19 01:02:54 +00:00
}
// Scale leading coefficient to 1
if submatrix.get(i, i) != Octet::one() {
let element_inverse = Octet::one() / submatrix.get(i, i);
submatrix.mul_assign_row(i, &element_inverse);
// Record the multiplication, in addition to multiplying the working submatrix
self.record_mul_row(row_offset + i, element_inverse);
2019-02-19 01:02:54 +00:00
}
// Zero out all following elements in i'th column
for j in (i + 1)..submatrix.height() {
if submatrix.get(j, i) != Octet::zero() {
let scalar = submatrix.get(j, i);
submatrix.fma_rows(j, i, &scalar);
// Record the FMA, in addition to applying it to the working submatrix
self.record_fma_rows(row_offset + i, row_offset + j, scalar);
2019-02-19 01:02:54 +00:00
}
}
}
return Some(submatrix);
2019-02-19 01:02:54 +00:00
}
// Performs backwards elimination in a size x size submatrix, starting at
// row_offset and col_offset as the upper left corner of the submatrix
#[inline(never)]
// TODO: optimize for sparse
// Applies the submatrix to the size-by-size lower right of A, and performs backwards
// elimination on it. "submatrix" must be in row echelon form.
fn backwards_elimination(
&mut self,
submatrix: DenseOctetMatrix,
row_offset: usize,
col_offset: usize,
size: usize,
) {
2019-02-19 01:02:54 +00:00
// Perform backwards elimination
for i in (0..size).rev() {
// Zero out all preceding elements in i'th column
for j in 0..i {
if submatrix.get(j, i) != Octet::zero() {
let scalar = submatrix.get(j, i);
// Record the FMA. No need to actually apply it to the submatrix,
// since it will be discarded, and we never read these values
self.record_fma_rows(row_offset + i, row_offset + j, scalar);
2019-02-19 01:02:54 +00:00
}
}
}
// Write the identity matrix into A, since that's the resulting value of this function
for row in row_offset..(row_offset + size) {
for col in col_offset..(col_offset + size) {
if row == col {
self.A.set(row, col, Octet::one());
} else {
self.A.set(row, col, Octet::zero());
}
}
}
2019-02-19 01:02:54 +00:00
}
#[allow(dead_code)]
pub fn get_symbol_mul_ops(&self) -> u32 {
self.debug_symbol_mul_ops
}
#[allow(dead_code)]
pub fn get_symbol_add_ops(&self) -> u32 {
self.debug_symbol_add_ops
}
#[allow(dead_code)]
pub fn get_symbol_mul_ops_by_phase(&self) -> Vec<u32> {
self.debug_symbol_mul_ops_by_phase.clone()
}
#[allow(dead_code)]
pub fn get_symbol_add_ops_by_phase(&self) -> Vec<u32> {
self.debug_symbol_add_ops_by_phase.clone()
}
2020-01-13 00:56:45 +00:00
#[cfg(feature = "benchmarking")]
pub fn get_non_symbol_bytes(&self) -> usize {
let mut bytes = size_of::<Self>();
bytes += self.A.size_in_bytes();
if let Some(ref hdpc) = self.A_hdpc_rows {
bytes += hdpc.size_in_bytes();
}
bytes += self.X.size_in_bytes();
// Skip self.D, since we're calculating non-Symbol bytes
bytes += size_of::<usize>() * self.c.len();
bytes += size_of::<usize>() * self.d.len();
bytes
}
// Record operation to apply operations to D.
fn record_mul_row(&mut self, i: usize, beta: Octet) {
2019-02-19 01:02:54 +00:00
self.debug_symbol_mul_ops += 1;
self.deferred_D_ops.push(SymbolOps::MulAssign {
dest: self.d[i],
scalar: beta,
});
assert!(self.A_hdpc_rows.is_none());
2019-02-19 01:02:54 +00:00
}
fn fma_rows(&mut self, i: usize, iprime: usize, beta: Octet) {
self.fma_rows_with_pi(i, iprime, beta, None, None);
}
fn record_fma_rows(&mut self, i: usize, iprime: usize, beta: Octet) {
2019-02-19 01:02:54 +00:00
if beta == Octet::one() {
self.debug_symbol_add_ops += 1;
self.deferred_D_ops.push(SymbolOps::AddAssign {
dest: self.d[iprime],
src: self.d[i],
});
2019-03-28 21:13:19 +00:00
} else {
2019-02-19 01:02:54 +00:00
self.debug_symbol_add_ops += 1;
self.debug_symbol_mul_ops += 1;
self.deferred_D_ops.push(SymbolOps::FMA {
dest: self.d[iprime],
src: self.d[i],
scalar: beta,
});
2019-02-19 01:02:54 +00:00
}
}
fn fma_rows_with_pi(
&mut self,
i: usize,
iprime: usize,
beta: Octet,
only_non_pi_nonzero_column: Option<usize>,
pi_octets: Option<&Vec<u8>>,
) {
self.record_fma_rows(i, iprime, beta.clone());
if let Some(ref mut hdpc) = self.A_hdpc_rows {
let first_hdpc_row = self.A.height() - hdpc.height();
// Adding HDPC rows to other rows isn't supported, since it should never happen
assert!(i < first_hdpc_row);
if iprime >= first_hdpc_row {
let col = only_non_pi_nonzero_column.unwrap();
let multiplicand = self.A.get(i, col);
let mut value = hdpc.get(iprime - first_hdpc_row, col);
value.fma(&multiplicand, &beta);
hdpc.set(iprime - first_hdpc_row, col, value);
// Handle this part separately, since it's in the dense U part of the matrix
let octets = pi_octets.unwrap();
hdpc.fma_sub_row(
iprime - first_hdpc_row,
self.A.width() - octets.len(),
&beta,
octets,
);
} else {
assert_eq!(&beta, &Octet::one());
self.A.add_assign_rows(iprime, i);
}
} else {
assert_eq!(&beta, &Octet::one());
self.A.add_assign_rows(iprime, i);
}
2019-02-19 01:02:54 +00:00
}
fn swap_rows(&mut self, i: usize, iprime: usize) {
if let Some(ref hdpc_rows) = self.A_hdpc_rows {
// Can't swap HDPC rows
assert!(i < self.A.height() - hdpc_rows.height());
assert!(iprime < self.A.height() - hdpc_rows.height());
}
2019-02-19 01:02:54 +00:00
self.A.swap_rows(i, iprime);
self.d.swap(i, iprime);
}
fn swap_columns(&mut self, j: usize, jprime: usize, start_row: usize) {
self.A.swap_columns(j, jprime, start_row);
self.A_hdpc_rows
.as_mut()
.unwrap()
.swap_columns(j, jprime, 0);
2019-02-19 01:02:54 +00:00
self.c.swap(j, jprime);
}
#[inline(never)]
pub fn execute(&mut self) -> Option<Vec<Symbol>> {
2020-01-06 06:33:30 +00:00
self.X.disable_column_acccess_acceleration();
2019-02-19 01:02:54 +00:00
if !self.first_phase() {
2019-03-28 21:13:19 +00:00
return None;
2019-02-19 01:02:54 +00:00
}
2019-04-08 04:02:06 +00:00
self.A.disable_column_acccess_acceleration();
2019-02-19 01:02:54 +00:00
if !self.second_phase() {
return None;
}
self.third_phase();
self.fourth_phase();
self.fifth_phase();
self.apply_deferred_symbol_ops();
2019-02-19 01:02:54 +00:00
// See end of section 5.4.2.1
let mut index_mapping = vec![0; self.L];
2019-02-19 01:02:54 +00:00
for i in 0..self.L {
index_mapping[self.c[i]] = self.d[i];
2019-02-19 01:02:54 +00:00
}
2020-01-06 02:33:25 +00:00
#[allow(non_snake_case)]
let mut removable_D: Vec<Option<Symbol>> = self.D.drain(..).map(Some).collect();
2019-02-19 01:02:54 +00:00
let mut result = Vec::with_capacity(self.L);
#[allow(clippy::needless_range_loop)]
2019-02-19 01:02:54 +00:00
for i in 0..self.L {
2020-01-06 02:33:25 +00:00
// push a None so it can be swapped in
removable_D.push(None);
result.push(removable_D.swap_remove(index_mapping[i]).unwrap());
2019-02-19 01:02:54 +00:00
}
Some(result)
}
}
// Fused implementation for self.inverse().mul_symbols(symbols)
// See section 5.4.2.1
pub fn fused_inverse_mul_symbols<T: BinaryMatrix>(
2019-04-06 04:18:25 +00:00
matrix: T,
hdpc_rows: DenseOctetMatrix,
2019-03-28 21:13:19 +00:00
symbols: Vec<Symbol>,
num_source_symbols: u32,
) -> Option<Vec<Symbol>> {
IntermediateSymbolDecoder::new(matrix, hdpc_rows, symbols, num_source_symbols).execute()
2019-02-19 01:02:54 +00:00
}
#[cfg(test)]
mod tests {
use super::IntermediateSymbolDecoder;
2019-03-23 01:25:07 +00:00
use crate::constraint_matrix::generate_constraint_matrix;
use crate::matrix::BinaryMatrix;
use crate::matrix::DenseBinaryMatrix;
2019-03-23 01:25:07 +00:00
use crate::symbol::Symbol;
2020-01-10 07:06:45 +00:00
use crate::systematic_constants::{
extended_source_block_symbols, num_ldpc_symbols, num_lt_symbols,
MAX_SOURCE_SYMBOLS_PER_BLOCK,
};
2019-02-19 01:02:54 +00:00
#[test]
fn operations_per_symbol() {
2019-03-28 21:13:19 +00:00
for &(elements, expected_mul_ops, expected_add_ops) in
[(10, 35.0, 50.0), (100, 16.0, 35.0)].iter()
{
2019-03-23 20:10:54 +00:00
let num_symbols = extended_source_block_symbols(elements);
2019-02-19 01:02:54 +00:00
let indices: Vec<u32> = (0..num_symbols).collect();
let (a, hdpc) = generate_constraint_matrix::<DenseBinaryMatrix>(num_symbols, &indices);
let symbols = vec![Symbol::zero(1usize); a.width()];
let mut decoder = IntermediateSymbolDecoder::new(a, hdpc, symbols, num_symbols);
2019-02-19 01:02:54 +00:00
decoder.execute();
2019-03-28 21:13:19 +00:00
assert!(
(decoder.get_symbol_mul_ops() as f64 / num_symbols as f64) < expected_mul_ops,
"mul ops per symbol = {}",
(decoder.get_symbol_mul_ops() as f64 / num_symbols as f64)
);
assert!(
(decoder.get_symbol_add_ops() as f64 / num_symbols as f64) < expected_add_ops,
"add ops per symbol = {}",
(decoder.get_symbol_add_ops() as f64 / num_symbols as f64)
);
2019-02-19 01:02:54 +00:00
}
}
2020-01-10 07:06:45 +00:00
#[test]
fn check_errata_3() {
// Check that the optimization of excluding HDPC rows from the X matrix during decoding is
// safe. This is described in RFC6330_ERRATA.md
for i in 0..=MAX_SOURCE_SYMBOLS_PER_BLOCK {
assert!(extended_source_block_symbols(i) + num_ldpc_symbols(i) >= num_lt_symbols(i));
}
}
2019-02-19 01:02:54 +00:00
}