Lines Matching full:v7

230 CPU_LE(	rev64		v7.16b, v7.16b			)
238 CPU_LE( ext v7.16b, v7.16b, v7.16b, #8 )
252 // While >= 128 data bytes remain (not counting v0-v7), fold the 128
253 // bytes v0-v7 into them, storing the result back into v0-v7.
258 fold_32_bytes \p, v6, v7
263 // Now fold the 112 bytes in v0-v6 into the 16 bytes in v7.
271 fold_16_bytes \p, v3, v7, 1
274 fold_16_bytes \p, v5, v7, 1
276 fold_16_bytes \p, v6, v7
279 // (not counting v7), following the previous extra subtraction by 128.
284 // While >= 16 data bytes remain (not counting v7), fold the 16 bytes v7
285 // into them, storing the result back into v7.
288 pmull16x64_\p fold_consts, v7, v8
289 eor v7.16b, v7.16b, v8.16b
293 eor v7.16b, v7.16b, v0.16b
299 // (not counting v7), following the previous extra subtraction by 16.
305 // 16 bytes are in v7 and the rest are the remaining data in 'buf'. To
316 // v1 = high order part of second chunk: v7 left-shifted by 'len' bytes.
320 tbl v1.16b, {v7.16b}, v2.16b
322 // v3 = first chunk: v7 right-shifted by '16-len' bytes.
325 tbl v3.16b, {v7.16b}, v2.16b
334 // Fold the first chunk into the second chunk, storing the result in v7.
336 eor v7.16b, v3.16b, v0.16b
337 eor v7.16b, v7.16b, v2.16b
347 CPU_LE( rev64 v7.16b, v7.16b )
348 CPU_LE( ext v7.16b, v7.16b, v7.16b, #8 )
353 eor v7.16b, v7.16b, v0.16b
385 CPU_LE( rev64 v7.16b, v7.16b )
386 CPU_LE( ext v7.16b, v7.16b, v7.16b, #8 )
402 // Reduce the 128-bit value M(x), stored in v7, to the final 16-bit CRC.
412 ext v0.16b, v2.16b, v7.16b, #8
413 pmull2 v7.1q, v7.2d, fold_consts.2d // high bits * x^48 * (x^80 mod G(x))
414 eor v0.16b, v0.16b, v7.16b // + low bits * x^64