Lines Matching full:dr
30 int16_t* dr = data + samples * 2; in xnn_cs16_fftr_ukernel__scalar_x4() local
40 dr[0] = vdcr - vdci; in xnn_cs16_fftr_ukernel__scalar_x4()
41 dr[1] = 0; in xnn_cs16_fftr_ukernel__scalar_x4()
46 dr -= 4 * 2; in xnn_cs16_fftr_ukernel__scalar_x4()
55 int32_t virr0 = (int32_t) dr[6]; in xnn_cs16_fftr_ukernel__scalar_x4()
56 int32_t viri0 = -(int32_t) dr[7]; in xnn_cs16_fftr_ukernel__scalar_x4()
57 int32_t virr1 = (int32_t) dr[4]; in xnn_cs16_fftr_ukernel__scalar_x4()
58 int32_t viri1 = -(int32_t) dr[5]; in xnn_cs16_fftr_ukernel__scalar_x4()
59 int32_t virr2 = (int32_t) dr[2]; in xnn_cs16_fftr_ukernel__scalar_x4()
60 int32_t viri2 = -(int32_t) dr[3]; in xnn_cs16_fftr_ukernel__scalar_x4()
61 int32_t virr3 = (int32_t) dr[0]; in xnn_cs16_fftr_ukernel__scalar_x4()
62 int32_t viri3 = -(int32_t) dr[1]; in xnn_cs16_fftr_ukernel__scalar_x4()
123 dr[6] = math_asr_s32(vacc1r0 - twr0, 1); in xnn_cs16_fftr_ukernel__scalar_x4()
124 dr[7] = math_asr_s32(twi0 - vacc1i0, 1); in xnn_cs16_fftr_ukernel__scalar_x4()
125 dr[4] = math_asr_s32(vacc1r1 - twr1, 1); in xnn_cs16_fftr_ukernel__scalar_x4()
126 dr[5] = math_asr_s32(twi1 - vacc1i1, 1); in xnn_cs16_fftr_ukernel__scalar_x4()
127 dr[2] = math_asr_s32(vacc1r2 - twr2, 1); in xnn_cs16_fftr_ukernel__scalar_x4()
128 dr[3] = math_asr_s32(twi2 - vacc1i2, 1); in xnn_cs16_fftr_ukernel__scalar_x4()
129 dr[0] = math_asr_s32(vacc1r3 - twr3, 1); in xnn_cs16_fftr_ukernel__scalar_x4()
130 dr[1] = math_asr_s32(twi3 - vacc1i3, 1); in xnn_cs16_fftr_ukernel__scalar_x4()
136 dr -= 2; in xnn_cs16_fftr_ukernel__scalar_x4()
139 int32_t virr = (int32_t) dr[0]; in xnn_cs16_fftr_ukernel__scalar_x4()
140 int32_t viri = -(int32_t) dr[1]; in xnn_cs16_fftr_ukernel__scalar_x4()
159 dr[0] = math_asr_s32(vacc1r - twr, 1); in xnn_cs16_fftr_ukernel__scalar_x4()
160 dr[1] = math_asr_s32(twi - vacc1i, 1); in xnn_cs16_fftr_ukernel__scalar_x4()