xref: /aosp_15_r20/external/XNNPACK/src/f32-ibilinear-chw/gen/sse-p8.c (revision 4bdc94577ba0e567308109d787f7fec7b531ce36)
1 // Auto-generated file. Do not edit!
2 //   Template: src/f32-ibilinear-chw/sse.c.in
3 //   Generator: tools/xngen
4 //
5 // Copyright 2021 Google LLC
6 //
7 // This source code is licensed under the BSD-style license found in the
8 // LICENSE file in the root directory of this source tree.
9 
10 #include <assert.h>
11 
12 #include <immintrin.h>
13 
14 #include <xnnpack/ibilinear.h>
15 
16 
xnn_f32_ibilinear_chw_ukernel__sse_p8(size_t output_pixels,size_t channels,const float ** restrict input,size_t input_offset,const float * restrict weights,float * restrict output,size_t input_increment)17 void xnn_f32_ibilinear_chw_ukernel__sse_p8(
18     size_t output_pixels,
19     size_t channels,
20     const float**restrict input,
21     size_t input_offset,
22     const float*restrict weights,
23     float*restrict output,
24     size_t input_increment) XNN_OOB_READS
25 {
26   assert(output_pixels != 0);
27   assert(channels != 0);
28   assert(input_increment % sizeof(float) == 0);
29 
30   do {
31     const float** i = input;
32     const float* w = weights;
33     size_t p = output_pixels;
34     for (; p >= 8; p -= 8) {
35       const float* itl0 = (const float*) ((uintptr_t) i[0] + input_offset);
36       const float* ibl0 = (const float*) ((uintptr_t) i[1] + input_offset);
37       const float* itl1 = (const float*) ((uintptr_t) i[2] + input_offset);
38       const float* ibl1 = (const float*) ((uintptr_t) i[3] + input_offset);
39       const float* itl2 = (const float*) ((uintptr_t) i[4] + input_offset);
40       const float* ibl2 = (const float*) ((uintptr_t) i[5] + input_offset);
41       const float* itl3 = (const float*) ((uintptr_t) i[6] + input_offset);
42       const float* ibl3 = (const float*) ((uintptr_t) i[7] + input_offset);
43       const float* itl4 = (const float*) ((uintptr_t) i[8] + input_offset);
44       const float* ibl4 = (const float*) ((uintptr_t) i[9] + input_offset);
45       const float* itl5 = (const float*) ((uintptr_t) i[10] + input_offset);
46       const float* ibl5 = (const float*) ((uintptr_t) i[11] + input_offset);
47       const float* itl6 = (const float*) ((uintptr_t) i[12] + input_offset);
48       const float* ibl6 = (const float*) ((uintptr_t) i[13] + input_offset);
49       const float* itl7 = (const float*) ((uintptr_t) i[14] + input_offset);
50       const float* ibl7 = (const float*) ((uintptr_t) i[15] + input_offset);
51       i += 2 * 8;
52 
53       const __m128 vw0123p0 = _mm_loadu_ps(w + 0);
54       const __m128 vw0123p1 = _mm_loadu_ps(w + 4);
55       const __m128 vw4567p0 = _mm_loadu_ps(w + 8);
56       const __m128 vw4567p1 = _mm_loadu_ps(w + 12);
57       w += 2 * 8;
58 
59       const __m128 vtltr0 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl0);
60       const __m128 vblbr0 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl0);
61       const __m128 vtltr2 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl2);
62       const __m128 vblbr2 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl2);
63       const __m128 vtltr4 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl4);
64       const __m128 vblbr4 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl4);
65       const __m128 vtltr6 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl6);
66       const __m128 vblbr6 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl6);
67 
68       const __m128 valphah0123 = _mm_shuffle_ps(vw0123p0, vw0123p1, _MM_SHUFFLE(2, 0, 2, 0));
69       const __m128 valphav0123 = _mm_shuffle_ps(vw0123p0, vw0123p1, _MM_SHUFFLE(3, 1, 3, 1));
70       const __m128 valphah4567 = _mm_shuffle_ps(vw4567p0, vw4567p1, _MM_SHUFFLE(2, 0, 2, 0));
71       const __m128 valphav4567 = _mm_shuffle_ps(vw4567p0, vw4567p1, _MM_SHUFFLE(3, 1, 3, 1));
72 
73       const __m128 vtltr01 = _mm_loadh_pi(vtltr0, (const __m64*) itl1);
74       const __m128 vblbr01 = _mm_loadh_pi(vblbr0, (const __m64*) ibl1);
75       const __m128 vtltr23 = _mm_loadh_pi(vtltr2, (const __m64*) itl3);
76       const __m128 vblbr23 = _mm_loadh_pi(vblbr2, (const __m64*) ibl3);
77       const __m128 vtltr45 = _mm_loadh_pi(vtltr4, (const __m64*) itl5);
78       const __m128 vblbr45 = _mm_loadh_pi(vblbr4, (const __m64*) ibl5);
79       const __m128 vtltr67 = _mm_loadh_pi(vtltr6, (const __m64*) itl7);
80       const __m128 vblbr67 = _mm_loadh_pi(vblbr6, (const __m64*) ibl7);
81 
82       const __m128 vldrd01 = _mm_sub_ps(vblbr01, vtltr01);
83       const __m128 vldrd23 = _mm_sub_ps(vblbr23, vtltr23);
84       const __m128 vldrd45 = _mm_sub_ps(vblbr45, vtltr45);
85       const __m128 vldrd67 = _mm_sub_ps(vblbr67, vtltr67);
86 
87       const __m128 vld0123 = _mm_shuffle_ps(vldrd01, vldrd23, _MM_SHUFFLE(2, 0, 2, 0));
88       const __m128 vrd0123 = _mm_shuffle_ps(vldrd01, vldrd23, _MM_SHUFFLE(3, 1, 3, 1));
89       const __m128 vld4567 = _mm_shuffle_ps(vldrd45, vldrd67, _MM_SHUFFLE(2, 0, 2, 0));
90       const __m128 vrd4567 = _mm_shuffle_ps(vldrd45, vldrd67, _MM_SHUFFLE(3, 1, 3, 1));
91 
92       const __m128 vtl0123 = _mm_shuffle_ps(vtltr01, vtltr23, _MM_SHUFFLE(2, 0, 2, 0));
93       const __m128 vtr0123 = _mm_shuffle_ps(vtltr01, vtltr23, _MM_SHUFFLE(3, 1, 3, 1));
94       const __m128 vtl4567 = _mm_shuffle_ps(vtltr45, vtltr67, _MM_SHUFFLE(2, 0, 2, 0));
95       const __m128 vtr4567 = _mm_shuffle_ps(vtltr45, vtltr67, _MM_SHUFFLE(3, 1, 3, 1));
96 
97       const __m128 vl0123 = _mm_add_ps(vtl0123, _mm_mul_ps(vld0123, valphav0123));
98       const __m128 vr0123 = _mm_add_ps(vtr0123, _mm_mul_ps(vrd0123, valphav0123));
99       const __m128 vl4567 = _mm_add_ps(vtl4567, _mm_mul_ps(vld4567, valphav4567));
100       const __m128 vr4567 = _mm_add_ps(vtr4567, _mm_mul_ps(vrd4567, valphav4567));
101 
102       const __m128 vd0123 = _mm_sub_ps(vr0123, vl0123);
103       const __m128 vd4567 = _mm_sub_ps(vr4567, vl4567);
104 
105       const __m128 vo0123 = _mm_add_ps(vl0123, _mm_mul_ps(vd0123, valphah0123));
106       const __m128 vo4567 = _mm_add_ps(vl4567, _mm_mul_ps(vd4567, valphah4567));
107 
108       _mm_storeu_ps(output + 0, vo0123);
109       _mm_storeu_ps(output + 4, vo4567);
110       output += 8;
111     }
112 
113     for (; p >= 4; p -= 4) {
114       const float* itl0 = (const float*) ((uintptr_t) i[0] + input_offset);
115       const float* ibl0 = (const float*) ((uintptr_t) i[1] + input_offset);
116       const float* itl1 = (const float*) ((uintptr_t) i[2] + input_offset);
117       const float* ibl1 = (const float*) ((uintptr_t) i[3] + input_offset);
118       const float* itl2 = (const float*) ((uintptr_t) i[4] + input_offset);
119       const float* ibl2 = (const float*) ((uintptr_t) i[5] + input_offset);
120       const float* itl3 = (const float*) ((uintptr_t) i[6] + input_offset);
121       const float* ibl3 = (const float*) ((uintptr_t) i[7] + input_offset);
122       i += 8;
123 
124       const __m128 vw0 = _mm_loadu_ps(w);
125       const __m128 vw1 = _mm_loadu_ps(w + 4);
126       w += 8;
127 
128       const __m128 vtltr0 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl0);
129       const __m128 vblbr0 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl0);
130       const __m128 vtltr2 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl2);
131       const __m128 vblbr2 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl2);
132 
133       const __m128 valphah = _mm_shuffle_ps(vw0, vw1, _MM_SHUFFLE(2, 0, 2, 0));
134       const __m128 valphav = _mm_shuffle_ps(vw0, vw1, _MM_SHUFFLE(3, 1, 3, 1));
135 
136       const __m128 vtltr01 = _mm_loadh_pi(vtltr0, (const __m64*) itl1);
137       const __m128 vblbr01 = _mm_loadh_pi(vblbr0, (const __m64*) ibl1);
138       const __m128 vtltr23 = _mm_loadh_pi(vtltr2, (const __m64*) itl3);
139       const __m128 vblbr23 = _mm_loadh_pi(vblbr2, (const __m64*) ibl3);
140 
141       const __m128 vldrd01 = _mm_sub_ps(vblbr01, vtltr01);
142       const __m128 vldrd23 = _mm_sub_ps(vblbr23, vtltr23);
143 
144       const __m128 vld = _mm_shuffle_ps(vldrd01, vldrd23, _MM_SHUFFLE(2, 0, 2, 0));
145       const __m128 vrd = _mm_shuffle_ps(vldrd01, vldrd23, _MM_SHUFFLE(3, 1, 3, 1));
146 
147       const __m128 vtl = _mm_shuffle_ps(vtltr01, vtltr23, _MM_SHUFFLE(2, 0, 2, 0));
148       const __m128 vtr = _mm_shuffle_ps(vtltr01, vtltr23, _MM_SHUFFLE(3, 1, 3, 1));
149 
150       const __m128 vl = _mm_add_ps(vtl, _mm_mul_ps(vld, valphav));
151       const __m128 vr = _mm_add_ps(vtr, _mm_mul_ps(vrd, valphav));
152 
153       const __m128 vd = _mm_sub_ps(vr, vl);
154       const __m128 vo = _mm_add_ps(vl, _mm_mul_ps(vd, valphah));
155 
156       _mm_storeu_ps(output, vo);
157       output += 4;
158     }
159 
160     if XNN_UNLIKELY(p != 0) {
161       if (p & 2) {
162         const __m128 vw = _mm_loadu_ps(w);
163         w += 4;
164 
165         const __m128 valphah = _mm_shuffle_ps(vw, vw, _MM_SHUFFLE(2, 0, 2, 0));
166         const __m128 valphav = _mm_shuffle_ps(vw, vw, _MM_SHUFFLE(3, 1, 3, 1));
167 
168         const float* itl0 = (const float*) ((uintptr_t) i[0] + input_offset);
169         const float* ibl0 = (const float*) ((uintptr_t) i[1] + input_offset);
170         const float* itl1 = (const float*) ((uintptr_t) i[2] + input_offset);
171         const float* ibl1 = (const float*) ((uintptr_t) i[3] + input_offset);
172         i += 4;
173 
174         const __m128 vtltr = _mm_loadh_pi(_mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl0), (const __m64*) itl1);
175         const __m128 vblbr = _mm_loadh_pi(_mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl0), (const __m64*) ibl1);
176 
177         const __m128 vldrd = _mm_sub_ps(vblbr, vtltr);
178         const __m128 vld = _mm_shuffle_ps(vldrd, vldrd, _MM_SHUFFLE(2, 0, 2, 0));
179         const __m128 vrd = _mm_shuffle_ps(vldrd, vldrd, _MM_SHUFFLE(3, 1, 3, 1));
180 
181         const __m128 vtl = _mm_shuffle_ps(vtltr, vtltr, _MM_SHUFFLE(2, 0, 2, 0));
182         const __m128 vtr = _mm_shuffle_ps(vtltr, vtltr, _MM_SHUFFLE(3, 1, 3, 1));
183 
184         const __m128 vl = _mm_add_ps(vtl, _mm_mul_ps(vld, valphav));
185         const __m128 vr = _mm_add_ps(vtr, _mm_mul_ps(vrd, valphav));
186 
187         const __m128 vd = _mm_sub_ps(vr, vl);
188         const __m128 vo = _mm_add_ps(vl, _mm_mul_ps(vd, valphah));
189 
190         _mm_storel_pi((__m64*) output, vo);
191         output += 2;
192       }
193 
194       if (p & 1) {
195         // We are computing the following formula:
196         //   result = (1 - alpha_h) * (1 - alpha_v) * top_left +
197         //                 alpha_h  * (1 - alpha_v) * top_right +
198         //            (1 - alpha_h) *      alpha_v  * bottom_left +
199         //                 alpha_h  *      alpha_v  * bottom_right.
200         //
201         // Rearranging gives
202         //   result =    left + alpha_h * (right        - left),
203         // where
204         //   left =  top_left + alpha_v * (bottom_left  - top_left),
205         //  right = top_right + alpha_v * (bottom_right - top_right).
206 
207         const float alphah = *w;
208         const __m128 valphav = _mm_load_ps1(w + 1);
209         w += 2;
210 
211         const float* itl = (const float*) ((uintptr_t) i[0] + input_offset);
212         const float* ibl = (const float*) ((uintptr_t) i[1] + input_offset);
213         i += 2;
214 
215         const __m128 vtltr = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl);
216         const __m128 vblbr = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl);
217 
218         // Compute at once
219         //    left_diff = bottom_left  - top_left
220         //   right_diff = bottom_right - top_right
221         const __m128 vldrd = _mm_sub_ps(vblbr, vtltr);
222         const __m128 vlr = _mm_add_ps(vtltr, _mm_mul_ps(vldrd, valphav));
223 
224         // Extract them and compute the result.
225         const float l = _mm_cvtss_f32(vlr);
226         const float r = _mm_cvtss_f32(_mm_shuffle_ps(vlr, vlr, 1));
227 
228         *output++ = l + alphah * (r - l);
229       }
230     }
231 
232     input_offset += input_increment;
233   } while (--channels != 0);
234 }
235