1 // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
2 /* Copyright (c) 2020 NVIDIA CORPORATION. All rights reserved. */
3 
4 #include <linux/types.h>
5 #include <linux/crc32.h>
6 #include "dr_ste.h"
7 
8 #define SVLAN_ETHERTYPE		0x88a8
9 #define DR_STE_ENABLE_FLOW_TAG	BIT(31)
10 
11 enum dr_ste_v0_entry_type {
12 	DR_STE_TYPE_TX          = 1,
13 	DR_STE_TYPE_RX          = 2,
14 	DR_STE_TYPE_MODIFY_PKT  = 6,
15 };
16 
17 enum dr_ste_v0_action_tunl {
18 	DR_STE_TUNL_ACTION_NONE		= 0,
19 	DR_STE_TUNL_ACTION_ENABLE	= 1,
20 	DR_STE_TUNL_ACTION_DECAP	= 2,
21 	DR_STE_TUNL_ACTION_L3_DECAP	= 3,
22 	DR_STE_TUNL_ACTION_POP_VLAN	= 4,
23 };
24 
25 enum dr_ste_v0_action_type {
26 	DR_STE_ACTION_TYPE_PUSH_VLAN	= 1,
27 	DR_STE_ACTION_TYPE_ENCAP_L3	= 3,
28 	DR_STE_ACTION_TYPE_ENCAP	= 4,
29 };
30 
31 enum dr_ste_v0_action_mdfy_op {
32 	DR_STE_ACTION_MDFY_OP_COPY	= 0x1,
33 	DR_STE_ACTION_MDFY_OP_SET	= 0x2,
34 	DR_STE_ACTION_MDFY_OP_ADD	= 0x3,
35 };
36 
37 #define DR_STE_CALC_LU_TYPE(lookup_type, rx, inner) \
38 	((inner) ? DR_STE_V0_LU_TYPE_##lookup_type##_I : \
39 		   (rx) ? DR_STE_V0_LU_TYPE_##lookup_type##_D : \
40 			  DR_STE_V0_LU_TYPE_##lookup_type##_O)
41 
42 enum {
43 	DR_STE_V0_LU_TYPE_NOP				= 0x00,
44 	DR_STE_V0_LU_TYPE_SRC_GVMI_AND_QP		= 0x05,
45 	DR_STE_V0_LU_TYPE_ETHL2_TUNNELING_I		= 0x0a,
46 	DR_STE_V0_LU_TYPE_ETHL2_DST_O			= 0x06,
47 	DR_STE_V0_LU_TYPE_ETHL2_DST_I			= 0x07,
48 	DR_STE_V0_LU_TYPE_ETHL2_DST_D			= 0x1b,
49 	DR_STE_V0_LU_TYPE_ETHL2_SRC_O			= 0x08,
50 	DR_STE_V0_LU_TYPE_ETHL2_SRC_I			= 0x09,
51 	DR_STE_V0_LU_TYPE_ETHL2_SRC_D			= 0x1c,
52 	DR_STE_V0_LU_TYPE_ETHL2_SRC_DST_O		= 0x36,
53 	DR_STE_V0_LU_TYPE_ETHL2_SRC_DST_I		= 0x37,
54 	DR_STE_V0_LU_TYPE_ETHL2_SRC_DST_D		= 0x38,
55 	DR_STE_V0_LU_TYPE_ETHL3_IPV6_DST_O		= 0x0d,
56 	DR_STE_V0_LU_TYPE_ETHL3_IPV6_DST_I		= 0x0e,
57 	DR_STE_V0_LU_TYPE_ETHL3_IPV6_DST_D		= 0x1e,
58 	DR_STE_V0_LU_TYPE_ETHL3_IPV6_SRC_O		= 0x0f,
59 	DR_STE_V0_LU_TYPE_ETHL3_IPV6_SRC_I		= 0x10,
60 	DR_STE_V0_LU_TYPE_ETHL3_IPV6_SRC_D		= 0x1f,
61 	DR_STE_V0_LU_TYPE_ETHL3_IPV4_5_TUPLE_O		= 0x11,
62 	DR_STE_V0_LU_TYPE_ETHL3_IPV4_5_TUPLE_I		= 0x12,
63 	DR_STE_V0_LU_TYPE_ETHL3_IPV4_5_TUPLE_D		= 0x20,
64 	DR_STE_V0_LU_TYPE_ETHL3_IPV4_MISC_O		= 0x29,
65 	DR_STE_V0_LU_TYPE_ETHL3_IPV4_MISC_I		= 0x2a,
66 	DR_STE_V0_LU_TYPE_ETHL3_IPV4_MISC_D		= 0x2b,
67 	DR_STE_V0_LU_TYPE_ETHL4_O			= 0x13,
68 	DR_STE_V0_LU_TYPE_ETHL4_I			= 0x14,
69 	DR_STE_V0_LU_TYPE_ETHL4_D			= 0x21,
70 	DR_STE_V0_LU_TYPE_ETHL4_MISC_O			= 0x2c,
71 	DR_STE_V0_LU_TYPE_ETHL4_MISC_I			= 0x2d,
72 	DR_STE_V0_LU_TYPE_ETHL4_MISC_D			= 0x2e,
73 	DR_STE_V0_LU_TYPE_MPLS_FIRST_O			= 0x15,
74 	DR_STE_V0_LU_TYPE_MPLS_FIRST_I			= 0x24,
75 	DR_STE_V0_LU_TYPE_MPLS_FIRST_D			= 0x25,
76 	DR_STE_V0_LU_TYPE_GRE				= 0x16,
77 	DR_STE_V0_LU_TYPE_FLEX_PARSER_0			= 0x22,
78 	DR_STE_V0_LU_TYPE_FLEX_PARSER_1			= 0x23,
79 	DR_STE_V0_LU_TYPE_FLEX_PARSER_TNL_HEADER	= 0x19,
80 	DR_STE_V0_LU_TYPE_GENERAL_PURPOSE		= 0x18,
81 	DR_STE_V0_LU_TYPE_STEERING_REGISTERS_0		= 0x2f,
82 	DR_STE_V0_LU_TYPE_STEERING_REGISTERS_1		= 0x30,
83 	DR_STE_V0_LU_TYPE_TUNNEL_HEADER			= 0x34,
84 	DR_STE_V0_LU_TYPE_DONT_CARE			= MLX5DR_STE_LU_TYPE_DONT_CARE,
85 };
86 
87 enum {
88 	DR_STE_V0_ACTION_MDFY_FLD_L2_0		= 0,
89 	DR_STE_V0_ACTION_MDFY_FLD_L2_1		= 1,
90 	DR_STE_V0_ACTION_MDFY_FLD_L2_2		= 2,
91 	DR_STE_V0_ACTION_MDFY_FLD_L3_0		= 3,
92 	DR_STE_V0_ACTION_MDFY_FLD_L3_1		= 4,
93 	DR_STE_V0_ACTION_MDFY_FLD_L3_2		= 5,
94 	DR_STE_V0_ACTION_MDFY_FLD_L3_3		= 6,
95 	DR_STE_V0_ACTION_MDFY_FLD_L3_4		= 7,
96 	DR_STE_V0_ACTION_MDFY_FLD_L4_0		= 8,
97 	DR_STE_V0_ACTION_MDFY_FLD_L4_1		= 9,
98 	DR_STE_V0_ACTION_MDFY_FLD_MPLS		= 10,
99 	DR_STE_V0_ACTION_MDFY_FLD_L2_TNL_0	= 11,
100 	DR_STE_V0_ACTION_MDFY_FLD_REG_0		= 12,
101 	DR_STE_V0_ACTION_MDFY_FLD_REG_1		= 13,
102 	DR_STE_V0_ACTION_MDFY_FLD_REG_2		= 14,
103 	DR_STE_V0_ACTION_MDFY_FLD_REG_3		= 15,
104 	DR_STE_V0_ACTION_MDFY_FLD_L4_2		= 16,
105 	DR_STE_V0_ACTION_MDFY_FLD_FLEX_0	= 17,
106 	DR_STE_V0_ACTION_MDFY_FLD_FLEX_1	= 18,
107 	DR_STE_V0_ACTION_MDFY_FLD_FLEX_2	= 19,
108 	DR_STE_V0_ACTION_MDFY_FLD_FLEX_3	= 20,
109 	DR_STE_V0_ACTION_MDFY_FLD_L2_TNL_1	= 21,
110 	DR_STE_V0_ACTION_MDFY_FLD_METADATA	= 22,
111 	DR_STE_V0_ACTION_MDFY_FLD_RESERVED	= 23,
112 };
113 
114 static const struct mlx5dr_ste_action_modify_field dr_ste_v0_action_modify_field_arr[] = {
115 	[MLX5_ACTION_IN_FIELD_OUT_SMAC_47_16] = {
116 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L2_1, .start = 16, .end = 47,
117 	},
118 	[MLX5_ACTION_IN_FIELD_OUT_SMAC_15_0] = {
119 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L2_1, .start = 0, .end = 15,
120 	},
121 	[MLX5_ACTION_IN_FIELD_OUT_ETHERTYPE] = {
122 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L2_2, .start = 32, .end = 47,
123 	},
124 	[MLX5_ACTION_IN_FIELD_OUT_DMAC_47_16] = {
125 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L2_0, .start = 16, .end = 47,
126 	},
127 	[MLX5_ACTION_IN_FIELD_OUT_DMAC_15_0] = {
128 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L2_0, .start = 0, .end = 15,
129 	},
130 	[MLX5_ACTION_IN_FIELD_OUT_IP_DSCP] = {
131 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L3_1, .start = 0, .end = 5,
132 	},
133 	[MLX5_ACTION_IN_FIELD_OUT_TCP_FLAGS] = {
134 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L4_0, .start = 48, .end = 56,
135 		.l4_type = DR_STE_ACTION_MDFY_TYPE_L4_TCP,
136 	},
137 	[MLX5_ACTION_IN_FIELD_OUT_TCP_SPORT] = {
138 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L4_0, .start = 0, .end = 15,
139 		.l4_type = DR_STE_ACTION_MDFY_TYPE_L4_TCP,
140 	},
141 	[MLX5_ACTION_IN_FIELD_OUT_TCP_DPORT] = {
142 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L4_0, .start = 16, .end = 31,
143 		.l4_type = DR_STE_ACTION_MDFY_TYPE_L4_TCP,
144 	},
145 	[MLX5_ACTION_IN_FIELD_OUT_IP_TTL] = {
146 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L3_1, .start = 8, .end = 15,
147 		.l3_type = DR_STE_ACTION_MDFY_TYPE_L3_IPV4,
148 	},
149 	[MLX5_ACTION_IN_FIELD_OUT_IPV6_HOPLIMIT] = {
150 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L3_1, .start = 8, .end = 15,
151 		.l3_type = DR_STE_ACTION_MDFY_TYPE_L3_IPV6,
152 	},
153 	[MLX5_ACTION_IN_FIELD_OUT_UDP_SPORT] = {
154 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L4_0, .start = 0, .end = 15,
155 		.l4_type = DR_STE_ACTION_MDFY_TYPE_L4_UDP,
156 	},
157 	[MLX5_ACTION_IN_FIELD_OUT_UDP_DPORT] = {
158 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L4_0, .start = 16, .end = 31,
159 		.l4_type = DR_STE_ACTION_MDFY_TYPE_L4_UDP,
160 	},
161 	[MLX5_ACTION_IN_FIELD_OUT_SIPV6_127_96] = {
162 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L3_3, .start = 32, .end = 63,
163 		.l3_type = DR_STE_ACTION_MDFY_TYPE_L3_IPV6,
164 	},
165 	[MLX5_ACTION_IN_FIELD_OUT_SIPV6_95_64] = {
166 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L3_3, .start = 0, .end = 31,
167 		.l3_type = DR_STE_ACTION_MDFY_TYPE_L3_IPV6,
168 	},
169 	[MLX5_ACTION_IN_FIELD_OUT_SIPV6_63_32] = {
170 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L3_4, .start = 32, .end = 63,
171 		.l3_type = DR_STE_ACTION_MDFY_TYPE_L3_IPV6,
172 	},
173 	[MLX5_ACTION_IN_FIELD_OUT_SIPV6_31_0] = {
174 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L3_4, .start = 0, .end = 31,
175 		.l3_type = DR_STE_ACTION_MDFY_TYPE_L3_IPV6,
176 	},
177 	[MLX5_ACTION_IN_FIELD_OUT_DIPV6_127_96] = {
178 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L3_0, .start = 32, .end = 63,
179 		.l3_type = DR_STE_ACTION_MDFY_TYPE_L3_IPV6,
180 	},
181 	[MLX5_ACTION_IN_FIELD_OUT_DIPV6_95_64] = {
182 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L3_0, .start = 0, .end = 31,
183 		.l3_type = DR_STE_ACTION_MDFY_TYPE_L3_IPV6,
184 	},
185 	[MLX5_ACTION_IN_FIELD_OUT_DIPV6_63_32] = {
186 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L3_2, .start = 32, .end = 63,
187 		.l3_type = DR_STE_ACTION_MDFY_TYPE_L3_IPV6,
188 	},
189 	[MLX5_ACTION_IN_FIELD_OUT_DIPV6_31_0] = {
190 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L3_2, .start = 0, .end = 31,
191 		.l3_type = DR_STE_ACTION_MDFY_TYPE_L3_IPV6,
192 	},
193 	[MLX5_ACTION_IN_FIELD_OUT_SIPV4] = {
194 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L3_0, .start = 0, .end = 31,
195 		.l3_type = DR_STE_ACTION_MDFY_TYPE_L3_IPV4,
196 	},
197 	[MLX5_ACTION_IN_FIELD_OUT_DIPV4] = {
198 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L3_0, .start = 32, .end = 63,
199 		.l3_type = DR_STE_ACTION_MDFY_TYPE_L3_IPV4,
200 	},
201 	[MLX5_ACTION_IN_FIELD_METADATA_REG_A] = {
202 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_METADATA, .start = 0, .end = 31,
203 	},
204 	[MLX5_ACTION_IN_FIELD_METADATA_REG_B] = {
205 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_METADATA, .start = 32, .end = 63,
206 	},
207 	[MLX5_ACTION_IN_FIELD_METADATA_REG_C_0] = {
208 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_REG_0, .start = 32, .end = 63,
209 	},
210 	[MLX5_ACTION_IN_FIELD_METADATA_REG_C_1] = {
211 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_REG_0, .start = 0, .end = 31,
212 	},
213 	[MLX5_ACTION_IN_FIELD_METADATA_REG_C_2] = {
214 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_REG_1, .start = 32, .end = 63,
215 	},
216 	[MLX5_ACTION_IN_FIELD_METADATA_REG_C_3] = {
217 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_REG_1, .start = 0, .end = 31,
218 	},
219 	[MLX5_ACTION_IN_FIELD_METADATA_REG_C_4] = {
220 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_REG_2, .start = 32, .end = 63,
221 	},
222 	[MLX5_ACTION_IN_FIELD_METADATA_REG_C_5] = {
223 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_REG_2, .start = 0, .end = 31,
224 	},
225 	[MLX5_ACTION_IN_FIELD_OUT_TCP_SEQ_NUM] = {
226 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L4_1, .start = 32, .end = 63,
227 	},
228 	[MLX5_ACTION_IN_FIELD_OUT_TCP_ACK_NUM] = {
229 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L4_1, .start = 0, .end = 31,
230 	},
231 	[MLX5_ACTION_IN_FIELD_OUT_FIRST_VID] = {
232 		.hw_field = DR_STE_V0_ACTION_MDFY_FLD_L2_2, .start = 0, .end = 15,
233 	},
234 };
235 
dr_ste_v0_set_entry_type(u8 * hw_ste_p,u8 entry_type)236 static void dr_ste_v0_set_entry_type(u8 *hw_ste_p, u8 entry_type)
237 {
238 	MLX5_SET(ste_general, hw_ste_p, entry_type, entry_type);
239 }
240 
dr_ste_v0_get_entry_type(u8 * hw_ste_p)241 static u8 dr_ste_v0_get_entry_type(u8 *hw_ste_p)
242 {
243 	return MLX5_GET(ste_general, hw_ste_p, entry_type);
244 }
245 
dr_ste_v0_set_miss_addr(u8 * hw_ste_p,u64 miss_addr)246 static void dr_ste_v0_set_miss_addr(u8 *hw_ste_p, u64 miss_addr)
247 {
248 	u64 index = miss_addr >> 6;
249 
250 	/* Miss address for TX and RX STEs located in the same offsets */
251 	MLX5_SET(ste_rx_steering_mult, hw_ste_p, miss_address_39_32, index >> 26);
252 	MLX5_SET(ste_rx_steering_mult, hw_ste_p, miss_address_31_6, index);
253 }
254 
dr_ste_v0_get_miss_addr(u8 * hw_ste_p)255 static u64 dr_ste_v0_get_miss_addr(u8 *hw_ste_p)
256 {
257 	u64 index =
258 		((u64)MLX5_GET(ste_rx_steering_mult, hw_ste_p, miss_address_31_6) |
259 		 ((u64)MLX5_GET(ste_rx_steering_mult, hw_ste_p, miss_address_39_32)) << 26);
260 
261 	return index << 6;
262 }
263 
dr_ste_v0_set_byte_mask(u8 * hw_ste_p,u16 byte_mask)264 static void dr_ste_v0_set_byte_mask(u8 *hw_ste_p, u16 byte_mask)
265 {
266 	MLX5_SET(ste_general, hw_ste_p, byte_mask, byte_mask);
267 }
268 
dr_ste_v0_get_byte_mask(u8 * hw_ste_p)269 static u16 dr_ste_v0_get_byte_mask(u8 *hw_ste_p)
270 {
271 	return MLX5_GET(ste_general, hw_ste_p, byte_mask);
272 }
273 
dr_ste_v0_set_lu_type(u8 * hw_ste_p,u16 lu_type)274 static void dr_ste_v0_set_lu_type(u8 *hw_ste_p, u16 lu_type)
275 {
276 	MLX5_SET(ste_general, hw_ste_p, entry_sub_type, lu_type);
277 }
278 
dr_ste_v0_set_next_lu_type(u8 * hw_ste_p,u16 lu_type)279 static void dr_ste_v0_set_next_lu_type(u8 *hw_ste_p, u16 lu_type)
280 {
281 	MLX5_SET(ste_general, hw_ste_p, next_lu_type, lu_type);
282 }
283 
dr_ste_v0_get_next_lu_type(u8 * hw_ste_p)284 static u16 dr_ste_v0_get_next_lu_type(u8 *hw_ste_p)
285 {
286 	return MLX5_GET(ste_general, hw_ste_p, next_lu_type);
287 }
288 
dr_ste_v0_set_hit_gvmi(u8 * hw_ste_p,u16 gvmi)289 static void dr_ste_v0_set_hit_gvmi(u8 *hw_ste_p, u16 gvmi)
290 {
291 	MLX5_SET(ste_general, hw_ste_p, next_table_base_63_48, gvmi);
292 }
293 
dr_ste_v0_set_hit_addr(u8 * hw_ste_p,u64 icm_addr,u32 ht_size)294 static void dr_ste_v0_set_hit_addr(u8 *hw_ste_p, u64 icm_addr, u32 ht_size)
295 {
296 	u64 index = (icm_addr >> 5) | ht_size;
297 
298 	MLX5_SET(ste_general, hw_ste_p, next_table_base_39_32_size, index >> 27);
299 	MLX5_SET(ste_general, hw_ste_p, next_table_base_31_5_size, index);
300 }
301 
dr_ste_v0_init_full(u8 * hw_ste_p,u16 lu_type,enum dr_ste_v0_entry_type entry_type,u16 gvmi)302 static void dr_ste_v0_init_full(u8 *hw_ste_p, u16 lu_type,
303 				enum dr_ste_v0_entry_type entry_type, u16 gvmi)
304 {
305 	dr_ste_v0_set_entry_type(hw_ste_p, entry_type);
306 	dr_ste_v0_set_lu_type(hw_ste_p, lu_type);
307 	dr_ste_v0_set_next_lu_type(hw_ste_p, MLX5DR_STE_LU_TYPE_DONT_CARE);
308 
309 	/* Set GVMI once, this is the same for RX/TX
310 	 * bits 63_48 of next table base / miss address encode the next GVMI
311 	 */
312 	MLX5_SET(ste_rx_steering_mult, hw_ste_p, gvmi, gvmi);
313 	MLX5_SET(ste_rx_steering_mult, hw_ste_p, next_table_base_63_48, gvmi);
314 	MLX5_SET(ste_rx_steering_mult, hw_ste_p, miss_address_63_48, gvmi);
315 }
316 
dr_ste_v0_init(u8 * hw_ste_p,u16 lu_type,bool is_rx,u16 gvmi)317 static void dr_ste_v0_init(u8 *hw_ste_p, u16 lu_type,
318 			   bool is_rx, u16 gvmi)
319 {
320 	enum dr_ste_v0_entry_type entry_type;
321 
322 	entry_type = is_rx ? DR_STE_TYPE_RX : DR_STE_TYPE_TX;
323 	dr_ste_v0_init_full(hw_ste_p, lu_type, entry_type, gvmi);
324 }
325 
dr_ste_v0_rx_set_flow_tag(u8 * hw_ste_p,u32 flow_tag)326 static void dr_ste_v0_rx_set_flow_tag(u8 *hw_ste_p, u32 flow_tag)
327 {
328 	MLX5_SET(ste_rx_steering_mult, hw_ste_p, qp_list_pointer,
329 		 DR_STE_ENABLE_FLOW_TAG | flow_tag);
330 }
331 
dr_ste_v0_set_counter_id(u8 * hw_ste_p,u32 ctr_id)332 static void dr_ste_v0_set_counter_id(u8 *hw_ste_p, u32 ctr_id)
333 {
334 	/* This can be used for both rx_steering_mult and for sx_transmit */
335 	MLX5_SET(ste_rx_steering_mult, hw_ste_p, counter_trigger_15_0, ctr_id);
336 	MLX5_SET(ste_rx_steering_mult, hw_ste_p, counter_trigger_23_16, ctr_id >> 16);
337 }
338 
dr_ste_v0_set_go_back_bit(u8 * hw_ste_p)339 static void dr_ste_v0_set_go_back_bit(u8 *hw_ste_p)
340 {
341 	MLX5_SET(ste_sx_transmit, hw_ste_p, go_back, 1);
342 }
343 
dr_ste_v0_set_tx_push_vlan(u8 * hw_ste_p,u32 vlan_hdr,bool go_back)344 static void dr_ste_v0_set_tx_push_vlan(u8 *hw_ste_p, u32 vlan_hdr,
345 				       bool go_back)
346 {
347 	MLX5_SET(ste_sx_transmit, hw_ste_p, action_type,
348 		 DR_STE_ACTION_TYPE_PUSH_VLAN);
349 	MLX5_SET(ste_sx_transmit, hw_ste_p, encap_pointer_vlan_data, vlan_hdr);
350 	/* Due to HW limitation we need to set this bit, otherwise reformat +
351 	 * push vlan will not work.
352 	 */
353 	if (go_back)
354 		dr_ste_v0_set_go_back_bit(hw_ste_p);
355 }
356 
dr_ste_v0_set_tx_encap(void * hw_ste_p,u32 reformat_id,int size,bool encap_l3)357 static void dr_ste_v0_set_tx_encap(void *hw_ste_p, u32 reformat_id,
358 				   int size, bool encap_l3)
359 {
360 	MLX5_SET(ste_sx_transmit, hw_ste_p, action_type,
361 		 encap_l3 ? DR_STE_ACTION_TYPE_ENCAP_L3 : DR_STE_ACTION_TYPE_ENCAP);
362 	/* The hardware expects here size in words (2 byte) */
363 	MLX5_SET(ste_sx_transmit, hw_ste_p, action_description, size / 2);
364 	MLX5_SET(ste_sx_transmit, hw_ste_p, encap_pointer_vlan_data, reformat_id);
365 }
366 
dr_ste_v0_set_rx_decap(u8 * hw_ste_p)367 static void dr_ste_v0_set_rx_decap(u8 *hw_ste_p)
368 {
369 	MLX5_SET(ste_rx_steering_mult, hw_ste_p, tunneling_action,
370 		 DR_STE_TUNL_ACTION_DECAP);
371 	MLX5_SET(ste_rx_steering_mult, hw_ste_p, fail_on_error, 1);
372 }
373 
dr_ste_v0_set_rx_pop_vlan(u8 * hw_ste_p)374 static void dr_ste_v0_set_rx_pop_vlan(u8 *hw_ste_p)
375 {
376 	MLX5_SET(ste_rx_steering_mult, hw_ste_p, tunneling_action,
377 		 DR_STE_TUNL_ACTION_POP_VLAN);
378 }
379 
dr_ste_v0_set_rx_decap_l3(u8 * hw_ste_p,bool vlan)380 static void dr_ste_v0_set_rx_decap_l3(u8 *hw_ste_p, bool vlan)
381 {
382 	MLX5_SET(ste_rx_steering_mult, hw_ste_p, tunneling_action,
383 		 DR_STE_TUNL_ACTION_L3_DECAP);
384 	MLX5_SET(ste_modify_packet, hw_ste_p, action_description, vlan ? 1 : 0);
385 	MLX5_SET(ste_rx_steering_mult, hw_ste_p, fail_on_error, 1);
386 }
387 
dr_ste_v0_set_rewrite_actions(u8 * hw_ste_p,u16 num_of_actions,u32 re_write_index)388 static void dr_ste_v0_set_rewrite_actions(u8 *hw_ste_p, u16 num_of_actions,
389 					  u32 re_write_index)
390 {
391 	MLX5_SET(ste_modify_packet, hw_ste_p, number_of_re_write_actions,
392 		 num_of_actions);
393 	MLX5_SET(ste_modify_packet, hw_ste_p, header_re_write_actions_pointer,
394 		 re_write_index);
395 }
396 
dr_ste_v0_arr_init_next(u8 ** last_ste,u32 * added_stes,enum dr_ste_v0_entry_type entry_type,u16 gvmi)397 static void dr_ste_v0_arr_init_next(u8 **last_ste,
398 				    u32 *added_stes,
399 				    enum dr_ste_v0_entry_type entry_type,
400 				    u16 gvmi)
401 {
402 	(*added_stes)++;
403 	*last_ste += DR_STE_SIZE;
404 	dr_ste_v0_init_full(*last_ste, MLX5DR_STE_LU_TYPE_DONT_CARE,
405 			    entry_type, gvmi);
406 }
407 
408 static void
dr_ste_v0_set_actions_tx(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_domain * dmn,u8 * action_type_set,u32 actions_caps,u8 * last_ste,struct mlx5dr_ste_actions_attr * attr,u32 * added_stes)409 dr_ste_v0_set_actions_tx(struct mlx5dr_ste_ctx *ste_ctx,
410 			 struct mlx5dr_domain *dmn,
411 			 u8 *action_type_set,
412 			 u32 actions_caps,
413 			 u8 *last_ste,
414 			 struct mlx5dr_ste_actions_attr *attr,
415 			 u32 *added_stes)
416 {
417 	bool encap = action_type_set[DR_ACTION_TYP_L2_TO_TNL_L2] ||
418 		action_type_set[DR_ACTION_TYP_L2_TO_TNL_L3];
419 
420 	/* We want to make sure the modify header comes before L2
421 	 * encapsulation. The reason for that is that we support
422 	 * modify headers for outer headers only
423 	 */
424 	if (action_type_set[DR_ACTION_TYP_MODIFY_HDR] && attr->modify_actions) {
425 		dr_ste_v0_set_entry_type(last_ste, DR_STE_TYPE_MODIFY_PKT);
426 		dr_ste_v0_set_rewrite_actions(last_ste,
427 					      attr->modify_actions,
428 					      attr->modify_index);
429 	}
430 
431 	if (action_type_set[DR_ACTION_TYP_PUSH_VLAN]) {
432 		int i;
433 
434 		for (i = 0; i < attr->vlans.count; i++) {
435 			if (i || action_type_set[DR_ACTION_TYP_MODIFY_HDR])
436 				dr_ste_v0_arr_init_next(&last_ste,
437 							added_stes,
438 							DR_STE_TYPE_TX,
439 							attr->gvmi);
440 
441 			dr_ste_v0_set_tx_push_vlan(last_ste,
442 						   attr->vlans.headers[i],
443 						   encap);
444 		}
445 	}
446 
447 	if (encap) {
448 		/* Modify header and encapsulation require a different STEs.
449 		 * Since modify header STE format doesn't support encapsulation
450 		 * tunneling_action.
451 		 */
452 		if (action_type_set[DR_ACTION_TYP_MODIFY_HDR] ||
453 		    action_type_set[DR_ACTION_TYP_PUSH_VLAN])
454 			dr_ste_v0_arr_init_next(&last_ste,
455 						added_stes,
456 						DR_STE_TYPE_TX,
457 						attr->gvmi);
458 
459 		dr_ste_v0_set_tx_encap(last_ste,
460 				       attr->reformat.id,
461 				       attr->reformat.size,
462 				       action_type_set[DR_ACTION_TYP_L2_TO_TNL_L3]);
463 		/* Whenever prio_tag_required enabled, we can be sure that the
464 		 * previous table (ACL) already push vlan to our packet,
465 		 * And due to HW limitation we need to set this bit, otherwise
466 		 * push vlan + reformat will not work.
467 		 */
468 		if (MLX5_CAP_GEN(dmn->mdev, prio_tag_required))
469 			dr_ste_v0_set_go_back_bit(last_ste);
470 	}
471 
472 	if (action_type_set[DR_ACTION_TYP_CTR])
473 		dr_ste_v0_set_counter_id(last_ste, attr->ctr_id);
474 
475 	dr_ste_v0_set_hit_gvmi(last_ste, attr->hit_gvmi);
476 	dr_ste_v0_set_hit_addr(last_ste, attr->final_icm_addr, 1);
477 }
478 
479 static void
dr_ste_v0_set_actions_rx(struct mlx5dr_ste_ctx * ste_ctx,struct mlx5dr_domain * dmn,u8 * action_type_set,u32 actions_caps,u8 * last_ste,struct mlx5dr_ste_actions_attr * attr,u32 * added_stes)480 dr_ste_v0_set_actions_rx(struct mlx5dr_ste_ctx *ste_ctx,
481 			 struct mlx5dr_domain *dmn,
482 			 u8 *action_type_set,
483 			 u32 actions_caps,
484 			 u8 *last_ste,
485 			 struct mlx5dr_ste_actions_attr *attr,
486 			 u32 *added_stes)
487 {
488 	if (action_type_set[DR_ACTION_TYP_CTR])
489 		dr_ste_v0_set_counter_id(last_ste, attr->ctr_id);
490 
491 	if (action_type_set[DR_ACTION_TYP_TNL_L3_TO_L2]) {
492 		dr_ste_v0_set_entry_type(last_ste, DR_STE_TYPE_MODIFY_PKT);
493 		dr_ste_v0_set_rx_decap_l3(last_ste, attr->decap_with_vlan);
494 		dr_ste_v0_set_rewrite_actions(last_ste,
495 					      attr->decap_actions,
496 					      attr->decap_index);
497 	}
498 
499 	if (action_type_set[DR_ACTION_TYP_TNL_L2_TO_L2])
500 		dr_ste_v0_set_rx_decap(last_ste);
501 
502 	if (action_type_set[DR_ACTION_TYP_POP_VLAN]) {
503 		int i;
504 
505 		for (i = 0; i < attr->vlans.count; i++) {
506 			if (i ||
507 			    action_type_set[DR_ACTION_TYP_TNL_L2_TO_L2] ||
508 			    action_type_set[DR_ACTION_TYP_TNL_L3_TO_L2])
509 				dr_ste_v0_arr_init_next(&last_ste,
510 							added_stes,
511 							DR_STE_TYPE_RX,
512 							attr->gvmi);
513 
514 			dr_ste_v0_set_rx_pop_vlan(last_ste);
515 		}
516 	}
517 
518 	if (action_type_set[DR_ACTION_TYP_MODIFY_HDR] && attr->modify_actions) {
519 		if (dr_ste_v0_get_entry_type(last_ste) == DR_STE_TYPE_MODIFY_PKT)
520 			dr_ste_v0_arr_init_next(&last_ste,
521 						added_stes,
522 						DR_STE_TYPE_MODIFY_PKT,
523 						attr->gvmi);
524 		else
525 			dr_ste_v0_set_entry_type(last_ste, DR_STE_TYPE_MODIFY_PKT);
526 
527 		dr_ste_v0_set_rewrite_actions(last_ste,
528 					      attr->modify_actions,
529 					      attr->modify_index);
530 	}
531 
532 	if (action_type_set[DR_ACTION_TYP_TAG]) {
533 		if (dr_ste_v0_get_entry_type(last_ste) == DR_STE_TYPE_MODIFY_PKT)
534 			dr_ste_v0_arr_init_next(&last_ste,
535 						added_stes,
536 						DR_STE_TYPE_RX,
537 						attr->gvmi);
538 
539 		dr_ste_v0_rx_set_flow_tag(last_ste, attr->flow_tag);
540 	}
541 
542 	dr_ste_v0_set_hit_gvmi(last_ste, attr->hit_gvmi);
543 	dr_ste_v0_set_hit_addr(last_ste, attr->final_icm_addr, 1);
544 }
545 
dr_ste_v0_set_action_set(u8 * hw_action,u8 hw_field,u8 shifter,u8 length,u32 data)546 static void dr_ste_v0_set_action_set(u8 *hw_action,
547 				     u8 hw_field,
548 				     u8 shifter,
549 				     u8 length,
550 				     u32 data)
551 {
552 	length = (length == 32) ? 0 : length;
553 	MLX5_SET(dr_action_hw_set, hw_action, opcode, DR_STE_ACTION_MDFY_OP_SET);
554 	MLX5_SET(dr_action_hw_set, hw_action, destination_field_code, hw_field);
555 	MLX5_SET(dr_action_hw_set, hw_action, destination_left_shifter, shifter);
556 	MLX5_SET(dr_action_hw_set, hw_action, destination_length, length);
557 	MLX5_SET(dr_action_hw_set, hw_action, inline_data, data);
558 }
559 
dr_ste_v0_set_action_add(u8 * hw_action,u8 hw_field,u8 shifter,u8 length,u32 data)560 static void dr_ste_v0_set_action_add(u8 *hw_action,
561 				     u8 hw_field,
562 				     u8 shifter,
563 				     u8 length,
564 				     u32 data)
565 {
566 	length = (length == 32) ? 0 : length;
567 	MLX5_SET(dr_action_hw_set, hw_action, opcode, DR_STE_ACTION_MDFY_OP_ADD);
568 	MLX5_SET(dr_action_hw_set, hw_action, destination_field_code, hw_field);
569 	MLX5_SET(dr_action_hw_set, hw_action, destination_left_shifter, shifter);
570 	MLX5_SET(dr_action_hw_set, hw_action, destination_length, length);
571 	MLX5_SET(dr_action_hw_set, hw_action, inline_data, data);
572 }
573 
dr_ste_v0_set_action_copy(u8 * hw_action,u8 dst_hw_field,u8 dst_shifter,u8 dst_len,u8 src_hw_field,u8 src_shifter)574 static void dr_ste_v0_set_action_copy(u8 *hw_action,
575 				      u8 dst_hw_field,
576 				      u8 dst_shifter,
577 				      u8 dst_len,
578 				      u8 src_hw_field,
579 				      u8 src_shifter)
580 {
581 	MLX5_SET(dr_action_hw_copy, hw_action, opcode, DR_STE_ACTION_MDFY_OP_COPY);
582 	MLX5_SET(dr_action_hw_copy, hw_action, destination_field_code, dst_hw_field);
583 	MLX5_SET(dr_action_hw_copy, hw_action, destination_left_shifter, dst_shifter);
584 	MLX5_SET(dr_action_hw_copy, hw_action, destination_length, dst_len);
585 	MLX5_SET(dr_action_hw_copy, hw_action, source_field_code, src_hw_field);
586 	MLX5_SET(dr_action_hw_copy, hw_action, source_left_shifter, src_shifter);
587 }
588 
589 #define DR_STE_DECAP_L3_MIN_ACTION_NUM	5
590 
591 static int
dr_ste_v0_set_action_decap_l3_list(void * data,u32 data_sz,u8 * hw_action,u32 hw_action_sz,u16 * used_hw_action_num)592 dr_ste_v0_set_action_decap_l3_list(void *data, u32 data_sz,
593 				   u8 *hw_action, u32 hw_action_sz,
594 				   u16 *used_hw_action_num)
595 {
596 	struct mlx5_ifc_l2_hdr_bits *l2_hdr = data;
597 	u32 hw_action_num;
598 	int required_actions;
599 	u32 hdr_fld_4b;
600 	u16 hdr_fld_2b;
601 	u16 vlan_type;
602 	bool vlan;
603 
604 	vlan = (data_sz != HDR_LEN_L2);
605 	hw_action_num = hw_action_sz / MLX5_ST_SZ_BYTES(dr_action_hw_set);
606 	required_actions = DR_STE_DECAP_L3_MIN_ACTION_NUM + !!vlan;
607 
608 	if (hw_action_num < required_actions)
609 		return -ENOMEM;
610 
611 	/* dmac_47_16 */
612 	MLX5_SET(dr_action_hw_set, hw_action,
613 		 opcode, DR_STE_ACTION_MDFY_OP_SET);
614 	MLX5_SET(dr_action_hw_set, hw_action,
615 		 destination_length, 0);
616 	MLX5_SET(dr_action_hw_set, hw_action,
617 		 destination_field_code, DR_STE_V0_ACTION_MDFY_FLD_L2_0);
618 	MLX5_SET(dr_action_hw_set, hw_action,
619 		 destination_left_shifter, 16);
620 	hdr_fld_4b = MLX5_GET(l2_hdr, l2_hdr, dmac_47_16);
621 	MLX5_SET(dr_action_hw_set, hw_action,
622 		 inline_data, hdr_fld_4b);
623 	hw_action += MLX5_ST_SZ_BYTES(dr_action_hw_set);
624 
625 	/* smac_47_16 */
626 	MLX5_SET(dr_action_hw_set, hw_action,
627 		 opcode, DR_STE_ACTION_MDFY_OP_SET);
628 	MLX5_SET(dr_action_hw_set, hw_action,
629 		 destination_length, 0);
630 	MLX5_SET(dr_action_hw_set, hw_action,
631 		 destination_field_code, DR_STE_V0_ACTION_MDFY_FLD_L2_1);
632 	MLX5_SET(dr_action_hw_set, hw_action, destination_left_shifter, 16);
633 	hdr_fld_4b = (MLX5_GET(l2_hdr, l2_hdr, smac_31_0) >> 16 |
634 		      MLX5_GET(l2_hdr, l2_hdr, smac_47_32) << 16);
635 	MLX5_SET(dr_action_hw_set, hw_action, inline_data, hdr_fld_4b);
636 	hw_action += MLX5_ST_SZ_BYTES(dr_action_hw_set);
637 
638 	/* dmac_15_0 */
639 	MLX5_SET(dr_action_hw_set, hw_action,
640 		 opcode, DR_STE_ACTION_MDFY_OP_SET);
641 	MLX5_SET(dr_action_hw_set, hw_action,
642 		 destination_length, 16);
643 	MLX5_SET(dr_action_hw_set, hw_action,
644 		 destination_field_code, DR_STE_V0_ACTION_MDFY_FLD_L2_0);
645 	MLX5_SET(dr_action_hw_set, hw_action,
646 		 destination_left_shifter, 0);
647 	hdr_fld_2b = MLX5_GET(l2_hdr, l2_hdr, dmac_15_0);
648 	MLX5_SET(dr_action_hw_set, hw_action,
649 		 inline_data, hdr_fld_2b);
650 	hw_action += MLX5_ST_SZ_BYTES(dr_action_hw_set);
651 
652 	/* ethertype + (optional) vlan */
653 	MLX5_SET(dr_action_hw_set, hw_action,
654 		 opcode, DR_STE_ACTION_MDFY_OP_SET);
655 	MLX5_SET(dr_action_hw_set, hw_action,
656 		 destination_field_code, DR_STE_V0_ACTION_MDFY_FLD_L2_2);
657 	MLX5_SET(dr_action_hw_set, hw_action,
658 		 destination_left_shifter, 32);
659 	if (!vlan) {
660 		hdr_fld_2b = MLX5_GET(l2_hdr, l2_hdr, ethertype);
661 		MLX5_SET(dr_action_hw_set, hw_action, inline_data, hdr_fld_2b);
662 		MLX5_SET(dr_action_hw_set, hw_action, destination_length, 16);
663 	} else {
664 		hdr_fld_2b = MLX5_GET(l2_hdr, l2_hdr, ethertype);
665 		vlan_type = hdr_fld_2b == SVLAN_ETHERTYPE ? DR_STE_SVLAN : DR_STE_CVLAN;
666 		hdr_fld_2b = MLX5_GET(l2_hdr, l2_hdr, vlan);
667 		hdr_fld_4b = (vlan_type << 16) | hdr_fld_2b;
668 		MLX5_SET(dr_action_hw_set, hw_action, inline_data, hdr_fld_4b);
669 		MLX5_SET(dr_action_hw_set, hw_action, destination_length, 18);
670 	}
671 	hw_action += MLX5_ST_SZ_BYTES(dr_action_hw_set);
672 
673 	/* smac_15_0 */
674 	MLX5_SET(dr_action_hw_set, hw_action,
675 		 opcode, DR_STE_ACTION_MDFY_OP_SET);
676 	MLX5_SET(dr_action_hw_set, hw_action,
677 		 destination_length, 16);
678 	MLX5_SET(dr_action_hw_set, hw_action,
679 		 destination_field_code, DR_STE_V0_ACTION_MDFY_FLD_L2_1);
680 	MLX5_SET(dr_action_hw_set, hw_action,
681 		 destination_left_shifter, 0);
682 	hdr_fld_2b = MLX5_GET(l2_hdr, l2_hdr, smac_31_0);
683 	MLX5_SET(dr_action_hw_set, hw_action, inline_data, hdr_fld_2b);
684 	hw_action += MLX5_ST_SZ_BYTES(dr_action_hw_set);
685 
686 	if (vlan) {
687 		MLX5_SET(dr_action_hw_set, hw_action,
688 			 opcode, DR_STE_ACTION_MDFY_OP_SET);
689 		hdr_fld_2b = MLX5_GET(l2_hdr, l2_hdr, vlan_type);
690 		MLX5_SET(dr_action_hw_set, hw_action,
691 			 inline_data, hdr_fld_2b);
692 		MLX5_SET(dr_action_hw_set, hw_action,
693 			 destination_length, 16);
694 		MLX5_SET(dr_action_hw_set, hw_action,
695 			 destination_field_code, DR_STE_V0_ACTION_MDFY_FLD_L2_2);
696 		MLX5_SET(dr_action_hw_set, hw_action,
697 			 destination_left_shifter, 0);
698 	}
699 
700 	*used_hw_action_num = required_actions;
701 
702 	return 0;
703 }
704 
705 static void
dr_ste_v0_build_eth_l2_src_dst_bit_mask(struct mlx5dr_match_param * value,bool inner,u8 * bit_mask)706 dr_ste_v0_build_eth_l2_src_dst_bit_mask(struct mlx5dr_match_param *value,
707 					bool inner, u8 *bit_mask)
708 {
709 	struct mlx5dr_match_spec *mask = inner ? &value->inner : &value->outer;
710 
711 	DR_STE_SET_TAG(eth_l2_src_dst, bit_mask, dmac_47_16, mask, dmac_47_16);
712 	DR_STE_SET_TAG(eth_l2_src_dst, bit_mask, dmac_15_0, mask, dmac_15_0);
713 
714 	if (mask->smac_47_16 || mask->smac_15_0) {
715 		MLX5_SET(ste_eth_l2_src_dst, bit_mask, smac_47_32,
716 			 mask->smac_47_16 >> 16);
717 		MLX5_SET(ste_eth_l2_src_dst, bit_mask, smac_31_0,
718 			 mask->smac_47_16 << 16 | mask->smac_15_0);
719 		mask->smac_47_16 = 0;
720 		mask->smac_15_0 = 0;
721 	}
722 
723 	DR_STE_SET_TAG(eth_l2_src_dst, bit_mask, first_vlan_id, mask, first_vid);
724 	DR_STE_SET_TAG(eth_l2_src_dst, bit_mask, first_cfi, mask, first_cfi);
725 	DR_STE_SET_TAG(eth_l2_src_dst, bit_mask, first_priority, mask, first_prio);
726 	DR_STE_SET_ONES(eth_l2_src_dst, bit_mask, l3_type, mask, ip_version);
727 
728 	if (mask->cvlan_tag) {
729 		MLX5_SET(ste_eth_l2_src_dst, bit_mask, first_vlan_qualifier, -1);
730 		mask->cvlan_tag = 0;
731 	} else if (mask->svlan_tag) {
732 		MLX5_SET(ste_eth_l2_src_dst, bit_mask, first_vlan_qualifier, -1);
733 		mask->svlan_tag = 0;
734 	}
735 }
736 
737 static int
dr_ste_v0_build_eth_l2_src_dst_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)738 dr_ste_v0_build_eth_l2_src_dst_tag(struct mlx5dr_match_param *value,
739 				   struct mlx5dr_ste_build *sb,
740 				   u8 *tag)
741 {
742 	struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
743 
744 	DR_STE_SET_TAG(eth_l2_src_dst, tag, dmac_47_16, spec, dmac_47_16);
745 	DR_STE_SET_TAG(eth_l2_src_dst, tag, dmac_15_0, spec, dmac_15_0);
746 
747 	if (spec->smac_47_16 || spec->smac_15_0) {
748 		MLX5_SET(ste_eth_l2_src_dst, tag, smac_47_32,
749 			 spec->smac_47_16 >> 16);
750 		MLX5_SET(ste_eth_l2_src_dst, tag, smac_31_0,
751 			 spec->smac_47_16 << 16 | spec->smac_15_0);
752 		spec->smac_47_16 = 0;
753 		spec->smac_15_0 = 0;
754 	}
755 
756 	if (spec->ip_version) {
757 		if (spec->ip_version == IP_VERSION_IPV4) {
758 			MLX5_SET(ste_eth_l2_src_dst, tag, l3_type, STE_IPV4);
759 			spec->ip_version = 0;
760 		} else if (spec->ip_version == IP_VERSION_IPV6) {
761 			MLX5_SET(ste_eth_l2_src_dst, tag, l3_type, STE_IPV6);
762 			spec->ip_version = 0;
763 		} else {
764 			return -EINVAL;
765 		}
766 	}
767 
768 	DR_STE_SET_TAG(eth_l2_src_dst, tag, first_vlan_id, spec, first_vid);
769 	DR_STE_SET_TAG(eth_l2_src_dst, tag, first_cfi, spec, first_cfi);
770 	DR_STE_SET_TAG(eth_l2_src_dst, tag, first_priority, spec, first_prio);
771 
772 	if (spec->cvlan_tag) {
773 		MLX5_SET(ste_eth_l2_src_dst, tag, first_vlan_qualifier, DR_STE_CVLAN);
774 		spec->cvlan_tag = 0;
775 	} else if (spec->svlan_tag) {
776 		MLX5_SET(ste_eth_l2_src_dst, tag, first_vlan_qualifier, DR_STE_SVLAN);
777 		spec->svlan_tag = 0;
778 	}
779 	return 0;
780 }
781 
782 static void
dr_ste_v0_build_eth_l2_src_dst_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)783 dr_ste_v0_build_eth_l2_src_dst_init(struct mlx5dr_ste_build *sb,
784 				    struct mlx5dr_match_param *mask)
785 {
786 	dr_ste_v0_build_eth_l2_src_dst_bit_mask(mask, sb->inner, sb->bit_mask);
787 
788 	sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL2_SRC_DST, sb->rx, sb->inner);
789 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
790 	sb->ste_build_tag_func = &dr_ste_v0_build_eth_l2_src_dst_tag;
791 }
792 
793 static int
dr_ste_v0_build_eth_l3_ipv6_dst_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)794 dr_ste_v0_build_eth_l3_ipv6_dst_tag(struct mlx5dr_match_param *value,
795 				    struct mlx5dr_ste_build *sb,
796 				    u8 *tag)
797 {
798 	struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
799 
800 	DR_STE_SET_TAG(eth_l3_ipv6_dst, tag, dst_ip_127_96, spec, dst_ip_127_96);
801 	DR_STE_SET_TAG(eth_l3_ipv6_dst, tag, dst_ip_95_64, spec, dst_ip_95_64);
802 	DR_STE_SET_TAG(eth_l3_ipv6_dst, tag, dst_ip_63_32, spec, dst_ip_63_32);
803 	DR_STE_SET_TAG(eth_l3_ipv6_dst, tag, dst_ip_31_0, spec, dst_ip_31_0);
804 
805 	return 0;
806 }
807 
808 static void
dr_ste_v0_build_eth_l3_ipv6_dst_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)809 dr_ste_v0_build_eth_l3_ipv6_dst_init(struct mlx5dr_ste_build *sb,
810 				     struct mlx5dr_match_param *mask)
811 {
812 	dr_ste_v0_build_eth_l3_ipv6_dst_tag(mask, sb, sb->bit_mask);
813 
814 	sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL3_IPV6_DST, sb->rx, sb->inner);
815 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
816 	sb->ste_build_tag_func = &dr_ste_v0_build_eth_l3_ipv6_dst_tag;
817 }
818 
819 static int
dr_ste_v0_build_eth_l3_ipv6_src_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)820 dr_ste_v0_build_eth_l3_ipv6_src_tag(struct mlx5dr_match_param *value,
821 				    struct mlx5dr_ste_build *sb,
822 				    u8 *tag)
823 {
824 	struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
825 
826 	DR_STE_SET_TAG(eth_l3_ipv6_src, tag, src_ip_127_96, spec, src_ip_127_96);
827 	DR_STE_SET_TAG(eth_l3_ipv6_src, tag, src_ip_95_64, spec, src_ip_95_64);
828 	DR_STE_SET_TAG(eth_l3_ipv6_src, tag, src_ip_63_32, spec, src_ip_63_32);
829 	DR_STE_SET_TAG(eth_l3_ipv6_src, tag, src_ip_31_0, spec, src_ip_31_0);
830 
831 	return 0;
832 }
833 
834 static void
dr_ste_v0_build_eth_l3_ipv6_src_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)835 dr_ste_v0_build_eth_l3_ipv6_src_init(struct mlx5dr_ste_build *sb,
836 				     struct mlx5dr_match_param *mask)
837 {
838 	dr_ste_v0_build_eth_l3_ipv6_src_tag(mask, sb, sb->bit_mask);
839 
840 	sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL3_IPV6_SRC, sb->rx, sb->inner);
841 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
842 	sb->ste_build_tag_func = &dr_ste_v0_build_eth_l3_ipv6_src_tag;
843 }
844 
845 static int
dr_ste_v0_build_eth_l3_ipv4_5_tuple_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)846 dr_ste_v0_build_eth_l3_ipv4_5_tuple_tag(struct mlx5dr_match_param *value,
847 					struct mlx5dr_ste_build *sb,
848 					u8 *tag)
849 {
850 	struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
851 
852 	DR_STE_SET_TAG(eth_l3_ipv4_5_tuple, tag, destination_address, spec, dst_ip_31_0);
853 	DR_STE_SET_TAG(eth_l3_ipv4_5_tuple, tag, source_address, spec, src_ip_31_0);
854 	DR_STE_SET_TAG(eth_l3_ipv4_5_tuple, tag, destination_port, spec, tcp_dport);
855 	DR_STE_SET_TAG(eth_l3_ipv4_5_tuple, tag, destination_port, spec, udp_dport);
856 	DR_STE_SET_TAG(eth_l3_ipv4_5_tuple, tag, source_port, spec, tcp_sport);
857 	DR_STE_SET_TAG(eth_l3_ipv4_5_tuple, tag, source_port, spec, udp_sport);
858 	DR_STE_SET_TAG(eth_l3_ipv4_5_tuple, tag, protocol, spec, ip_protocol);
859 	DR_STE_SET_TAG(eth_l3_ipv4_5_tuple, tag, fragmented, spec, frag);
860 	DR_STE_SET_TAG(eth_l3_ipv4_5_tuple, tag, dscp, spec, ip_dscp);
861 	DR_STE_SET_TAG(eth_l3_ipv4_5_tuple, tag, ecn, spec, ip_ecn);
862 
863 	if (spec->tcp_flags) {
864 		DR_STE_SET_TCP_FLAGS(eth_l3_ipv4_5_tuple, tag, spec);
865 		spec->tcp_flags = 0;
866 	}
867 
868 	return 0;
869 }
870 
871 static void
dr_ste_v0_build_eth_l3_ipv4_5_tuple_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)872 dr_ste_v0_build_eth_l3_ipv4_5_tuple_init(struct mlx5dr_ste_build *sb,
873 					 struct mlx5dr_match_param *mask)
874 {
875 	dr_ste_v0_build_eth_l3_ipv4_5_tuple_tag(mask, sb, sb->bit_mask);
876 
877 	sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL3_IPV4_5_TUPLE, sb->rx, sb->inner);
878 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
879 	sb->ste_build_tag_func = &dr_ste_v0_build_eth_l3_ipv4_5_tuple_tag;
880 }
881 
882 static void
dr_ste_v0_build_eth_l2_src_or_dst_bit_mask(struct mlx5dr_match_param * value,bool inner,u8 * bit_mask)883 dr_ste_v0_build_eth_l2_src_or_dst_bit_mask(struct mlx5dr_match_param *value,
884 					   bool inner, u8 *bit_mask)
885 {
886 	struct mlx5dr_match_spec *mask = inner ? &value->inner : &value->outer;
887 	struct mlx5dr_match_misc *misc_mask = &value->misc;
888 
889 	DR_STE_SET_TAG(eth_l2_src, bit_mask, first_vlan_id, mask, first_vid);
890 	DR_STE_SET_TAG(eth_l2_src, bit_mask, first_cfi, mask, first_cfi);
891 	DR_STE_SET_TAG(eth_l2_src, bit_mask, first_priority, mask, first_prio);
892 	DR_STE_SET_TAG(eth_l2_src, bit_mask, ip_fragmented, mask, frag);
893 	DR_STE_SET_TAG(eth_l2_src, bit_mask, l3_ethertype, mask, ethertype);
894 	DR_STE_SET_ONES(eth_l2_src, bit_mask, l3_type, mask, ip_version);
895 
896 	if (mask->svlan_tag || mask->cvlan_tag) {
897 		MLX5_SET(ste_eth_l2_src, bit_mask, first_vlan_qualifier, -1);
898 		mask->cvlan_tag = 0;
899 		mask->svlan_tag = 0;
900 	}
901 
902 	if (inner) {
903 		if (misc_mask->inner_second_cvlan_tag ||
904 		    misc_mask->inner_second_svlan_tag) {
905 			MLX5_SET(ste_eth_l2_src, bit_mask, second_vlan_qualifier, -1);
906 			misc_mask->inner_second_cvlan_tag = 0;
907 			misc_mask->inner_second_svlan_tag = 0;
908 		}
909 
910 		DR_STE_SET_TAG(eth_l2_src, bit_mask,
911 			       second_vlan_id, misc_mask, inner_second_vid);
912 		DR_STE_SET_TAG(eth_l2_src, bit_mask,
913 			       second_cfi, misc_mask, inner_second_cfi);
914 		DR_STE_SET_TAG(eth_l2_src, bit_mask,
915 			       second_priority, misc_mask, inner_second_prio);
916 	} else {
917 		if (misc_mask->outer_second_cvlan_tag ||
918 		    misc_mask->outer_second_svlan_tag) {
919 			MLX5_SET(ste_eth_l2_src, bit_mask, second_vlan_qualifier, -1);
920 			misc_mask->outer_second_cvlan_tag = 0;
921 			misc_mask->outer_second_svlan_tag = 0;
922 		}
923 
924 		DR_STE_SET_TAG(eth_l2_src, bit_mask,
925 			       second_vlan_id, misc_mask, outer_second_vid);
926 		DR_STE_SET_TAG(eth_l2_src, bit_mask,
927 			       second_cfi, misc_mask, outer_second_cfi);
928 		DR_STE_SET_TAG(eth_l2_src, bit_mask,
929 			       second_priority, misc_mask, outer_second_prio);
930 	}
931 }
932 
933 static int
dr_ste_v0_build_eth_l2_src_or_dst_tag(struct mlx5dr_match_param * value,bool inner,u8 * tag)934 dr_ste_v0_build_eth_l2_src_or_dst_tag(struct mlx5dr_match_param *value,
935 				      bool inner, u8 *tag)
936 {
937 	struct mlx5dr_match_spec *spec = inner ? &value->inner : &value->outer;
938 	struct mlx5dr_match_misc *misc_spec = &value->misc;
939 
940 	DR_STE_SET_TAG(eth_l2_src, tag, first_vlan_id, spec, first_vid);
941 	DR_STE_SET_TAG(eth_l2_src, tag, first_cfi, spec, first_cfi);
942 	DR_STE_SET_TAG(eth_l2_src, tag, first_priority, spec, first_prio);
943 	DR_STE_SET_TAG(eth_l2_src, tag, ip_fragmented, spec, frag);
944 	DR_STE_SET_TAG(eth_l2_src, tag, l3_ethertype, spec, ethertype);
945 
946 	if (spec->ip_version) {
947 		if (spec->ip_version == IP_VERSION_IPV4) {
948 			MLX5_SET(ste_eth_l2_src, tag, l3_type, STE_IPV4);
949 			spec->ip_version = 0;
950 		} else if (spec->ip_version == IP_VERSION_IPV6) {
951 			MLX5_SET(ste_eth_l2_src, tag, l3_type, STE_IPV6);
952 			spec->ip_version = 0;
953 		} else {
954 			return -EINVAL;
955 		}
956 	}
957 
958 	if (spec->cvlan_tag) {
959 		MLX5_SET(ste_eth_l2_src, tag, first_vlan_qualifier, DR_STE_CVLAN);
960 		spec->cvlan_tag = 0;
961 	} else if (spec->svlan_tag) {
962 		MLX5_SET(ste_eth_l2_src, tag, first_vlan_qualifier, DR_STE_SVLAN);
963 		spec->svlan_tag = 0;
964 	}
965 
966 	if (inner) {
967 		if (misc_spec->inner_second_cvlan_tag) {
968 			MLX5_SET(ste_eth_l2_src, tag, second_vlan_qualifier, DR_STE_CVLAN);
969 			misc_spec->inner_second_cvlan_tag = 0;
970 		} else if (misc_spec->inner_second_svlan_tag) {
971 			MLX5_SET(ste_eth_l2_src, tag, second_vlan_qualifier, DR_STE_SVLAN);
972 			misc_spec->inner_second_svlan_tag = 0;
973 		}
974 
975 		DR_STE_SET_TAG(eth_l2_src, tag, second_vlan_id, misc_spec, inner_second_vid);
976 		DR_STE_SET_TAG(eth_l2_src, tag, second_cfi, misc_spec, inner_second_cfi);
977 		DR_STE_SET_TAG(eth_l2_src, tag, second_priority, misc_spec, inner_second_prio);
978 	} else {
979 		if (misc_spec->outer_second_cvlan_tag) {
980 			MLX5_SET(ste_eth_l2_src, tag, second_vlan_qualifier, DR_STE_CVLAN);
981 			misc_spec->outer_second_cvlan_tag = 0;
982 		} else if (misc_spec->outer_second_svlan_tag) {
983 			MLX5_SET(ste_eth_l2_src, tag, second_vlan_qualifier, DR_STE_SVLAN);
984 			misc_spec->outer_second_svlan_tag = 0;
985 		}
986 		DR_STE_SET_TAG(eth_l2_src, tag, second_vlan_id, misc_spec, outer_second_vid);
987 		DR_STE_SET_TAG(eth_l2_src, tag, second_cfi, misc_spec, outer_second_cfi);
988 		DR_STE_SET_TAG(eth_l2_src, tag, second_priority, misc_spec, outer_second_prio);
989 	}
990 
991 	return 0;
992 }
993 
994 static void
dr_ste_v0_build_eth_l2_src_bit_mask(struct mlx5dr_match_param * value,bool inner,u8 * bit_mask)995 dr_ste_v0_build_eth_l2_src_bit_mask(struct mlx5dr_match_param *value,
996 				    bool inner, u8 *bit_mask)
997 {
998 	struct mlx5dr_match_spec *mask = inner ? &value->inner : &value->outer;
999 
1000 	DR_STE_SET_TAG(eth_l2_src, bit_mask, smac_47_16, mask, smac_47_16);
1001 	DR_STE_SET_TAG(eth_l2_src, bit_mask, smac_15_0, mask, smac_15_0);
1002 
1003 	dr_ste_v0_build_eth_l2_src_or_dst_bit_mask(value, inner, bit_mask);
1004 }
1005 
1006 static int
dr_ste_v0_build_eth_l2_src_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1007 dr_ste_v0_build_eth_l2_src_tag(struct mlx5dr_match_param *value,
1008 			       struct mlx5dr_ste_build *sb,
1009 			       u8 *tag)
1010 {
1011 	struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
1012 
1013 	DR_STE_SET_TAG(eth_l2_src, tag, smac_47_16, spec, smac_47_16);
1014 	DR_STE_SET_TAG(eth_l2_src, tag, smac_15_0, spec, smac_15_0);
1015 
1016 	return dr_ste_v0_build_eth_l2_src_or_dst_tag(value, sb->inner, tag);
1017 }
1018 
1019 static void
dr_ste_v0_build_eth_l2_src_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1020 dr_ste_v0_build_eth_l2_src_init(struct mlx5dr_ste_build *sb,
1021 				struct mlx5dr_match_param *mask)
1022 {
1023 	dr_ste_v0_build_eth_l2_src_bit_mask(mask, sb->inner, sb->bit_mask);
1024 	sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL2_SRC, sb->rx, sb->inner);
1025 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1026 	sb->ste_build_tag_func = &dr_ste_v0_build_eth_l2_src_tag;
1027 }
1028 
1029 static void
dr_ste_v0_build_eth_l2_dst_bit_mask(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * bit_mask)1030 dr_ste_v0_build_eth_l2_dst_bit_mask(struct mlx5dr_match_param *value,
1031 				    struct mlx5dr_ste_build *sb,
1032 				    u8 *bit_mask)
1033 {
1034 	struct mlx5dr_match_spec *mask = sb->inner ? &value->inner : &value->outer;
1035 
1036 	DR_STE_SET_TAG(eth_l2_dst, bit_mask, dmac_47_16, mask, dmac_47_16);
1037 	DR_STE_SET_TAG(eth_l2_dst, bit_mask, dmac_15_0, mask, dmac_15_0);
1038 
1039 	dr_ste_v0_build_eth_l2_src_or_dst_bit_mask(value, sb->inner, bit_mask);
1040 }
1041 
1042 static int
dr_ste_v0_build_eth_l2_dst_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1043 dr_ste_v0_build_eth_l2_dst_tag(struct mlx5dr_match_param *value,
1044 			       struct mlx5dr_ste_build *sb,
1045 			       u8 *tag)
1046 {
1047 	struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
1048 
1049 	DR_STE_SET_TAG(eth_l2_dst, tag, dmac_47_16, spec, dmac_47_16);
1050 	DR_STE_SET_TAG(eth_l2_dst, tag, dmac_15_0, spec, dmac_15_0);
1051 
1052 	return dr_ste_v0_build_eth_l2_src_or_dst_tag(value, sb->inner, tag);
1053 }
1054 
1055 static void
dr_ste_v0_build_eth_l2_dst_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1056 dr_ste_v0_build_eth_l2_dst_init(struct mlx5dr_ste_build *sb,
1057 				struct mlx5dr_match_param *mask)
1058 {
1059 	dr_ste_v0_build_eth_l2_dst_bit_mask(mask, sb, sb->bit_mask);
1060 
1061 	sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL2_DST, sb->rx, sb->inner);
1062 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1063 	sb->ste_build_tag_func = &dr_ste_v0_build_eth_l2_dst_tag;
1064 }
1065 
1066 static void
dr_ste_v0_build_eth_l2_tnl_bit_mask(struct mlx5dr_match_param * value,bool inner,u8 * bit_mask)1067 dr_ste_v0_build_eth_l2_tnl_bit_mask(struct mlx5dr_match_param *value,
1068 				    bool inner, u8 *bit_mask)
1069 {
1070 	struct mlx5dr_match_spec *mask = inner ? &value->inner : &value->outer;
1071 	struct mlx5dr_match_misc *misc = &value->misc;
1072 
1073 	DR_STE_SET_TAG(eth_l2_tnl, bit_mask, dmac_47_16, mask, dmac_47_16);
1074 	DR_STE_SET_TAG(eth_l2_tnl, bit_mask, dmac_15_0, mask, dmac_15_0);
1075 	DR_STE_SET_TAG(eth_l2_tnl, bit_mask, first_vlan_id, mask, first_vid);
1076 	DR_STE_SET_TAG(eth_l2_tnl, bit_mask, first_cfi, mask, first_cfi);
1077 	DR_STE_SET_TAG(eth_l2_tnl, bit_mask, first_priority, mask, first_prio);
1078 	DR_STE_SET_TAG(eth_l2_tnl, bit_mask, ip_fragmented, mask, frag);
1079 	DR_STE_SET_TAG(eth_l2_tnl, bit_mask, l3_ethertype, mask, ethertype);
1080 	DR_STE_SET_ONES(eth_l2_tnl, bit_mask, l3_type, mask, ip_version);
1081 
1082 	if (misc->vxlan_vni) {
1083 		MLX5_SET(ste_eth_l2_tnl, bit_mask,
1084 			 l2_tunneling_network_id, (misc->vxlan_vni << 8));
1085 		misc->vxlan_vni = 0;
1086 	}
1087 
1088 	if (mask->svlan_tag || mask->cvlan_tag) {
1089 		MLX5_SET(ste_eth_l2_tnl, bit_mask, first_vlan_qualifier, -1);
1090 		mask->cvlan_tag = 0;
1091 		mask->svlan_tag = 0;
1092 	}
1093 }
1094 
1095 static int
dr_ste_v0_build_eth_l2_tnl_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1096 dr_ste_v0_build_eth_l2_tnl_tag(struct mlx5dr_match_param *value,
1097 			       struct mlx5dr_ste_build *sb,
1098 			       u8 *tag)
1099 {
1100 	struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
1101 	struct mlx5dr_match_misc *misc = &value->misc;
1102 
1103 	DR_STE_SET_TAG(eth_l2_tnl, tag, dmac_47_16, spec, dmac_47_16);
1104 	DR_STE_SET_TAG(eth_l2_tnl, tag, dmac_15_0, spec, dmac_15_0);
1105 	DR_STE_SET_TAG(eth_l2_tnl, tag, first_vlan_id, spec, first_vid);
1106 	DR_STE_SET_TAG(eth_l2_tnl, tag, first_cfi, spec, first_cfi);
1107 	DR_STE_SET_TAG(eth_l2_tnl, tag, ip_fragmented, spec, frag);
1108 	DR_STE_SET_TAG(eth_l2_tnl, tag, first_priority, spec, first_prio);
1109 	DR_STE_SET_TAG(eth_l2_tnl, tag, l3_ethertype, spec, ethertype);
1110 
1111 	if (misc->vxlan_vni) {
1112 		MLX5_SET(ste_eth_l2_tnl, tag, l2_tunneling_network_id,
1113 			 (misc->vxlan_vni << 8));
1114 		misc->vxlan_vni = 0;
1115 	}
1116 
1117 	if (spec->cvlan_tag) {
1118 		MLX5_SET(ste_eth_l2_tnl, tag, first_vlan_qualifier, DR_STE_CVLAN);
1119 		spec->cvlan_tag = 0;
1120 	} else if (spec->svlan_tag) {
1121 		MLX5_SET(ste_eth_l2_tnl, tag, first_vlan_qualifier, DR_STE_SVLAN);
1122 		spec->svlan_tag = 0;
1123 	}
1124 
1125 	if (spec->ip_version) {
1126 		if (spec->ip_version == IP_VERSION_IPV4) {
1127 			MLX5_SET(ste_eth_l2_tnl, tag, l3_type, STE_IPV4);
1128 			spec->ip_version = 0;
1129 		} else if (spec->ip_version == IP_VERSION_IPV6) {
1130 			MLX5_SET(ste_eth_l2_tnl, tag, l3_type, STE_IPV6);
1131 			spec->ip_version = 0;
1132 		} else {
1133 			return -EINVAL;
1134 		}
1135 	}
1136 
1137 	return 0;
1138 }
1139 
1140 static void
dr_ste_v0_build_eth_l2_tnl_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1141 dr_ste_v0_build_eth_l2_tnl_init(struct mlx5dr_ste_build *sb,
1142 				struct mlx5dr_match_param *mask)
1143 {
1144 	dr_ste_v0_build_eth_l2_tnl_bit_mask(mask, sb->inner, sb->bit_mask);
1145 
1146 	sb->lu_type = DR_STE_V0_LU_TYPE_ETHL2_TUNNELING_I;
1147 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1148 	sb->ste_build_tag_func = &dr_ste_v0_build_eth_l2_tnl_tag;
1149 }
1150 
1151 static int
dr_ste_v0_build_eth_l3_ipv4_misc_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1152 dr_ste_v0_build_eth_l3_ipv4_misc_tag(struct mlx5dr_match_param *value,
1153 				     struct mlx5dr_ste_build *sb,
1154 				     u8 *tag)
1155 {
1156 	struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
1157 
1158 	DR_STE_SET_TAG(eth_l3_ipv4_misc, tag, time_to_live, spec, ttl_hoplimit);
1159 	DR_STE_SET_TAG(eth_l3_ipv4_misc, tag, ihl, spec, ipv4_ihl);
1160 
1161 	return 0;
1162 }
1163 
1164 static void
dr_ste_v0_build_eth_l3_ipv4_misc_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1165 dr_ste_v0_build_eth_l3_ipv4_misc_init(struct mlx5dr_ste_build *sb,
1166 				      struct mlx5dr_match_param *mask)
1167 {
1168 	dr_ste_v0_build_eth_l3_ipv4_misc_tag(mask, sb, sb->bit_mask);
1169 
1170 	sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL3_IPV4_MISC, sb->rx, sb->inner);
1171 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1172 	sb->ste_build_tag_func = &dr_ste_v0_build_eth_l3_ipv4_misc_tag;
1173 }
1174 
1175 static int
dr_ste_v0_build_eth_ipv6_l3_l4_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1176 dr_ste_v0_build_eth_ipv6_l3_l4_tag(struct mlx5dr_match_param *value,
1177 				   struct mlx5dr_ste_build *sb,
1178 				   u8 *tag)
1179 {
1180 	struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
1181 	struct mlx5dr_match_misc *misc = &value->misc;
1182 
1183 	DR_STE_SET_TAG(eth_l4, tag, dst_port, spec, tcp_dport);
1184 	DR_STE_SET_TAG(eth_l4, tag, src_port, spec, tcp_sport);
1185 	DR_STE_SET_TAG(eth_l4, tag, dst_port, spec, udp_dport);
1186 	DR_STE_SET_TAG(eth_l4, tag, src_port, spec, udp_sport);
1187 	DR_STE_SET_TAG(eth_l4, tag, protocol, spec, ip_protocol);
1188 	DR_STE_SET_TAG(eth_l4, tag, fragmented, spec, frag);
1189 	DR_STE_SET_TAG(eth_l4, tag, dscp, spec, ip_dscp);
1190 	DR_STE_SET_TAG(eth_l4, tag, ecn, spec, ip_ecn);
1191 	DR_STE_SET_TAG(eth_l4, tag, ipv6_hop_limit, spec, ttl_hoplimit);
1192 
1193 	if (sb->inner)
1194 		DR_STE_SET_TAG(eth_l4, tag, flow_label, misc, inner_ipv6_flow_label);
1195 	else
1196 		DR_STE_SET_TAG(eth_l4, tag, flow_label, misc, outer_ipv6_flow_label);
1197 
1198 	if (spec->tcp_flags) {
1199 		DR_STE_SET_TCP_FLAGS(eth_l4, tag, spec);
1200 		spec->tcp_flags = 0;
1201 	}
1202 
1203 	return 0;
1204 }
1205 
1206 static void
dr_ste_v0_build_eth_ipv6_l3_l4_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1207 dr_ste_v0_build_eth_ipv6_l3_l4_init(struct mlx5dr_ste_build *sb,
1208 				    struct mlx5dr_match_param *mask)
1209 {
1210 	dr_ste_v0_build_eth_ipv6_l3_l4_tag(mask, sb, sb->bit_mask);
1211 
1212 	sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL4, sb->rx, sb->inner);
1213 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1214 	sb->ste_build_tag_func = &dr_ste_v0_build_eth_ipv6_l3_l4_tag;
1215 }
1216 
1217 static int
dr_ste_v0_build_mpls_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1218 dr_ste_v0_build_mpls_tag(struct mlx5dr_match_param *value,
1219 			 struct mlx5dr_ste_build *sb,
1220 			 u8 *tag)
1221 {
1222 	struct mlx5dr_match_misc2 *misc2 = &value->misc2;
1223 
1224 	if (sb->inner)
1225 		DR_STE_SET_MPLS(mpls, misc2, inner, tag);
1226 	else
1227 		DR_STE_SET_MPLS(mpls, misc2, outer, tag);
1228 
1229 	return 0;
1230 }
1231 
1232 static void
dr_ste_v0_build_mpls_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1233 dr_ste_v0_build_mpls_init(struct mlx5dr_ste_build *sb,
1234 			  struct mlx5dr_match_param *mask)
1235 {
1236 	dr_ste_v0_build_mpls_tag(mask, sb, sb->bit_mask);
1237 
1238 	sb->lu_type = DR_STE_CALC_LU_TYPE(MPLS_FIRST, sb->rx, sb->inner);
1239 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1240 	sb->ste_build_tag_func = &dr_ste_v0_build_mpls_tag;
1241 }
1242 
1243 static int
dr_ste_v0_build_tnl_gre_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1244 dr_ste_v0_build_tnl_gre_tag(struct mlx5dr_match_param *value,
1245 			    struct mlx5dr_ste_build *sb,
1246 			    u8 *tag)
1247 {
1248 	struct  mlx5dr_match_misc *misc = &value->misc;
1249 
1250 	DR_STE_SET_TAG(gre, tag, gre_protocol, misc, gre_protocol);
1251 
1252 	DR_STE_SET_TAG(gre, tag, gre_k_present, misc, gre_k_present);
1253 	DR_STE_SET_TAG(gre, tag, gre_key_h, misc, gre_key_h);
1254 	DR_STE_SET_TAG(gre, tag, gre_key_l, misc, gre_key_l);
1255 
1256 	DR_STE_SET_TAG(gre, tag, gre_c_present, misc, gre_c_present);
1257 
1258 	DR_STE_SET_TAG(gre, tag, gre_s_present, misc, gre_s_present);
1259 
1260 	return 0;
1261 }
1262 
1263 static void
dr_ste_v0_build_tnl_gre_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1264 dr_ste_v0_build_tnl_gre_init(struct mlx5dr_ste_build *sb,
1265 			     struct mlx5dr_match_param *mask)
1266 {
1267 	dr_ste_v0_build_tnl_gre_tag(mask, sb, sb->bit_mask);
1268 
1269 	sb->lu_type = DR_STE_V0_LU_TYPE_GRE;
1270 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1271 	sb->ste_build_tag_func = &dr_ste_v0_build_tnl_gre_tag;
1272 }
1273 
1274 static int
dr_ste_v0_build_tnl_mpls_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1275 dr_ste_v0_build_tnl_mpls_tag(struct mlx5dr_match_param *value,
1276 			     struct mlx5dr_ste_build *sb,
1277 			     u8 *tag)
1278 {
1279 	struct mlx5dr_match_misc2 *misc_2 = &value->misc2;
1280 	u32 mpls_hdr;
1281 
1282 	if (DR_STE_IS_OUTER_MPLS_OVER_GRE_SET(misc_2)) {
1283 		mpls_hdr = misc_2->outer_first_mpls_over_gre_label << HDR_MPLS_OFFSET_LABEL;
1284 		misc_2->outer_first_mpls_over_gre_label = 0;
1285 		mpls_hdr |= misc_2->outer_first_mpls_over_gre_exp << HDR_MPLS_OFFSET_EXP;
1286 		misc_2->outer_first_mpls_over_gre_exp = 0;
1287 		mpls_hdr |= misc_2->outer_first_mpls_over_gre_s_bos << HDR_MPLS_OFFSET_S_BOS;
1288 		misc_2->outer_first_mpls_over_gre_s_bos = 0;
1289 		mpls_hdr |= misc_2->outer_first_mpls_over_gre_ttl << HDR_MPLS_OFFSET_TTL;
1290 		misc_2->outer_first_mpls_over_gre_ttl = 0;
1291 	} else {
1292 		mpls_hdr = misc_2->outer_first_mpls_over_udp_label << HDR_MPLS_OFFSET_LABEL;
1293 		misc_2->outer_first_mpls_over_udp_label = 0;
1294 		mpls_hdr |= misc_2->outer_first_mpls_over_udp_exp << HDR_MPLS_OFFSET_EXP;
1295 		misc_2->outer_first_mpls_over_udp_exp = 0;
1296 		mpls_hdr |= misc_2->outer_first_mpls_over_udp_s_bos << HDR_MPLS_OFFSET_S_BOS;
1297 		misc_2->outer_first_mpls_over_udp_s_bos = 0;
1298 		mpls_hdr |= misc_2->outer_first_mpls_over_udp_ttl << HDR_MPLS_OFFSET_TTL;
1299 		misc_2->outer_first_mpls_over_udp_ttl = 0;
1300 	}
1301 
1302 	MLX5_SET(ste_flex_parser_0, tag, flex_parser_3, mpls_hdr);
1303 	return 0;
1304 }
1305 
1306 static void
dr_ste_v0_build_tnl_mpls_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1307 dr_ste_v0_build_tnl_mpls_init(struct mlx5dr_ste_build *sb,
1308 			      struct mlx5dr_match_param *mask)
1309 {
1310 	dr_ste_v0_build_tnl_mpls_tag(mask, sb, sb->bit_mask);
1311 
1312 	sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_0;
1313 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1314 	sb->ste_build_tag_func = &dr_ste_v0_build_tnl_mpls_tag;
1315 }
1316 
1317 static int
dr_ste_v0_build_tnl_mpls_over_udp_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1318 dr_ste_v0_build_tnl_mpls_over_udp_tag(struct mlx5dr_match_param *value,
1319 				      struct mlx5dr_ste_build *sb,
1320 				      u8 *tag)
1321 {
1322 	struct mlx5dr_match_misc2 *misc2 = &value->misc2;
1323 	u8 *parser_ptr;
1324 	u8 parser_id;
1325 	u32 mpls_hdr;
1326 
1327 	mpls_hdr = misc2->outer_first_mpls_over_udp_label << HDR_MPLS_OFFSET_LABEL;
1328 	misc2->outer_first_mpls_over_udp_label = 0;
1329 	mpls_hdr |= misc2->outer_first_mpls_over_udp_exp << HDR_MPLS_OFFSET_EXP;
1330 	misc2->outer_first_mpls_over_udp_exp = 0;
1331 	mpls_hdr |= misc2->outer_first_mpls_over_udp_s_bos << HDR_MPLS_OFFSET_S_BOS;
1332 	misc2->outer_first_mpls_over_udp_s_bos = 0;
1333 	mpls_hdr |= misc2->outer_first_mpls_over_udp_ttl << HDR_MPLS_OFFSET_TTL;
1334 	misc2->outer_first_mpls_over_udp_ttl = 0;
1335 
1336 	parser_id = sb->caps->flex_parser_id_mpls_over_udp;
1337 	parser_ptr = dr_ste_calc_flex_parser_offset(tag, parser_id);
1338 	*(__be32 *)parser_ptr = cpu_to_be32(mpls_hdr);
1339 
1340 	return 0;
1341 }
1342 
1343 static void
dr_ste_v0_build_tnl_mpls_over_udp_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1344 dr_ste_v0_build_tnl_mpls_over_udp_init(struct mlx5dr_ste_build *sb,
1345 				       struct mlx5dr_match_param *mask)
1346 {
1347 	dr_ste_v0_build_tnl_mpls_over_udp_tag(mask, sb, sb->bit_mask);
1348 	/* STEs with lookup type FLEX_PARSER_{0/1} includes
1349 	 * flex parsers_{0-3}/{4-7} respectively.
1350 	 */
1351 	sb->lu_type = sb->caps->flex_parser_id_mpls_over_udp > DR_STE_MAX_FLEX_0_ID ?
1352 		      DR_STE_V0_LU_TYPE_FLEX_PARSER_1 :
1353 		      DR_STE_V0_LU_TYPE_FLEX_PARSER_0;
1354 
1355 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1356 	sb->ste_build_tag_func = &dr_ste_v0_build_tnl_mpls_over_udp_tag;
1357 }
1358 
1359 static int
dr_ste_v0_build_tnl_mpls_over_gre_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1360 dr_ste_v0_build_tnl_mpls_over_gre_tag(struct mlx5dr_match_param *value,
1361 				      struct mlx5dr_ste_build *sb,
1362 				      u8 *tag)
1363 {
1364 	struct mlx5dr_match_misc2 *misc2 = &value->misc2;
1365 	u8 *parser_ptr;
1366 	u8 parser_id;
1367 	u32 mpls_hdr;
1368 
1369 	mpls_hdr = misc2->outer_first_mpls_over_gre_label << HDR_MPLS_OFFSET_LABEL;
1370 	misc2->outer_first_mpls_over_gre_label = 0;
1371 	mpls_hdr |= misc2->outer_first_mpls_over_gre_exp << HDR_MPLS_OFFSET_EXP;
1372 	misc2->outer_first_mpls_over_gre_exp = 0;
1373 	mpls_hdr |= misc2->outer_first_mpls_over_gre_s_bos << HDR_MPLS_OFFSET_S_BOS;
1374 	misc2->outer_first_mpls_over_gre_s_bos = 0;
1375 	mpls_hdr |= misc2->outer_first_mpls_over_gre_ttl << HDR_MPLS_OFFSET_TTL;
1376 	misc2->outer_first_mpls_over_gre_ttl = 0;
1377 
1378 	parser_id = sb->caps->flex_parser_id_mpls_over_gre;
1379 	parser_ptr = dr_ste_calc_flex_parser_offset(tag, parser_id);
1380 	*(__be32 *)parser_ptr = cpu_to_be32(mpls_hdr);
1381 
1382 	return 0;
1383 }
1384 
1385 static void
dr_ste_v0_build_tnl_mpls_over_gre_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1386 dr_ste_v0_build_tnl_mpls_over_gre_init(struct mlx5dr_ste_build *sb,
1387 				       struct mlx5dr_match_param *mask)
1388 {
1389 	dr_ste_v0_build_tnl_mpls_over_gre_tag(mask, sb, sb->bit_mask);
1390 
1391 	/* STEs with lookup type FLEX_PARSER_{0/1} includes
1392 	 * flex parsers_{0-3}/{4-7} respectively.
1393 	 */
1394 	sb->lu_type = sb->caps->flex_parser_id_mpls_over_gre > DR_STE_MAX_FLEX_0_ID ?
1395 		      DR_STE_V0_LU_TYPE_FLEX_PARSER_1 :
1396 		      DR_STE_V0_LU_TYPE_FLEX_PARSER_0;
1397 
1398 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1399 	sb->ste_build_tag_func = &dr_ste_v0_build_tnl_mpls_over_gre_tag;
1400 }
1401 
1402 #define ICMP_TYPE_OFFSET_FIRST_DW	24
1403 #define ICMP_CODE_OFFSET_FIRST_DW	16
1404 
1405 static int
dr_ste_v0_build_icmp_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1406 dr_ste_v0_build_icmp_tag(struct mlx5dr_match_param *value,
1407 			 struct mlx5dr_ste_build *sb,
1408 			 u8 *tag)
1409 {
1410 	struct mlx5dr_match_misc3 *misc_3 = &value->misc3;
1411 	u32 *icmp_header_data;
1412 	int dw0_location;
1413 	int dw1_location;
1414 	u8 *parser_ptr;
1415 	u8 *icmp_type;
1416 	u8 *icmp_code;
1417 	bool is_ipv4;
1418 	u32 icmp_hdr;
1419 
1420 	is_ipv4 = DR_MASK_IS_ICMPV4_SET(misc_3);
1421 	if (is_ipv4) {
1422 		icmp_header_data	= &misc_3->icmpv4_header_data;
1423 		icmp_type		= &misc_3->icmpv4_type;
1424 		icmp_code		= &misc_3->icmpv4_code;
1425 		dw0_location		= sb->caps->flex_parser_id_icmp_dw0;
1426 		dw1_location		= sb->caps->flex_parser_id_icmp_dw1;
1427 	} else {
1428 		icmp_header_data	= &misc_3->icmpv6_header_data;
1429 		icmp_type		= &misc_3->icmpv6_type;
1430 		icmp_code		= &misc_3->icmpv6_code;
1431 		dw0_location		= sb->caps->flex_parser_id_icmpv6_dw0;
1432 		dw1_location		= sb->caps->flex_parser_id_icmpv6_dw1;
1433 	}
1434 
1435 	parser_ptr = dr_ste_calc_flex_parser_offset(tag, dw0_location);
1436 	icmp_hdr = (*icmp_type << ICMP_TYPE_OFFSET_FIRST_DW) |
1437 		   (*icmp_code << ICMP_CODE_OFFSET_FIRST_DW);
1438 	*(__be32 *)parser_ptr = cpu_to_be32(icmp_hdr);
1439 	*icmp_code = 0;
1440 	*icmp_type = 0;
1441 
1442 	parser_ptr = dr_ste_calc_flex_parser_offset(tag, dw1_location);
1443 	*(__be32 *)parser_ptr = cpu_to_be32(*icmp_header_data);
1444 	*icmp_header_data = 0;
1445 
1446 	return 0;
1447 }
1448 
1449 static void
dr_ste_v0_build_icmp_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1450 dr_ste_v0_build_icmp_init(struct mlx5dr_ste_build *sb,
1451 			  struct mlx5dr_match_param *mask)
1452 {
1453 	u8 parser_id;
1454 	bool is_ipv4;
1455 
1456 	dr_ste_v0_build_icmp_tag(mask, sb, sb->bit_mask);
1457 
1458 	/* STEs with lookup type FLEX_PARSER_{0/1} includes
1459 	 * flex parsers_{0-3}/{4-7} respectively.
1460 	 */
1461 	is_ipv4 = DR_MASK_IS_ICMPV4_SET(&mask->misc3);
1462 	parser_id = is_ipv4 ? sb->caps->flex_parser_id_icmp_dw0 :
1463 		    sb->caps->flex_parser_id_icmpv6_dw0;
1464 	sb->lu_type = parser_id > DR_STE_MAX_FLEX_0_ID ?
1465 		      DR_STE_V0_LU_TYPE_FLEX_PARSER_1 :
1466 		      DR_STE_V0_LU_TYPE_FLEX_PARSER_0;
1467 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1468 	sb->ste_build_tag_func = &dr_ste_v0_build_icmp_tag;
1469 }
1470 
1471 static int
dr_ste_v0_build_general_purpose_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1472 dr_ste_v0_build_general_purpose_tag(struct mlx5dr_match_param *value,
1473 				    struct mlx5dr_ste_build *sb,
1474 				    u8 *tag)
1475 {
1476 	struct mlx5dr_match_misc2 *misc_2 = &value->misc2;
1477 
1478 	DR_STE_SET_TAG(general_purpose, tag, general_purpose_lookup_field,
1479 		       misc_2, metadata_reg_a);
1480 
1481 	return 0;
1482 }
1483 
1484 static void
dr_ste_v0_build_general_purpose_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1485 dr_ste_v0_build_general_purpose_init(struct mlx5dr_ste_build *sb,
1486 				     struct mlx5dr_match_param *mask)
1487 {
1488 	dr_ste_v0_build_general_purpose_tag(mask, sb, sb->bit_mask);
1489 
1490 	sb->lu_type = DR_STE_V0_LU_TYPE_GENERAL_PURPOSE;
1491 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1492 	sb->ste_build_tag_func = &dr_ste_v0_build_general_purpose_tag;
1493 }
1494 
1495 static int
dr_ste_v0_build_eth_l4_misc_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1496 dr_ste_v0_build_eth_l4_misc_tag(struct mlx5dr_match_param *value,
1497 				struct mlx5dr_ste_build *sb,
1498 				u8 *tag)
1499 {
1500 	struct mlx5dr_match_misc3 *misc3 = &value->misc3;
1501 
1502 	if (sb->inner) {
1503 		DR_STE_SET_TAG(eth_l4_misc, tag, seq_num, misc3, inner_tcp_seq_num);
1504 		DR_STE_SET_TAG(eth_l4_misc, tag, ack_num, misc3, inner_tcp_ack_num);
1505 	} else {
1506 		DR_STE_SET_TAG(eth_l4_misc, tag, seq_num, misc3, outer_tcp_seq_num);
1507 		DR_STE_SET_TAG(eth_l4_misc, tag, ack_num, misc3, outer_tcp_ack_num);
1508 	}
1509 
1510 	return 0;
1511 }
1512 
1513 static void
dr_ste_v0_build_eth_l4_misc_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1514 dr_ste_v0_build_eth_l4_misc_init(struct mlx5dr_ste_build *sb,
1515 				 struct mlx5dr_match_param *mask)
1516 {
1517 	dr_ste_v0_build_eth_l4_misc_tag(mask, sb, sb->bit_mask);
1518 
1519 	sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL4_MISC, sb->rx, sb->inner);
1520 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1521 	sb->ste_build_tag_func = &dr_ste_v0_build_eth_l4_misc_tag;
1522 }
1523 
1524 static int
dr_ste_v0_build_flex_parser_tnl_vxlan_gpe_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1525 dr_ste_v0_build_flex_parser_tnl_vxlan_gpe_tag(struct mlx5dr_match_param *value,
1526 					      struct mlx5dr_ste_build *sb,
1527 					      u8 *tag)
1528 {
1529 	struct mlx5dr_match_misc3 *misc3 = &value->misc3;
1530 
1531 	DR_STE_SET_TAG(flex_parser_tnl_vxlan_gpe, tag,
1532 		       outer_vxlan_gpe_flags, misc3,
1533 		       outer_vxlan_gpe_flags);
1534 	DR_STE_SET_TAG(flex_parser_tnl_vxlan_gpe, tag,
1535 		       outer_vxlan_gpe_next_protocol, misc3,
1536 		       outer_vxlan_gpe_next_protocol);
1537 	DR_STE_SET_TAG(flex_parser_tnl_vxlan_gpe, tag,
1538 		       outer_vxlan_gpe_vni, misc3,
1539 		       outer_vxlan_gpe_vni);
1540 
1541 	return 0;
1542 }
1543 
1544 static void
dr_ste_v0_build_flex_parser_tnl_vxlan_gpe_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1545 dr_ste_v0_build_flex_parser_tnl_vxlan_gpe_init(struct mlx5dr_ste_build *sb,
1546 					       struct mlx5dr_match_param *mask)
1547 {
1548 	dr_ste_v0_build_flex_parser_tnl_vxlan_gpe_tag(mask, sb, sb->bit_mask);
1549 	sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_TNL_HEADER;
1550 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1551 	sb->ste_build_tag_func = &dr_ste_v0_build_flex_parser_tnl_vxlan_gpe_tag;
1552 }
1553 
1554 static int
dr_ste_v0_build_flex_parser_tnl_geneve_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1555 dr_ste_v0_build_flex_parser_tnl_geneve_tag(struct mlx5dr_match_param *value,
1556 					   struct mlx5dr_ste_build *sb,
1557 					   u8 *tag)
1558 {
1559 	struct mlx5dr_match_misc *misc = &value->misc;
1560 
1561 	DR_STE_SET_TAG(flex_parser_tnl_geneve, tag,
1562 		       geneve_protocol_type, misc, geneve_protocol_type);
1563 	DR_STE_SET_TAG(flex_parser_tnl_geneve, tag,
1564 		       geneve_oam, misc, geneve_oam);
1565 	DR_STE_SET_TAG(flex_parser_tnl_geneve, tag,
1566 		       geneve_opt_len, misc, geneve_opt_len);
1567 	DR_STE_SET_TAG(flex_parser_tnl_geneve, tag,
1568 		       geneve_vni, misc, geneve_vni);
1569 
1570 	return 0;
1571 }
1572 
1573 static void
dr_ste_v0_build_flex_parser_tnl_geneve_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1574 dr_ste_v0_build_flex_parser_tnl_geneve_init(struct mlx5dr_ste_build *sb,
1575 					    struct mlx5dr_match_param *mask)
1576 {
1577 	dr_ste_v0_build_flex_parser_tnl_geneve_tag(mask, sb, sb->bit_mask);
1578 	sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_TNL_HEADER;
1579 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1580 	sb->ste_build_tag_func = &dr_ste_v0_build_flex_parser_tnl_geneve_tag;
1581 }
1582 
1583 static int
dr_ste_v0_build_register_0_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1584 dr_ste_v0_build_register_0_tag(struct mlx5dr_match_param *value,
1585 			       struct mlx5dr_ste_build *sb,
1586 			       u8 *tag)
1587 {
1588 	struct mlx5dr_match_misc2 *misc2 = &value->misc2;
1589 
1590 	DR_STE_SET_TAG(register_0, tag, register_0_h, misc2, metadata_reg_c_0);
1591 	DR_STE_SET_TAG(register_0, tag, register_0_l, misc2, metadata_reg_c_1);
1592 	DR_STE_SET_TAG(register_0, tag, register_1_h, misc2, metadata_reg_c_2);
1593 	DR_STE_SET_TAG(register_0, tag, register_1_l, misc2, metadata_reg_c_3);
1594 
1595 	return 0;
1596 }
1597 
1598 static void
dr_ste_v0_build_register_0_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1599 dr_ste_v0_build_register_0_init(struct mlx5dr_ste_build *sb,
1600 				struct mlx5dr_match_param *mask)
1601 {
1602 	dr_ste_v0_build_register_0_tag(mask, sb, sb->bit_mask);
1603 
1604 	sb->lu_type = DR_STE_V0_LU_TYPE_STEERING_REGISTERS_0;
1605 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1606 	sb->ste_build_tag_func = &dr_ste_v0_build_register_0_tag;
1607 }
1608 
1609 static int
dr_ste_v0_build_register_1_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1610 dr_ste_v0_build_register_1_tag(struct mlx5dr_match_param *value,
1611 			       struct mlx5dr_ste_build *sb,
1612 			       u8 *tag)
1613 {
1614 	struct mlx5dr_match_misc2 *misc2 = &value->misc2;
1615 
1616 	DR_STE_SET_TAG(register_1, tag, register_2_h, misc2, metadata_reg_c_4);
1617 	DR_STE_SET_TAG(register_1, tag, register_2_l, misc2, metadata_reg_c_5);
1618 	DR_STE_SET_TAG(register_1, tag, register_3_h, misc2, metadata_reg_c_6);
1619 	DR_STE_SET_TAG(register_1, tag, register_3_l, misc2, metadata_reg_c_7);
1620 
1621 	return 0;
1622 }
1623 
1624 static void
dr_ste_v0_build_register_1_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1625 dr_ste_v0_build_register_1_init(struct mlx5dr_ste_build *sb,
1626 				struct mlx5dr_match_param *mask)
1627 {
1628 	dr_ste_v0_build_register_1_tag(mask, sb, sb->bit_mask);
1629 
1630 	sb->lu_type = DR_STE_V0_LU_TYPE_STEERING_REGISTERS_1;
1631 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1632 	sb->ste_build_tag_func = &dr_ste_v0_build_register_1_tag;
1633 }
1634 
1635 static void
dr_ste_v0_build_src_gvmi_qpn_bit_mask(struct mlx5dr_match_param * value,u8 * bit_mask)1636 dr_ste_v0_build_src_gvmi_qpn_bit_mask(struct mlx5dr_match_param *value,
1637 				      u8 *bit_mask)
1638 {
1639 	struct mlx5dr_match_misc *misc_mask = &value->misc;
1640 
1641 	DR_STE_SET_ONES(src_gvmi_qp, bit_mask, source_gvmi, misc_mask, source_port);
1642 	DR_STE_SET_ONES(src_gvmi_qp, bit_mask, source_qp, misc_mask, source_sqn);
1643 	misc_mask->source_eswitch_owner_vhca_id = 0;
1644 }
1645 
1646 static int
dr_ste_v0_build_src_gvmi_qpn_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1647 dr_ste_v0_build_src_gvmi_qpn_tag(struct mlx5dr_match_param *value,
1648 				 struct mlx5dr_ste_build *sb,
1649 				 u8 *tag)
1650 {
1651 	struct mlx5dr_match_misc *misc = &value->misc;
1652 	int id = misc->source_eswitch_owner_vhca_id;
1653 	struct mlx5dr_cmd_vport_cap *vport_cap;
1654 	struct mlx5dr_domain *dmn = sb->dmn;
1655 	struct mlx5dr_domain *vport_dmn;
1656 	u8 *bit_mask = sb->bit_mask;
1657 	struct mlx5dr_domain *peer;
1658 	bool source_gvmi_set;
1659 
1660 	DR_STE_SET_TAG(src_gvmi_qp, tag, source_qp, misc, source_sqn);
1661 
1662 	if (sb->vhca_id_valid) {
1663 		peer = xa_load(&dmn->peer_dmn_xa, id);
1664 		/* Find port GVMI based on the eswitch_owner_vhca_id */
1665 		if (id == dmn->info.caps.gvmi)
1666 			vport_dmn = dmn;
1667 		else if (peer && (id == peer->info.caps.gvmi))
1668 			vport_dmn = peer;
1669 		else
1670 			return -EINVAL;
1671 
1672 		misc->source_eswitch_owner_vhca_id = 0;
1673 	} else {
1674 		vport_dmn = dmn;
1675 	}
1676 
1677 	source_gvmi_set = MLX5_GET(ste_src_gvmi_qp, bit_mask, source_gvmi);
1678 	if (source_gvmi_set) {
1679 		vport_cap = mlx5dr_domain_get_vport_cap(vport_dmn,
1680 							misc->source_port);
1681 		if (!vport_cap) {
1682 			mlx5dr_err(dmn, "Vport 0x%x is disabled or invalid\n",
1683 				   misc->source_port);
1684 			return -EINVAL;
1685 		}
1686 
1687 		if (vport_cap->vport_gvmi)
1688 			MLX5_SET(ste_src_gvmi_qp, tag, source_gvmi, vport_cap->vport_gvmi);
1689 
1690 		misc->source_port = 0;
1691 	}
1692 
1693 	return 0;
1694 }
1695 
1696 static void
dr_ste_v0_build_src_gvmi_qpn_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1697 dr_ste_v0_build_src_gvmi_qpn_init(struct mlx5dr_ste_build *sb,
1698 				  struct mlx5dr_match_param *mask)
1699 {
1700 	dr_ste_v0_build_src_gvmi_qpn_bit_mask(mask, sb->bit_mask);
1701 
1702 	sb->lu_type = DR_STE_V0_LU_TYPE_SRC_GVMI_AND_QP;
1703 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1704 	sb->ste_build_tag_func = &dr_ste_v0_build_src_gvmi_qpn_tag;
1705 }
1706 
dr_ste_v0_set_flex_parser(u32 * misc4_field_id,u32 * misc4_field_value,bool * parser_is_used,u8 * tag)1707 static void dr_ste_v0_set_flex_parser(u32 *misc4_field_id,
1708 				      u32 *misc4_field_value,
1709 				      bool *parser_is_used,
1710 				      u8 *tag)
1711 {
1712 	u32 id = *misc4_field_id;
1713 	u8 *parser_ptr;
1714 
1715 	if (id >= DR_NUM_OF_FLEX_PARSERS || parser_is_used[id])
1716 		return;
1717 
1718 	parser_is_used[id] = true;
1719 	parser_ptr = dr_ste_calc_flex_parser_offset(tag, id);
1720 
1721 	*(__be32 *)parser_ptr = cpu_to_be32(*misc4_field_value);
1722 	*misc4_field_id = 0;
1723 	*misc4_field_value = 0;
1724 }
1725 
dr_ste_v0_build_flex_parser_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1726 static int dr_ste_v0_build_flex_parser_tag(struct mlx5dr_match_param *value,
1727 					   struct mlx5dr_ste_build *sb,
1728 					   u8 *tag)
1729 {
1730 	struct mlx5dr_match_misc4 *misc_4_mask = &value->misc4;
1731 	bool parser_is_used[DR_NUM_OF_FLEX_PARSERS] = {};
1732 
1733 	dr_ste_v0_set_flex_parser(&misc_4_mask->prog_sample_field_id_0,
1734 				  &misc_4_mask->prog_sample_field_value_0,
1735 				  parser_is_used, tag);
1736 
1737 	dr_ste_v0_set_flex_parser(&misc_4_mask->prog_sample_field_id_1,
1738 				  &misc_4_mask->prog_sample_field_value_1,
1739 				  parser_is_used, tag);
1740 
1741 	dr_ste_v0_set_flex_parser(&misc_4_mask->prog_sample_field_id_2,
1742 				  &misc_4_mask->prog_sample_field_value_2,
1743 				  parser_is_used, tag);
1744 
1745 	dr_ste_v0_set_flex_parser(&misc_4_mask->prog_sample_field_id_3,
1746 				  &misc_4_mask->prog_sample_field_value_3,
1747 				  parser_is_used, tag);
1748 
1749 	return 0;
1750 }
1751 
dr_ste_v0_build_flex_parser_0_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1752 static void dr_ste_v0_build_flex_parser_0_init(struct mlx5dr_ste_build *sb,
1753 					       struct mlx5dr_match_param *mask)
1754 {
1755 	sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_0;
1756 	dr_ste_v0_build_flex_parser_tag(mask, sb, sb->bit_mask);
1757 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1758 	sb->ste_build_tag_func = &dr_ste_v0_build_flex_parser_tag;
1759 }
1760 
dr_ste_v0_build_flex_parser_1_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1761 static void dr_ste_v0_build_flex_parser_1_init(struct mlx5dr_ste_build *sb,
1762 					       struct mlx5dr_match_param *mask)
1763 {
1764 	sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_1;
1765 	dr_ste_v0_build_flex_parser_tag(mask, sb, sb->bit_mask);
1766 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1767 	sb->ste_build_tag_func = &dr_ste_v0_build_flex_parser_tag;
1768 }
1769 
1770 static int
dr_ste_v0_build_flex_parser_tnl_geneve_tlv_opt_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1771 dr_ste_v0_build_flex_parser_tnl_geneve_tlv_opt_tag(struct mlx5dr_match_param *value,
1772 						   struct mlx5dr_ste_build *sb,
1773 						   u8 *tag)
1774 {
1775 	struct mlx5dr_match_misc3 *misc3 = &value->misc3;
1776 	u8 parser_id = sb->caps->flex_parser_id_geneve_tlv_option_0;
1777 	u8 *parser_ptr = dr_ste_calc_flex_parser_offset(tag, parser_id);
1778 
1779 	MLX5_SET(ste_flex_parser_0, parser_ptr, flex_parser_3,
1780 		 misc3->geneve_tlv_option_0_data);
1781 	misc3->geneve_tlv_option_0_data = 0;
1782 
1783 	return 0;
1784 }
1785 
1786 static void
dr_ste_v0_build_flex_parser_tnl_geneve_tlv_opt_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1787 dr_ste_v0_build_flex_parser_tnl_geneve_tlv_opt_init(struct mlx5dr_ste_build *sb,
1788 						    struct mlx5dr_match_param *mask)
1789 {
1790 	dr_ste_v0_build_flex_parser_tnl_geneve_tlv_opt_tag(mask, sb, sb->bit_mask);
1791 
1792 	/* STEs with lookup type FLEX_PARSER_{0/1} includes
1793 	 * flex parsers_{0-3}/{4-7} respectively.
1794 	 */
1795 	sb->lu_type = sb->caps->flex_parser_id_geneve_tlv_option_0 > 3 ?
1796 		DR_STE_V0_LU_TYPE_FLEX_PARSER_1 :
1797 		DR_STE_V0_LU_TYPE_FLEX_PARSER_0;
1798 
1799 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1800 	sb->ste_build_tag_func = &dr_ste_v0_build_flex_parser_tnl_geneve_tlv_opt_tag;
1801 }
1802 
dr_ste_v0_build_flex_parser_tnl_gtpu_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1803 static int dr_ste_v0_build_flex_parser_tnl_gtpu_tag(struct mlx5dr_match_param *value,
1804 						    struct mlx5dr_ste_build *sb,
1805 						    u8 *tag)
1806 {
1807 	struct mlx5dr_match_misc3 *misc3 = &value->misc3;
1808 
1809 	DR_STE_SET_TAG(flex_parser_tnl_gtpu, tag,
1810 		       gtpu_msg_flags, misc3,
1811 		       gtpu_msg_flags);
1812 	DR_STE_SET_TAG(flex_parser_tnl_gtpu, tag,
1813 		       gtpu_msg_type, misc3,
1814 		       gtpu_msg_type);
1815 	DR_STE_SET_TAG(flex_parser_tnl_gtpu, tag,
1816 		       gtpu_teid, misc3,
1817 		       gtpu_teid);
1818 
1819 	return 0;
1820 }
1821 
dr_ste_v0_build_flex_parser_tnl_gtpu_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1822 static void dr_ste_v0_build_flex_parser_tnl_gtpu_init(struct mlx5dr_ste_build *sb,
1823 						      struct mlx5dr_match_param *mask)
1824 {
1825 	dr_ste_v0_build_flex_parser_tnl_gtpu_tag(mask, sb, sb->bit_mask);
1826 
1827 	sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_TNL_HEADER;
1828 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1829 	sb->ste_build_tag_func = &dr_ste_v0_build_flex_parser_tnl_gtpu_tag;
1830 }
1831 
1832 static int
dr_ste_v0_build_tnl_gtpu_flex_parser_0_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1833 dr_ste_v0_build_tnl_gtpu_flex_parser_0_tag(struct mlx5dr_match_param *value,
1834 					   struct mlx5dr_ste_build *sb,
1835 					   u8 *tag)
1836 {
1837 	if (dr_is_flex_parser_0_id(sb->caps->flex_parser_id_gtpu_dw_0))
1838 		DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_dw_0, sb->caps, &value->misc3);
1839 	if (dr_is_flex_parser_0_id(sb->caps->flex_parser_id_gtpu_teid))
1840 		DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_teid, sb->caps, &value->misc3);
1841 	if (dr_is_flex_parser_0_id(sb->caps->flex_parser_id_gtpu_dw_2))
1842 		DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_dw_2, sb->caps, &value->misc3);
1843 	if (dr_is_flex_parser_0_id(sb->caps->flex_parser_id_gtpu_first_ext_dw_0))
1844 		DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_first_ext_dw_0, sb->caps, &value->misc3);
1845 	return 0;
1846 }
1847 
1848 static void
dr_ste_v0_build_tnl_gtpu_flex_parser_0_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1849 dr_ste_v0_build_tnl_gtpu_flex_parser_0_init(struct mlx5dr_ste_build *sb,
1850 					    struct mlx5dr_match_param *mask)
1851 {
1852 	dr_ste_v0_build_tnl_gtpu_flex_parser_0_tag(mask, sb, sb->bit_mask);
1853 
1854 	sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_0;
1855 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1856 	sb->ste_build_tag_func = &dr_ste_v0_build_tnl_gtpu_flex_parser_0_tag;
1857 }
1858 
1859 static int
dr_ste_v0_build_tnl_gtpu_flex_parser_1_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1860 dr_ste_v0_build_tnl_gtpu_flex_parser_1_tag(struct mlx5dr_match_param *value,
1861 					   struct mlx5dr_ste_build *sb,
1862 					   u8 *tag)
1863 {
1864 	if (dr_is_flex_parser_1_id(sb->caps->flex_parser_id_gtpu_dw_0))
1865 		DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_dw_0, sb->caps, &value->misc3);
1866 	if (dr_is_flex_parser_1_id(sb->caps->flex_parser_id_gtpu_teid))
1867 		DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_teid, sb->caps, &value->misc3);
1868 	if (dr_is_flex_parser_1_id(sb->caps->flex_parser_id_gtpu_dw_2))
1869 		DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_dw_2, sb->caps, &value->misc3);
1870 	if (dr_is_flex_parser_1_id(sb->caps->flex_parser_id_gtpu_first_ext_dw_0))
1871 		DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_first_ext_dw_0, sb->caps, &value->misc3);
1872 	return 0;
1873 }
1874 
1875 static void
dr_ste_v0_build_tnl_gtpu_flex_parser_1_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1876 dr_ste_v0_build_tnl_gtpu_flex_parser_1_init(struct mlx5dr_ste_build *sb,
1877 					    struct mlx5dr_match_param *mask)
1878 {
1879 	dr_ste_v0_build_tnl_gtpu_flex_parser_1_tag(mask, sb, sb->bit_mask);
1880 
1881 	sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_1;
1882 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1883 	sb->ste_build_tag_func = &dr_ste_v0_build_tnl_gtpu_flex_parser_1_tag;
1884 }
1885 
dr_ste_v0_build_tnl_header_0_1_tag(struct mlx5dr_match_param * value,struct mlx5dr_ste_build * sb,u8 * tag)1886 static int dr_ste_v0_build_tnl_header_0_1_tag(struct mlx5dr_match_param *value,
1887 					      struct mlx5dr_ste_build *sb,
1888 					      u8 *tag)
1889 {
1890 	struct mlx5dr_match_misc5 *misc5 = &value->misc5;
1891 
1892 	DR_STE_SET_TAG(tunnel_header, tag, tunnel_header_0, misc5, tunnel_header_0);
1893 	DR_STE_SET_TAG(tunnel_header, tag, tunnel_header_1, misc5, tunnel_header_1);
1894 
1895 	return 0;
1896 }
1897 
dr_ste_v0_build_tnl_header_0_1_init(struct mlx5dr_ste_build * sb,struct mlx5dr_match_param * mask)1898 static void dr_ste_v0_build_tnl_header_0_1_init(struct mlx5dr_ste_build *sb,
1899 						struct mlx5dr_match_param *mask)
1900 {
1901 	sb->lu_type = DR_STE_V0_LU_TYPE_TUNNEL_HEADER;
1902 	dr_ste_v0_build_tnl_header_0_1_tag(mask, sb, sb->bit_mask);
1903 	sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
1904 	sb->ste_build_tag_func = &dr_ste_v0_build_tnl_header_0_1_tag;
1905 }
1906 
1907 static struct mlx5dr_ste_ctx ste_ctx_v0 = {
1908 	/* Builders */
1909 	.build_eth_l2_src_dst_init	= &dr_ste_v0_build_eth_l2_src_dst_init,
1910 	.build_eth_l3_ipv6_src_init	= &dr_ste_v0_build_eth_l3_ipv6_src_init,
1911 	.build_eth_l3_ipv6_dst_init	= &dr_ste_v0_build_eth_l3_ipv6_dst_init,
1912 	.build_eth_l3_ipv4_5_tuple_init	= &dr_ste_v0_build_eth_l3_ipv4_5_tuple_init,
1913 	.build_eth_l2_src_init		= &dr_ste_v0_build_eth_l2_src_init,
1914 	.build_eth_l2_dst_init		= &dr_ste_v0_build_eth_l2_dst_init,
1915 	.build_eth_l2_tnl_init		= &dr_ste_v0_build_eth_l2_tnl_init,
1916 	.build_eth_l3_ipv4_misc_init	= &dr_ste_v0_build_eth_l3_ipv4_misc_init,
1917 	.build_eth_ipv6_l3_l4_init	= &dr_ste_v0_build_eth_ipv6_l3_l4_init,
1918 	.build_mpls_init		= &dr_ste_v0_build_mpls_init,
1919 	.build_tnl_gre_init		= &dr_ste_v0_build_tnl_gre_init,
1920 	.build_tnl_mpls_init		= &dr_ste_v0_build_tnl_mpls_init,
1921 	.build_tnl_mpls_over_udp_init	= &dr_ste_v0_build_tnl_mpls_over_udp_init,
1922 	.build_tnl_mpls_over_gre_init	= &dr_ste_v0_build_tnl_mpls_over_gre_init,
1923 	.build_icmp_init		= &dr_ste_v0_build_icmp_init,
1924 	.build_general_purpose_init	= &dr_ste_v0_build_general_purpose_init,
1925 	.build_eth_l4_misc_init		= &dr_ste_v0_build_eth_l4_misc_init,
1926 	.build_tnl_vxlan_gpe_init	= &dr_ste_v0_build_flex_parser_tnl_vxlan_gpe_init,
1927 	.build_tnl_geneve_init		= &dr_ste_v0_build_flex_parser_tnl_geneve_init,
1928 	.build_tnl_geneve_tlv_opt_init	= &dr_ste_v0_build_flex_parser_tnl_geneve_tlv_opt_init,
1929 	.build_register_0_init		= &dr_ste_v0_build_register_0_init,
1930 	.build_register_1_init		= &dr_ste_v0_build_register_1_init,
1931 	.build_src_gvmi_qpn_init	= &dr_ste_v0_build_src_gvmi_qpn_init,
1932 	.build_flex_parser_0_init	= &dr_ste_v0_build_flex_parser_0_init,
1933 	.build_flex_parser_1_init	= &dr_ste_v0_build_flex_parser_1_init,
1934 	.build_tnl_gtpu_init		= &dr_ste_v0_build_flex_parser_tnl_gtpu_init,
1935 	.build_tnl_header_0_1_init	= &dr_ste_v0_build_tnl_header_0_1_init,
1936 	.build_tnl_gtpu_flex_parser_0_init   = &dr_ste_v0_build_tnl_gtpu_flex_parser_0_init,
1937 	.build_tnl_gtpu_flex_parser_1_init   = &dr_ste_v0_build_tnl_gtpu_flex_parser_1_init,
1938 
1939 	/* Getters and Setters */
1940 	.ste_init			= &dr_ste_v0_init,
1941 	.set_next_lu_type		= &dr_ste_v0_set_next_lu_type,
1942 	.get_next_lu_type		= &dr_ste_v0_get_next_lu_type,
1943 	.set_miss_addr			= &dr_ste_v0_set_miss_addr,
1944 	.get_miss_addr			= &dr_ste_v0_get_miss_addr,
1945 	.set_hit_addr			= &dr_ste_v0_set_hit_addr,
1946 	.set_byte_mask			= &dr_ste_v0_set_byte_mask,
1947 	.get_byte_mask			= &dr_ste_v0_get_byte_mask,
1948 
1949 	/* Actions */
1950 	.actions_caps			= DR_STE_CTX_ACTION_CAP_NONE,
1951 	.set_actions_rx			= &dr_ste_v0_set_actions_rx,
1952 	.set_actions_tx			= &dr_ste_v0_set_actions_tx,
1953 	.modify_field_arr_sz		= ARRAY_SIZE(dr_ste_v0_action_modify_field_arr),
1954 	.modify_field_arr		= dr_ste_v0_action_modify_field_arr,
1955 	.set_action_set			= &dr_ste_v0_set_action_set,
1956 	.set_action_add			= &dr_ste_v0_set_action_add,
1957 	.set_action_copy		= &dr_ste_v0_set_action_copy,
1958 	.set_action_decap_l3_list	= &dr_ste_v0_set_action_decap_l3_list,
1959 };
1960 
mlx5dr_ste_get_ctx_v0(void)1961 struct mlx5dr_ste_ctx *mlx5dr_ste_get_ctx_v0(void)
1962 {
1963 	return &ste_ctx_v0;
1964 }
1965