Merge branch 'for-v5.13-rc3' of git://git.kernel.org/pub/scm/linux/kernel/git/ebieder...
[linux-2.6-microblaze.git] / drivers / net / ethernet / mellanox / mlxsw / spectrum_matchall.c
1 // SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0
2 /* Copyright (c) 2017-2020 Mellanox Technologies. All rights reserved */
3
4 #include <linux/kernel.h>
5 #include <linux/errno.h>
6 #include <linux/netdevice.h>
7 #include <net/flow_offload.h>
8
9 #include "spectrum.h"
10 #include "spectrum_span.h"
11 #include "reg.h"
12
13 static struct mlxsw_sp_mall_entry *
14 mlxsw_sp_mall_entry_find(struct mlxsw_sp_flow_block *block, unsigned long cookie)
15 {
16         struct mlxsw_sp_mall_entry *mall_entry;
17
18         list_for_each_entry(mall_entry, &block->mall.list, list)
19                 if (mall_entry->cookie == cookie)
20                         return mall_entry;
21
22         return NULL;
23 }
24
25 static int
26 mlxsw_sp_mall_port_mirror_add(struct mlxsw_sp_port *mlxsw_sp_port,
27                               struct mlxsw_sp_mall_entry *mall_entry,
28                               struct netlink_ext_ack *extack)
29 {
30         struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
31         struct mlxsw_sp_span_agent_parms agent_parms = {};
32         struct mlxsw_sp_span_trigger_parms parms;
33         enum mlxsw_sp_span_trigger trigger;
34         int err;
35
36         if (!mall_entry->mirror.to_dev) {
37                 NL_SET_ERR_MSG(extack, "Could not find requested device");
38                 return -EINVAL;
39         }
40
41         agent_parms.to_dev = mall_entry->mirror.to_dev;
42         err = mlxsw_sp_span_agent_get(mlxsw_sp, &mall_entry->mirror.span_id,
43                                       &agent_parms);
44         if (err) {
45                 NL_SET_ERR_MSG(extack, "Failed to get SPAN agent");
46                 return err;
47         }
48
49         err = mlxsw_sp_span_analyzed_port_get(mlxsw_sp_port,
50                                               mall_entry->ingress);
51         if (err) {
52                 NL_SET_ERR_MSG(extack, "Failed to get analyzed port");
53                 goto err_analyzed_port_get;
54         }
55
56         trigger = mall_entry->ingress ? MLXSW_SP_SPAN_TRIGGER_INGRESS :
57                                         MLXSW_SP_SPAN_TRIGGER_EGRESS;
58         parms.span_id = mall_entry->mirror.span_id;
59         parms.probability_rate = 1;
60         err = mlxsw_sp_span_agent_bind(mlxsw_sp, trigger, mlxsw_sp_port,
61                                        &parms);
62         if (err) {
63                 NL_SET_ERR_MSG(extack, "Failed to bind SPAN agent");
64                 goto err_agent_bind;
65         }
66
67         return 0;
68
69 err_agent_bind:
70         mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, mall_entry->ingress);
71 err_analyzed_port_get:
72         mlxsw_sp_span_agent_put(mlxsw_sp, mall_entry->mirror.span_id);
73         return err;
74 }
75
76 static void
77 mlxsw_sp_mall_port_mirror_del(struct mlxsw_sp_port *mlxsw_sp_port,
78                               struct mlxsw_sp_mall_entry *mall_entry)
79 {
80         struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
81         struct mlxsw_sp_span_trigger_parms parms;
82         enum mlxsw_sp_span_trigger trigger;
83
84         trigger = mall_entry->ingress ? MLXSW_SP_SPAN_TRIGGER_INGRESS :
85                                         MLXSW_SP_SPAN_TRIGGER_EGRESS;
86         parms.span_id = mall_entry->mirror.span_id;
87         mlxsw_sp_span_agent_unbind(mlxsw_sp, trigger, mlxsw_sp_port, &parms);
88         mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, mall_entry->ingress);
89         mlxsw_sp_span_agent_put(mlxsw_sp, mall_entry->mirror.span_id);
90 }
91
92 static int mlxsw_sp_mall_port_sample_set(struct mlxsw_sp_port *mlxsw_sp_port,
93                                          bool enable, u32 rate)
94 {
95         struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
96         char mpsc_pl[MLXSW_REG_MPSC_LEN];
97
98         mlxsw_reg_mpsc_pack(mpsc_pl, mlxsw_sp_port->local_port, enable, rate);
99         return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(mpsc), mpsc_pl);
100 }
101
102 static int
103 mlxsw_sp_mall_port_sample_add(struct mlxsw_sp_port *mlxsw_sp_port,
104                               struct mlxsw_sp_mall_entry *mall_entry,
105                               struct netlink_ext_ack *extack)
106 {
107         struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
108         struct mlxsw_sp_sample_trigger trigger;
109         int err;
110
111         if (mall_entry->ingress)
112                 trigger.type = MLXSW_SP_SAMPLE_TRIGGER_TYPE_INGRESS;
113         else
114                 trigger.type = MLXSW_SP_SAMPLE_TRIGGER_TYPE_EGRESS;
115         trigger.local_port = mlxsw_sp_port->local_port;
116         err = mlxsw_sp_sample_trigger_params_set(mlxsw_sp, &trigger,
117                                                  &mall_entry->sample.params,
118                                                  extack);
119         if (err)
120                 return err;
121
122         err = mlxsw_sp->mall_ops->sample_add(mlxsw_sp, mlxsw_sp_port,
123                                              mall_entry, extack);
124         if (err)
125                 goto err_port_sample_set;
126         return 0;
127
128 err_port_sample_set:
129         mlxsw_sp_sample_trigger_params_unset(mlxsw_sp, &trigger);
130         return err;
131 }
132
133 static void
134 mlxsw_sp_mall_port_sample_del(struct mlxsw_sp_port *mlxsw_sp_port,
135                               struct mlxsw_sp_mall_entry *mall_entry)
136 {
137         struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp;
138         struct mlxsw_sp_sample_trigger trigger;
139
140         if (mall_entry->ingress)
141                 trigger.type = MLXSW_SP_SAMPLE_TRIGGER_TYPE_INGRESS;
142         else
143                 trigger.type = MLXSW_SP_SAMPLE_TRIGGER_TYPE_EGRESS;
144         trigger.local_port = mlxsw_sp_port->local_port;
145
146         mlxsw_sp->mall_ops->sample_del(mlxsw_sp, mlxsw_sp_port, mall_entry);
147         mlxsw_sp_sample_trigger_params_unset(mlxsw_sp, &trigger);
148 }
149
150 static int
151 mlxsw_sp_mall_port_rule_add(struct mlxsw_sp_port *mlxsw_sp_port,
152                             struct mlxsw_sp_mall_entry *mall_entry,
153                             struct netlink_ext_ack *extack)
154 {
155         switch (mall_entry->type) {
156         case MLXSW_SP_MALL_ACTION_TYPE_MIRROR:
157                 return mlxsw_sp_mall_port_mirror_add(mlxsw_sp_port, mall_entry,
158                                                      extack);
159         case MLXSW_SP_MALL_ACTION_TYPE_SAMPLE:
160                 return mlxsw_sp_mall_port_sample_add(mlxsw_sp_port, mall_entry,
161                                                      extack);
162         default:
163                 WARN_ON(1);
164                 return -EINVAL;
165         }
166 }
167
168 static void
169 mlxsw_sp_mall_port_rule_del(struct mlxsw_sp_port *mlxsw_sp_port,
170                             struct mlxsw_sp_mall_entry *mall_entry)
171 {
172         switch (mall_entry->type) {
173         case MLXSW_SP_MALL_ACTION_TYPE_MIRROR:
174                 mlxsw_sp_mall_port_mirror_del(mlxsw_sp_port, mall_entry);
175                 break;
176         case MLXSW_SP_MALL_ACTION_TYPE_SAMPLE:
177                 mlxsw_sp_mall_port_sample_del(mlxsw_sp_port, mall_entry);
178                 break;
179         default:
180                 WARN_ON(1);
181         }
182 }
183
184 static void mlxsw_sp_mall_prio_update(struct mlxsw_sp_flow_block *block)
185 {
186         struct mlxsw_sp_mall_entry *mall_entry;
187
188         if (list_empty(&block->mall.list))
189                 return;
190         block->mall.min_prio = UINT_MAX;
191         block->mall.max_prio = 0;
192         list_for_each_entry(mall_entry, &block->mall.list, list) {
193                 if (mall_entry->priority < block->mall.min_prio)
194                         block->mall.min_prio = mall_entry->priority;
195                 if (mall_entry->priority > block->mall.max_prio)
196                         block->mall.max_prio = mall_entry->priority;
197         }
198 }
199
200 int mlxsw_sp_mall_replace(struct mlxsw_sp *mlxsw_sp,
201                           struct mlxsw_sp_flow_block *block,
202                           struct tc_cls_matchall_offload *f)
203 {
204         struct mlxsw_sp_flow_block_binding *binding;
205         struct mlxsw_sp_mall_entry *mall_entry;
206         __be16 protocol = f->common.protocol;
207         struct flow_action_entry *act;
208         unsigned int flower_min_prio;
209         unsigned int flower_max_prio;
210         bool flower_prio_valid;
211         int err;
212
213         if (!flow_offload_has_one_action(&f->rule->action)) {
214                 NL_SET_ERR_MSG(f->common.extack, "Only singular actions are supported");
215                 return -EOPNOTSUPP;
216         }
217
218         if (f->common.chain_index) {
219                 NL_SET_ERR_MSG(f->common.extack, "Only chain 0 is supported");
220                 return -EOPNOTSUPP;
221         }
222
223         if (mlxsw_sp_flow_block_is_mixed_bound(block)) {
224                 NL_SET_ERR_MSG(f->common.extack, "Only not mixed bound blocks are supported");
225                 return -EOPNOTSUPP;
226         }
227
228         err = mlxsw_sp_flower_prio_get(mlxsw_sp, block, f->common.chain_index,
229                                        &flower_min_prio, &flower_max_prio);
230         if (err) {
231                 if (err != -ENOENT) {
232                         NL_SET_ERR_MSG(f->common.extack, "Failed to get flower priorities");
233                         return err;
234                 }
235                 flower_prio_valid = false;
236                 /* No flower filters are installed in specified chain. */
237         } else {
238                 flower_prio_valid = true;
239         }
240
241         if (protocol != htons(ETH_P_ALL)) {
242                 NL_SET_ERR_MSG(f->common.extack, "matchall rules only supported with 'all' protocol");
243                 return -EOPNOTSUPP;
244         }
245
246         mall_entry = kzalloc(sizeof(*mall_entry), GFP_KERNEL);
247         if (!mall_entry)
248                 return -ENOMEM;
249         mall_entry->cookie = f->cookie;
250         mall_entry->priority = f->common.prio;
251         mall_entry->ingress = mlxsw_sp_flow_block_is_ingress_bound(block);
252
253         if (flower_prio_valid && mall_entry->ingress &&
254             mall_entry->priority >= flower_min_prio) {
255                 NL_SET_ERR_MSG(f->common.extack, "Failed to add behind existing flower rules");
256                 err = -EOPNOTSUPP;
257                 goto errout;
258         }
259         if (flower_prio_valid && !mall_entry->ingress &&
260             mall_entry->priority <= flower_max_prio) {
261                 NL_SET_ERR_MSG(f->common.extack, "Failed to add in front of existing flower rules");
262                 err = -EOPNOTSUPP;
263                 goto errout;
264         }
265
266         act = &f->rule->action.entries[0];
267
268         switch (act->id) {
269         case FLOW_ACTION_MIRRED:
270                 mall_entry->type = MLXSW_SP_MALL_ACTION_TYPE_MIRROR;
271                 mall_entry->mirror.to_dev = act->dev;
272                 break;
273         case FLOW_ACTION_SAMPLE:
274                 mall_entry->type = MLXSW_SP_MALL_ACTION_TYPE_SAMPLE;
275                 mall_entry->sample.params.psample_group = act->sample.psample_group;
276                 mall_entry->sample.params.truncate = act->sample.truncate;
277                 mall_entry->sample.params.trunc_size = act->sample.trunc_size;
278                 mall_entry->sample.params.rate = act->sample.rate;
279                 break;
280         default:
281                 err = -EOPNOTSUPP;
282                 goto errout;
283         }
284
285         list_for_each_entry(binding, &block->binding_list, list) {
286                 err = mlxsw_sp_mall_port_rule_add(binding->mlxsw_sp_port,
287                                                   mall_entry, f->common.extack);
288                 if (err)
289                         goto rollback;
290         }
291
292         block->rule_count++;
293         if (mall_entry->ingress)
294                 block->egress_blocker_rule_count++;
295         else
296                 block->ingress_blocker_rule_count++;
297         list_add_tail(&mall_entry->list, &block->mall.list);
298         mlxsw_sp_mall_prio_update(block);
299         return 0;
300
301 rollback:
302         list_for_each_entry_continue_reverse(binding, &block->binding_list,
303                                              list)
304                 mlxsw_sp_mall_port_rule_del(binding->mlxsw_sp_port, mall_entry);
305 errout:
306         kfree(mall_entry);
307         return err;
308 }
309
310 void mlxsw_sp_mall_destroy(struct mlxsw_sp_flow_block *block,
311                            struct tc_cls_matchall_offload *f)
312 {
313         struct mlxsw_sp_flow_block_binding *binding;
314         struct mlxsw_sp_mall_entry *mall_entry;
315
316         mall_entry = mlxsw_sp_mall_entry_find(block, f->cookie);
317         if (!mall_entry) {
318                 NL_SET_ERR_MSG(f->common.extack, "Entry not found");
319                 return;
320         }
321
322         list_del(&mall_entry->list);
323         if (mall_entry->ingress)
324                 block->egress_blocker_rule_count--;
325         else
326                 block->ingress_blocker_rule_count--;
327         block->rule_count--;
328         list_for_each_entry(binding, &block->binding_list, list)
329                 mlxsw_sp_mall_port_rule_del(binding->mlxsw_sp_port, mall_entry);
330         kfree_rcu(mall_entry, rcu); /* sample RX packets may be in-flight */
331         mlxsw_sp_mall_prio_update(block);
332 }
333
334 int mlxsw_sp_mall_port_bind(struct mlxsw_sp_flow_block *block,
335                             struct mlxsw_sp_port *mlxsw_sp_port,
336                             struct netlink_ext_ack *extack)
337 {
338         struct mlxsw_sp_mall_entry *mall_entry;
339         int err;
340
341         list_for_each_entry(mall_entry, &block->mall.list, list) {
342                 err = mlxsw_sp_mall_port_rule_add(mlxsw_sp_port, mall_entry,
343                                                   extack);
344                 if (err)
345                         goto rollback;
346         }
347         return 0;
348
349 rollback:
350         list_for_each_entry_continue_reverse(mall_entry, &block->mall.list,
351                                              list)
352                 mlxsw_sp_mall_port_rule_del(mlxsw_sp_port, mall_entry);
353         return err;
354 }
355
356 void mlxsw_sp_mall_port_unbind(struct mlxsw_sp_flow_block *block,
357                                struct mlxsw_sp_port *mlxsw_sp_port)
358 {
359         struct mlxsw_sp_mall_entry *mall_entry;
360
361         list_for_each_entry(mall_entry, &block->mall.list, list)
362                 mlxsw_sp_mall_port_rule_del(mlxsw_sp_port, mall_entry);
363 }
364
365 int mlxsw_sp_mall_prio_get(struct mlxsw_sp_flow_block *block, u32 chain_index,
366                            unsigned int *p_min_prio, unsigned int *p_max_prio)
367 {
368         if (chain_index || list_empty(&block->mall.list))
369                 /* In case there are no matchall rules, the caller
370                  * receives -ENOENT to indicate there is no need
371                  * to check the priorities.
372                  */
373                 return -ENOENT;
374         *p_min_prio = block->mall.min_prio;
375         *p_max_prio = block->mall.max_prio;
376         return 0;
377 }
378
379 static int mlxsw_sp1_mall_sample_add(struct mlxsw_sp *mlxsw_sp,
380                                      struct mlxsw_sp_port *mlxsw_sp_port,
381                                      struct mlxsw_sp_mall_entry *mall_entry,
382                                      struct netlink_ext_ack *extack)
383 {
384         u32 rate = mall_entry->sample.params.rate;
385
386         if (!mall_entry->ingress) {
387                 NL_SET_ERR_MSG(extack, "Sampling is not supported on egress");
388                 return -EOPNOTSUPP;
389         }
390
391         if (rate > MLXSW_REG_MPSC_RATE_MAX) {
392                 NL_SET_ERR_MSG(extack, "Unsupported sampling rate");
393                 return -EOPNOTSUPP;
394         }
395
396         return mlxsw_sp_mall_port_sample_set(mlxsw_sp_port, true, rate);
397 }
398
399 static void mlxsw_sp1_mall_sample_del(struct mlxsw_sp *mlxsw_sp,
400                                       struct mlxsw_sp_port *mlxsw_sp_port,
401                                       struct mlxsw_sp_mall_entry *mall_entry)
402 {
403         mlxsw_sp_mall_port_sample_set(mlxsw_sp_port, false, 1);
404 }
405
406 const struct mlxsw_sp_mall_ops mlxsw_sp1_mall_ops = {
407         .sample_add = mlxsw_sp1_mall_sample_add,
408         .sample_del = mlxsw_sp1_mall_sample_del,
409 };
410
411 static int mlxsw_sp2_mall_sample_add(struct mlxsw_sp *mlxsw_sp,
412                                      struct mlxsw_sp_port *mlxsw_sp_port,
413                                      struct mlxsw_sp_mall_entry *mall_entry,
414                                      struct netlink_ext_ack *extack)
415 {
416         struct mlxsw_sp_span_trigger_parms trigger_parms = {};
417         struct mlxsw_sp_span_agent_parms agent_parms = {
418                 .to_dev = NULL, /* Mirror to CPU. */
419                 .session_id = MLXSW_SP_SPAN_SESSION_ID_SAMPLING,
420         };
421         u32 rate = mall_entry->sample.params.rate;
422         enum mlxsw_sp_span_trigger span_trigger;
423         int err;
424
425         err = mlxsw_sp_span_agent_get(mlxsw_sp, &mall_entry->sample.span_id,
426                                       &agent_parms);
427         if (err) {
428                 NL_SET_ERR_MSG(extack, "Failed to get SPAN agent");
429                 return err;
430         }
431
432         err = mlxsw_sp_span_analyzed_port_get(mlxsw_sp_port,
433                                               mall_entry->ingress);
434         if (err) {
435                 NL_SET_ERR_MSG(extack, "Failed to get analyzed port");
436                 goto err_analyzed_port_get;
437         }
438
439         span_trigger = mall_entry->ingress ? MLXSW_SP_SPAN_TRIGGER_INGRESS :
440                                              MLXSW_SP_SPAN_TRIGGER_EGRESS;
441         trigger_parms.span_id = mall_entry->sample.span_id;
442         trigger_parms.probability_rate = rate;
443         err = mlxsw_sp_span_agent_bind(mlxsw_sp, span_trigger, mlxsw_sp_port,
444                                        &trigger_parms);
445         if (err) {
446                 NL_SET_ERR_MSG(extack, "Failed to bind SPAN agent");
447                 goto err_agent_bind;
448         }
449
450         return 0;
451
452 err_agent_bind:
453         mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, mall_entry->ingress);
454 err_analyzed_port_get:
455         mlxsw_sp_span_agent_put(mlxsw_sp, mall_entry->sample.span_id);
456         return err;
457 }
458
459 static void mlxsw_sp2_mall_sample_del(struct mlxsw_sp *mlxsw_sp,
460                                       struct mlxsw_sp_port *mlxsw_sp_port,
461                                       struct mlxsw_sp_mall_entry *mall_entry)
462 {
463         struct mlxsw_sp_span_trigger_parms trigger_parms = {};
464         enum mlxsw_sp_span_trigger span_trigger;
465
466         span_trigger = mall_entry->ingress ? MLXSW_SP_SPAN_TRIGGER_INGRESS :
467                                              MLXSW_SP_SPAN_TRIGGER_EGRESS;
468         trigger_parms.span_id = mall_entry->sample.span_id;
469         mlxsw_sp_span_agent_unbind(mlxsw_sp, span_trigger, mlxsw_sp_port,
470                                    &trigger_parms);
471         mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, mall_entry->ingress);
472         mlxsw_sp_span_agent_put(mlxsw_sp, mall_entry->sample.span_id);
473 }
474
475 const struct mlxsw_sp_mall_ops mlxsw_sp2_mall_ops = {
476         .sample_add = mlxsw_sp2_mall_sample_add,
477         .sample_del = mlxsw_sp2_mall_sample_del,
478 };