582 lines
16 KiB
C
582 lines
16 KiB
C
// SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0
|
|
/* Copyright (c) 2020 Marvell International Ltd. All rights reserved */
|
|
|
|
#include "prestera.h"
|
|
#include "prestera_acl.h"
|
|
#include "prestera_flow.h"
|
|
#include "prestera_flower.h"
|
|
#include "prestera_matchall.h"
|
|
|
|
struct prestera_flower_template {
|
|
struct prestera_acl_ruleset *ruleset;
|
|
struct list_head list;
|
|
u32 chain_index;
|
|
};
|
|
|
|
static void
|
|
prestera_flower_template_free(struct prestera_flower_template *template)
|
|
{
|
|
prestera_acl_ruleset_put(template->ruleset);
|
|
list_del(&template->list);
|
|
kfree(template);
|
|
}
|
|
|
|
void prestera_flower_template_cleanup(struct prestera_flow_block *block)
|
|
{
|
|
struct prestera_flower_template *template, *tmp;
|
|
|
|
/* put the reference to all rulesets kept in tmpl create */
|
|
list_for_each_entry_safe(template, tmp, &block->template_list, list)
|
|
prestera_flower_template_free(template);
|
|
}
|
|
|
|
static int
|
|
prestera_flower_parse_goto_action(struct prestera_flow_block *block,
|
|
struct prestera_acl_rule *rule,
|
|
u32 chain_index,
|
|
const struct flow_action_entry *act)
|
|
{
|
|
struct prestera_acl_ruleset *ruleset;
|
|
|
|
if (act->chain_index <= chain_index)
|
|
/* we can jump only forward */
|
|
return -EINVAL;
|
|
|
|
if (rule->re_arg.jump.valid)
|
|
return -EEXIST;
|
|
|
|
ruleset = prestera_acl_ruleset_get(block->sw->acl, block,
|
|
act->chain_index);
|
|
if (IS_ERR(ruleset))
|
|
return PTR_ERR(ruleset);
|
|
|
|
rule->re_arg.jump.valid = 1;
|
|
rule->re_arg.jump.i.index = prestera_acl_ruleset_index_get(ruleset);
|
|
|
|
rule->jump_ruleset = ruleset;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int prestera_flower_parse_actions(struct prestera_flow_block *block,
|
|
struct prestera_acl_rule *rule,
|
|
struct flow_action *flow_action,
|
|
u32 chain_index,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
const struct flow_action_entry *act;
|
|
int err, i;
|
|
|
|
/* whole struct (rule->re_arg) must be initialized with 0 */
|
|
if (!flow_action_has_entries(flow_action))
|
|
return 0;
|
|
|
|
if (!flow_action_mixed_hw_stats_check(flow_action, extack))
|
|
return -EOPNOTSUPP;
|
|
|
|
act = flow_action_first_entry_get(flow_action);
|
|
if (act->hw_stats & FLOW_ACTION_HW_STATS_DISABLED) {
|
|
/* Nothing to do */
|
|
} else if (act->hw_stats & FLOW_ACTION_HW_STATS_DELAYED) {
|
|
/* setup counter first */
|
|
rule->re_arg.count.valid = true;
|
|
err = prestera_acl_chain_to_client(chain_index, block->ingress,
|
|
&rule->re_arg.count.client);
|
|
if (err)
|
|
return err;
|
|
} else {
|
|
NL_SET_ERR_MSG_MOD(extack, "Unsupported action HW stats type");
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
flow_action_for_each(i, act, flow_action) {
|
|
switch (act->id) {
|
|
case FLOW_ACTION_ACCEPT:
|
|
if (rule->re_arg.accept.valid)
|
|
return -EEXIST;
|
|
|
|
rule->re_arg.accept.valid = 1;
|
|
break;
|
|
case FLOW_ACTION_DROP:
|
|
if (rule->re_arg.drop.valid)
|
|
return -EEXIST;
|
|
|
|
rule->re_arg.drop.valid = 1;
|
|
break;
|
|
case FLOW_ACTION_TRAP:
|
|
if (rule->re_arg.trap.valid)
|
|
return -EEXIST;
|
|
|
|
rule->re_arg.trap.valid = 1;
|
|
break;
|
|
case FLOW_ACTION_POLICE:
|
|
if (rule->re_arg.police.valid)
|
|
return -EEXIST;
|
|
|
|
rule->re_arg.police.valid = 1;
|
|
rule->re_arg.police.rate =
|
|
act->police.rate_bytes_ps;
|
|
rule->re_arg.police.burst = act->police.burst;
|
|
rule->re_arg.police.ingress = block->ingress;
|
|
break;
|
|
case FLOW_ACTION_GOTO:
|
|
err = prestera_flower_parse_goto_action(block, rule,
|
|
chain_index,
|
|
act);
|
|
if (err)
|
|
return err;
|
|
break;
|
|
default:
|
|
NL_SET_ERR_MSG_MOD(extack, "Unsupported action");
|
|
pr_err("Unsupported action\n");
|
|
return -EOPNOTSUPP;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int prestera_flower_parse_meta(struct prestera_acl_rule *rule,
|
|
struct flow_cls_offload *f,
|
|
struct prestera_flow_block *block)
|
|
{
|
|
struct flow_rule *f_rule = flow_cls_offload_flow_rule(f);
|
|
struct prestera_acl_match *r_match = &rule->re_key.match;
|
|
struct prestera_port *port;
|
|
struct net_device *ingress_dev;
|
|
struct flow_match_meta match;
|
|
__be16 key, mask;
|
|
|
|
flow_rule_match_meta(f_rule, &match);
|
|
|
|
if (match.mask->l2_miss) {
|
|
NL_SET_ERR_MSG_MOD(f->common.extack, "Can't match on \"l2_miss\"");
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
if (match.mask->ingress_ifindex != 0xFFFFFFFF) {
|
|
NL_SET_ERR_MSG_MOD(f->common.extack,
|
|
"Unsupported ingress ifindex mask");
|
|
return -EINVAL;
|
|
}
|
|
|
|
ingress_dev = __dev_get_by_index(block->net,
|
|
match.key->ingress_ifindex);
|
|
if (!ingress_dev) {
|
|
NL_SET_ERR_MSG_MOD(f->common.extack,
|
|
"Can't find specified ingress port to match on");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!prestera_netdev_check(ingress_dev)) {
|
|
NL_SET_ERR_MSG_MOD(f->common.extack,
|
|
"Can't match on switchdev ingress port");
|
|
return -EINVAL;
|
|
}
|
|
port = netdev_priv(ingress_dev);
|
|
|
|
mask = htons(0x1FFF << 3);
|
|
key = htons(port->hw_id << 3);
|
|
rule_match_set(r_match->key, SYS_PORT, key);
|
|
rule_match_set(r_match->mask, SYS_PORT, mask);
|
|
|
|
mask = htons(0x3FF);
|
|
key = htons(port->dev_id);
|
|
rule_match_set(r_match->key, SYS_DEV, key);
|
|
rule_match_set(r_match->mask, SYS_DEV, mask);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int prestera_flower_parse(struct prestera_flow_block *block,
|
|
struct prestera_acl_rule *rule,
|
|
struct flow_cls_offload *f)
|
|
{
|
|
struct flow_rule *f_rule = flow_cls_offload_flow_rule(f);
|
|
struct flow_dissector *dissector = f_rule->match.dissector;
|
|
struct prestera_acl_match *r_match = &rule->re_key.match;
|
|
__be16 n_proto_mask = 0;
|
|
__be16 n_proto_key = 0;
|
|
u16 addr_type = 0;
|
|
u8 ip_proto = 0;
|
|
int err;
|
|
|
|
if (dissector->used_keys &
|
|
~(BIT(FLOW_DISSECTOR_KEY_META) |
|
|
BIT(FLOW_DISSECTOR_KEY_CONTROL) |
|
|
BIT(FLOW_DISSECTOR_KEY_BASIC) |
|
|
BIT(FLOW_DISSECTOR_KEY_ETH_ADDRS) |
|
|
BIT(FLOW_DISSECTOR_KEY_IPV4_ADDRS) |
|
|
BIT(FLOW_DISSECTOR_KEY_IPV6_ADDRS) |
|
|
BIT(FLOW_DISSECTOR_KEY_ICMP) |
|
|
BIT(FLOW_DISSECTOR_KEY_PORTS) |
|
|
BIT(FLOW_DISSECTOR_KEY_PORTS_RANGE) |
|
|
BIT(FLOW_DISSECTOR_KEY_VLAN))) {
|
|
NL_SET_ERR_MSG_MOD(f->common.extack, "Unsupported key");
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
prestera_acl_rule_priority_set(rule, f->common.prio);
|
|
|
|
if (flow_rule_match_key(f_rule, FLOW_DISSECTOR_KEY_META)) {
|
|
err = prestera_flower_parse_meta(rule, f, block);
|
|
if (err)
|
|
return err;
|
|
}
|
|
|
|
if (flow_rule_match_key(f_rule, FLOW_DISSECTOR_KEY_CONTROL)) {
|
|
struct flow_match_control match;
|
|
|
|
flow_rule_match_control(f_rule, &match);
|
|
addr_type = match.key->addr_type;
|
|
}
|
|
|
|
if (flow_rule_match_key(f_rule, FLOW_DISSECTOR_KEY_BASIC)) {
|
|
struct flow_match_basic match;
|
|
|
|
flow_rule_match_basic(f_rule, &match);
|
|
n_proto_key = match.key->n_proto;
|
|
n_proto_mask = match.mask->n_proto;
|
|
|
|
if (ntohs(match.key->n_proto) == ETH_P_ALL) {
|
|
n_proto_key = 0;
|
|
n_proto_mask = 0;
|
|
}
|
|
|
|
rule_match_set(r_match->key, ETH_TYPE, n_proto_key);
|
|
rule_match_set(r_match->mask, ETH_TYPE, n_proto_mask);
|
|
|
|
rule_match_set(r_match->key, IP_PROTO, match.key->ip_proto);
|
|
rule_match_set(r_match->mask, IP_PROTO, match.mask->ip_proto);
|
|
ip_proto = match.key->ip_proto;
|
|
}
|
|
|
|
if (flow_rule_match_key(f_rule, FLOW_DISSECTOR_KEY_ETH_ADDRS)) {
|
|
struct flow_match_eth_addrs match;
|
|
|
|
flow_rule_match_eth_addrs(f_rule, &match);
|
|
|
|
/* DA key, mask */
|
|
rule_match_set_n(r_match->key,
|
|
ETH_DMAC_0, &match.key->dst[0], 4);
|
|
rule_match_set_n(r_match->key,
|
|
ETH_DMAC_1, &match.key->dst[4], 2);
|
|
|
|
rule_match_set_n(r_match->mask,
|
|
ETH_DMAC_0, &match.mask->dst[0], 4);
|
|
rule_match_set_n(r_match->mask,
|
|
ETH_DMAC_1, &match.mask->dst[4], 2);
|
|
|
|
/* SA key, mask */
|
|
rule_match_set_n(r_match->key,
|
|
ETH_SMAC_0, &match.key->src[0], 4);
|
|
rule_match_set_n(r_match->key,
|
|
ETH_SMAC_1, &match.key->src[4], 2);
|
|
|
|
rule_match_set_n(r_match->mask,
|
|
ETH_SMAC_0, &match.mask->src[0], 4);
|
|
rule_match_set_n(r_match->mask,
|
|
ETH_SMAC_1, &match.mask->src[4], 2);
|
|
}
|
|
|
|
if (addr_type == FLOW_DISSECTOR_KEY_IPV4_ADDRS) {
|
|
struct flow_match_ipv4_addrs match;
|
|
|
|
flow_rule_match_ipv4_addrs(f_rule, &match);
|
|
|
|
rule_match_set(r_match->key, IP_SRC, match.key->src);
|
|
rule_match_set(r_match->mask, IP_SRC, match.mask->src);
|
|
|
|
rule_match_set(r_match->key, IP_DST, match.key->dst);
|
|
rule_match_set(r_match->mask, IP_DST, match.mask->dst);
|
|
}
|
|
|
|
if (flow_rule_match_key(f_rule, FLOW_DISSECTOR_KEY_PORTS)) {
|
|
struct flow_match_ports match;
|
|
|
|
if (ip_proto != IPPROTO_TCP && ip_proto != IPPROTO_UDP) {
|
|
NL_SET_ERR_MSG_MOD
|
|
(f->common.extack,
|
|
"Only UDP and TCP keys are supported");
|
|
return -EINVAL;
|
|
}
|
|
|
|
flow_rule_match_ports(f_rule, &match);
|
|
|
|
rule_match_set(r_match->key, L4_PORT_SRC, match.key->src);
|
|
rule_match_set(r_match->mask, L4_PORT_SRC, match.mask->src);
|
|
|
|
rule_match_set(r_match->key, L4_PORT_DST, match.key->dst);
|
|
rule_match_set(r_match->mask, L4_PORT_DST, match.mask->dst);
|
|
}
|
|
|
|
if (flow_rule_match_key(f_rule, FLOW_DISSECTOR_KEY_PORTS_RANGE)) {
|
|
struct flow_match_ports_range match;
|
|
__be32 tp_key, tp_mask;
|
|
|
|
flow_rule_match_ports_range(f_rule, &match);
|
|
|
|
/* src port range (min, max) */
|
|
tp_key = htonl(ntohs(match.key->tp_min.src) |
|
|
(ntohs(match.key->tp_max.src) << 16));
|
|
tp_mask = htonl(ntohs(match.mask->tp_min.src) |
|
|
(ntohs(match.mask->tp_max.src) << 16));
|
|
rule_match_set(r_match->key, L4_PORT_RANGE_SRC, tp_key);
|
|
rule_match_set(r_match->mask, L4_PORT_RANGE_SRC, tp_mask);
|
|
|
|
/* dst port range (min, max) */
|
|
tp_key = htonl(ntohs(match.key->tp_min.dst) |
|
|
(ntohs(match.key->tp_max.dst) << 16));
|
|
tp_mask = htonl(ntohs(match.mask->tp_min.dst) |
|
|
(ntohs(match.mask->tp_max.dst) << 16));
|
|
rule_match_set(r_match->key, L4_PORT_RANGE_DST, tp_key);
|
|
rule_match_set(r_match->mask, L4_PORT_RANGE_DST, tp_mask);
|
|
}
|
|
|
|
if (flow_rule_match_key(f_rule, FLOW_DISSECTOR_KEY_VLAN)) {
|
|
struct flow_match_vlan match;
|
|
|
|
flow_rule_match_vlan(f_rule, &match);
|
|
|
|
if (match.mask->vlan_id != 0) {
|
|
__be16 key = cpu_to_be16(match.key->vlan_id);
|
|
__be16 mask = cpu_to_be16(match.mask->vlan_id);
|
|
|
|
rule_match_set(r_match->key, VLAN_ID, key);
|
|
rule_match_set(r_match->mask, VLAN_ID, mask);
|
|
}
|
|
|
|
rule_match_set(r_match->key, VLAN_TPID, match.key->vlan_tpid);
|
|
rule_match_set(r_match->mask, VLAN_TPID, match.mask->vlan_tpid);
|
|
}
|
|
|
|
if (flow_rule_match_key(f_rule, FLOW_DISSECTOR_KEY_ICMP)) {
|
|
struct flow_match_icmp match;
|
|
|
|
flow_rule_match_icmp(f_rule, &match);
|
|
|
|
rule_match_set(r_match->key, ICMP_TYPE, match.key->type);
|
|
rule_match_set(r_match->mask, ICMP_TYPE, match.mask->type);
|
|
|
|
rule_match_set(r_match->key, ICMP_CODE, match.key->code);
|
|
rule_match_set(r_match->mask, ICMP_CODE, match.mask->code);
|
|
}
|
|
|
|
return prestera_flower_parse_actions(block, rule, &f->rule->action,
|
|
f->common.chain_index,
|
|
f->common.extack);
|
|
}
|
|
|
|
static int prestera_flower_prio_check(struct prestera_flow_block *block,
|
|
struct flow_cls_offload *f)
|
|
{
|
|
u32 mall_prio_min;
|
|
u32 mall_prio_max;
|
|
int err;
|
|
|
|
err = prestera_mall_prio_get(block, &mall_prio_min, &mall_prio_max);
|
|
if (err == -ENOENT)
|
|
/* No matchall filters installed on this chain. */
|
|
return 0;
|
|
|
|
if (err) {
|
|
NL_SET_ERR_MSG(f->common.extack, "Failed to get matchall priorities");
|
|
return err;
|
|
}
|
|
|
|
if (f->common.prio <= mall_prio_max && block->ingress) {
|
|
NL_SET_ERR_MSG(f->common.extack,
|
|
"Failed to add in front of existing matchall rules");
|
|
return -EOPNOTSUPP;
|
|
}
|
|
if (f->common.prio >= mall_prio_min && !block->ingress) {
|
|
NL_SET_ERR_MSG(f->common.extack, "Failed to add behind of existing matchall rules");
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int prestera_flower_prio_get(struct prestera_flow_block *block, u32 chain_index,
|
|
u32 *prio_min, u32 *prio_max)
|
|
{
|
|
struct prestera_acl_ruleset *ruleset;
|
|
|
|
ruleset = prestera_acl_ruleset_lookup(block->sw->acl, block, chain_index);
|
|
if (IS_ERR(ruleset))
|
|
return PTR_ERR(ruleset);
|
|
|
|
prestera_acl_ruleset_prio_get(ruleset, prio_min, prio_max);
|
|
return 0;
|
|
}
|
|
|
|
int prestera_flower_replace(struct prestera_flow_block *block,
|
|
struct flow_cls_offload *f)
|
|
{
|
|
struct prestera_acl_ruleset *ruleset;
|
|
struct prestera_acl *acl = block->sw->acl;
|
|
struct prestera_acl_rule *rule;
|
|
int err;
|
|
|
|
err = prestera_flower_prio_check(block, f);
|
|
if (err)
|
|
return err;
|
|
|
|
ruleset = prestera_acl_ruleset_get(acl, block, f->common.chain_index);
|
|
if (IS_ERR(ruleset))
|
|
return PTR_ERR(ruleset);
|
|
|
|
/* increments the ruleset reference */
|
|
rule = prestera_acl_rule_create(ruleset, f->cookie,
|
|
f->common.chain_index);
|
|
if (IS_ERR(rule)) {
|
|
err = PTR_ERR(rule);
|
|
goto err_rule_create;
|
|
}
|
|
|
|
err = prestera_flower_parse(block, rule, f);
|
|
if (err)
|
|
goto err_rule_add;
|
|
|
|
if (!prestera_acl_ruleset_is_offload(ruleset)) {
|
|
err = prestera_acl_ruleset_offload(ruleset);
|
|
if (err)
|
|
goto err_ruleset_offload;
|
|
}
|
|
|
|
err = prestera_acl_rule_add(block->sw, rule);
|
|
if (err)
|
|
goto err_rule_add;
|
|
|
|
prestera_acl_ruleset_put(ruleset);
|
|
return 0;
|
|
|
|
err_ruleset_offload:
|
|
err_rule_add:
|
|
prestera_acl_rule_destroy(rule);
|
|
err_rule_create:
|
|
prestera_acl_ruleset_put(ruleset);
|
|
return err;
|
|
}
|
|
|
|
void prestera_flower_destroy(struct prestera_flow_block *block,
|
|
struct flow_cls_offload *f)
|
|
{
|
|
struct prestera_acl_ruleset *ruleset;
|
|
struct prestera_acl_rule *rule;
|
|
|
|
ruleset = prestera_acl_ruleset_lookup(block->sw->acl, block,
|
|
f->common.chain_index);
|
|
if (IS_ERR(ruleset))
|
|
return;
|
|
|
|
rule = prestera_acl_rule_lookup(ruleset, f->cookie);
|
|
if (rule) {
|
|
prestera_acl_rule_del(block->sw, rule);
|
|
prestera_acl_rule_destroy(rule);
|
|
}
|
|
prestera_acl_ruleset_put(ruleset);
|
|
}
|
|
|
|
int prestera_flower_tmplt_create(struct prestera_flow_block *block,
|
|
struct flow_cls_offload *f)
|
|
{
|
|
struct prestera_flower_template *template;
|
|
struct prestera_acl_ruleset *ruleset;
|
|
struct prestera_acl_rule rule;
|
|
int err;
|
|
|
|
memset(&rule, 0, sizeof(rule));
|
|
err = prestera_flower_parse(block, &rule, f);
|
|
if (err)
|
|
return err;
|
|
|
|
template = kmalloc(sizeof(*template), GFP_KERNEL);
|
|
if (!template) {
|
|
err = -ENOMEM;
|
|
goto err_malloc;
|
|
}
|
|
|
|
prestera_acl_rule_keymask_pcl_id_set(&rule, 0);
|
|
ruleset = prestera_acl_ruleset_get(block->sw->acl, block,
|
|
f->common.chain_index);
|
|
if (IS_ERR_OR_NULL(ruleset)) {
|
|
err = -EINVAL;
|
|
goto err_ruleset_get;
|
|
}
|
|
|
|
/* preserve keymask/template to this ruleset */
|
|
err = prestera_acl_ruleset_keymask_set(ruleset, rule.re_key.match.mask);
|
|
if (err)
|
|
goto err_ruleset_keymask_set;
|
|
|
|
/* skip error, as it is not possible to reject template operation,
|
|
* so, keep the reference to the ruleset for rules to be added
|
|
* to that ruleset later. In case of offload fail, the ruleset
|
|
* will be offloaded again during adding a new rule. Also,
|
|
* unlikly possble that ruleset is already offloaded at this staage.
|
|
*/
|
|
prestera_acl_ruleset_offload(ruleset);
|
|
|
|
/* keep the reference to the ruleset */
|
|
template->ruleset = ruleset;
|
|
template->chain_index = f->common.chain_index;
|
|
list_add_rcu(&template->list, &block->template_list);
|
|
return 0;
|
|
|
|
err_ruleset_keymask_set:
|
|
prestera_acl_ruleset_put(ruleset);
|
|
err_ruleset_get:
|
|
kfree(template);
|
|
err_malloc:
|
|
NL_SET_ERR_MSG_MOD(f->common.extack, "Create chain template failed");
|
|
return err;
|
|
}
|
|
|
|
void prestera_flower_tmplt_destroy(struct prestera_flow_block *block,
|
|
struct flow_cls_offload *f)
|
|
{
|
|
struct prestera_flower_template *template, *tmp;
|
|
|
|
list_for_each_entry_safe(template, tmp, &block->template_list, list)
|
|
if (template->chain_index == f->common.chain_index) {
|
|
/* put the reference to the ruleset kept in create */
|
|
prestera_flower_template_free(template);
|
|
return;
|
|
}
|
|
}
|
|
|
|
int prestera_flower_stats(struct prestera_flow_block *block,
|
|
struct flow_cls_offload *f)
|
|
{
|
|
struct prestera_acl_ruleset *ruleset;
|
|
struct prestera_acl_rule *rule;
|
|
u64 packets;
|
|
u64 lastuse;
|
|
u64 bytes;
|
|
int err;
|
|
|
|
ruleset = prestera_acl_ruleset_lookup(block->sw->acl, block,
|
|
f->common.chain_index);
|
|
if (IS_ERR(ruleset))
|
|
return PTR_ERR(ruleset);
|
|
|
|
rule = prestera_acl_rule_lookup(ruleset, f->cookie);
|
|
if (!rule) {
|
|
err = -EINVAL;
|
|
goto err_rule_get_stats;
|
|
}
|
|
|
|
err = prestera_acl_rule_get_stats(block->sw->acl, rule, &packets,
|
|
&bytes, &lastuse);
|
|
if (err)
|
|
goto err_rule_get_stats;
|
|
|
|
flow_stats_update(&f->stats, bytes, packets, 0, lastuse,
|
|
FLOW_ACTION_HW_STATS_DELAYED);
|
|
|
|
err_rule_get_stats:
|
|
prestera_acl_ruleset_put(ruleset);
|
|
return err;
|
|
}
|