X-Git-Url: http://git.onelab.eu/?a=blobdiff_plain;f=lib%2Fclassifier.c;h=658be8086395105c7c26fffaa484c7a7e65a2e01;hb=d31f1109f10e5ffb9bf266306b913ebf23781666;hp=42f026ff9d859b15766edd9e1293d6bdfeb0dc1c;hpb=f9bfea1402fb4776e505eba7bb712c8355e01e70;p=sliver-openvswitch.git diff --git a/lib/classifier.c b/lib/classifier.c index 42f026ff9..658be8086 100644 --- a/lib/classifier.c +++ b/lib/classifier.c @@ -1,5 +1,5 @@ /* - * Copyright (c) 2009, 2010 Nicira Networks. + * Copyright (c) 2009, 2010, 2011 Nicira Networks. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. @@ -19,9 +19,12 @@ #include #include #include +#include "byte-order.h" #include "dynamic-string.h" #include "flow.h" #include "hash.h" +#include "odp-util.h" +#include "ofp-util.h" #include "packets.h" static struct cls_table *find_table(const struct classifier *, @@ -34,8 +37,6 @@ static struct cls_table *classifier_next_table(const struct classifier *, const struct cls_table *); static void destroy_table(struct classifier *, struct cls_table *); -static bool should_include(const struct cls_table *, int include); - static struct cls_rule *find_match(const struct cls_table *, const struct flow *); static struct cls_rule *find_equal(struct cls_table *, const struct flow *, @@ -54,6 +55,7 @@ static void zero_wildcards(struct flow *, const struct flow_wildcards *); (RULE) != NULL && ((NEXT) = next_rule_in_list(RULE), true); \ (RULE) = (NEXT)) +static struct cls_rule *next_rule_in_list__(struct cls_rule *); static struct cls_rule *next_rule_in_list(struct cls_rule *); static struct cls_table * @@ -62,90 +64,540 @@ cls_table_from_hmap_node(const struct hmap_node *node) return node ? CONTAINER_OF(node, struct cls_table, hmap_node) : NULL; } -static struct cls_rule * -cls_rule_from_hmap_node(const struct hmap_node *node) +/* Converts the flow in 'flow' into a cls_rule in 'rule', with the given + * 'wildcards' and 'priority'. */ +void +cls_rule_init(const struct flow *flow, const struct flow_wildcards *wildcards, + unsigned int priority, struct cls_rule *rule) { - return node ? CONTAINER_OF(node, struct cls_rule, hmap_node) : NULL; + rule->flow = *flow; + rule->wc = *wildcards; + rule->priority = priority; + cls_rule_zero_wildcarded_fields(rule); } -/* Returns the cls_table within 'cls' that has no wildcards, or NULL if there - * is none. */ -struct cls_table * -classifier_exact_table(const struct classifier *cls) +/* Converts the flow in 'flow' into an exact-match cls_rule in 'rule', with the + * given 'priority'. (For OpenFlow 1.0, exact-match rule are always highest + * priority, so 'priority' should be at least 65535.) */ +void +cls_rule_init_exact(const struct flow *flow, + unsigned int priority, struct cls_rule *rule) { - struct flow_wildcards exact_wc; - flow_wildcards_init_exact(&exact_wc); - return find_table(cls, &exact_wc); + rule->flow = *flow; + flow_wildcards_init_exact(&rule->wc); + rule->priority = priority; } -/* Returns the first rule in 'table', or a null pointer if 'table' is NULL. */ -struct cls_rule * -cls_table_first_rule(const struct cls_table *table) +/* Initializes 'rule' as a "catch-all" rule that matches every packet, with + * priority 'priority'. */ +void +cls_rule_init_catchall(struct cls_rule *rule, unsigned int priority) { - return table ? cls_rule_from_hmap_node(hmap_first(&table->rules)) : NULL; + memset(&rule->flow, 0, sizeof rule->flow); + flow_wildcards_init_catchall(&rule->wc); + rule->priority = priority; } -/* Returns the next rule in 'table' following 'rule', or a null pointer if - * 'rule' is the last rule in 'table'. */ -struct cls_rule * -cls_table_next_rule(const struct cls_table *table, const struct cls_rule *rule) +/* For each bit or field wildcarded in 'rule', sets the corresponding bit or + * field in 'flow' to all-0-bits. It is important to maintain this invariant + * in a clr_rule that might be inserted into a classifier. + * + * It is never necessary to call this function directly for a cls_rule that is + * initialized or modified only by cls_rule_*() functions. It is useful to + * restore the invariant in a cls_rule whose 'wc' member is modified by hand. + */ +void +cls_rule_zero_wildcarded_fields(struct cls_rule *rule) { - struct cls_rule *next - = CONTAINER_OF(rule->list.next, struct cls_rule, hmap_node); + zero_wildcards(&rule->flow, &rule->wc); +} - return (next->priority < rule->priority - ? next - : cls_rule_from_hmap_node(hmap_next(&table->rules, - &next->hmap_node))); +void +cls_rule_set_reg(struct cls_rule *rule, unsigned int reg_idx, uint32_t value) +{ + cls_rule_set_reg_masked(rule, reg_idx, value, UINT32_MAX); } -static void -cls_rule_init__(struct cls_rule *rule, - const struct flow *flow, uint32_t wildcards) +void +cls_rule_set_reg_masked(struct cls_rule *rule, unsigned int reg_idx, + uint32_t value, uint32_t mask) { - rule->flow = *flow; - flow_wildcards_init(&rule->wc, wildcards); - cls_rule_zero_wildcards(rule); + assert(reg_idx < FLOW_N_REGS); + flow_wildcards_set_reg_mask(&rule->wc, reg_idx, mask); + rule->flow.regs[reg_idx] = value & mask; } -/* Converts the flow in 'flow' into a cls_rule in 'rule', with the given - * 'wildcards' and 'priority'.*/ void -cls_rule_from_flow(const struct flow *flow, uint32_t wildcards, - unsigned int priority, struct cls_rule *rule) +cls_rule_set_tun_id(struct cls_rule *rule, ovs_be64 tun_id) { - cls_rule_init__(rule, flow, wildcards); - rule->priority = priority; + cls_rule_set_tun_id_masked(rule, tun_id, htonll(UINT64_MAX)); } -/* Converts the ofp_match in 'match' (with format 'flow_format', one of NXFF_*) - * into a cls_rule in 'rule', with the given 'priority'. 'cookie' is used - * when 'flow_format' is NXFF_TUN_ID_FROM_COOKIE. */ void -cls_rule_from_match(const struct ofp_match *match, unsigned int priority, - int flow_format, uint64_t cookie, - struct cls_rule *rule) +cls_rule_set_tun_id_masked(struct cls_rule *rule, + ovs_be64 tun_id, ovs_be64 mask) { - uint32_t wildcards; - struct flow flow; + rule->wc.tun_id_mask = mask; + rule->flow.tun_id = tun_id & mask; +} - flow_from_match(match, flow_format, cookie, &flow, &wildcards); - cls_rule_init__(rule, &flow, wildcards); - rule->priority = rule->wc.wildcards ? priority : UINT16_MAX; +void +cls_rule_set_in_port(struct cls_rule *rule, uint16_t odp_port) +{ + rule->wc.wildcards &= ~FWW_IN_PORT; + rule->flow.in_port = odp_port; } -/* For each bit or field wildcarded in 'rule', sets the corresponding bit or - * field in 'flow' to all-0-bits. It is important to maintain this invariant - * in a clr_rule that might be inserted into a classifier. +void +cls_rule_set_dl_type(struct cls_rule *rule, ovs_be16 dl_type) +{ + rule->wc.wildcards &= ~FWW_DL_TYPE; + rule->flow.dl_type = dl_type; +} + +void +cls_rule_set_dl_src(struct cls_rule *rule, const uint8_t dl_src[ETH_ADDR_LEN]) +{ + rule->wc.wildcards &= ~FWW_DL_SRC; + memcpy(rule->flow.dl_src, dl_src, ETH_ADDR_LEN); +} + +void +cls_rule_set_dl_dst(struct cls_rule *rule, const uint8_t dl_dst[ETH_ADDR_LEN]) +{ + rule->wc.wildcards &= ~(FWW_DL_DST | FWW_ETH_MCAST); + memcpy(rule->flow.dl_dst, dl_dst, ETH_ADDR_LEN); +} + +void +cls_rule_set_dl_tci(struct cls_rule *rule, ovs_be16 tci) +{ + cls_rule_set_dl_tci_masked(rule, tci, htons(0xffff)); +} + +void +cls_rule_set_dl_tci_masked(struct cls_rule *rule, ovs_be16 tci, ovs_be16 mask) +{ + rule->flow.vlan_tci = tci & mask; + rule->wc.vlan_tci_mask = mask; +} + +/* Modifies 'rule' so that the VLAN VID is wildcarded. If the PCP is already + * wildcarded, then 'rule' will match a packet regardless of whether it has an + * 802.1Q header or not. */ +void +cls_rule_set_any_vid(struct cls_rule *rule) +{ + if (rule->wc.vlan_tci_mask & htons(VLAN_PCP_MASK)) { + rule->wc.vlan_tci_mask &= ~htons(VLAN_VID_MASK); + rule->flow.vlan_tci &= ~htons(VLAN_VID_MASK); + } else { + cls_rule_set_dl_tci_masked(rule, htons(0), htons(0)); + } +} + +/* Modifies 'rule' depending on 'dl_vlan': * - * It is never necessary to call this function directly for a cls_rule that is - * initialized or modified only by cls_rule_*() functions. It is useful to - * restore the invariant in a cls_rule whose 'wc' member is modified by hand. + * - If 'dl_vlan' is htons(OFP_VLAN_NONE), makes 'rule' match only packets + * without an 802.1Q header. + * + * - Otherwise, makes 'rule' match only packets with an 802.1Q header whose + * VID equals the low 12 bits of 'dl_vlan'. */ void -cls_rule_zero_wildcards(struct cls_rule *rule) +cls_rule_set_dl_vlan(struct cls_rule *rule, ovs_be16 dl_vlan) { - zero_wildcards(&rule->flow, &rule->wc); + if (dl_vlan == htons(OFP_VLAN_NONE)) { + cls_rule_set_dl_tci(rule, htons(0)); + } else { + dl_vlan &= htons(VLAN_VID_MASK); + rule->flow.vlan_tci &= ~htons(VLAN_VID_MASK); + rule->flow.vlan_tci |= htons(VLAN_CFI) | dl_vlan; + rule->wc.vlan_tci_mask |= htons(VLAN_VID_MASK | VLAN_CFI); + } +} + +/* Modifies 'rule' so that the VLAN PCP is wildcarded. If the VID is already + * wildcarded, then 'rule' will match a packet regardless of whether it has an + * 802.1Q header or not. */ +void +cls_rule_set_any_pcp(struct cls_rule *rule) +{ + if (rule->wc.vlan_tci_mask & htons(VLAN_VID_MASK)) { + rule->wc.vlan_tci_mask &= ~htons(VLAN_PCP_MASK); + rule->flow.vlan_tci &= ~htons(VLAN_PCP_MASK); + } else { + cls_rule_set_dl_tci_masked(rule, htons(0), htons(0)); + } +} + +/* Modifies 'rule' so that it matches only packets with an 802.1Q header whose + * PCP equals the low 3 bits of 'dl_vlan_pcp'. */ +void +cls_rule_set_dl_vlan_pcp(struct cls_rule *rule, uint8_t dl_vlan_pcp) +{ + dl_vlan_pcp &= 0x07; + rule->flow.vlan_tci &= ~htons(VLAN_PCP_MASK); + rule->flow.vlan_tci |= htons((dl_vlan_pcp << VLAN_PCP_SHIFT) | VLAN_CFI); + rule->wc.vlan_tci_mask |= htons(VLAN_CFI | VLAN_PCP_MASK); +} + +void +cls_rule_set_tp_src(struct cls_rule *rule, ovs_be16 tp_src) +{ + rule->wc.wildcards &= ~FWW_TP_SRC; + rule->flow.tp_src = tp_src; +} + +void +cls_rule_set_tp_dst(struct cls_rule *rule, ovs_be16 tp_dst) +{ + rule->wc.wildcards &= ~FWW_TP_DST; + rule->flow.tp_dst = tp_dst; +} + +void +cls_rule_set_nw_proto(struct cls_rule *rule, uint8_t nw_proto) +{ + rule->wc.wildcards &= ~FWW_NW_PROTO; + rule->flow.nw_proto = nw_proto; +} + +void +cls_rule_set_nw_src(struct cls_rule *rule, ovs_be32 nw_src) +{ + cls_rule_set_nw_src_masked(rule, nw_src, htonl(UINT32_MAX)); +} + +bool +cls_rule_set_nw_src_masked(struct cls_rule *rule, ovs_be32 ip, ovs_be32 mask) +{ + if (flow_wildcards_set_nw_src_mask(&rule->wc, mask)) { + rule->flow.nw_src = ip & mask; + return true; + } else { + return false; + } +} + +void +cls_rule_set_nw_dst(struct cls_rule *rule, ovs_be32 nw_dst) +{ + cls_rule_set_nw_dst_masked(rule, nw_dst, htonl(UINT32_MAX)); +} + +bool +cls_rule_set_nw_dst_masked(struct cls_rule *rule, ovs_be32 ip, ovs_be32 mask) +{ + if (flow_wildcards_set_nw_dst_mask(&rule->wc, mask)) { + rule->flow.nw_dst = ip & mask; + return true; + } else { + return false; + } +} + +void +cls_rule_set_nw_tos(struct cls_rule *rule, uint8_t nw_tos) +{ + rule->wc.wildcards &= ~FWW_NW_TOS; + rule->flow.nw_tos = nw_tos & IP_DSCP_MASK; +} + +void +cls_rule_set_icmp_type(struct cls_rule *rule, uint8_t icmp_type) +{ + rule->wc.wildcards &= ~FWW_TP_SRC; + rule->flow.icmp_type = htons(icmp_type); + +} + +void +cls_rule_set_icmp_code(struct cls_rule *rule, uint8_t icmp_code) +{ + rule->wc.wildcards &= ~FWW_TP_DST; + rule->flow.icmp_code = htons(icmp_code); +} + +void +cls_rule_set_arp_sha(struct cls_rule *rule, const uint8_t sha[ETH_ADDR_LEN]) +{ + rule->wc.wildcards &= ~FWW_ARP_SHA; + memcpy(rule->flow.arp_sha, sha, ETH_ADDR_LEN); +} + +void +cls_rule_set_arp_tha(struct cls_rule *rule, const uint8_t tha[ETH_ADDR_LEN]) +{ + rule->wc.wildcards &= ~FWW_ARP_THA; + memcpy(rule->flow.arp_tha, tha, ETH_ADDR_LEN); +} + +void +cls_rule_set_ipv6_src(struct cls_rule *rule, const struct in6_addr *src) +{ + cls_rule_set_ipv6_src_masked(rule, src, &in6addr_exact); +} + +bool +cls_rule_set_ipv6_src_masked(struct cls_rule *rule, const struct in6_addr *src, + const struct in6_addr *mask) +{ + if (flow_wildcards_set_ipv6_src_mask(&rule->wc, mask)) { + rule->flow.ipv6_src = ipv6_addr_bitand(src, mask); + return true; + } else { + return false; + } +} + +void +cls_rule_set_ipv6_dst(struct cls_rule *rule, const struct in6_addr *dst) +{ + cls_rule_set_ipv6_dst_masked(rule, dst, &in6addr_exact); +} + +bool +cls_rule_set_ipv6_dst_masked(struct cls_rule *rule, const struct in6_addr *dst, + const struct in6_addr *mask) +{ + if (flow_wildcards_set_ipv6_dst_mask(&rule->wc, mask)) { + rule->flow.ipv6_dst = ipv6_addr_bitand(dst, mask); + return true; + } else { + return false; + } +} + +/* Returns true if 'a' and 'b' have the same priority, wildcard the same + * fields, and have the same values for fixed fields, otherwise false. */ +bool +cls_rule_equal(const struct cls_rule *a, const struct cls_rule *b) +{ + return (a->priority == b->priority + && flow_wildcards_equal(&a->wc, &b->wc) + && flow_equal(&a->flow, &b->flow)); +} + +static void +format_ip_netmask(struct ds *s, const char *name, ovs_be32 ip, + ovs_be32 netmask) +{ + if (netmask) { + ds_put_format(s, "%s="IP_FMT, name, IP_ARGS(&ip)); + if (netmask != htonl(UINT32_MAX)) { + if (ip_is_cidr(netmask)) { + int wcbits = ofputil_netmask_to_wcbits(netmask); + ds_put_format(s, "/%d", 32 - wcbits); + } else { + ds_put_format(s, "/"IP_FMT, IP_ARGS(&netmask)); + } + } + ds_put_char(s, ','); + } +} + +static void +format_ipv6_netmask(struct ds *s, const char *name, + const struct in6_addr *addr, + const struct in6_addr *netmask) +{ + if (!ipv6_mask_is_any(netmask)) { + ds_put_format(s, "%s=", name); + print_ipv6_addr(s, addr); + if (!ipv6_mask_is_exact(netmask)) { + if (ipv6_is_cidr(netmask)) { + int cidr_bits = ipv6_count_cidr_bits(netmask); + ds_put_format(s, "/%d", cidr_bits); + } else { + ds_put_char(s, '/'); + print_ipv6_addr(s, netmask); + } + } + ds_put_char(s, ','); + } +} + +void +cls_rule_format(const struct cls_rule *rule, struct ds *s) +{ + const struct flow_wildcards *wc = &rule->wc; + size_t start_len = s->length; + flow_wildcards_t w = wc->wildcards; + const struct flow *f = &rule->flow; + bool skip_type = false; + bool skip_proto = false; + + int i; + + if (rule->priority != OFP_DEFAULT_PRIORITY) { + ds_put_format(s, "priority=%d,", rule->priority); + } + + if (!(w & FWW_DL_TYPE)) { + skip_type = true; + if (f->dl_type == htons(ETH_TYPE_IP)) { + if (!(w & FWW_NW_PROTO)) { + skip_proto = true; + if (f->nw_proto == IPPROTO_ICMP) { + ds_put_cstr(s, "icmp,"); + } else if (f->nw_proto == IPPROTO_TCP) { + ds_put_cstr(s, "tcp,"); + } else if (f->nw_proto == IPPROTO_UDP) { + ds_put_cstr(s, "udp,"); + } else { + ds_put_cstr(s, "ip,"); + skip_proto = false; + } + } else { + ds_put_cstr(s, "ip,"); + } + } else if (f->dl_type == htons(ETH_TYPE_IPV6)) { + if (!(w & FWW_NW_PROTO)) { + skip_proto = true; + if (f->nw_proto == IPPROTO_ICMPV6) { + ds_put_cstr(s, "icmp6,"); + } else if (f->nw_proto == IPPROTO_TCP) { + ds_put_cstr(s, "tcp6,"); + } else if (f->nw_proto == IPPROTO_UDP) { + ds_put_cstr(s, "udp6,"); + } else { + ds_put_cstr(s, "ipv6,"); + skip_proto = false; + } + } else { + ds_put_cstr(s, "ipv6,"); + } + } else if (f->dl_type == htons(ETH_TYPE_ARP)) { + ds_put_cstr(s, "arp,"); + } else { + skip_type = false; + } + } + for (i = 0; i < FLOW_N_REGS; i++) { + switch (wc->reg_masks[i]) { + case 0: + break; + case UINT32_MAX: + ds_put_format(s, "reg%d=0x%"PRIx32",", i, f->regs[i]); + break; + default: + ds_put_format(s, "reg%d=0x%"PRIx32"/0x%"PRIx32",", + i, f->regs[i], wc->reg_masks[i]); + break; + } + } + switch (wc->tun_id_mask) { + case 0: + break; + case CONSTANT_HTONLL(UINT64_MAX): + ds_put_format(s, "tun_id=%#"PRIx64",", ntohll(f->tun_id)); + break; + default: + ds_put_format(s, "tun_id=%#"PRIx64"/%#"PRIx64",", + ntohll(f->tun_id), ntohll(wc->tun_id_mask)); + break; + } + if (!(w & FWW_IN_PORT)) { + ds_put_format(s, "in_port=%"PRIu16",", + odp_port_to_ofp_port(f->in_port)); + } + if (wc->vlan_tci_mask) { + ovs_be16 vid_mask = wc->vlan_tci_mask & htons(VLAN_VID_MASK); + ovs_be16 pcp_mask = wc->vlan_tci_mask & htons(VLAN_PCP_MASK); + ovs_be16 cfi = wc->vlan_tci_mask & htons(VLAN_CFI); + + if (cfi && f->vlan_tci & htons(VLAN_CFI) + && (!vid_mask || vid_mask == htons(VLAN_VID_MASK)) + && (!pcp_mask || pcp_mask == htons(VLAN_PCP_MASK)) + && (vid_mask || pcp_mask)) { + if (vid_mask) { + ds_put_format(s, "dl_vlan=%"PRIu16",", + vlan_tci_to_vid(f->vlan_tci)); + } + if (pcp_mask) { + ds_put_format(s, "dl_vlan_pcp=%d,", + vlan_tci_to_pcp(f->vlan_tci)); + } + } else { + ds_put_format(s, "vlan_tci=0x%04"PRIx16"/0x%04"PRIx16",", + ntohs(f->vlan_tci), ntohs(wc->vlan_tci_mask)); + } + } + if (!(w & FWW_DL_SRC)) { + ds_put_format(s, "dl_src="ETH_ADDR_FMT",", ETH_ADDR_ARGS(f->dl_src)); + } + switch (w & (FWW_DL_DST | FWW_ETH_MCAST)) { + case 0: + ds_put_format(s, "dl_dst="ETH_ADDR_FMT",", ETH_ADDR_ARGS(f->dl_dst)); + break; + case FWW_DL_DST: + ds_put_format(s, "dl_dst="ETH_ADDR_FMT"/01:00:00:00:00:00,", + ETH_ADDR_ARGS(f->dl_dst)); + break; + case FWW_ETH_MCAST: + ds_put_format(s, "dl_dst="ETH_ADDR_FMT"/fe:ff:ff:ff:ff:ff,", + ETH_ADDR_ARGS(f->dl_dst)); + break; + case FWW_DL_DST | FWW_ETH_MCAST: + break; + } + if (!skip_type && !(w & FWW_DL_TYPE)) { + ds_put_format(s, "dl_type=0x%04"PRIx16",", ntohs(f->dl_type)); + } + if (f->dl_type == htons(ETH_TYPE_IPV6)) { + format_ipv6_netmask(s, "ipv6_src", &f->ipv6_src, &wc->ipv6_src_mask); + format_ipv6_netmask(s, "ipv6_dst", &f->ipv6_dst, &wc->ipv6_dst_mask); + } else { + format_ip_netmask(s, "nw_src", f->nw_src, wc->nw_src_mask); + format_ip_netmask(s, "nw_dst", f->nw_dst, wc->nw_dst_mask); + } + if (!skip_proto && !(w & FWW_NW_PROTO)) { + if (f->dl_type == htons(ETH_TYPE_ARP)) { + ds_put_format(s, "opcode=%"PRIu8",", f->nw_proto); + } else { + ds_put_format(s, "nw_proto=%"PRIu8",", f->nw_proto); + } + } + if (f->dl_type == htons(ETH_TYPE_ARP)) { + if (!(w & FWW_ARP_SHA)) { + ds_put_format(s, "arp_sha="ETH_ADDR_FMT",", + ETH_ADDR_ARGS(f->arp_sha)); + } + if (!(w & FWW_ARP_THA)) { + ds_put_format(s, "arp_tha="ETH_ADDR_FMT",", + ETH_ADDR_ARGS(f->arp_tha)); + } + } + if (!(w & FWW_NW_TOS)) { + ds_put_format(s, "nw_tos=%"PRIu8",", f->nw_tos); + } + if (f->nw_proto == IPPROTO_ICMP) { + if (!(w & FWW_TP_SRC)) { + ds_put_format(s, "icmp_type=%"PRIu16",", ntohs(f->tp_src)); + } + if (!(w & FWW_TP_DST)) { + ds_put_format(s, "icmp_code=%"PRIu16",", ntohs(f->tp_dst)); + } + } else if (f->nw_proto == IPPROTO_ICMPV6) { + if (!(w & FWW_TP_SRC)) { + ds_put_format(s, "icmp_type=%"PRIu16",", ntohs(f->tp_src)); + } + if (!(w & FWW_TP_DST)) { + ds_put_format(s, "icmp_code=%"PRIu16",", ntohs(f->tp_dst)); + } + } else { + if (!(w & FWW_TP_SRC)) { + ds_put_format(s, "tp_src=%"PRIu16",", ntohs(f->tp_src)); + } + if (!(w & FWW_TP_DST)) { + ds_put_format(s, "tp_dst=%"PRIu16",", ntohs(f->tp_dst)); + } + } + + if (s->length > start_len && ds_last(s) == ',') { + s->length--; + } } /* Converts 'rule' to a string and returns the string. The caller must free @@ -154,22 +606,16 @@ char * cls_rule_to_string(const struct cls_rule *rule) { struct ds s = DS_EMPTY_INITIALIZER; - ds_put_format(&s, "wildcards=%x priority=%u ", - rule->wc.wildcards, rule->priority); - flow_format(&s, &rule->flow); - return ds_cstr(&s); + cls_rule_format(rule, &s); + return ds_steal_cstr(&s); } -/* Prints cls_rule 'rule', for debugging. - * - * (The output could be improved and expanded, but this was good enough to - * debug the classifier.) */ void cls_rule_print(const struct cls_rule *rule) { - printf("wildcards=%x priority=%u ", rule->wc.wildcards, rule->priority); - flow_print(stdout, &rule->flow); - putc('\n', stdout); + char *s = cls_rule_to_string(rule); + puts(s); + free(s); } /* Initializes 'cls' as a classifier that initially contains no classification @@ -212,14 +658,6 @@ classifier_count(const struct classifier *cls) return cls->n_rules; } -/* Returns the number of rules in 'classifier' that have no wildcards. */ -int -classifier_count_exact(const struct classifier *cls) -{ - struct cls_table *exact_table = classifier_exact_table(cls); - return exact_table ? exact_table->n_table_rules : 0; -} - /* Inserts 'rule' into 'cls'. Until 'rule' is removed from 'cls', the caller * must not modify or free it. * @@ -273,7 +711,7 @@ classifier_remove(struct classifier *cls, struct cls_rule *rule) hmap_replace(&table->rules, &rule->hmap_node, &next->hmap_node); } - if (--table->n_table_rules == 0 && !table->n_refs) { + if (--table->n_table_rules == 0) { destroy_table(cls, table); } @@ -282,24 +720,18 @@ classifier_remove(struct classifier *cls, struct cls_rule *rule) /* Finds and returns the highest-priority rule in 'cls' that matches 'flow'. * Returns a null pointer if no rules in 'cls' match 'flow'. If multiple rules - * of equal priority match 'flow', returns one arbitrarily. - * - * 'include' is a combination of CLS_INC_* values that specify tables to - * include in the search. */ + * of equal priority match 'flow', returns one arbitrarily. */ struct cls_rule * -classifier_lookup(const struct classifier *cls, const struct flow *flow, - int include) +classifier_lookup(const struct classifier *cls, const struct flow *flow) { struct cls_table *table; struct cls_rule *best; best = NULL; HMAP_FOR_EACH (table, hmap_node, &cls->tables) { - if (should_include(table, include)) { - struct cls_rule *rule = find_match(table, flow); - if (rule && (!best || rule->priority > best->priority)) { - best = rule; - } + struct cls_rule *rule = find_match(table, flow); + if (rule && (!best || rule->priority > best->priority)) { + best = rule; } } return best; @@ -324,7 +756,7 @@ classifier_find_rule_exactly(const struct classifier *cls, } head = find_equal(table, &target->flow, flow_hash(&target->flow, 0)); - if (!target->wc.wildcards) { + if (flow_wildcards_is_exact(&target->wc)) { return head; } FOR_EACH_RULE_IN_LIST (rule, head) { @@ -363,10 +795,34 @@ classifier_rule_overlaps(const struct classifier *cls, return false; } + +/* Iteration. */ + +static bool +rule_matches(const struct cls_rule *rule, const struct cls_rule *target) +{ + return (!target + || flow_equal_except(&rule->flow, &target->flow, &target->wc)); +} + +static struct cls_rule * +search_table(const struct cls_table *table, const struct cls_rule *target) +{ + if (!target || !flow_wildcards_has_extra(&table->wc, &target->wc)) { + struct cls_rule *rule; -/* Searches 'cls' for rules that exactly match 'target' or are more specific - * than 'target'. That is, a given 'rule' matches 'target' if, for every - * field: + HMAP_FOR_EACH (rule, hmap_node, &table->rules) { + if (rule_matches(rule, target)) { + return rule; + } + } + } + return NULL; +} + +/* Initializes 'cursor' for iterating through 'cls' rules that exactly match + * 'target' or are more specific than 'target'. That is, a given 'rule' + * matches 'target' if, for every field: * * - 'target' and 'rule' specify the same (non-wildcarded) value for the * field, or @@ -397,80 +853,67 @@ classifier_rule_overlaps(const struct classifier *cls, * * Ignores target->priority. * - * 'callback' is allowed to delete the rule that is passed as its argument, but - * it must not delete (or move) any other rules in 'cls' that have the same - * wildcards as the argument rule. */ -void -classifier_for_each_match(const struct classifier *cls_, - const struct cls_rule *target, - int include, cls_cb_func *callback, void *aux) -{ - struct classifier *cls = (struct classifier *) cls_; - struct cls_table *table, *next_table; - - for (table = classifier_first_table(cls); table; table = next_table) { - if (should_include(table, include) - && !flow_wildcards_has_extra(&table->wc, &target->wc)) { - /* We have eliminated the "no" case in the truth table above. Two - * of the three remaining cases are trivial. We only need to check - * the fourth case, where both 'rule' and 'target' require an exact - * match. */ - struct cls_rule *head, *next_head; - - table->n_refs++; - HMAP_FOR_EACH_SAFE (head, next_head, hmap_node, &table->rules) { - if (flow_equal_except(&head->flow, &target->flow, - &target->wc)) { - struct cls_rule *rule, *next_rule; - - FOR_EACH_RULE_IN_LIST_SAFE (rule, next_rule, head) { - callback(rule, aux); - } - } - } - next_table = classifier_next_table(cls, table); - if (!--table->n_refs && !table->n_table_rules) { - destroy_table(cls, table); - } - } else { - next_table = classifier_next_table(cls, table); + * 'target' may be NULL to iterate over every rule in 'cls'. */ +void +cls_cursor_init(struct cls_cursor *cursor, const struct classifier *cls, + const struct cls_rule *target) +{ + cursor->cls = cls; + cursor->target = target; +} + +/* Returns the first matching cls_rule in 'cursor''s iteration, or a null + * pointer if there are no matches. */ +struct cls_rule * +cls_cursor_first(struct cls_cursor *cursor) +{ + struct cls_table *table; + + for (table = classifier_first_table(cursor->cls); table; + table = classifier_next_table(cursor->cls, table)) { + struct cls_rule *rule = search_table(table, cursor->target); + if (rule) { + cursor->table = table; + return rule; } } + + return NULL; } -/* 'callback' is allowed to delete the rule that is passed as its argument, but - * it must not delete (or move) any other rules in 'cls' that have the same - * wildcards as the argument rule. - * - * If 'include' is CLS_INC_EXACT then CLASSIFIER_FOR_EACH_EXACT_RULE is - * probably easier to use. */ -void -classifier_for_each(const struct classifier *cls_, int include, - cls_cb_func *callback, void *aux) +/* Returns the next matching cls_rule in 'cursor''s iteration, or a null + * pointer if there are no more matches. */ +struct cls_rule * +cls_cursor_next(struct cls_cursor *cursor, struct cls_rule *rule) { - struct classifier *cls = (struct classifier *) cls_; - struct cls_table *table, *next_table; + const struct cls_table *table; + struct cls_rule *next; - for (table = classifier_first_table(cls); table; table = next_table) { - if (should_include(table, include)) { - struct cls_rule *head, *next_head; + next = next_rule_in_list__(rule); + if (next->priority < rule->priority) { + return next; + } - table->n_refs++; - HMAP_FOR_EACH_SAFE (head, next_head, hmap_node, &table->rules) { - struct cls_rule *rule, *next_rule; + /* 'next' is the head of the list, that is, the rule that is included in + * the table's hmap. (This is important when the classifier contains rules + * that differ only in priority.) */ + rule = next; + HMAP_FOR_EACH_CONTINUE (rule, hmap_node, &cursor->table->rules) { + if (rule_matches(rule, cursor->target)) { + return rule; + } + } - FOR_EACH_RULE_IN_LIST_SAFE (rule, next_rule, head) { - callback(rule, aux); - } - } - next_table = classifier_next_table(cls, table); - if (!--table->n_refs && !table->n_table_rules) { - destroy_table(cls, table); - } - } else { - next_table = classifier_next_table(cls, table); + for (table = classifier_next_table(cursor->cls, cursor->table); table; + table = classifier_next_table(cursor->cls, table)) { + rule = search_table(table, cursor->target); + if (rule) { + cursor->table = table; + return rule; } } + + return NULL; } static struct cls_table * @@ -522,14 +965,6 @@ destroy_table(struct classifier *cls, struct cls_table *table) free(table); } -/* Returns true if 'table' should be included by an operation with the - * specified 'include' (a combination of CLS_INC_*). */ -static bool -should_include(const struct cls_table *table, int include) -{ - return include & (table->wc.wildcards ? CLS_INC_WILD : CLS_INC_EXACT); -} - static struct cls_rule * find_match(const struct cls_table *table, const struct flow *flow) { @@ -601,75 +1036,137 @@ insert_rule(struct cls_table *table, struct cls_rule *new) } static struct cls_rule * -next_rule_in_list(struct cls_rule *rule) +next_rule_in_list__(struct cls_rule *rule) { struct cls_rule *next = OBJECT_CONTAINING(rule->list.next, next, list); + return next; +} + +static struct cls_rule * +next_rule_in_list(struct cls_rule *rule) +{ + struct cls_rule *next = next_rule_in_list__(rule); return next->priority < rule->priority ? next : NULL; } +static bool +ipv6_equal_except(const struct in6_addr *a, const struct in6_addr *b, + const struct in6_addr *mask) +{ + int i; + +#ifdef s6_addr32 + for (i=0; i<4; i++) { + if ((a->s6_addr32[i] ^ b->s6_addr32[i]) & mask->s6_addr32[i]) { + return false; + } + } +#else + for (i=0; i<16; i++) { + if ((a->s6_addr[i] ^ b->s6_addr[i]) & mask->s6_addr[i]) { + return false; + } + } +#endif + + return true; +} + + static bool flow_equal_except(const struct flow *a, const struct flow *b, const struct flow_wildcards *wildcards) { - const uint32_t wc = wildcards->wildcards; + const flow_wildcards_t wc = wildcards->wildcards; + int i; - BUILD_ASSERT_DECL(FLOW_SIG_SIZE == 37); + BUILD_ASSERT_DECL(FLOW_SIG_SIZE == 84 + FLOW_N_REGS * 4); - return ((wc & NXFW_TUN_ID || a->tun_id == b->tun_id) + for (i = 0; i < FLOW_N_REGS; i++) { + if ((a->regs[i] ^ b->regs[i]) & wildcards->reg_masks[i]) { + return false; + } + } + + return (!((a->tun_id ^ b->tun_id) & wildcards->tun_id_mask) && !((a->nw_src ^ b->nw_src) & wildcards->nw_src_mask) && !((a->nw_dst ^ b->nw_dst) & wildcards->nw_dst_mask) - && (wc & OFPFW_IN_PORT || a->in_port == b->in_port) - && (wc & OFPFW_DL_VLAN || a->dl_vlan == b->dl_vlan) - && (wc & OFPFW_DL_TYPE || a->dl_type == b->dl_type) - && (wc & OFPFW_TP_SRC || a->tp_src == b->tp_src) - && (wc & OFPFW_TP_DST || a->tp_dst == b->tp_dst) - && (wc & OFPFW_DL_SRC || eth_addr_equals(a->dl_src, b->dl_src)) - && (wc & OFPFW_DL_DST || eth_addr_equals(a->dl_dst, b->dl_dst)) - && (wc & OFPFW_NW_PROTO || a->nw_proto == b->nw_proto) - && (wc & OFPFW_DL_VLAN_PCP || a->dl_vlan_pcp == b->dl_vlan_pcp) - && (wc & OFPFW_NW_TOS || a->nw_tos == b->nw_tos)); + && (wc & FWW_IN_PORT || a->in_port == b->in_port) + && !((a->vlan_tci ^ b->vlan_tci) & wildcards->vlan_tci_mask) + && (wc & FWW_DL_TYPE || a->dl_type == b->dl_type) + && (wc & FWW_TP_SRC || a->tp_src == b->tp_src) + && (wc & FWW_TP_DST || a->tp_dst == b->tp_dst) + && (wc & FWW_DL_SRC || eth_addr_equals(a->dl_src, b->dl_src)) + && (wc & FWW_DL_DST + || (!((a->dl_dst[0] ^ b->dl_dst[0]) & 0xfe) + && a->dl_dst[1] == b->dl_dst[1] + && a->dl_dst[2] == b->dl_dst[2] + && a->dl_dst[3] == b->dl_dst[3] + && a->dl_dst[4] == b->dl_dst[4] + && a->dl_dst[5] == b->dl_dst[5])) + && (wc & FWW_ETH_MCAST + || !((a->dl_dst[0] ^ b->dl_dst[0]) & 0x01)) + && (wc & FWW_NW_PROTO || a->nw_proto == b->nw_proto) + && (wc & FWW_NW_TOS || a->nw_tos == b->nw_tos) + && (wc & FWW_ARP_SHA || eth_addr_equals(a->arp_sha, b->arp_sha)) + && (wc & FWW_ARP_THA || eth_addr_equals(a->arp_tha, b->arp_tha)) + && ipv6_equal_except(&a->ipv6_src, &b->ipv6_src, + &wildcards->ipv6_src_mask) + && ipv6_equal_except(&a->ipv6_dst, &b->ipv6_dst, + &wildcards->ipv6_dst_mask)); } static void zero_wildcards(struct flow *flow, const struct flow_wildcards *wildcards) { - const uint32_t wc = wildcards->wildcards; + const flow_wildcards_t wc = wildcards->wildcards; + int i; - BUILD_ASSERT_DECL(FLOW_SIG_SIZE == 37); + BUILD_ASSERT_DECL(FLOW_SIG_SIZE == 84 + 4 * FLOW_N_REGS); - if (wc & NXFW_TUN_ID) { - flow->tun_id = 0; + for (i = 0; i < FLOW_N_REGS; i++) { + flow->regs[i] &= wildcards->reg_masks[i]; } + flow->tun_id &= wildcards->tun_id_mask; flow->nw_src &= wildcards->nw_src_mask; flow->nw_dst &= wildcards->nw_dst_mask; - if (wc & OFPFW_IN_PORT) { + if (wc & FWW_IN_PORT) { flow->in_port = 0; } - if (wc & OFPFW_DL_VLAN) { - flow->dl_vlan = 0; - } - if (wc & OFPFW_DL_TYPE) { + flow->vlan_tci &= wildcards->vlan_tci_mask; + if (wc & FWW_DL_TYPE) { flow->dl_type = 0; } - if (wc & OFPFW_TP_SRC) { + if (wc & FWW_TP_SRC) { flow->tp_src = 0; } - if (wc & OFPFW_TP_DST) { + if (wc & FWW_TP_DST) { flow->tp_dst = 0; } - if (wc & OFPFW_DL_SRC) { + if (wc & FWW_DL_SRC) { memset(flow->dl_src, 0, sizeof flow->dl_src); } - if (wc & OFPFW_DL_DST) { - memset(flow->dl_dst, 0, sizeof flow->dl_dst); + if (wc & FWW_DL_DST) { + flow->dl_dst[0] &= 0x01; + memset(&flow->dl_dst[1], 0, 5); } - if (wc & OFPFW_NW_PROTO) { - flow->nw_proto = 0; + if (wc & FWW_ETH_MCAST) { + flow->dl_dst[0] &= 0xfe; } - if (wc & OFPFW_DL_VLAN_PCP) { - flow->dl_vlan_pcp = 0; + if (wc & FWW_NW_PROTO) { + flow->nw_proto = 0; } - if (wc & OFPFW_NW_TOS) { + if (wc & FWW_NW_TOS) { flow->nw_tos = 0; } + if (wc & FWW_ARP_SHA) { + memset(flow->arp_sha, 0, sizeof flow->arp_sha); + } + if (wc & FWW_ARP_THA) { + memset(flow->arp_tha, 0, sizeof flow->arp_tha); + } + flow->ipv6_src = ipv6_addr_bitand(&flow->ipv6_src, + &wildcards->ipv6_src_mask); + flow->ipv6_dst = ipv6_addr_bitand(&flow->ipv6_dst, + &wildcards->ipv6_dst_mask); }