dpif_linux_enumerate(struct sset *all_dps)
{
struct nl_dump dump;
- struct ofpbuf msg;
+ uint64_t reply_stub[NL_DUMP_BUFSIZE / 8];
+ struct ofpbuf msg, buf;
int error;
error = dpif_linux_init();
return error;
}
+ ofpbuf_use_stub(&buf, reply_stub, sizeof reply_stub);
dpif_linux_dp_dump_start(&dump);
- while (nl_dump_next(&dump, &msg)) {
+ while (nl_dump_next(&dump, &msg, &buf)) {
struct dpif_linux_dp dp;
if (!dpif_linux_dp_from_ofpbuf(&dp, &msg)) {
sset_add(all_dps, dp.name);
}
}
+ ofpbuf_uninit(&buf);
return nl_dump_done(&dump);
}
struct dpif_linux_port_state {
struct nl_dump dump;
+ struct ofpbuf buf;
};
static void
*statep = state = xmalloc(sizeof *state);
dpif_linux_port_dump_start__(dpif, &state->dump);
+ ofpbuf_init(&state->buf, NL_DUMP_BUFSIZE);
return 0;
}
static int
dpif_linux_port_dump_next__(const struct dpif *dpif_, struct nl_dump *dump,
- struct dpif_linux_vport *vport)
+ struct dpif_linux_vport *vport,
+ struct ofpbuf *buffer)
{
struct dpif_linux *dpif = dpif_linux_cast(dpif_);
struct ofpbuf buf;
int error;
- if (!nl_dump_next(dump, &buf)) {
+ if (!nl_dump_next(dump, &buf, buffer)) {
return EOF;
}
struct dpif_linux_vport vport;
int error;
- error = dpif_linux_port_dump_next__(dpif, &state->dump, &vport);
+ error = dpif_linux_port_dump_next__(dpif, &state->dump, &vport,
+ &state->buf);
if (error) {
return error;
}
struct dpif_linux_port_state *state = state_;
int error = nl_dump_done(&state->dump);
+ ofpbuf_uninit(&state->buf);
free(state);
return error;
}
struct nl_dump dump;
struct dpif_linux_flow flow;
struct dpif_flow_stats stats;
- struct ofpbuf *buf;
+ struct ofpbuf buffer; /* Always used to store flows. */
+ struct ofpbuf *tmp; /* Used if kernel does not supply actions. */
};
static int
nl_dump_start(&state->dump, NETLINK_GENERIC, buf);
ofpbuf_delete(buf);
- state->buf = NULL;
+ ofpbuf_init(&state->buffer, NL_DUMP_BUFSIZE);
+ state->tmp = NULL;
return 0;
}
int error;
do {
- ofpbuf_delete(state->buf);
- state->buf = NULL;
+ ofpbuf_delete(state->tmp);
+ state->tmp = NULL;
- if (!nl_dump_next(&state->dump, &buf)) {
+ if (!nl_dump_next(&state->dump, &buf, &state->buffer)) {
return EOF;
}
if (actions && !state->flow.actions) {
error = dpif_linux_flow_get__(dpif_, state->flow.key,
state->flow.key_len,
- &state->flow, &state->buf);
+ &state->flow, &state->tmp);
if (error == ENOENT) {
VLOG_DBG("dumped flow disappeared on get");
} else if (error) {
{
struct dpif_linux_flow_state *state = state_;
int error = nl_dump_done(&state->dump);
- ofpbuf_delete(state->buf);
+ ofpbuf_uninit(&state->buffer);
+ ofpbuf_delete(state->tmp);
free(state);
return error;
}
struct dpif_linux_vport vport;
size_t keep_channels_nbits;
struct nl_dump dump;
+ uint64_t reply_stub[NL_DUMP_BUFSIZE / 8];
+ struct ofpbuf buf;
int retval = 0;
size_t i;
dpif->n_events = dpif->event_offset = 0;
+ ofpbuf_use_stub(&buf, reply_stub, sizeof reply_stub);
dpif_linux_port_dump_start__(dpif_, &dump);
- while (!dpif_linux_port_dump_next__(dpif_, &dump, &vport)) {
+ while (!dpif_linux_port_dump_next__(dpif_, &dump, &vport, &buf)) {
uint32_t port_no = odp_to_u32(vport.port_no);
struct nl_sock *sock = (port_no < dpif->uc_array_size
? dpif->channels[port_no].sock
nl_sock_destroy(sock);
}
nl_dump_done(&dump);
+ ofpbuf_uninit(&buf);
/* Discard any saved channels that we didn't reuse. */
for (i = 0; i < keep_channels_nbits; i++) {
return error;
}
+struct queue_dump_state {
+ struct nl_dump dump;
+ struct ofpbuf buf;
+};
+
static bool
-start_queue_dump(const struct netdev *netdev, struct nl_dump *dump)
+start_queue_dump(const struct netdev *netdev, struct queue_dump_state *state)
{
struct ofpbuf request;
struct tcmsg *tcmsg;
return false;
}
tcmsg->tcm_parent = 0;
- nl_dump_start(dump, NETLINK_ROUTE, &request);
+ nl_dump_start(&state->dump, NETLINK_ROUTE, &request);
ofpbuf_uninit(&request);
+
+ ofpbuf_init(&state->buf, NL_DUMP_BUFSIZE);
return true;
}
+static int
+finish_queue_dump(struct queue_dump_state *state)
+{
+ ofpbuf_uninit(&state->buf);
+ return nl_dump_done(&state->dump);
+}
+
struct netdev_linux_queue_state {
unsigned int *queues;
size_t cur_queue;
ovs_mutex_lock(&netdev->mutex);
error = tc_query_qdisc(netdev_);
if (!error) {
- struct nl_dump dump;
+ struct queue_dump_state state;
if (!netdev->tc->ops->class_dump_stats) {
error = EOPNOTSUPP;
- } else if (!start_queue_dump(netdev_, &dump)) {
+ } else if (!start_queue_dump(netdev_, &state)) {
error = ENODEV;
} else {
struct ofpbuf msg;
int retval;
- while (nl_dump_next(&dump, &msg)) {
+ while (nl_dump_next(&state.dump, &msg, &state.buf)) {
retval = netdev->tc->ops->class_dump_stats(netdev_, &msg,
cb, aux);
if (retval) {
}
}
- retval = nl_dump_done(&dump);
+ retval = finish_queue_dump(&state);
if (retval) {
error = retval;
}
htb_tc_load(struct netdev *netdev, struct ofpbuf *nlmsg OVS_UNUSED)
{
struct ofpbuf msg;
- struct nl_dump dump;
+ struct queue_dump_state state;
struct htb_class hc;
/* Get qdisc options. */
htb_install__(netdev, hc.max_rate);
/* Get queues. */
- if (!start_queue_dump(netdev, &dump)) {
+ if (!start_queue_dump(netdev, &state)) {
return ENODEV;
}
- while (nl_dump_next(&dump, &msg)) {
+ while (nl_dump_next(&state.dump, &msg, &state.buf)) {
unsigned int queue_id;
if (!htb_parse_tcmsg__(&msg, &queue_id, &hc, NULL)) {
htb_update_queue__(netdev, queue_id, &hc);
}
}
- nl_dump_done(&dump);
+ finish_queue_dump(&state);
return 0;
}
hfsc_tc_load(struct netdev *netdev, struct ofpbuf *nlmsg OVS_UNUSED)
{
struct ofpbuf msg;
- struct nl_dump dump;
+ struct queue_dump_state state;
struct hfsc_class hc;
hc.max_rate = 0;
hfsc_query_class__(netdev, tc_make_handle(1, 0xfffe), 0, &hc, NULL);
hfsc_install__(netdev, hc.max_rate);
- if (!start_queue_dump(netdev, &dump)) {
+ if (!start_queue_dump(netdev, &state)) {
return ENODEV;
}
- while (nl_dump_next(&dump, &msg)) {
+ while (nl_dump_next(&state.dump, &msg, &state.buf)) {
unsigned int queue_id;
if (!hfsc_parse_tcmsg__(&msg, &queue_id, &hc, NULL)) {
}
}
- nl_dump_done(&dump);
+ finish_queue_dump(&state);
return 0;
}
void
nl_dump_start(struct nl_dump *dump, int protocol, const struct ofpbuf *request)
{
- ofpbuf_init(&dump->buffer, 4096);
dump->status = nl_pool_alloc(protocol, &dump->sock);
if (dump->status) {
return;
/* Helper function for nl_dump_next(). */
static int
-nl_dump_recv(struct nl_dump *dump)
+nl_dump_recv(struct nl_dump *dump, struct ofpbuf *buffer)
{
struct nlmsghdr *nlmsghdr;
int retval;
- retval = nl_sock_recv__(dump->sock, &dump->buffer, true);
+ retval = nl_sock_recv__(dump->sock, buffer, true);
if (retval) {
return retval == EINTR ? EAGAIN : retval;
}
- nlmsghdr = nl_msg_nlmsghdr(&dump->buffer);
+ nlmsghdr = nl_msg_nlmsghdr(buffer);
if (dump->nl_seq != nlmsghdr->nlmsg_seq) {
VLOG_DBG_RL(&rl, "ignoring seq %#"PRIx32" != expected %#"PRIx32,
nlmsghdr->nlmsg_seq, dump->nl_seq);
return EAGAIN;
}
- if (nl_msg_nlmsgerr(&dump->buffer, &retval)) {
+ if (nl_msg_nlmsgerr(buffer, &retval)) {
VLOG_INFO_RL(&rl, "netlink dump request error (%s)",
ovs_strerror(retval));
return retval && retval != EAGAIN ? retval : EPROTO;
return 0;
}
-/* Attempts to retrieve another reply from 'dump', which must have been
- * initialized with nl_dump_start().
+/* Attempts to retrieve another reply from 'dump' into 'buffer'. 'dump' must
+ * have been initialized with nl_dump_start(), and 'buffer' must have been
+ * initialized. 'buffer' should be at least NL_DUMP_BUFSIZE bytes long.
*
* If successful, returns true and points 'reply->data' and 'reply->size' to
- * the message that was retrieved. The caller must not modify 'reply' (because
- * it points into the middle of a larger buffer).
+ * the message that was retrieved. The caller must not modify 'reply' (because
+ * it points within 'buffer', which will be used by future calls to this
+ * function).
*
* On failure, returns false and sets 'reply->data' to NULL and 'reply->size'
* to 0. Failure might indicate an actual error or merely the end of replies.
* completed by calling nl_dump_done().
*/
bool
-nl_dump_next(struct nl_dump *dump, struct ofpbuf *reply)
+nl_dump_next(struct nl_dump *dump, struct ofpbuf *reply, struct ofpbuf *buffer)
{
struct nlmsghdr *nlmsghdr;
return false;
}
- while (!dump->buffer.size) {
- int retval = nl_dump_recv(dump);
+ while (!buffer->size) {
+ int retval = nl_dump_recv(dump, buffer);
if (retval) {
- ofpbuf_clear(&dump->buffer);
+ ofpbuf_clear(buffer);
if (retval != EAGAIN) {
dump->status = retval;
return false;
}
}
- nlmsghdr = nl_msg_next(&dump->buffer, reply);
+ nlmsghdr = nl_msg_next(buffer, reply);
if (!nlmsghdr) {
VLOG_WARN_RL(&rl, "netlink dump reply contains message fragment");
dump->status = EPROTO;
int
nl_dump_done(struct nl_dump *dump)
{
+ uint64_t tmp_reply_stub[NL_DUMP_BUFSIZE / 8];
+ struct ofpbuf buf;
+
/* Drain any remaining messages that the client didn't read. Otherwise the
* kernel will continue to queue them up and waste buffer space.
*
* XXX We could just destroy and discard the socket in this case. */
+ ofpbuf_use_stub(&buf, tmp_reply_stub, sizeof tmp_reply_stub);
while (!dump->status) {
struct ofpbuf reply;
- if (!nl_dump_next(dump, &reply)) {
+ if (!nl_dump_next(dump, &reply, &buf)) {
ovs_assert(dump->status);
}
}
nl_pool_release(dump->sock);
- ofpbuf_uninit(&dump->buffer);
+ ofpbuf_uninit(&buf);
return dump->status == EOF ? 0 : dump->status;
}
void nl_transact_multiple(int protocol, struct nl_transaction **, size_t n);
/* Table dumping. */
+#define NL_DUMP_BUFSIZE 4096
+
struct nl_dump {
struct nl_sock *sock; /* Socket being dumped. */
uint32_t nl_seq; /* Expected nlmsg_seq for replies. */
- struct ofpbuf buffer; /* Receive buffer currently being iterated. */
int status; /* 0=OK, EOF=done, or positive errno value. */
};
void nl_dump_start(struct nl_dump *, int protocol,
const struct ofpbuf *request);
-bool nl_dump_next(struct nl_dump *, struct ofpbuf *reply);
+bool nl_dump_next(struct nl_dump *, struct ofpbuf *reply, struct ofpbuf *buf);
int nl_dump_done(struct nl_dump *);
/* Miscellaneous */
{
struct nl_dump dump;
struct rtgenmsg *rtmsg;
- struct ofpbuf request, reply;
+ uint64_t reply_stub[NL_DUMP_BUFSIZE / 8];
+ struct ofpbuf request, reply, buf;
route_map_clear();
route_table_valid = true;
nl_dump_start(&dump, NETLINK_ROUTE, &request);
ofpbuf_uninit(&request);
- while (nl_dump_next(&dump, &reply)) {
+ ofpbuf_use_stub(&buf, reply_stub, sizeof reply_stub);
+ while (nl_dump_next(&dump, &reply, &buf)) {
struct route_table_msg msg;
if (route_table_parse(&reply, &msg)) {
route_table_handle_msg(&msg);
}
}
+ ofpbuf_uninit(&buf);
return nl_dump_done(&dump);
}
{
struct nl_dump dump;
struct rtgenmsg *rtmsg;
- struct ofpbuf request, reply;
+ uint64_t reply_stub[NL_DUMP_BUFSIZE / 8];
+ struct ofpbuf request, reply, buf;
name_table_valid = true;
name_map_clear();
nl_dump_start(&dump, NETLINK_ROUTE, &request);
ofpbuf_uninit(&request);
- while (nl_dump_next(&dump, &reply)) {
+ ofpbuf_use_stub(&buf, reply_stub, sizeof reply_stub);
+ while (nl_dump_next(&dump, &reply, &buf)) {
struct rtnetlink_link_change change;
if (rtnetlink_link_parse(&reply, &change)
hmap_insert(&name_map, &nn->node, hash_int(nn->ifi_index, 0));
}
}
+ ofpbuf_uninit(&buf);
return nl_dump_done(&dump);
}