/*
- * Copyright (c) 2008, 2009, 2010, 2011, 2012, 2013 Nicira, Inc.
+ * Copyright (c) 2008, 2009, 2010, 2011, 2012, 2013, 2014 Nicira, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
*
* See the large comment in rconn.h for more information. */
struct rconn {
+ struct ovs_mutex mutex;
+
enum state state;
time_t state_entered;
int probe_interval; /* Secs of inactivity before sending probe. */
time_t last_activity; /* Last time we saw some activity. */
- /* When we create a vconn we obtain these values, to save them past the end
- * of the vconn's lifetime. Otherwise, in-band control will only allow
- * traffic when a vconn is actually open, but it is nice to allow ARP to
- * complete even between connection attempts, and it is also polite to
- * allow traffic from other switches to go through to the controller
- * whether or not we are connected.
- *
- * We don't cache the local port, because that changes from one connection
- * attempt to the next. */
- ovs_be32 local_ip, remote_ip;
- ovs_be16 remote_port;
uint8_t dscp;
/* Messages sent or received are copied to the monitor connections. */
return rconn->allowed_versions;
}
-static unsigned int elapsed_in_this_state(const struct rconn *);
-static unsigned int timeout(const struct rconn *);
-static bool timed_out(const struct rconn *);
-static void state_transition(struct rconn *, enum state);
-static void rconn_set_target__(struct rconn *,
- const char *target, const char *name);
-static int try_send(struct rconn *);
-static void reconnect(struct rconn *);
-static void report_error(struct rconn *, int error);
-static void disconnect(struct rconn *, int error);
-static void flush_queue(struct rconn *);
-static void close_monitor(struct rconn *, size_t idx, int retval);
+static unsigned int elapsed_in_this_state(const struct rconn *rc)
+ OVS_REQUIRES(rc->mutex);
+static unsigned int timeout(const struct rconn *rc) OVS_REQUIRES(rc->mutex);
+static bool timed_out(const struct rconn *rc) OVS_REQUIRES(rc->mutex);
+static void state_transition(struct rconn *rc, enum state)
+ OVS_REQUIRES(rc->mutex);
+static void rconn_set_target__(struct rconn *rc,
+ const char *target, const char *name)
+ OVS_REQUIRES(rc->mutex);
+static int rconn_send__(struct rconn *rc, struct ofpbuf *,
+ struct rconn_packet_counter *)
+ OVS_REQUIRES(rc->mutex);
+static int try_send(struct rconn *rc) OVS_REQUIRES(rc->mutex);
+static void reconnect(struct rconn *rc) OVS_REQUIRES(rc->mutex);
+static void report_error(struct rconn *rc, int error) OVS_REQUIRES(rc->mutex);
+static void rconn_disconnect__(struct rconn *rc) OVS_REQUIRES(rc->mutex);
+static void disconnect(struct rconn *rc, int error) OVS_REQUIRES(rc->mutex);
+static void flush_queue(struct rconn *rc) OVS_REQUIRES(rc->mutex);
+static void close_monitor(struct rconn *rc, size_t idx, int retval)
+ OVS_REQUIRES(rc->mutex);
static void copy_to_monitor(struct rconn *, const struct ofpbuf *);
static bool is_connected_state(enum state);
static bool is_admitted_msg(const struct ofpbuf *);
-static bool rconn_logging_connection_attempts__(const struct rconn *);
+static bool rconn_logging_connection_attempts__(const struct rconn *rc)
+ OVS_REQUIRES(rc->mutex);
+static int rconn_get_version__(const struct rconn *rconn)
+ OVS_REQUIRES(rconn->mutex);
+
+/* The following prototypes duplicate those in rconn.h, but there we weren't
+ * able to add the OVS_EXCLUDED annotations because the definition of struct
+ * rconn was not visible. */
+
+void rconn_set_max_backoff(struct rconn *rc, int max_backoff)
+ OVS_EXCLUDED(rc->mutex);
+void rconn_connect(struct rconn *rc, const char *target, const char *name)
+ OVS_EXCLUDED(rc->mutex);
+void rconn_connect_unreliably(struct rconn *rc,
+ struct vconn *vconn, const char *name)
+ OVS_EXCLUDED(rc->mutex);
+void rconn_reconnect(struct rconn *rc) OVS_EXCLUDED(rc->mutex);
+void rconn_disconnect(struct rconn *rc) OVS_EXCLUDED(rc->mutex);
+void rconn_run(struct rconn *rc) OVS_EXCLUDED(rc->mutex);
+void rconn_run_wait(struct rconn *rc) OVS_EXCLUDED(rc->mutex);
+struct ofpbuf *rconn_recv(struct rconn *rc) OVS_EXCLUDED(rc->mutex);
+void rconn_recv_wait(struct rconn *rc) OVS_EXCLUDED(rc->mutex);
+int rconn_send(struct rconn *rc, struct ofpbuf *b,
+ struct rconn_packet_counter *counter)
+ OVS_EXCLUDED(rc->mutex);
+int rconn_send_with_limit(struct rconn *rc, struct ofpbuf *b,
+ struct rconn_packet_counter *counter,
+ int queue_limit)
+ OVS_EXCLUDED(rc->mutex);
+void rconn_add_monitor(struct rconn *rc, struct vconn *vconn)
+ OVS_EXCLUDED(rc->mutex);
+void rconn_set_name(struct rconn *rc, const char *new_name)
+ OVS_EXCLUDED(rc->mutex);
+bool rconn_is_admitted(const struct rconn *rconn) OVS_EXCLUDED(rconn->mutex);
+int rconn_failure_duration(const struct rconn *rconn)
+ OVS_EXCLUDED(rconn->mutex);
+ovs_be16 rconn_get_local_port(const struct rconn *rconn)
+ OVS_EXCLUDED(rconn->mutex);
+int rconn_get_version(const struct rconn *rconn) OVS_EXCLUDED(rconn->mutex);
+unsigned int rconn_count_txqlen(const struct rconn *rc)
+ OVS_EXCLUDED(rc->mutex);
+
/* Creates and returns a new rconn.
*
{
struct rconn *rc = xzalloc(sizeof *rc);
+ ovs_mutex_init(&rc->mutex);
+
rc->state = S_VOID;
rc->state_entered = time_now();
void
rconn_set_max_backoff(struct rconn *rc, int max_backoff)
+ OVS_EXCLUDED(rc->mutex)
{
+ ovs_mutex_lock(&rc->mutex);
rc->max_backoff = MAX(1, max_backoff);
if (rc->state == S_BACKOFF && rc->backoff > max_backoff) {
rc->backoff = max_backoff;
rc->backoff_deadline = time_now() + max_backoff;
}
}
+ ovs_mutex_unlock(&rc->mutex);
}
int
* but it need not be acceptable to vconn_open(). */
void
rconn_connect(struct rconn *rc, const char *target, const char *name)
+ OVS_EXCLUDED(rc->mutex)
{
- rconn_disconnect(rc);
+ ovs_mutex_lock(&rc->mutex);
+ rconn_disconnect__(rc);
rconn_set_target__(rc, target, name);
rc->reliable = true;
reconnect(rc);
+ ovs_mutex_unlock(&rc->mutex);
}
/* Drops any existing connection on 'rc', then configures 'rc' to use
void
rconn_connect_unreliably(struct rconn *rc,
struct vconn *vconn, const char *name)
+ OVS_EXCLUDED(rc->mutex)
{
ovs_assert(vconn != NULL);
- rconn_disconnect(rc);
+
+ ovs_mutex_lock(&rc->mutex);
+ rconn_disconnect__(rc);
rconn_set_target__(rc, vconn_get_name(vconn), name);
rc->reliable = false;
rc->vconn = vconn;
rc->last_connected = time_now();
state_transition(rc, S_ACTIVE);
+ ovs_mutex_unlock(&rc->mutex);
}
/* If 'rc' is connected, forces it to drop the connection and reconnect. */
void
rconn_reconnect(struct rconn *rc)
+ OVS_EXCLUDED(rc->mutex)
{
+ ovs_mutex_lock(&rc->mutex);
if (rc->state & (S_ACTIVE | S_IDLE)) {
VLOG_INFO("%s: disconnecting", rc->name);
disconnect(rc, 0);
}
+ ovs_mutex_unlock(&rc->mutex);
}
-void
-rconn_disconnect(struct rconn *rc)
+static void
+rconn_disconnect__(struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
if (rc->state != S_VOID) {
if (rc->vconn) {
}
}
+void
+rconn_disconnect(struct rconn *rc)
+ OVS_EXCLUDED(rc->mutex)
+{
+ ovs_mutex_lock(&rc->mutex);
+ rconn_disconnect__(rc);
+ ovs_mutex_unlock(&rc->mutex);
+}
+
/* Disconnects 'rc' and frees the underlying storage. */
void
rconn_destroy(struct rconn *rc)
if (rc) {
size_t i;
+ ovs_mutex_lock(&rc->mutex);
free(rc->name);
free(rc->target);
vconn_close(rc->vconn);
for (i = 0; i < rc->n_monitors; i++) {
vconn_close(rc->monitors[i]);
}
+ ovs_mutex_unlock(&rc->mutex);
+ ovs_mutex_destroy(&rc->mutex);
+
free(rc);
}
}
static unsigned int
timeout_VOID(const struct rconn *rc OVS_UNUSED)
+ OVS_REQUIRES(rc->mutex)
{
return UINT_MAX;
}
static void
run_VOID(struct rconn *rc OVS_UNUSED)
+ OVS_REQUIRES(rc->mutex)
{
/* Nothing to do. */
}
static void
reconnect(struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
int retval;
retval = vconn_open(rc->target, rc->allowed_versions, rc->dscp,
&rc->vconn);
if (!retval) {
- rc->remote_ip = vconn_get_remote_ip(rc->vconn);
- rc->local_ip = vconn_get_local_ip(rc->vconn);
- rc->remote_port = vconn_get_remote_port(rc->vconn);
rc->backoff_deadline = time_now() + rc->backoff;
state_transition(rc, S_CONNECTING);
} else {
static unsigned int
timeout_BACKOFF(const struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
return rc->backoff;
}
static void
run_BACKOFF(struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
if (timed_out(rc)) {
reconnect(rc);
static unsigned int
timeout_CONNECTING(const struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
return MAX(1, rc->backoff);
}
static void
run_CONNECTING(struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
int retval = vconn_connect(rc->vconn);
if (!retval) {
static void
do_tx_work(struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
if (list_is_empty(&rc->txq)) {
return;
static unsigned int
timeout_ACTIVE(const struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
if (rc->probe_interval) {
unsigned int base = MAX(rc->last_activity, rc->state_entered);
static void
run_ACTIVE(struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
if (timed_out(rc)) {
unsigned int base = MAX(rc->last_activity, rc->state_entered);
VLOG_DBG("%s: idle %u seconds, sending inactivity probe",
rc->name, (unsigned int) (time_now() - base));
- version = rconn_get_version(rc);
+ version = rconn_get_version__(rc);
ovs_assert(version >= 0 && version <= 0xff);
/* Ordering is important here: rconn_send() can transition to BACKOFF,
* and we don't want to transition back to IDLE if so, because then we
* can end up queuing a packet with vconn == NULL and then *boom*. */
state_transition(rc, S_IDLE);
- rconn_send(rc, make_echo_request(version), NULL);
+ rconn_send__(rc, make_echo_request(version), NULL);
return;
}
static unsigned int
timeout_IDLE(const struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
return rc->probe_interval;
}
static void
run_IDLE(struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
if (timed_out(rc)) {
VLOG_ERR("%s: no response to inactivity probe after %u "
* connected, attempts to send packets in the send queue, if any. */
void
rconn_run(struct rconn *rc)
+ OVS_EXCLUDED(rc->mutex)
{
int old_state;
size_t i;
+ ovs_mutex_lock(&rc->mutex);
if (rc->vconn) {
+ int error;
+
vconn_run(rc->vconn);
+
+ error = vconn_get_status(rc->vconn);
+ if (error) {
+ report_error(rc, error);
+ disconnect(rc, error);
+ }
}
for (i = 0; i < rc->n_monitors; ) {
struct ofpbuf *msg;
STATES
#undef STATE
default:
- NOT_REACHED();
+ OVS_NOT_REACHED();
}
} while (rc->state != old_state);
+ ovs_mutex_unlock(&rc->mutex);
}
/* Causes the next call to poll_block() to wake up when rconn_run() should be
* called on 'rc'. */
void
rconn_run_wait(struct rconn *rc)
+ OVS_EXCLUDED(rc->mutex)
{
unsigned int timeo;
size_t i;
+ ovs_mutex_lock(&rc->mutex);
if (rc->vconn) {
vconn_run_wait(rc->vconn);
if ((rc->state & (S_ACTIVE | S_IDLE)) && !list_is_empty(&rc->txq)) {
long long int expires = sat_add(rc->state_entered, timeo);
poll_timer_wait_until(expires * 1000);
}
+ ovs_mutex_unlock(&rc->mutex);
}
/* Attempts to receive a packet from 'rc'. If successful, returns the packet;
* the packet (with ofpbuf_delete()). */
struct ofpbuf *
rconn_recv(struct rconn *rc)
+ OVS_EXCLUDED(rc->mutex)
{
+ struct ofpbuf *buffer = NULL;
+
+ ovs_mutex_lock(&rc->mutex);
if (rc->state & (S_ACTIVE | S_IDLE)) {
- struct ofpbuf *buffer;
int error = vconn_recv(rc->vconn, &buffer);
if (!error) {
copy_to_monitor(rc, buffer);
if (rc->state == S_IDLE) {
state_transition(rc, S_ACTIVE);
}
- return buffer;
} else if (error != EAGAIN) {
report_error(rc, error);
disconnect(rc, error);
}
}
- return NULL;
+ ovs_mutex_unlock(&rc->mutex);
+
+ return buffer;
}
/* Causes the next call to poll_block() to wake up when a packet may be ready
* to be received by vconn_recv() on 'rc'. */
void
rconn_recv_wait(struct rconn *rc)
+ OVS_EXCLUDED(rc->mutex)
{
+ ovs_mutex_lock(&rc->mutex);
if (rc->vconn) {
vconn_wait(rc->vconn, WAIT_RECV);
}
+ ovs_mutex_unlock(&rc->mutex);
}
-/* Sends 'b' on 'rc'. Returns 0 if successful, or ENOTCONN if 'rc' is not
- * currently connected. Takes ownership of 'b'.
- *
- * If 'counter' is non-null, then 'counter' will be incremented while the
- * packet is in flight, then decremented when it has been sent (or discarded
- * due to disconnection). Because 'b' may be sent (or discarded) before this
- * function returns, the caller may not be able to observe any change in
- * 'counter'.
- *
- * There is no rconn_send_wait() function: an rconn has a send queue that it
- * takes care of sending if you call rconn_run(), which will have the side
- * effect of waking up poll_block(). */
-int
-rconn_send(struct rconn *rc, struct ofpbuf *b,
+static int
+rconn_send__(struct rconn *rc, struct ofpbuf *b,
struct rconn_packet_counter *counter)
+ OVS_REQUIRES(rc->mutex)
{
if (rconn_is_connected(rc)) {
COVERAGE_INC(rconn_queued);
}
}
+/* Sends 'b' on 'rc'. Returns 0 if successful, or ENOTCONN if 'rc' is not
+ * currently connected. Takes ownership of 'b'.
+ *
+ * If 'counter' is non-null, then 'counter' will be incremented while the
+ * packet is in flight, then decremented when it has been sent (or discarded
+ * due to disconnection). Because 'b' may be sent (or discarded) before this
+ * function returns, the caller may not be able to observe any change in
+ * 'counter'.
+ *
+ * There is no rconn_send_wait() function: an rconn has a send queue that it
+ * takes care of sending if you call rconn_run(), which will have the side
+ * effect of waking up poll_block(). */
+int
+rconn_send(struct rconn *rc, struct ofpbuf *b,
+ struct rconn_packet_counter *counter)
+ OVS_EXCLUDED(rc->mutex)
+{
+ int error;
+
+ ovs_mutex_lock(&rc->mutex);
+ error = rconn_send__(rc, b, counter);
+ ovs_mutex_unlock(&rc->mutex);
+
+ return error;
+}
+
/* Sends 'b' on 'rc'. Increments 'counter' while the packet is in flight; it
* will be decremented when it has been sent (or discarded due to
* disconnection). Returns 0 if successful, EAGAIN if 'counter->n' is already
int
rconn_send_with_limit(struct rconn *rc, struct ofpbuf *b,
struct rconn_packet_counter *counter, int queue_limit)
+ OVS_EXCLUDED(rc->mutex)
{
- if (counter->n_packets < queue_limit) {
- return rconn_send(rc, b, counter);
+ int error;
+
+ ovs_mutex_lock(&rc->mutex);
+ if (rconn_packet_counter_n_packets(counter) < queue_limit) {
+ error = rconn_send__(rc, b, counter);
} else {
COVERAGE_INC(rconn_overflow);
ofpbuf_delete(b);
- return EAGAIN;
+ error = EAGAIN;
}
+ ovs_mutex_unlock(&rc->mutex);
+
+ return error;
}
/* Returns the total number of packets successfully sent on the underlying
* and received on 'rconn' will be copied. 'rc' takes ownership of 'vconn'. */
void
rconn_add_monitor(struct rconn *rc, struct vconn *vconn)
+ OVS_EXCLUDED(rc->mutex)
{
+ ovs_mutex_lock(&rc->mutex);
if (rc->n_monitors < ARRAY_SIZE(rc->monitors)) {
VLOG_INFO("new monitor connection from %s", vconn_get_name(vconn));
rc->monitors[rc->n_monitors++] = vconn;
vconn_get_name(vconn));
vconn_close(vconn);
}
+ ovs_mutex_unlock(&rc->mutex);
}
/* Returns 'rc''s name. This is a name for human consumption, appropriate for
/* Sets 'rc''s name to 'new_name'. */
void
rconn_set_name(struct rconn *rc, const char *new_name)
+ OVS_EXCLUDED(rc->mutex)
{
+ ovs_mutex_lock(&rc->mutex);
free(rc->name);
rc->name = xstrdup(new_name);
+ ovs_mutex_unlock(&rc->mutex);
}
/* Returns 'rc''s target. This is intended to be a string that may be passed
return is_connected_state(rconn->state);
}
+static bool
+rconn_is_admitted__(const struct rconn *rconn)
+ OVS_REQUIRES(rconn->mutex)
+{
+ return (rconn_is_connected(rconn)
+ && rconn->last_admitted >= rconn->last_connected);
+}
+
/* Returns true if 'rconn' is connected and thought to have been accepted by
* the peer's admission-control policy. */
bool
rconn_is_admitted(const struct rconn *rconn)
+ OVS_EXCLUDED(rconn->mutex)
{
- return (rconn_is_connected(rconn)
- && rconn->last_admitted >= rconn->last_connected);
+ bool admitted;
+
+ ovs_mutex_lock(&rconn->mutex);
+ admitted = rconn_is_admitted__(rconn);
+ ovs_mutex_unlock(&rconn->mutex);
+
+ return admitted;
}
/* Returns 0 if 'rconn' is currently connected and considered to have been
* seconds since 'rconn' was last in such a state. */
int
rconn_failure_duration(const struct rconn *rconn)
+ OVS_EXCLUDED(rconn->mutex)
{
- return rconn_is_admitted(rconn) ? 0 : time_now() - rconn->last_admitted;
-}
+ int duration;
-/* Returns the IP address of the peer, or 0 if the peer's IP address is not
- * known. */
-ovs_be32
-rconn_get_remote_ip(const struct rconn *rconn)
-{
- return rconn->remote_ip;
-}
+ ovs_mutex_lock(&rconn->mutex);
+ duration = (rconn_is_admitted__(rconn)
+ ? 0
+ : time_now() - rconn->last_admitted);
+ ovs_mutex_unlock(&rconn->mutex);
-/* Returns the transport port of the peer, or 0 if the peer's port is not
- * known. */
-ovs_be16
-rconn_get_remote_port(const struct rconn *rconn)
-{
- return rconn->remote_port;
+ return duration;
}
-/* Returns the IP address used to connect to the peer, or 0 if the
- * connection is not an IP-based protocol or if its IP address is not
- * known. */
-ovs_be32
-rconn_get_local_ip(const struct rconn *rconn)
-{
- return rconn->local_ip;
-}
-
-/* Returns the transport port used to connect to the peer, or 0 if the
- * connection does not contain a port or if the port is not known. */
-ovs_be16
-rconn_get_local_port(const struct rconn *rconn)
+static int
+rconn_get_version__(const struct rconn *rconn)
+ OVS_REQUIRES(rconn->mutex)
{
- return rconn->vconn ? vconn_get_local_port(rconn->vconn) : 0;
+ return rconn->vconn ? vconn_get_version(rconn->vconn) : -1;
}
/* Returns the OpenFlow version negotiated with the peer, or -1 if there is
* currently no connection or if version negotiation is not yet complete. */
int
rconn_get_version(const struct rconn *rconn)
+ OVS_EXCLUDED(rconn->mutex)
{
- return rconn->vconn ? vconn_get_version(rconn->vconn) : -1;
+ int version;
+
+ ovs_mutex_lock(&rconn->mutex);
+ version = rconn_get_version__(rconn);
+ ovs_mutex_unlock(&rconn->mutex);
+
+ return version;
}
/* Returns the total number of packets successfully received by the underlying
/* Returns the number of messages queued for transmission on 'rc'. */
unsigned int
rconn_count_txqlen(const struct rconn *rc)
+ OVS_EXCLUDED(rc->mutex)
{
- return list_size(&rc->txq);
+ unsigned int len;
+
+ ovs_mutex_lock(&rc->mutex);
+ len = list_size(&rc->txq);
+ ovs_mutex_unlock(&rc->mutex);
+
+ return len;
}
\f
struct rconn_packet_counter *
rconn_packet_counter_create(void)
{
struct rconn_packet_counter *c = xzalloc(sizeof *c);
+ ovs_mutex_init(&c->mutex);
+ ovs_mutex_lock(&c->mutex);
c->ref_cnt = 1;
+ ovs_mutex_unlock(&c->mutex);
return c;
}
rconn_packet_counter_destroy(struct rconn_packet_counter *c)
{
if (c) {
+ bool dead;
+
+ ovs_mutex_lock(&c->mutex);
ovs_assert(c->ref_cnt > 0);
- if (!--c->ref_cnt && !c->n_packets) {
+ dead = !--c->ref_cnt && !c->n_packets;
+ ovs_mutex_unlock(&c->mutex);
+
+ if (dead) {
+ ovs_mutex_destroy(&c->mutex);
free(c);
}
}
void
rconn_packet_counter_inc(struct rconn_packet_counter *c, unsigned int n_bytes)
{
+ ovs_mutex_lock(&c->mutex);
c->n_packets++;
c->n_bytes += n_bytes;
+ ovs_mutex_unlock(&c->mutex);
}
void
rconn_packet_counter_dec(struct rconn_packet_counter *c, unsigned int n_bytes)
{
- ovs_assert(c->n_packets > 0);
- ovs_assert(c->n_bytes >= n_bytes);
+ bool dead = false;
- c->n_bytes -= n_bytes;
+ ovs_mutex_lock(&c->mutex);
+ ovs_assert(c->n_packets > 0);
+ ovs_assert(c->n_packets == 1
+ ? c->n_bytes == n_bytes
+ : c->n_bytes > n_bytes);
c->n_packets--;
- if (!c->n_packets) {
- ovs_assert(!c->n_bytes);
- if (!c->ref_cnt) {
- free(c);
- }
+ c->n_bytes -= n_bytes;
+ dead = !c->n_packets && !c->ref_cnt;
+ ovs_mutex_unlock(&c->mutex);
+
+ if (dead) {
+ ovs_mutex_destroy(&c->mutex);
+ free(c);
}
}
+
+unsigned int
+rconn_packet_counter_n_packets(const struct rconn_packet_counter *c)
+{
+ unsigned int n;
+
+ ovs_mutex_lock(&c->mutex);
+ n = c->n_packets;
+ ovs_mutex_unlock(&c->mutex);
+
+ return n;
+}
+
+unsigned int
+rconn_packet_counter_n_bytes(const struct rconn_packet_counter *c)
+{
+ unsigned int n;
+
+ ovs_mutex_lock(&c->mutex);
+ n = c->n_bytes;
+ ovs_mutex_unlock(&c->mutex);
+
+ return n;
+}
\f
/* Set rc->target and rc->name to 'target' and 'name', respectively. If 'name'
* is null, 'target' is used.
* the target also likely changes these values. */
static void
rconn_set_target__(struct rconn *rc, const char *target, const char *name)
+ OVS_REQUIRES(rc->mutex)
{
free(rc->name);
rc->name = xstrdup(name ? name : target);
free(rc->target);
rc->target = xstrdup(target);
- rc->local_ip = 0;
- rc->remote_ip = 0;
- rc->remote_port = 0;
}
/* Tries to send a packet from 'rc''s send buffer. Returns 0 if successful,
* otherwise a positive errno value. */
static int
try_send(struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
struct ofpbuf *msg = ofpbuf_from_list(rc->txq.next);
unsigned int n_bytes = msg->size;
* normally. */
static void
report_error(struct rconn *rc, int error)
+ OVS_REQUIRES(rc->mutex)
{
if (error == EOF) {
/* If 'rc' isn't reliable, then we don't really expect this connection
*/
static void
disconnect(struct rconn *rc, int error)
+ OVS_REQUIRES(rc->mutex)
{
rc->last_error = error;
if (rc->reliable) {
state_transition(rc, S_BACKOFF);
} else {
rc->last_disconnected = time_now();
- rconn_disconnect(rc);
+ rconn_disconnect__(rc);
}
}
* counts. */
static void
flush_queue(struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
if (list_is_empty(&rc->txq)) {
return;
static unsigned int
elapsed_in_this_state(const struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
return time_now() - rc->state_entered;
}
static unsigned int
timeout(const struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
switch (rc->state) {
#define STATE(NAME, VALUE) case S_##NAME: return timeout_##NAME(rc);
STATES
#undef STATE
default:
- NOT_REACHED();
+ OVS_NOT_REACHED();
}
}
static bool
timed_out(const struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
return time_now() >= sat_add(rc->state_entered, timeout(rc));
}
static void
state_transition(struct rconn *rc, enum state state)
+ OVS_REQUIRES(rc->mutex)
{
rc->seqno += (rc->state == S_ACTIVE) != (state == S_ACTIVE);
if (is_connected_state(state) && !is_connected_state(rc->state)) {
static void
close_monitor(struct rconn *rc, size_t idx, int retval)
+ OVS_REQUIRES(rc->mutex)
{
VLOG_DBG("%s: closing monitor connection to %s: %s",
rconn_get_name(rc), vconn_get_name(rc->monitors[idx]),
static void
copy_to_monitor(struct rconn *rc, const struct ofpbuf *b)
+ OVS_REQUIRES(rc->mutex)
{
struct ofpbuf *clone = NULL;
int retval;
return (state & (S_ACTIVE | S_IDLE)) != 0;
}
+/* When a switch initially connects to a controller, the controller may spend a
+ * little time examining the switch, looking at, for example, its datapath ID,
+ * before it decides whether it is willing to control that switch. At that
+ * point, it either disconnects or starts controlling the switch.
+ *
+ * This function returns a guess to its caller about whether 'b' is OpenFlow
+ * message that indicates that the controller has decided to control the
+ * switch. It returns false if the message is one that a controller typically
+ * uses to determine whether a switch is admissible, true if the message is one
+ * that would typically be used only after the controller has admitted the
+ * switch. */
static bool
is_admitted_msg(const struct ofpbuf *b)
{
case OFPTYPE_GET_CONFIG_REQUEST:
case OFPTYPE_GET_CONFIG_REPLY:
case OFPTYPE_SET_CONFIG:
- /* FIXME: Change the following once they are implemented: */
case OFPTYPE_QUEUE_GET_CONFIG_REQUEST:
case OFPTYPE_QUEUE_GET_CONFIG_REPLY:
case OFPTYPE_GET_ASYNC_REQUEST:
case OFPTYPE_GET_ASYNC_REPLY:
- case OFPTYPE_GROUP_REQUEST:
- case OFPTYPE_GROUP_REPLY:
- case OFPTYPE_GROUP_DESC_REQUEST:
- case OFPTYPE_GROUP_DESC_REPLY:
- case OFPTYPE_GROUP_FEATURES_REQUEST:
- case OFPTYPE_GROUP_FEATURES_REPLY:
- case OFPTYPE_TABLE_FEATURES_REQUEST:
- case OFPTYPE_TABLE_FEATURES_REPLY:
+ case OFPTYPE_GROUP_STATS_REQUEST:
+ case OFPTYPE_GROUP_STATS_REPLY:
+ case OFPTYPE_GROUP_DESC_STATS_REQUEST:
+ case OFPTYPE_GROUP_DESC_STATS_REPLY:
+ case OFPTYPE_GROUP_FEATURES_STATS_REQUEST:
+ case OFPTYPE_GROUP_FEATURES_STATS_REPLY:
+ case OFPTYPE_TABLE_FEATURES_STATS_REQUEST:
+ case OFPTYPE_TABLE_FEATURES_STATS_REPLY:
return false;
case OFPTYPE_PACKET_IN:
case OFPTYPE_PORT_STATUS:
case OFPTYPE_PACKET_OUT:
case OFPTYPE_FLOW_MOD:
+ case OFPTYPE_GROUP_MOD:
case OFPTYPE_PORT_MOD:
+ case OFPTYPE_TABLE_MOD:
case OFPTYPE_METER_MOD:
case OFPTYPE_BARRIER_REQUEST:
case OFPTYPE_BARRIER_REPLY:
case OFPTYPE_QUEUE_STATS_REPLY:
case OFPTYPE_PORT_DESC_STATS_REQUEST:
case OFPTYPE_PORT_DESC_STATS_REPLY:
- case OFPTYPE_METER_REQUEST:
- case OFPTYPE_METER_REPLY:
- case OFPTYPE_METER_CONFIG_REQUEST:
- case OFPTYPE_METER_CONFIG_REPLY:
- case OFPTYPE_METER_FEATURES_REQUEST:
- case OFPTYPE_METER_FEATURES_REPLY:
+ case OFPTYPE_METER_STATS_REQUEST:
+ case OFPTYPE_METER_STATS_REPLY:
+ case OFPTYPE_METER_CONFIG_STATS_REQUEST:
+ case OFPTYPE_METER_CONFIG_STATS_REPLY:
+ case OFPTYPE_METER_FEATURES_STATS_REQUEST:
+ case OFPTYPE_METER_FEATURES_STATS_REPLY:
case OFPTYPE_ROLE_REQUEST:
case OFPTYPE_ROLE_REPLY:
+ case OFPTYPE_ROLE_STATUS:
case OFPTYPE_SET_FLOW_FORMAT:
case OFPTYPE_FLOW_MOD_TABLE_ID:
case OFPTYPE_SET_PACKET_IN_FORMAT:
* successuflly connected in too long. */
static bool
rconn_logging_connection_attempts__(const struct rconn *rc)
+ OVS_REQUIRES(rc->mutex)
{
return rc->backoff < rc->max_backoff;
}