/* * Copyright (c) 2014 Nicira, Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include #include "ovs-rcu.h" #include "guarded-list.h" #include "list.h" #include "ovs-thread.h" #include "poll-loop.h" #include "seq.h" #include "timeval.h" #include "vlog.h" VLOG_DEFINE_THIS_MODULE(ovs_rcu); struct ovsrcu_cb { void (*function)(void *aux); void *aux; }; struct ovsrcu_cbset { struct list list_node; struct ovsrcu_cb cbs[16]; int n_cbs; }; struct ovsrcu_perthread { struct list list_node; /* In global list. */ struct ovs_mutex mutex; uint64_t seqno; struct ovsrcu_cbset *cbset; char name[16]; /* This thread's name. */ }; static struct seq *global_seqno; static pthread_key_t perthread_key; static struct list ovsrcu_threads; static struct ovs_mutex ovsrcu_threads_mutex; static struct guarded_list flushed_cbsets; static struct seq *flushed_cbsets_seq; static void ovsrcu_init(void); static void ovsrcu_flush_cbset(struct ovsrcu_perthread *); static void ovsrcu_unregister__(struct ovsrcu_perthread *); static bool ovsrcu_call_postponed(void); static void *ovsrcu_postpone_thread(void *arg OVS_UNUSED); static void ovsrcu_synchronize(void); static struct ovsrcu_perthread * ovsrcu_perthread_get(void) { struct ovsrcu_perthread *perthread; ovsrcu_init(); perthread = pthread_getspecific(perthread_key); if (!perthread) { const char *name = get_subprogram_name(); perthread = xmalloc(sizeof *perthread); ovs_mutex_init(&perthread->mutex); perthread->seqno = seq_read(global_seqno); perthread->cbset = NULL; ovs_strlcpy(perthread->name, name[0] ? name : "main", sizeof perthread->name); ovs_mutex_lock(&ovsrcu_threads_mutex); list_push_back(&ovsrcu_threads, &perthread->list_node); ovs_mutex_unlock(&ovsrcu_threads_mutex); pthread_setspecific(perthread_key, perthread); } return perthread; } /* Indicates the end of a quiescent state. See "Details" near the top of * ovs-rcu.h. * * Quiescent states don't stack or nest, so this always ends a quiescent state * even if ovsrcu_quiesce_start() was called multiple times in a row. */ void ovsrcu_quiesce_end(void) { ovsrcu_perthread_get(); } static void ovsrcu_quiesced(void) { if (single_threaded()) { ovsrcu_call_postponed(); } else { static struct ovsthread_once once = OVSTHREAD_ONCE_INITIALIZER; if (ovsthread_once_start(&once)) { ovs_thread_create("urcu", ovsrcu_postpone_thread, NULL); ovsthread_once_done(&once); } } } /* Indicates the beginning of a quiescent state. See "Details" near the top of * ovs-rcu.h. */ void ovsrcu_quiesce_start(void) { struct ovsrcu_perthread *perthread; ovsrcu_init(); perthread = pthread_getspecific(perthread_key); if (perthread) { pthread_setspecific(perthread_key, NULL); ovsrcu_unregister__(perthread); } ovsrcu_quiesced(); } /* Indicates a momentary quiescent state. See "Details" near the top of * ovs-rcu.h. */ void ovsrcu_quiesce(void) { ovsrcu_init(); ovsrcu_perthread_get()->seqno = seq_read(global_seqno); seq_change(global_seqno); ovsrcu_quiesced(); } bool ovsrcu_is_quiescent(void) { ovsrcu_init(); return pthread_getspecific(perthread_key) == NULL; } static void ovsrcu_synchronize(void) { unsigned int warning_threshold = 1000; uint64_t target_seqno; long long int start; if (single_threaded()) { return; } target_seqno = seq_read(global_seqno); ovsrcu_quiesce_start(); start = time_msec(); for (;;) { uint64_t cur_seqno = seq_read(global_seqno); struct ovsrcu_perthread *perthread; char stalled_thread[16]; unsigned int elapsed; bool done = true; ovs_mutex_lock(&ovsrcu_threads_mutex); LIST_FOR_EACH (perthread, list_node, &ovsrcu_threads) { if (perthread->seqno <= target_seqno) { ovs_strlcpy(stalled_thread, perthread->name, sizeof stalled_thread); done = false; break; } } ovs_mutex_unlock(&ovsrcu_threads_mutex); if (done) { break; } elapsed = time_msec() - start; if (elapsed >= warning_threshold) { VLOG_WARN("blocked %u ms waiting for %s to quiesce", elapsed, stalled_thread); warning_threshold *= 2; } poll_timer_wait_until(start + warning_threshold); seq_wait(global_seqno, cur_seqno); poll_block(); } ovsrcu_quiesce_end(); } /* Registers 'function' to be called, passing 'aux' as argument, after the * next grace period. * * This function is more conveniently called through the ovsrcu_postpone() * macro, which provides a type-safe way to allow 'function''s parameter to be * any pointer type. */ void ovsrcu_postpone__(void (*function)(void *aux), void *aux) { struct ovsrcu_perthread *perthread = ovsrcu_perthread_get(); struct ovsrcu_cbset *cbset; struct ovsrcu_cb *cb; cbset = perthread->cbset; if (!cbset) { cbset = perthread->cbset = xmalloc(sizeof *perthread->cbset); cbset->n_cbs = 0; } cb = &cbset->cbs[cbset->n_cbs++]; cb->function = function; cb->aux = aux; if (cbset->n_cbs >= ARRAY_SIZE(cbset->cbs)) { ovsrcu_flush_cbset(perthread); } } static bool ovsrcu_call_postponed(void) { struct ovsrcu_cbset *cbset, *next_cbset; struct list cbsets; guarded_list_pop_all(&flushed_cbsets, &cbsets); if (list_is_empty(&cbsets)) { return false; } ovsrcu_synchronize(); LIST_FOR_EACH_SAFE (cbset, next_cbset, list_node, &cbsets) { struct ovsrcu_cb *cb; for (cb = cbset->cbs; cb < &cbset->cbs[cbset->n_cbs]; cb++) { cb->function(cb->aux); } list_remove(&cbset->list_node); free(cbset); } return true; } static void * ovsrcu_postpone_thread(void *arg OVS_UNUSED) { pthread_detach(pthread_self()); for (;;) { uint64_t seqno = seq_read(flushed_cbsets_seq); if (!ovsrcu_call_postponed()) { seq_wait(flushed_cbsets_seq, seqno); poll_block(); } } OVS_NOT_REACHED(); } static void ovsrcu_flush_cbset(struct ovsrcu_perthread *perthread) { struct ovsrcu_cbset *cbset = perthread->cbset; if (cbset) { guarded_list_push_back(&flushed_cbsets, &cbset->list_node, SIZE_MAX); perthread->cbset = NULL; seq_change(flushed_cbsets_seq); } } static void ovsrcu_unregister__(struct ovsrcu_perthread *perthread) { if (perthread->cbset) { ovsrcu_flush_cbset(perthread); } ovs_mutex_lock(&ovsrcu_threads_mutex); list_remove(&perthread->list_node); ovs_mutex_unlock(&ovsrcu_threads_mutex); ovs_mutex_destroy(&perthread->mutex); free(perthread); seq_change(global_seqno); } static void ovsrcu_thread_exit_cb(void *perthread) { ovsrcu_unregister__(perthread); } static void ovsrcu_init(void) { static struct ovsthread_once once = OVSTHREAD_ONCE_INITIALIZER; if (ovsthread_once_start(&once)) { global_seqno = seq_create(); xpthread_key_create(&perthread_key, ovsrcu_thread_exit_cb); list_init(&ovsrcu_threads); ovs_mutex_init(&ovsrcu_threads_mutex); guarded_list_init(&flushed_cbsets); flushed_cbsets_seq = seq_create(); ovsthread_once_done(&once); } }