2
0
mirror of https://github.com/openvswitch/ovs synced 2025-08-23 18:37:36 +00:00
ovs/lib/cmap.c

836 lines
27 KiB
C
Raw Normal View History

cmap: New module for cuckoo hash table. This implements an "optimistic concurrent cuckoo hash", a single-writer, multiple-reader hash table data structure. The point of this data structure is performance, so this commit message focuses on performance. I tested the performance of cmap with the test-cmap utility included in this commit. It takes three parameters for benchmarking: - n, the number of elements to insert. - n_threads, the number of threads to use for searching and mutating the hash table. - mutations, the percentage of operations that should modify the hash table, from 0% to 100%. e.g. "test-cmap 1000000 16 1" inserts one million elements, uses 16 threads, and 1% of the operations modify the hash table. Any given run does the following for both hmap and cmap implementations: - Inserts n elements into a hash table. - Iterates over all of the elements. - Spawns n_threads threads, each of which searches for each of the elements in the hash table, once, and removes the specified percentage of them. - Removes each of the (remaining) elements and destroys the hash table. and reports the time taken by each step, The tables below report results for various parameters with a draft version of this library. The tests were not formally rerun for the final version, but the intermediate changes should only have improved performance, and this seemed to be the case in some informal testing. n_threads=16 was used each time, on a 16-core x86-64 machine. The compiler used was Clang 3.5. (GCC yields different numbers but similar relative results.) The results show: - Insertion is generally 3x to 5x faster in an hmap. - Iteration is generally about 3x faster in a cmap. - Search and mutation is 4x faster with .1% mutations and the advantage grows as the fraction of mutations grows. This is because a cmap does not require locking for read operations, even in the presence of a writer. With no mutations, however, no locking is required in the hmap case, and the hmap is somewhat faster. This is because raw hmap search is somewhat simpler and faster than raw cmap search. - Destruction is faster, usually by less than 2x, in an hmap. n=10,000,000: .1% mutations 1% mutations 10% mutations no mutations cmap hmap cmap hmap cmap hmap cmap hmap insert: 6132 2182 6136 2178 6111 2174 6124 2180 iterate: 370 1203 378 1201 370 1200 371 1202 search: 1375 8692 2393 28197 18402 80379 1281 1097 destroy: 1382 1187 1197 1034 324 241 1405 1205 n=1,000,000: .1% mutations 1% mutations 10% mutations no mutations cmap hmap cmap hmap cmap hmap cmap hmap insert: 311 25 310 60 311 59 310 60 iterate: 25 62 25 64 25 57 25 60 search: 115 637 197 2266 1803 7284 101 67 destroy: 103 64 90 59 25 13 104 66 n=100,000: .1% mutations 1% mutations 10% mutations no mutations cmap hmap cmap hmap cmap hmap cmap hmap insert: 25 6 26 5 25 5 25 5 iterate: 1 3 1 3 1 3 2 3 search: 12 57 27 219 164 735 10 5 destroy: 5 3 6 3 2 1 6 4 Signed-off-by: Ben Pfaff <blp@nicira.com> Acked-by: Jarno Rajahalme <jrajahalme@nicira.com>
2014-05-20 16:51:42 -07:00
/*
* Copyright (c) 2014 Nicira, Inc.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at:
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include <config.h>
#include "cmap.h"
#include "hash.h"
#include "ovs-rcu.h"
#include "random.h"
#include "util.h"
/* Optimistic Concurrent Cuckoo Hash
* =================================
*
* A "cuckoo hash" is an open addressing hash table schema, designed such that
* a given element can be in one of only a small number of buckets 'd', each of
* which holds up to a small number 'k' elements. Thus, the expected and
* worst-case lookup times are O(1) because they require comparing no more than
* a fixed number of elements (k * d). Inserting a new element can require
* moving around existing elements, but it is also O(1) amortized expected
* time.
*
* An optimistic concurrent hash table goes one step further, making it
* possible for a single writer to execute concurrently with any number of
* readers without requiring the readers to take any locks.
*
* This cuckoo hash implementation uses:
*
* - Two hash functions (d=2). More hash functions allow for a higher load
* factor, but increasing 'k' is easier and the benefits of increasing 'd'
* quickly fall off with the 'k' values used here. Also, the method of
* generating hashes used in this implementation is hard to reasonably
* extend beyond d=2. Finally, each additional hash function means that a
* lookup has to look at least one extra cache line.
*
* - 5 or 7 elements per bucket (k=5 or k=7), chosen to make buckets
* exactly one cache line in size.
*
* According to Erlingsson [4], these parameters suggest a maximum load factor
* of about 93%. The current implementation is conservative, expanding the
* hash table when it is over 85% full.
*
*
* Hash Functions
* ==============
*
* A cuckoo hash requires multiple hash functions. When reorganizing the hash
* becomes too difficult, it also requires the ability to change the hash
* functions. Requiring the client to provide multiple hashes and to be able
* to change them to new hashes upon insertion is inconvenient.
*
* This implementation takes another approach. The client provides a single,
* fixed hash. The cuckoo hash internally "rehashes" this hash against a
* randomly selected basis value (see rehash()). This rehashed value is one of
* the two hashes. The other hash is computed by 16-bit circular rotation of
* the rehashed value. Updating the basis changes the hash functions.
*
* To work properly, the hash functions used by a cuckoo hash must be
* independent. If one hash function is a function of the other (e.g. h2(x) =
* h1(x) + 1, or h2(x) = hash(h1(x))), then insertion will eventually fail
* catastrophically (loop forever) because of collisions. With this rehashing
* technique, the two hashes are completely independent for masks up to 16 bits
* wide. For masks wider than 16 bits, only 32-n bits are independent between
* the two hashes. Thus, it becomes risky to grow a cuckoo hash table beyond
* about 2**24 buckets (about 71 million elements with k=5 and maximum load
* 85%). Fortunately, Open vSwitch does not normally deal with hash tables
* this large.
*
*
* Handling Duplicates
* ===================
*
* This cuckoo hash table implementation deals with duplicate client-provided
* hash values by chaining: the second and subsequent cmap_nodes with a given
* hash are chained off the initially inserted node's 'next' member. The hash
* table maintains the invariant that a single client-provided hash value
* exists in only a single chain in a single bucket (even though that hash
* could be stored in two buckets).
*
*
* References
* ==========
*
* [1] D. Zhou, B. Fan, H. Lim, M. Kaminsky, D. G. Andersen, "Scalable, High
* Performance Ethernet Forwarding with CuckooSwitch". In Proc. 9th
* CoNEXT, Dec. 2013.
*
* [2] B. Fan, D. G. Andersen, and M. Kaminsky. "MemC3: Compact and concurrent
* memcache with dumber caching and smarter hashing". In Proc. 10th USENIX
* NSDI, Apr. 2013
*
* [3] R. Pagh and F. Rodler. "Cuckoo hashing". Journal of Algorithms, 51(2):
* 122-144, May 2004.
*
* [4] U. Erlingsson, M. Manasse, F. McSherry, "A Cool and Practical
* Alternative to Traditional Hash Tables". In Proc. 7th Workshop on
* Distributed Data and Structures (WDAS'06), 2006.
*/
/* An entry is an int and a pointer: 8 bytes on 32-bit, 12 bytes on 64-bit. */
#define CMAP_ENTRY_SIZE (4 + (UINTPTR_MAX == UINT32_MAX ? 4 : 8))
/* Number of entries per bucket: 7 on 32-bit, 5 on 64-bit. */
#define CMAP_K ((CACHE_LINE_SIZE - 4) / CMAP_ENTRY_SIZE)
/* Pad to make a bucket a full cache line in size: 4 on 32-bit, 0 on 64-bit. */
#define CMAP_PADDING ((CACHE_LINE_SIZE - 4) - (CMAP_K * CMAP_ENTRY_SIZE))
/* A cuckoo hash bucket. Designed to be cache-aligned and exactly one cache
* line long. */
struct cmap_bucket {
/* Allows readers to track in-progress changes. Initially zero, each
* writer increments this value just before and just after each change (see
* cmap_set_bucket()). Thus, a reader can ensure that it gets a consistent
* snapshot by waiting for the counter to become even (see
* read_even_counter()), then checking that its value does not change while
* examining the bucket (see cmap_find()). */
atomic_uint32_t counter;
/* (hash, node) slots. They are parallel arrays instead of an array of
* structs to reduce the amount of space lost to padding.
*
* The slots are in no particular order. A null pointer indicates that a
* pair is unused. In-use slots are not necessarily in the earliest
* slots. */
uint32_t hashes[CMAP_K];
struct cmap_node *nodes[CMAP_K];
/* Padding to make cmap_bucket exactly one cache line long. */
#if CMAP_PADDING > 0
uint8_t pad[CMAP_PADDING];
#endif
};
BUILD_ASSERT_DECL(sizeof(struct cmap_bucket) == CACHE_LINE_SIZE);
/* Default maximum load factor (as a fraction of UINT32_MAX + 1) before
* enlarging a cmap. Reasonable values lie between about 75% and 93%. Smaller
* values waste memory; larger values increase the average insertion time. */
#define CMAP_MAX_LOAD ((uint32_t) (UINT32_MAX * .85))
/* The implementation of a concurrent hash map. */
struct cmap_impl {
unsigned int n; /* Number of in-use elements. */
unsigned int max_n; /* Max elements before enlarging. */
uint32_t mask; /* Number of 'buckets', minus one. */
uint32_t basis; /* Basis for rehashing client's hash values. */
/* Padding to make cmap_impl exactly one cache line long. */
uint8_t pad[CACHE_LINE_SIZE - sizeof(unsigned int) * 4];
struct cmap_bucket buckets[];
};
BUILD_ASSERT_DECL(sizeof(struct cmap_impl) == CACHE_LINE_SIZE);
static struct cmap_impl *cmap_rehash(struct cmap *, uint32_t mask);
/* Given a rehashed value 'hash', returns the other hash for that rehashed
* value. This is symmetric: other_hash(other_hash(x)) == x. (See also "Hash
* Functions" at the top of this file.) */
static uint32_t
other_hash(uint32_t hash)
{
return (hash << 16) | (hash >> 16);
}
/* Returns the rehashed value for 'hash' within 'impl'. (See also "Hash
* Functions" at the top of this file.) */
static uint32_t
rehash(const struct cmap_impl *impl, uint32_t hash)
{
return mhash_finish(impl->basis, hash);
}
static struct cmap_impl *
cmap_get_impl(const struct cmap *cmap)
{
return ovsrcu_get(struct cmap_impl *, &cmap->impl);
}
static uint32_t
calc_max_n(uint32_t mask)
{
return ((uint64_t) (mask + 1) * CMAP_K * CMAP_MAX_LOAD) >> 32;
}
static struct cmap_impl *
cmap_impl_create(uint32_t mask)
{
struct cmap_impl *impl;
ovs_assert(is_pow2(mask + 1));
impl = xzalloc_cacheline(sizeof *impl
+ (mask + 1) * sizeof *impl->buckets);
impl->n = 0;
impl->max_n = calc_max_n(mask);
impl->mask = mask;
impl->basis = random_uint32();
return impl;
}
/* Initializes 'cmap' as an empty concurrent hash map. */
void
cmap_init(struct cmap *cmap)
{
ovsrcu_set(&cmap->impl, cmap_impl_create(0));
}
/* Destroys 'cmap'.
*
* The client is responsible for destroying any data previously held in
* 'cmap'. */
void
cmap_destroy(struct cmap *cmap)
{
if (cmap) {
free_cacheline(cmap_get_impl(cmap));
}
}
/* Returns the number of elements in 'cmap'. */
size_t
cmap_count(const struct cmap *cmap)
{
return cmap_get_impl(cmap)->n;
}
/* Returns true if 'cmap' is empty, false otherwise. */
bool
cmap_is_empty(const struct cmap *cmap)
{
return cmap_count(cmap) == 0;
}
static uint32_t
read_counter(struct cmap_bucket *bucket, memory_order order)
{
uint32_t counter;
atomic_read_explicit(&bucket->counter, &counter, order);
return counter;
}
static uint32_t
read_even_counter(struct cmap_bucket *bucket)
{
uint32_t counter;
do {
counter = read_counter(bucket, memory_order_acquire);
} while (OVS_UNLIKELY(counter & 1));
return counter;
}
static bool
counter_changed(struct cmap_bucket *b, uint32_t c)
{
return OVS_UNLIKELY(read_counter(b, memory_order_release) != c);
}
/* Searches 'cmap' for an element with the specified 'hash'. If one or more is
* found, returns a pointer to the first one, otherwise a null pointer. All of
* the nodes on the returned list are guaranteed to have exactly the given
* 'hash'.
*
* This function works even if 'cmap' is changing concurrently. If 'cmap' is
* not changing, then cmap_find_protected() is slightly faster.
*
* CMAP_FOR_EACH_WITH_HASH is usually more convenient. */
struct cmap_node *
cmap_find(const struct cmap *cmap, uint32_t hash)
{
struct cmap_impl *impl = cmap_get_impl(cmap);
uint32_t h1 = rehash(impl, hash);
uint32_t h2 = other_hash(h1);
struct cmap_bucket *b1;
struct cmap_bucket *b2;
uint32_t c1, c2;
int i;
retry:
b1 = &impl->buckets[h1 & impl->mask];
c1 = read_even_counter(b1);
for (i = 0; i < CMAP_K; i++) {
struct cmap_node *node = b1->nodes[i];
if (node && b1->hashes[i] == hash) {
if (counter_changed(b1, c1)) {
goto retry;
}
return node;
}
}
b2 = &impl->buckets[h2 & impl->mask];
c2 = read_even_counter(b2);
for (i = 0; i < CMAP_K; i++) {
struct cmap_node *node = b2->nodes[i];
if (node && b2->hashes[i] == hash) {
if (counter_changed(b2, c2)) {
goto retry;
}
return node;
}
}
if (counter_changed(b1, c1) || counter_changed(b2, c2)) {
goto retry;
}
return NULL;
}
static int
cmap_find_slot(struct cmap_bucket *b, uint32_t hash)
{
int i;
for (i = 0; i < CMAP_K; i++) {
struct cmap_node *node = b->nodes[i];
if (node && b->hashes[i] == hash) {
return i;
}
}
return -1;
}
static struct cmap_node *
cmap_find_bucket_protected(struct cmap_impl *impl, uint32_t hash, uint32_t h)
{
struct cmap_bucket *b = &impl->buckets[h & impl->mask];
int i;
for (i = 0; i < CMAP_K; i++) {
struct cmap_node *node = b->nodes[i];
if (node && b->hashes[i] == hash) {
return node;
}
}
return NULL;
}
/* Like cmap_find(), but only for use if 'cmap' cannot change concurrently.
*
* CMAP_FOR_EACH_WITH_HASH_PROTECTED is usually more convenient. */
struct cmap_node *
cmap_find_protected(const struct cmap *cmap, uint32_t hash)
{
struct cmap_impl *impl = cmap_get_impl(cmap);
uint32_t h1 = rehash(impl, hash);
uint32_t h2 = other_hash(hash);
struct cmap_node *node;
node = cmap_find_bucket_protected(impl, hash, h1);
if (node) {
return node;
}
return cmap_find_bucket_protected(impl, hash, h2);
}
static int
cmap_find_empty_slot(const struct cmap_bucket *b)
{
int i;
for (i = 0; i < CMAP_K; i++) {
if (!b->nodes[i]) {
return i;
}
}
return -1;
}
static void
cmap_set_bucket(struct cmap_bucket *b, int i,
struct cmap_node *node, uint32_t hash)
{
uint32_t c;
atomic_read_explicit(&b->counter, &c, memory_order_acquire);
atomic_store_explicit(&b->counter, c + 1, memory_order_release);
b->nodes[i] = node;
b->hashes[i] = hash;
atomic_store_explicit(&b->counter, c + 2, memory_order_release);
}
/* Searches 'b' for a node with the given 'hash'. If it finds one, adds
* 'new_node' to the node's linked list and returns true. If it does not find
* one, returns false. */
static bool
cmap_insert_dup(struct cmap_node *new_node, uint32_t hash,
struct cmap_bucket *b)
{
int i;
for (i = 0; i < CMAP_K; i++) {
struct cmap_node *node = b->nodes[i];
if (b->hashes[i] == hash) {
if (node) {
struct cmap_node *p;
/* The common case is that 'new_node' is a singleton, with a
* null 'next' pointer, but rehashing can add a longer chain.
* Find the end of the chain starting at 'new_node', then
* splice 'node' to the end of that chain. */
p = new_node;
for (;;) {
struct cmap_node *next = cmap_node_next_protected(p);
if (!next) {
break;
}
p = next;
}
ovsrcu_set(&p->next, node);
} else {
/* The hash value is there from some previous insertion, but
* the associated node has been removed. We're not really
* inserting a duplicate, but we can still reuse the slot.
* Carry on. */
}
/* Change the bucket to point to 'new_node'. This is a degenerate
* form of cmap_set_bucket() that doesn't update the counter since
* we're only touching one field and in a way that doesn't change
* the bucket's meaning for readers. */
b->nodes[i] = new_node;
atomic_thread_fence(memory_order_release);
return true;
}
}
return false;
}
/* Searches 'b' for an empty slot. If successful, stores 'node' and 'hash' in
* the slot and returns true. Otherwise, returns false. */
static bool
cmap_insert_bucket(struct cmap_node *node, uint32_t hash,
struct cmap_bucket *b)
{
int i;
for (i = 0; i < CMAP_K; i++) {
if (!b->nodes[i]) {
cmap_set_bucket(b, i, node, hash);
return true;
}
}
return false;
}
/* Returns the other bucket that b->nodes[slot] could occupy in 'impl'. (This
* might be the same as 'b'.) */
static struct cmap_bucket *
other_bucket(struct cmap_impl *impl, struct cmap_bucket *b, int slot)
{
uint32_t h1 = rehash(impl, b->hashes[slot]);
uint32_t h2 = other_hash(h1);
uint32_t b_idx = b - impl->buckets;
uint32_t other_h = (h1 & impl->mask) == b_idx ? h2 : h1;
return &impl->buckets[other_h & impl->mask];
}
/* 'new_node' is to be inserted into 'impl', but both candidate buckets 'b1'
* and 'b2' are full. This function attempts to rearrange buckets within
* 'impl' to make room for 'new_node'.
*
* The implementation is a general-purpose breadth-first search. At first
* glance, this is more complex than a random walk through 'impl' (suggested by
* some references), but random walks have a tendency to loop back through a
* single bucket. We have to move nodes backward along the path that we find,
* so that no node actually disappears from the hash table, which means a
* random walk would have to be careful to deal with loops. By contrast, a
* successful breadth-first search always finds a *shortest* path through the
* hash table, and a shortest path will never contain loops, so it avoids that
* problem entirely.
*/
static bool
cmap_insert_bfs(struct cmap_impl *impl, struct cmap_node *new_node,
uint32_t hash, struct cmap_bucket *b1, struct cmap_bucket *b2)
{
enum { MAX_PATH = 4 };
/* A path from 'start' to 'end' via the 'n' steps in 'slots[]'.
*
* One can follow the path via:
*
* struct cmap_bucket *b;
* int i;
*
* b = path->start;
* for (i = 0; i < path->n; i++) {
* b = other_bucket(impl, b, path->slots[i]);
* }
* ovs_assert(b == path->end);
*/
struct cmap_path {
struct cmap_bucket *start; /* First bucket along the path. */
struct cmap_bucket *end; /* Last bucket on the path. */
uint8_t slots[MAX_PATH]; /* Slots used for each hop. */
int n; /* Number of slots[]. */
};
/* We need to limit the amount of work we do trying to find a path. It
* might actually be impossible to rearrange the cmap, and after some time
* it is likely to be easier to rehash the entire cmap.
*
* This value of MAX_QUEUE is an arbitrary limit suggested by one of the
* references. Empirically, it seems to work OK. */
enum { MAX_QUEUE = 500 };
struct cmap_path queue[MAX_QUEUE];
int head = 0;
int tail = 0;
/* Add 'b1' and 'b2' as starting points for the search. */
queue[head].start = b1;
queue[head].end = b1;
queue[head].n = 0;
head++;
if (b1 != b2) {
queue[head].start = b2;
queue[head].end = b2;
queue[head].n = 0;
head++;
}
while (tail < head) {
const struct cmap_path *path = &queue[tail++];
struct cmap_bucket *this = path->end;
int i;
for (i = 0; i < CMAP_K; i++) {
struct cmap_bucket *next = other_bucket(impl, this, i);
int j;
if (this == next) {
continue;
}
j = cmap_find_empty_slot(next);
if (j >= 0) {
/* We've found a path along which we can rearrange the hash
* table: Start at path->start, follow all the slots in
* path->slots[], then follow slot 'i', then the bucket you
* arrive at has slot 'j' empty. */
struct cmap_bucket *buckets[MAX_PATH + 2];
int slots[MAX_PATH + 2];
int k;
/* Figure out the full sequence of slots. */
for (k = 0; k < path->n; k++) {
slots[k] = path->slots[k];
}
slots[path->n] = i;
slots[path->n + 1] = j;
/* Figure out the full sequence of buckets. */
buckets[0] = path->start;
for (k = 0; k <= path->n; k++) {
buckets[k + 1] = other_bucket(impl, buckets[k], slots[k]);
}
/* Now the path is fully expressed. One can start from
* buckets[0], go via slots[0] to buckets[1], via slots[1] to
* buckets[2], and so on.
*
* Move all the nodes across the path "backward". After each
* step some node appears in two buckets. Thus, every node is
* always visible to a concurrent search. */
for (k = path->n + 1; k > 0; k--) {
int slot = slots[k - 1];
cmap_set_bucket(buckets[k], slots[k],
buckets[k - 1]->nodes[slot],
buckets[k - 1]->hashes[slot]);
}
/* Finally, replace the first node on the path by
* 'new_node'. */
cmap_set_bucket(buckets[0], slots[0], new_node, hash);
return true;
}
if (path->n < MAX_PATH && head < MAX_QUEUE) {
struct cmap_path *new_path = &queue[head++];
*new_path = *path;
new_path->end = next;
new_path->slots[new_path->n++] = i;
}
}
}
return false;
}
/* Adds 'node', with the given 'hash', to 'impl'.
*
* 'node' is ordinarily a single node, with a null 'next' pointer. When
* rehashing, however, it may be a longer chain of nodes. */
static bool
cmap_try_insert(struct cmap_impl *impl, struct cmap_node *node, uint32_t hash)
{
uint32_t h1 = rehash(impl, hash);
uint32_t h2 = other_hash(h1);
struct cmap_bucket *b1 = &impl->buckets[h1 & impl->mask];
struct cmap_bucket *b2 = &impl->buckets[h2 & impl->mask];
return (OVS_UNLIKELY(cmap_insert_dup(node, hash, b1) ||
cmap_insert_dup(node, hash, b2)) ||
OVS_LIKELY(cmap_insert_bucket(node, hash, b1) ||
cmap_insert_bucket(node, hash, b2)) ||
cmap_insert_bfs(impl, node, hash, b1, b2));
}
/* Inserts 'node', with the given 'hash', into 'cmap'. The caller must ensure
* that 'cmap' cannot change concurrently (from another thread). If duplicates
* are undesirable, the caller must have already verified that 'cmap' does not
* contain a duplicate of 'node'. */
void
cmap_insert(struct cmap *cmap, struct cmap_node *node, uint32_t hash)
{
struct cmap_impl *impl = cmap_get_impl(cmap);
ovsrcu_set(&node->next, NULL);
if (OVS_UNLIKELY(impl->n >= impl->max_n)) {
impl = cmap_rehash(cmap, (impl->mask << 1) | 1);
}
while (OVS_UNLIKELY(!cmap_try_insert(impl, node, hash))) {
impl = cmap_rehash(cmap, impl->mask);
}
impl->n++;
}
static bool
cmap_remove__(struct cmap_impl *impl, struct cmap_node *node,
uint32_t hash, uint32_t h)
{
struct cmap_bucket *b = &impl->buckets[h & impl->mask];
int slot;
slot = cmap_find_slot(b, hash);
if (slot < 0) {
return false;
}
if (b->nodes[slot] == node) {
b->nodes[slot] = cmap_node_next_protected(node);
atomic_thread_fence(memory_order_release);
} else {
struct cmap_node *iter = b->nodes[slot];
for (;;) {
struct cmap_node *next = cmap_node_next_protected(iter);
if (next == node) {
break;
}
iter = next;
}
ovsrcu_set(&iter->next, cmap_node_next_protected(node));
}
return true;
}
/* Removes 'node' from 'cmap'. The caller must ensure that 'cmap' cannot
* change concurrently (from another thread).
*
* 'node' must not be destroyed or modified or inserted back into 'cmap' or
* into any other concurrent hash map while any other thread might be accessing
* it. One correct way to do this is to free it from an RCU callback with
* ovsrcu_postpone(). */
void
cmap_remove(struct cmap *cmap, struct cmap_node *node, uint32_t hash)
{
struct cmap_impl *impl = cmap_get_impl(cmap);
uint32_t h1 = rehash(impl, hash);
uint32_t h2 = other_hash(h1);
bool ok;
ok = (cmap_remove__(impl, node, hash, h1) ||
cmap_remove__(impl, node, hash, h2));
ovs_assert(ok);
impl->n--;
}
static bool
cmap_try_rehash(const struct cmap_impl *old, struct cmap_impl *new)
{
const struct cmap_bucket *b;
for (b = old->buckets; b <= &old->buckets[old->mask]; b++) {
int i;
for (i = 0; i < CMAP_K; i++) {
/* possible optimization here because we know the hashes are
* unique */
struct cmap_node *node = b->nodes[i];
if (node && !cmap_try_insert(new, node, b->hashes[i])) {
return false;
}
}
}
return true;
}
static struct cmap_impl *
cmap_rehash(struct cmap *cmap, uint32_t mask)
{
struct cmap_impl *old = cmap_get_impl(cmap);
struct cmap_impl *new;
new = cmap_impl_create(mask);
ovs_assert(old->n < new->max_n);
while (!cmap_try_rehash(old, new)) {
memset(new->buckets, 0, (mask + 1) * sizeof *new->buckets);
new->basis = random_uint32();
}
new->n = old->n;
ovsrcu_set(&cmap->impl, new);
ovsrcu_postpone(free_cacheline, old);
return new;
}
/* Initializes 'cursor' for iterating through 'cmap'.
*
* Use via CMAP_FOR_EACH. */
void
cmap_cursor_init(struct cmap_cursor *cursor, const struct cmap *cmap)
{
cursor->impl = cmap_get_impl(cmap);
cursor->bucket_idx = 0;
cursor->entry_idx = 0;
}
/* Returns the next node for 'cursor' to visit, following 'node', or NULL if
* the last node has been visited.
*
* Use via CMAP_FOR_EACH. */
struct cmap_node *
cmap_cursor_next(struct cmap_cursor *cursor, const struct cmap_node *node)
{
const struct cmap_impl *impl = cursor->impl;
if (node) {
struct cmap_node *next = cmap_node_next(node);
if (next) {
return next;
}
}
while (cursor->bucket_idx <= impl->mask) {
const struct cmap_bucket *b = &impl->buckets[cursor->bucket_idx];
while (cursor->entry_idx < CMAP_K) {
struct cmap_node *node = b->nodes[cursor->entry_idx++];
if (node) {
return node;
}
}
cursor->bucket_idx++;
cursor->entry_idx = 0;
}
return NULL;
}
/* Returns the next node in 'cmap' in hash order, or NULL if no nodes remain in
* 'cmap'. Uses '*pos' to determine where to begin iteration, and updates
* '*pos' to pass on the next iteration into them before returning.
*
* It's better to use plain CMAP_FOR_EACH and related functions, since they are
* faster and better at dealing with cmaps that change during iteration.
*
* Before beginning iteration, set '*pos' to all zeros. */
struct cmap_node *
cmap_next_position(const struct cmap *cmap,
struct cmap_position *pos)
{
struct cmap_impl *impl = cmap_get_impl(cmap);
unsigned int bucket = pos->bucket;
unsigned int entry = pos->entry;
unsigned int offset = pos->offset;
while (bucket <= impl->mask) {
const struct cmap_bucket *b = &impl->buckets[bucket];
while (entry < CMAP_K) {
const struct cmap_node *node = b->nodes[entry];
unsigned int i;
for (i = 0; node; i++) {
if (i == offset) {
if (cmap_node_next(node)) {
offset++;
} else {
entry++;
offset = 0;
}
pos->bucket = bucket;
pos->entry = entry;
pos->offset = offset;
return CONST_CAST(struct cmap_node *, node);
}
}
entry++;
offset = 0;
}
bucket++;
entry = offset = 0;
}
pos->bucket = pos->entry = pos->offset = 0;
return NULL;
}