Skip to content

Commit e909fe6

Browse files
committed
a
1 parent c93cba5 commit e909fe6

File tree

11 files changed

+233
-131
lines changed

11 files changed

+233
-131
lines changed

cmake/helpers.cmake

Lines changed: 8 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -273,7 +273,14 @@ function(add_umf_target_compile_options name)
273273
# disable warning 6326: Potential comparison of a constant
274274
# with another constant
275275
/wd6326
276-
# disable 4200 warning: nonstandard extension used:
276+
# disable warning 28112: a variable (ptr) which is accessed
277+
# via an Interlocked function must always be accessed via an
278+
# Interlocked function
279+
/wd28112
280+
# disable warning 4324: structure was padded due to
281+
# alignment specifier
282+
/wd4324
283+
# disable warning 4200: nonstandard extension used:
277284
# zero-sized array in struct/union
278285
/wd4200)
279286
if(UMF_DEVELOPER_MODE)

src/critnib/critnib.c

Lines changed: 29 additions & 43 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
/*
22
*
3-
* Copyright (C) 2023-2024 Intel Corporation
3+
* Copyright (C) 2023-2025 Intel Corporation
44
*
55
* Under the Apache License v2.0 with LLVM Exceptions. See LICENSE.TXT.
66
* SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
@@ -133,24 +133,6 @@ struct critnib {
133133
struct utils_mutex_t mutex; /* writes/removes */
134134
};
135135

136-
/*
137-
* atomic load
138-
*/
139-
static void load(void *src, void *dst) {
140-
utils_atomic_load_acquire((word *)src, (word *)dst);
141-
}
142-
143-
static void load64(uint64_t *src, uint64_t *dst) {
144-
utils_atomic_load_acquire(src, dst);
145-
}
146-
147-
/*
148-
* atomic store
149-
*/
150-
static void store(void *dst, void *src) {
151-
utils_atomic_store_release((word *)dst, (word)src);
152-
}
153-
154136
/*
155137
* internal: is_leaf -- check tagged pointer for leafness
156138
*/
@@ -343,10 +325,8 @@ int critnib_insert(struct critnib *c, word key, void *value, int update) {
343325

344326
struct critnib_node *n = c->root;
345327
if (!n) {
346-
store(&c->root, kn);
347-
328+
utils_atomic_store_release_ptr((void **)&c->root, kn);
348329
utils_mutex_unlock(&c->mutex);
349-
350330
return 0;
351331
}
352332

@@ -361,7 +341,8 @@ int critnib_insert(struct critnib *c, word key, void *value, int update) {
361341

362342
if (!n) {
363343
n = prev;
364-
store(&n->child[slice_index(key, n->shift)], kn);
344+
utils_atomic_store_release_ptr(
345+
(void **)&n->child[slice_index(key, n->shift)], kn);
365346

366347
utils_mutex_unlock(&c->mutex);
367348

@@ -406,7 +387,7 @@ int critnib_insert(struct critnib *c, word key, void *value, int update) {
406387
m->child[slice_index(path, sh)] = n;
407388
m->shift = sh;
408389
m->path = key & path_mask(sh);
409-
store(parent, m);
390+
utils_atomic_store_release_ptr((void **)parent, m);
410391

411392
utils_mutex_unlock(&c->mutex);
412393

@@ -427,7 +408,8 @@ void *critnib_remove(struct critnib *c, word key) {
427408
goto not_found;
428409
}
429410

430-
word del = (utils_atomic_increment(&c->remove_count) - 1) % DELETED_LIFE;
411+
word del =
412+
(utils_atomic_increment_u64(&c->remove_count) - 1) % DELETED_LIFE;
431413
free_node(c, c->pending_del_nodes[del]);
432414
free_leaf(c, c->pending_del_leaves[del]);
433415
c->pending_del_nodes[del] = NULL;
@@ -436,7 +418,7 @@ void *critnib_remove(struct critnib *c, word key) {
436418
if (is_leaf(n)) {
437419
k = to_leaf(n);
438420
if (k->key == key) {
439-
store(&c->root, NULL);
421+
utils_atomic_store_release_ptr((void **)&c->root, NULL);
440422
goto del_leaf;
441423
}
442424

@@ -466,7 +448,8 @@ void *critnib_remove(struct critnib *c, word key) {
466448
goto not_found;
467449
}
468450

469-
store(&n->child[slice_index(key, n->shift)], NULL);
451+
utils_atomic_store_release_ptr(
452+
(void **)&n->child[slice_index(key, n->shift)], NULL);
470453

471454
/* Remove the node if there's only one remaining child. */
472455
int ochild = -1;
@@ -482,7 +465,7 @@ void *critnib_remove(struct critnib *c, word key) {
482465

483466
ASSERTne(ochild, -1);
484467

485-
store(n_parent, n->child[ochild]);
468+
utils_atomic_store_release_ptr((void **)n_parent, n->child[ochild]);
486469
c->pending_del_nodes[del] = n;
487470

488471
del_leaf:
@@ -511,22 +494,23 @@ void *critnib_get(struct critnib *c, word key) {
511494
do {
512495
struct critnib_node *n;
513496

514-
load64(&c->remove_count, &wrs1);
515-
load(&c->root, &n);
497+
utils_atomic_load_acquire_u64(&c->remove_count, &wrs1);
498+
utils_atomic_load_acquire_ptr((void **)&c->root, (void **)&n);
516499

517500
/*
518501
* critbit algorithm: dive into the tree, looking at nothing but
519502
* each node's critical bit^H^H^Hnibble. This means we risk
520503
* going wrong way if our path is missing, but that's ok...
521504
*/
522505
while (n && !is_leaf(n)) {
523-
load(&n->child[slice_index(key, n->shift)], &n);
506+
utils_atomic_load_acquire_ptr(
507+
(void **)&n->child[slice_index(key, n->shift)], (void **)&n);
524508
}
525509

526510
/* ... as we check it at the end. */
527511
struct critnib_leaf *k = to_leaf(n);
528512
res = (n && k->key == key) ? k->value : NULL;
529-
load64(&c->remove_count, &wrs2);
513+
utils_atomic_load_acquire_u64(&c->remove_count, &wrs2);
530514
} while (wrs1 + DELETED_LIFE <= wrs2);
531515

532516
return res;
@@ -597,7 +581,7 @@ static struct critnib_leaf *find_le(struct critnib_node *__restrict n,
597581
/* recursive call: follow the path */
598582
{
599583
struct critnib_node *m;
600-
load(&n->child[nib], &m);
584+
utils_atomic_load_acquire_ptr((void **)&n->child[nib], (void **)&m);
601585
struct critnib_leaf *k = find_le(m, key);
602586
if (k) {
603587
return k;
@@ -611,7 +595,7 @@ static struct critnib_leaf *find_le(struct critnib_node *__restrict n,
611595
*/
612596
for (; nib > 0; nib--) {
613597
struct critnib_node *m;
614-
load(&n->child[nib - 1], &m);
598+
utils_atomic_load_acquire_ptr((void **)&n->child[nib - 1], (void **)&m);
615599
if (m) {
616600
n = m;
617601
if (is_leaf(n)) {
@@ -635,12 +619,12 @@ void *critnib_find_le(struct critnib *c, word key) {
635619
void *res;
636620

637621
do {
638-
load64(&c->remove_count, &wrs1);
622+
utils_atomic_load_acquire_u64(&c->remove_count, &wrs1);
639623
struct critnib_node *n; /* avoid a subtle TOCTOU */
640-
load(&c->root, &n);
624+
utils_atomic_load_acquire_ptr((void **)&c->root, (void **)&n);
641625
struct critnib_leaf *k = n ? find_le(n, key) : NULL;
642626
res = k ? k->value : NULL;
643-
load64(&c->remove_count, &wrs2);
627+
utils_atomic_load_acquire_u64(&c->remove_count, &wrs2);
644628
} while (wrs1 + DELETED_LIFE <= wrs2);
645629

646630
return res;
@@ -694,7 +678,7 @@ static struct critnib_leaf *find_ge(struct critnib_node *__restrict n,
694678
unsigned nib = slice_index(key, n->shift);
695679
{
696680
struct critnib_node *m;
697-
load(&n->child[nib], &m);
681+
utils_atomic_load_acquire_ptr((void **)&n->child[nib], (void **)&m);
698682
struct critnib_leaf *k = find_ge(m, key);
699683
if (k) {
700684
return k;
@@ -703,7 +687,7 @@ static struct critnib_leaf *find_ge(struct critnib_node *__restrict n,
703687

704688
for (; nib < NIB; nib++) {
705689
struct critnib_node *m;
706-
load(&n->child[nib + 1], &m);
690+
utils_atomic_load_acquire_ptr((void **)&n->child[nib + 1], (void **)&m);
707691
if (m) {
708692
n = m;
709693
if (is_leaf(n)) {
@@ -741,17 +725,19 @@ int critnib_find(struct critnib *c, uintptr_t key, enum find_dir_t dir,
741725
}
742726

743727
do {
744-
load64(&c->remove_count, &wrs1);
728+
utils_atomic_load_acquire_u64(&c->remove_count, &wrs1);
745729
struct critnib_node *n;
746-
load(&c->root, &n);
730+
utils_atomic_load_acquire_ptr((void **)&c->root, (void **)&n);
747731

748732
if (dir < 0) {
749733
k = find_le(n, key);
750734
} else if (dir > 0) {
751735
k = find_ge(n, key);
752736
} else {
753737
while (n && !is_leaf(n)) {
754-
load(&n->child[slice_index(key, n->shift)], &n);
738+
utils_atomic_load_acquire_ptr(
739+
(void **)&n->child[slice_index(key, n->shift)],
740+
(void **)&n);
755741
}
756742

757743
struct critnib_leaf *kk = to_leaf(n);
@@ -761,7 +747,7 @@ int critnib_find(struct critnib *c, uintptr_t key, enum find_dir_t dir,
761747
_rkey = k->key;
762748
_rvalue = k->value;
763749
}
764-
load64(&c->remove_count, &wrs2);
750+
utils_atomic_load_acquire_u64(&c->remove_count, &wrs2);
765751
} while (wrs1 + DELETED_LIFE <= wrs2);
766752

767753
if (k) {

src/ipc_cache.c

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,10 +28,10 @@ typedef struct ipc_opened_cache_entry_t *hash_map_t;
2828
typedef struct ipc_opened_cache_entry_t *lru_list_t;
2929

3030
typedef struct ipc_opened_cache_entry_t {
31+
uint64_t ref_count;
3132
UT_hash_handle hh;
3233
struct ipc_opened_cache_entry_t *next, *prev;
3334
ipc_opened_cache_key_t key;
34-
uint64_t ref_count;
3535
uint64_t handle_id;
3636
hash_map_t
3737
*hash_table; // pointer to the hash table to which the entry belongs
@@ -70,6 +70,7 @@ umf_result_t umfIpcCacheGlobalInit(void) {
7070
goto err_cache_global_free;
7171
}
7272

73+
// TODO: create aligned-allocator
7374
cache_global->cache_allocator =
7475
umf_ba_create(sizeof(ipc_opened_cache_entry_t));
7576
if (!cache_global->cache_allocator) {
@@ -232,7 +233,7 @@ umf_result_t umfIpcOpenedCacheGet(ipc_opened_cache_handle_t cache,
232233

233234
exit:
234235
if (ret == UMF_RESULT_SUCCESS) {
235-
utils_atomic_increment(&entry->ref_count);
236+
utils_atomic_increment_u64(&entry->ref_count);
236237
*retEntry = &entry->value;
237238
}
238239

src/libumf.c

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -24,10 +24,10 @@
2424

2525
umf_memory_tracker_handle_t TRACKER = NULL;
2626

27-
static unsigned long long umfRefCount = 0;
27+
CACHE_ALIGNED static uint64_t umfRefCount = 0;
2828

2929
int umfInit(void) {
30-
if (utils_fetch_and_add64(&umfRefCount, 1) == 0) {
30+
if (utils_fetch_and_add_u64(&umfRefCount, 1) == 0) {
3131
utils_log_init();
3232
TRACKER = umfMemoryTrackerCreate();
3333
if (!TRACKER) {
@@ -54,7 +54,7 @@ int umfInit(void) {
5454
}
5555

5656
void umfTearDown(void) {
57-
if (utils_fetch_and_add64(&umfRefCount, -1) == 1) {
57+
if (utils_fetch_and_sub_u64(&umfRefCount, 1) == 1) {
5858
#if !defined(_WIN32) && !defined(UMF_NO_HWLOC)
5959
umfMemspaceHostAllDestroy();
6060
umfMemspaceHighestCapacityDestroy();

src/pool/pool_disjoint.c

Lines changed: 25 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -20,6 +20,7 @@
2020
#include "provider/provider_tracking.h"
2121
#include "uthash/utlist.h"
2222
#include "utils_common.h"
23+
#include "utils_concurrency.h"
2324
#include "utils_log.h"
2425
#include "utils_math.h"
2526

@@ -34,7 +35,6 @@
3435
// Forward declarations
3536
static void bucket_update_stats(bucket_t *bucket, int in_use, int in_pool);
3637
static bool bucket_can_pool(bucket_t *bucket);
37-
static void bucket_decrement_pool(bucket_t *bucket);
3838
static slab_list_item_t *bucket_get_avail_slab(bucket_t *bucket,
3939
bool *from_pool);
4040

@@ -316,6 +316,7 @@ static void bucket_free_chunk(bucket_t *bucket, void *ptr, slab_t *slab,
316316
assert(slab_it->val != NULL);
317317
pool_unregister_slab(bucket->pool, slab_it->val);
318318
DL_DELETE(bucket->available_slabs, slab_it);
319+
assert(bucket->available_slabs_num > 0);
319320
bucket->available_slabs_num--;
320321
destroy_slab(slab_it->val);
321322
}
@@ -381,10 +382,17 @@ static slab_list_item_t *bucket_get_avail_slab(bucket_t *bucket,
381382
// Allocation from existing slab is treated as from pool for statistics.
382383
*from_pool = true;
383384
if (slab->num_chunks_allocated == 0) {
385+
assert(bucket->chunked_slabs_in_pool > 0);
386+
assert(bucket->shared_limits->total_size >=
387+
bucket_slab_alloc_size(bucket));
384388
// If this was an empty slab, it was in the pool.
385389
// Now it is no longer in the pool, so update count.
386390
--bucket->chunked_slabs_in_pool;
387-
bucket_decrement_pool(bucket);
391+
CACHE_ALIGNED size_t size_to_add =
392+
-(long long)bucket_slab_alloc_size(bucket);
393+
utils_fetch_and_add_u64(&bucket->shared_limits->total_size,
394+
size_to_add);
395+
bucket_update_stats(bucket, 1, -1);
388396
}
389397
}
390398

@@ -420,36 +428,26 @@ static void bucket_update_stats(bucket_t *bucket, int in_use, int in_pool) {
420428
in_pool * bucket_slab_alloc_size(bucket);
421429
}
422430

423-
static void bucket_decrement_pool(bucket_t *bucket) {
424-
bucket_update_stats(bucket, 1, -1);
425-
utils_fetch_and_add64(&bucket->shared_limits->total_size,
426-
-(long long)bucket_slab_alloc_size(bucket));
427-
}
428-
429431
static bool bucket_can_pool(bucket_t *bucket) {
430432
size_t new_free_slabs_in_bucket;
431433

432434
new_free_slabs_in_bucket = bucket->chunked_slabs_in_pool + 1;
433435

434436
// we keep at most params.capacity slabs in the pool
435437
if (bucket_max_pooled_slabs(bucket) >= new_free_slabs_in_bucket) {
436-
size_t pool_size = 0;
437-
utils_atomic_load_acquire(&bucket->shared_limits->total_size,
438-
&pool_size);
439-
while (true) {
440-
size_t new_pool_size = pool_size + bucket_slab_alloc_size(bucket);
441-
442-
if (bucket->shared_limits->max_size < new_pool_size) {
443-
break;
444-
}
445-
446-
if (utils_compare_exchange(&bucket->shared_limits->total_size,
447-
&pool_size, &new_pool_size)) {
448-
++bucket->chunked_slabs_in_pool;
449-
450-
bucket_update_stats(bucket, -1, 1);
451-
return true;
452-
}
438+
439+
CACHE_ALIGNED size_t size_to_add = bucket_slab_alloc_size(bucket);
440+
size_t previous_size = utils_fetch_and_add_u64(
441+
&bucket->shared_limits->total_size, size_to_add);
442+
443+
if (previous_size + size_to_add <= bucket->shared_limits->max_size) {
444+
445+
++bucket->chunked_slabs_in_pool;
446+
bucket_update_stats(bucket, -1, 1);
447+
return true;
448+
} else {
449+
utils_fetch_and_sub_u64(&bucket->shared_limits->total_size,
450+
size_to_add);
453451
}
454452
}
455453

@@ -895,7 +893,8 @@ umf_memory_pool_ops_t *umfDisjointPoolOps(void) {
895893

896894
umf_disjoint_pool_shared_limits_t *
897895
umfDisjointPoolSharedLimitsCreate(size_t max_size) {
898-
umf_disjoint_pool_shared_limits_t *ptr = umf_ba_global_alloc(sizeof(*ptr));
896+
umf_disjoint_pool_shared_limits_t *ptr =
897+
umf_ba_global_aligned_alloc(sizeof(*ptr), 64);
899898
if (ptr == NULL) {
900899
LOG_ERR("cannot allocate memory for disjoint pool shared limits");
901900
return NULL;

src/pool/pool_disjoint_internal.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -101,8 +101,8 @@ typedef struct slab_t {
101101
} slab_t;
102102

103103
typedef struct umf_disjoint_pool_shared_limits_t {
104+
CACHE_ALIGNED size_t total_size; // requires atomic access
104105
size_t max_size;
105-
size_t total_size; // requires atomic access
106106
} umf_disjoint_pool_shared_limits_t;
107107

108108
typedef struct umf_disjoint_pool_params_t {

0 commit comments

Comments
 (0)