int64_t blocking_timeout_ms;
if (config->alloc == RING_BUFFER_ALLOC_PER_CPU)
- nr_streams = num_possible_cpus();
+ nr_streams = get_possible_cpus_array_len();
else
nr_streams = 1;
return NULL;
/* Allocate table for channel + per-cpu buffers */
- handle->table = shm_object_table_create(1 + num_possible_cpus());
+ handle->table = shm_object_table_create(1 + get_possible_cpus_array_len());
if (!handle->table)
goto error_table_alloc;
return NULL;
/* Allocate table for channel + per-cpu buffers */
- handle->table = shm_object_table_create(1 + num_possible_cpus());
+ handle->table = shm_object_table_create(1 + get_possible_cpus_array_len());
if (!handle->table)
goto error_table_alloc;
/* Add channel object */
if (config->alloc == RING_BUFFER_ALLOC_GLOBAL) {
cpu = 0;
} else {
- if (cpu >= num_possible_cpus())
+ if (cpu >= get_possible_cpus_array_len())
return NULL;
}
ref = &chan->backend.buf[cpu].shmp._ref;
if (config->alloc == RING_BUFFER_ALLOC_GLOBAL) {
cpu = 0;
} else {
- if (cpu >= num_possible_cpus())
+ if (cpu >= get_possible_cpus_array_len())
return -EINVAL;
}
ref = &chan->backend.buf[cpu].shmp._ref;
if (config->alloc == RING_BUFFER_ALLOC_GLOBAL) {
cpu = 0;
} else {
- if (cpu >= num_possible_cpus())
+ if (cpu >= get_possible_cpus_array_len())
return -EINVAL;
}
ref = &chan->backend.buf[cpu].shmp._ref;