trigger_status = lttng_trigger_get_owner_uid(trigger,
trigger_owner_uid);
- assert(trigger_status == LTTNG_TRIGGER_STATUS_OK);
+ LTTNG_ASSERT(trigger_status == LTTNG_TRIGGER_STATUS_OK);
}
static inline
{
enum event_notifier_error_accounting_status status;
- assert(state);
+ LTTNG_ASSERT(state);
state->number_indices = index_count;
static
void fini_error_accounting_state(struct error_accounting_state *state)
{
- assert(state);
+ LTTNG_ASSERT(state);
/*
* Will assert if some error counter indices were not released (an
lttng_ust_ctl_destroy_counter(daemon_counter);
error_create_daemon_counter:
error_shm_alloc:
- /* Error occured before per-cpu SHMs were handed-off to ustctl. */
+ /* Error occurred before per-cpu SHMs were handed-off to ustctl. */
if (cpu_counter_fds) {
for (i = 0; i < entry->nr_counter_cpu_fds; i++) {
if (cpu_counter_fds[i] < 0) {
uint64_t local_error_counter_index;
enum event_notifier_error_accounting_status status;
- assert(state);
+ LTTNG_ASSERT(state);
/* Allocate a new index for that counter. */
index_alloc_status = lttng_index_allocator_alloc(state->index_allocator,
counter_aggregate.index.number_dimensions = 1;
counter_aggregate.index.dimension_indexes[0] = error_counter_index;
- assert(kernel_error_accounting_entry.error_counter_fd);
+ LTTNG_ASSERT(kernel_error_accounting_entry.error_counter_fd);
ret = kernctl_counter_get_aggregate_value(
kernel_error_accounting_entry.error_counter_fd,
}
/* Error count can't be negative. */
- assert(counter_aggregate.value.value >= 0);
+ LTTNG_ASSERT(counter_aggregate.value.value >= 0);
*count = (uint64_t) counter_aggregate.value.value;
status = EVENT_NOTIFIER_ERROR_ACCOUNTING_STATUS_OK;
}
del_ret = lttng_ht_del(state->indices_ht, &iter);
- assert(!del_ret);
+ LTTNG_ASSERT(!del_ret);
call_rcu(&index_entry->rcu_head, free_index_ht_entry);
}