#include "context-internal.h"
#include "lttng-bytecode.h"
#include "ust-events-internal.h"
+#include "ust-helper.h"
static const char *opnames[] = {
[ BYTECODE_OP_UNKNOWN ] = "UNKNOWN",
}
static
-int apply_field_reloc(const struct lttng_event_desc *event_desc,
+int apply_field_reloc(const struct lttng_ust_event_desc *event_desc,
struct bytecode_runtime *runtime,
uint32_t runtime_len,
uint32_t reloc_offset,
const char *field_name,
enum bytecode_op bytecode_op)
{
- const struct lttng_event_field *fields, *field = NULL;
+ const struct lttng_ust_event_field **fields, *field = NULL;
unsigned int nr_fields, i;
struct load_op *op;
uint32_t field_offset = 0;
return -EINVAL;
nr_fields = event_desc->nr_fields;
for (i = 0; i < nr_fields; i++) {
- if (fields[i].u.ext.nofilter) {
+ if (fields[i]->nofilter) {
continue;
}
- if (!strcmp(fields[i].name, field_name)) {
- field = &fields[i];
+ if (!strcmp(fields[i]->name, field_name)) {
+ field = fields[i];
break;
}
/* compute field offset */
- switch (fields[i].type.atype) {
+ switch (fields[i]->type.atype) {
case atype_integer:
case atype_enum_nestable:
field_offset += sizeof(int64_t);
return -EINVAL;
/* Check if field offset is too large for 16-bit offset */
- if (field_offset > FILTER_BYTECODE_MAX_LEN - 1)
+ if (field_offset > LTTNG_UST_ABI_FILTER_BYTECODE_MAX_LEN - 1)
return -EINVAL;
/* set type */
}
}
/* Check if idx is too large for 16-bit offset */
- if (idx > FILTER_BYTECODE_MAX_LEN - 1)
+ if (idx > LTTNG_UST_ABI_FILTER_BYTECODE_MAX_LEN - 1)
return -EINVAL;
/* Get context return type */
}
static
-int apply_reloc(const struct lttng_event_desc *event_desc,
+int apply_reloc(const struct lttng_ust_event_desc *event_desc,
struct bytecode_runtime *runtime,
uint32_t runtime_len,
uint32_t reloc_offset,
int bytecode_is_linked(struct lttng_ust_bytecode_node *bytecode,
struct cds_list_head *bytecode_runtime_head)
{
- struct lttng_bytecode_runtime *bc_runtime;
+ struct lttng_ust_bytecode_runtime *bc_runtime;
cds_list_for_each_entry(bc_runtime, bytecode_runtime_head, node) {
if (bc_runtime->priv->bc == bytecode)
* bytecode runtime.
*/
static
-int link_bytecode(const struct lttng_event_desc *event_desc,
+int link_bytecode(const struct lttng_ust_event_desc *event_desc,
struct lttng_ctx **ctx,
struct lttng_ust_bytecode_node *bytecode,
struct cds_list_head *bytecode_runtime_head,
goto alloc_error;
}
runtime->p.priv = runtime_priv;
+ runtime->p.struct_size = sizeof(struct lttng_ust_bytecode_runtime);
runtime_priv->pub = runtime;
runtime_priv->bc = bytecode;
runtime_priv->pctx = ctx;
return ret;
}
-void lttng_bytecode_filter_sync_state(struct lttng_bytecode_runtime *runtime)
+void lttng_bytecode_filter_sync_state(struct lttng_ust_bytecode_runtime *runtime)
{
struct lttng_ust_bytecode_node *bc = runtime->priv->bc;
runtime->interpreter_funcs.filter = lttng_bytecode_filter_interpret;
}
-void lttng_bytecode_capture_sync_state(struct lttng_bytecode_runtime *runtime)
+void lttng_bytecode_capture_sync_state(struct lttng_ust_bytecode_runtime *runtime)
{
struct lttng_ust_bytecode_node *bc = runtime->priv->bc;
* This function is called after we confirmed that name enabler and the
* instance are name matching (or glob pattern matching).
*/
-void lttng_enabler_link_bytecode(const struct lttng_event_desc *event_desc,
+void lttng_enabler_link_bytecode(const struct lttng_ust_event_desc *event_desc,
struct lttng_ctx **ctx,
struct cds_list_head *instance_bytecode_head,
struct cds_list_head *enabler_bytecode_head)
{
struct lttng_ust_bytecode_node *enabler_bc;
- struct lttng_bytecode_runtime *runtime;
+ struct lttng_ust_bytecode_runtime *runtime;
assert(event_desc);
}
}
-void lttng_free_event_filter_runtime(struct lttng_event *event)
+void lttng_free_event_filter_runtime(struct lttng_ust_event_common *event)
{
- free_filter_runtime(&event->parent->filter_bytecode_runtime_head);
-}
-
-void lttng_free_event_notifier_filter_runtime(
- struct lttng_event_notifier *event_notifier)
-{
- free_filter_runtime(&event_notifier->filter_bytecode_runtime_head);
+ free_filter_runtime(&event->filter_bytecode_runtime_head);
}