2 * SPDX-License-Identifier: MIT
4 * Copyright (C) 2010-2016 Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
6 * LTTng UST bytecode interpreter.
13 #include <lttng/urcu/pointer.h>
14 #include <lttng/ust-endian.h>
15 #include <lttng/ust-events.h>
16 #include "ust-events-internal.h"
18 #include "lttng-bytecode.h"
19 #include "string-utils.h"
24 * -2: unknown escape char.
29 int parse_char(const char **p
)
49 * Returns SIZE_MAX if the string is null-terminated, or the number of
53 size_t get_str_or_seq_len(const struct estack_entry
*entry
)
55 return entry
->u
.s
.seq_len
;
59 int stack_star_glob_match(struct estack
*stack
, int top
, const char *cmp_type
)
62 const char *candidate
;
66 /* Find out which side is the pattern vs. the candidate. */
67 if (estack_ax(stack
, top
)->u
.s
.literal_type
== ESTACK_STRING_LITERAL_TYPE_STAR_GLOB
) {
68 pattern
= estack_ax(stack
, top
)->u
.s
.str
;
69 pattern_len
= get_str_or_seq_len(estack_ax(stack
, top
));
70 candidate
= estack_bx(stack
, top
)->u
.s
.str
;
71 candidate_len
= get_str_or_seq_len(estack_bx(stack
, top
));
73 pattern
= estack_bx(stack
, top
)->u
.s
.str
;
74 pattern_len
= get_str_or_seq_len(estack_bx(stack
, top
));
75 candidate
= estack_ax(stack
, top
)->u
.s
.str
;
76 candidate_len
= get_str_or_seq_len(estack_ax(stack
, top
));
79 /* Perform the match. Returns 0 when the result is true. */
80 return !strutils_star_glob_match(pattern
, pattern_len
, candidate
,
85 int stack_strcmp(struct estack
*stack
, int top
, const char *cmp_type
)
87 const char *p
= estack_bx(stack
, top
)->u
.s
.str
, *q
= estack_ax(stack
, top
)->u
.s
.str
;
94 if (unlikely(p
- estack_bx(stack
, top
)->u
.s
.str
>= estack_bx(stack
, top
)->u
.s
.seq_len
|| *p
== '\0')) {
95 if (q
- estack_ax(stack
, top
)->u
.s
.str
>= estack_ax(stack
, top
)->u
.s
.seq_len
|| *q
== '\0') {
98 if (estack_ax(stack
, top
)->u
.s
.literal_type
==
99 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
100 ret
= parse_char(&q
);
107 if (unlikely(q
- estack_ax(stack
, top
)->u
.s
.str
>= estack_ax(stack
, top
)->u
.s
.seq_len
|| *q
== '\0')) {
108 if (estack_bx(stack
, top
)->u
.s
.literal_type
==
109 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
110 ret
= parse_char(&p
);
116 if (estack_bx(stack
, top
)->u
.s
.literal_type
==
117 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
118 ret
= parse_char(&p
);
121 } else if (ret
== -2) {
124 /* else compare both char */
126 if (estack_ax(stack
, top
)->u
.s
.literal_type
==
127 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
128 ret
= parse_char(&q
);
131 } else if (ret
== -2) {
151 int lttng_bytecode_interpret_error(struct lttng_ust_bytecode_runtime
*bytecode_runtime
,
152 const char *stack_data
,
155 return LTTNG_UST_BYTECODE_INTERPRETER_ERROR
;
158 #ifdef INTERPRETER_USE_SWITCH
161 * Fallback for compilers that do not support taking address of labels.
165 start_pc = &bytecode->data[0]; \
166 for (pc = next_pc = start_pc; pc - start_pc < bytecode->len; \
168 dbg_printf("Executing op %s (%u)\n", \
169 lttng_bytecode_print_op((unsigned int) *(bytecode_opcode_t *) pc), \
170 (unsigned int) *(bytecode_opcode_t *) pc); \
171 switch (*(bytecode_opcode_t *) pc) {
173 #define OP(name) jump_target_##name: __attribute__((unused)); \
181 #define JUMP_TO(name) \
182 goto jump_target_##name
187 * Dispatch-table based interpreter.
191 start_pc = &bytecode->code[0]; \
192 pc = next_pc = start_pc; \
193 if (unlikely(pc - start_pc >= bytecode->len)) \
195 goto *dispatch[*(bytecode_opcode_t *) pc];
202 goto *dispatch[*(bytecode_opcode_t *) pc];
206 #define JUMP_TO(name) \
211 #define IS_INTEGER_REGISTER(reg_type) \
212 (reg_type == REG_U64 || reg_type == REG_S64)
214 static int context_get_index(struct lttng_ust_ctx
*ctx
,
215 struct load_ptr
*ptr
,
219 struct lttng_ust_ctx_field
*ctx_field
;
220 struct lttng_ust_event_field
*field
;
221 struct lttng_ust_ctx_value v
;
223 ctx_field
= ctx
->fields
[idx
];
224 field
= ctx_field
->event_field
;
225 ptr
->type
= LOAD_OBJECT
;
228 switch (field
->type
->type
) {
229 case lttng_ust_type_integer
:
230 ctx_field
->get_value(ctx_field
, &v
);
231 if (lttng_ust_get_type_integer(field
->type
)->signedness
) {
232 ptr
->object_type
= OBJECT_TYPE_S64
;
233 ptr
->u
.s64
= v
.u
.s64
;
234 ptr
->ptr
= &ptr
->u
.s64
;
236 ptr
->object_type
= OBJECT_TYPE_U64
;
237 ptr
->u
.u64
= v
.u
.s64
; /* Cast. */
238 ptr
->ptr
= &ptr
->u
.u64
;
241 case lttng_ust_type_enum
:
243 const struct lttng_ust_type_integer
*itype
;
245 itype
= lttng_ust_get_type_integer(lttng_ust_get_type_enum(field
->type
)->container_type
);
246 ctx_field
->get_value(ctx_field
, &v
);
247 if (itype
->signedness
) {
248 ptr
->object_type
= OBJECT_TYPE_SIGNED_ENUM
;
249 ptr
->u
.s64
= v
.u
.s64
;
250 ptr
->ptr
= &ptr
->u
.s64
;
252 ptr
->object_type
= OBJECT_TYPE_UNSIGNED_ENUM
;
253 ptr
->u
.u64
= v
.u
.s64
; /* Cast. */
254 ptr
->ptr
= &ptr
->u
.u64
;
258 case lttng_ust_type_array
:
259 if (lttng_ust_get_type_array(field
->type
)->elem_type
->type
!= lttng_ust_type_integer
) {
260 ERR("Array nesting only supports integer types.");
263 if (lttng_ust_get_type_array(field
->type
)->encoding
== lttng_ust_string_encoding_none
) {
264 ERR("Only string arrays are supported for contexts.");
267 ptr
->object_type
= OBJECT_TYPE_STRING
;
268 ctx_field
->get_value(ctx_field
, &v
);
271 case lttng_ust_type_sequence
:
272 if (lttng_ust_get_type_sequence(field
->type
)->elem_type
->type
!= lttng_ust_type_integer
) {
273 ERR("Sequence nesting only supports integer types.");
276 if (lttng_ust_get_type_sequence(field
->type
)->encoding
== lttng_ust_string_encoding_none
) {
277 ERR("Only string sequences are supported for contexts.");
280 ptr
->object_type
= OBJECT_TYPE_STRING
;
281 ctx_field
->get_value(ctx_field
, &v
);
284 case lttng_ust_type_string
:
285 ptr
->object_type
= OBJECT_TYPE_STRING
;
286 ctx_field
->get_value(ctx_field
, &v
);
289 case lttng_ust_type_float
:
290 ptr
->object_type
= OBJECT_TYPE_DOUBLE
;
291 ctx_field
->get_value(ctx_field
, &v
);
293 ptr
->ptr
= &ptr
->u
.d
;
295 case lttng_ust_type_dynamic
:
296 ctx_field
->get_value(ctx_field
, &v
);
298 case LTTNG_UST_DYNAMIC_TYPE_NONE
:
300 case LTTNG_UST_DYNAMIC_TYPE_U8
:
301 case LTTNG_UST_DYNAMIC_TYPE_U16
:
302 case LTTNG_UST_DYNAMIC_TYPE_U32
:
303 case LTTNG_UST_DYNAMIC_TYPE_U64
:
304 ptr
->object_type
= OBJECT_TYPE_U64
;
305 ptr
->u
.u64
= v
.u
.u64
;
306 ptr
->ptr
= &ptr
->u
.u64
;
307 dbg_printf("context get index dynamic u64 %" PRIi64
"\n", ptr
->u
.u64
);
309 case LTTNG_UST_DYNAMIC_TYPE_S8
:
310 case LTTNG_UST_DYNAMIC_TYPE_S16
:
311 case LTTNG_UST_DYNAMIC_TYPE_S32
:
312 case LTTNG_UST_DYNAMIC_TYPE_S64
:
313 ptr
->object_type
= OBJECT_TYPE_S64
;
314 ptr
->u
.s64
= v
.u
.s64
;
315 ptr
->ptr
= &ptr
->u
.s64
;
316 dbg_printf("context get index dynamic s64 %" PRIi64
"\n", ptr
->u
.s64
);
318 case LTTNG_UST_DYNAMIC_TYPE_FLOAT
:
319 case LTTNG_UST_DYNAMIC_TYPE_DOUBLE
:
320 ptr
->object_type
= OBJECT_TYPE_DOUBLE
;
322 ptr
->ptr
= &ptr
->u
.d
;
323 dbg_printf("context get index dynamic double %g\n", ptr
->u
.d
);
325 case LTTNG_UST_DYNAMIC_TYPE_STRING
:
326 ptr
->object_type
= OBJECT_TYPE_STRING
;
328 dbg_printf("context get index dynamic string %s\n", (const char *) ptr
->ptr
);
331 dbg_printf("Interpreter warning: unknown dynamic type (%d).\n", (int) v
.sel
);
336 ERR("Unknown type: %d", (int) field
->type
->type
);
342 static int dynamic_get_index(struct lttng_ust_ctx
*ctx
,
343 struct bytecode_runtime
*runtime
,
344 uint64_t index
, struct estack_entry
*stack_top
)
347 const struct bytecode_get_index_data
*gid
;
349 gid
= (const struct bytecode_get_index_data
*) &runtime
->data
[index
];
350 switch (stack_top
->u
.ptr
.type
) {
352 switch (stack_top
->u
.ptr
.object_type
) {
353 case OBJECT_TYPE_ARRAY
:
357 assert(gid
->offset
< gid
->array_len
);
358 /* Skip count (unsigned long) */
359 ptr
= *(const char **) (stack_top
->u
.ptr
.ptr
+ sizeof(unsigned long));
360 ptr
= ptr
+ gid
->offset
;
361 stack_top
->u
.ptr
.ptr
= ptr
;
362 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
363 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
364 assert(stack_top
->u
.ptr
.field
->type
->type
== lttng_ust_type_array
);
365 stack_top
->u
.ptr
.field
= NULL
;
368 case OBJECT_TYPE_SEQUENCE
:
373 ptr
= *(const char **) (stack_top
->u
.ptr
.ptr
+ sizeof(unsigned long));
374 ptr_seq_len
= *(unsigned long *) stack_top
->u
.ptr
.ptr
;
375 if (gid
->offset
>= gid
->elem
.len
* ptr_seq_len
) {
379 ptr
= ptr
+ gid
->offset
;
380 stack_top
->u
.ptr
.ptr
= ptr
;
381 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
382 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
383 assert(stack_top
->u
.ptr
.field
->type
->type
== lttng_ust_type_sequence
);
384 stack_top
->u
.ptr
.field
= NULL
;
387 case OBJECT_TYPE_STRUCT
:
388 ERR("Nested structures are not supported yet.");
391 case OBJECT_TYPE_VARIANT
:
393 ERR("Unexpected get index type %d",
394 (int) stack_top
->u
.ptr
.object_type
);
399 case LOAD_ROOT_CONTEXT
:
400 case LOAD_ROOT_APP_CONTEXT
: /* Fall-through */
402 ret
= context_get_index(ctx
,
410 case LOAD_ROOT_PAYLOAD
:
411 stack_top
->u
.ptr
.ptr
+= gid
->offset
;
412 if (gid
->elem
.type
== OBJECT_TYPE_STRING
)
413 stack_top
->u
.ptr
.ptr
= *(const char * const *) stack_top
->u
.ptr
.ptr
;
414 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
415 stack_top
->u
.ptr
.type
= LOAD_OBJECT
;
416 stack_top
->u
.ptr
.field
= gid
->field
;
417 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
421 stack_top
->type
= REG_PTR
;
429 static int dynamic_load_field(struct estack_entry
*stack_top
)
433 switch (stack_top
->u
.ptr
.type
) {
436 case LOAD_ROOT_CONTEXT
:
437 case LOAD_ROOT_APP_CONTEXT
:
438 case LOAD_ROOT_PAYLOAD
:
440 dbg_printf("Interpreter warning: cannot load root, missing field name.\n");
444 switch (stack_top
->u
.ptr
.object_type
) {
446 dbg_printf("op load field s8\n");
447 stack_top
->u
.v
= *(int8_t *) stack_top
->u
.ptr
.ptr
;
448 stack_top
->type
= REG_S64
;
450 case OBJECT_TYPE_S16
:
454 dbg_printf("op load field s16\n");
455 tmp
= *(int16_t *) stack_top
->u
.ptr
.ptr
;
456 if (stack_top
->u
.ptr
.rev_bo
)
458 stack_top
->u
.v
= tmp
;
459 stack_top
->type
= REG_S64
;
462 case OBJECT_TYPE_S32
:
466 dbg_printf("op load field s32\n");
467 tmp
= *(int32_t *) stack_top
->u
.ptr
.ptr
;
468 if (stack_top
->u
.ptr
.rev_bo
)
470 stack_top
->u
.v
= tmp
;
471 stack_top
->type
= REG_S64
;
474 case OBJECT_TYPE_S64
:
478 dbg_printf("op load field s64\n");
479 tmp
= *(int64_t *) stack_top
->u
.ptr
.ptr
;
480 if (stack_top
->u
.ptr
.rev_bo
)
482 stack_top
->u
.v
= tmp
;
483 stack_top
->type
= REG_S64
;
486 case OBJECT_TYPE_SIGNED_ENUM
:
490 dbg_printf("op load field signed enumeration\n");
491 tmp
= *(int64_t *) stack_top
->u
.ptr
.ptr
;
492 if (stack_top
->u
.ptr
.rev_bo
)
494 stack_top
->u
.v
= tmp
;
495 stack_top
->type
= REG_S64
;
499 dbg_printf("op load field u8\n");
500 stack_top
->u
.v
= *(uint8_t *) stack_top
->u
.ptr
.ptr
;
501 stack_top
->type
= REG_U64
;
503 case OBJECT_TYPE_U16
:
507 dbg_printf("op load field u16\n");
508 tmp
= *(uint16_t *) stack_top
->u
.ptr
.ptr
;
509 if (stack_top
->u
.ptr
.rev_bo
)
511 stack_top
->u
.v
= tmp
;
512 stack_top
->type
= REG_U64
;
515 case OBJECT_TYPE_U32
:
519 dbg_printf("op load field u32\n");
520 tmp
= *(uint32_t *) stack_top
->u
.ptr
.ptr
;
521 if (stack_top
->u
.ptr
.rev_bo
)
523 stack_top
->u
.v
= tmp
;
524 stack_top
->type
= REG_U64
;
527 case OBJECT_TYPE_U64
:
531 dbg_printf("op load field u64\n");
532 tmp
= *(uint64_t *) stack_top
->u
.ptr
.ptr
;
533 if (stack_top
->u
.ptr
.rev_bo
)
535 stack_top
->u
.v
= tmp
;
536 stack_top
->type
= REG_U64
;
539 case OBJECT_TYPE_UNSIGNED_ENUM
:
543 dbg_printf("op load field unsigned enumeration\n");
544 tmp
= *(uint64_t *) stack_top
->u
.ptr
.ptr
;
545 if (stack_top
->u
.ptr
.rev_bo
)
547 stack_top
->u
.v
= tmp
;
548 stack_top
->type
= REG_U64
;
551 case OBJECT_TYPE_DOUBLE
:
552 memcpy(&stack_top
->u
.d
,
553 stack_top
->u
.ptr
.ptr
,
554 sizeof(struct literal_double
));
555 stack_top
->type
= REG_DOUBLE
;
557 case OBJECT_TYPE_STRING
:
561 dbg_printf("op load field string\n");
562 str
= (const char *) stack_top
->u
.ptr
.ptr
;
563 stack_top
->u
.s
.str
= str
;
564 if (unlikely(!stack_top
->u
.s
.str
)) {
565 dbg_printf("Interpreter warning: loading a NULL string.\n");
569 stack_top
->u
.s
.seq_len
= SIZE_MAX
;
570 stack_top
->u
.s
.literal_type
=
571 ESTACK_STRING_LITERAL_TYPE_NONE
;
572 stack_top
->type
= REG_STRING
;
575 case OBJECT_TYPE_STRING_SEQUENCE
:
579 dbg_printf("op load field string sequence\n");
580 ptr
= stack_top
->u
.ptr
.ptr
;
581 stack_top
->u
.s
.seq_len
= *(unsigned long *) ptr
;
582 stack_top
->u
.s
.str
= *(const char **) (ptr
+ sizeof(unsigned long));
583 stack_top
->type
= REG_STRING
;
584 if (unlikely(!stack_top
->u
.s
.str
)) {
585 dbg_printf("Interpreter warning: loading a NULL sequence.\n");
589 stack_top
->u
.s
.literal_type
=
590 ESTACK_STRING_LITERAL_TYPE_NONE
;
593 case OBJECT_TYPE_DYNAMIC
:
595 * Dynamic types in context are looked up
596 * by context get index.
600 case OBJECT_TYPE_SEQUENCE
:
601 case OBJECT_TYPE_ARRAY
:
602 case OBJECT_TYPE_STRUCT
:
603 case OBJECT_TYPE_VARIANT
:
604 ERR("Sequences, arrays, struct and variant cannot be loaded (nested types).");
615 int lttng_bytecode_interpret_format_output(struct estack_entry
*ax
,
616 struct lttng_interpreter_output
*output
)
623 output
->type
= LTTNG_INTERPRETER_TYPE_S64
;
624 output
->u
.s
= ax
->u
.v
;
627 output
->type
= LTTNG_INTERPRETER_TYPE_U64
;
628 output
->u
.u
= (uint64_t) ax
->u
.v
;
631 output
->type
= LTTNG_INTERPRETER_TYPE_DOUBLE
;
632 output
->u
.d
= ax
->u
.d
;
635 output
->type
= LTTNG_INTERPRETER_TYPE_STRING
;
636 output
->u
.str
.str
= ax
->u
.s
.str
;
637 output
->u
.str
.len
= ax
->u
.s
.seq_len
;
640 switch (ax
->u
.ptr
.object_type
) {
642 case OBJECT_TYPE_S16
:
643 case OBJECT_TYPE_S32
:
644 case OBJECT_TYPE_S64
:
646 case OBJECT_TYPE_U16
:
647 case OBJECT_TYPE_U32
:
648 case OBJECT_TYPE_U64
:
649 case OBJECT_TYPE_DOUBLE
:
650 case OBJECT_TYPE_STRING
:
651 case OBJECT_TYPE_STRING_SEQUENCE
:
652 ret
= dynamic_load_field(ax
);
655 /* Retry after loading ptr into stack top. */
657 case OBJECT_TYPE_SEQUENCE
:
658 output
->type
= LTTNG_INTERPRETER_TYPE_SEQUENCE
;
659 output
->u
.sequence
.ptr
= *(const char **) (ax
->u
.ptr
.ptr
+ sizeof(unsigned long));
660 output
->u
.sequence
.nr_elem
= *(unsigned long *) ax
->u
.ptr
.ptr
;
661 output
->u
.sequence
.nested_type
= lttng_ust_get_type_sequence(ax
->u
.ptr
.field
->type
)->elem_type
;
663 case OBJECT_TYPE_ARRAY
:
664 /* Skip count (unsigned long) */
665 output
->type
= LTTNG_INTERPRETER_TYPE_SEQUENCE
;
666 output
->u
.sequence
.ptr
= *(const char **) (ax
->u
.ptr
.ptr
+ sizeof(unsigned long));
667 output
->u
.sequence
.nr_elem
= lttng_ust_get_type_array(ax
->u
.ptr
.field
->type
)->length
;
668 output
->u
.sequence
.nested_type
= lttng_ust_get_type_array(ax
->u
.ptr
.field
->type
)->elem_type
;
670 case OBJECT_TYPE_SIGNED_ENUM
:
671 ret
= dynamic_load_field(ax
);
674 output
->type
= LTTNG_INTERPRETER_TYPE_SIGNED_ENUM
;
675 output
->u
.s
= ax
->u
.v
;
677 case OBJECT_TYPE_UNSIGNED_ENUM
:
678 ret
= dynamic_load_field(ax
);
681 output
->type
= LTTNG_INTERPRETER_TYPE_UNSIGNED_ENUM
;
682 output
->u
.u
= ax
->u
.v
;
684 case OBJECT_TYPE_STRUCT
:
685 case OBJECT_TYPE_VARIANT
:
691 case REG_STAR_GLOB_STRING
:
701 * For `output` equal to NULL:
702 * Return 0 (discard), or raise the 0x1 flag (log event).
703 * Currently, other flags are kept for future extensions and have no
705 * For `output` not equal to NULL:
706 * Return 0 on success, negative error value on error.
708 int lttng_bytecode_interpret(struct lttng_ust_bytecode_runtime
*ust_bytecode
,
709 const char *interpreter_stack_data
,
712 struct bytecode_runtime
*bytecode
= caa_container_of(ust_bytecode
, struct bytecode_runtime
, p
);
713 struct lttng_ust_ctx
*ctx
= lttng_ust_rcu_dereference(*ust_bytecode
->priv
->pctx
);
714 void *pc
, *next_pc
, *start_pc
;
715 int ret
= -EINVAL
, retval
= 0;
716 struct estack _stack
;
717 struct estack
*stack
= &_stack
;
718 register int64_t ax
= 0, bx
= 0;
719 register enum entry_type ax_t
= REG_UNKNOWN
, bx_t
= REG_UNKNOWN
;
720 register int top
= INTERPRETER_STACK_EMPTY
;
721 #ifndef INTERPRETER_USE_SWITCH
722 static void *dispatch
[NR_BYTECODE_OPS
] = {
723 [ BYTECODE_OP_UNKNOWN
] = &&LABEL_BYTECODE_OP_UNKNOWN
,
725 [ BYTECODE_OP_RETURN
] = &&LABEL_BYTECODE_OP_RETURN
,
728 [ BYTECODE_OP_MUL
] = &&LABEL_BYTECODE_OP_MUL
,
729 [ BYTECODE_OP_DIV
] = &&LABEL_BYTECODE_OP_DIV
,
730 [ BYTECODE_OP_MOD
] = &&LABEL_BYTECODE_OP_MOD
,
731 [ BYTECODE_OP_PLUS
] = &&LABEL_BYTECODE_OP_PLUS
,
732 [ BYTECODE_OP_MINUS
] = &&LABEL_BYTECODE_OP_MINUS
,
733 [ BYTECODE_OP_BIT_RSHIFT
] = &&LABEL_BYTECODE_OP_BIT_RSHIFT
,
734 [ BYTECODE_OP_BIT_LSHIFT
] = &&LABEL_BYTECODE_OP_BIT_LSHIFT
,
735 [ BYTECODE_OP_BIT_AND
] = &&LABEL_BYTECODE_OP_BIT_AND
,
736 [ BYTECODE_OP_BIT_OR
] = &&LABEL_BYTECODE_OP_BIT_OR
,
737 [ BYTECODE_OP_BIT_XOR
] = &&LABEL_BYTECODE_OP_BIT_XOR
,
739 /* binary comparators */
740 [ BYTECODE_OP_EQ
] = &&LABEL_BYTECODE_OP_EQ
,
741 [ BYTECODE_OP_NE
] = &&LABEL_BYTECODE_OP_NE
,
742 [ BYTECODE_OP_GT
] = &&LABEL_BYTECODE_OP_GT
,
743 [ BYTECODE_OP_LT
] = &&LABEL_BYTECODE_OP_LT
,
744 [ BYTECODE_OP_GE
] = &&LABEL_BYTECODE_OP_GE
,
745 [ BYTECODE_OP_LE
] = &&LABEL_BYTECODE_OP_LE
,
747 /* string binary comparator */
748 [ BYTECODE_OP_EQ_STRING
] = &&LABEL_BYTECODE_OP_EQ_STRING
,
749 [ BYTECODE_OP_NE_STRING
] = &&LABEL_BYTECODE_OP_NE_STRING
,
750 [ BYTECODE_OP_GT_STRING
] = &&LABEL_BYTECODE_OP_GT_STRING
,
751 [ BYTECODE_OP_LT_STRING
] = &&LABEL_BYTECODE_OP_LT_STRING
,
752 [ BYTECODE_OP_GE_STRING
] = &&LABEL_BYTECODE_OP_GE_STRING
,
753 [ BYTECODE_OP_LE_STRING
] = &&LABEL_BYTECODE_OP_LE_STRING
,
755 /* globbing pattern binary comparator */
756 [ BYTECODE_OP_EQ_STAR_GLOB_STRING
] = &&LABEL_BYTECODE_OP_EQ_STAR_GLOB_STRING
,
757 [ BYTECODE_OP_NE_STAR_GLOB_STRING
] = &&LABEL_BYTECODE_OP_NE_STAR_GLOB_STRING
,
759 /* s64 binary comparator */
760 [ BYTECODE_OP_EQ_S64
] = &&LABEL_BYTECODE_OP_EQ_S64
,
761 [ BYTECODE_OP_NE_S64
] = &&LABEL_BYTECODE_OP_NE_S64
,
762 [ BYTECODE_OP_GT_S64
] = &&LABEL_BYTECODE_OP_GT_S64
,
763 [ BYTECODE_OP_LT_S64
] = &&LABEL_BYTECODE_OP_LT_S64
,
764 [ BYTECODE_OP_GE_S64
] = &&LABEL_BYTECODE_OP_GE_S64
,
765 [ BYTECODE_OP_LE_S64
] = &&LABEL_BYTECODE_OP_LE_S64
,
767 /* double binary comparator */
768 [ BYTECODE_OP_EQ_DOUBLE
] = &&LABEL_BYTECODE_OP_EQ_DOUBLE
,
769 [ BYTECODE_OP_NE_DOUBLE
] = &&LABEL_BYTECODE_OP_NE_DOUBLE
,
770 [ BYTECODE_OP_GT_DOUBLE
] = &&LABEL_BYTECODE_OP_GT_DOUBLE
,
771 [ BYTECODE_OP_LT_DOUBLE
] = &&LABEL_BYTECODE_OP_LT_DOUBLE
,
772 [ BYTECODE_OP_GE_DOUBLE
] = &&LABEL_BYTECODE_OP_GE_DOUBLE
,
773 [ BYTECODE_OP_LE_DOUBLE
] = &&LABEL_BYTECODE_OP_LE_DOUBLE
,
775 /* Mixed S64-double binary comparators */
776 [ BYTECODE_OP_EQ_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_EQ_DOUBLE_S64
,
777 [ BYTECODE_OP_NE_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_NE_DOUBLE_S64
,
778 [ BYTECODE_OP_GT_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_GT_DOUBLE_S64
,
779 [ BYTECODE_OP_LT_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_LT_DOUBLE_S64
,
780 [ BYTECODE_OP_GE_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_GE_DOUBLE_S64
,
781 [ BYTECODE_OP_LE_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_LE_DOUBLE_S64
,
783 [ BYTECODE_OP_EQ_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_EQ_S64_DOUBLE
,
784 [ BYTECODE_OP_NE_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_NE_S64_DOUBLE
,
785 [ BYTECODE_OP_GT_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_GT_S64_DOUBLE
,
786 [ BYTECODE_OP_LT_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_LT_S64_DOUBLE
,
787 [ BYTECODE_OP_GE_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_GE_S64_DOUBLE
,
788 [ BYTECODE_OP_LE_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_LE_S64_DOUBLE
,
791 [ BYTECODE_OP_UNARY_PLUS
] = &&LABEL_BYTECODE_OP_UNARY_PLUS
,
792 [ BYTECODE_OP_UNARY_MINUS
] = &&LABEL_BYTECODE_OP_UNARY_MINUS
,
793 [ BYTECODE_OP_UNARY_NOT
] = &&LABEL_BYTECODE_OP_UNARY_NOT
,
794 [ BYTECODE_OP_UNARY_PLUS_S64
] = &&LABEL_BYTECODE_OP_UNARY_PLUS_S64
,
795 [ BYTECODE_OP_UNARY_MINUS_S64
] = &&LABEL_BYTECODE_OP_UNARY_MINUS_S64
,
796 [ BYTECODE_OP_UNARY_NOT_S64
] = &&LABEL_BYTECODE_OP_UNARY_NOT_S64
,
797 [ BYTECODE_OP_UNARY_PLUS_DOUBLE
] = &&LABEL_BYTECODE_OP_UNARY_PLUS_DOUBLE
,
798 [ BYTECODE_OP_UNARY_MINUS_DOUBLE
] = &&LABEL_BYTECODE_OP_UNARY_MINUS_DOUBLE
,
799 [ BYTECODE_OP_UNARY_NOT_DOUBLE
] = &&LABEL_BYTECODE_OP_UNARY_NOT_DOUBLE
,
802 [ BYTECODE_OP_AND
] = &&LABEL_BYTECODE_OP_AND
,
803 [ BYTECODE_OP_OR
] = &&LABEL_BYTECODE_OP_OR
,
806 [ BYTECODE_OP_LOAD_FIELD_REF
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF
,
807 [ BYTECODE_OP_LOAD_FIELD_REF_STRING
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_STRING
,
808 [ BYTECODE_OP_LOAD_FIELD_REF_SEQUENCE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_SEQUENCE
,
809 [ BYTECODE_OP_LOAD_FIELD_REF_S64
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_S64
,
810 [ BYTECODE_OP_LOAD_FIELD_REF_DOUBLE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_DOUBLE
,
812 /* load from immediate operand */
813 [ BYTECODE_OP_LOAD_STRING
] = &&LABEL_BYTECODE_OP_LOAD_STRING
,
814 [ BYTECODE_OP_LOAD_STAR_GLOB_STRING
] = &&LABEL_BYTECODE_OP_LOAD_STAR_GLOB_STRING
,
815 [ BYTECODE_OP_LOAD_S64
] = &&LABEL_BYTECODE_OP_LOAD_S64
,
816 [ BYTECODE_OP_LOAD_DOUBLE
] = &&LABEL_BYTECODE_OP_LOAD_DOUBLE
,
819 [ BYTECODE_OP_CAST_TO_S64
] = &&LABEL_BYTECODE_OP_CAST_TO_S64
,
820 [ BYTECODE_OP_CAST_DOUBLE_TO_S64
] = &&LABEL_BYTECODE_OP_CAST_DOUBLE_TO_S64
,
821 [ BYTECODE_OP_CAST_NOP
] = &&LABEL_BYTECODE_OP_CAST_NOP
,
823 /* get context ref */
824 [ BYTECODE_OP_GET_CONTEXT_REF
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF
,
825 [ BYTECODE_OP_GET_CONTEXT_REF_STRING
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF_STRING
,
826 [ BYTECODE_OP_GET_CONTEXT_REF_S64
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF_S64
,
827 [ BYTECODE_OP_GET_CONTEXT_REF_DOUBLE
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF_DOUBLE
,
829 /* Instructions for recursive traversal through composed types. */
830 [ BYTECODE_OP_GET_CONTEXT_ROOT
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_ROOT
,
831 [ BYTECODE_OP_GET_APP_CONTEXT_ROOT
] = &&LABEL_BYTECODE_OP_GET_APP_CONTEXT_ROOT
,
832 [ BYTECODE_OP_GET_PAYLOAD_ROOT
] = &&LABEL_BYTECODE_OP_GET_PAYLOAD_ROOT
,
834 [ BYTECODE_OP_GET_SYMBOL
] = &&LABEL_BYTECODE_OP_GET_SYMBOL
,
835 [ BYTECODE_OP_GET_SYMBOL_FIELD
] = &&LABEL_BYTECODE_OP_GET_SYMBOL_FIELD
,
836 [ BYTECODE_OP_GET_INDEX_U16
] = &&LABEL_BYTECODE_OP_GET_INDEX_U16
,
837 [ BYTECODE_OP_GET_INDEX_U64
] = &&LABEL_BYTECODE_OP_GET_INDEX_U64
,
839 [ BYTECODE_OP_LOAD_FIELD
] = &&LABEL_BYTECODE_OP_LOAD_FIELD
,
840 [ BYTECODE_OP_LOAD_FIELD_S8
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S8
,
841 [ BYTECODE_OP_LOAD_FIELD_S16
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S16
,
842 [ BYTECODE_OP_LOAD_FIELD_S32
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S32
,
843 [ BYTECODE_OP_LOAD_FIELD_S64
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S64
,
844 [ BYTECODE_OP_LOAD_FIELD_U8
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U8
,
845 [ BYTECODE_OP_LOAD_FIELD_U16
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U16
,
846 [ BYTECODE_OP_LOAD_FIELD_U32
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U32
,
847 [ BYTECODE_OP_LOAD_FIELD_U64
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U64
,
848 [ BYTECODE_OP_LOAD_FIELD_STRING
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_STRING
,
849 [ BYTECODE_OP_LOAD_FIELD_SEQUENCE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_SEQUENCE
,
850 [ BYTECODE_OP_LOAD_FIELD_DOUBLE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_DOUBLE
,
852 [ BYTECODE_OP_UNARY_BIT_NOT
] = &&LABEL_BYTECODE_OP_UNARY_BIT_NOT
,
854 [ BYTECODE_OP_RETURN_S64
] = &&LABEL_BYTECODE_OP_RETURN_S64
,
856 #endif /* #ifndef INTERPRETER_USE_SWITCH */
860 OP(BYTECODE_OP_UNKNOWN
):
861 OP(BYTECODE_OP_LOAD_FIELD_REF
):
862 #ifdef INTERPRETER_USE_SWITCH
864 #endif /* INTERPRETER_USE_SWITCH */
865 ERR("unknown bytecode op %u",
866 (unsigned int) *(bytecode_opcode_t
*) pc
);
870 OP(BYTECODE_OP_RETURN
):
871 /* LTTNG_UST_BYTECODE_INTERPRETER_ERROR or LTTNG_UST_BYTECODE_INTERPRETER_OK */
872 /* Handle dynamic typing. */
873 switch (estack_ax_t
) {
876 retval
= !!estack_ax_v
;
881 if (ust_bytecode
->priv
->type
!= LTTNG_UST_BYTECODE_TYPE_CAPTURE
) {
887 case REG_STAR_GLOB_STRING
:
896 OP(BYTECODE_OP_RETURN_S64
):
897 /* LTTNG_UST_BYTECODE_INTERPRETER_ERROR or LTTNG_UST_BYTECODE_INTERPRETER_OK */
898 retval
= !!estack_ax_v
;
906 OP(BYTECODE_OP_PLUS
):
907 OP(BYTECODE_OP_MINUS
):
908 ERR("unsupported bytecode op %u",
909 (unsigned int) *(bytecode_opcode_t
*) pc
);
915 /* Dynamic typing. */
916 switch (estack_ax_t
) {
917 case REG_S64
: /* Fall-through */
919 switch (estack_bx_t
) {
920 case REG_S64
: /* Fall-through */
922 JUMP_TO(BYTECODE_OP_EQ_S64
);
924 JUMP_TO(BYTECODE_OP_EQ_DOUBLE_S64
);
925 case REG_STRING
: /* Fall-through */
926 case REG_STAR_GLOB_STRING
:
930 ERR("Unknown interpreter register type (%d)",
937 switch (estack_bx_t
) {
938 case REG_S64
: /* Fall-through */
940 JUMP_TO(BYTECODE_OP_EQ_S64_DOUBLE
);
942 JUMP_TO(BYTECODE_OP_EQ_DOUBLE
);
943 case REG_STRING
: /* Fall-through */
944 case REG_STAR_GLOB_STRING
:
948 ERR("Unknown interpreter register type (%d)",
955 switch (estack_bx_t
) {
956 case REG_S64
: /* Fall-through */
957 case REG_U64
: /* Fall-through */
962 JUMP_TO(BYTECODE_OP_EQ_STRING
);
963 case REG_STAR_GLOB_STRING
:
964 JUMP_TO(BYTECODE_OP_EQ_STAR_GLOB_STRING
);
966 ERR("Unknown interpreter register type (%d)",
972 case REG_STAR_GLOB_STRING
:
973 switch (estack_bx_t
) {
974 case REG_S64
: /* Fall-through */
975 case REG_U64
: /* Fall-through */
980 JUMP_TO(BYTECODE_OP_EQ_STAR_GLOB_STRING
);
981 case REG_STAR_GLOB_STRING
:
985 ERR("Unknown interpreter register type (%d)",
992 ERR("Unknown interpreter register type (%d)",
1000 /* Dynamic typing. */
1001 switch (estack_ax_t
) {
1002 case REG_S64
: /* Fall-through */
1004 switch (estack_bx_t
) {
1005 case REG_S64
: /* Fall-through */
1007 JUMP_TO(BYTECODE_OP_NE_S64
);
1009 JUMP_TO(BYTECODE_OP_NE_DOUBLE_S64
);
1010 case REG_STRING
: /* Fall-through */
1011 case REG_STAR_GLOB_STRING
:
1015 ERR("Unknown interpreter register type (%d)",
1022 switch (estack_bx_t
) {
1023 case REG_S64
: /* Fall-through */
1025 JUMP_TO(BYTECODE_OP_NE_S64_DOUBLE
);
1027 JUMP_TO(BYTECODE_OP_NE_DOUBLE
);
1028 case REG_STRING
: /* Fall-through */
1029 case REG_STAR_GLOB_STRING
:
1033 ERR("Unknown interpreter register type (%d)",
1040 switch (estack_bx_t
) {
1041 case REG_S64
: /* Fall-through */
1047 JUMP_TO(BYTECODE_OP_NE_STRING
);
1048 case REG_STAR_GLOB_STRING
:
1049 JUMP_TO(BYTECODE_OP_NE_STAR_GLOB_STRING
);
1051 ERR("Unknown interpreter register type (%d)",
1057 case REG_STAR_GLOB_STRING
:
1058 switch (estack_bx_t
) {
1059 case REG_S64
: /* Fall-through */
1065 JUMP_TO(BYTECODE_OP_NE_STAR_GLOB_STRING
);
1066 case REG_STAR_GLOB_STRING
:
1070 ERR("Unknown interpreter register type (%d)",
1077 ERR("Unknown interpreter register type (%d)",
1085 /* Dynamic typing. */
1086 switch (estack_ax_t
) {
1087 case REG_S64
: /* Fall-through */
1089 switch (estack_bx_t
) {
1090 case REG_S64
: /* Fall-through */
1092 JUMP_TO(BYTECODE_OP_GT_S64
);
1094 JUMP_TO(BYTECODE_OP_GT_DOUBLE_S64
);
1095 case REG_STRING
: /* Fall-through */
1096 case REG_STAR_GLOB_STRING
:
1100 ERR("Unknown interpreter register type (%d)",
1107 switch (estack_bx_t
) {
1108 case REG_S64
: /* Fall-through */
1110 JUMP_TO(BYTECODE_OP_GT_S64_DOUBLE
);
1112 JUMP_TO(BYTECODE_OP_GT_DOUBLE
);
1113 case REG_STRING
: /* Fall-through */
1114 case REG_STAR_GLOB_STRING
:
1118 ERR("Unknown interpreter register type (%d)",
1125 switch (estack_bx_t
) {
1126 case REG_S64
: /* Fall-through */
1127 case REG_U64
: /* Fall-through */
1128 case REG_DOUBLE
: /* Fall-through */
1129 case REG_STAR_GLOB_STRING
:
1133 JUMP_TO(BYTECODE_OP_GT_STRING
);
1135 ERR("Unknown interpreter register type (%d)",
1142 ERR("Unknown interpreter register type (%d)",
1150 /* Dynamic typing. */
1151 switch (estack_ax_t
) {
1152 case REG_S64
: /* Fall-through */
1154 switch (estack_bx_t
) {
1155 case REG_S64
: /* Fall-through */
1157 JUMP_TO(BYTECODE_OP_LT_S64
);
1159 JUMP_TO(BYTECODE_OP_LT_DOUBLE_S64
);
1160 case REG_STRING
: /* Fall-through */
1161 case REG_STAR_GLOB_STRING
:
1165 ERR("Unknown interpreter register type (%d)",
1172 switch (estack_bx_t
) {
1173 case REG_S64
: /* Fall-through */
1175 JUMP_TO(BYTECODE_OP_LT_S64_DOUBLE
);
1177 JUMP_TO(BYTECODE_OP_LT_DOUBLE
);
1178 case REG_STRING
: /* Fall-through */
1179 case REG_STAR_GLOB_STRING
:
1183 ERR("Unknown interpreter register type (%d)",
1190 switch (estack_bx_t
) {
1191 case REG_S64
: /* Fall-through */
1192 case REG_U64
: /* Fall-through */
1193 case REG_DOUBLE
: /* Fall-through */
1194 case REG_STAR_GLOB_STRING
:
1198 JUMP_TO(BYTECODE_OP_LT_STRING
);
1200 ERR("Unknown interpreter register type (%d)",
1207 ERR("Unknown interpreter register type (%d)",
1215 /* Dynamic typing. */
1216 switch (estack_ax_t
) {
1217 case REG_S64
: /* Fall-through */
1219 switch (estack_bx_t
) {
1220 case REG_S64
: /* Fall-through */
1222 JUMP_TO(BYTECODE_OP_GE_S64
);
1224 JUMP_TO(BYTECODE_OP_GE_DOUBLE_S64
);
1225 case REG_STRING
: /* Fall-through */
1226 case REG_STAR_GLOB_STRING
:
1230 ERR("Unknown interpreter register type (%d)",
1237 switch (estack_bx_t
) {
1238 case REG_S64
: /* Fall-through */
1240 JUMP_TO(BYTECODE_OP_GE_S64_DOUBLE
);
1242 JUMP_TO(BYTECODE_OP_GE_DOUBLE
);
1243 case REG_STRING
: /* Fall-through */
1244 case REG_STAR_GLOB_STRING
:
1248 ERR("Unknown interpreter register type (%d)",
1255 switch (estack_bx_t
) {
1256 case REG_S64
: /* Fall-through */
1257 case REG_U64
: /* Fall-through */
1258 case REG_DOUBLE
: /* Fall-through */
1259 case REG_STAR_GLOB_STRING
:
1263 JUMP_TO(BYTECODE_OP_GE_STRING
);
1265 ERR("Unknown interpreter register type (%d)",
1272 ERR("Unknown interpreter register type (%d)",
1280 /* Dynamic typing. */
1281 switch (estack_ax_t
) {
1282 case REG_S64
: /* Fall-through */
1284 switch (estack_bx_t
) {
1285 case REG_S64
: /* Fall-through */
1287 JUMP_TO(BYTECODE_OP_LE_S64
);
1289 JUMP_TO(BYTECODE_OP_LE_DOUBLE_S64
);
1290 case REG_STRING
: /* Fall-through */
1291 case REG_STAR_GLOB_STRING
:
1295 ERR("Unknown interpreter register type (%d)",
1302 switch (estack_bx_t
) {
1303 case REG_S64
: /* Fall-through */
1305 JUMP_TO(BYTECODE_OP_LE_S64_DOUBLE
);
1307 JUMP_TO(BYTECODE_OP_LE_DOUBLE
);
1308 case REG_STRING
: /* Fall-through */
1309 case REG_STAR_GLOB_STRING
:
1313 ERR("Unknown interpreter register type (%d)",
1320 switch (estack_bx_t
) {
1321 case REG_S64
: /* Fall-through */
1322 case REG_U64
: /* Fall-through */
1323 case REG_DOUBLE
: /* Fall-through */
1324 case REG_STAR_GLOB_STRING
:
1328 JUMP_TO(BYTECODE_OP_LE_STRING
);
1330 ERR("Unknown interpreter register type (%d)",
1337 ERR("Unknown interpreter register type (%d)",
1344 OP(BYTECODE_OP_EQ_STRING
):
1348 res
= (stack_strcmp(stack
, top
, "==") == 0);
1349 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1351 estack_ax_t
= REG_S64
;
1352 next_pc
+= sizeof(struct binary_op
);
1355 OP(BYTECODE_OP_NE_STRING
):
1359 res
= (stack_strcmp(stack
, top
, "!=") != 0);
1360 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1362 estack_ax_t
= REG_S64
;
1363 next_pc
+= sizeof(struct binary_op
);
1366 OP(BYTECODE_OP_GT_STRING
):
1370 res
= (stack_strcmp(stack
, top
, ">") > 0);
1371 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1373 estack_ax_t
= REG_S64
;
1374 next_pc
+= sizeof(struct binary_op
);
1377 OP(BYTECODE_OP_LT_STRING
):
1381 res
= (stack_strcmp(stack
, top
, "<") < 0);
1382 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1384 estack_ax_t
= REG_S64
;
1385 next_pc
+= sizeof(struct binary_op
);
1388 OP(BYTECODE_OP_GE_STRING
):
1392 res
= (stack_strcmp(stack
, top
, ">=") >= 0);
1393 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1395 estack_ax_t
= REG_S64
;
1396 next_pc
+= sizeof(struct binary_op
);
1399 OP(BYTECODE_OP_LE_STRING
):
1403 res
= (stack_strcmp(stack
, top
, "<=") <= 0);
1404 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1406 estack_ax_t
= REG_S64
;
1407 next_pc
+= sizeof(struct binary_op
);
1411 OP(BYTECODE_OP_EQ_STAR_GLOB_STRING
):
1415 res
= (stack_star_glob_match(stack
, top
, "==") == 0);
1416 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1418 estack_ax_t
= REG_S64
;
1419 next_pc
+= sizeof(struct binary_op
);
1422 OP(BYTECODE_OP_NE_STAR_GLOB_STRING
):
1426 res
= (stack_star_glob_match(stack
, top
, "!=") != 0);
1427 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1429 estack_ax_t
= REG_S64
;
1430 next_pc
+= sizeof(struct binary_op
);
1434 OP(BYTECODE_OP_EQ_S64
):
1438 res
= (estack_bx_v
== estack_ax_v
);
1439 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1441 estack_ax_t
= REG_S64
;
1442 next_pc
+= sizeof(struct binary_op
);
1445 OP(BYTECODE_OP_NE_S64
):
1449 res
= (estack_bx_v
!= estack_ax_v
);
1450 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1452 estack_ax_t
= REG_S64
;
1453 next_pc
+= sizeof(struct binary_op
);
1456 OP(BYTECODE_OP_GT_S64
):
1460 res
= (estack_bx_v
> estack_ax_v
);
1461 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1463 estack_ax_t
= REG_S64
;
1464 next_pc
+= sizeof(struct binary_op
);
1467 OP(BYTECODE_OP_LT_S64
):
1471 res
= (estack_bx_v
< estack_ax_v
);
1472 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1474 estack_ax_t
= REG_S64
;
1475 next_pc
+= sizeof(struct binary_op
);
1478 OP(BYTECODE_OP_GE_S64
):
1482 res
= (estack_bx_v
>= estack_ax_v
);
1483 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1485 estack_ax_t
= REG_S64
;
1486 next_pc
+= sizeof(struct binary_op
);
1489 OP(BYTECODE_OP_LE_S64
):
1493 res
= (estack_bx_v
<= estack_ax_v
);
1494 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1496 estack_ax_t
= REG_S64
;
1497 next_pc
+= sizeof(struct binary_op
);
1501 OP(BYTECODE_OP_EQ_DOUBLE
):
1505 res
= (estack_bx(stack
, top
)->u
.d
== estack_ax(stack
, top
)->u
.d
);
1506 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1508 estack_ax_t
= REG_S64
;
1509 next_pc
+= sizeof(struct binary_op
);
1512 OP(BYTECODE_OP_NE_DOUBLE
):
1516 res
= (estack_bx(stack
, top
)->u
.d
!= estack_ax(stack
, top
)->u
.d
);
1517 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1519 estack_ax_t
= REG_S64
;
1520 next_pc
+= sizeof(struct binary_op
);
1523 OP(BYTECODE_OP_GT_DOUBLE
):
1527 res
= (estack_bx(stack
, top
)->u
.d
> estack_ax(stack
, top
)->u
.d
);
1528 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1530 estack_ax_t
= REG_S64
;
1531 next_pc
+= sizeof(struct binary_op
);
1534 OP(BYTECODE_OP_LT_DOUBLE
):
1538 res
= (estack_bx(stack
, top
)->u
.d
< estack_ax(stack
, top
)->u
.d
);
1539 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1541 estack_ax_t
= REG_S64
;
1542 next_pc
+= sizeof(struct binary_op
);
1545 OP(BYTECODE_OP_GE_DOUBLE
):
1549 res
= (estack_bx(stack
, top
)->u
.d
>= estack_ax(stack
, top
)->u
.d
);
1550 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1552 estack_ax_t
= REG_S64
;
1553 next_pc
+= sizeof(struct binary_op
);
1556 OP(BYTECODE_OP_LE_DOUBLE
):
1560 res
= (estack_bx(stack
, top
)->u
.d
<= estack_ax(stack
, top
)->u
.d
);
1561 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1563 estack_ax_t
= REG_S64
;
1564 next_pc
+= sizeof(struct binary_op
);
1568 /* Mixed S64-double binary comparators */
1569 OP(BYTECODE_OP_EQ_DOUBLE_S64
):
1573 res
= (estack_bx(stack
, top
)->u
.d
== estack_ax_v
);
1574 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1576 estack_ax_t
= REG_S64
;
1577 next_pc
+= sizeof(struct binary_op
);
1580 OP(BYTECODE_OP_NE_DOUBLE_S64
):
1584 res
= (estack_bx(stack
, top
)->u
.d
!= estack_ax_v
);
1585 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1587 estack_ax_t
= REG_S64
;
1588 next_pc
+= sizeof(struct binary_op
);
1591 OP(BYTECODE_OP_GT_DOUBLE_S64
):
1595 res
= (estack_bx(stack
, top
)->u
.d
> estack_ax_v
);
1596 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1598 estack_ax_t
= REG_S64
;
1599 next_pc
+= sizeof(struct binary_op
);
1602 OP(BYTECODE_OP_LT_DOUBLE_S64
):
1606 res
= (estack_bx(stack
, top
)->u
.d
< estack_ax_v
);
1607 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1609 estack_ax_t
= REG_S64
;
1610 next_pc
+= sizeof(struct binary_op
);
1613 OP(BYTECODE_OP_GE_DOUBLE_S64
):
1617 res
= (estack_bx(stack
, top
)->u
.d
>= estack_ax_v
);
1618 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1620 estack_ax_t
= REG_S64
;
1621 next_pc
+= sizeof(struct binary_op
);
1624 OP(BYTECODE_OP_LE_DOUBLE_S64
):
1628 res
= (estack_bx(stack
, top
)->u
.d
<= estack_ax_v
);
1629 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1631 estack_ax_t
= REG_S64
;
1632 next_pc
+= sizeof(struct binary_op
);
1636 OP(BYTECODE_OP_EQ_S64_DOUBLE
):
1640 res
= (estack_bx_v
== estack_ax(stack
, top
)->u
.d
);
1641 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1643 estack_ax_t
= REG_S64
;
1644 next_pc
+= sizeof(struct binary_op
);
1647 OP(BYTECODE_OP_NE_S64_DOUBLE
):
1651 res
= (estack_bx_v
!= estack_ax(stack
, top
)->u
.d
);
1652 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1654 estack_ax_t
= REG_S64
;
1655 next_pc
+= sizeof(struct binary_op
);
1658 OP(BYTECODE_OP_GT_S64_DOUBLE
):
1662 res
= (estack_bx_v
> estack_ax(stack
, top
)->u
.d
);
1663 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1665 estack_ax_t
= REG_S64
;
1666 next_pc
+= sizeof(struct binary_op
);
1669 OP(BYTECODE_OP_LT_S64_DOUBLE
):
1673 res
= (estack_bx_v
< estack_ax(stack
, top
)->u
.d
);
1674 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1676 estack_ax_t
= REG_S64
;
1677 next_pc
+= sizeof(struct binary_op
);
1680 OP(BYTECODE_OP_GE_S64_DOUBLE
):
1684 res
= (estack_bx_v
>= estack_ax(stack
, top
)->u
.d
);
1685 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1687 estack_ax_t
= REG_S64
;
1688 next_pc
+= sizeof(struct binary_op
);
1691 OP(BYTECODE_OP_LE_S64_DOUBLE
):
1695 res
= (estack_bx_v
<= estack_ax(stack
, top
)->u
.d
);
1696 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1698 estack_ax_t
= REG_S64
;
1699 next_pc
+= sizeof(struct binary_op
);
1702 OP(BYTECODE_OP_BIT_RSHIFT
):
1706 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1711 /* Catch undefined behavior. */
1712 if (caa_unlikely(estack_ax_v
< 0 || estack_ax_v
>= 64)) {
1716 res
= ((uint64_t) estack_bx_v
>> (uint32_t) estack_ax_v
);
1717 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1719 estack_ax_t
= REG_U64
;
1720 next_pc
+= sizeof(struct binary_op
);
1723 OP(BYTECODE_OP_BIT_LSHIFT
):
1727 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1732 /* Catch undefined behavior. */
1733 if (caa_unlikely(estack_ax_v
< 0 || estack_ax_v
>= 64)) {
1737 res
= ((uint64_t) estack_bx_v
<< (uint32_t) estack_ax_v
);
1738 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1740 estack_ax_t
= REG_U64
;
1741 next_pc
+= sizeof(struct binary_op
);
1744 OP(BYTECODE_OP_BIT_AND
):
1748 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1753 res
= ((uint64_t) estack_bx_v
& (uint64_t) estack_ax_v
);
1754 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1756 estack_ax_t
= REG_U64
;
1757 next_pc
+= sizeof(struct binary_op
);
1760 OP(BYTECODE_OP_BIT_OR
):
1764 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1769 res
= ((uint64_t) estack_bx_v
| (uint64_t) estack_ax_v
);
1770 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1772 estack_ax_t
= REG_U64
;
1773 next_pc
+= sizeof(struct binary_op
);
1776 OP(BYTECODE_OP_BIT_XOR
):
1780 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1785 res
= ((uint64_t) estack_bx_v
^ (uint64_t) estack_ax_v
);
1786 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1788 estack_ax_t
= REG_U64
;
1789 next_pc
+= sizeof(struct binary_op
);
1794 OP(BYTECODE_OP_UNARY_PLUS
):
1796 /* Dynamic typing. */
1797 switch (estack_ax_t
) {
1798 case REG_S64
: /* Fall-through. */
1800 JUMP_TO(BYTECODE_OP_UNARY_PLUS_S64
);
1802 JUMP_TO(BYTECODE_OP_UNARY_PLUS_DOUBLE
);
1803 case REG_STRING
: /* Fall-through */
1804 case REG_STAR_GLOB_STRING
:
1808 ERR("Unknown interpreter register type (%d)",
1814 OP(BYTECODE_OP_UNARY_MINUS
):
1816 /* Dynamic typing. */
1817 switch (estack_ax_t
) {
1818 case REG_S64
: /* Fall-through. */
1820 JUMP_TO(BYTECODE_OP_UNARY_MINUS_S64
);
1822 JUMP_TO(BYTECODE_OP_UNARY_MINUS_DOUBLE
);
1823 case REG_STRING
: /* Fall-through */
1824 case REG_STAR_GLOB_STRING
:
1828 ERR("Unknown interpreter register type (%d)",
1834 OP(BYTECODE_OP_UNARY_NOT
):
1836 /* Dynamic typing. */
1837 switch (estack_ax_t
) {
1838 case REG_S64
: /* Fall-through. */
1840 JUMP_TO(BYTECODE_OP_UNARY_NOT_S64
);
1842 JUMP_TO(BYTECODE_OP_UNARY_NOT_DOUBLE
);
1843 case REG_STRING
: /* Fall-through */
1844 case REG_STAR_GLOB_STRING
:
1848 ERR("Unknown interpreter register type (%d)",
1853 next_pc
+= sizeof(struct unary_op
);
1857 OP(BYTECODE_OP_UNARY_BIT_NOT
):
1859 /* Dynamic typing. */
1860 if (!IS_INTEGER_REGISTER(estack_ax_t
)) {
1865 estack_ax_v
= ~(uint64_t) estack_ax_v
;
1866 estack_ax_t
= REG_U64
;
1867 next_pc
+= sizeof(struct unary_op
);
1871 OP(BYTECODE_OP_UNARY_PLUS_S64
):
1872 OP(BYTECODE_OP_UNARY_PLUS_DOUBLE
):
1874 next_pc
+= sizeof(struct unary_op
);
1877 OP(BYTECODE_OP_UNARY_MINUS_S64
):
1879 estack_ax_v
= -estack_ax_v
;
1880 next_pc
+= sizeof(struct unary_op
);
1883 OP(BYTECODE_OP_UNARY_MINUS_DOUBLE
):
1885 estack_ax(stack
, top
)->u
.d
= -estack_ax(stack
, top
)->u
.d
;
1886 next_pc
+= sizeof(struct unary_op
);
1889 OP(BYTECODE_OP_UNARY_NOT_S64
):
1891 estack_ax_v
= !estack_ax_v
;
1892 estack_ax_t
= REG_S64
;
1893 next_pc
+= sizeof(struct unary_op
);
1896 OP(BYTECODE_OP_UNARY_NOT_DOUBLE
):
1898 estack_ax_v
= !estack_ax(stack
, top
)->u
.d
;
1899 estack_ax_t
= REG_S64
;
1900 next_pc
+= sizeof(struct unary_op
);
1905 OP(BYTECODE_OP_AND
):
1907 struct logical_op
*insn
= (struct logical_op
*) pc
;
1909 if (estack_ax_t
!= REG_S64
&& estack_ax_t
!= REG_U64
) {
1913 /* If AX is 0, skip and evaluate to 0 */
1914 if (unlikely(estack_ax_v
== 0)) {
1915 dbg_printf("Jumping to bytecode offset %u\n",
1916 (unsigned int) insn
->skip_offset
);
1917 next_pc
= start_pc
+ insn
->skip_offset
;
1919 /* Pop 1 when jump not taken */
1920 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1921 next_pc
+= sizeof(struct logical_op
);
1927 struct logical_op
*insn
= (struct logical_op
*) pc
;
1929 if (estack_ax_t
!= REG_S64
&& estack_ax_t
!= REG_U64
) {
1933 /* If AX is nonzero, skip and evaluate to 1 */
1934 if (unlikely(estack_ax_v
!= 0)) {
1936 dbg_printf("Jumping to bytecode offset %u\n",
1937 (unsigned int) insn
->skip_offset
);
1938 next_pc
= start_pc
+ insn
->skip_offset
;
1940 /* Pop 1 when jump not taken */
1941 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1942 next_pc
+= sizeof(struct logical_op
);
1948 /* load field ref */
1949 OP(BYTECODE_OP_LOAD_FIELD_REF_STRING
):
1951 struct load_op
*insn
= (struct load_op
*) pc
;
1952 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
1954 dbg_printf("load field ref offset %u type string\n",
1956 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1957 estack_ax(stack
, top
)->u
.s
.str
=
1958 *(const char * const *) &interpreter_stack_data
[ref
->offset
];
1959 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
1960 dbg_printf("Interpreter warning: loading a NULL string.\n");
1964 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
1965 estack_ax(stack
, top
)->u
.s
.literal_type
=
1966 ESTACK_STRING_LITERAL_TYPE_NONE
;
1967 estack_ax_t
= REG_STRING
;
1968 dbg_printf("ref load string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
1969 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
1973 OP(BYTECODE_OP_LOAD_FIELD_REF_SEQUENCE
):
1975 struct load_op
*insn
= (struct load_op
*) pc
;
1976 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
1978 dbg_printf("load field ref offset %u type sequence\n",
1980 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1981 estack_ax(stack
, top
)->u
.s
.seq_len
=
1982 *(unsigned long *) &interpreter_stack_data
[ref
->offset
];
1983 estack_ax(stack
, top
)->u
.s
.str
=
1984 *(const char **) (&interpreter_stack_data
[ref
->offset
1985 + sizeof(unsigned long)]);
1986 estack_ax_t
= REG_STRING
;
1987 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
1988 dbg_printf("Interpreter warning: loading a NULL sequence.\n");
1992 estack_ax(stack
, top
)->u
.s
.literal_type
=
1993 ESTACK_STRING_LITERAL_TYPE_NONE
;
1994 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
1998 OP(BYTECODE_OP_LOAD_FIELD_REF_S64
):
2000 struct load_op
*insn
= (struct load_op
*) pc
;
2001 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2003 dbg_printf("load field ref offset %u type s64\n",
2005 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2007 ((struct literal_numeric
*) &interpreter_stack_data
[ref
->offset
])->v
;
2008 estack_ax_t
= REG_S64
;
2009 dbg_printf("ref load s64 %" PRIi64
"\n", estack_ax_v
);
2010 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2014 OP(BYTECODE_OP_LOAD_FIELD_REF_DOUBLE
):
2016 struct load_op
*insn
= (struct load_op
*) pc
;
2017 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2019 dbg_printf("load field ref offset %u type double\n",
2021 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2022 memcpy(&estack_ax(stack
, top
)->u
.d
, &interpreter_stack_data
[ref
->offset
],
2023 sizeof(struct literal_double
));
2024 estack_ax_t
= REG_DOUBLE
;
2025 dbg_printf("ref load double %g\n", estack_ax(stack
, top
)->u
.d
);
2026 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2030 /* load from immediate operand */
2031 OP(BYTECODE_OP_LOAD_STRING
):
2033 struct load_op
*insn
= (struct load_op
*) pc
;
2035 dbg_printf("load string %s\n", insn
->data
);
2036 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2037 estack_ax(stack
, top
)->u
.s
.str
= insn
->data
;
2038 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2039 estack_ax(stack
, top
)->u
.s
.literal_type
=
2040 ESTACK_STRING_LITERAL_TYPE_PLAIN
;
2041 estack_ax_t
= REG_STRING
;
2042 next_pc
+= sizeof(struct load_op
) + strlen(insn
->data
) + 1;
2046 OP(BYTECODE_OP_LOAD_STAR_GLOB_STRING
):
2048 struct load_op
*insn
= (struct load_op
*) pc
;
2050 dbg_printf("load globbing pattern %s\n", insn
->data
);
2051 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2052 estack_ax(stack
, top
)->u
.s
.str
= insn
->data
;
2053 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2054 estack_ax(stack
, top
)->u
.s
.literal_type
=
2055 ESTACK_STRING_LITERAL_TYPE_STAR_GLOB
;
2056 estack_ax_t
= REG_STAR_GLOB_STRING
;
2057 next_pc
+= sizeof(struct load_op
) + strlen(insn
->data
) + 1;
2061 OP(BYTECODE_OP_LOAD_S64
):
2063 struct load_op
*insn
= (struct load_op
*) pc
;
2065 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2066 estack_ax_v
= ((struct literal_numeric
*) insn
->data
)->v
;
2067 estack_ax_t
= REG_S64
;
2068 dbg_printf("load s64 %" PRIi64
"\n", estack_ax_v
);
2069 next_pc
+= sizeof(struct load_op
)
2070 + sizeof(struct literal_numeric
);
2074 OP(BYTECODE_OP_LOAD_DOUBLE
):
2076 struct load_op
*insn
= (struct load_op
*) pc
;
2078 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2079 memcpy(&estack_ax(stack
, top
)->u
.d
, insn
->data
,
2080 sizeof(struct literal_double
));
2081 estack_ax_t
= REG_DOUBLE
;
2082 dbg_printf("load double %g\n", estack_ax(stack
, top
)->u
.d
);
2083 next_pc
+= sizeof(struct load_op
)
2084 + sizeof(struct literal_double
);
2089 OP(BYTECODE_OP_CAST_TO_S64
):
2091 /* Dynamic typing. */
2092 switch (estack_ax_t
) {
2094 JUMP_TO(BYTECODE_OP_CAST_NOP
);
2096 JUMP_TO(BYTECODE_OP_CAST_DOUBLE_TO_S64
);
2098 estack_ax_t
= REG_S64
;
2099 next_pc
+= sizeof(struct cast_op
);
2100 case REG_STRING
: /* Fall-through */
2101 case REG_STAR_GLOB_STRING
:
2105 ERR("Unknown interpreter register type (%d)",
2112 OP(BYTECODE_OP_CAST_DOUBLE_TO_S64
):
2114 estack_ax_v
= (int64_t) estack_ax(stack
, top
)->u
.d
;
2115 estack_ax_t
= REG_S64
;
2116 next_pc
+= sizeof(struct cast_op
);
2120 OP(BYTECODE_OP_CAST_NOP
):
2122 next_pc
+= sizeof(struct cast_op
);
2126 /* get context ref */
2127 OP(BYTECODE_OP_GET_CONTEXT_REF
):
2129 struct load_op
*insn
= (struct load_op
*) pc
;
2130 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2131 struct lttng_ust_ctx_field
*ctx_field
;
2132 struct lttng_ust_ctx_value v
;
2134 dbg_printf("get context ref offset %u type dynamic\n",
2136 ctx_field
= ctx
->fields
[ref
->offset
];
2137 ctx_field
->get_value(ctx_field
, &v
);
2138 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2140 case LTTNG_UST_DYNAMIC_TYPE_NONE
:
2143 case LTTNG_UST_DYNAMIC_TYPE_S64
:
2144 estack_ax_v
= v
.u
.s64
;
2145 estack_ax_t
= REG_S64
;
2146 dbg_printf("ref get context dynamic s64 %" PRIi64
"\n", estack_ax_v
);
2148 case LTTNG_UST_DYNAMIC_TYPE_DOUBLE
:
2149 estack_ax(stack
, top
)->u
.d
= v
.u
.d
;
2150 estack_ax_t
= REG_DOUBLE
;
2151 dbg_printf("ref get context dynamic double %g\n", estack_ax(stack
, top
)->u
.d
);
2153 case LTTNG_UST_DYNAMIC_TYPE_STRING
:
2154 estack_ax(stack
, top
)->u
.s
.str
= v
.u
.str
;
2155 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2156 dbg_printf("Interpreter warning: loading a NULL string.\n");
2160 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2161 estack_ax(stack
, top
)->u
.s
.literal_type
=
2162 ESTACK_STRING_LITERAL_TYPE_NONE
;
2163 dbg_printf("ref get context dynamic string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
2164 estack_ax_t
= REG_STRING
;
2167 dbg_printf("Interpreter warning: unknown dynamic type (%d).\n", (int) v
.sel
);
2171 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2175 OP(BYTECODE_OP_GET_CONTEXT_REF_STRING
):
2177 struct load_op
*insn
= (struct load_op
*) pc
;
2178 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2179 struct lttng_ust_ctx_field
*ctx_field
;
2180 struct lttng_ust_ctx_value v
;
2182 dbg_printf("get context ref offset %u type string\n",
2184 ctx_field
= ctx
->fields
[ref
->offset
];
2185 ctx_field
->get_value(ctx_field
, &v
);
2186 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2187 estack_ax(stack
, top
)->u
.s
.str
= v
.u
.str
;
2188 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2189 dbg_printf("Interpreter warning: loading a NULL string.\n");
2193 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2194 estack_ax(stack
, top
)->u
.s
.literal_type
=
2195 ESTACK_STRING_LITERAL_TYPE_NONE
;
2196 estack_ax_t
= REG_STRING
;
2197 dbg_printf("ref get context string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
2198 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2202 OP(BYTECODE_OP_GET_CONTEXT_REF_S64
):
2204 struct load_op
*insn
= (struct load_op
*) pc
;
2205 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2206 struct lttng_ust_ctx_field
*ctx_field
;
2207 struct lttng_ust_ctx_value v
;
2209 dbg_printf("get context ref offset %u type s64\n",
2211 ctx_field
= ctx
->fields
[ref
->offset
];
2212 ctx_field
->get_value(ctx_field
, &v
);
2213 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2214 estack_ax_v
= v
.u
.s64
;
2215 estack_ax_t
= REG_S64
;
2216 dbg_printf("ref get context s64 %" PRIi64
"\n", estack_ax_v
);
2217 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2221 OP(BYTECODE_OP_GET_CONTEXT_REF_DOUBLE
):
2223 struct load_op
*insn
= (struct load_op
*) pc
;
2224 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2225 struct lttng_ust_ctx_field
*ctx_field
;
2226 struct lttng_ust_ctx_value v
;
2228 dbg_printf("get context ref offset %u type double\n",
2230 ctx_field
= ctx
->fields
[ref
->offset
];
2231 ctx_field
->get_value(ctx_field
, &v
);
2232 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2233 memcpy(&estack_ax(stack
, top
)->u
.d
, &v
.u
.d
, sizeof(struct literal_double
));
2234 estack_ax_t
= REG_DOUBLE
;
2235 dbg_printf("ref get context double %g\n", estack_ax(stack
, top
)->u
.d
);
2236 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2240 OP(BYTECODE_OP_GET_CONTEXT_ROOT
):
2242 dbg_printf("op get context root\n");
2243 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2244 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_CONTEXT
;
2245 /* "field" only needed for variants. */
2246 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
2247 estack_ax_t
= REG_PTR
;
2248 next_pc
+= sizeof(struct load_op
);
2252 OP(BYTECODE_OP_GET_APP_CONTEXT_ROOT
):
2254 dbg_printf("op get app context root\n");
2255 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2256 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_APP_CONTEXT
;
2257 /* "field" only needed for variants. */
2258 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
2259 estack_ax_t
= REG_PTR
;
2260 next_pc
+= sizeof(struct load_op
);
2264 OP(BYTECODE_OP_GET_PAYLOAD_ROOT
):
2266 dbg_printf("op get app payload root\n");
2267 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2268 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_PAYLOAD
;
2269 estack_ax(stack
, top
)->u
.ptr
.ptr
= interpreter_stack_data
;
2270 /* "field" only needed for variants. */
2271 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
2272 estack_ax_t
= REG_PTR
;
2273 next_pc
+= sizeof(struct load_op
);
2277 OP(BYTECODE_OP_GET_SYMBOL
):
2279 dbg_printf("op get symbol\n");
2280 switch (estack_ax(stack
, top
)->u
.ptr
.type
) {
2282 ERR("Nested fields not implemented yet.");
2285 case LOAD_ROOT_CONTEXT
:
2286 case LOAD_ROOT_APP_CONTEXT
:
2287 case LOAD_ROOT_PAYLOAD
:
2289 * symbol lookup is performed by
2295 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_symbol
);
2299 OP(BYTECODE_OP_GET_SYMBOL_FIELD
):
2302 * Used for first variant encountered in a
2303 * traversal. Variants are not implemented yet.
2309 OP(BYTECODE_OP_GET_INDEX_U16
):
2311 struct load_op
*insn
= (struct load_op
*) pc
;
2312 struct get_index_u16
*index
= (struct get_index_u16
*) insn
->data
;
2314 dbg_printf("op get index u16\n");
2315 ret
= dynamic_get_index(ctx
, bytecode
, index
->index
, estack_ax(stack
, top
));
2318 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
2319 estack_ax_t
= estack_ax(stack
, top
)->type
;
2320 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_index_u16
);
2324 OP(BYTECODE_OP_GET_INDEX_U64
):
2326 struct load_op
*insn
= (struct load_op
*) pc
;
2327 struct get_index_u64
*index
= (struct get_index_u64
*) insn
->data
;
2329 dbg_printf("op get index u64\n");
2330 ret
= dynamic_get_index(ctx
, bytecode
, index
->index
, estack_ax(stack
, top
));
2333 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
2334 estack_ax_t
= estack_ax(stack
, top
)->type
;
2335 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_index_u64
);
2339 OP(BYTECODE_OP_LOAD_FIELD
):
2341 dbg_printf("op load field\n");
2342 ret
= dynamic_load_field(estack_ax(stack
, top
));
2345 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
2346 estack_ax_t
= estack_ax(stack
, top
)->type
;
2347 next_pc
+= sizeof(struct load_op
);
2351 OP(BYTECODE_OP_LOAD_FIELD_S8
):
2353 dbg_printf("op load field s8\n");
2355 estack_ax_v
= *(int8_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2356 estack_ax_t
= REG_S64
;
2357 next_pc
+= sizeof(struct load_op
);
2360 OP(BYTECODE_OP_LOAD_FIELD_S16
):
2362 dbg_printf("op load field s16\n");
2364 estack_ax_v
= *(int16_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2365 estack_ax_t
= REG_S64
;
2366 next_pc
+= sizeof(struct load_op
);
2369 OP(BYTECODE_OP_LOAD_FIELD_S32
):
2371 dbg_printf("op load field s32\n");
2373 estack_ax_v
= *(int32_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2374 estack_ax_t
= REG_S64
;
2375 next_pc
+= sizeof(struct load_op
);
2378 OP(BYTECODE_OP_LOAD_FIELD_S64
):
2380 dbg_printf("op load field s64\n");
2382 estack_ax_v
= *(int64_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2383 estack_ax_t
= REG_S64
;
2384 next_pc
+= sizeof(struct load_op
);
2387 OP(BYTECODE_OP_LOAD_FIELD_U8
):
2389 dbg_printf("op load field u8\n");
2391 estack_ax_v
= *(uint8_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2392 estack_ax_t
= REG_U64
;
2393 next_pc
+= sizeof(struct load_op
);
2396 OP(BYTECODE_OP_LOAD_FIELD_U16
):
2398 dbg_printf("op load field u16\n");
2400 estack_ax_v
= *(uint16_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2401 estack_ax_t
= REG_U64
;
2402 next_pc
+= sizeof(struct load_op
);
2405 OP(BYTECODE_OP_LOAD_FIELD_U32
):
2407 dbg_printf("op load field u32\n");
2409 estack_ax_v
= *(uint32_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2410 estack_ax_t
= REG_U64
;
2411 next_pc
+= sizeof(struct load_op
);
2414 OP(BYTECODE_OP_LOAD_FIELD_U64
):
2416 dbg_printf("op load field u64\n");
2418 estack_ax_v
= *(uint64_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2419 estack_ax_t
= REG_U64
;
2420 next_pc
+= sizeof(struct load_op
);
2423 OP(BYTECODE_OP_LOAD_FIELD_DOUBLE
):
2425 dbg_printf("op load field double\n");
2427 memcpy(&estack_ax(stack
, top
)->u
.d
,
2428 estack_ax(stack
, top
)->u
.ptr
.ptr
,
2429 sizeof(struct literal_double
));
2430 estack_ax(stack
, top
)->type
= REG_DOUBLE
;
2431 next_pc
+= sizeof(struct load_op
);
2435 OP(BYTECODE_OP_LOAD_FIELD_STRING
):
2439 dbg_printf("op load field string\n");
2440 str
= (const char *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2441 estack_ax(stack
, top
)->u
.s
.str
= str
;
2442 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2443 dbg_printf("Interpreter warning: loading a NULL string.\n");
2447 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2448 estack_ax(stack
, top
)->u
.s
.literal_type
=
2449 ESTACK_STRING_LITERAL_TYPE_NONE
;
2450 estack_ax(stack
, top
)->type
= REG_STRING
;
2451 next_pc
+= sizeof(struct load_op
);
2455 OP(BYTECODE_OP_LOAD_FIELD_SEQUENCE
):
2459 dbg_printf("op load field string sequence\n");
2460 ptr
= estack_ax(stack
, top
)->u
.ptr
.ptr
;
2461 estack_ax(stack
, top
)->u
.s
.seq_len
= *(unsigned long *) ptr
;
2462 estack_ax(stack
, top
)->u
.s
.str
= *(const char **) (ptr
+ sizeof(unsigned long));
2463 estack_ax(stack
, top
)->type
= REG_STRING
;
2464 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2465 dbg_printf("Interpreter warning: loading a NULL sequence.\n");
2469 estack_ax(stack
, top
)->u
.s
.literal_type
=
2470 ESTACK_STRING_LITERAL_TYPE_NONE
;
2471 next_pc
+= sizeof(struct load_op
);
2477 /* No need to prepare output if an error occurred. */
2479 return LTTNG_UST_BYTECODE_INTERPRETER_ERROR
;
2481 /* Prepare output. */
2482 switch (ust_bytecode
->priv
->type
) {
2483 case LTTNG_UST_BYTECODE_TYPE_FILTER
:
2485 struct lttng_ust_bytecode_filter_ctx
*filter_ctx
=
2486 (struct lttng_ust_bytecode_filter_ctx
*) caller_ctx
;
2488 filter_ctx
->result
= LTTNG_UST_BYTECODE_FILTER_ACCEPT
;
2490 filter_ctx
->result
= LTTNG_UST_BYTECODE_FILTER_REJECT
;
2493 case LTTNG_UST_BYTECODE_TYPE_CAPTURE
:
2494 ret
= lttng_bytecode_interpret_format_output(estack_ax(stack
, top
),
2495 (struct lttng_interpreter_output
*) caller_ctx
);
2502 return LTTNG_UST_BYTECODE_INTERPRETER_ERROR
;
2504 return LTTNG_UST_BYTECODE_INTERPRETER_OK
;