2 * SPDX-License-Identifier: MIT
4 * Copyright (C) 2010-2016 Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
6 * LTTng UST bytecode interpreter.
13 #include <lttng/urcu/pointer.h>
14 #include <urcu/rculist.h>
15 #include <lttng/ust-endian.h>
16 #include <lttng/ust-events.h>
17 #include "lib/lttng-ust/events.h"
19 #include "lttng-bytecode.h"
20 #include "common/strutils.h"
25 * -2: unknown escape char.
30 int parse_char(const char **p
)
50 * Returns SIZE_MAX if the string is null-terminated, or the number of
54 size_t get_str_or_seq_len(const struct estack_entry
*entry
)
56 return entry
->u
.s
.seq_len
;
60 int stack_star_glob_match(struct estack
*stack
, int top
,
61 const char *cmp_type
__attribute__((unused
)))
64 const char *candidate
;
68 /* Find out which side is the pattern vs. the candidate. */
69 if (estack_ax(stack
, top
)->u
.s
.literal_type
== ESTACK_STRING_LITERAL_TYPE_STAR_GLOB
) {
70 pattern
= estack_ax(stack
, top
)->u
.s
.str
;
71 pattern_len
= get_str_or_seq_len(estack_ax(stack
, top
));
72 candidate
= estack_bx(stack
, top
)->u
.s
.str
;
73 candidate_len
= get_str_or_seq_len(estack_bx(stack
, top
));
75 pattern
= estack_bx(stack
, top
)->u
.s
.str
;
76 pattern_len
= get_str_or_seq_len(estack_bx(stack
, top
));
77 candidate
= estack_ax(stack
, top
)->u
.s
.str
;
78 candidate_len
= get_str_or_seq_len(estack_ax(stack
, top
));
81 /* Perform the match. Returns 0 when the result is true. */
82 return !strutils_star_glob_match(pattern
, pattern_len
, candidate
,
87 int stack_strcmp(struct estack
*stack
, int top
, const char *cmp_type
__attribute__((unused
)))
89 const char *p
= estack_bx(stack
, top
)->u
.s
.str
, *q
= estack_ax(stack
, top
)->u
.s
.str
;
96 if (unlikely(p
- estack_bx(stack
, top
)->u
.s
.str
>= estack_bx(stack
, top
)->u
.s
.seq_len
|| *p
== '\0')) {
97 if (q
- estack_ax(stack
, top
)->u
.s
.str
>= estack_ax(stack
, top
)->u
.s
.seq_len
|| *q
== '\0') {
100 if (estack_ax(stack
, top
)->u
.s
.literal_type
==
101 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
102 ret
= parse_char(&q
);
109 if (unlikely(q
- estack_ax(stack
, top
)->u
.s
.str
>= estack_ax(stack
, top
)->u
.s
.seq_len
|| *q
== '\0')) {
110 if (estack_bx(stack
, top
)->u
.s
.literal_type
==
111 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
112 ret
= parse_char(&p
);
118 if (estack_bx(stack
, top
)->u
.s
.literal_type
==
119 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
120 ret
= parse_char(&p
);
123 } else if (ret
== -2) {
126 /* else compare both char */
128 if (estack_ax(stack
, top
)->u
.s
.literal_type
==
129 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
130 ret
= parse_char(&q
);
133 } else if (ret
== -2) {
153 int lttng_bytecode_interpret_error(
154 struct lttng_ust_bytecode_runtime
*bytecode_runtime
__attribute__((unused
)),
155 const char *stack_data
__attribute__((unused
)),
156 struct lttng_ust_probe_ctx
*probe_ctx
__attribute__((unused
)),
157 void *ctx
__attribute__((unused
)))
159 return LTTNG_UST_BYTECODE_INTERPRETER_ERROR
;
162 #ifdef INTERPRETER_USE_SWITCH
165 * Fallback for compilers that do not support taking address of labels.
169 start_pc = &bytecode->data[0]; \
170 for (pc = next_pc = start_pc; pc - start_pc < bytecode->len; \
172 dbg_printf("Executing op %s (%u)\n", \
173 lttng_bytecode_print_op((unsigned int) *(bytecode_opcode_t *) pc), \
174 (unsigned int) *(bytecode_opcode_t *) pc); \
175 switch (*(bytecode_opcode_t *) pc) {
177 #define OP(name) jump_target_##name: __attribute__((unused)); \
185 #define JUMP_TO(name) \
186 goto jump_target_##name
191 * Dispatch-table based interpreter.
195 start_pc = &bytecode->code[0]; \
196 pc = next_pc = start_pc; \
197 if (unlikely(pc - start_pc >= bytecode->len)) \
199 goto *dispatch[*(bytecode_opcode_t *) pc];
206 goto *dispatch[*(bytecode_opcode_t *) pc];
210 #define JUMP_TO(name) \
215 #define IS_INTEGER_REGISTER(reg_type) \
216 (reg_type == REG_U64 || reg_type == REG_S64)
218 static int context_get_index(struct lttng_ust_ctx
*ctx
,
219 struct lttng_ust_probe_ctx
*probe_ctx
,
220 struct load_ptr
*ptr
,
224 const struct lttng_ust_ctx_field
*ctx_field
;
225 const struct lttng_ust_event_field
*field
;
226 struct lttng_ust_ctx_value v
;
228 ctx_field
= &ctx
->fields
[idx
];
229 field
= ctx_field
->event_field
;
230 ptr
->type
= LOAD_OBJECT
;
233 switch (field
->type
->type
) {
234 case lttng_ust_type_integer
:
235 ctx_field
->get_value(ctx_field
->priv
, probe_ctx
, &v
);
236 if (lttng_ust_get_type_integer(field
->type
)->signedness
) {
237 ptr
->object_type
= OBJECT_TYPE_S64
;
238 ptr
->u
.s64
= v
.u
.s64
;
239 ptr
->ptr
= &ptr
->u
.s64
;
241 ptr
->object_type
= OBJECT_TYPE_U64
;
242 ptr
->u
.u64
= v
.u
.s64
; /* Cast. */
243 ptr
->ptr
= &ptr
->u
.u64
;
246 case lttng_ust_type_enum
:
248 const struct lttng_ust_type_integer
*itype
;
250 itype
= lttng_ust_get_type_integer(lttng_ust_get_type_enum(field
->type
)->container_type
);
251 ctx_field
->get_value(ctx_field
->priv
, probe_ctx
, &v
);
252 if (itype
->signedness
) {
253 ptr
->object_type
= OBJECT_TYPE_SIGNED_ENUM
;
254 ptr
->u
.s64
= v
.u
.s64
;
255 ptr
->ptr
= &ptr
->u
.s64
;
257 ptr
->object_type
= OBJECT_TYPE_UNSIGNED_ENUM
;
258 ptr
->u
.u64
= v
.u
.s64
; /* Cast. */
259 ptr
->ptr
= &ptr
->u
.u64
;
263 case lttng_ust_type_array
:
264 if (lttng_ust_get_type_array(field
->type
)->elem_type
->type
!= lttng_ust_type_integer
) {
265 ERR("Array nesting only supports integer types.");
268 if (lttng_ust_get_type_array(field
->type
)->encoding
== lttng_ust_string_encoding_none
) {
269 ERR("Only string arrays are supported for contexts.");
272 ptr
->object_type
= OBJECT_TYPE_STRING
;
273 ctx_field
->get_value(ctx_field
->priv
, probe_ctx
, &v
);
276 case lttng_ust_type_sequence
:
277 if (lttng_ust_get_type_sequence(field
->type
)->elem_type
->type
!= lttng_ust_type_integer
) {
278 ERR("Sequence nesting only supports integer types.");
281 if (lttng_ust_get_type_sequence(field
->type
)->encoding
== lttng_ust_string_encoding_none
) {
282 ERR("Only string sequences are supported for contexts.");
285 ptr
->object_type
= OBJECT_TYPE_STRING
;
286 ctx_field
->get_value(ctx_field
->priv
, probe_ctx
, &v
);
289 case lttng_ust_type_string
:
290 ptr
->object_type
= OBJECT_TYPE_STRING
;
291 ctx_field
->get_value(ctx_field
->priv
, probe_ctx
, &v
);
294 case lttng_ust_type_float
:
295 ptr
->object_type
= OBJECT_TYPE_DOUBLE
;
296 ctx_field
->get_value(ctx_field
->priv
, probe_ctx
, &v
);
298 ptr
->ptr
= &ptr
->u
.d
;
300 case lttng_ust_type_dynamic
:
301 ctx_field
->get_value(ctx_field
->priv
, probe_ctx
, &v
);
303 case LTTNG_UST_DYNAMIC_TYPE_NONE
:
305 case LTTNG_UST_DYNAMIC_TYPE_U8
:
306 case LTTNG_UST_DYNAMIC_TYPE_U16
:
307 case LTTNG_UST_DYNAMIC_TYPE_U32
:
308 case LTTNG_UST_DYNAMIC_TYPE_U64
:
309 ptr
->object_type
= OBJECT_TYPE_U64
;
310 ptr
->u
.u64
= v
.u
.u64
;
311 ptr
->ptr
= &ptr
->u
.u64
;
312 dbg_printf("context get index dynamic u64 %" PRIi64
"\n", ptr
->u
.u64
);
314 case LTTNG_UST_DYNAMIC_TYPE_S8
:
315 case LTTNG_UST_DYNAMIC_TYPE_S16
:
316 case LTTNG_UST_DYNAMIC_TYPE_S32
:
317 case LTTNG_UST_DYNAMIC_TYPE_S64
:
318 ptr
->object_type
= OBJECT_TYPE_S64
;
319 ptr
->u
.s64
= v
.u
.s64
;
320 ptr
->ptr
= &ptr
->u
.s64
;
321 dbg_printf("context get index dynamic s64 %" PRIi64
"\n", ptr
->u
.s64
);
323 case LTTNG_UST_DYNAMIC_TYPE_FLOAT
:
324 case LTTNG_UST_DYNAMIC_TYPE_DOUBLE
:
325 ptr
->object_type
= OBJECT_TYPE_DOUBLE
;
327 ptr
->ptr
= &ptr
->u
.d
;
328 dbg_printf("context get index dynamic double %g\n", ptr
->u
.d
);
330 case LTTNG_UST_DYNAMIC_TYPE_STRING
:
331 ptr
->object_type
= OBJECT_TYPE_STRING
;
333 dbg_printf("context get index dynamic string %s\n", (const char *) ptr
->ptr
);
336 dbg_printf("Interpreter warning: unknown dynamic type (%d).\n", (int) v
.sel
);
341 ERR("Unknown type: %d", (int) field
->type
->type
);
347 static int dynamic_get_index(struct lttng_ust_ctx
*ctx
,
348 struct lttng_ust_probe_ctx
*probe_ctx
,
349 struct bytecode_runtime
*runtime
,
350 uint64_t index
, struct estack_entry
*stack_top
)
353 const struct bytecode_get_index_data
*gid
;
355 gid
= (const struct bytecode_get_index_data
*) &runtime
->data
[index
];
356 switch (stack_top
->u
.ptr
.type
) {
358 switch (stack_top
->u
.ptr
.object_type
) {
359 case OBJECT_TYPE_ARRAY
:
363 assert(gid
->offset
< gid
->array_len
);
364 /* Skip count (unsigned long) */
365 ptr
= *(const char **) (stack_top
->u
.ptr
.ptr
+ sizeof(unsigned long));
366 ptr
= ptr
+ gid
->offset
;
367 stack_top
->u
.ptr
.ptr
= ptr
;
368 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
369 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
370 assert(stack_top
->u
.ptr
.field
->type
->type
== lttng_ust_type_array
);
371 stack_top
->u
.ptr
.field
= NULL
;
374 case OBJECT_TYPE_SEQUENCE
:
379 ptr
= *(const char **) (stack_top
->u
.ptr
.ptr
+ sizeof(unsigned long));
380 ptr_seq_len
= *(unsigned long *) stack_top
->u
.ptr
.ptr
;
381 if (gid
->offset
>= gid
->elem
.len
* ptr_seq_len
) {
385 ptr
= ptr
+ gid
->offset
;
386 stack_top
->u
.ptr
.ptr
= ptr
;
387 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
388 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
389 assert(stack_top
->u
.ptr
.field
->type
->type
== lttng_ust_type_sequence
);
390 stack_top
->u
.ptr
.field
= NULL
;
393 case OBJECT_TYPE_STRUCT
:
394 ERR("Nested structures are not supported yet.");
397 case OBJECT_TYPE_VARIANT
:
399 ERR("Unexpected get index type %d",
400 (int) stack_top
->u
.ptr
.object_type
);
405 case LOAD_ROOT_CONTEXT
:
406 case LOAD_ROOT_APP_CONTEXT
: /* Fall-through */
408 ret
= context_get_index(ctx
,
417 case LOAD_ROOT_PAYLOAD
:
418 stack_top
->u
.ptr
.ptr
+= gid
->offset
;
419 if (gid
->elem
.type
== OBJECT_TYPE_STRING
)
420 stack_top
->u
.ptr
.ptr
= *(const char * const *) stack_top
->u
.ptr
.ptr
;
421 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
422 stack_top
->u
.ptr
.type
= LOAD_OBJECT
;
423 stack_top
->u
.ptr
.field
= gid
->field
;
424 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
428 stack_top
->type
= REG_PTR
;
436 static int dynamic_load_field(struct estack_entry
*stack_top
)
440 switch (stack_top
->u
.ptr
.type
) {
443 case LOAD_ROOT_CONTEXT
:
444 case LOAD_ROOT_APP_CONTEXT
:
445 case LOAD_ROOT_PAYLOAD
:
447 dbg_printf("Interpreter warning: cannot load root, missing field name.\n");
451 switch (stack_top
->u
.ptr
.object_type
) {
453 dbg_printf("op load field s8\n");
454 stack_top
->u
.v
= *(int8_t *) stack_top
->u
.ptr
.ptr
;
455 stack_top
->type
= REG_S64
;
457 case OBJECT_TYPE_S16
:
461 dbg_printf("op load field s16\n");
462 tmp
= *(int16_t *) stack_top
->u
.ptr
.ptr
;
463 if (stack_top
->u
.ptr
.rev_bo
)
464 tmp
= lttng_ust_bswap_16(tmp
);
465 stack_top
->u
.v
= tmp
;
466 stack_top
->type
= REG_S64
;
469 case OBJECT_TYPE_S32
:
473 dbg_printf("op load field s32\n");
474 tmp
= *(int32_t *) stack_top
->u
.ptr
.ptr
;
475 if (stack_top
->u
.ptr
.rev_bo
)
476 tmp
= lttng_ust_bswap_32(tmp
);
477 stack_top
->u
.v
= tmp
;
478 stack_top
->type
= REG_S64
;
481 case OBJECT_TYPE_S64
:
485 dbg_printf("op load field s64\n");
486 tmp
= *(int64_t *) stack_top
->u
.ptr
.ptr
;
487 if (stack_top
->u
.ptr
.rev_bo
)
488 tmp
= lttng_ust_bswap_64(tmp
);
489 stack_top
->u
.v
= tmp
;
490 stack_top
->type
= REG_S64
;
493 case OBJECT_TYPE_SIGNED_ENUM
:
497 dbg_printf("op load field signed enumeration\n");
498 tmp
= *(int64_t *) stack_top
->u
.ptr
.ptr
;
499 if (stack_top
->u
.ptr
.rev_bo
)
500 tmp
= lttng_ust_bswap_64(tmp
);
501 stack_top
->u
.v
= tmp
;
502 stack_top
->type
= REG_S64
;
506 dbg_printf("op load field u8\n");
507 stack_top
->u
.v
= *(uint8_t *) stack_top
->u
.ptr
.ptr
;
508 stack_top
->type
= REG_U64
;
510 case OBJECT_TYPE_U16
:
514 dbg_printf("op load field u16\n");
515 tmp
= *(uint16_t *) stack_top
->u
.ptr
.ptr
;
516 if (stack_top
->u
.ptr
.rev_bo
)
517 tmp
= lttng_ust_bswap_16(tmp
);
518 stack_top
->u
.v
= tmp
;
519 stack_top
->type
= REG_U64
;
522 case OBJECT_TYPE_U32
:
526 dbg_printf("op load field u32\n");
527 tmp
= *(uint32_t *) stack_top
->u
.ptr
.ptr
;
528 if (stack_top
->u
.ptr
.rev_bo
)
529 tmp
= lttng_ust_bswap_32(tmp
);
530 stack_top
->u
.v
= tmp
;
531 stack_top
->type
= REG_U64
;
534 case OBJECT_TYPE_U64
:
538 dbg_printf("op load field u64\n");
539 tmp
= *(uint64_t *) stack_top
->u
.ptr
.ptr
;
540 if (stack_top
->u
.ptr
.rev_bo
)
541 tmp
= lttng_ust_bswap_64(tmp
);
542 stack_top
->u
.v
= tmp
;
543 stack_top
->type
= REG_U64
;
546 case OBJECT_TYPE_UNSIGNED_ENUM
:
550 dbg_printf("op load field unsigned enumeration\n");
551 tmp
= *(uint64_t *) stack_top
->u
.ptr
.ptr
;
552 if (stack_top
->u
.ptr
.rev_bo
)
553 tmp
= lttng_ust_bswap_64(tmp
);
554 stack_top
->u
.v
= tmp
;
555 stack_top
->type
= REG_U64
;
558 case OBJECT_TYPE_DOUBLE
:
559 memcpy(&stack_top
->u
.d
,
560 stack_top
->u
.ptr
.ptr
,
561 sizeof(struct literal_double
));
562 stack_top
->type
= REG_DOUBLE
;
564 case OBJECT_TYPE_STRING
:
568 dbg_printf("op load field string\n");
569 str
= (const char *) stack_top
->u
.ptr
.ptr
;
570 stack_top
->u
.s
.str
= str
;
571 if (unlikely(!stack_top
->u
.s
.str
)) {
572 dbg_printf("Interpreter warning: loading a NULL string.\n");
576 stack_top
->u
.s
.seq_len
= SIZE_MAX
;
577 stack_top
->u
.s
.literal_type
=
578 ESTACK_STRING_LITERAL_TYPE_NONE
;
579 stack_top
->type
= REG_STRING
;
582 case OBJECT_TYPE_STRING_SEQUENCE
:
586 dbg_printf("op load field string sequence\n");
587 ptr
= stack_top
->u
.ptr
.ptr
;
588 stack_top
->u
.s
.seq_len
= *(unsigned long *) ptr
;
589 stack_top
->u
.s
.str
= *(const char **) (ptr
+ sizeof(unsigned long));
590 stack_top
->type
= REG_STRING
;
591 if (unlikely(!stack_top
->u
.s
.str
)) {
592 dbg_printf("Interpreter warning: loading a NULL sequence.\n");
596 stack_top
->u
.s
.literal_type
=
597 ESTACK_STRING_LITERAL_TYPE_NONE
;
600 case OBJECT_TYPE_DYNAMIC
:
602 * Dynamic types in context are looked up
603 * by context get index.
607 case OBJECT_TYPE_SEQUENCE
:
608 case OBJECT_TYPE_ARRAY
:
609 case OBJECT_TYPE_STRUCT
:
610 case OBJECT_TYPE_VARIANT
:
611 ERR("Sequences, arrays, struct and variant cannot be loaded (nested types).");
622 int lttng_bytecode_interpret_format_output(struct estack_entry
*ax
,
623 struct lttng_interpreter_output
*output
)
630 output
->type
= LTTNG_INTERPRETER_TYPE_S64
;
631 output
->u
.s
= ax
->u
.v
;
634 output
->type
= LTTNG_INTERPRETER_TYPE_U64
;
635 output
->u
.u
= (uint64_t) ax
->u
.v
;
638 output
->type
= LTTNG_INTERPRETER_TYPE_DOUBLE
;
639 output
->u
.d
= ax
->u
.d
;
642 output
->type
= LTTNG_INTERPRETER_TYPE_STRING
;
643 output
->u
.str
.str
= ax
->u
.s
.str
;
644 output
->u
.str
.len
= ax
->u
.s
.seq_len
;
647 switch (ax
->u
.ptr
.object_type
) {
649 case OBJECT_TYPE_S16
:
650 case OBJECT_TYPE_S32
:
651 case OBJECT_TYPE_S64
:
653 case OBJECT_TYPE_U16
:
654 case OBJECT_TYPE_U32
:
655 case OBJECT_TYPE_U64
:
656 case OBJECT_TYPE_DOUBLE
:
657 case OBJECT_TYPE_STRING
:
658 case OBJECT_TYPE_STRING_SEQUENCE
:
659 ret
= dynamic_load_field(ax
);
662 /* Retry after loading ptr into stack top. */
664 case OBJECT_TYPE_SEQUENCE
:
665 output
->type
= LTTNG_INTERPRETER_TYPE_SEQUENCE
;
666 output
->u
.sequence
.ptr
= *(const char **) (ax
->u
.ptr
.ptr
+ sizeof(unsigned long));
667 output
->u
.sequence
.nr_elem
= *(unsigned long *) ax
->u
.ptr
.ptr
;
668 output
->u
.sequence
.nested_type
= lttng_ust_get_type_sequence(ax
->u
.ptr
.field
->type
)->elem_type
;
670 case OBJECT_TYPE_ARRAY
:
671 /* Skip count (unsigned long) */
672 output
->type
= LTTNG_INTERPRETER_TYPE_SEQUENCE
;
673 output
->u
.sequence
.ptr
= *(const char **) (ax
->u
.ptr
.ptr
+ sizeof(unsigned long));
674 output
->u
.sequence
.nr_elem
= lttng_ust_get_type_array(ax
->u
.ptr
.field
->type
)->length
;
675 output
->u
.sequence
.nested_type
= lttng_ust_get_type_array(ax
->u
.ptr
.field
->type
)->elem_type
;
677 case OBJECT_TYPE_SIGNED_ENUM
:
678 ret
= dynamic_load_field(ax
);
681 output
->type
= LTTNG_INTERPRETER_TYPE_SIGNED_ENUM
;
682 output
->u
.s
= ax
->u
.v
;
684 case OBJECT_TYPE_UNSIGNED_ENUM
:
685 ret
= dynamic_load_field(ax
);
688 output
->type
= LTTNG_INTERPRETER_TYPE_UNSIGNED_ENUM
;
689 output
->u
.u
= ax
->u
.v
;
691 case OBJECT_TYPE_STRUCT
:
692 case OBJECT_TYPE_VARIANT
:
698 case REG_STAR_GLOB_STRING
:
708 * Return LTTNG_UST_BYTECODE_INTERPRETER_OK on success.
709 * Return LTTNG_UST_BYTECODE_INTERPRETER_ERROR on error.
711 * For FILTER bytecode: expect a struct lttng_ust_bytecode_filter_ctx *
713 * For CAPTURE bytecode: expect a struct lttng_interpreter_output *
716 int lttng_bytecode_interpret(struct lttng_ust_bytecode_runtime
*ust_bytecode
,
717 const char *interpreter_stack_data
,
718 struct lttng_ust_probe_ctx
*probe_ctx
,
721 struct bytecode_runtime
*bytecode
= caa_container_of(ust_bytecode
, struct bytecode_runtime
, p
);
722 struct lttng_ust_ctx
*ctx
= lttng_ust_rcu_dereference(*ust_bytecode
->pctx
);
723 void *pc
, *next_pc
, *start_pc
;
724 int ret
= -EINVAL
, retval
= 0;
725 struct estack _stack
;
726 struct estack
*stack
= &_stack
;
727 register int64_t ax
= 0, bx
= 0;
728 register enum entry_type ax_t
= REG_UNKNOWN
, bx_t
= REG_UNKNOWN
;
729 register int top
= INTERPRETER_STACK_EMPTY
;
730 #ifndef INTERPRETER_USE_SWITCH
731 static void *dispatch
[NR_BYTECODE_OPS
] = {
732 [ BYTECODE_OP_UNKNOWN
] = &&LABEL_BYTECODE_OP_UNKNOWN
,
734 [ BYTECODE_OP_RETURN
] = &&LABEL_BYTECODE_OP_RETURN
,
737 [ BYTECODE_OP_MUL
] = &&LABEL_BYTECODE_OP_MUL
,
738 [ BYTECODE_OP_DIV
] = &&LABEL_BYTECODE_OP_DIV
,
739 [ BYTECODE_OP_MOD
] = &&LABEL_BYTECODE_OP_MOD
,
740 [ BYTECODE_OP_PLUS
] = &&LABEL_BYTECODE_OP_PLUS
,
741 [ BYTECODE_OP_MINUS
] = &&LABEL_BYTECODE_OP_MINUS
,
742 [ BYTECODE_OP_BIT_RSHIFT
] = &&LABEL_BYTECODE_OP_BIT_RSHIFT
,
743 [ BYTECODE_OP_BIT_LSHIFT
] = &&LABEL_BYTECODE_OP_BIT_LSHIFT
,
744 [ BYTECODE_OP_BIT_AND
] = &&LABEL_BYTECODE_OP_BIT_AND
,
745 [ BYTECODE_OP_BIT_OR
] = &&LABEL_BYTECODE_OP_BIT_OR
,
746 [ BYTECODE_OP_BIT_XOR
] = &&LABEL_BYTECODE_OP_BIT_XOR
,
748 /* binary comparators */
749 [ BYTECODE_OP_EQ
] = &&LABEL_BYTECODE_OP_EQ
,
750 [ BYTECODE_OP_NE
] = &&LABEL_BYTECODE_OP_NE
,
751 [ BYTECODE_OP_GT
] = &&LABEL_BYTECODE_OP_GT
,
752 [ BYTECODE_OP_LT
] = &&LABEL_BYTECODE_OP_LT
,
753 [ BYTECODE_OP_GE
] = &&LABEL_BYTECODE_OP_GE
,
754 [ BYTECODE_OP_LE
] = &&LABEL_BYTECODE_OP_LE
,
756 /* string binary comparator */
757 [ BYTECODE_OP_EQ_STRING
] = &&LABEL_BYTECODE_OP_EQ_STRING
,
758 [ BYTECODE_OP_NE_STRING
] = &&LABEL_BYTECODE_OP_NE_STRING
,
759 [ BYTECODE_OP_GT_STRING
] = &&LABEL_BYTECODE_OP_GT_STRING
,
760 [ BYTECODE_OP_LT_STRING
] = &&LABEL_BYTECODE_OP_LT_STRING
,
761 [ BYTECODE_OP_GE_STRING
] = &&LABEL_BYTECODE_OP_GE_STRING
,
762 [ BYTECODE_OP_LE_STRING
] = &&LABEL_BYTECODE_OP_LE_STRING
,
764 /* globbing pattern binary comparator */
765 [ BYTECODE_OP_EQ_STAR_GLOB_STRING
] = &&LABEL_BYTECODE_OP_EQ_STAR_GLOB_STRING
,
766 [ BYTECODE_OP_NE_STAR_GLOB_STRING
] = &&LABEL_BYTECODE_OP_NE_STAR_GLOB_STRING
,
768 /* s64 binary comparator */
769 [ BYTECODE_OP_EQ_S64
] = &&LABEL_BYTECODE_OP_EQ_S64
,
770 [ BYTECODE_OP_NE_S64
] = &&LABEL_BYTECODE_OP_NE_S64
,
771 [ BYTECODE_OP_GT_S64
] = &&LABEL_BYTECODE_OP_GT_S64
,
772 [ BYTECODE_OP_LT_S64
] = &&LABEL_BYTECODE_OP_LT_S64
,
773 [ BYTECODE_OP_GE_S64
] = &&LABEL_BYTECODE_OP_GE_S64
,
774 [ BYTECODE_OP_LE_S64
] = &&LABEL_BYTECODE_OP_LE_S64
,
776 /* double binary comparator */
777 [ BYTECODE_OP_EQ_DOUBLE
] = &&LABEL_BYTECODE_OP_EQ_DOUBLE
,
778 [ BYTECODE_OP_NE_DOUBLE
] = &&LABEL_BYTECODE_OP_NE_DOUBLE
,
779 [ BYTECODE_OP_GT_DOUBLE
] = &&LABEL_BYTECODE_OP_GT_DOUBLE
,
780 [ BYTECODE_OP_LT_DOUBLE
] = &&LABEL_BYTECODE_OP_LT_DOUBLE
,
781 [ BYTECODE_OP_GE_DOUBLE
] = &&LABEL_BYTECODE_OP_GE_DOUBLE
,
782 [ BYTECODE_OP_LE_DOUBLE
] = &&LABEL_BYTECODE_OP_LE_DOUBLE
,
784 /* Mixed S64-double binary comparators */
785 [ BYTECODE_OP_EQ_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_EQ_DOUBLE_S64
,
786 [ BYTECODE_OP_NE_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_NE_DOUBLE_S64
,
787 [ BYTECODE_OP_GT_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_GT_DOUBLE_S64
,
788 [ BYTECODE_OP_LT_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_LT_DOUBLE_S64
,
789 [ BYTECODE_OP_GE_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_GE_DOUBLE_S64
,
790 [ BYTECODE_OP_LE_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_LE_DOUBLE_S64
,
792 [ BYTECODE_OP_EQ_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_EQ_S64_DOUBLE
,
793 [ BYTECODE_OP_NE_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_NE_S64_DOUBLE
,
794 [ BYTECODE_OP_GT_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_GT_S64_DOUBLE
,
795 [ BYTECODE_OP_LT_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_LT_S64_DOUBLE
,
796 [ BYTECODE_OP_GE_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_GE_S64_DOUBLE
,
797 [ BYTECODE_OP_LE_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_LE_S64_DOUBLE
,
800 [ BYTECODE_OP_UNARY_PLUS
] = &&LABEL_BYTECODE_OP_UNARY_PLUS
,
801 [ BYTECODE_OP_UNARY_MINUS
] = &&LABEL_BYTECODE_OP_UNARY_MINUS
,
802 [ BYTECODE_OP_UNARY_NOT
] = &&LABEL_BYTECODE_OP_UNARY_NOT
,
803 [ BYTECODE_OP_UNARY_PLUS_S64
] = &&LABEL_BYTECODE_OP_UNARY_PLUS_S64
,
804 [ BYTECODE_OP_UNARY_MINUS_S64
] = &&LABEL_BYTECODE_OP_UNARY_MINUS_S64
,
805 [ BYTECODE_OP_UNARY_NOT_S64
] = &&LABEL_BYTECODE_OP_UNARY_NOT_S64
,
806 [ BYTECODE_OP_UNARY_PLUS_DOUBLE
] = &&LABEL_BYTECODE_OP_UNARY_PLUS_DOUBLE
,
807 [ BYTECODE_OP_UNARY_MINUS_DOUBLE
] = &&LABEL_BYTECODE_OP_UNARY_MINUS_DOUBLE
,
808 [ BYTECODE_OP_UNARY_NOT_DOUBLE
] = &&LABEL_BYTECODE_OP_UNARY_NOT_DOUBLE
,
811 [ BYTECODE_OP_AND
] = &&LABEL_BYTECODE_OP_AND
,
812 [ BYTECODE_OP_OR
] = &&LABEL_BYTECODE_OP_OR
,
815 [ BYTECODE_OP_LOAD_FIELD_REF
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF
,
816 [ BYTECODE_OP_LOAD_FIELD_REF_STRING
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_STRING
,
817 [ BYTECODE_OP_LOAD_FIELD_REF_SEQUENCE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_SEQUENCE
,
818 [ BYTECODE_OP_LOAD_FIELD_REF_S64
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_S64
,
819 [ BYTECODE_OP_LOAD_FIELD_REF_DOUBLE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_DOUBLE
,
821 /* load from immediate operand */
822 [ BYTECODE_OP_LOAD_STRING
] = &&LABEL_BYTECODE_OP_LOAD_STRING
,
823 [ BYTECODE_OP_LOAD_STAR_GLOB_STRING
] = &&LABEL_BYTECODE_OP_LOAD_STAR_GLOB_STRING
,
824 [ BYTECODE_OP_LOAD_S64
] = &&LABEL_BYTECODE_OP_LOAD_S64
,
825 [ BYTECODE_OP_LOAD_DOUBLE
] = &&LABEL_BYTECODE_OP_LOAD_DOUBLE
,
828 [ BYTECODE_OP_CAST_TO_S64
] = &&LABEL_BYTECODE_OP_CAST_TO_S64
,
829 [ BYTECODE_OP_CAST_DOUBLE_TO_S64
] = &&LABEL_BYTECODE_OP_CAST_DOUBLE_TO_S64
,
830 [ BYTECODE_OP_CAST_NOP
] = &&LABEL_BYTECODE_OP_CAST_NOP
,
832 /* get context ref */
833 [ BYTECODE_OP_GET_CONTEXT_REF
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF
,
834 [ BYTECODE_OP_GET_CONTEXT_REF_STRING
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF_STRING
,
835 [ BYTECODE_OP_GET_CONTEXT_REF_S64
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF_S64
,
836 [ BYTECODE_OP_GET_CONTEXT_REF_DOUBLE
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF_DOUBLE
,
838 /* Instructions for recursive traversal through composed types. */
839 [ BYTECODE_OP_GET_CONTEXT_ROOT
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_ROOT
,
840 [ BYTECODE_OP_GET_APP_CONTEXT_ROOT
] = &&LABEL_BYTECODE_OP_GET_APP_CONTEXT_ROOT
,
841 [ BYTECODE_OP_GET_PAYLOAD_ROOT
] = &&LABEL_BYTECODE_OP_GET_PAYLOAD_ROOT
,
843 [ BYTECODE_OP_GET_SYMBOL
] = &&LABEL_BYTECODE_OP_GET_SYMBOL
,
844 [ BYTECODE_OP_GET_SYMBOL_FIELD
] = &&LABEL_BYTECODE_OP_GET_SYMBOL_FIELD
,
845 [ BYTECODE_OP_GET_INDEX_U16
] = &&LABEL_BYTECODE_OP_GET_INDEX_U16
,
846 [ BYTECODE_OP_GET_INDEX_U64
] = &&LABEL_BYTECODE_OP_GET_INDEX_U64
,
848 [ BYTECODE_OP_LOAD_FIELD
] = &&LABEL_BYTECODE_OP_LOAD_FIELD
,
849 [ BYTECODE_OP_LOAD_FIELD_S8
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S8
,
850 [ BYTECODE_OP_LOAD_FIELD_S16
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S16
,
851 [ BYTECODE_OP_LOAD_FIELD_S32
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S32
,
852 [ BYTECODE_OP_LOAD_FIELD_S64
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S64
,
853 [ BYTECODE_OP_LOAD_FIELD_U8
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U8
,
854 [ BYTECODE_OP_LOAD_FIELD_U16
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U16
,
855 [ BYTECODE_OP_LOAD_FIELD_U32
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U32
,
856 [ BYTECODE_OP_LOAD_FIELD_U64
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U64
,
857 [ BYTECODE_OP_LOAD_FIELD_STRING
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_STRING
,
858 [ BYTECODE_OP_LOAD_FIELD_SEQUENCE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_SEQUENCE
,
859 [ BYTECODE_OP_LOAD_FIELD_DOUBLE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_DOUBLE
,
861 [ BYTECODE_OP_UNARY_BIT_NOT
] = &&LABEL_BYTECODE_OP_UNARY_BIT_NOT
,
863 [ BYTECODE_OP_RETURN_S64
] = &&LABEL_BYTECODE_OP_RETURN_S64
,
865 #endif /* #ifndef INTERPRETER_USE_SWITCH */
869 OP(BYTECODE_OP_UNKNOWN
):
870 OP(BYTECODE_OP_LOAD_FIELD_REF
):
871 #ifdef INTERPRETER_USE_SWITCH
873 #endif /* INTERPRETER_USE_SWITCH */
874 ERR("unknown bytecode op %u",
875 (unsigned int) *(bytecode_opcode_t
*) pc
);
879 OP(BYTECODE_OP_RETURN
):
880 /* LTTNG_UST_BYTECODE_INTERPRETER_ERROR or LTTNG_UST_BYTECODE_INTERPRETER_OK */
881 /* Handle dynamic typing. */
882 switch (estack_ax_t
) {
885 retval
= !!estack_ax_v
;
890 if (ust_bytecode
->type
!= LTTNG_UST_BYTECODE_TYPE_CAPTURE
) {
896 case REG_STAR_GLOB_STRING
:
905 OP(BYTECODE_OP_RETURN_S64
):
906 /* LTTNG_UST_BYTECODE_INTERPRETER_ERROR or LTTNG_UST_BYTECODE_INTERPRETER_OK */
907 retval
= !!estack_ax_v
;
915 OP(BYTECODE_OP_PLUS
):
916 OP(BYTECODE_OP_MINUS
):
917 ERR("unsupported bytecode op %u",
918 (unsigned int) *(bytecode_opcode_t
*) pc
);
924 /* Dynamic typing. */
925 switch (estack_ax_t
) {
926 case REG_S64
: /* Fall-through */
928 switch (estack_bx_t
) {
929 case REG_S64
: /* Fall-through */
931 JUMP_TO(BYTECODE_OP_EQ_S64
);
933 JUMP_TO(BYTECODE_OP_EQ_DOUBLE_S64
);
934 case REG_STRING
: /* Fall-through */
935 case REG_STAR_GLOB_STRING
:
939 ERR("Unknown interpreter register type (%d)",
946 switch (estack_bx_t
) {
947 case REG_S64
: /* Fall-through */
949 JUMP_TO(BYTECODE_OP_EQ_S64_DOUBLE
);
951 JUMP_TO(BYTECODE_OP_EQ_DOUBLE
);
952 case REG_STRING
: /* Fall-through */
953 case REG_STAR_GLOB_STRING
:
957 ERR("Unknown interpreter register type (%d)",
964 switch (estack_bx_t
) {
965 case REG_S64
: /* Fall-through */
966 case REG_U64
: /* Fall-through */
971 JUMP_TO(BYTECODE_OP_EQ_STRING
);
972 case REG_STAR_GLOB_STRING
:
973 JUMP_TO(BYTECODE_OP_EQ_STAR_GLOB_STRING
);
975 ERR("Unknown interpreter register type (%d)",
981 case REG_STAR_GLOB_STRING
:
982 switch (estack_bx_t
) {
983 case REG_S64
: /* Fall-through */
984 case REG_U64
: /* Fall-through */
989 JUMP_TO(BYTECODE_OP_EQ_STAR_GLOB_STRING
);
990 case REG_STAR_GLOB_STRING
:
994 ERR("Unknown interpreter register type (%d)",
1001 ERR("Unknown interpreter register type (%d)",
1009 /* Dynamic typing. */
1010 switch (estack_ax_t
) {
1011 case REG_S64
: /* Fall-through */
1013 switch (estack_bx_t
) {
1014 case REG_S64
: /* Fall-through */
1016 JUMP_TO(BYTECODE_OP_NE_S64
);
1018 JUMP_TO(BYTECODE_OP_NE_DOUBLE_S64
);
1019 case REG_STRING
: /* Fall-through */
1020 case REG_STAR_GLOB_STRING
:
1024 ERR("Unknown interpreter register type (%d)",
1031 switch (estack_bx_t
) {
1032 case REG_S64
: /* Fall-through */
1034 JUMP_TO(BYTECODE_OP_NE_S64_DOUBLE
);
1036 JUMP_TO(BYTECODE_OP_NE_DOUBLE
);
1037 case REG_STRING
: /* Fall-through */
1038 case REG_STAR_GLOB_STRING
:
1042 ERR("Unknown interpreter register type (%d)",
1049 switch (estack_bx_t
) {
1050 case REG_S64
: /* Fall-through */
1056 JUMP_TO(BYTECODE_OP_NE_STRING
);
1057 case REG_STAR_GLOB_STRING
:
1058 JUMP_TO(BYTECODE_OP_NE_STAR_GLOB_STRING
);
1060 ERR("Unknown interpreter register type (%d)",
1066 case REG_STAR_GLOB_STRING
:
1067 switch (estack_bx_t
) {
1068 case REG_S64
: /* Fall-through */
1074 JUMP_TO(BYTECODE_OP_NE_STAR_GLOB_STRING
);
1075 case REG_STAR_GLOB_STRING
:
1079 ERR("Unknown interpreter register type (%d)",
1086 ERR("Unknown interpreter register type (%d)",
1094 /* Dynamic typing. */
1095 switch (estack_ax_t
) {
1096 case REG_S64
: /* Fall-through */
1098 switch (estack_bx_t
) {
1099 case REG_S64
: /* Fall-through */
1101 JUMP_TO(BYTECODE_OP_GT_S64
);
1103 JUMP_TO(BYTECODE_OP_GT_DOUBLE_S64
);
1104 case REG_STRING
: /* Fall-through */
1105 case REG_STAR_GLOB_STRING
:
1109 ERR("Unknown interpreter register type (%d)",
1116 switch (estack_bx_t
) {
1117 case REG_S64
: /* Fall-through */
1119 JUMP_TO(BYTECODE_OP_GT_S64_DOUBLE
);
1121 JUMP_TO(BYTECODE_OP_GT_DOUBLE
);
1122 case REG_STRING
: /* Fall-through */
1123 case REG_STAR_GLOB_STRING
:
1127 ERR("Unknown interpreter register type (%d)",
1134 switch (estack_bx_t
) {
1135 case REG_S64
: /* Fall-through */
1136 case REG_U64
: /* Fall-through */
1137 case REG_DOUBLE
: /* Fall-through */
1138 case REG_STAR_GLOB_STRING
:
1142 JUMP_TO(BYTECODE_OP_GT_STRING
);
1144 ERR("Unknown interpreter register type (%d)",
1151 ERR("Unknown interpreter register type (%d)",
1159 /* Dynamic typing. */
1160 switch (estack_ax_t
) {
1161 case REG_S64
: /* Fall-through */
1163 switch (estack_bx_t
) {
1164 case REG_S64
: /* Fall-through */
1166 JUMP_TO(BYTECODE_OP_LT_S64
);
1168 JUMP_TO(BYTECODE_OP_LT_DOUBLE_S64
);
1169 case REG_STRING
: /* Fall-through */
1170 case REG_STAR_GLOB_STRING
:
1174 ERR("Unknown interpreter register type (%d)",
1181 switch (estack_bx_t
) {
1182 case REG_S64
: /* Fall-through */
1184 JUMP_TO(BYTECODE_OP_LT_S64_DOUBLE
);
1186 JUMP_TO(BYTECODE_OP_LT_DOUBLE
);
1187 case REG_STRING
: /* Fall-through */
1188 case REG_STAR_GLOB_STRING
:
1192 ERR("Unknown interpreter register type (%d)",
1199 switch (estack_bx_t
) {
1200 case REG_S64
: /* Fall-through */
1201 case REG_U64
: /* Fall-through */
1202 case REG_DOUBLE
: /* Fall-through */
1203 case REG_STAR_GLOB_STRING
:
1207 JUMP_TO(BYTECODE_OP_LT_STRING
);
1209 ERR("Unknown interpreter register type (%d)",
1216 ERR("Unknown interpreter register type (%d)",
1224 /* Dynamic typing. */
1225 switch (estack_ax_t
) {
1226 case REG_S64
: /* Fall-through */
1228 switch (estack_bx_t
) {
1229 case REG_S64
: /* Fall-through */
1231 JUMP_TO(BYTECODE_OP_GE_S64
);
1233 JUMP_TO(BYTECODE_OP_GE_DOUBLE_S64
);
1234 case REG_STRING
: /* Fall-through */
1235 case REG_STAR_GLOB_STRING
:
1239 ERR("Unknown interpreter register type (%d)",
1246 switch (estack_bx_t
) {
1247 case REG_S64
: /* Fall-through */
1249 JUMP_TO(BYTECODE_OP_GE_S64_DOUBLE
);
1251 JUMP_TO(BYTECODE_OP_GE_DOUBLE
);
1252 case REG_STRING
: /* Fall-through */
1253 case REG_STAR_GLOB_STRING
:
1257 ERR("Unknown interpreter register type (%d)",
1264 switch (estack_bx_t
) {
1265 case REG_S64
: /* Fall-through */
1266 case REG_U64
: /* Fall-through */
1267 case REG_DOUBLE
: /* Fall-through */
1268 case REG_STAR_GLOB_STRING
:
1272 JUMP_TO(BYTECODE_OP_GE_STRING
);
1274 ERR("Unknown interpreter register type (%d)",
1281 ERR("Unknown interpreter register type (%d)",
1289 /* Dynamic typing. */
1290 switch (estack_ax_t
) {
1291 case REG_S64
: /* Fall-through */
1293 switch (estack_bx_t
) {
1294 case REG_S64
: /* Fall-through */
1296 JUMP_TO(BYTECODE_OP_LE_S64
);
1298 JUMP_TO(BYTECODE_OP_LE_DOUBLE_S64
);
1299 case REG_STRING
: /* Fall-through */
1300 case REG_STAR_GLOB_STRING
:
1304 ERR("Unknown interpreter register type (%d)",
1311 switch (estack_bx_t
) {
1312 case REG_S64
: /* Fall-through */
1314 JUMP_TO(BYTECODE_OP_LE_S64_DOUBLE
);
1316 JUMP_TO(BYTECODE_OP_LE_DOUBLE
);
1317 case REG_STRING
: /* Fall-through */
1318 case REG_STAR_GLOB_STRING
:
1322 ERR("Unknown interpreter register type (%d)",
1329 switch (estack_bx_t
) {
1330 case REG_S64
: /* Fall-through */
1331 case REG_U64
: /* Fall-through */
1332 case REG_DOUBLE
: /* Fall-through */
1333 case REG_STAR_GLOB_STRING
:
1337 JUMP_TO(BYTECODE_OP_LE_STRING
);
1339 ERR("Unknown interpreter register type (%d)",
1346 ERR("Unknown interpreter register type (%d)",
1353 OP(BYTECODE_OP_EQ_STRING
):
1357 res
= (stack_strcmp(stack
, top
, "==") == 0);
1358 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1360 estack_ax_t
= REG_S64
;
1361 next_pc
+= sizeof(struct binary_op
);
1364 OP(BYTECODE_OP_NE_STRING
):
1368 res
= (stack_strcmp(stack
, top
, "!=") != 0);
1369 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1371 estack_ax_t
= REG_S64
;
1372 next_pc
+= sizeof(struct binary_op
);
1375 OP(BYTECODE_OP_GT_STRING
):
1379 res
= (stack_strcmp(stack
, top
, ">") > 0);
1380 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1382 estack_ax_t
= REG_S64
;
1383 next_pc
+= sizeof(struct binary_op
);
1386 OP(BYTECODE_OP_LT_STRING
):
1390 res
= (stack_strcmp(stack
, top
, "<") < 0);
1391 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1393 estack_ax_t
= REG_S64
;
1394 next_pc
+= sizeof(struct binary_op
);
1397 OP(BYTECODE_OP_GE_STRING
):
1401 res
= (stack_strcmp(stack
, top
, ">=") >= 0);
1402 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1404 estack_ax_t
= REG_S64
;
1405 next_pc
+= sizeof(struct binary_op
);
1408 OP(BYTECODE_OP_LE_STRING
):
1412 res
= (stack_strcmp(stack
, top
, "<=") <= 0);
1413 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1415 estack_ax_t
= REG_S64
;
1416 next_pc
+= sizeof(struct binary_op
);
1420 OP(BYTECODE_OP_EQ_STAR_GLOB_STRING
):
1424 res
= (stack_star_glob_match(stack
, top
, "==") == 0);
1425 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1427 estack_ax_t
= REG_S64
;
1428 next_pc
+= sizeof(struct binary_op
);
1431 OP(BYTECODE_OP_NE_STAR_GLOB_STRING
):
1435 res
= (stack_star_glob_match(stack
, top
, "!=") != 0);
1436 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1438 estack_ax_t
= REG_S64
;
1439 next_pc
+= sizeof(struct binary_op
);
1443 OP(BYTECODE_OP_EQ_S64
):
1447 res
= (estack_bx_v
== estack_ax_v
);
1448 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1450 estack_ax_t
= REG_S64
;
1451 next_pc
+= sizeof(struct binary_op
);
1454 OP(BYTECODE_OP_NE_S64
):
1458 res
= (estack_bx_v
!= estack_ax_v
);
1459 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1461 estack_ax_t
= REG_S64
;
1462 next_pc
+= sizeof(struct binary_op
);
1465 OP(BYTECODE_OP_GT_S64
):
1469 res
= (estack_bx_v
> estack_ax_v
);
1470 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1472 estack_ax_t
= REG_S64
;
1473 next_pc
+= sizeof(struct binary_op
);
1476 OP(BYTECODE_OP_LT_S64
):
1480 res
= (estack_bx_v
< estack_ax_v
);
1481 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1483 estack_ax_t
= REG_S64
;
1484 next_pc
+= sizeof(struct binary_op
);
1487 OP(BYTECODE_OP_GE_S64
):
1491 res
= (estack_bx_v
>= estack_ax_v
);
1492 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1494 estack_ax_t
= REG_S64
;
1495 next_pc
+= sizeof(struct binary_op
);
1498 OP(BYTECODE_OP_LE_S64
):
1502 res
= (estack_bx_v
<= estack_ax_v
);
1503 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1505 estack_ax_t
= REG_S64
;
1506 next_pc
+= sizeof(struct binary_op
);
1510 OP(BYTECODE_OP_EQ_DOUBLE
):
1514 res
= (estack_bx(stack
, top
)->u
.d
== estack_ax(stack
, top
)->u
.d
);
1515 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1517 estack_ax_t
= REG_S64
;
1518 next_pc
+= sizeof(struct binary_op
);
1521 OP(BYTECODE_OP_NE_DOUBLE
):
1525 res
= (estack_bx(stack
, top
)->u
.d
!= estack_ax(stack
, top
)->u
.d
);
1526 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1528 estack_ax_t
= REG_S64
;
1529 next_pc
+= sizeof(struct binary_op
);
1532 OP(BYTECODE_OP_GT_DOUBLE
):
1536 res
= (estack_bx(stack
, top
)->u
.d
> estack_ax(stack
, top
)->u
.d
);
1537 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1539 estack_ax_t
= REG_S64
;
1540 next_pc
+= sizeof(struct binary_op
);
1543 OP(BYTECODE_OP_LT_DOUBLE
):
1547 res
= (estack_bx(stack
, top
)->u
.d
< estack_ax(stack
, top
)->u
.d
);
1548 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1550 estack_ax_t
= REG_S64
;
1551 next_pc
+= sizeof(struct binary_op
);
1554 OP(BYTECODE_OP_GE_DOUBLE
):
1558 res
= (estack_bx(stack
, top
)->u
.d
>= estack_ax(stack
, top
)->u
.d
);
1559 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1561 estack_ax_t
= REG_S64
;
1562 next_pc
+= sizeof(struct binary_op
);
1565 OP(BYTECODE_OP_LE_DOUBLE
):
1569 res
= (estack_bx(stack
, top
)->u
.d
<= estack_ax(stack
, top
)->u
.d
);
1570 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1572 estack_ax_t
= REG_S64
;
1573 next_pc
+= sizeof(struct binary_op
);
1577 /* Mixed S64-double binary comparators */
1578 OP(BYTECODE_OP_EQ_DOUBLE_S64
):
1582 res
= (estack_bx(stack
, top
)->u
.d
== estack_ax_v
);
1583 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1585 estack_ax_t
= REG_S64
;
1586 next_pc
+= sizeof(struct binary_op
);
1589 OP(BYTECODE_OP_NE_DOUBLE_S64
):
1593 res
= (estack_bx(stack
, top
)->u
.d
!= estack_ax_v
);
1594 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1596 estack_ax_t
= REG_S64
;
1597 next_pc
+= sizeof(struct binary_op
);
1600 OP(BYTECODE_OP_GT_DOUBLE_S64
):
1604 res
= (estack_bx(stack
, top
)->u
.d
> estack_ax_v
);
1605 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1607 estack_ax_t
= REG_S64
;
1608 next_pc
+= sizeof(struct binary_op
);
1611 OP(BYTECODE_OP_LT_DOUBLE_S64
):
1615 res
= (estack_bx(stack
, top
)->u
.d
< estack_ax_v
);
1616 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1618 estack_ax_t
= REG_S64
;
1619 next_pc
+= sizeof(struct binary_op
);
1622 OP(BYTECODE_OP_GE_DOUBLE_S64
):
1626 res
= (estack_bx(stack
, top
)->u
.d
>= estack_ax_v
);
1627 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1629 estack_ax_t
= REG_S64
;
1630 next_pc
+= sizeof(struct binary_op
);
1633 OP(BYTECODE_OP_LE_DOUBLE_S64
):
1637 res
= (estack_bx(stack
, top
)->u
.d
<= estack_ax_v
);
1638 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1640 estack_ax_t
= REG_S64
;
1641 next_pc
+= sizeof(struct binary_op
);
1645 OP(BYTECODE_OP_EQ_S64_DOUBLE
):
1649 res
= (estack_bx_v
== estack_ax(stack
, top
)->u
.d
);
1650 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1652 estack_ax_t
= REG_S64
;
1653 next_pc
+= sizeof(struct binary_op
);
1656 OP(BYTECODE_OP_NE_S64_DOUBLE
):
1660 res
= (estack_bx_v
!= estack_ax(stack
, top
)->u
.d
);
1661 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1663 estack_ax_t
= REG_S64
;
1664 next_pc
+= sizeof(struct binary_op
);
1667 OP(BYTECODE_OP_GT_S64_DOUBLE
):
1671 res
= (estack_bx_v
> estack_ax(stack
, top
)->u
.d
);
1672 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1674 estack_ax_t
= REG_S64
;
1675 next_pc
+= sizeof(struct binary_op
);
1678 OP(BYTECODE_OP_LT_S64_DOUBLE
):
1682 res
= (estack_bx_v
< estack_ax(stack
, top
)->u
.d
);
1683 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1685 estack_ax_t
= REG_S64
;
1686 next_pc
+= sizeof(struct binary_op
);
1689 OP(BYTECODE_OP_GE_S64_DOUBLE
):
1693 res
= (estack_bx_v
>= estack_ax(stack
, top
)->u
.d
);
1694 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1696 estack_ax_t
= REG_S64
;
1697 next_pc
+= sizeof(struct binary_op
);
1700 OP(BYTECODE_OP_LE_S64_DOUBLE
):
1704 res
= (estack_bx_v
<= estack_ax(stack
, top
)->u
.d
);
1705 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1707 estack_ax_t
= REG_S64
;
1708 next_pc
+= sizeof(struct binary_op
);
1711 OP(BYTECODE_OP_BIT_RSHIFT
):
1715 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1720 /* Catch undefined behavior. */
1721 if (caa_unlikely(estack_ax_v
< 0 || estack_ax_v
>= 64)) {
1725 res
= ((uint64_t) estack_bx_v
>> (uint32_t) estack_ax_v
);
1726 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1728 estack_ax_t
= REG_U64
;
1729 next_pc
+= sizeof(struct binary_op
);
1732 OP(BYTECODE_OP_BIT_LSHIFT
):
1736 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1741 /* Catch undefined behavior. */
1742 if (caa_unlikely(estack_ax_v
< 0 || estack_ax_v
>= 64)) {
1746 res
= ((uint64_t) estack_bx_v
<< (uint32_t) estack_ax_v
);
1747 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1749 estack_ax_t
= REG_U64
;
1750 next_pc
+= sizeof(struct binary_op
);
1753 OP(BYTECODE_OP_BIT_AND
):
1757 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1762 res
= ((uint64_t) estack_bx_v
& (uint64_t) estack_ax_v
);
1763 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1765 estack_ax_t
= REG_U64
;
1766 next_pc
+= sizeof(struct binary_op
);
1769 OP(BYTECODE_OP_BIT_OR
):
1773 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1778 res
= ((uint64_t) estack_bx_v
| (uint64_t) estack_ax_v
);
1779 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1781 estack_ax_t
= REG_U64
;
1782 next_pc
+= sizeof(struct binary_op
);
1785 OP(BYTECODE_OP_BIT_XOR
):
1789 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1794 res
= ((uint64_t) estack_bx_v
^ (uint64_t) estack_ax_v
);
1795 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1797 estack_ax_t
= REG_U64
;
1798 next_pc
+= sizeof(struct binary_op
);
1803 OP(BYTECODE_OP_UNARY_PLUS
):
1805 /* Dynamic typing. */
1806 switch (estack_ax_t
) {
1807 case REG_S64
: /* Fall-through. */
1809 JUMP_TO(BYTECODE_OP_UNARY_PLUS_S64
);
1811 JUMP_TO(BYTECODE_OP_UNARY_PLUS_DOUBLE
);
1812 case REG_STRING
: /* Fall-through */
1813 case REG_STAR_GLOB_STRING
:
1817 ERR("Unknown interpreter register type (%d)",
1823 OP(BYTECODE_OP_UNARY_MINUS
):
1825 /* Dynamic typing. */
1826 switch (estack_ax_t
) {
1827 case REG_S64
: /* Fall-through. */
1829 JUMP_TO(BYTECODE_OP_UNARY_MINUS_S64
);
1831 JUMP_TO(BYTECODE_OP_UNARY_MINUS_DOUBLE
);
1832 case REG_STRING
: /* Fall-through */
1833 case REG_STAR_GLOB_STRING
:
1837 ERR("Unknown interpreter register type (%d)",
1843 OP(BYTECODE_OP_UNARY_NOT
):
1845 /* Dynamic typing. */
1846 switch (estack_ax_t
) {
1847 case REG_S64
: /* Fall-through. */
1849 JUMP_TO(BYTECODE_OP_UNARY_NOT_S64
);
1851 JUMP_TO(BYTECODE_OP_UNARY_NOT_DOUBLE
);
1852 case REG_STRING
: /* Fall-through */
1853 case REG_STAR_GLOB_STRING
:
1857 ERR("Unknown interpreter register type (%d)",
1862 next_pc
+= sizeof(struct unary_op
);
1866 OP(BYTECODE_OP_UNARY_BIT_NOT
):
1868 /* Dynamic typing. */
1869 if (!IS_INTEGER_REGISTER(estack_ax_t
)) {
1874 estack_ax_v
= ~(uint64_t) estack_ax_v
;
1875 estack_ax_t
= REG_U64
;
1876 next_pc
+= sizeof(struct unary_op
);
1880 OP(BYTECODE_OP_UNARY_PLUS_S64
):
1881 OP(BYTECODE_OP_UNARY_PLUS_DOUBLE
):
1883 next_pc
+= sizeof(struct unary_op
);
1886 OP(BYTECODE_OP_UNARY_MINUS_S64
):
1888 estack_ax_v
= -estack_ax_v
;
1889 next_pc
+= sizeof(struct unary_op
);
1892 OP(BYTECODE_OP_UNARY_MINUS_DOUBLE
):
1894 estack_ax(stack
, top
)->u
.d
= -estack_ax(stack
, top
)->u
.d
;
1895 next_pc
+= sizeof(struct unary_op
);
1898 OP(BYTECODE_OP_UNARY_NOT_S64
):
1900 estack_ax_v
= !estack_ax_v
;
1901 estack_ax_t
= REG_S64
;
1902 next_pc
+= sizeof(struct unary_op
);
1905 OP(BYTECODE_OP_UNARY_NOT_DOUBLE
):
1907 estack_ax_v
= !estack_ax(stack
, top
)->u
.d
;
1908 estack_ax_t
= REG_S64
;
1909 next_pc
+= sizeof(struct unary_op
);
1914 OP(BYTECODE_OP_AND
):
1916 struct logical_op
*insn
= (struct logical_op
*) pc
;
1918 if (estack_ax_t
!= REG_S64
&& estack_ax_t
!= REG_U64
) {
1922 /* If AX is 0, skip and evaluate to 0 */
1923 if (unlikely(estack_ax_v
== 0)) {
1924 dbg_printf("Jumping to bytecode offset %u\n",
1925 (unsigned int) insn
->skip_offset
);
1926 next_pc
= start_pc
+ insn
->skip_offset
;
1928 /* Pop 1 when jump not taken */
1929 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1930 next_pc
+= sizeof(struct logical_op
);
1936 struct logical_op
*insn
= (struct logical_op
*) pc
;
1938 if (estack_ax_t
!= REG_S64
&& estack_ax_t
!= REG_U64
) {
1942 /* If AX is nonzero, skip and evaluate to 1 */
1943 if (unlikely(estack_ax_v
!= 0)) {
1945 dbg_printf("Jumping to bytecode offset %u\n",
1946 (unsigned int) insn
->skip_offset
);
1947 next_pc
= start_pc
+ insn
->skip_offset
;
1949 /* Pop 1 when jump not taken */
1950 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1951 next_pc
+= sizeof(struct logical_op
);
1957 /* load field ref */
1958 OP(BYTECODE_OP_LOAD_FIELD_REF_STRING
):
1960 struct load_op
*insn
= (struct load_op
*) pc
;
1961 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
1963 dbg_printf("load field ref offset %u type string\n",
1965 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1966 estack_ax(stack
, top
)->u
.s
.str
=
1967 *(const char * const *) &interpreter_stack_data
[ref
->offset
];
1968 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
1969 dbg_printf("Interpreter warning: loading a NULL string.\n");
1973 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
1974 estack_ax(stack
, top
)->u
.s
.literal_type
=
1975 ESTACK_STRING_LITERAL_TYPE_NONE
;
1976 estack_ax_t
= REG_STRING
;
1977 dbg_printf("ref load string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
1978 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
1982 OP(BYTECODE_OP_LOAD_FIELD_REF_SEQUENCE
):
1984 struct load_op
*insn
= (struct load_op
*) pc
;
1985 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
1987 dbg_printf("load field ref offset %u type sequence\n",
1989 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1990 estack_ax(stack
, top
)->u
.s
.seq_len
=
1991 *(unsigned long *) &interpreter_stack_data
[ref
->offset
];
1992 estack_ax(stack
, top
)->u
.s
.str
=
1993 *(const char **) (&interpreter_stack_data
[ref
->offset
1994 + sizeof(unsigned long)]);
1995 estack_ax_t
= REG_STRING
;
1996 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
1997 dbg_printf("Interpreter warning: loading a NULL sequence.\n");
2001 estack_ax(stack
, top
)->u
.s
.literal_type
=
2002 ESTACK_STRING_LITERAL_TYPE_NONE
;
2003 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2007 OP(BYTECODE_OP_LOAD_FIELD_REF_S64
):
2009 struct load_op
*insn
= (struct load_op
*) pc
;
2010 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2012 dbg_printf("load field ref offset %u type s64\n",
2014 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2016 ((struct literal_numeric
*) &interpreter_stack_data
[ref
->offset
])->v
;
2017 estack_ax_t
= REG_S64
;
2018 dbg_printf("ref load s64 %" PRIi64
"\n", estack_ax_v
);
2019 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2023 OP(BYTECODE_OP_LOAD_FIELD_REF_DOUBLE
):
2025 struct load_op
*insn
= (struct load_op
*) pc
;
2026 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2028 dbg_printf("load field ref offset %u type double\n",
2030 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2031 memcpy(&estack_ax(stack
, top
)->u
.d
, &interpreter_stack_data
[ref
->offset
],
2032 sizeof(struct literal_double
));
2033 estack_ax_t
= REG_DOUBLE
;
2034 dbg_printf("ref load double %g\n", estack_ax(stack
, top
)->u
.d
);
2035 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2039 /* load from immediate operand */
2040 OP(BYTECODE_OP_LOAD_STRING
):
2042 struct load_op
*insn
= (struct load_op
*) pc
;
2044 dbg_printf("load string %s\n", insn
->data
);
2045 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2046 estack_ax(stack
, top
)->u
.s
.str
= insn
->data
;
2047 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2048 estack_ax(stack
, top
)->u
.s
.literal_type
=
2049 ESTACK_STRING_LITERAL_TYPE_PLAIN
;
2050 estack_ax_t
= REG_STRING
;
2051 next_pc
+= sizeof(struct load_op
) + strlen(insn
->data
) + 1;
2055 OP(BYTECODE_OP_LOAD_STAR_GLOB_STRING
):
2057 struct load_op
*insn
= (struct load_op
*) pc
;
2059 dbg_printf("load globbing pattern %s\n", insn
->data
);
2060 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2061 estack_ax(stack
, top
)->u
.s
.str
= insn
->data
;
2062 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2063 estack_ax(stack
, top
)->u
.s
.literal_type
=
2064 ESTACK_STRING_LITERAL_TYPE_STAR_GLOB
;
2065 estack_ax_t
= REG_STAR_GLOB_STRING
;
2066 next_pc
+= sizeof(struct load_op
) + strlen(insn
->data
) + 1;
2070 OP(BYTECODE_OP_LOAD_S64
):
2072 struct load_op
*insn
= (struct load_op
*) pc
;
2074 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2075 estack_ax_v
= ((struct literal_numeric
*) insn
->data
)->v
;
2076 estack_ax_t
= REG_S64
;
2077 dbg_printf("load s64 %" PRIi64
"\n", estack_ax_v
);
2078 next_pc
+= sizeof(struct load_op
)
2079 + sizeof(struct literal_numeric
);
2083 OP(BYTECODE_OP_LOAD_DOUBLE
):
2085 struct load_op
*insn
= (struct load_op
*) pc
;
2087 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2088 memcpy(&estack_ax(stack
, top
)->u
.d
, insn
->data
,
2089 sizeof(struct literal_double
));
2090 estack_ax_t
= REG_DOUBLE
;
2091 dbg_printf("load double %g\n", estack_ax(stack
, top
)->u
.d
);
2092 next_pc
+= sizeof(struct load_op
)
2093 + sizeof(struct literal_double
);
2098 OP(BYTECODE_OP_CAST_TO_S64
):
2100 /* Dynamic typing. */
2101 switch (estack_ax_t
) {
2103 JUMP_TO(BYTECODE_OP_CAST_NOP
);
2105 JUMP_TO(BYTECODE_OP_CAST_DOUBLE_TO_S64
);
2107 estack_ax_t
= REG_S64
;
2108 next_pc
+= sizeof(struct cast_op
); /* Fall-through */
2109 case REG_STRING
: /* Fall-through */
2110 case REG_STAR_GLOB_STRING
:
2114 ERR("Unknown interpreter register type (%d)",
2121 OP(BYTECODE_OP_CAST_DOUBLE_TO_S64
):
2123 estack_ax_v
= (int64_t) estack_ax(stack
, top
)->u
.d
;
2124 estack_ax_t
= REG_S64
;
2125 next_pc
+= sizeof(struct cast_op
);
2129 OP(BYTECODE_OP_CAST_NOP
):
2131 next_pc
+= sizeof(struct cast_op
);
2135 /* get context ref */
2136 OP(BYTECODE_OP_GET_CONTEXT_REF
):
2138 struct load_op
*insn
= (struct load_op
*) pc
;
2139 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2140 const struct lttng_ust_ctx_field
*ctx_field
;
2141 struct lttng_ust_ctx_value v
;
2143 dbg_printf("get context ref offset %u type dynamic\n",
2145 ctx_field
= &ctx
->fields
[ref
->offset
];
2146 ctx_field
->get_value(ctx_field
->priv
, probe_ctx
, &v
);
2147 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2149 case LTTNG_UST_DYNAMIC_TYPE_NONE
:
2152 case LTTNG_UST_DYNAMIC_TYPE_S64
:
2153 estack_ax_v
= v
.u
.s64
;
2154 estack_ax_t
= REG_S64
;
2155 dbg_printf("ref get context dynamic s64 %" PRIi64
"\n", estack_ax_v
);
2157 case LTTNG_UST_DYNAMIC_TYPE_DOUBLE
:
2158 estack_ax(stack
, top
)->u
.d
= v
.u
.d
;
2159 estack_ax_t
= REG_DOUBLE
;
2160 dbg_printf("ref get context dynamic double %g\n", estack_ax(stack
, top
)->u
.d
);
2162 case LTTNG_UST_DYNAMIC_TYPE_STRING
:
2163 estack_ax(stack
, top
)->u
.s
.str
= v
.u
.str
;
2164 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2165 dbg_printf("Interpreter warning: loading a NULL string.\n");
2169 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2170 estack_ax(stack
, top
)->u
.s
.literal_type
=
2171 ESTACK_STRING_LITERAL_TYPE_NONE
;
2172 dbg_printf("ref get context dynamic string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
2173 estack_ax_t
= REG_STRING
;
2176 dbg_printf("Interpreter warning: unknown dynamic type (%d).\n", (int) v
.sel
);
2180 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2184 OP(BYTECODE_OP_GET_CONTEXT_REF_STRING
):
2186 struct load_op
*insn
= (struct load_op
*) pc
;
2187 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2188 const struct lttng_ust_ctx_field
*ctx_field
;
2189 struct lttng_ust_ctx_value v
;
2191 dbg_printf("get context ref offset %u type string\n",
2193 ctx_field
= &ctx
->fields
[ref
->offset
];
2194 ctx_field
->get_value(ctx_field
->priv
, probe_ctx
, &v
);
2195 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2196 estack_ax(stack
, top
)->u
.s
.str
= v
.u
.str
;
2197 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2198 dbg_printf("Interpreter warning: loading a NULL string.\n");
2202 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2203 estack_ax(stack
, top
)->u
.s
.literal_type
=
2204 ESTACK_STRING_LITERAL_TYPE_NONE
;
2205 estack_ax_t
= REG_STRING
;
2206 dbg_printf("ref get context string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
2207 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2211 OP(BYTECODE_OP_GET_CONTEXT_REF_S64
):
2213 struct load_op
*insn
= (struct load_op
*) pc
;
2214 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2215 const struct lttng_ust_ctx_field
*ctx_field
;
2216 struct lttng_ust_ctx_value v
;
2218 dbg_printf("get context ref offset %u type s64\n",
2220 ctx_field
= &ctx
->fields
[ref
->offset
];
2221 ctx_field
->get_value(ctx_field
->priv
, probe_ctx
, &v
);
2222 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2223 estack_ax_v
= v
.u
.s64
;
2224 estack_ax_t
= REG_S64
;
2225 dbg_printf("ref get context s64 %" PRIi64
"\n", estack_ax_v
);
2226 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2230 OP(BYTECODE_OP_GET_CONTEXT_REF_DOUBLE
):
2232 struct load_op
*insn
= (struct load_op
*) pc
;
2233 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2234 const struct lttng_ust_ctx_field
*ctx_field
;
2235 struct lttng_ust_ctx_value v
;
2237 dbg_printf("get context ref offset %u type double\n",
2239 ctx_field
= &ctx
->fields
[ref
->offset
];
2240 ctx_field
->get_value(ctx_field
->priv
, probe_ctx
, &v
);
2241 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2242 memcpy(&estack_ax(stack
, top
)->u
.d
, &v
.u
.d
, sizeof(struct literal_double
));
2243 estack_ax_t
= REG_DOUBLE
;
2244 dbg_printf("ref get context double %g\n", estack_ax(stack
, top
)->u
.d
);
2245 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2249 OP(BYTECODE_OP_GET_CONTEXT_ROOT
):
2251 dbg_printf("op get context root\n");
2252 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2253 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_CONTEXT
;
2254 /* "field" only needed for variants. */
2255 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
2256 estack_ax_t
= REG_PTR
;
2257 next_pc
+= sizeof(struct load_op
);
2261 OP(BYTECODE_OP_GET_APP_CONTEXT_ROOT
):
2263 dbg_printf("op get app context root\n");
2264 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2265 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_APP_CONTEXT
;
2266 /* "field" only needed for variants. */
2267 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
2268 estack_ax_t
= REG_PTR
;
2269 next_pc
+= sizeof(struct load_op
);
2273 OP(BYTECODE_OP_GET_PAYLOAD_ROOT
):
2275 dbg_printf("op get app payload root\n");
2276 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2277 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_PAYLOAD
;
2278 estack_ax(stack
, top
)->u
.ptr
.ptr
= interpreter_stack_data
;
2279 /* "field" only needed for variants. */
2280 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
2281 estack_ax_t
= REG_PTR
;
2282 next_pc
+= sizeof(struct load_op
);
2286 OP(BYTECODE_OP_GET_SYMBOL
):
2288 dbg_printf("op get symbol\n");
2289 switch (estack_ax(stack
, top
)->u
.ptr
.type
) {
2291 ERR("Nested fields not implemented yet.");
2294 case LOAD_ROOT_CONTEXT
:
2295 case LOAD_ROOT_APP_CONTEXT
:
2296 case LOAD_ROOT_PAYLOAD
:
2298 * symbol lookup is performed by
2304 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_symbol
);
2308 OP(BYTECODE_OP_GET_SYMBOL_FIELD
):
2311 * Used for first variant encountered in a
2312 * traversal. Variants are not implemented yet.
2318 OP(BYTECODE_OP_GET_INDEX_U16
):
2320 struct load_op
*insn
= (struct load_op
*) pc
;
2321 struct get_index_u16
*index
= (struct get_index_u16
*) insn
->data
;
2323 dbg_printf("op get index u16\n");
2324 ret
= dynamic_get_index(ctx
, probe_ctx
, bytecode
, index
->index
, estack_ax(stack
, top
));
2327 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
2328 estack_ax_t
= estack_ax(stack
, top
)->type
;
2329 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_index_u16
);
2333 OP(BYTECODE_OP_GET_INDEX_U64
):
2335 struct load_op
*insn
= (struct load_op
*) pc
;
2336 struct get_index_u64
*index
= (struct get_index_u64
*) insn
->data
;
2338 dbg_printf("op get index u64\n");
2339 ret
= dynamic_get_index(ctx
, probe_ctx
, bytecode
, index
->index
, estack_ax(stack
, top
));
2342 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
2343 estack_ax_t
= estack_ax(stack
, top
)->type
;
2344 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_index_u64
);
2348 OP(BYTECODE_OP_LOAD_FIELD
):
2350 dbg_printf("op load field\n");
2351 ret
= dynamic_load_field(estack_ax(stack
, top
));
2354 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
2355 estack_ax_t
= estack_ax(stack
, top
)->type
;
2356 next_pc
+= sizeof(struct load_op
);
2360 OP(BYTECODE_OP_LOAD_FIELD_S8
):
2362 dbg_printf("op load field s8\n");
2364 estack_ax_v
= *(int8_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2365 estack_ax_t
= REG_S64
;
2366 next_pc
+= sizeof(struct load_op
);
2369 OP(BYTECODE_OP_LOAD_FIELD_S16
):
2371 dbg_printf("op load field s16\n");
2373 estack_ax_v
= *(int16_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2374 estack_ax_t
= REG_S64
;
2375 next_pc
+= sizeof(struct load_op
);
2378 OP(BYTECODE_OP_LOAD_FIELD_S32
):
2380 dbg_printf("op load field s32\n");
2382 estack_ax_v
= *(int32_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2383 estack_ax_t
= REG_S64
;
2384 next_pc
+= sizeof(struct load_op
);
2387 OP(BYTECODE_OP_LOAD_FIELD_S64
):
2389 dbg_printf("op load field s64\n");
2391 estack_ax_v
= *(int64_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2392 estack_ax_t
= REG_S64
;
2393 next_pc
+= sizeof(struct load_op
);
2396 OP(BYTECODE_OP_LOAD_FIELD_U8
):
2398 dbg_printf("op load field u8\n");
2400 estack_ax_v
= *(uint8_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2401 estack_ax_t
= REG_U64
;
2402 next_pc
+= sizeof(struct load_op
);
2405 OP(BYTECODE_OP_LOAD_FIELD_U16
):
2407 dbg_printf("op load field u16\n");
2409 estack_ax_v
= *(uint16_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2410 estack_ax_t
= REG_U64
;
2411 next_pc
+= sizeof(struct load_op
);
2414 OP(BYTECODE_OP_LOAD_FIELD_U32
):
2416 dbg_printf("op load field u32\n");
2418 estack_ax_v
= *(uint32_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2419 estack_ax_t
= REG_U64
;
2420 next_pc
+= sizeof(struct load_op
);
2423 OP(BYTECODE_OP_LOAD_FIELD_U64
):
2425 dbg_printf("op load field u64\n");
2427 estack_ax_v
= *(uint64_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2428 estack_ax_t
= REG_U64
;
2429 next_pc
+= sizeof(struct load_op
);
2432 OP(BYTECODE_OP_LOAD_FIELD_DOUBLE
):
2434 dbg_printf("op load field double\n");
2436 memcpy(&estack_ax(stack
, top
)->u
.d
,
2437 estack_ax(stack
, top
)->u
.ptr
.ptr
,
2438 sizeof(struct literal_double
));
2439 estack_ax(stack
, top
)->type
= REG_DOUBLE
;
2440 next_pc
+= sizeof(struct load_op
);
2444 OP(BYTECODE_OP_LOAD_FIELD_STRING
):
2448 dbg_printf("op load field string\n");
2449 str
= (const char *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2450 estack_ax(stack
, top
)->u
.s
.str
= str
;
2451 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2452 dbg_printf("Interpreter warning: loading a NULL string.\n");
2456 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2457 estack_ax(stack
, top
)->u
.s
.literal_type
=
2458 ESTACK_STRING_LITERAL_TYPE_NONE
;
2459 estack_ax(stack
, top
)->type
= REG_STRING
;
2460 next_pc
+= sizeof(struct load_op
);
2464 OP(BYTECODE_OP_LOAD_FIELD_SEQUENCE
):
2468 dbg_printf("op load field string sequence\n");
2469 ptr
= estack_ax(stack
, top
)->u
.ptr
.ptr
;
2470 estack_ax(stack
, top
)->u
.s
.seq_len
= *(unsigned long *) ptr
;
2471 estack_ax(stack
, top
)->u
.s
.str
= *(const char **) (ptr
+ sizeof(unsigned long));
2472 estack_ax(stack
, top
)->type
= REG_STRING
;
2473 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2474 dbg_printf("Interpreter warning: loading a NULL sequence.\n");
2478 estack_ax(stack
, top
)->u
.s
.literal_type
=
2479 ESTACK_STRING_LITERAL_TYPE_NONE
;
2480 next_pc
+= sizeof(struct load_op
);
2486 /* No need to prepare output if an error occurred. */
2488 return LTTNG_UST_BYTECODE_INTERPRETER_ERROR
;
2490 /* Prepare output. */
2491 switch (ust_bytecode
->type
) {
2492 case LTTNG_UST_BYTECODE_TYPE_FILTER
:
2494 struct lttng_ust_bytecode_filter_ctx
*filter_ctx
=
2495 (struct lttng_ust_bytecode_filter_ctx
*) caller_ctx
;
2497 filter_ctx
->result
= LTTNG_UST_BYTECODE_FILTER_ACCEPT
;
2499 filter_ctx
->result
= LTTNG_UST_BYTECODE_FILTER_REJECT
;
2502 case LTTNG_UST_BYTECODE_TYPE_CAPTURE
:
2503 ret
= lttng_bytecode_interpret_format_output(estack_ax(stack
, top
),
2504 (struct lttng_interpreter_output
*) caller_ctx
);
2511 return LTTNG_UST_BYTECODE_INTERPRETER_ERROR
;
2513 return LTTNG_UST_BYTECODE_INTERPRETER_OK
;
2517 * Return LTTNG_UST_EVENT_FILTER_ACCEPT or LTTNG_UST_EVENT_FILTER_REJECT.
2519 int lttng_ust_interpret_event_filter(const struct lttng_ust_event_common
*event
,
2520 const char *interpreter_stack_data
,
2521 struct lttng_ust_probe_ctx
*probe_ctx
,
2522 void *event_filter_ctx
__attribute__((unused
)))
2524 struct lttng_ust_bytecode_runtime
*filter_bc_runtime
;
2525 struct cds_list_head
*filter_bytecode_runtime_head
= &event
->priv
->filter_bytecode_runtime_head
;
2526 struct lttng_ust_bytecode_filter_ctx bytecode_filter_ctx
;
2527 bool filter_record
= false;
2529 cds_list_for_each_entry_rcu(filter_bc_runtime
, filter_bytecode_runtime_head
, node
) {
2530 if (caa_likely(filter_bc_runtime
->interpreter_func(filter_bc_runtime
,
2531 interpreter_stack_data
, probe_ctx
, &bytecode_filter_ctx
) == LTTNG_UST_BYTECODE_INTERPRETER_OK
)) {
2532 if (caa_unlikely(bytecode_filter_ctx
.result
== LTTNG_UST_BYTECODE_FILTER_ACCEPT
)) {
2533 filter_record
= true;
2539 return LTTNG_UST_EVENT_FILTER_ACCEPT
;
2541 return LTTNG_UST_EVENT_FILTER_REJECT
;