2 * SPDX-License-Identifier: MIT
4 * Copyright (C) 2010-2016 Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
6 * LTTng UST bytecode interpreter.
13 #include <lttng/urcu/pointer.h>
14 #include <urcu/rculist.h>
15 #include <lttng/ust-endian.h>
16 #include <lttng/ust-events.h>
17 #include "lib/lttng-ust/events.h"
19 #include "lttng-bytecode.h"
20 #include "common/strutils.h"
25 * -2: unknown escape char.
30 int parse_char(const char **p
)
50 * Returns SIZE_MAX if the string is null-terminated, or the number of
54 size_t get_str_or_seq_len(const struct estack_entry
*entry
)
56 return entry
->u
.s
.seq_len
;
60 int stack_star_glob_match(struct estack
*stack
, int top
,
61 const char *cmp_type
__attribute__((unused
)))
64 const char *candidate
;
68 /* Find out which side is the pattern vs. the candidate. */
69 if (estack_ax(stack
, top
)->u
.s
.literal_type
== ESTACK_STRING_LITERAL_TYPE_STAR_GLOB
) {
70 pattern
= estack_ax(stack
, top
)->u
.s
.str
;
71 pattern_len
= get_str_or_seq_len(estack_ax(stack
, top
));
72 candidate
= estack_bx(stack
, top
)->u
.s
.str
;
73 candidate_len
= get_str_or_seq_len(estack_bx(stack
, top
));
75 pattern
= estack_bx(stack
, top
)->u
.s
.str
;
76 pattern_len
= get_str_or_seq_len(estack_bx(stack
, top
));
77 candidate
= estack_ax(stack
, top
)->u
.s
.str
;
78 candidate_len
= get_str_or_seq_len(estack_ax(stack
, top
));
81 /* Perform the match. Returns 0 when the result is true. */
82 return !strutils_star_glob_match(pattern
, pattern_len
, candidate
,
87 int stack_strcmp(struct estack
*stack
, int top
, const char *cmp_type
__attribute__((unused
)))
89 const char *p
= estack_bx(stack
, top
)->u
.s
.str
, *q
= estack_ax(stack
, top
)->u
.s
.str
;
96 if (unlikely(p
- estack_bx(stack
, top
)->u
.s
.str
>= estack_bx(stack
, top
)->u
.s
.seq_len
|| *p
== '\0')) {
97 if (q
- estack_ax(stack
, top
)->u
.s
.str
>= estack_ax(stack
, top
)->u
.s
.seq_len
|| *q
== '\0') {
100 if (estack_ax(stack
, top
)->u
.s
.literal_type
==
101 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
102 ret
= parse_char(&q
);
109 if (unlikely(q
- estack_ax(stack
, top
)->u
.s
.str
>= estack_ax(stack
, top
)->u
.s
.seq_len
|| *q
== '\0')) {
110 if (estack_bx(stack
, top
)->u
.s
.literal_type
==
111 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
112 ret
= parse_char(&p
);
118 if (estack_bx(stack
, top
)->u
.s
.literal_type
==
119 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
120 ret
= parse_char(&p
);
123 } else if (ret
== -2) {
126 /* else compare both char */
128 if (estack_ax(stack
, top
)->u
.s
.literal_type
==
129 ESTACK_STRING_LITERAL_TYPE_PLAIN
) {
130 ret
= parse_char(&q
);
133 } else if (ret
== -2) {
153 int lttng_bytecode_interpret_error(
154 struct lttng_ust_bytecode_runtime
*bytecode_runtime
__attribute__((unused
)),
155 const char *stack_data
__attribute__((unused
)),
156 void *ctx
__attribute__((unused
)))
158 return LTTNG_UST_BYTECODE_INTERPRETER_ERROR
;
161 #ifdef INTERPRETER_USE_SWITCH
164 * Fallback for compilers that do not support taking address of labels.
168 start_pc = &bytecode->data[0]; \
169 for (pc = next_pc = start_pc; pc - start_pc < bytecode->len; \
171 dbg_printf("Executing op %s (%u)\n", \
172 lttng_bytecode_print_op((unsigned int) *(bytecode_opcode_t *) pc), \
173 (unsigned int) *(bytecode_opcode_t *) pc); \
174 switch (*(bytecode_opcode_t *) pc) {
176 #define OP(name) jump_target_##name: __attribute__((unused)); \
184 #define JUMP_TO(name) \
185 goto jump_target_##name
190 * Dispatch-table based interpreter.
194 start_pc = &bytecode->code[0]; \
195 pc = next_pc = start_pc; \
196 if (unlikely(pc - start_pc >= bytecode->len)) \
198 goto *dispatch[*(bytecode_opcode_t *) pc];
205 goto *dispatch[*(bytecode_opcode_t *) pc];
209 #define JUMP_TO(name) \
214 #define IS_INTEGER_REGISTER(reg_type) \
215 (reg_type == REG_U64 || reg_type == REG_S64)
217 static int context_get_index(struct lttng_ust_ctx
*ctx
,
218 struct load_ptr
*ptr
,
222 const struct lttng_ust_ctx_field
*ctx_field
;
223 const struct lttng_ust_event_field
*field
;
224 struct lttng_ust_ctx_value v
;
226 ctx_field
= &ctx
->fields
[idx
];
227 field
= ctx_field
->event_field
;
228 ptr
->type
= LOAD_OBJECT
;
231 switch (field
->type
->type
) {
232 case lttng_ust_type_integer
:
233 ctx_field
->get_value(ctx_field
->priv
, &v
);
234 if (lttng_ust_get_type_integer(field
->type
)->signedness
) {
235 ptr
->object_type
= OBJECT_TYPE_S64
;
236 ptr
->u
.s64
= v
.u
.s64
;
237 ptr
->ptr
= &ptr
->u
.s64
;
239 ptr
->object_type
= OBJECT_TYPE_U64
;
240 ptr
->u
.u64
= v
.u
.s64
; /* Cast. */
241 ptr
->ptr
= &ptr
->u
.u64
;
244 case lttng_ust_type_enum
:
246 const struct lttng_ust_type_integer
*itype
;
248 itype
= lttng_ust_get_type_integer(lttng_ust_get_type_enum(field
->type
)->container_type
);
249 ctx_field
->get_value(ctx_field
->priv
, &v
);
250 if (itype
->signedness
) {
251 ptr
->object_type
= OBJECT_TYPE_SIGNED_ENUM
;
252 ptr
->u
.s64
= v
.u
.s64
;
253 ptr
->ptr
= &ptr
->u
.s64
;
255 ptr
->object_type
= OBJECT_TYPE_UNSIGNED_ENUM
;
256 ptr
->u
.u64
= v
.u
.s64
; /* Cast. */
257 ptr
->ptr
= &ptr
->u
.u64
;
261 case lttng_ust_type_array
:
262 if (lttng_ust_get_type_array(field
->type
)->elem_type
->type
!= lttng_ust_type_integer
) {
263 ERR("Array nesting only supports integer types.");
266 if (lttng_ust_get_type_array(field
->type
)->encoding
== lttng_ust_string_encoding_none
) {
267 ERR("Only string arrays are supported for contexts.");
270 ptr
->object_type
= OBJECT_TYPE_STRING
;
271 ctx_field
->get_value(ctx_field
->priv
, &v
);
274 case lttng_ust_type_sequence
:
275 if (lttng_ust_get_type_sequence(field
->type
)->elem_type
->type
!= lttng_ust_type_integer
) {
276 ERR("Sequence nesting only supports integer types.");
279 if (lttng_ust_get_type_sequence(field
->type
)->encoding
== lttng_ust_string_encoding_none
) {
280 ERR("Only string sequences are supported for contexts.");
283 ptr
->object_type
= OBJECT_TYPE_STRING
;
284 ctx_field
->get_value(ctx_field
->priv
, &v
);
287 case lttng_ust_type_string
:
288 ptr
->object_type
= OBJECT_TYPE_STRING
;
289 ctx_field
->get_value(ctx_field
->priv
, &v
);
292 case lttng_ust_type_float
:
293 ptr
->object_type
= OBJECT_TYPE_DOUBLE
;
294 ctx_field
->get_value(ctx_field
->priv
, &v
);
296 ptr
->ptr
= &ptr
->u
.d
;
298 case lttng_ust_type_dynamic
:
299 ctx_field
->get_value(ctx_field
->priv
, &v
);
301 case LTTNG_UST_DYNAMIC_TYPE_NONE
:
303 case LTTNG_UST_DYNAMIC_TYPE_U8
:
304 case LTTNG_UST_DYNAMIC_TYPE_U16
:
305 case LTTNG_UST_DYNAMIC_TYPE_U32
:
306 case LTTNG_UST_DYNAMIC_TYPE_U64
:
307 ptr
->object_type
= OBJECT_TYPE_U64
;
308 ptr
->u
.u64
= v
.u
.u64
;
309 ptr
->ptr
= &ptr
->u
.u64
;
310 dbg_printf("context get index dynamic u64 %" PRIi64
"\n", ptr
->u
.u64
);
312 case LTTNG_UST_DYNAMIC_TYPE_S8
:
313 case LTTNG_UST_DYNAMIC_TYPE_S16
:
314 case LTTNG_UST_DYNAMIC_TYPE_S32
:
315 case LTTNG_UST_DYNAMIC_TYPE_S64
:
316 ptr
->object_type
= OBJECT_TYPE_S64
;
317 ptr
->u
.s64
= v
.u
.s64
;
318 ptr
->ptr
= &ptr
->u
.s64
;
319 dbg_printf("context get index dynamic s64 %" PRIi64
"\n", ptr
->u
.s64
);
321 case LTTNG_UST_DYNAMIC_TYPE_FLOAT
:
322 case LTTNG_UST_DYNAMIC_TYPE_DOUBLE
:
323 ptr
->object_type
= OBJECT_TYPE_DOUBLE
;
325 ptr
->ptr
= &ptr
->u
.d
;
326 dbg_printf("context get index dynamic double %g\n", ptr
->u
.d
);
328 case LTTNG_UST_DYNAMIC_TYPE_STRING
:
329 ptr
->object_type
= OBJECT_TYPE_STRING
;
331 dbg_printf("context get index dynamic string %s\n", (const char *) ptr
->ptr
);
334 dbg_printf("Interpreter warning: unknown dynamic type (%d).\n", (int) v
.sel
);
339 ERR("Unknown type: %d", (int) field
->type
->type
);
345 static int dynamic_get_index(struct lttng_ust_ctx
*ctx
,
346 struct bytecode_runtime
*runtime
,
347 uint64_t index
, struct estack_entry
*stack_top
)
350 const struct bytecode_get_index_data
*gid
;
352 gid
= (const struct bytecode_get_index_data
*) &runtime
->data
[index
];
353 switch (stack_top
->u
.ptr
.type
) {
355 switch (stack_top
->u
.ptr
.object_type
) {
356 case OBJECT_TYPE_ARRAY
:
360 assert(gid
->offset
< gid
->array_len
);
361 /* Skip count (unsigned long) */
362 ptr
= *(const char **) (stack_top
->u
.ptr
.ptr
+ sizeof(unsigned long));
363 ptr
= ptr
+ gid
->offset
;
364 stack_top
->u
.ptr
.ptr
= ptr
;
365 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
366 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
367 assert(stack_top
->u
.ptr
.field
->type
->type
== lttng_ust_type_array
);
368 stack_top
->u
.ptr
.field
= NULL
;
371 case OBJECT_TYPE_SEQUENCE
:
376 ptr
= *(const char **) (stack_top
->u
.ptr
.ptr
+ sizeof(unsigned long));
377 ptr_seq_len
= *(unsigned long *) stack_top
->u
.ptr
.ptr
;
378 if (gid
->offset
>= gid
->elem
.len
* ptr_seq_len
) {
382 ptr
= ptr
+ gid
->offset
;
383 stack_top
->u
.ptr
.ptr
= ptr
;
384 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
385 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
386 assert(stack_top
->u
.ptr
.field
->type
->type
== lttng_ust_type_sequence
);
387 stack_top
->u
.ptr
.field
= NULL
;
390 case OBJECT_TYPE_STRUCT
:
391 ERR("Nested structures are not supported yet.");
394 case OBJECT_TYPE_VARIANT
:
396 ERR("Unexpected get index type %d",
397 (int) stack_top
->u
.ptr
.object_type
);
402 case LOAD_ROOT_CONTEXT
:
403 case LOAD_ROOT_APP_CONTEXT
: /* Fall-through */
405 ret
= context_get_index(ctx
,
413 case LOAD_ROOT_PAYLOAD
:
414 stack_top
->u
.ptr
.ptr
+= gid
->offset
;
415 if (gid
->elem
.type
== OBJECT_TYPE_STRING
)
416 stack_top
->u
.ptr
.ptr
= *(const char * const *) stack_top
->u
.ptr
.ptr
;
417 stack_top
->u
.ptr
.object_type
= gid
->elem
.type
;
418 stack_top
->u
.ptr
.type
= LOAD_OBJECT
;
419 stack_top
->u
.ptr
.field
= gid
->field
;
420 stack_top
->u
.ptr
.rev_bo
= gid
->elem
.rev_bo
;
424 stack_top
->type
= REG_PTR
;
432 static int dynamic_load_field(struct estack_entry
*stack_top
)
436 switch (stack_top
->u
.ptr
.type
) {
439 case LOAD_ROOT_CONTEXT
:
440 case LOAD_ROOT_APP_CONTEXT
:
441 case LOAD_ROOT_PAYLOAD
:
443 dbg_printf("Interpreter warning: cannot load root, missing field name.\n");
447 switch (stack_top
->u
.ptr
.object_type
) {
449 dbg_printf("op load field s8\n");
450 stack_top
->u
.v
= *(int8_t *) stack_top
->u
.ptr
.ptr
;
451 stack_top
->type
= REG_S64
;
453 case OBJECT_TYPE_S16
:
457 dbg_printf("op load field s16\n");
458 tmp
= *(int16_t *) stack_top
->u
.ptr
.ptr
;
459 if (stack_top
->u
.ptr
.rev_bo
)
460 tmp
= lttng_ust_bswap_16(tmp
);
461 stack_top
->u
.v
= tmp
;
462 stack_top
->type
= REG_S64
;
465 case OBJECT_TYPE_S32
:
469 dbg_printf("op load field s32\n");
470 tmp
= *(int32_t *) stack_top
->u
.ptr
.ptr
;
471 if (stack_top
->u
.ptr
.rev_bo
)
472 tmp
= lttng_ust_bswap_32(tmp
);
473 stack_top
->u
.v
= tmp
;
474 stack_top
->type
= REG_S64
;
477 case OBJECT_TYPE_S64
:
481 dbg_printf("op load field s64\n");
482 tmp
= *(int64_t *) stack_top
->u
.ptr
.ptr
;
483 if (stack_top
->u
.ptr
.rev_bo
)
484 tmp
= lttng_ust_bswap_64(tmp
);
485 stack_top
->u
.v
= tmp
;
486 stack_top
->type
= REG_S64
;
489 case OBJECT_TYPE_SIGNED_ENUM
:
493 dbg_printf("op load field signed enumeration\n");
494 tmp
= *(int64_t *) stack_top
->u
.ptr
.ptr
;
495 if (stack_top
->u
.ptr
.rev_bo
)
496 tmp
= lttng_ust_bswap_64(tmp
);
497 stack_top
->u
.v
= tmp
;
498 stack_top
->type
= REG_S64
;
502 dbg_printf("op load field u8\n");
503 stack_top
->u
.v
= *(uint8_t *) stack_top
->u
.ptr
.ptr
;
504 stack_top
->type
= REG_U64
;
506 case OBJECT_TYPE_U16
:
510 dbg_printf("op load field u16\n");
511 tmp
= *(uint16_t *) stack_top
->u
.ptr
.ptr
;
512 if (stack_top
->u
.ptr
.rev_bo
)
513 tmp
= lttng_ust_bswap_16(tmp
);
514 stack_top
->u
.v
= tmp
;
515 stack_top
->type
= REG_U64
;
518 case OBJECT_TYPE_U32
:
522 dbg_printf("op load field u32\n");
523 tmp
= *(uint32_t *) stack_top
->u
.ptr
.ptr
;
524 if (stack_top
->u
.ptr
.rev_bo
)
525 tmp
= lttng_ust_bswap_32(tmp
);
526 stack_top
->u
.v
= tmp
;
527 stack_top
->type
= REG_U64
;
530 case OBJECT_TYPE_U64
:
534 dbg_printf("op load field u64\n");
535 tmp
= *(uint64_t *) stack_top
->u
.ptr
.ptr
;
536 if (stack_top
->u
.ptr
.rev_bo
)
537 tmp
= lttng_ust_bswap_64(tmp
);
538 stack_top
->u
.v
= tmp
;
539 stack_top
->type
= REG_U64
;
542 case OBJECT_TYPE_UNSIGNED_ENUM
:
546 dbg_printf("op load field unsigned enumeration\n");
547 tmp
= *(uint64_t *) stack_top
->u
.ptr
.ptr
;
548 if (stack_top
->u
.ptr
.rev_bo
)
549 tmp
= lttng_ust_bswap_64(tmp
);
550 stack_top
->u
.v
= tmp
;
551 stack_top
->type
= REG_U64
;
554 case OBJECT_TYPE_DOUBLE
:
555 memcpy(&stack_top
->u
.d
,
556 stack_top
->u
.ptr
.ptr
,
557 sizeof(struct literal_double
));
558 stack_top
->type
= REG_DOUBLE
;
560 case OBJECT_TYPE_STRING
:
564 dbg_printf("op load field string\n");
565 str
= (const char *) stack_top
->u
.ptr
.ptr
;
566 stack_top
->u
.s
.str
= str
;
567 if (unlikely(!stack_top
->u
.s
.str
)) {
568 dbg_printf("Interpreter warning: loading a NULL string.\n");
572 stack_top
->u
.s
.seq_len
= SIZE_MAX
;
573 stack_top
->u
.s
.literal_type
=
574 ESTACK_STRING_LITERAL_TYPE_NONE
;
575 stack_top
->type
= REG_STRING
;
578 case OBJECT_TYPE_STRING_SEQUENCE
:
582 dbg_printf("op load field string sequence\n");
583 ptr
= stack_top
->u
.ptr
.ptr
;
584 stack_top
->u
.s
.seq_len
= *(unsigned long *) ptr
;
585 stack_top
->u
.s
.str
= *(const char **) (ptr
+ sizeof(unsigned long));
586 stack_top
->type
= REG_STRING
;
587 if (unlikely(!stack_top
->u
.s
.str
)) {
588 dbg_printf("Interpreter warning: loading a NULL sequence.\n");
592 stack_top
->u
.s
.literal_type
=
593 ESTACK_STRING_LITERAL_TYPE_NONE
;
596 case OBJECT_TYPE_DYNAMIC
:
598 * Dynamic types in context are looked up
599 * by context get index.
603 case OBJECT_TYPE_SEQUENCE
:
604 case OBJECT_TYPE_ARRAY
:
605 case OBJECT_TYPE_STRUCT
:
606 case OBJECT_TYPE_VARIANT
:
607 ERR("Sequences, arrays, struct and variant cannot be loaded (nested types).");
618 int lttng_bytecode_interpret_format_output(struct estack_entry
*ax
,
619 struct lttng_interpreter_output
*output
)
626 output
->type
= LTTNG_INTERPRETER_TYPE_S64
;
627 output
->u
.s
= ax
->u
.v
;
630 output
->type
= LTTNG_INTERPRETER_TYPE_U64
;
631 output
->u
.u
= (uint64_t) ax
->u
.v
;
634 output
->type
= LTTNG_INTERPRETER_TYPE_DOUBLE
;
635 output
->u
.d
= ax
->u
.d
;
638 output
->type
= LTTNG_INTERPRETER_TYPE_STRING
;
639 output
->u
.str
.str
= ax
->u
.s
.str
;
640 output
->u
.str
.len
= ax
->u
.s
.seq_len
;
643 switch (ax
->u
.ptr
.object_type
) {
645 case OBJECT_TYPE_S16
:
646 case OBJECT_TYPE_S32
:
647 case OBJECT_TYPE_S64
:
649 case OBJECT_TYPE_U16
:
650 case OBJECT_TYPE_U32
:
651 case OBJECT_TYPE_U64
:
652 case OBJECT_TYPE_DOUBLE
:
653 case OBJECT_TYPE_STRING
:
654 case OBJECT_TYPE_STRING_SEQUENCE
:
655 ret
= dynamic_load_field(ax
);
658 /* Retry after loading ptr into stack top. */
660 case OBJECT_TYPE_SEQUENCE
:
661 output
->type
= LTTNG_INTERPRETER_TYPE_SEQUENCE
;
662 output
->u
.sequence
.ptr
= *(const char **) (ax
->u
.ptr
.ptr
+ sizeof(unsigned long));
663 output
->u
.sequence
.nr_elem
= *(unsigned long *) ax
->u
.ptr
.ptr
;
664 output
->u
.sequence
.nested_type
= lttng_ust_get_type_sequence(ax
->u
.ptr
.field
->type
)->elem_type
;
666 case OBJECT_TYPE_ARRAY
:
667 /* Skip count (unsigned long) */
668 output
->type
= LTTNG_INTERPRETER_TYPE_SEQUENCE
;
669 output
->u
.sequence
.ptr
= *(const char **) (ax
->u
.ptr
.ptr
+ sizeof(unsigned long));
670 output
->u
.sequence
.nr_elem
= lttng_ust_get_type_array(ax
->u
.ptr
.field
->type
)->length
;
671 output
->u
.sequence
.nested_type
= lttng_ust_get_type_array(ax
->u
.ptr
.field
->type
)->elem_type
;
673 case OBJECT_TYPE_SIGNED_ENUM
:
674 ret
= dynamic_load_field(ax
);
677 output
->type
= LTTNG_INTERPRETER_TYPE_SIGNED_ENUM
;
678 output
->u
.s
= ax
->u
.v
;
680 case OBJECT_TYPE_UNSIGNED_ENUM
:
681 ret
= dynamic_load_field(ax
);
684 output
->type
= LTTNG_INTERPRETER_TYPE_UNSIGNED_ENUM
;
685 output
->u
.u
= ax
->u
.v
;
687 case OBJECT_TYPE_STRUCT
:
688 case OBJECT_TYPE_VARIANT
:
694 case REG_STAR_GLOB_STRING
:
704 * Return LTTNG_UST_BYTECODE_INTERPRETER_OK on success.
705 * Return LTTNG_UST_BYTECODE_INTERPRETER_ERROR on error.
707 * For FILTER bytecode: expect a struct lttng_ust_bytecode_filter_ctx *
709 * For CAPTURE bytecode: expect a struct lttng_interpreter_output *
712 int lttng_bytecode_interpret(struct lttng_ust_bytecode_runtime
*ust_bytecode
,
713 const char *interpreter_stack_data
,
716 struct bytecode_runtime
*bytecode
= caa_container_of(ust_bytecode
, struct bytecode_runtime
, p
);
717 struct lttng_ust_ctx
*ctx
= lttng_ust_rcu_dereference(*ust_bytecode
->pctx
);
718 void *pc
, *next_pc
, *start_pc
;
719 int ret
= -EINVAL
, retval
= 0;
720 struct estack _stack
;
721 struct estack
*stack
= &_stack
;
722 register int64_t ax
= 0, bx
= 0;
723 register enum entry_type ax_t
= REG_UNKNOWN
, bx_t
= REG_UNKNOWN
;
724 register int top
= INTERPRETER_STACK_EMPTY
;
725 #ifndef INTERPRETER_USE_SWITCH
726 static void *dispatch
[NR_BYTECODE_OPS
] = {
727 [ BYTECODE_OP_UNKNOWN
] = &&LABEL_BYTECODE_OP_UNKNOWN
,
729 [ BYTECODE_OP_RETURN
] = &&LABEL_BYTECODE_OP_RETURN
,
732 [ BYTECODE_OP_MUL
] = &&LABEL_BYTECODE_OP_MUL
,
733 [ BYTECODE_OP_DIV
] = &&LABEL_BYTECODE_OP_DIV
,
734 [ BYTECODE_OP_MOD
] = &&LABEL_BYTECODE_OP_MOD
,
735 [ BYTECODE_OP_PLUS
] = &&LABEL_BYTECODE_OP_PLUS
,
736 [ BYTECODE_OP_MINUS
] = &&LABEL_BYTECODE_OP_MINUS
,
737 [ BYTECODE_OP_BIT_RSHIFT
] = &&LABEL_BYTECODE_OP_BIT_RSHIFT
,
738 [ BYTECODE_OP_BIT_LSHIFT
] = &&LABEL_BYTECODE_OP_BIT_LSHIFT
,
739 [ BYTECODE_OP_BIT_AND
] = &&LABEL_BYTECODE_OP_BIT_AND
,
740 [ BYTECODE_OP_BIT_OR
] = &&LABEL_BYTECODE_OP_BIT_OR
,
741 [ BYTECODE_OP_BIT_XOR
] = &&LABEL_BYTECODE_OP_BIT_XOR
,
743 /* binary comparators */
744 [ BYTECODE_OP_EQ
] = &&LABEL_BYTECODE_OP_EQ
,
745 [ BYTECODE_OP_NE
] = &&LABEL_BYTECODE_OP_NE
,
746 [ BYTECODE_OP_GT
] = &&LABEL_BYTECODE_OP_GT
,
747 [ BYTECODE_OP_LT
] = &&LABEL_BYTECODE_OP_LT
,
748 [ BYTECODE_OP_GE
] = &&LABEL_BYTECODE_OP_GE
,
749 [ BYTECODE_OP_LE
] = &&LABEL_BYTECODE_OP_LE
,
751 /* string binary comparator */
752 [ BYTECODE_OP_EQ_STRING
] = &&LABEL_BYTECODE_OP_EQ_STRING
,
753 [ BYTECODE_OP_NE_STRING
] = &&LABEL_BYTECODE_OP_NE_STRING
,
754 [ BYTECODE_OP_GT_STRING
] = &&LABEL_BYTECODE_OP_GT_STRING
,
755 [ BYTECODE_OP_LT_STRING
] = &&LABEL_BYTECODE_OP_LT_STRING
,
756 [ BYTECODE_OP_GE_STRING
] = &&LABEL_BYTECODE_OP_GE_STRING
,
757 [ BYTECODE_OP_LE_STRING
] = &&LABEL_BYTECODE_OP_LE_STRING
,
759 /* globbing pattern binary comparator */
760 [ BYTECODE_OP_EQ_STAR_GLOB_STRING
] = &&LABEL_BYTECODE_OP_EQ_STAR_GLOB_STRING
,
761 [ BYTECODE_OP_NE_STAR_GLOB_STRING
] = &&LABEL_BYTECODE_OP_NE_STAR_GLOB_STRING
,
763 /* s64 binary comparator */
764 [ BYTECODE_OP_EQ_S64
] = &&LABEL_BYTECODE_OP_EQ_S64
,
765 [ BYTECODE_OP_NE_S64
] = &&LABEL_BYTECODE_OP_NE_S64
,
766 [ BYTECODE_OP_GT_S64
] = &&LABEL_BYTECODE_OP_GT_S64
,
767 [ BYTECODE_OP_LT_S64
] = &&LABEL_BYTECODE_OP_LT_S64
,
768 [ BYTECODE_OP_GE_S64
] = &&LABEL_BYTECODE_OP_GE_S64
,
769 [ BYTECODE_OP_LE_S64
] = &&LABEL_BYTECODE_OP_LE_S64
,
771 /* double binary comparator */
772 [ BYTECODE_OP_EQ_DOUBLE
] = &&LABEL_BYTECODE_OP_EQ_DOUBLE
,
773 [ BYTECODE_OP_NE_DOUBLE
] = &&LABEL_BYTECODE_OP_NE_DOUBLE
,
774 [ BYTECODE_OP_GT_DOUBLE
] = &&LABEL_BYTECODE_OP_GT_DOUBLE
,
775 [ BYTECODE_OP_LT_DOUBLE
] = &&LABEL_BYTECODE_OP_LT_DOUBLE
,
776 [ BYTECODE_OP_GE_DOUBLE
] = &&LABEL_BYTECODE_OP_GE_DOUBLE
,
777 [ BYTECODE_OP_LE_DOUBLE
] = &&LABEL_BYTECODE_OP_LE_DOUBLE
,
779 /* Mixed S64-double binary comparators */
780 [ BYTECODE_OP_EQ_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_EQ_DOUBLE_S64
,
781 [ BYTECODE_OP_NE_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_NE_DOUBLE_S64
,
782 [ BYTECODE_OP_GT_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_GT_DOUBLE_S64
,
783 [ BYTECODE_OP_LT_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_LT_DOUBLE_S64
,
784 [ BYTECODE_OP_GE_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_GE_DOUBLE_S64
,
785 [ BYTECODE_OP_LE_DOUBLE_S64
] = &&LABEL_BYTECODE_OP_LE_DOUBLE_S64
,
787 [ BYTECODE_OP_EQ_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_EQ_S64_DOUBLE
,
788 [ BYTECODE_OP_NE_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_NE_S64_DOUBLE
,
789 [ BYTECODE_OP_GT_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_GT_S64_DOUBLE
,
790 [ BYTECODE_OP_LT_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_LT_S64_DOUBLE
,
791 [ BYTECODE_OP_GE_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_GE_S64_DOUBLE
,
792 [ BYTECODE_OP_LE_S64_DOUBLE
] = &&LABEL_BYTECODE_OP_LE_S64_DOUBLE
,
795 [ BYTECODE_OP_UNARY_PLUS
] = &&LABEL_BYTECODE_OP_UNARY_PLUS
,
796 [ BYTECODE_OP_UNARY_MINUS
] = &&LABEL_BYTECODE_OP_UNARY_MINUS
,
797 [ BYTECODE_OP_UNARY_NOT
] = &&LABEL_BYTECODE_OP_UNARY_NOT
,
798 [ BYTECODE_OP_UNARY_PLUS_S64
] = &&LABEL_BYTECODE_OP_UNARY_PLUS_S64
,
799 [ BYTECODE_OP_UNARY_MINUS_S64
] = &&LABEL_BYTECODE_OP_UNARY_MINUS_S64
,
800 [ BYTECODE_OP_UNARY_NOT_S64
] = &&LABEL_BYTECODE_OP_UNARY_NOT_S64
,
801 [ BYTECODE_OP_UNARY_PLUS_DOUBLE
] = &&LABEL_BYTECODE_OP_UNARY_PLUS_DOUBLE
,
802 [ BYTECODE_OP_UNARY_MINUS_DOUBLE
] = &&LABEL_BYTECODE_OP_UNARY_MINUS_DOUBLE
,
803 [ BYTECODE_OP_UNARY_NOT_DOUBLE
] = &&LABEL_BYTECODE_OP_UNARY_NOT_DOUBLE
,
806 [ BYTECODE_OP_AND
] = &&LABEL_BYTECODE_OP_AND
,
807 [ BYTECODE_OP_OR
] = &&LABEL_BYTECODE_OP_OR
,
810 [ BYTECODE_OP_LOAD_FIELD_REF
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF
,
811 [ BYTECODE_OP_LOAD_FIELD_REF_STRING
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_STRING
,
812 [ BYTECODE_OP_LOAD_FIELD_REF_SEQUENCE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_SEQUENCE
,
813 [ BYTECODE_OP_LOAD_FIELD_REF_S64
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_S64
,
814 [ BYTECODE_OP_LOAD_FIELD_REF_DOUBLE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_REF_DOUBLE
,
816 /* load from immediate operand */
817 [ BYTECODE_OP_LOAD_STRING
] = &&LABEL_BYTECODE_OP_LOAD_STRING
,
818 [ BYTECODE_OP_LOAD_STAR_GLOB_STRING
] = &&LABEL_BYTECODE_OP_LOAD_STAR_GLOB_STRING
,
819 [ BYTECODE_OP_LOAD_S64
] = &&LABEL_BYTECODE_OP_LOAD_S64
,
820 [ BYTECODE_OP_LOAD_DOUBLE
] = &&LABEL_BYTECODE_OP_LOAD_DOUBLE
,
823 [ BYTECODE_OP_CAST_TO_S64
] = &&LABEL_BYTECODE_OP_CAST_TO_S64
,
824 [ BYTECODE_OP_CAST_DOUBLE_TO_S64
] = &&LABEL_BYTECODE_OP_CAST_DOUBLE_TO_S64
,
825 [ BYTECODE_OP_CAST_NOP
] = &&LABEL_BYTECODE_OP_CAST_NOP
,
827 /* get context ref */
828 [ BYTECODE_OP_GET_CONTEXT_REF
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF
,
829 [ BYTECODE_OP_GET_CONTEXT_REF_STRING
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF_STRING
,
830 [ BYTECODE_OP_GET_CONTEXT_REF_S64
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF_S64
,
831 [ BYTECODE_OP_GET_CONTEXT_REF_DOUBLE
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_REF_DOUBLE
,
833 /* Instructions for recursive traversal through composed types. */
834 [ BYTECODE_OP_GET_CONTEXT_ROOT
] = &&LABEL_BYTECODE_OP_GET_CONTEXT_ROOT
,
835 [ BYTECODE_OP_GET_APP_CONTEXT_ROOT
] = &&LABEL_BYTECODE_OP_GET_APP_CONTEXT_ROOT
,
836 [ BYTECODE_OP_GET_PAYLOAD_ROOT
] = &&LABEL_BYTECODE_OP_GET_PAYLOAD_ROOT
,
838 [ BYTECODE_OP_GET_SYMBOL
] = &&LABEL_BYTECODE_OP_GET_SYMBOL
,
839 [ BYTECODE_OP_GET_SYMBOL_FIELD
] = &&LABEL_BYTECODE_OP_GET_SYMBOL_FIELD
,
840 [ BYTECODE_OP_GET_INDEX_U16
] = &&LABEL_BYTECODE_OP_GET_INDEX_U16
,
841 [ BYTECODE_OP_GET_INDEX_U64
] = &&LABEL_BYTECODE_OP_GET_INDEX_U64
,
843 [ BYTECODE_OP_LOAD_FIELD
] = &&LABEL_BYTECODE_OP_LOAD_FIELD
,
844 [ BYTECODE_OP_LOAD_FIELD_S8
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S8
,
845 [ BYTECODE_OP_LOAD_FIELD_S16
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S16
,
846 [ BYTECODE_OP_LOAD_FIELD_S32
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S32
,
847 [ BYTECODE_OP_LOAD_FIELD_S64
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_S64
,
848 [ BYTECODE_OP_LOAD_FIELD_U8
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U8
,
849 [ BYTECODE_OP_LOAD_FIELD_U16
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U16
,
850 [ BYTECODE_OP_LOAD_FIELD_U32
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U32
,
851 [ BYTECODE_OP_LOAD_FIELD_U64
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_U64
,
852 [ BYTECODE_OP_LOAD_FIELD_STRING
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_STRING
,
853 [ BYTECODE_OP_LOAD_FIELD_SEQUENCE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_SEQUENCE
,
854 [ BYTECODE_OP_LOAD_FIELD_DOUBLE
] = &&LABEL_BYTECODE_OP_LOAD_FIELD_DOUBLE
,
856 [ BYTECODE_OP_UNARY_BIT_NOT
] = &&LABEL_BYTECODE_OP_UNARY_BIT_NOT
,
858 [ BYTECODE_OP_RETURN_S64
] = &&LABEL_BYTECODE_OP_RETURN_S64
,
860 #endif /* #ifndef INTERPRETER_USE_SWITCH */
864 OP(BYTECODE_OP_UNKNOWN
):
865 OP(BYTECODE_OP_LOAD_FIELD_REF
):
866 #ifdef INTERPRETER_USE_SWITCH
868 #endif /* INTERPRETER_USE_SWITCH */
869 ERR("unknown bytecode op %u",
870 (unsigned int) *(bytecode_opcode_t
*) pc
);
874 OP(BYTECODE_OP_RETURN
):
875 /* LTTNG_UST_BYTECODE_INTERPRETER_ERROR or LTTNG_UST_BYTECODE_INTERPRETER_OK */
876 /* Handle dynamic typing. */
877 switch (estack_ax_t
) {
880 retval
= !!estack_ax_v
;
885 if (ust_bytecode
->type
!= LTTNG_UST_BYTECODE_TYPE_CAPTURE
) {
891 case REG_STAR_GLOB_STRING
:
900 OP(BYTECODE_OP_RETURN_S64
):
901 /* LTTNG_UST_BYTECODE_INTERPRETER_ERROR or LTTNG_UST_BYTECODE_INTERPRETER_OK */
902 retval
= !!estack_ax_v
;
910 OP(BYTECODE_OP_PLUS
):
911 OP(BYTECODE_OP_MINUS
):
912 ERR("unsupported bytecode op %u",
913 (unsigned int) *(bytecode_opcode_t
*) pc
);
919 /* Dynamic typing. */
920 switch (estack_ax_t
) {
921 case REG_S64
: /* Fall-through */
923 switch (estack_bx_t
) {
924 case REG_S64
: /* Fall-through */
926 JUMP_TO(BYTECODE_OP_EQ_S64
);
928 JUMP_TO(BYTECODE_OP_EQ_DOUBLE_S64
);
929 case REG_STRING
: /* Fall-through */
930 case REG_STAR_GLOB_STRING
:
934 ERR("Unknown interpreter register type (%d)",
941 switch (estack_bx_t
) {
942 case REG_S64
: /* Fall-through */
944 JUMP_TO(BYTECODE_OP_EQ_S64_DOUBLE
);
946 JUMP_TO(BYTECODE_OP_EQ_DOUBLE
);
947 case REG_STRING
: /* Fall-through */
948 case REG_STAR_GLOB_STRING
:
952 ERR("Unknown interpreter register type (%d)",
959 switch (estack_bx_t
) {
960 case REG_S64
: /* Fall-through */
961 case REG_U64
: /* Fall-through */
966 JUMP_TO(BYTECODE_OP_EQ_STRING
);
967 case REG_STAR_GLOB_STRING
:
968 JUMP_TO(BYTECODE_OP_EQ_STAR_GLOB_STRING
);
970 ERR("Unknown interpreter register type (%d)",
976 case REG_STAR_GLOB_STRING
:
977 switch (estack_bx_t
) {
978 case REG_S64
: /* Fall-through */
979 case REG_U64
: /* Fall-through */
984 JUMP_TO(BYTECODE_OP_EQ_STAR_GLOB_STRING
);
985 case REG_STAR_GLOB_STRING
:
989 ERR("Unknown interpreter register type (%d)",
996 ERR("Unknown interpreter register type (%d)",
1004 /* Dynamic typing. */
1005 switch (estack_ax_t
) {
1006 case REG_S64
: /* Fall-through */
1008 switch (estack_bx_t
) {
1009 case REG_S64
: /* Fall-through */
1011 JUMP_TO(BYTECODE_OP_NE_S64
);
1013 JUMP_TO(BYTECODE_OP_NE_DOUBLE_S64
);
1014 case REG_STRING
: /* Fall-through */
1015 case REG_STAR_GLOB_STRING
:
1019 ERR("Unknown interpreter register type (%d)",
1026 switch (estack_bx_t
) {
1027 case REG_S64
: /* Fall-through */
1029 JUMP_TO(BYTECODE_OP_NE_S64_DOUBLE
);
1031 JUMP_TO(BYTECODE_OP_NE_DOUBLE
);
1032 case REG_STRING
: /* Fall-through */
1033 case REG_STAR_GLOB_STRING
:
1037 ERR("Unknown interpreter register type (%d)",
1044 switch (estack_bx_t
) {
1045 case REG_S64
: /* Fall-through */
1051 JUMP_TO(BYTECODE_OP_NE_STRING
);
1052 case REG_STAR_GLOB_STRING
:
1053 JUMP_TO(BYTECODE_OP_NE_STAR_GLOB_STRING
);
1055 ERR("Unknown interpreter register type (%d)",
1061 case REG_STAR_GLOB_STRING
:
1062 switch (estack_bx_t
) {
1063 case REG_S64
: /* Fall-through */
1069 JUMP_TO(BYTECODE_OP_NE_STAR_GLOB_STRING
);
1070 case REG_STAR_GLOB_STRING
:
1074 ERR("Unknown interpreter register type (%d)",
1081 ERR("Unknown interpreter register type (%d)",
1089 /* Dynamic typing. */
1090 switch (estack_ax_t
) {
1091 case REG_S64
: /* Fall-through */
1093 switch (estack_bx_t
) {
1094 case REG_S64
: /* Fall-through */
1096 JUMP_TO(BYTECODE_OP_GT_S64
);
1098 JUMP_TO(BYTECODE_OP_GT_DOUBLE_S64
);
1099 case REG_STRING
: /* Fall-through */
1100 case REG_STAR_GLOB_STRING
:
1104 ERR("Unknown interpreter register type (%d)",
1111 switch (estack_bx_t
) {
1112 case REG_S64
: /* Fall-through */
1114 JUMP_TO(BYTECODE_OP_GT_S64_DOUBLE
);
1116 JUMP_TO(BYTECODE_OP_GT_DOUBLE
);
1117 case REG_STRING
: /* Fall-through */
1118 case REG_STAR_GLOB_STRING
:
1122 ERR("Unknown interpreter register type (%d)",
1129 switch (estack_bx_t
) {
1130 case REG_S64
: /* Fall-through */
1131 case REG_U64
: /* Fall-through */
1132 case REG_DOUBLE
: /* Fall-through */
1133 case REG_STAR_GLOB_STRING
:
1137 JUMP_TO(BYTECODE_OP_GT_STRING
);
1139 ERR("Unknown interpreter register type (%d)",
1146 ERR("Unknown interpreter register type (%d)",
1154 /* Dynamic typing. */
1155 switch (estack_ax_t
) {
1156 case REG_S64
: /* Fall-through */
1158 switch (estack_bx_t
) {
1159 case REG_S64
: /* Fall-through */
1161 JUMP_TO(BYTECODE_OP_LT_S64
);
1163 JUMP_TO(BYTECODE_OP_LT_DOUBLE_S64
);
1164 case REG_STRING
: /* Fall-through */
1165 case REG_STAR_GLOB_STRING
:
1169 ERR("Unknown interpreter register type (%d)",
1176 switch (estack_bx_t
) {
1177 case REG_S64
: /* Fall-through */
1179 JUMP_TO(BYTECODE_OP_LT_S64_DOUBLE
);
1181 JUMP_TO(BYTECODE_OP_LT_DOUBLE
);
1182 case REG_STRING
: /* Fall-through */
1183 case REG_STAR_GLOB_STRING
:
1187 ERR("Unknown interpreter register type (%d)",
1194 switch (estack_bx_t
) {
1195 case REG_S64
: /* Fall-through */
1196 case REG_U64
: /* Fall-through */
1197 case REG_DOUBLE
: /* Fall-through */
1198 case REG_STAR_GLOB_STRING
:
1202 JUMP_TO(BYTECODE_OP_LT_STRING
);
1204 ERR("Unknown interpreter register type (%d)",
1211 ERR("Unknown interpreter register type (%d)",
1219 /* Dynamic typing. */
1220 switch (estack_ax_t
) {
1221 case REG_S64
: /* Fall-through */
1223 switch (estack_bx_t
) {
1224 case REG_S64
: /* Fall-through */
1226 JUMP_TO(BYTECODE_OP_GE_S64
);
1228 JUMP_TO(BYTECODE_OP_GE_DOUBLE_S64
);
1229 case REG_STRING
: /* Fall-through */
1230 case REG_STAR_GLOB_STRING
:
1234 ERR("Unknown interpreter register type (%d)",
1241 switch (estack_bx_t
) {
1242 case REG_S64
: /* Fall-through */
1244 JUMP_TO(BYTECODE_OP_GE_S64_DOUBLE
);
1246 JUMP_TO(BYTECODE_OP_GE_DOUBLE
);
1247 case REG_STRING
: /* Fall-through */
1248 case REG_STAR_GLOB_STRING
:
1252 ERR("Unknown interpreter register type (%d)",
1259 switch (estack_bx_t
) {
1260 case REG_S64
: /* Fall-through */
1261 case REG_U64
: /* Fall-through */
1262 case REG_DOUBLE
: /* Fall-through */
1263 case REG_STAR_GLOB_STRING
:
1267 JUMP_TO(BYTECODE_OP_GE_STRING
);
1269 ERR("Unknown interpreter register type (%d)",
1276 ERR("Unknown interpreter register type (%d)",
1284 /* Dynamic typing. */
1285 switch (estack_ax_t
) {
1286 case REG_S64
: /* Fall-through */
1288 switch (estack_bx_t
) {
1289 case REG_S64
: /* Fall-through */
1291 JUMP_TO(BYTECODE_OP_LE_S64
);
1293 JUMP_TO(BYTECODE_OP_LE_DOUBLE_S64
);
1294 case REG_STRING
: /* Fall-through */
1295 case REG_STAR_GLOB_STRING
:
1299 ERR("Unknown interpreter register type (%d)",
1306 switch (estack_bx_t
) {
1307 case REG_S64
: /* Fall-through */
1309 JUMP_TO(BYTECODE_OP_LE_S64_DOUBLE
);
1311 JUMP_TO(BYTECODE_OP_LE_DOUBLE
);
1312 case REG_STRING
: /* Fall-through */
1313 case REG_STAR_GLOB_STRING
:
1317 ERR("Unknown interpreter register type (%d)",
1324 switch (estack_bx_t
) {
1325 case REG_S64
: /* Fall-through */
1326 case REG_U64
: /* Fall-through */
1327 case REG_DOUBLE
: /* Fall-through */
1328 case REG_STAR_GLOB_STRING
:
1332 JUMP_TO(BYTECODE_OP_LE_STRING
);
1334 ERR("Unknown interpreter register type (%d)",
1341 ERR("Unknown interpreter register type (%d)",
1348 OP(BYTECODE_OP_EQ_STRING
):
1352 res
= (stack_strcmp(stack
, top
, "==") == 0);
1353 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1355 estack_ax_t
= REG_S64
;
1356 next_pc
+= sizeof(struct binary_op
);
1359 OP(BYTECODE_OP_NE_STRING
):
1363 res
= (stack_strcmp(stack
, top
, "!=") != 0);
1364 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1366 estack_ax_t
= REG_S64
;
1367 next_pc
+= sizeof(struct binary_op
);
1370 OP(BYTECODE_OP_GT_STRING
):
1374 res
= (stack_strcmp(stack
, top
, ">") > 0);
1375 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1377 estack_ax_t
= REG_S64
;
1378 next_pc
+= sizeof(struct binary_op
);
1381 OP(BYTECODE_OP_LT_STRING
):
1385 res
= (stack_strcmp(stack
, top
, "<") < 0);
1386 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1388 estack_ax_t
= REG_S64
;
1389 next_pc
+= sizeof(struct binary_op
);
1392 OP(BYTECODE_OP_GE_STRING
):
1396 res
= (stack_strcmp(stack
, top
, ">=") >= 0);
1397 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1399 estack_ax_t
= REG_S64
;
1400 next_pc
+= sizeof(struct binary_op
);
1403 OP(BYTECODE_OP_LE_STRING
):
1407 res
= (stack_strcmp(stack
, top
, "<=") <= 0);
1408 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1410 estack_ax_t
= REG_S64
;
1411 next_pc
+= sizeof(struct binary_op
);
1415 OP(BYTECODE_OP_EQ_STAR_GLOB_STRING
):
1419 res
= (stack_star_glob_match(stack
, top
, "==") == 0);
1420 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1422 estack_ax_t
= REG_S64
;
1423 next_pc
+= sizeof(struct binary_op
);
1426 OP(BYTECODE_OP_NE_STAR_GLOB_STRING
):
1430 res
= (stack_star_glob_match(stack
, top
, "!=") != 0);
1431 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1433 estack_ax_t
= REG_S64
;
1434 next_pc
+= sizeof(struct binary_op
);
1438 OP(BYTECODE_OP_EQ_S64
):
1442 res
= (estack_bx_v
== estack_ax_v
);
1443 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1445 estack_ax_t
= REG_S64
;
1446 next_pc
+= sizeof(struct binary_op
);
1449 OP(BYTECODE_OP_NE_S64
):
1453 res
= (estack_bx_v
!= estack_ax_v
);
1454 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1456 estack_ax_t
= REG_S64
;
1457 next_pc
+= sizeof(struct binary_op
);
1460 OP(BYTECODE_OP_GT_S64
):
1464 res
= (estack_bx_v
> estack_ax_v
);
1465 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1467 estack_ax_t
= REG_S64
;
1468 next_pc
+= sizeof(struct binary_op
);
1471 OP(BYTECODE_OP_LT_S64
):
1475 res
= (estack_bx_v
< estack_ax_v
);
1476 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1478 estack_ax_t
= REG_S64
;
1479 next_pc
+= sizeof(struct binary_op
);
1482 OP(BYTECODE_OP_GE_S64
):
1486 res
= (estack_bx_v
>= estack_ax_v
);
1487 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1489 estack_ax_t
= REG_S64
;
1490 next_pc
+= sizeof(struct binary_op
);
1493 OP(BYTECODE_OP_LE_S64
):
1497 res
= (estack_bx_v
<= estack_ax_v
);
1498 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1500 estack_ax_t
= REG_S64
;
1501 next_pc
+= sizeof(struct binary_op
);
1505 OP(BYTECODE_OP_EQ_DOUBLE
):
1509 res
= (estack_bx(stack
, top
)->u
.d
== estack_ax(stack
, top
)->u
.d
);
1510 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1512 estack_ax_t
= REG_S64
;
1513 next_pc
+= sizeof(struct binary_op
);
1516 OP(BYTECODE_OP_NE_DOUBLE
):
1520 res
= (estack_bx(stack
, top
)->u
.d
!= estack_ax(stack
, top
)->u
.d
);
1521 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1523 estack_ax_t
= REG_S64
;
1524 next_pc
+= sizeof(struct binary_op
);
1527 OP(BYTECODE_OP_GT_DOUBLE
):
1531 res
= (estack_bx(stack
, top
)->u
.d
> estack_ax(stack
, top
)->u
.d
);
1532 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1534 estack_ax_t
= REG_S64
;
1535 next_pc
+= sizeof(struct binary_op
);
1538 OP(BYTECODE_OP_LT_DOUBLE
):
1542 res
= (estack_bx(stack
, top
)->u
.d
< estack_ax(stack
, top
)->u
.d
);
1543 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1545 estack_ax_t
= REG_S64
;
1546 next_pc
+= sizeof(struct binary_op
);
1549 OP(BYTECODE_OP_GE_DOUBLE
):
1553 res
= (estack_bx(stack
, top
)->u
.d
>= estack_ax(stack
, top
)->u
.d
);
1554 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1556 estack_ax_t
= REG_S64
;
1557 next_pc
+= sizeof(struct binary_op
);
1560 OP(BYTECODE_OP_LE_DOUBLE
):
1564 res
= (estack_bx(stack
, top
)->u
.d
<= estack_ax(stack
, top
)->u
.d
);
1565 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1567 estack_ax_t
= REG_S64
;
1568 next_pc
+= sizeof(struct binary_op
);
1572 /* Mixed S64-double binary comparators */
1573 OP(BYTECODE_OP_EQ_DOUBLE_S64
):
1577 res
= (estack_bx(stack
, top
)->u
.d
== estack_ax_v
);
1578 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1580 estack_ax_t
= REG_S64
;
1581 next_pc
+= sizeof(struct binary_op
);
1584 OP(BYTECODE_OP_NE_DOUBLE_S64
):
1588 res
= (estack_bx(stack
, top
)->u
.d
!= estack_ax_v
);
1589 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1591 estack_ax_t
= REG_S64
;
1592 next_pc
+= sizeof(struct binary_op
);
1595 OP(BYTECODE_OP_GT_DOUBLE_S64
):
1599 res
= (estack_bx(stack
, top
)->u
.d
> estack_ax_v
);
1600 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1602 estack_ax_t
= REG_S64
;
1603 next_pc
+= sizeof(struct binary_op
);
1606 OP(BYTECODE_OP_LT_DOUBLE_S64
):
1610 res
= (estack_bx(stack
, top
)->u
.d
< estack_ax_v
);
1611 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1613 estack_ax_t
= REG_S64
;
1614 next_pc
+= sizeof(struct binary_op
);
1617 OP(BYTECODE_OP_GE_DOUBLE_S64
):
1621 res
= (estack_bx(stack
, top
)->u
.d
>= estack_ax_v
);
1622 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1624 estack_ax_t
= REG_S64
;
1625 next_pc
+= sizeof(struct binary_op
);
1628 OP(BYTECODE_OP_LE_DOUBLE_S64
):
1632 res
= (estack_bx(stack
, top
)->u
.d
<= estack_ax_v
);
1633 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1635 estack_ax_t
= REG_S64
;
1636 next_pc
+= sizeof(struct binary_op
);
1640 OP(BYTECODE_OP_EQ_S64_DOUBLE
):
1644 res
= (estack_bx_v
== estack_ax(stack
, top
)->u
.d
);
1645 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1647 estack_ax_t
= REG_S64
;
1648 next_pc
+= sizeof(struct binary_op
);
1651 OP(BYTECODE_OP_NE_S64_DOUBLE
):
1655 res
= (estack_bx_v
!= estack_ax(stack
, top
)->u
.d
);
1656 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1658 estack_ax_t
= REG_S64
;
1659 next_pc
+= sizeof(struct binary_op
);
1662 OP(BYTECODE_OP_GT_S64_DOUBLE
):
1666 res
= (estack_bx_v
> estack_ax(stack
, top
)->u
.d
);
1667 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1669 estack_ax_t
= REG_S64
;
1670 next_pc
+= sizeof(struct binary_op
);
1673 OP(BYTECODE_OP_LT_S64_DOUBLE
):
1677 res
= (estack_bx_v
< estack_ax(stack
, top
)->u
.d
);
1678 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1680 estack_ax_t
= REG_S64
;
1681 next_pc
+= sizeof(struct binary_op
);
1684 OP(BYTECODE_OP_GE_S64_DOUBLE
):
1688 res
= (estack_bx_v
>= estack_ax(stack
, top
)->u
.d
);
1689 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1691 estack_ax_t
= REG_S64
;
1692 next_pc
+= sizeof(struct binary_op
);
1695 OP(BYTECODE_OP_LE_S64_DOUBLE
):
1699 res
= (estack_bx_v
<= estack_ax(stack
, top
)->u
.d
);
1700 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1702 estack_ax_t
= REG_S64
;
1703 next_pc
+= sizeof(struct binary_op
);
1706 OP(BYTECODE_OP_BIT_RSHIFT
):
1710 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1715 /* Catch undefined behavior. */
1716 if (caa_unlikely(estack_ax_v
< 0 || estack_ax_v
>= 64)) {
1720 res
= ((uint64_t) estack_bx_v
>> (uint32_t) estack_ax_v
);
1721 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1723 estack_ax_t
= REG_U64
;
1724 next_pc
+= sizeof(struct binary_op
);
1727 OP(BYTECODE_OP_BIT_LSHIFT
):
1731 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1736 /* Catch undefined behavior. */
1737 if (caa_unlikely(estack_ax_v
< 0 || estack_ax_v
>= 64)) {
1741 res
= ((uint64_t) estack_bx_v
<< (uint32_t) estack_ax_v
);
1742 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1744 estack_ax_t
= REG_U64
;
1745 next_pc
+= sizeof(struct binary_op
);
1748 OP(BYTECODE_OP_BIT_AND
):
1752 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1757 res
= ((uint64_t) estack_bx_v
& (uint64_t) estack_ax_v
);
1758 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1760 estack_ax_t
= REG_U64
;
1761 next_pc
+= sizeof(struct binary_op
);
1764 OP(BYTECODE_OP_BIT_OR
):
1768 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1773 res
= ((uint64_t) estack_bx_v
| (uint64_t) estack_ax_v
);
1774 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1776 estack_ax_t
= REG_U64
;
1777 next_pc
+= sizeof(struct binary_op
);
1780 OP(BYTECODE_OP_BIT_XOR
):
1784 if (!IS_INTEGER_REGISTER(estack_ax_t
) || !IS_INTEGER_REGISTER(estack_bx_t
)) {
1789 res
= ((uint64_t) estack_bx_v
^ (uint64_t) estack_ax_v
);
1790 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1792 estack_ax_t
= REG_U64
;
1793 next_pc
+= sizeof(struct binary_op
);
1798 OP(BYTECODE_OP_UNARY_PLUS
):
1800 /* Dynamic typing. */
1801 switch (estack_ax_t
) {
1802 case REG_S64
: /* Fall-through. */
1804 JUMP_TO(BYTECODE_OP_UNARY_PLUS_S64
);
1806 JUMP_TO(BYTECODE_OP_UNARY_PLUS_DOUBLE
);
1807 case REG_STRING
: /* Fall-through */
1808 case REG_STAR_GLOB_STRING
:
1812 ERR("Unknown interpreter register type (%d)",
1818 OP(BYTECODE_OP_UNARY_MINUS
):
1820 /* Dynamic typing. */
1821 switch (estack_ax_t
) {
1822 case REG_S64
: /* Fall-through. */
1824 JUMP_TO(BYTECODE_OP_UNARY_MINUS_S64
);
1826 JUMP_TO(BYTECODE_OP_UNARY_MINUS_DOUBLE
);
1827 case REG_STRING
: /* Fall-through */
1828 case REG_STAR_GLOB_STRING
:
1832 ERR("Unknown interpreter register type (%d)",
1838 OP(BYTECODE_OP_UNARY_NOT
):
1840 /* Dynamic typing. */
1841 switch (estack_ax_t
) {
1842 case REG_S64
: /* Fall-through. */
1844 JUMP_TO(BYTECODE_OP_UNARY_NOT_S64
);
1846 JUMP_TO(BYTECODE_OP_UNARY_NOT_DOUBLE
);
1847 case REG_STRING
: /* Fall-through */
1848 case REG_STAR_GLOB_STRING
:
1852 ERR("Unknown interpreter register type (%d)",
1857 next_pc
+= sizeof(struct unary_op
);
1861 OP(BYTECODE_OP_UNARY_BIT_NOT
):
1863 /* Dynamic typing. */
1864 if (!IS_INTEGER_REGISTER(estack_ax_t
)) {
1869 estack_ax_v
= ~(uint64_t) estack_ax_v
;
1870 estack_ax_t
= REG_U64
;
1871 next_pc
+= sizeof(struct unary_op
);
1875 OP(BYTECODE_OP_UNARY_PLUS_S64
):
1876 OP(BYTECODE_OP_UNARY_PLUS_DOUBLE
):
1878 next_pc
+= sizeof(struct unary_op
);
1881 OP(BYTECODE_OP_UNARY_MINUS_S64
):
1883 estack_ax_v
= -estack_ax_v
;
1884 next_pc
+= sizeof(struct unary_op
);
1887 OP(BYTECODE_OP_UNARY_MINUS_DOUBLE
):
1889 estack_ax(stack
, top
)->u
.d
= -estack_ax(stack
, top
)->u
.d
;
1890 next_pc
+= sizeof(struct unary_op
);
1893 OP(BYTECODE_OP_UNARY_NOT_S64
):
1895 estack_ax_v
= !estack_ax_v
;
1896 estack_ax_t
= REG_S64
;
1897 next_pc
+= sizeof(struct unary_op
);
1900 OP(BYTECODE_OP_UNARY_NOT_DOUBLE
):
1902 estack_ax_v
= !estack_ax(stack
, top
)->u
.d
;
1903 estack_ax_t
= REG_S64
;
1904 next_pc
+= sizeof(struct unary_op
);
1909 OP(BYTECODE_OP_AND
):
1911 struct logical_op
*insn
= (struct logical_op
*) pc
;
1913 if (estack_ax_t
!= REG_S64
&& estack_ax_t
!= REG_U64
) {
1917 /* If AX is 0, skip and evaluate to 0 */
1918 if (unlikely(estack_ax_v
== 0)) {
1919 dbg_printf("Jumping to bytecode offset %u\n",
1920 (unsigned int) insn
->skip_offset
);
1921 next_pc
= start_pc
+ insn
->skip_offset
;
1923 /* Pop 1 when jump not taken */
1924 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1925 next_pc
+= sizeof(struct logical_op
);
1931 struct logical_op
*insn
= (struct logical_op
*) pc
;
1933 if (estack_ax_t
!= REG_S64
&& estack_ax_t
!= REG_U64
) {
1937 /* If AX is nonzero, skip and evaluate to 1 */
1938 if (unlikely(estack_ax_v
!= 0)) {
1940 dbg_printf("Jumping to bytecode offset %u\n",
1941 (unsigned int) insn
->skip_offset
);
1942 next_pc
= start_pc
+ insn
->skip_offset
;
1944 /* Pop 1 when jump not taken */
1945 estack_pop(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1946 next_pc
+= sizeof(struct logical_op
);
1952 /* load field ref */
1953 OP(BYTECODE_OP_LOAD_FIELD_REF_STRING
):
1955 struct load_op
*insn
= (struct load_op
*) pc
;
1956 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
1958 dbg_printf("load field ref offset %u type string\n",
1960 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1961 estack_ax(stack
, top
)->u
.s
.str
=
1962 *(const char * const *) &interpreter_stack_data
[ref
->offset
];
1963 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
1964 dbg_printf("Interpreter warning: loading a NULL string.\n");
1968 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
1969 estack_ax(stack
, top
)->u
.s
.literal_type
=
1970 ESTACK_STRING_LITERAL_TYPE_NONE
;
1971 estack_ax_t
= REG_STRING
;
1972 dbg_printf("ref load string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
1973 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
1977 OP(BYTECODE_OP_LOAD_FIELD_REF_SEQUENCE
):
1979 struct load_op
*insn
= (struct load_op
*) pc
;
1980 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
1982 dbg_printf("load field ref offset %u type sequence\n",
1984 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
1985 estack_ax(stack
, top
)->u
.s
.seq_len
=
1986 *(unsigned long *) &interpreter_stack_data
[ref
->offset
];
1987 estack_ax(stack
, top
)->u
.s
.str
=
1988 *(const char **) (&interpreter_stack_data
[ref
->offset
1989 + sizeof(unsigned long)]);
1990 estack_ax_t
= REG_STRING
;
1991 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
1992 dbg_printf("Interpreter warning: loading a NULL sequence.\n");
1996 estack_ax(stack
, top
)->u
.s
.literal_type
=
1997 ESTACK_STRING_LITERAL_TYPE_NONE
;
1998 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2002 OP(BYTECODE_OP_LOAD_FIELD_REF_S64
):
2004 struct load_op
*insn
= (struct load_op
*) pc
;
2005 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2007 dbg_printf("load field ref offset %u type s64\n",
2009 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2011 ((struct literal_numeric
*) &interpreter_stack_data
[ref
->offset
])->v
;
2012 estack_ax_t
= REG_S64
;
2013 dbg_printf("ref load s64 %" PRIi64
"\n", estack_ax_v
);
2014 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2018 OP(BYTECODE_OP_LOAD_FIELD_REF_DOUBLE
):
2020 struct load_op
*insn
= (struct load_op
*) pc
;
2021 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2023 dbg_printf("load field ref offset %u type double\n",
2025 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2026 memcpy(&estack_ax(stack
, top
)->u
.d
, &interpreter_stack_data
[ref
->offset
],
2027 sizeof(struct literal_double
));
2028 estack_ax_t
= REG_DOUBLE
;
2029 dbg_printf("ref load double %g\n", estack_ax(stack
, top
)->u
.d
);
2030 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2034 /* load from immediate operand */
2035 OP(BYTECODE_OP_LOAD_STRING
):
2037 struct load_op
*insn
= (struct load_op
*) pc
;
2039 dbg_printf("load string %s\n", insn
->data
);
2040 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2041 estack_ax(stack
, top
)->u
.s
.str
= insn
->data
;
2042 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2043 estack_ax(stack
, top
)->u
.s
.literal_type
=
2044 ESTACK_STRING_LITERAL_TYPE_PLAIN
;
2045 estack_ax_t
= REG_STRING
;
2046 next_pc
+= sizeof(struct load_op
) + strlen(insn
->data
) + 1;
2050 OP(BYTECODE_OP_LOAD_STAR_GLOB_STRING
):
2052 struct load_op
*insn
= (struct load_op
*) pc
;
2054 dbg_printf("load globbing pattern %s\n", insn
->data
);
2055 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2056 estack_ax(stack
, top
)->u
.s
.str
= insn
->data
;
2057 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2058 estack_ax(stack
, top
)->u
.s
.literal_type
=
2059 ESTACK_STRING_LITERAL_TYPE_STAR_GLOB
;
2060 estack_ax_t
= REG_STAR_GLOB_STRING
;
2061 next_pc
+= sizeof(struct load_op
) + strlen(insn
->data
) + 1;
2065 OP(BYTECODE_OP_LOAD_S64
):
2067 struct load_op
*insn
= (struct load_op
*) pc
;
2069 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2070 estack_ax_v
= ((struct literal_numeric
*) insn
->data
)->v
;
2071 estack_ax_t
= REG_S64
;
2072 dbg_printf("load s64 %" PRIi64
"\n", estack_ax_v
);
2073 next_pc
+= sizeof(struct load_op
)
2074 + sizeof(struct literal_numeric
);
2078 OP(BYTECODE_OP_LOAD_DOUBLE
):
2080 struct load_op
*insn
= (struct load_op
*) pc
;
2082 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2083 memcpy(&estack_ax(stack
, top
)->u
.d
, insn
->data
,
2084 sizeof(struct literal_double
));
2085 estack_ax_t
= REG_DOUBLE
;
2086 dbg_printf("load double %g\n", estack_ax(stack
, top
)->u
.d
);
2087 next_pc
+= sizeof(struct load_op
)
2088 + sizeof(struct literal_double
);
2093 OP(BYTECODE_OP_CAST_TO_S64
):
2095 /* Dynamic typing. */
2096 switch (estack_ax_t
) {
2098 JUMP_TO(BYTECODE_OP_CAST_NOP
);
2100 JUMP_TO(BYTECODE_OP_CAST_DOUBLE_TO_S64
);
2102 estack_ax_t
= REG_S64
;
2103 next_pc
+= sizeof(struct cast_op
); /* Fall-through */
2104 case REG_STRING
: /* Fall-through */
2105 case REG_STAR_GLOB_STRING
:
2109 ERR("Unknown interpreter register type (%d)",
2116 OP(BYTECODE_OP_CAST_DOUBLE_TO_S64
):
2118 estack_ax_v
= (int64_t) estack_ax(stack
, top
)->u
.d
;
2119 estack_ax_t
= REG_S64
;
2120 next_pc
+= sizeof(struct cast_op
);
2124 OP(BYTECODE_OP_CAST_NOP
):
2126 next_pc
+= sizeof(struct cast_op
);
2130 /* get context ref */
2131 OP(BYTECODE_OP_GET_CONTEXT_REF
):
2133 struct load_op
*insn
= (struct load_op
*) pc
;
2134 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2135 const struct lttng_ust_ctx_field
*ctx_field
;
2136 struct lttng_ust_ctx_value v
;
2138 dbg_printf("get context ref offset %u type dynamic\n",
2140 ctx_field
= &ctx
->fields
[ref
->offset
];
2141 ctx_field
->get_value(ctx_field
->priv
, &v
);
2142 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2144 case LTTNG_UST_DYNAMIC_TYPE_NONE
:
2147 case LTTNG_UST_DYNAMIC_TYPE_S64
:
2148 estack_ax_v
= v
.u
.s64
;
2149 estack_ax_t
= REG_S64
;
2150 dbg_printf("ref get context dynamic s64 %" PRIi64
"\n", estack_ax_v
);
2152 case LTTNG_UST_DYNAMIC_TYPE_DOUBLE
:
2153 estack_ax(stack
, top
)->u
.d
= v
.u
.d
;
2154 estack_ax_t
= REG_DOUBLE
;
2155 dbg_printf("ref get context dynamic double %g\n", estack_ax(stack
, top
)->u
.d
);
2157 case LTTNG_UST_DYNAMIC_TYPE_STRING
:
2158 estack_ax(stack
, top
)->u
.s
.str
= v
.u
.str
;
2159 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2160 dbg_printf("Interpreter warning: loading a NULL string.\n");
2164 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2165 estack_ax(stack
, top
)->u
.s
.literal_type
=
2166 ESTACK_STRING_LITERAL_TYPE_NONE
;
2167 dbg_printf("ref get context dynamic string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
2168 estack_ax_t
= REG_STRING
;
2171 dbg_printf("Interpreter warning: unknown dynamic type (%d).\n", (int) v
.sel
);
2175 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2179 OP(BYTECODE_OP_GET_CONTEXT_REF_STRING
):
2181 struct load_op
*insn
= (struct load_op
*) pc
;
2182 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2183 const struct lttng_ust_ctx_field
*ctx_field
;
2184 struct lttng_ust_ctx_value v
;
2186 dbg_printf("get context ref offset %u type string\n",
2188 ctx_field
= &ctx
->fields
[ref
->offset
];
2189 ctx_field
->get_value(ctx_field
->priv
, &v
);
2190 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2191 estack_ax(stack
, top
)->u
.s
.str
= v
.u
.str
;
2192 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2193 dbg_printf("Interpreter warning: loading a NULL string.\n");
2197 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2198 estack_ax(stack
, top
)->u
.s
.literal_type
=
2199 ESTACK_STRING_LITERAL_TYPE_NONE
;
2200 estack_ax_t
= REG_STRING
;
2201 dbg_printf("ref get context string %s\n", estack_ax(stack
, top
)->u
.s
.str
);
2202 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2206 OP(BYTECODE_OP_GET_CONTEXT_REF_S64
):
2208 struct load_op
*insn
= (struct load_op
*) pc
;
2209 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2210 const struct lttng_ust_ctx_field
*ctx_field
;
2211 struct lttng_ust_ctx_value v
;
2213 dbg_printf("get context ref offset %u type s64\n",
2215 ctx_field
= &ctx
->fields
[ref
->offset
];
2216 ctx_field
->get_value(ctx_field
->priv
, &v
);
2217 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2218 estack_ax_v
= v
.u
.s64
;
2219 estack_ax_t
= REG_S64
;
2220 dbg_printf("ref get context s64 %" PRIi64
"\n", estack_ax_v
);
2221 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2225 OP(BYTECODE_OP_GET_CONTEXT_REF_DOUBLE
):
2227 struct load_op
*insn
= (struct load_op
*) pc
;
2228 struct field_ref
*ref
= (struct field_ref
*) insn
->data
;
2229 const struct lttng_ust_ctx_field
*ctx_field
;
2230 struct lttng_ust_ctx_value v
;
2232 dbg_printf("get context ref offset %u type double\n",
2234 ctx_field
= &ctx
->fields
[ref
->offset
];
2235 ctx_field
->get_value(ctx_field
->priv
, &v
);
2236 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2237 memcpy(&estack_ax(stack
, top
)->u
.d
, &v
.u
.d
, sizeof(struct literal_double
));
2238 estack_ax_t
= REG_DOUBLE
;
2239 dbg_printf("ref get context double %g\n", estack_ax(stack
, top
)->u
.d
);
2240 next_pc
+= sizeof(struct load_op
) + sizeof(struct field_ref
);
2244 OP(BYTECODE_OP_GET_CONTEXT_ROOT
):
2246 dbg_printf("op get context root\n");
2247 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2248 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_CONTEXT
;
2249 /* "field" only needed for variants. */
2250 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
2251 estack_ax_t
= REG_PTR
;
2252 next_pc
+= sizeof(struct load_op
);
2256 OP(BYTECODE_OP_GET_APP_CONTEXT_ROOT
):
2258 dbg_printf("op get app context root\n");
2259 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2260 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_APP_CONTEXT
;
2261 /* "field" only needed for variants. */
2262 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
2263 estack_ax_t
= REG_PTR
;
2264 next_pc
+= sizeof(struct load_op
);
2268 OP(BYTECODE_OP_GET_PAYLOAD_ROOT
):
2270 dbg_printf("op get app payload root\n");
2271 estack_push(stack
, top
, ax
, bx
, ax_t
, bx_t
);
2272 estack_ax(stack
, top
)->u
.ptr
.type
= LOAD_ROOT_PAYLOAD
;
2273 estack_ax(stack
, top
)->u
.ptr
.ptr
= interpreter_stack_data
;
2274 /* "field" only needed for variants. */
2275 estack_ax(stack
, top
)->u
.ptr
.field
= NULL
;
2276 estack_ax_t
= REG_PTR
;
2277 next_pc
+= sizeof(struct load_op
);
2281 OP(BYTECODE_OP_GET_SYMBOL
):
2283 dbg_printf("op get symbol\n");
2284 switch (estack_ax(stack
, top
)->u
.ptr
.type
) {
2286 ERR("Nested fields not implemented yet.");
2289 case LOAD_ROOT_CONTEXT
:
2290 case LOAD_ROOT_APP_CONTEXT
:
2291 case LOAD_ROOT_PAYLOAD
:
2293 * symbol lookup is performed by
2299 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_symbol
);
2303 OP(BYTECODE_OP_GET_SYMBOL_FIELD
):
2306 * Used for first variant encountered in a
2307 * traversal. Variants are not implemented yet.
2313 OP(BYTECODE_OP_GET_INDEX_U16
):
2315 struct load_op
*insn
= (struct load_op
*) pc
;
2316 struct get_index_u16
*index
= (struct get_index_u16
*) insn
->data
;
2318 dbg_printf("op get index u16\n");
2319 ret
= dynamic_get_index(ctx
, bytecode
, index
->index
, estack_ax(stack
, top
));
2322 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
2323 estack_ax_t
= estack_ax(stack
, top
)->type
;
2324 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_index_u16
);
2328 OP(BYTECODE_OP_GET_INDEX_U64
):
2330 struct load_op
*insn
= (struct load_op
*) pc
;
2331 struct get_index_u64
*index
= (struct get_index_u64
*) insn
->data
;
2333 dbg_printf("op get index u64\n");
2334 ret
= dynamic_get_index(ctx
, bytecode
, index
->index
, estack_ax(stack
, top
));
2337 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
2338 estack_ax_t
= estack_ax(stack
, top
)->type
;
2339 next_pc
+= sizeof(struct load_op
) + sizeof(struct get_index_u64
);
2343 OP(BYTECODE_OP_LOAD_FIELD
):
2345 dbg_printf("op load field\n");
2346 ret
= dynamic_load_field(estack_ax(stack
, top
));
2349 estack_ax_v
= estack_ax(stack
, top
)->u
.v
;
2350 estack_ax_t
= estack_ax(stack
, top
)->type
;
2351 next_pc
+= sizeof(struct load_op
);
2355 OP(BYTECODE_OP_LOAD_FIELD_S8
):
2357 dbg_printf("op load field s8\n");
2359 estack_ax_v
= *(int8_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2360 estack_ax_t
= REG_S64
;
2361 next_pc
+= sizeof(struct load_op
);
2364 OP(BYTECODE_OP_LOAD_FIELD_S16
):
2366 dbg_printf("op load field s16\n");
2368 estack_ax_v
= *(int16_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2369 estack_ax_t
= REG_S64
;
2370 next_pc
+= sizeof(struct load_op
);
2373 OP(BYTECODE_OP_LOAD_FIELD_S32
):
2375 dbg_printf("op load field s32\n");
2377 estack_ax_v
= *(int32_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2378 estack_ax_t
= REG_S64
;
2379 next_pc
+= sizeof(struct load_op
);
2382 OP(BYTECODE_OP_LOAD_FIELD_S64
):
2384 dbg_printf("op load field s64\n");
2386 estack_ax_v
= *(int64_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2387 estack_ax_t
= REG_S64
;
2388 next_pc
+= sizeof(struct load_op
);
2391 OP(BYTECODE_OP_LOAD_FIELD_U8
):
2393 dbg_printf("op load field u8\n");
2395 estack_ax_v
= *(uint8_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2396 estack_ax_t
= REG_U64
;
2397 next_pc
+= sizeof(struct load_op
);
2400 OP(BYTECODE_OP_LOAD_FIELD_U16
):
2402 dbg_printf("op load field u16\n");
2404 estack_ax_v
= *(uint16_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2405 estack_ax_t
= REG_U64
;
2406 next_pc
+= sizeof(struct load_op
);
2409 OP(BYTECODE_OP_LOAD_FIELD_U32
):
2411 dbg_printf("op load field u32\n");
2413 estack_ax_v
= *(uint32_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2414 estack_ax_t
= REG_U64
;
2415 next_pc
+= sizeof(struct load_op
);
2418 OP(BYTECODE_OP_LOAD_FIELD_U64
):
2420 dbg_printf("op load field u64\n");
2422 estack_ax_v
= *(uint64_t *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2423 estack_ax_t
= REG_U64
;
2424 next_pc
+= sizeof(struct load_op
);
2427 OP(BYTECODE_OP_LOAD_FIELD_DOUBLE
):
2429 dbg_printf("op load field double\n");
2431 memcpy(&estack_ax(stack
, top
)->u
.d
,
2432 estack_ax(stack
, top
)->u
.ptr
.ptr
,
2433 sizeof(struct literal_double
));
2434 estack_ax(stack
, top
)->type
= REG_DOUBLE
;
2435 next_pc
+= sizeof(struct load_op
);
2439 OP(BYTECODE_OP_LOAD_FIELD_STRING
):
2443 dbg_printf("op load field string\n");
2444 str
= (const char *) estack_ax(stack
, top
)->u
.ptr
.ptr
;
2445 estack_ax(stack
, top
)->u
.s
.str
= str
;
2446 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2447 dbg_printf("Interpreter warning: loading a NULL string.\n");
2451 estack_ax(stack
, top
)->u
.s
.seq_len
= SIZE_MAX
;
2452 estack_ax(stack
, top
)->u
.s
.literal_type
=
2453 ESTACK_STRING_LITERAL_TYPE_NONE
;
2454 estack_ax(stack
, top
)->type
= REG_STRING
;
2455 next_pc
+= sizeof(struct load_op
);
2459 OP(BYTECODE_OP_LOAD_FIELD_SEQUENCE
):
2463 dbg_printf("op load field string sequence\n");
2464 ptr
= estack_ax(stack
, top
)->u
.ptr
.ptr
;
2465 estack_ax(stack
, top
)->u
.s
.seq_len
= *(unsigned long *) ptr
;
2466 estack_ax(stack
, top
)->u
.s
.str
= *(const char **) (ptr
+ sizeof(unsigned long));
2467 estack_ax(stack
, top
)->type
= REG_STRING
;
2468 if (unlikely(!estack_ax(stack
, top
)->u
.s
.str
)) {
2469 dbg_printf("Interpreter warning: loading a NULL sequence.\n");
2473 estack_ax(stack
, top
)->u
.s
.literal_type
=
2474 ESTACK_STRING_LITERAL_TYPE_NONE
;
2475 next_pc
+= sizeof(struct load_op
);
2481 /* No need to prepare output if an error occurred. */
2483 return LTTNG_UST_BYTECODE_INTERPRETER_ERROR
;
2485 /* Prepare output. */
2486 switch (ust_bytecode
->type
) {
2487 case LTTNG_UST_BYTECODE_TYPE_FILTER
:
2489 struct lttng_ust_bytecode_filter_ctx
*filter_ctx
=
2490 (struct lttng_ust_bytecode_filter_ctx
*) caller_ctx
;
2492 filter_ctx
->result
= LTTNG_UST_BYTECODE_FILTER_ACCEPT
;
2494 filter_ctx
->result
= LTTNG_UST_BYTECODE_FILTER_REJECT
;
2497 case LTTNG_UST_BYTECODE_TYPE_CAPTURE
:
2498 ret
= lttng_bytecode_interpret_format_output(estack_ax(stack
, top
),
2499 (struct lttng_interpreter_output
*) caller_ctx
);
2506 return LTTNG_UST_BYTECODE_INTERPRETER_ERROR
;
2508 return LTTNG_UST_BYTECODE_INTERPRETER_OK
;
2512 * Return LTTNG_UST_EVENT_FILTER_ACCEPT or LTTNG_UST_EVENT_FILTER_REJECT.
2514 int lttng_ust_interpret_event_filter(const struct lttng_ust_event_common
*event
,
2515 const char *interpreter_stack_data
,
2516 void *event_filter_ctx
__attribute__((unused
)))
2518 struct lttng_ust_bytecode_runtime
*filter_bc_runtime
;
2519 struct cds_list_head
*filter_bytecode_runtime_head
= &event
->priv
->filter_bytecode_runtime_head
;
2520 struct lttng_ust_bytecode_filter_ctx bytecode_filter_ctx
;
2521 bool filter_record
= false;
2523 cds_list_for_each_entry_rcu(filter_bc_runtime
, filter_bytecode_runtime_head
, node
) {
2524 if (caa_likely(filter_bc_runtime
->interpreter_func(filter_bc_runtime
,
2525 interpreter_stack_data
, &bytecode_filter_ctx
) == LTTNG_UST_BYTECODE_INTERPRETER_OK
)) {
2526 if (caa_unlikely(bytecode_filter_ctx
.result
== LTTNG_UST_BYTECODE_FILTER_ACCEPT
)) {
2527 filter_record
= true;
2533 return LTTNG_UST_EVENT_FILTER_ACCEPT
;
2535 return LTTNG_UST_EVENT_FILTER_REJECT
;