projects
/
lttng-modules.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
fix: mm: move kvmalloc-related functions to slab.h (v5.16)
[lttng-modules.git]
/
src
/
lttng-bytecode.c
diff --git
a/src/lttng-bytecode.c
b/src/lttng-bytecode.c
index 6ccd72c7cc8c664318f66fcb78b420f107af99d0..2ad0b390cb4ac2dc904e8f2eef89021016c54846 100644
(file)
--- a/
src/lttng-bytecode.c
+++ b/
src/lttng-bytecode.c
@@
-175,7
+175,7
@@
int apply_field_reloc(const struct lttng_kernel_event_desc *event_desc,
const char *field_name,
enum bytecode_op bytecode_op)
{
const char *field_name,
enum bytecode_op bytecode_op)
{
- const struct lttng_kernel_event_field **fields, *field = NULL;
+ const struct lttng_kernel_event_field *
const
*fields, *field = NULL;
unsigned int nr_fields, i;
struct load_op *op;
uint32_t field_offset = 0;
unsigned int nr_fields, i;
struct load_op *op;
uint32_t field_offset = 0;
@@
-185,10
+185,10
@@
int apply_field_reloc(const struct lttng_kernel_event_desc *event_desc,
/* Lookup event by name */
if (!event_desc)
return -EINVAL;
/* Lookup event by name */
if (!event_desc)
return -EINVAL;
- fields = event_desc->fields;
+ fields = event_desc->
tp_class->
fields;
if (!fields)
return -EINVAL;
if (!fields)
return -EINVAL;
- nr_fields = event_desc->nr_fields;
+ nr_fields = event_desc->
tp_class->
nr_fields;
for (i = 0; i < nr_fields; i++) {
if (fields[i]->nofilter)
continue;
for (i = 0; i < nr_fields; i++) {
if (fields[i]->nofilter)
continue;
@@
-411,7
+411,7
@@
static
int bytecode_is_linked(struct lttng_kernel_bytecode_node *bytecode,
struct list_head *bytecode_runtime_head)
{
int bytecode_is_linked(struct lttng_kernel_bytecode_node *bytecode,
struct list_head *bytecode_runtime_head)
{
- struct lttng_bytecode_runtime *bc_runtime;
+ struct lttng_
kernel_
bytecode_runtime *bc_runtime;
list_for_each_entry(bc_runtime, bytecode_runtime_head, node) {
if (bc_runtime->bc == bytecode)
list_for_each_entry(bc_runtime, bytecode_runtime_head, node) {
if (bc_runtime->bc == bytecode)
@@
-500,7
+500,7
@@
alloc_error:
return ret;
}
return ret;
}
-void lttng_bytecode_sync_state(struct lttng_bytecode_runtime *runtime)
+void lttng_bytecode_sync_state(struct lttng_
kernel_
bytecode_runtime *runtime)
{
struct lttng_kernel_bytecode_node *bc = runtime->bc;
{
struct lttng_kernel_bytecode_node *bc = runtime->bc;
@@
-524,7
+524,7
@@
void lttng_enabler_link_bytecode(const struct lttng_kernel_event_desc *event_des
struct list_head *enabler_bytecode_head)
{
struct lttng_kernel_bytecode_node *enabler_bc;
struct list_head *enabler_bytecode_head)
{
struct lttng_kernel_bytecode_node *enabler_bc;
- struct lttng_bytecode_runtime *runtime;
+ struct lttng_
kernel_
bytecode_runtime *runtime;
WARN_ON_ONCE(!event_desc);
WARN_ON_ONCE(!event_desc);
@@
-578,14
+578,14
@@
void lttng_enabler_link_bytecode(const struct lttng_kernel_event_desc *event_des
/*
* We own the filter_bytecode if we return success.
*/
/*
* We own the filter_bytecode if we return success.
*/
-int lttng_filter_enabler_attach_bytecode(struct lttng_e
nabler
*enabler,
+int lttng_filter_enabler_attach_bytecode(struct lttng_e
vent_enabler_common
*enabler,
struct lttng_kernel_bytecode_node *filter_bytecode)
{
list_add(&filter_bytecode->node, &enabler->filter_bytecode_head);
return 0;
}
struct lttng_kernel_bytecode_node *filter_bytecode)
{
list_add(&filter_bytecode->node, &enabler->filter_bytecode_head);
return 0;
}
-void lttng_free_enabler_filter_bytecode(struct lttng_e
nabler
*enabler)
+void lttng_free_enabler_filter_bytecode(struct lttng_e
vent_enabler_common
*enabler)
{
struct lttng_kernel_bytecode_node *filter_bytecode, *tmp;
{
struct lttng_kernel_bytecode_node *filter_bytecode, *tmp;
This page took
0.025414 seconds
and
4
git commands to generate.