1 /* This file is part of the Linux Trace Toolkit viewer
2 * Copyright (C) 2005 Mathieu Desnoyers
4 * Complete rewrite from the original version made by XangXiu Yang.
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License Version 2.1 as published by the Free Software Foundation.
10 * This library is distributed in the hope that it will be useful,
11 * but WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
13 * Lesser General Public License for more details.
15 * You should have received a copy of the GNU Lesser General Public
16 * License along with this library; if not, write to the
17 * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
18 * Boston, MA 02111-1307, USA.
30 #include <sys/types.h>
35 #include <glib/gprintf.h>
48 #include "ltt-private.h"
49 #include <ltt/trace.h>
50 #include <ltt/event.h>
51 #include <ltt/ltt-types.h>
52 #include <ltt/marker.h>
55 extern long marker_update_fields_offsets(struct marker_info
*info
, const char *data
);
57 /* Tracefile names used in this file */
59 GQuark LTT_TRACEFILE_NAME_METADATA
;
66 #define __UNUSED__ __attribute__((__unused__))
68 #define g_info(format...) g_log (G_LOG_DOMAIN, G_LOG_LEVEL_INFO, format)
71 #define g_debug(format...) g_log (G_LOG_DOMAIN, G_LOG_LEVEL_DEBUG, format)
76 /* Those macros must be called from within a function where page_size is a known
78 #define PAGE_MASK (~(page_size-1))
79 #define PAGE_ALIGN(addr) (((addr)+page_size-1)&PAGE_MASK)
81 LttTrace
*father_trace
= NULL
;
83 /* set the offset of the fields belonging to the event,
84 need the information of the archecture */
85 //void set_fields_offsets(LttTracefile *tf, LttEventType *event_type);
86 //size_t get_fields_offsets(LttTracefile *tf, LttEventType *event_type, void *data);
89 /* get the size of the field type according to
90 * The facility size information. */
91 static inline void preset_field_type_size(LttTracefile
*tf
,
92 LttEventType
*event_type
,
93 off_t offset_root
, off_t offset_parent
,
94 enum field_status
*fixed_root
, enum field_status
*fixed_parent
,
98 /* map a fixed size or a block information from the file (fd) */
99 static gint
map_block(LttTracefile
* tf
, guint block_num
);
101 /* calculate nsec per cycles for current block */
103 static guint32
calc_nsecs_per_cycle(LttTracefile
* t
);
104 static guint64
cycles_2_ns(LttTracefile
*tf
, guint64 cycles
);
107 /* go to the next event */
108 static int ltt_seek_next_event(LttTracefile
*tf
);
110 static int open_tracefiles(LttTrace
*trace
, gchar
*root_path
,
111 gchar
*relative_path
);
112 static int ltt_process_metadata_tracefile(LttTracefile
*tf
);
113 static void ltt_tracefile_time_span_get(LttTracefile
*tf
,
114 LttTime
*start
, LttTime
*end
);
115 static void group_time_span_get(GQuark name
, gpointer data
, gpointer user_data
);
116 static gint
map_block(LttTracefile
* tf
, guint block_num
);
117 static void ltt_update_event_size(LttTracefile
*tf
);
119 /* Enable event debugging */
120 static int a_event_debug
= 0;
122 void ltt_event_debug(int state
)
124 a_event_debug
= state
;
129 * Return value : 0 success, 1 bad tracefile
131 static int parse_trace_header(ltt_subbuffer_header_t
*header
,
132 LttTracefile
*tf
, LttTrace
*t
)
134 if (header
->magic_number
== LTT_MAGIC_NUMBER
)
136 else if(header
->magic_number
== LTT_REV_MAGIC_NUMBER
)
138 else /* invalid magic number, bad tracefile ! */
142 t
->ltt_major_version
= header
->major_version
;
143 t
->ltt_minor_version
= header
->minor_version
;
144 t
->arch_size
= header
->arch_size
;
146 tf
->alignment
= header
->alignment
;
148 /* Get float byte order : might be different from int byte order
149 * (or is set to 0 if the trace has no float (kernel trace)) */
150 tf
->float_word_order
= 0;
152 switch(header
->major_version
) {
155 g_warning("Unsupported trace version : %hhu.%hhu",
156 header
->major_version
, header
->minor_version
);
160 switch(header
->minor_version
) {
163 struct ltt_subbuffer_header_2_3
*vheader
= header
;
164 tf
->buffer_header_size
= ltt_subbuffer_header_size();
167 tf
->tsc_mask
= ((1ULL << tf
->tscbits
) - 1);
168 tf
->tsc_mask_next_bit
= (1ULL << tf
->tscbits
);
171 t
->start_freq
= ltt_get_uint64(LTT_GET_BO(tf
),
172 &vheader
->start_freq
);
173 t
->freq_scale
= ltt_get_uint32(LTT_GET_BO(tf
),
174 &vheader
->freq_scale
);
176 t
->start_freq
= father_trace
->start_freq
;
177 t
->freq_scale
= father_trace
->freq_scale
;
181 t
->start_tsc
= ltt_get_uint64(LTT_GET_BO(tf
),
182 &vheader
->cycle_count_begin
);
183 t
->start_monotonic
= 0;
184 t
->start_time
.tv_sec
= ltt_get_uint64(LTT_GET_BO(tf
),
185 &vheader
->start_time_sec
);
186 t
->start_time
.tv_nsec
= ltt_get_uint64(LTT_GET_BO(tf
),
187 &vheader
->start_time_usec
);
188 t
->start_time
.tv_nsec
*= 1000; /* microsec to nanosec */
190 t
->start_time_from_tsc
= ltt_time_from_uint64(
192 * 1000000000.0 * tf
->trace
->freq_scale
193 / (double)t
->start_freq
);
198 g_warning("Unsupported trace version : %hhu.%hhu",
199 header
->major_version
, header
->minor_version
);
204 g_warning("Unsupported trace version : %hhu.%hhu",
205 header
->major_version
, header
->minor_version
);
213 /*****************************************************************************
215 * ltt_tracefile_open : open a trace file, construct a LttTracefile
217 * t : the trace containing the tracefile
218 * fileName : path name of the trace file
219 * tf : the tracefile structure
221 * : 0 for success, -1 otherwise.
222 ****************************************************************************/
224 static gint
ltt_tracefile_open(LttTrace
*t
, gchar
* fileName
, LttTracefile
*tf
)
226 struct stat lTDFStat
; /* Trace data file status */
227 ltt_subbuffer_header_t
*header
;
228 int page_size
= getpagesize();
231 tf
->long_name
= g_quark_from_string(fileName
);
233 tf
->fd
= open(fileName
, O_RDONLY
);
235 g_warning("Unable to open input data file %s\n", fileName
);
239 // Get the file's status
240 if(fstat(tf
->fd
, &lTDFStat
) < 0){
241 g_warning("Unable to get the status of the input data file %s\n", fileName
);
245 // Is the file large enough to contain a trace
246 if(lTDFStat
.st_size
<
247 (off_t
)(ltt_subbuffer_header_size())){
248 g_print("The input data file %s does not contain a trace\n", fileName
);
252 /* Temporarily map the buffer start header to get trace information */
253 /* Multiple of pages aligned head */
254 tf
->buffer
.head
= mmap(0,
255 PAGE_ALIGN(ltt_subbuffer_header_size()), PROT_READ
,
256 MAP_PRIVATE
, tf
->fd
, 0);
257 if(tf
->buffer
.head
== MAP_FAILED
) {
258 perror("Error in allocating memory for buffer of tracefile");
261 g_assert( ( (gulong
)tf
->buffer
.head
&(8-1) ) == 0); // make sure it's aligned.
263 header
= (ltt_subbuffer_header_t
*)tf
->buffer
.head
;
265 if(parse_trace_header(header
, tf
, NULL
)) {
266 g_warning("parse_trace_header error");
270 //store the size of the file
271 tf
->file_size
= lTDFStat
.st_size
;
272 tf
->buf_size
= ltt_get_uint32(LTT_GET_BO(tf
), &header
->buf_size
);
273 tf
->num_blocks
= tf
->file_size
/ tf
->buf_size
;
275 tf
->subbuf_corrupt
= 0;
277 if(munmap(tf
->buffer
.head
,
278 PAGE_ALIGN(ltt_subbuffer_header_size()))) {
279 g_warning("unmap size : %zu\n",
280 PAGE_ALIGN(ltt_subbuffer_header_size()));
281 perror("munmap error");
284 tf
->buffer
.head
= NULL
;
286 //read the first block
287 if(map_block(tf
,0)) {
288 perror("Cannot map block for tracefile");
296 if(munmap(tf
->buffer
.head
,
297 PAGE_ALIGN(ltt_subbuffer_header_size()))) {
298 g_warning("unmap size : %zu\n",
299 PAGE_ALIGN(ltt_subbuffer_header_size()));
300 perror("munmap error");
310 /*****************************************************************************
312 * ltt_tracefile_close: close a trace file,
314 * t : tracefile which will be closed
315 ****************************************************************************/
317 static void ltt_tracefile_close(LttTracefile
*t
)
319 int page_size
= getpagesize();
321 if(t
->buffer
.head
!= NULL
)
322 if(munmap(t
->buffer
.head
, PAGE_ALIGN(t
->buf_size
))) {
323 g_warning("unmap size : %u\n",
324 PAGE_ALIGN(t
->buf_size
));
325 perror("munmap error");
332 /****************************************************************************
333 * get_absolute_pathname
335 * return the unique pathname in the system
337 * MD : Fixed this function so it uses realpath, dealing well with
338 * forgotten cases (.. were not used correctly before).
340 ****************************************************************************/
341 void get_absolute_pathname(const gchar
*pathname
, gchar
* abs_pathname
)
343 abs_pathname
[0] = '\0';
345 if (realpath(pathname
, abs_pathname
) != NULL
)
349 /* error, return the original path unmodified */
350 strcpy(abs_pathname
, pathname
);
356 /* Search for something like : .*_.*
358 * The left side is the name, the right side is the number.
360 * Exclude flight- prefix.
363 static int get_tracefile_name_number(gchar
*raw_name
,
370 guint raw_name_len
= strlen(raw_name
);
371 gchar char_name
[PATH_MAX
];
379 for(i
= 0; i
< raw_name_len
-1;i
++) {
380 if(raw_name
[i
] != '/')
383 raw_name
= &raw_name
[i
];
384 raw_name_len
= strlen(raw_name
);
386 for(i
=raw_name_len
-1;i
>=0;i
--) {
387 if(raw_name
[i
] == '_') break;
389 if(i
==-1) { /* Either not found or name length is 0 */
390 /* This is a userspace tracefile */
391 strncpy(char_name
, raw_name
, raw_name_len
);
392 char_name
[raw_name_len
] = '\0';
393 *name
= g_quark_from_string(char_name
);
394 *num
= 0; /* unknown cpu */
395 for(i
=0;i
<raw_name_len
;i
++) {
396 if(raw_name
[i
] == '/') {
401 for(;i
<raw_name_len
;i
++) {
402 if(raw_name
[i
] == '/') {
407 for(;i
<raw_name_len
;i
++) {
408 if(raw_name
[i
] == '-') {
412 if(i
== raw_name_len
) return -1;
414 tmpptr
= &raw_name
[i
];
415 for(;i
<raw_name_len
;i
++) {
416 if(raw_name
[i
] == '.') {
421 *tid
= strtoul(tmpptr
, &endptr
, 10);
423 return -1; /* No digit */
424 if(*tid
== ULONG_MAX
)
425 return -1; /* underflow / overflow */
427 tmpptr
= &raw_name
[i
];
428 for(;i
<raw_name_len
;i
++) {
429 if(raw_name
[i
] == '.') {
434 *pgid
= strtoul(tmpptr
, &endptr
, 10);
436 return -1; /* No digit */
437 if(*pgid
== ULONG_MAX
)
438 return -1; /* underflow / overflow */
440 tmpptr
= &raw_name
[i
];
441 *creation
= strtoull(tmpptr
, &endptr
, 10);
443 return -1; /* No digit */
444 if(*creation
== G_MAXUINT64
)
445 return -1; /* underflow / overflow */
449 cpu_num
= strtol(raw_name
+underscore_pos
+1, &endptr
, 10);
451 if(endptr
== raw_name
+underscore_pos
+1)
452 return -1; /* No digit */
453 if(cpu_num
== LONG_MIN
|| cpu_num
== LONG_MAX
)
454 return -1; /* underflow / overflow */
456 if (!strncmp(raw_name
, "flight-", sizeof("flight-") - 1)) {
457 raw_name
+= sizeof("flight-") - 1;
458 underscore_pos
-= sizeof("flight-") - 1;
460 strncpy(char_name
, raw_name
, underscore_pos
);
461 char_name
[underscore_pos
] = '\0';
462 *name
= g_quark_from_string(char_name
);
471 GData
**ltt_trace_get_tracefiles_groups(LttTrace
*trace
)
473 return &trace
->tracefiles
;
477 void compute_tracefile_group(GQuark key_id
,
479 struct compute_tracefile_group_args
*args
)
484 for(i
=0; i
<group
->len
; i
++) {
485 tf
= &g_array_index (group
, LttTracefile
, i
);
487 args
->func(tf
, args
->func_args
);
492 static void ltt_tracefile_group_destroy(gpointer data
)
494 GArray
*group
= (GArray
*)data
;
499 destroy_marker_data(g_array_index (group
, LttTracefile
, 0).mdata
);
500 for(i
=0; i
<group
->len
; i
++) {
501 tf
= &g_array_index (group
, LttTracefile
, i
);
503 ltt_tracefile_close(tf
);
505 g_array_free(group
, TRUE
);
508 static __attribute__ ((__unused__
)) gboolean
ltt_tracefile_group_has_cpu_online(gpointer data
)
510 GArray
*group
= (GArray
*)data
;
514 for(i
=0; i
<group
->len
; i
++) {
515 tf
= &g_array_index (group
, LttTracefile
, i
);
523 /* Open each tracefile under a specific directory. Put them in a
524 * GData : permits to access them using their tracefile group pathname.
525 * i.e. access control/modules tracefile group by index :
528 * relative path is the path relative to the trace root
529 * root path is the full path
531 * A tracefile group is simply an array where all the per cpu tracefiles sit.
534 static int open_tracefiles(LttTrace
*trace
, gchar
*root_path
, gchar
*relative_path
)
536 DIR *dir
= opendir(root_path
);
537 struct dirent
*entry
;
538 struct stat stat_buf
;
540 struct marker_data
*mdata
;
542 gchar path
[PATH_MAX
];
547 gchar rel_path
[PATH_MAX
];
556 strncpy(path
, root_path
, PATH_MAX
-1);
557 path_len
= strlen(path
);
558 path
[path_len
] = '/';
560 path_ptr
= path
+ path_len
;
562 strncpy(rel_path
, relative_path
, PATH_MAX
-1);
563 rel_path_len
= strlen(rel_path
);
564 rel_path
[rel_path_len
] = '/';
566 rel_path_ptr
= rel_path
+ rel_path_len
;
568 while((entry
= readdir(dir
)) != NULL
) {
570 if(entry
->d_name
[0] == '.') continue;
572 strncpy(path_ptr
, entry
->d_name
, PATH_MAX
- path_len
);
573 strncpy(rel_path_ptr
, entry
->d_name
, PATH_MAX
- rel_path_len
);
575 ret
= stat(path
, &stat_buf
);
581 g_debug("Tracefile file or directory : %s\n", path
);
583 // if(strcmp(rel_path, "/eventdefs") == 0) continue;
585 if(S_ISDIR(stat_buf
.st_mode
)) {
587 g_debug("Entering subdirectory...\n");
588 ret
= open_tracefiles(trace
, path
, rel_path
);
589 if(ret
< 0) continue;
590 } else if(S_ISREG(stat_buf
.st_mode
)) {
599 if(get_tracefile_name_number(rel_path
, &name
, &num
, &tid
, &pgid
, &creation
))
600 continue; /* invalid name */
602 g_debug("Opening file.\n");
603 if(ltt_tracefile_open(trace
, path
, &tmp_tf
)) {
604 g_info("Error opening tracefile %s", path
);
606 continue; /* error opening the tracefile : bad magic number ? */
609 g_debug("Tracefile name is %s and number is %u",
610 g_quark_to_string(name
), num
);
613 tmp_tf
.cpu_online
= 1;
614 tmp_tf
.cpu_num
= num
;
618 tmp_tf
.creation
= creation
;
619 group
= g_datalist_id_get_data(&trace
->tracefiles
, name
);
621 /* Elements are automatically cleared when the array is allocated.
622 * It makes the cpu_online variable set to 0 : cpu offline, by default.
624 group
= g_array_sized_new (FALSE
, TRUE
, sizeof(LttTracefile
), 10);
625 g_datalist_id_set_data_full(&trace
->tracefiles
, name
,
626 group
, ltt_tracefile_group_destroy
);
627 mdata
= allocate_marker_data();
629 g_error("Error in allocating marker data");
632 /* Add the per cpu tracefile to the named group */
633 unsigned int old_len
= group
->len
;
635 group
= g_array_set_size(group
, num
+1);
637 g_assert(group
->len
> 0);
639 mdata
= g_array_index (group
, LttTracefile
, 0).mdata
;
641 g_array_index (group
, LttTracefile
, num
) = tmp_tf
;
642 g_array_index (group
, LttTracefile
, num
).event
.tracefile
=
643 &g_array_index (group
, LttTracefile
, num
);
644 for (i
= 0; i
< group
->len
; i
++)
645 g_array_index (group
, LttTracefile
, i
).mdata
= mdata
;
655 /* Presumes the tracefile is already seeked at the beginning. It makes sense,
656 * because it must be done just after the opening */
657 static int ltt_process_metadata_tracefile(LttTracefile
*tf
)
662 err
= ltt_tracefile_read_seek(tf
);
663 if(err
== EPERM
) goto seek_error
;
664 else if(err
== ERANGE
) break; /* End of tracefile */
666 err
= ltt_tracefile_read_update_event(tf
);
667 if(err
) goto update_error
;
670 * It contains only core events :
672 * 1 : set_marker_format
674 if(tf
->event
.event_id
>= MARKER_CORE_IDS
) {
675 /* Should only contain core events */
676 g_warning("Error in processing metadata file %s, "
677 "should not contain event id %u.", g_quark_to_string(tf
->name
),
683 const char *channel_name
, *marker_name
, *format
;
685 guint8 int_size
, long_size
, pointer_size
, size_t_size
, alignment
;
687 switch((enum marker_id
)tf
->event
.event_id
) {
688 case MARKER_ID_SET_MARKER_ID
:
689 channel_name
= pos
= tf
->event
.data
;
690 pos
+= strlen(channel_name
) + 1;
692 g_debug("Doing MARKER_ID_SET_MARKER_ID of marker %s.%s",
693 channel_name
, marker_name
);
694 pos
+= strlen(marker_name
) + 1;
695 pos
+= ltt_align((size_t)pos
, sizeof(guint16
), tf
->alignment
);
696 id
= ltt_get_uint16(LTT_GET_BO(tf
), pos
);
697 g_debug("In MARKER_ID_SET_MARKER_ID of marker %s.%s id %hu",
698 channel_name
, marker_name
, id
);
699 pos
+= sizeof(guint16
);
700 int_size
= *(guint8
*)pos
;
701 pos
+= sizeof(guint8
);
702 long_size
= *(guint8
*)pos
;
703 pos
+= sizeof(guint8
);
704 pointer_size
= *(guint8
*)pos
;
705 pos
+= sizeof(guint8
);
706 size_t_size
= *(guint8
*)pos
;
707 pos
+= sizeof(guint8
);
708 alignment
= *(guint8
*)pos
;
709 pos
+= sizeof(guint8
);
710 marker_id_event(tf
->trace
,
711 g_quark_from_string(channel_name
),
712 g_quark_from_string(marker_name
),
713 id
, int_size
, long_size
,
714 pointer_size
, size_t_size
, alignment
);
716 case MARKER_ID_SET_MARKER_FORMAT
:
717 channel_name
= pos
= tf
->event
.data
;
718 pos
+= strlen(channel_name
) + 1;
720 g_debug("Doing MARKER_ID_SET_MARKER_FORMAT of marker %s.%s",
721 channel_name
, marker_name
);
722 pos
+= strlen(marker_name
) + 1;
724 pos
+= strlen(format
) + 1;
725 marker_format_event(tf
->trace
,
726 g_quark_from_string(channel_name
),
727 g_quark_from_string(marker_name
),
729 /* get information from dictionary TODO */
732 g_warning("Error in processing metadata file %s, "
733 "unknown event id %hhu.",
734 g_quark_to_string(tf
->name
),
747 g_warning("An error occured in metadata tracefile parsing");
752 * Open a trace and return its LttTrace handle.
754 * pathname must be the directory of the trace
757 LttTrace
*ltt_trace_open(const gchar
*pathname
)
759 gchar abs_path
[PATH_MAX
];
765 ltt_subbuffer_header_t
*header
;
767 struct dirent
*entry
;
768 struct stat stat_buf
;
769 gchar path
[PATH_MAX
];
771 t
= g_new(LttTrace
, 1);
772 if(!t
) goto alloc_error
;
774 get_absolute_pathname(pathname
, abs_path
);
775 t
->pathname
= g_quark_from_string(abs_path
);
777 g_datalist_init(&t
->tracefiles
);
779 /* Test to see if it looks like a trace */
780 dir
= opendir(abs_path
);
785 while((entry
= readdir(dir
)) != NULL
) {
786 strcpy(path
, abs_path
);
788 strcat(path
, entry
->d_name
);
789 ret
= stat(path
, &stat_buf
);
797 /* Open all the tracefiles */
798 if(open_tracefiles(t
, abs_path
, "")) {
799 g_warning("Error opening tracefile %s", abs_path
);
803 /* Parse each trace metadata_N files : get runtime fac. info */
804 group
= g_datalist_id_get_data(&t
->tracefiles
, LTT_TRACEFILE_NAME_METADATA
);
806 g_error("Trace %s has no metadata tracefile", abs_path
);
812 * Get the trace information for the metadata_0 tracefile.
813 * Getting a correct trace start_time and start_tsc is insured by the fact
814 * that no subbuffers are supposed to be lost in the metadata channel.
815 * Therefore, the first subbuffer contains the start_tsc timestamp in its
818 g_assert(group
->len
> 0);
819 tf
= &g_array_index (group
, LttTracefile
, 0);
820 header
= (ltt_subbuffer_header_t
*)tf
->buffer
.head
;
821 ret
= parse_trace_header(header
, tf
, t
);
824 t
->num_cpu
= group
->len
;
826 //ret = allocate_marker_data(t);
828 // g_error("Error in allocating marker data");
830 for(i
=0; i
<group
->len
; i
++) {
831 tf
= &g_array_index (group
, LttTracefile
, i
);
833 if(ltt_process_metadata_tracefile(tf
))
835 // goto metadata_error;
842 // destroy_marker_data(t);
844 g_datalist_clear(&t
->tracefiles
);
852 /* Open another, completely independant, instance of a trace.
854 * A read on this new instance will read the first event of the trace.
856 * When we copy a trace, we want all the opening actions to happen again :
857 * the trace will be reopened and totally independant from the original.
858 * That's why we call ltt_trace_open.
860 LttTrace
*ltt_trace_copy(LttTrace
*self
)
862 return ltt_trace_open(g_quark_to_string(self
->pathname
));
869 void ltt_trace_close(LttTrace
*t
)
871 g_datalist_clear(&t
->tracefiles
);
876 /*****************************************************************************
877 * Get the start time and end time of the trace
878 ****************************************************************************/
880 void ltt_tracefile_time_span_get(LttTracefile
*tf
,
881 LttTime
*start
, LttTime
*end
)
885 err
= map_block(tf
, 0);
887 g_error("Can not map block");
888 *start
= ltt_time_infinite
;
890 *start
= tf
->buffer
.begin
.timestamp
;
892 err
= map_block(tf
, tf
->num_blocks
- 1); /* Last block */
894 g_error("Can not map block");
895 *end
= ltt_time_zero
;
897 *end
= tf
->buffer
.end
.timestamp
;
900 struct tracefile_time_span_get_args
{
906 static void group_time_span_get(GQuark name
, gpointer data
, gpointer user_data
)
908 struct tracefile_time_span_get_args
*args
=
909 (struct tracefile_time_span_get_args
*)user_data
;
911 GArray
*group
= (GArray
*)data
;
917 for(i
=0; i
<group
->len
; i
++) {
918 tf
= &g_array_index (group
, LttTracefile
, i
);
920 ltt_tracefile_time_span_get(tf
, &tmp_start
, &tmp_end
);
921 if(ltt_time_compare(*args
->start
, tmp_start
)>0) *args
->start
= tmp_start
;
922 if(ltt_time_compare(*args
->end
, tmp_end
)<0) *args
->end
= tmp_end
;
927 /* return the start and end time of a trace */
929 void ltt_trace_time_span_get(LttTrace
*t
, LttTime
*start
, LttTime
*end
)
931 LttTime min_start
= ltt_time_infinite
;
932 LttTime max_end
= ltt_time_zero
;
933 struct tracefile_time_span_get_args args
= { t
, &min_start
, &max_end
};
935 g_datalist_foreach(&t
->tracefiles
, &group_time_span_get
, &args
);
937 if(start
!= NULL
) *start
= min_start
;
938 if(end
!= NULL
) *end
= max_end
;
943 /* Seek to the first event in a tracefile that has a time equal or greater than
944 * the time passed in parameter.
946 * If the time parameter is outside the tracefile time span, seek to the first
947 * event or if after, return ERANGE.
949 * If the time parameter is before the first event, we have to seek specially to
952 * If the time is after the end of the trace, return ERANGE.
954 * Do a binary search to find the right block, then a sequential search in the
955 * block to find the event.
957 * In the special case where the time requested fits inside a block that has no
958 * event corresponding to the requested time, the first event of the next block
961 * IMPORTANT NOTE : // FIXME everywhere...
963 * You MUST NOT do a ltt_tracefile_read right after a ltt_tracefile_seek_time :
964 * you will jump over an event if you do.
966 * Return value : 0 : no error, the tf->event can be used
967 * ERANGE : time if after the last event of the trace
968 * otherwise : this is an error.
972 int ltt_tracefile_seek_time(LttTracefile
*tf
, LttTime time
)
976 unsigned int block_num
, high
, low
;
978 /* seek at the beginning of trace */
979 err
= map_block(tf
, 0); /* First block */
981 g_error("Can not map block");
985 /* If the time is lower or equal the beginning of the trace,
986 * go to the first event. */
987 if(ltt_time_compare(time
, tf
->buffer
.begin
.timestamp
) <= 0) {
988 ret
= ltt_tracefile_read(tf
);
989 if(ret
== ERANGE
) goto range
;
990 else if (ret
) goto fail
;
991 goto found
; /* There is either no event in the trace or the event points
992 to the first event in the trace */
995 err
= map_block(tf
, tf
->num_blocks
- 1); /* Last block */
997 g_error("Can not map block");
1001 /* If the time is after the end of the trace, return ERANGE. */
1002 if(ltt_time_compare(time
, tf
->buffer
.end
.timestamp
) > 0) {
1006 /* Binary search the block */
1007 high
= tf
->num_blocks
- 1;
1011 block_num
= ((high
-low
) / 2) + low
;
1013 err
= map_block(tf
, block_num
);
1015 g_error("Can not map block");
1019 /* We cannot divide anymore : this is what would happen if the time
1020 * requested was exactly between two consecutive buffers'end and start
1021 * timestamps. This is also what would happend if we didn't deal with out
1022 * of span cases prior in this function. */
1023 /* The event is right in the buffer!
1024 * (or in the next buffer first event) */
1026 ret
= ltt_tracefile_read(tf
);
1027 if(ret
== ERANGE
) goto range
; /* ERANGE or EPERM */
1028 else if(ret
) goto fail
;
1030 if(ltt_time_compare(time
, tf
->event
.event_time
) <= 0)
1034 } else if(ltt_time_compare(time
, tf
->buffer
.begin
.timestamp
) < 0) {
1035 /* go to lower part */
1036 high
= block_num
- 1;
1037 } else if(ltt_time_compare(time
, tf
->buffer
.end
.timestamp
) > 0) {
1038 /* go to higher part */
1039 low
= block_num
+ 1;
1040 } else {/* The event is right in the buffer!
1041 (or in the next buffer first event) */
1043 ret
= ltt_tracefile_read(tf
);
1044 if(ret
== ERANGE
) goto range
; /* ERANGE or EPERM */
1045 else if(ret
) goto fail
;
1047 if(ltt_time_compare(time
, tf
->event
.event_time
) <= 0)
1059 /* Error handling */
1061 g_error("ltt_tracefile_seek_time failed on tracefile %s",
1062 g_quark_to_string(tf
->name
));
1066 /* Seek to a position indicated by an LttEventPosition
1069 int ltt_tracefile_seek_position(LttTracefile
*tf
, const LttEventPosition
*ep
)
1073 if(ep
->tracefile
!= tf
) {
1077 err
= map_block(tf
, ep
->block
);
1079 g_error("Can not map block");
1083 tf
->event
.offset
= ep
->offset
;
1085 /* Put back the event real tsc */
1086 tf
->event
.tsc
= ep
->tsc
;
1087 tf
->buffer
.tsc
= ep
->tsc
;
1089 err
= ltt_tracefile_read_update_event(tf
);
1092 /* deactivate this, as it does nothing for now
1093 err = ltt_tracefile_read_op(tf);
1100 g_error("ltt_tracefile_seek_time failed on tracefile %s",
1101 g_quark_to_string(tf
->name
));
1105 /* Given a TSC value, return the LttTime (seconds,nanoseconds) it
1109 LttTime
ltt_interpolate_time_from_tsc(LttTracefile
*tf
, guint64 tsc
)
1113 if(tsc
> tf
->trace
->start_tsc
) {
1114 time
= ltt_time_from_uint64(
1115 (double)(tsc
- tf
->trace
->start_tsc
)
1116 * 1000000000.0 * tf
->trace
->freq_scale
1117 / (double)tf
->trace
->start_freq
);
1118 time
= ltt_time_add(tf
->trace
->start_time_from_tsc
, time
);
1120 time
= ltt_time_from_uint64(
1121 (double)(tf
->trace
->start_tsc
- tsc
)
1122 * 1000000000.0 * tf
->trace
->freq_scale
1123 / (double)tf
->trace
->start_freq
);
1124 time
= ltt_time_sub(tf
->trace
->start_time_from_tsc
, time
);
1129 /* Calculate the real event time based on the buffer boundaries */
1130 LttTime
ltt_interpolate_time(LttTracefile
*tf
, LttEvent
*event
)
1132 return ltt_interpolate_time_from_tsc(tf
, tf
->buffer
.tsc
);
1136 /* Get the current event of the tracefile : valid until the next read */
1137 LttEvent
*ltt_tracefile_get_event(LttTracefile
*tf
)
1144 /*****************************************************************************
1146 * ltt_tracefile_read : Read the next event in the tracefile
1151 * Returns 0 if an event can be used in tf->event.
1152 * Returns ERANGE on end of trace. The event in tf->event still can be used
1153 * (if the last block was not empty).
1154 * Returns EPERM on error.
1156 * This function does make the tracefile event structure point to the event
1157 * currently pointed to by the tf->event.
1159 * Note : you must call a ltt_tracefile_seek to the beginning of the trace to
1160 * reinitialize it after an error if you want results to be coherent.
1161 * It would be the case if a end of trace last buffer has no event : the end
1162 * of trace wouldn't be returned, but an error.
1163 * We make the assumption there is at least one event per buffer.
1164 ****************************************************************************/
1166 int ltt_tracefile_read(LttTracefile
*tf
)
1170 err
= ltt_tracefile_read_seek(tf
);
1172 err
= ltt_tracefile_read_update_event(tf
);
1175 /* deactivate this, as it does nothing for now
1176 err = ltt_tracefile_read_op(tf);
1183 int ltt_tracefile_read_seek(LttTracefile
*tf
)
1187 /* Get next buffer until we finally have an event, or end of trace */
1189 err
= ltt_seek_next_event(tf
);
1190 if(unlikely(err
== ENOPROTOOPT
)) {
1194 /* Are we at the end of the buffer ? */
1196 if(unlikely(tf
->buffer
.index
== tf
->num_blocks
-1)){ /* end of trace ? */
1199 /* get next block */
1200 err
= map_block(tf
, tf
->buffer
.index
+ 1);
1202 g_error("Can not map block");
1206 } else break; /* We found an event ! */
1212 /* do an operation when reading a new event */
1214 /* This function does nothing for now */
1216 int ltt_tracefile_read_op(LttTracefile
*tf
)
1222 /* do event specific operation */
1230 static void print_debug_event_header(LttEvent
*ev
, void *start_pos
, void *end_pos
)
1232 unsigned int offset
= 0;
1235 g_printf("Event header (tracefile %s offset %" PRIx64
"):\n",
1236 g_quark_to_string(ev
->tracefile
->long_name
),
1237 ((uint64_t)ev
->tracefile
->buffer
.index
* ev
->tracefile
->buf_size
)
1238 + (long)start_pos
- (long)ev
->tracefile
->buffer
.head
);
1240 while (offset
< (long)end_pos
- (long)start_pos
) {
1241 g_printf("%8lx", (long)start_pos
- (long)ev
->tracefile
->buffer
.head
+ offset
);
1244 for (i
= 0; i
< 4 ; i
++) {
1245 for (j
= 0; j
< 4; j
++) {
1246 if (offset
+ ((i
* 4) + j
) <
1247 (long)end_pos
- (long)start_pos
)
1249 ((char*)start_pos
)[offset
+ ((i
* 4) + j
)]);
1263 /* same as ltt_tracefile_read, but does not seek to the next event nor call
1264 * event specific operation. */
1265 int ltt_tracefile_read_update_event(LttTracefile
*tf
)
1272 pos
= tf
->buffer
.head
+ event
->offset
;
1274 /* Read event header */
1276 /* Align the head */
1277 pos
+= ltt_align((size_t)pos
, sizeof(guint32
), tf
->alignment
);
1280 event
->timestamp
= ltt_get_uint32(LTT_GET_BO(tf
), pos
);
1281 event
->event_id
= event
->timestamp
>> tf
->tscbits
;
1282 event
->timestamp
= event
->timestamp
& tf
->tsc_mask
;
1283 pos
+= sizeof(guint32
);
1285 switch (event
->event_id
) {
1286 case 29: /* LTT_RFLAG_ID_SIZE_TSC */
1287 event
->event_id
= ltt_get_uint16(LTT_GET_BO(tf
), pos
);
1288 pos
+= sizeof(guint16
);
1289 event
->event_size
= ltt_get_uint16(LTT_GET_BO(tf
), pos
);
1290 pos
+= sizeof(guint16
);
1291 if (event
->event_size
== 0xFFFF) {
1292 event
->event_size
= ltt_get_uint32(LTT_GET_BO(tf
), pos
);
1293 pos
+= sizeof(guint32
);
1295 pos
+= ltt_align((size_t)pos
, sizeof(guint64
), tf
->alignment
);
1296 tf
->buffer
.tsc
= ltt_get_uint64(LTT_GET_BO(tf
), pos
);
1297 pos
+= sizeof(guint64
);
1299 case 30: /* LTT_RFLAG_ID_SIZE */
1300 event
->event_id
= ltt_get_uint16(LTT_GET_BO(tf
), pos
);
1301 pos
+= sizeof(guint16
);
1302 event
->event_size
= ltt_get_uint16(LTT_GET_BO(tf
), pos
);
1303 pos
+= sizeof(guint16
);
1304 if (event
->event_size
== 0xFFFF) {
1305 event
->event_size
= ltt_get_uint32(LTT_GET_BO(tf
), pos
);
1306 pos
+= sizeof(guint32
);
1309 case 31: /* LTT_RFLAG_ID */
1310 event
->event_id
= ltt_get_uint16(LTT_GET_BO(tf
), pos
);
1311 pos
+= sizeof(guint16
);
1312 event
->event_size
= G_MAXUINT
;
1315 event
->event_size
= G_MAXUINT
;
1319 if (likely(event
->event_id
!= 29)) {
1320 /* No extended timestamp */
1321 if (event
->timestamp
< (tf
->buffer
.tsc
& tf
->tsc_mask
))
1322 tf
->buffer
.tsc
= ((tf
->buffer
.tsc
& ~tf
->tsc_mask
) /* overflow */
1323 + tf
->tsc_mask_next_bit
)
1324 | (guint64
)event
->timestamp
;
1326 tf
->buffer
.tsc
= (tf
->buffer
.tsc
& ~tf
->tsc_mask
) /* no overflow */
1327 | (guint64
)event
->timestamp
;
1329 event
->tsc
= tf
->buffer
.tsc
;
1331 event
->event_time
= ltt_interpolate_time(tf
, event
);
1334 print_debug_event_header(event
, pos_aligned
, pos
);
1339 * Let ltt_update_event_size update event->data according to the largest
1340 * alignment within the payload.
1341 * Get the data size and update the event fields with the current
1343 ltt_update_event_size(tf
);
1349 /****************************************************************************
1351 * map_block : map a block from the file
1353 * lttdes : ltt trace file
1354 * whichBlock : the block which will be read
1357 * EINVAL : lseek fail
1358 * EIO : can not read from the file
1359 ****************************************************************************/
1361 static gint
map_block(LttTracefile
* tf
, guint block_num
)
1363 int page_size
= getpagesize();
1364 ltt_subbuffer_header_t
*header
;
1366 g_assert(block_num
< tf
->num_blocks
);
1368 if(tf
->buffer
.head
!= NULL
) {
1369 if(munmap(tf
->buffer
.head
, PAGE_ALIGN(tf
->buf_size
))) {
1370 g_warning("unmap size : %u\n",
1371 PAGE_ALIGN(tf
->buf_size
));
1372 perror("munmap error");
1377 /* Multiple of pages aligned head */
1378 tf
->buffer
.head
= mmap(0,
1379 PAGE_ALIGN(tf
->buf_size
),
1380 PROT_READ
, MAP_PRIVATE
, tf
->fd
,
1381 PAGE_ALIGN((off_t
)tf
->buf_size
* (off_t
)block_num
));
1383 if(tf
->buffer
.head
== MAP_FAILED
) {
1384 perror("Error in allocating memory for buffer of tracefile");
1388 g_assert( ( (gulong
)tf
->buffer
.head
&(8-1) ) == 0); // make sure it's aligned.
1391 tf
->buffer
.index
= block_num
;
1393 header
= (ltt_subbuffer_header_t
*)tf
->buffer
.head
;
1395 tf
->buffer
.begin
.cycle_count
= ltt_get_uint64(LTT_GET_BO(tf
),
1396 &header
->cycle_count_begin
);
1397 tf
->buffer
.begin
.freq
= tf
->trace
->start_freq
;
1399 tf
->buffer
.begin
.timestamp
= ltt_interpolate_time_from_tsc(tf
,
1400 tf
->buffer
.begin
.cycle_count
);
1401 tf
->buffer
.end
.cycle_count
= ltt_get_uint64(LTT_GET_BO(tf
),
1402 &header
->cycle_count_end
);
1403 tf
->buffer
.end
.freq
= tf
->trace
->start_freq
;
1405 tf
->buffer
.lost_size
= ltt_get_uint32(LTT_GET_BO(tf
),
1406 &header
->lost_size
);
1407 tf
->buffer
.end
.timestamp
= ltt_interpolate_time_from_tsc(tf
,
1408 tf
->buffer
.end
.cycle_count
);
1409 tf
->buffer
.tsc
= tf
->buffer
.begin
.cycle_count
;
1410 tf
->event
.tsc
= tf
->buffer
.tsc
;
1411 tf
->buffer
.freq
= tf
->buffer
.begin
.freq
;
1414 * eventually support variable buffer size : will need a partial pre-read of
1415 * the headers to create an index when we open the trace... eventually. */
1416 g_assert(tf
->buf_size
== ltt_get_uint32(LTT_GET_BO(tf
),
1417 &header
->buf_size
));
1419 /* Make the current event point to the beginning of the buffer :
1420 * it means that the event read must get the first event. */
1421 tf
->event
.tracefile
= tf
;
1422 tf
->event
.block
= block_num
;
1423 tf
->event
.offset
= 0;
1425 if (header
->events_lost
) {
1426 g_warning("%d events lost so far in tracefile %s at block %u",
1427 (guint
)header
->events_lost
,
1428 g_quark_to_string(tf
->long_name
),
1430 tf
->events_lost
= header
->events_lost
;
1432 if (header
->subbuf_corrupt
) {
1433 g_warning("%d subbuffer(s) corrupted so far in tracefile %s at block %u",
1434 (guint
)header
->subbuf_corrupt
,
1435 g_quark_to_string(tf
->long_name
),
1437 tf
->subbuf_corrupt
= header
->subbuf_corrupt
;
1446 static void print_debug_event_data(LttEvent
*ev
)
1448 unsigned int offset
= 0;
1451 if (!max(ev
->event_size
, ev
->data_size
))
1454 g_printf("Event data (tracefile %s offset %" PRIx64
"):\n",
1455 g_quark_to_string(ev
->tracefile
->long_name
),
1456 ((uint64_t)ev
->tracefile
->buffer
.index
* ev
->tracefile
->buf_size
)
1457 + (long)ev
->data
- (long)ev
->tracefile
->buffer
.head
);
1459 while (offset
< max(ev
->event_size
, ev
->data_size
)) {
1460 g_printf("%8lx", (long)ev
->data
+ offset
1461 - (long)ev
->tracefile
->buffer
.head
);
1464 for (i
= 0; i
< 4 ; i
++) {
1465 for (j
= 0; j
< 4; j
++) {
1466 if (offset
+ ((i
* 4) + j
) < max(ev
->event_size
, ev
->data_size
))
1467 g_printf("%02hhX", ((char*)ev
->data
)[offset
+ ((i
* 4) + j
)]);
1478 for (i
= 0; i
< 4; i
++) {
1479 for (j
= 0; j
< 4; j
++) {
1480 if (offset
+ ((i
* 4) + j
) < max(ev
->event_size
, ev
->data_size
)) {
1481 if (isprint(((char*)ev
->data
)[offset
+ ((i
* 4) + j
)]))
1482 g_printf("%c", ((char*)ev
->data
)[offset
+ ((i
* 4) + j
)]);
1494 /* It will update the fields offsets too */
1495 void ltt_update_event_size(LttTracefile
*tf
)
1498 struct marker_info
*info
;
1500 if (tf
->name
== LTT_TRACEFILE_NAME_METADATA
) {
1501 switch((enum marker_id
)tf
->event
.event_id
) {
1502 case MARKER_ID_SET_MARKER_ID
:
1503 size
= strlen((char*)tf
->event
.data
) + 1;
1504 g_debug("marker %s id set", (char*)tf
->event
.data
+ size
);
1505 size
+= strlen((char*)tf
->event
.data
+ size
) + 1;
1506 size
+= ltt_align(size
, sizeof(guint16
), tf
->alignment
);
1507 size
+= sizeof(guint16
);
1508 size
+= sizeof(guint8
);
1509 size
+= sizeof(guint8
);
1510 size
+= sizeof(guint8
);
1511 size
+= sizeof(guint8
);
1512 size
+= sizeof(guint8
);
1514 case MARKER_ID_SET_MARKER_FORMAT
:
1515 size
= strlen((char*)tf
->event
.data
) + 1;
1516 g_debug("marker %s format set", (char*)tf
->event
.data
);
1517 size
+= strlen((char*)tf
->event
.data
+ size
) + 1;
1518 size
+= strlen((char*)tf
->event
.data
+ size
) + 1;
1523 info
= marker_get_info_from_id(tf
->mdata
, tf
->event
.event_id
);
1525 if (tf
->event
.event_id
>= MARKER_CORE_IDS
)
1526 g_assert(info
!= NULL
);
1528 /* Do not update field offsets of core markers when initially reading the
1529 * metadata tracefile when the infos about these markers do not exist yet.
1531 if (likely(info
&& info
->fields
)) {
1533 tf
->event
.data
+= ltt_align((off_t
)(unsigned long)tf
->event
.data
,
1534 info
->largest_align
,
1536 /* size, dynamically computed */
1537 if (info
->size
!= -1)
1540 size
= marker_update_fields_offsets(marker_get_info_from_id(tf
->mdata
,
1541 tf
->event
.event_id
), tf
->event
.data
);
1544 tf
->event
.data_size
= size
;
1546 /* Check consistency between kernel and LTTV structure sizes */
1547 if(tf
->event
.event_size
== G_MAXUINT
) {
1548 /* Event size too big to fit in the event size field */
1549 tf
->event
.event_size
= tf
->event
.data_size
;
1553 print_debug_event_data(&tf
->event
);
1555 if (tf
->event
.data_size
!= tf
->event
.event_size
) {
1556 struct marker_info
*info
= marker_get_info_from_id(tf
->mdata
,
1557 tf
->event
.event_id
);
1559 g_error("Undescribed event %hhu in channel %s", tf
->event
.event_id
,
1560 g_quark_to_string(tf
->name
));
1561 g_error("Kernel/LTTV event size differs for event %s: kernel %u, LTTV %u",
1562 g_quark_to_string(info
->name
),
1563 tf
->event
.event_size
, tf
->event
.data_size
);
1569 /* Take the tf current event offset and use the event id to figure out where is
1570 * the next event offset.
1572 * This is an internal function not aiming at being used elsewhere : it will
1573 * not jump over the current block limits. Please consider using
1574 * ltt_tracefile_read to do this.
1576 * Returns 0 on success
1577 * ERANGE if we are at the end of the buffer.
1578 * ENOPROTOOPT if an error occured when getting the current event size.
1580 static int ltt_seek_next_event(LttTracefile
*tf
)
1585 /* seek over the buffer header if we are at the buffer start */
1586 if(tf
->event
.offset
== 0) {
1587 tf
->event
.offset
+= tf
->buffer_header_size
;
1589 if(tf
->event
.offset
== tf
->buf_size
- tf
->buffer
.lost_size
) {
1595 pos
= tf
->event
.data
;
1597 if(tf
->event
.data_size
< 0) goto error
;
1599 pos
+= (size_t)tf
->event
.data_size
;
1601 tf
->event
.offset
= pos
- tf
->buffer
.head
;
1603 if(tf
->event
.offset
== tf
->buf_size
- tf
->buffer
.lost_size
) {
1607 g_assert(tf
->event
.offset
< tf
->buf_size
- tf
->buffer
.lost_size
);
1613 g_error("Error in ltt_seek_next_event for tracefile %s",
1614 g_quark_to_string(tf
->name
));
1619 /*****************************************************************************
1621 * set_fields_offsets : set the precomputable offset of the fields
1623 * tracefile : opened trace file
1624 * event_type : the event type
1625 ****************************************************************************/
1627 void set_fields_offsets(LttTracefile
*tf
, LttEventType
*event_type
)
1629 LttField
*field
= event_type
->root_field
;
1630 enum field_status fixed_root
= FIELD_FIXED
, fixed_parent
= FIELD_FIXED
;
1633 preset_field_type_size(tf
, event_type
, 0, 0,
1634 &fixed_root
, &fixed_parent
,
1641 /*****************************************************************************
1643 * get_alignment : Get the alignment needed for a field.
1647 * returns : The size on which it must be aligned.
1649 ****************************************************************************/
1651 off_t
get_alignment(LttField
*field
)
1653 LttType
*type
= &field
->field_type
;
1655 switch(type
->type_class
) {
1657 case LTT_UINT_FIXED
:
1672 /* Align offset on type size */
1673 g_assert(field
->field_size
!= 0);
1674 return field
->field_size
;
1680 g_assert(type
->fields
->len
== 1);
1682 LttField
*child
= &g_array_index(type
->fields
, LttField
, 0);
1683 return get_alignment(child
);
1687 g_assert(type
->fields
->len
== 2);
1690 LttField
*child
= &g_array_index(type
->fields
, LttField
, 0);
1692 localign
= max(localign
, get_alignment(child
));
1694 child
= &g_array_index(type
->fields
, LttField
, 1);
1695 localign
= max(localign
, get_alignment(child
));
1706 for(i
=0; i
<type
->fields
->len
; i
++) {
1707 LttField
*child
= &g_array_index(type
->fields
, LttField
, i
);
1708 localign
= max(localign
, get_alignment(child
));
1715 g_error("get_alignment : unknown type");
1722 /*****************************************************************************
1724 * field_compute_static_size : Determine the size of fields known by their
1725 * sole definition. Unions, arrays and struct sizes might be known, but
1726 * the parser does not give that information.
1731 ****************************************************************************/
1733 void field_compute_static_size(LttFacility
*fac
, LttField
*field
)
1735 LttType
*type
= &field
->field_type
;
1737 switch(type
->type_class
) {
1739 case LTT_UINT_FIXED
:
1758 /* note this : array type size is the number of elements in the array,
1759 * while array field size of the length of the array in bytes */
1760 g_assert(type
->fields
->len
== 1);
1762 LttField
*child
= &g_array_index(type
->fields
, LttField
, 0);
1763 field_compute_static_size(fac
, child
);
1765 if(child
->field_size
!= 0) {
1766 field
->field_size
= type
->size
* child
->field_size
;
1767 field
->dynamic_offsets
= g_array_sized_new(FALSE
, TRUE
,
1768 sizeof(off_t
), type
->size
);
1770 field
->field_size
= 0;
1775 g_assert(type
->fields
->len
== 2);
1777 off_t local_offset
= 0;
1778 LttField
*child
= &g_array_index(type
->fields
, LttField
, 1);
1779 field_compute_static_size(fac
, child
);
1780 field
->field_size
= 0;
1782 if(child
->field_size
!= 0) {
1783 field
->dynamic_offsets
= g_array_sized_new(FALSE
, TRUE
,
1784 sizeof(off_t
), SEQUENCE_AVG_ELEMENTS
);
1792 for(i
=0;i
<type
->fields
->len
;i
++) {
1793 LttField
*child
= &g_array_index(type
->fields
, LttField
, i
);
1794 field_compute_static_size(fac
, child
);
1795 if(child
->field_size
!= 0) {
1796 type
->size
+= ltt_align(type
->size
, get_alignment(child
),
1798 type
->size
+= child
->field_size
;
1800 /* As soon as we find a child with variable size, we have
1801 * a variable size */
1806 field
->field_size
= type
->size
;
1810 g_error("field_static_size : unknown type");
1817 /*****************************************************************************
1819 * precompute_fields_offsets : set the precomputable offset of the fields
1823 * offset : pointer to the current offset, must be incremented
1825 * return : 1 : found a variable length field, stop the processing.
1827 ****************************************************************************/
1830 gint
precompute_fields_offsets(LttFacility
*fac
, LttField
*field
, off_t
*offset
, gint is_compact
)
1832 LttType
*type
= &field
->field_type
;
1834 if(unlikely(is_compact
)) {
1835 g_assert(field
->field_size
!= 0);
1836 /* FIXME THIS IS A HUUUUUGE hack :
1837 * offset is between the compact_data field in struct LttEvent
1838 * and the address of the field root in the memory map.
1839 * ark. Both will stay at the same addresses while the event
1840 * is readable, so it's ok.
1842 field
->offset_root
= 0;
1843 field
->fixed_root
= FIELD_FIXED
;
1847 switch(type
->type_class
) {
1849 case LTT_UINT_FIXED
:
1864 g_assert(field
->field_size
!= 0);
1865 /* Align offset on type size */
1866 *offset
+= ltt_align(*offset
, get_alignment(field
),
1868 /* remember offset */
1869 field
->offset_root
= *offset
;
1870 field
->fixed_root
= FIELD_FIXED
;
1871 /* Increment offset */
1872 *offset
+= field
->field_size
;
1876 field
->offset_root
= *offset
;
1877 field
->fixed_root
= FIELD_FIXED
;
1881 g_assert(type
->fields
->len
== 1);
1883 LttField
*child
= &g_array_index(type
->fields
, LttField
, 0);
1885 *offset
+= ltt_align(*offset
, get_alignment(field
),
1888 /* remember offset */
1889 field
->offset_root
= *offset
;
1890 field
->array_offset
= *offset
;
1891 field
->fixed_root
= FIELD_FIXED
;
1893 /* Let the child be variable */
1894 //precompute_fields_offsets(tf, child, offset);
1896 if(field
->field_size
!= 0) {
1897 /* Increment offset */
1898 /* field_size is the array size in bytes */
1899 *offset
+= field
->field_size
;
1907 g_assert(type
->fields
->len
== 2);
1912 *offset
+= ltt_align(*offset
, get_alignment(field
),
1915 /* remember offset */
1916 field
->offset_root
= *offset
;
1917 field
->fixed_root
= FIELD_FIXED
;
1919 child
= &g_array_index(type
->fields
, LttField
, 0);
1920 ret
= precompute_fields_offsets(fac
, child
, offset
, is_compact
);
1921 g_assert(ret
== 0); /* Seq len cannot have variable len */
1923 child
= &g_array_index(type
->fields
, LttField
, 1);
1924 *offset
+= ltt_align(*offset
, get_alignment(child
),
1926 field
->array_offset
= *offset
;
1927 /* Let the child be variable. */
1928 //ret = precompute_fields_offsets(fac, child, offset);
1930 /* Cannot precompute fields offsets of sequence members, and has
1931 * variable length. */
1941 *offset
+= ltt_align(*offset
, get_alignment(field
),
1943 /* remember offset */
1944 field
->offset_root
= *offset
;
1945 field
->fixed_root
= FIELD_FIXED
;
1947 for(i
=0; i
< type
->fields
->len
; i
++) {
1948 child
= &g_array_index(type
->fields
, LttField
, i
);
1949 ret
= precompute_fields_offsets(fac
, child
, offset
, is_compact
);
1962 *offset
+= ltt_align(*offset
, get_alignment(field
),
1964 /* remember offset */
1965 field
->offset_root
= *offset
;
1966 field
->fixed_root
= FIELD_FIXED
;
1968 for(i
=0; i
< type
->fields
->len
; i
++) {
1969 *offset
= field
->offset_root
;
1970 child
= &g_array_index(type
->fields
, LttField
, i
);
1971 ret
= precompute_fields_offsets(fac
, child
, offset
, is_compact
);
1975 *offset
= field
->offset_root
+ field
->field_size
;
1982 g_error("precompute_fields_offsets : unknown type");
1991 /*****************************************************************************
1993 * precompute_offsets : set the precomputable offset of an event type
1996 * event : event type
1998 ****************************************************************************/
1999 void precompute_offsets(LttFacility
*fac
, LttEventType
*event
)
2005 /* First, compute the size of fixed size fields. Will determine size for
2006 * arrays, struct and unions, which is not done by the parser */
2007 for(i
=0; i
<event
->fields
->len
; i
++) {
2008 LttField
*field
= &g_array_index(event
->fields
, LttField
, i
);
2009 field_compute_static_size(fac
, field
);
2012 /* Precompute all known offsets */
2013 for(i
=0; i
<event
->fields
->len
; i
++) {
2014 LttField
*field
= &g_array_index(event
->fields
, LttField
, i
);
2015 if(event
->has_compact_data
&& i
== 0)
2016 ret
= precompute_fields_offsets(fac
, field
, &offset
, 1);
2018 ret
= precompute_fields_offsets(fac
, field
, &offset
, 0);
2026 /*****************************************************************************
2028 * preset_field_type_size : set the fixed sizes of the field type
2031 * event_type : event type
2032 * offset_root : offset from the root
2033 * offset_parent : offset from the parent
2034 * fixed_root : Do we know a fixed offset to the root ?
2035 * fixed_parent : Do we know a fixed offset to the parent ?
2037 ****************************************************************************/
2041 // preset the fixed size offsets. Calculate them just like genevent-new : an
2042 // increment of a *to value that represents the offset from the start of the
2044 // The preset information is : offsets up to (and including) the first element
2045 // of variable size. All subsequent fields must be flagged "VARIABLE OFFSET".
2047 void preset_field_type_size(LttTracefile
*tf
, LttEventType
*event_type
,
2048 off_t offset_root
, off_t offset_parent
,
2049 enum field_status
*fixed_root
, enum field_status
*fixed_parent
,
2052 enum field_status local_fixed_root
, local_fixed_parent
;
2056 g_assert(field
->fixed_root
== FIELD_UNKNOWN
);
2057 g_assert(field
->fixed_parent
== FIELD_UNKNOWN
);
2058 g_assert(field
->fixed_size
== FIELD_UNKNOWN
);
2060 type
= field
->field_type
;
2062 field
->fixed_root
= *fixed_root
;
2063 if(field
->fixed_root
== FIELD_FIXED
)
2064 field
->offset_root
= offset_root
;
2066 field
->offset_root
= 0;
2068 field
->fixed_parent
= *fixed_parent
;
2069 if(field
->fixed_parent
== FIELD_FIXED
)
2070 field
->offset_parent
= offset_parent
;
2072 field
->offset_parent
= 0;
2074 size_t current_root_offset
;
2075 size_t current_offset
;
2076 enum field_status current_child_status
, final_child_status
;
2079 switch(type
->type_class
) {
2081 case LTT_UINT_FIXED
:
2090 field
->field_size
= ltt_type_size(tf
->trace
, type
);
2091 field
->fixed_size
= FIELD_FIXED
;
2094 field
->field_size
= (off_t
)event_type
->facility
->pointer_size
;
2095 field
->fixed_size
= FIELD_FIXED
;
2099 field
->field_size
= (off_t
)event_type
->facility
->long_size
;
2100 field
->fixed_size
= FIELD_FIXED
;
2105 field
->field_size
= (off_t
)event_type
->facility
->size_t_size
;
2106 field
->fixed_size
= FIELD_FIXED
;
2109 local_fixed_root
= FIELD_VARIABLE
;
2110 local_fixed_parent
= FIELD_VARIABLE
;
2111 preset_field_type_size(tf
, event_type
,
2113 &local_fixed_root
, &local_fixed_parent
,
2115 field
->fixed_size
= FIELD_VARIABLE
;
2116 field
->field_size
= 0;
2117 *fixed_root
= FIELD_VARIABLE
;
2118 *fixed_parent
= FIELD_VARIABLE
;
2121 field
->fixed_size
= FIELD_VARIABLE
;
2122 field
->field_size
= 0;
2123 *fixed_root
= FIELD_VARIABLE
;
2124 *fixed_parent
= FIELD_VARIABLE
;
2127 local_fixed_root
= FIELD_VARIABLE
;
2128 local_fixed_parent
= FIELD_VARIABLE
;
2129 preset_field_type_size(tf
, event_type
,
2131 &local_fixed_root
, &local_fixed_parent
,
2133 field
->fixed_size
= field
->child
[0]->fixed_size
;
2134 if(field
->fixed_size
== FIELD_FIXED
) {
2135 field
->field_size
= type
->element_number
* field
->child
[0]->field_size
;
2137 field
->field_size
= 0;
2138 *fixed_root
= FIELD_VARIABLE
;
2139 *fixed_parent
= FIELD_VARIABLE
;
2143 current_root_offset
= field
->offset_root
;
2145 current_child_status
= FIELD_FIXED
;
2146 for(i
=0;i
<type
->element_number
;i
++) {
2147 preset_field_type_size(tf
, event_type
,
2148 current_root_offset
, current_offset
,
2149 fixed_root
, ¤t_child_status
,
2151 if(current_child_status
== FIELD_FIXED
) {
2152 current_root_offset
+= field
->child
[i
]->field_size
;
2153 current_offset
+= field
->child
[i
]->field_size
;
2155 current_root_offset
= 0;
2159 if(current_child_status
!= FIELD_FIXED
) {
2160 *fixed_parent
= current_child_status
;
2161 field
->field_size
= 0;
2162 field
->fixed_size
= current_child_status
;
2164 field
->field_size
= current_offset
;
2165 field
->fixed_size
= FIELD_FIXED
;
2169 current_root_offset
= field
->offset_root
;
2172 final_child_status
= FIELD_FIXED
;
2173 for(i
=0;i
<type
->element_number
;i
++) {
2174 enum field_status current_root_child_status
= FIELD_FIXED
;
2175 enum field_status current_child_status
= FIELD_FIXED
;
2176 preset_field_type_size(tf
, event_type
,
2177 current_root_offset
, current_offset
,
2178 ¤t_root_child_status
, ¤t_child_status
,
2180 if(current_child_status
!= FIELD_FIXED
)
2181 final_child_status
= current_child_status
;
2183 max_size
= max(max_size
, field
->child
[i
]->field_size
);
2185 if(final_child_status
!= FIELD_FIXED
) {
2186 g_error("LTTV does not support variable size fields in unions.");
2187 /* This will stop the application. */
2188 *fixed_root
= final_child_status
;
2189 *fixed_parent
= final_child_status
;
2190 field
->field_size
= 0;
2191 field
->fixed_size
= current_child_status
;
2193 field
->field_size
= max_size
;
2194 field
->fixed_size
= FIELD_FIXED
;
2198 g_error("unexpected type NONE");
2205 /*****************************************************************************
2207 * check_fields_compatibility : Check for compatibility between two fields :
2208 * do they use the same inner structure ?
2210 * event_type1 : event type
2211 * event_type2 : event type
2214 *Returns : 0 if identical
2216 ****************************************************************************/
2217 // this function checks for equality of field types. Therefore, it does not use
2218 // per se offsets. For instance, an aligned version of a structure is
2219 // compatible with an unaligned version of the same structure.
2221 gint
check_fields_compatibility(LttEventType
*event_type1
,
2222 LttEventType
*event_type2
,
2223 LttField
*field1
, LttField
*field2
)
2225 guint different
= 0;
2229 if(field1
== NULL
) {
2230 if(field2
== NULL
) goto end
;
2235 } else if(field2
== NULL
) {
2240 type1
= &field1
->field_type
;
2241 type2
= &field2
->field_type
;
2243 if(type1
->type_class
!= type2
->type_class
) {
2247 if(type1
->network
!= type2
->network
) {
2252 switch(type1
->type_class
) {
2254 case LTT_UINT_FIXED
:
2269 if(field1
->field_size
!= field2
->field_size
)
2276 LttField
*child1
= &g_array_index(type1
->fields
, LttField
, 0);
2277 LttField
*child2
= &g_array_index(type2
->fields
, LttField
, 0);
2279 if(type1
->size
!= type2
->size
)
2281 if(check_fields_compatibility(event_type1
, event_type2
, child1
, child2
))
2287 LttField
*child1
= &g_array_index(type1
->fields
, LttField
, 1);
2288 LttField
*child2
= &g_array_index(type2
->fields
, LttField
, 1);
2290 if(check_fields_compatibility(event_type1
, event_type2
, child1
, child2
))
2300 if(type1
->fields
->len
!= type2
->fields
->len
) {
2305 for(i
=0; i
< type1
->fields
->len
; i
++) {
2308 child1
= &g_array_index(type1
->fields
, LttField
, i
);
2309 child2
= &g_array_index(type2
->fields
, LttField
, i
);
2310 different
= check_fields_compatibility(event_type1
,
2311 event_type2
, child1
, child2
);
2313 if(different
) break;
2319 g_error("check_fields_compatibility : unknown type");
2328 gint
check_fields_compatibility(LttEventType
*event_type1
,
2329 LttEventType
*event_type2
,
2330 LttField
*field1
, LttField
*field2
)
2332 guint different
= 0;
2337 if(field1
== NULL
) {
2338 if(field2
== NULL
) goto end
;
2343 } else if(field2
== NULL
) {
2348 g_assert(field1
->fixed_root
!= FIELD_UNKNOWN
);
2349 g_assert(field2
->fixed_root
!= FIELD_UNKNOWN
);
2350 g_assert(field1
->fixed_parent
!= FIELD_UNKNOWN
);
2351 g_assert(field2
->fixed_parent
!= FIELD_UNKNOWN
);
2352 g_assert(field1
->fixed_size
!= FIELD_UNKNOWN
);
2353 g_assert(field2
->fixed_size
!= FIELD_UNKNOWN
);
2355 type1
= field1
->field_type
;
2356 type2
= field2
->field_type
;
2358 if(type1
->type_class
!= type2
->type_class
) {
2362 if(type1
->element_name
!= type2
->element_name
) {
2367 switch(type1
->type_class
) {
2369 case LTT_UINT_FIXED
:
2384 if(field1
->field_size
!= field2
->field_size
) {
2390 if(type1
->element_number
!= type2
->element_number
) {
2394 for(i
=0;i
<type1
->element_number
;i
++) {
2395 if(type1
->enum_strings
[i
] != type2
->enum_strings
[i
]) {
2402 /* Two elements : size and child */
2403 g_assert(type1
->element_number
!= type2
->element_number
);
2404 for(i
=0;i
<type1
->element_number
;i
++) {
2405 if(check_fields_compatibility(event_type1
, event_type2
,
2406 field1
->child
[0], field2
->child
[0])) {
2415 if(field1
->field_size
!= field2
->field_size
) {
2419 /* Two elements : size and child */
2420 g_assert(type1
->element_number
!= type2
->element_number
);
2421 for(i
=0;i
<type1
->element_number
;i
++) {
2422 if(check_fields_compatibility(event_type1
, event_type2
,
2423 field1
->child
[0], field2
->child
[0])) {
2431 if(type1
->element_number
!= type2
->element_number
) {
2435 for(i
=0;i
<type1
->element_number
;i
++) {
2436 if(check_fields_compatibility(event_type1
, event_type2
,
2437 field1
->child
[0], field2
->child
[0])) {
2450 /*****************************************************************************
2452 * ltt_get_int : get an integer number
2454 * reverse_byte_order: must we reverse the byte order ?
2455 * size : the size of the integer
2456 * ptr : the data pointer
2458 * gint64 : a 64 bits integer
2459 ****************************************************************************/
2461 gint64
ltt_get_int(gboolean reverse_byte_order
, gint size
, void *data
)
2466 case 1: val
= *((gint8
*)data
); break;
2467 case 2: val
= ltt_get_int16(reverse_byte_order
, data
); break;
2468 case 4: val
= ltt_get_int32(reverse_byte_order
, data
); break;
2469 case 8: val
= ltt_get_int64(reverse_byte_order
, data
); break;
2470 default: val
= ltt_get_int64(reverse_byte_order
, data
);
2471 g_critical("get_int : integer size %d unknown", size
);
2478 /*****************************************************************************
2480 * ltt_get_uint : get an unsigned integer number
2482 * reverse_byte_order: must we reverse the byte order ?
2483 * size : the size of the integer
2484 * ptr : the data pointer
2486 * guint64 : a 64 bits unsigned integer
2487 ****************************************************************************/
2489 guint64
ltt_get_uint(gboolean reverse_byte_order
, gint size
, void *data
)
2494 case 1: val
= *((gint8
*)data
); break;
2495 case 2: val
= ltt_get_uint16(reverse_byte_order
, data
); break;
2496 case 4: val
= ltt_get_uint32(reverse_byte_order
, data
); break;
2497 case 8: val
= ltt_get_uint64(reverse_byte_order
, data
); break;
2498 default: val
= ltt_get_uint64(reverse_byte_order
, data
);
2499 g_critical("get_uint : unsigned integer size %d unknown",
2508 /* get the node name of the system */
2510 char * ltt_trace_system_description_node_name (LttSystemDescription
* s
)
2512 return s
->node_name
;
2516 /* get the domain name of the system */
2518 char * ltt_trace_system_description_domain_name (LttSystemDescription
* s
)
2520 return s
->domain_name
;
2524 /* get the description of the system */
2526 char * ltt_trace_system_description_description (LttSystemDescription
* s
)
2528 return s
->description
;
2532 /* get the NTP corrected start time of the trace */
2533 LttTime
ltt_trace_start_time(LttTrace
*t
)
2535 return t
->start_time
;
2538 /* get the monotonic start time of the trace */
2539 LttTime
ltt_trace_start_time_monotonic(LttTrace
*t
)
2541 return t
->start_time_from_tsc
;
2544 static __attribute__ ((__unused__
)) LttTracefile
*ltt_tracefile_new()
2547 tf
= g_new(LttTracefile
, 1);
2548 tf
->event
.tracefile
= tf
;
2552 static __attribute__ ((__unused__
)) void ltt_tracefile_destroy(LttTracefile
*tf
)
2557 static __attribute__ ((__unused__
)) void ltt_tracefile_copy(LttTracefile
*dest
, const LttTracefile
*src
)
2562 /* Before library loading... */
2564 static __attribute__((constructor
)) void init(void)
2566 LTT_TRACEFILE_NAME_METADATA
= g_quark_from_string("metadata");