only facilities missing
authorcompudj <compudj@04897980-b3bd-0310-b5e0-8ef037075253>
Wed, 8 Mar 2006 17:08:48 +0000 (17:08 +0000)
committercompudj <compudj@04897980-b3bd-0310-b5e0-8ef037075253>
Wed, 8 Mar 2006 17:08:48 +0000 (17:08 +0000)
git-svn-id: http://ltt.polymtl.ca/svn@1616 04897980-b3bd-0310-b5e0-8ef037075253

usertrace-fast/ltt-usertrace-fast.c
usertrace-fast/ltt-usertrace-fast.h
usertrace-fast/test.c

index 990b0211bbf0b5025c3d0ccad5c8446fe39988f2..fc919c348452d8a52dffcdadea560daa7c462d2a 100644 (file)
@@ -318,22 +318,6 @@ static inline int ltt_buffer_put(struct ltt_buf *ltt_buf,
        }
 }
 
-/* In the writer :
- *
- * if(buffer full condition) {
- *   put myself in the wait queue
- *   ltt_buf->full = 1;
- *   schedule
- * }
- *{
-       if(buffer_is_full) {
-               atomic_set(&ltt_buf->full, 1);
-               ret = do_futex((unsigned long)&ltt_buf->full, 1, 0, 0, 0);
-       }
-}
-
- */
-
 static int read_subbuffer(struct ltt_buf *ltt_buf, int fd)
 {
        unsigned int consumed_old;
@@ -342,8 +326,8 @@ static int read_subbuffer(struct ltt_buf *ltt_buf, int fd)
 
 
        err = ltt_buffer_get(ltt_buf, &consumed_old);
-       if(err != -EAGAIN && err != 0) {
-               printf("LTT Reserving sub buffer failed\n");
+       if(err != 0) {
+               if(err != -EAGAIN) printf("LTT Reserving sub buffer failed\n");
                goto get_error;
        }
 
@@ -509,11 +493,16 @@ void ltt_rw_init(void)
        shared_trace_info->channel.facilities.subbuf_size = LTT_SUBBUF_SIZE_FACILITIES;
        shared_trace_info->channel.facilities.start =
                shared_trace_info->channel.facilities_buf;
+       ltt_buffer_begin_callback(&shared_trace_info->channel.facilities,
+                       ltt_get_timestamp(), 0);
 
        atomic_set(&shared_trace_info->channel.cpu.full, 0);
        shared_trace_info->channel.cpu.alloc_size = LTT_BUF_SIZE_CPU;
        shared_trace_info->channel.cpu.subbuf_size = LTT_SUBBUF_SIZE_CPU;
        shared_trace_info->channel.cpu.start = shared_trace_info->channel.cpu_buf;
+       ltt_buffer_begin_callback(&shared_trace_info->channel.cpu,
+                       ltt_get_timestamp(), 0);
+       
        shared_trace_info->init = 1;
 
        /* Disable signals */
index 9953c15a62cb0eb7579ffce11d1cb3e4b2a3058a..5f7915306171e82278af1e0447d5d35a7467bbec 100644 (file)
 #include <linux/futex.h>
 
 #ifndef futex
-static inline _syscall6(long, futex, unsigned long, uaddr, int, op, int, val,
+static inline __attribute__((no_instrument_function))
+       _syscall6(long, futex, unsigned long, uaddr, int, op, int, val,
                unsigned long, timeout, unsigned long, uaddr2, int, val2)
 #endif //futex
 
 
+
 #ifndef        LTT_N_SUBBUFS
 #define LTT_N_SUBBUFS 2
 #endif //LTT_N_SUBBUFS
@@ -61,6 +63,9 @@ static inline _syscall6(long, futex, unsigned long, uaddr, int, op, int, val,
 #ifndef atomic_cmpxchg
 #define atomic_cmpxchg(v, old, new) ((int)cmpxchg(&((v)->counter), old, new))
 #endif //atomic_cmpxchg
+
+typedef unsigned int ltt_facility_t;
+       
 struct ltt_trace_header {
        uint32_t                                magic_number;
        uint32_t                                arch_type;
@@ -126,28 +131,173 @@ struct ltt_trace_info {
 };
 
 
+struct ltt_event_header_nohb {
+  uint64_t      timestamp;
+  unsigned char facility_id;
+  unsigned char event_id;
+  uint16_t      event_size;
+} __attribute((packed));
 
 extern __thread struct ltt_trace_info *thread_trace_info;
 
 void ltt_thread_init(void);
 
-void ltt_usertrace_fast_buffer_switch(void);
+void __attribute__((no_instrument_function))
+       ltt_usertrace_fast_buffer_switch(void);
+
+#ifndef LTT_PACK
+/* Calculate the offset needed to align the type */
+static inline unsigned int __attribute__((no_instrument_function))
+                            ltt_align(size_t align_drift,
+                                      size_t size_of_type)
+{
+  size_t alignment = min(sizeof(void*), size_of_type);
+
+  return ((alignment - align_drift) & (alignment-1));
+}
+#define LTT_ALIGN
+#else
+static inline unsigned int __attribute__((no_instrument_function))
+                            ltt_align(size_t align_drift,
+                                      size_t size_of_type)
+{
+  return 0;
+}
+#define LTT_ALIGN __attribute__((packed))
+#endif //LTT_PACK
+
+/* Get the offset of the channel in the ltt_trace_struct */
+#define GET_CHANNEL_INDEX(chan) \
+  (unsigned int)&((struct ltt_trace_info*)NULL)->channel.chan
+
+/* ltt_get_index_from_facility
+ *
+ * Get channel index from facility and event id.
+ * 
+ * @fID : facility ID
+ * @eID : event number
+ *
+ * Get the channel index into which events must be written for the given
+ * facility and event number. We get this structure offset as soon as possible
+ * and remember it so we pass through this logic only once per trace call (not
+ * for every trace).
+ */
+static inline unsigned int __attribute__((no_instrument_function))
+               ltt_get_index_from_facility(ltt_facility_t fID,
+                                                                                                                               uint8_t eID)
+{
+
+       if(fID == ltt_facility_core) {
+               switch(eID) {
+                       case event_core_facility_load:
+                       case event_core_facility_unload:
+                       case event_core_state_dump_facility_load:
+                               return GET_CHANNEL_INDEX(facilities);
+                       default:
+                               return GET_CHANNEL_INDEX(cpu);
+               }
+       }
+       return GET_CHANNEL_INDEX(cpu);
+}
+
+
+static inline struct ltt_buf * __attribute__((no_instrument_function))
+       ltt_get_channel_from_index(
+               struct ltt_trace_info *trace, unsigned int index)
+{
+       return *(struct ltt_buf **)((void*)trace+index);
+}
+
+
+/*
+ * ltt_get_header_size
+ *
+ * Calculate alignment offset for arch size void*. This is the
+ * alignment offset of the event header.
+ *
+ * Important note :
+ * The event header must be a size multiple of the void* size. This is necessary
+ * to be able to calculate statically the alignment offset of the variable
+ * length data fields that follows. The total offset calculated here :
+ *
+ *   Alignment of header struct on arch size
+ * + sizeof(header struct)
+ * + padding added to end of struct to align on arch size.
+ * */
+static inline unsigned char __attribute__((no_instrument_function))
+                                                                                                               ltt_get_header_size(struct ltt_trace_info *trace,
+                                                void *address,
+                                                size_t *before_hdr_pad,
+                                                size_t *after_hdr_pad,
+                                                size_t *header_size)
+{
+  unsigned int padding;
+  unsigned int header;
+
+  header = sizeof(struct ltt_event_header_nohb);
+
+  /* Padding before the header. Calculated dynamically */
+  *before_hdr_pad = ltt_align((unsigned long)address, header);
+  padding = *before_hdr_pad;
+
+  /* Padding after header, considering header aligned on ltt_align.
+   * Calculated statically if header size if known. */
+  *after_hdr_pad = ltt_align(header, sizeof(void*));
+  padding += *after_hdr_pad;
+
+  *header_size = header;
+
+  return header+padding;
+}
+
+
+/* ltt_write_event_header
+ *
+ * Writes the event header to the pointer.
+ *
+ * @channel : pointer to the channel structure
+ * @ptr : buffer pointer
+ * @fID : facility ID
+ * @eID : event ID
+ * @event_size : size of the event, excluding the event header.
+ * @offset : offset of the beginning of the header, for alignment.
+ *                                      Calculated by ltt_get_event_header_size.
+ * @tsc : time stamp counter.
+ */
+static inline void __attribute__((no_instrument_function))
+       ltt_write_event_header(
+               struct ltt_trace_info *trace, struct ltt_buf *buf,
+               void *ptr, ltt_facility_t fID, uint32_t eID, size_t event_size,
+               size_t offset, uint64_t tsc)
+{
+       struct ltt_event_header_nohb *nohb;
+       
+       event_size = min(event_size, 0xFFFFU);
+       nohb = (struct ltt_event_header_nohb *)(ptr+offset);
+       nohb->timestamp = (uint64_t)tsc;
+       nohb->facility_id = fID;
+       nohb->event_id = eID;
+       nohb->event_size = (uint16_t)event_size;
+}
 
 
 
-static inline uint64_t ltt_get_timestamp()
+static inline uint64_t __attribute__((no_instrument_function))
+ltt_get_timestamp()
 {
        return get_cycles();
 }
 
-static inline unsigned int ltt_subbuf_header_len(struct ltt_buf *buf)
+static inline unsigned int __attribute__((no_instrument_function))
+ltt_subbuf_header_len(struct ltt_buf *buf)
 {
        return sizeof(struct ltt_block_start_header);
 }
 
 
 
-static inline void ltt_write_trace_header(struct ltt_trace_header *header)
+static inline void __attribute__((no_instrument_function))
+ltt_write_trace_header(struct ltt_trace_header *header)
 {
        header->magic_number = LTT_TRACER_MAGIC_NUMBER;
        header->major_version = LTT_TRACER_VERSION_MAJOR;
@@ -159,7 +309,7 @@ static inline void ltt_write_trace_header(struct ltt_trace_header *header)
        header->flight_recorder = 0;
        header->has_heartbeat = 0;
 
-#ifdef CONFIG_LTT_ALIGNMENT
+#ifndef LTT_PACK
        header->has_alignment = sizeof(void*);
 #else
        header->has_alignment = 0;
@@ -175,7 +325,8 @@ static inline void ltt_write_trace_header(struct ltt_trace_header *header)
 }
 
 
-static inline void ltt_buffer_begin_callback(struct ltt_buf *buf,
+static inline void __attribute__((no_instrument_function))
+ltt_buffer_begin_callback(struct ltt_buf *buf,
                      uint64_t tsc, unsigned int subbuf_idx)
 {
        struct ltt_block_start_header *header = 
@@ -195,7 +346,8 @@ static inline void ltt_buffer_begin_callback(struct ltt_buf *buf,
 
 
 
-static inline void ltt_buffer_end_callback(struct ltt_buf *buf,
+static inline void __attribute__((no_instrument_function))
+ltt_buffer_end_callback(struct ltt_buf *buf,
                      uint64_t tsc, unsigned int offset, unsigned int subbuf_idx)
 {
        struct ltt_block_start_header *header = 
@@ -211,10 +363,302 @@ static inline void ltt_buffer_end_callback(struct ltt_buf *buf,
 }
 
 
-static inline void ltt_deliver_callback(struct ltt_buf *buf,
+static inline void __attribute__((no_instrument_function))
+ltt_deliver_callback(struct ltt_buf *buf,
     unsigned subbuf_idx,
     void *subbuf)
 {
        ltt_usertrace_fast_buffer_switch();
 }
+
+
+/* ltt_reserve_slot
+ *
+ * Atomic slot reservation in a LTTng buffer. It will take care of
+ * sub-buffer switching.
+ *
+ * Parameters:
+ *
+ * @trace : the trace structure to log to.
+ * @buf : the buffer to reserve space into.
+ * @data_size : size of the variable length data to log.
+ * @slot_size : pointer to total size of the slot (out)
+ * @tsc : pointer to the tsc at the slot reservation (out)
+ * @before_hdr_pad : dynamic padding before the event header.
+ * @after_hdr_pad : dynamic padding after the event header.
+ *
+ * Return : NULL if not enough space, else returns the pointer
+ *                                     to the beginning of the reserved slot. */
+static inline void * __attribute__((no_instrument_function)) ltt_reserve_slot(
+                                                                                                                       struct ltt_trace_info *trace,
+                                                                                                                       struct ltt_buf *ltt_buf,
+                                                                                                                       unsigned int data_size,
+                                                                                                                       unsigned int *slot_size,
+                                                                                                                       uint64_t *tsc,
+                                                                                                                       size_t *before_hdr_pad,
+                                                                                                                       size_t *after_hdr_pad,
+                                                                                                                       size_t *header_size)
+{
+       int offset_begin, offset_end, offset_old;
+       //int has_switch;
+       int begin_switch, end_switch_current, end_switch_old;
+       int reserve_commit_diff = 0;
+       unsigned int size;
+       int consumed_old, consumed_new;
+       int commit_count, reserve_count;
+       int ret;
+
+       do {
+               offset_old = atomic_read(&ltt_buf->offset);
+               offset_begin = offset_old;
+               //has_switch = 0;
+               begin_switch = 0;
+               end_switch_current = 0;
+               end_switch_old = 0;
+               *tsc = ltt_get_timestamp();
+               if(*tsc == 0) {
+                       /* Error in getting the timestamp, event lost */
+                       atomic_inc(&ltt_buf->events_lost);
+                       return NULL;
+               }
+
+               if(SUBBUF_OFFSET(offset_begin, ltt_buf) == 0) {
+                       begin_switch = 1; /* For offset_begin */
+               } else {
+                       size = ltt_get_header_size(trace, ltt_buf->start + offset_begin,
+                                                                                                                                before_hdr_pad, after_hdr_pad, header_size)
+                                                + data_size;
+
+                       if((SUBBUF_OFFSET(offset_begin, ltt_buf)+size)>ltt_buf->subbuf_size) {
+                               //has_switch = 1;
+                               end_switch_old = 1;     /* For offset_old */
+                               begin_switch = 1;       /* For offset_begin */
+                       }
+               }
+
+               if(begin_switch) {
+                       if(end_switch_old) {
+                               offset_begin = SUBBUF_ALIGN(offset_begin, ltt_buf);
+                       }
+                       offset_begin = offset_begin + ltt_subbuf_header_len(ltt_buf);
+                       /* Test new buffer integrity */
+                       reserve_commit_diff =
+                               atomic_read(&ltt_buf->reserve_count[SUBBUF_INDEX(offset_begin,
+                                                                                                                                                                                                                                ltt_buf)])
+                               - atomic_read(&ltt_buf->commit_count[SUBBUF_INDEX(offset_begin,
+                                                                                                                                                                                ltt_buf)]);
+                       if(reserve_commit_diff == 0) {
+                               /* Next buffer not corrupted. */
+                               if((SUBBUF_TRUNC(offset_begin, ltt_buf) 
+                                                               - SUBBUF_TRUNC(atomic_read(&ltt_buf->consumed), ltt_buf))
+                                                                       >= ltt_buf->alloc_size) {
+                                       /* We block until the reader unblocks us */
+                                       atomic_set(&ltt_buf->full, 1);
+                                       /* We block until the reader tells us to wake up.
+                                                Signals will simply cause this loop to restart.
+                                                */
+                                       do {
+                                               ret = futex((unsigned long)&ltt_buf->full, FUTEX_WAIT, 1, 0, 0, 0);
+                                       } while(ret != 0 && ret != EWOULDBLOCK);
+                                       /* go on with the write */
+
+                               } else {
+                                       /* next buffer not corrupted, we are either in overwrite mode or
+                                        * the buffer is not full. It's safe to write in this new subbuffer.*/
+                               }
+                       } else {
+                               /* Next subbuffer corrupted. Force pushing reader even in normal
+                                * mode. It's safe to write in this new subbuffer. */
+                       }
+                       size = ltt_get_header_size(trace, ltt_buf->start + offset_begin,
+                                       before_hdr_pad, after_hdr_pad, header_size) + data_size;
+                       if((SUBBUF_OFFSET(offset_begin,ltt_buf)+size)>ltt_buf->subbuf_size) {
+                               /* Event too big for subbuffers, report error, don't complete 
+                                * the sub-buffer switch. */
+                               atomic_inc(&ltt_buf->events_lost);
+                               return NULL;
+                       } else {
+                               /* We just made a successful buffer switch and the event fits in the
+                                * new subbuffer. Let's write. */
+                       }
+               } else {
+                       /* Event fits in the current buffer and we are not on a switch boundary.
+                        * It's safe to write */
+               }
+               offset_end = offset_begin + size;
+
+               if((SUBBUF_OFFSET(offset_end, ltt_buf)) == 0) {
+                       /* The offset_end will fall at the very beginning of the next subbuffer.
+                        */
+                       end_switch_current = 1; /* For offset_begin */
+               }
+
+       } while(atomic_cmpxchg(&ltt_buf->offset, offset_old, offset_end)
+                                                       != offset_old);
+
+
+       /* Push the reader if necessary */
+       do {
+               consumed_old = atomic_read(&ltt_buf->consumed);
+               /* If buffer is in overwrite mode, push the reader consumed count if
+                        the write position has reached it and we are not at the first
+                        iteration (don't push the reader farther than the writer). 
+                        This operation can be done concurrently by many writers in the
+                        same buffer, the writer being at the fartest write position sub-buffer
+                        index in the buffer being the one which will win this loop. */
+               /* If the buffer is not in overwrite mode, pushing the reader only
+                        happen if a sub-buffer is corrupted */
+               if((SUBBUF_TRUNC(offset_end, ltt_buf) 
+                                       - SUBBUF_TRUNC(consumed_old, ltt_buf)) 
+                                                       >= ltt_buf->alloc_size)
+                       consumed_new = SUBBUF_ALIGN(consumed_old, ltt_buf);
+               else {
+                       consumed_new = consumed_old;
+                       break;
+               }
+       } while(atomic_cmpxchg(&ltt_buf->consumed, consumed_old, consumed_new)
+                                               != consumed_old);
+
+       if(consumed_old != consumed_new) {
+               /* Reader pushed : we are the winner of the push, we can therefore
+                        reequilibrate reserve and commit. Atomic increment of the commit
+                        count permits other writers to play around with this variable
+                        before us. We keep track of corrupted_subbuffers even in overwrite mode :
+                        we never want to write over a non completely committed sub-buffer : 
+                        possible causes : the buffer size is too low compared to the unordered
+                        data input, or there is a writer who died between the reserve and the
+                        commit. */
+               if(reserve_commit_diff) {
+                       /* We have to alter the sub-buffer commit count : a sub-buffer is
+                                corrupted. We do not deliver it. */
+                       atomic_add(reserve_commit_diff,
+                                                               &ltt_buf->commit_count[SUBBUF_INDEX(offset_begin, ltt_buf)]);
+                       atomic_inc(&ltt_buf->corrupted_subbuffers);
+               }
+       }
+
+
+       if(end_switch_old) {
+               /* old subbuffer */
+               /* Concurrency safe because we are the last and only thread to alter this
+                        sub-buffer. As long as it is not delivered and read, no other thread can
+                        alter the offset, alter the reserve_count or call the
+                        client_buffer_end_callback on this sub-buffer.
+                        The only remaining threads could be the ones with pending commits. They
+                        will have to do the deliver themself.
+                        Not concurrency safe in overwrite mode. We detect corrupted subbuffers 
+                        with commit and reserve counts. We keep a corrupted sub-buffers count
+                        and push the readers across these sub-buffers.
+                        Not concurrency safe if a writer is stalled in a subbuffer and
+                        another writer switches in, finding out it's corrupted. The result will
+                        be than the old (uncommited) subbuffer will be declared corrupted, and
+                        that the new subbuffer will be declared corrupted too because of the
+                        commit count adjustment.
+                        Note : offset_old should never be 0 here.*/
+               ltt_buffer_end_callback(ltt_buf, *tsc, offset_old, 
+                                                                                                               SUBBUF_INDEX((offset_old-1), ltt_buf));
+               /* Setting this reserve_count will allow the sub-buffer to be delivered by
+                        the last committer. */
+               reserve_count = 
+                                                atomic_add_return((SUBBUF_OFFSET((offset_old-1), ltt_buf)+1),
+                                                &ltt_buf->reserve_count[SUBBUF_INDEX((offset_old-1), ltt_buf)]);
+               if(reserve_count 
+                                       == atomic_read(&ltt_buf->commit_count[SUBBUF_INDEX((offset_old-1),
+                                                                                                                                                                                                                                                       ltt_buf)])) {
+                       ltt_deliver_callback(ltt_buf, SUBBUF_INDEX((offset_old-1), ltt_buf),
+                                                                                                                NULL);
+               }
+       }
+
+       if(begin_switch) {
+               /* New sub-buffer */
+               /* This code can be executed unordered : writers may already have written
+                        to the sub-buffer before this code gets executed, caution. */
+               /* The commit makes sure that this code is executed before the deliver
+                        of this sub-buffer */
+               ltt_buffer_begin_callback(ltt_buf, *tsc, SUBBUF_INDEX(offset_begin, ltt_buf));
+               commit_count = atomic_add_return(ltt_subbuf_header_len(ltt_buf),
+                                                        &ltt_buf->commit_count[SUBBUF_INDEX(offset_begin, ltt_buf)]);
+               /* Check if the written buffer has to be delivered */
+               if(commit_count
+                                       == atomic_read(&ltt_buf->reserve_count[SUBBUF_INDEX(offset_begin,
+                                                                                                                                                                                                                                                       ltt_buf)])) {
+                       ltt_deliver_callback(ltt_buf, SUBBUF_INDEX(offset_begin, ltt_buf), NULL);
+               }
+       }
+
+       if(end_switch_current) {
+               /* current subbuffer */
+               /* Concurrency safe because we are the last and only thread to alter this
+                        sub-buffer. As long as it is not delivered and read, no other thread can
+                        alter the offset, alter the reserve_count or call the
+                        client_buffer_end_callback on this sub-buffer.
+                        The only remaining threads could be the ones with pending commits. They
+                        will have to do the deliver themself.
+                        Not concurrency safe in overwrite mode. We detect corrupted subbuffers 
+                        with commit and reserve counts. We keep a corrupted sub-buffers count
+                        and push the readers across these sub-buffers.
+                        Not concurrency safe if a writer is stalled in a subbuffer and
+                        another writer switches in, finding out it's corrupted. The result will
+                        be than the old (uncommited) subbuffer will be declared corrupted, and
+                        that the new subbuffer will be declared corrupted too because of the
+                        commit count adjustment. */
+               ltt_buffer_end_callback(ltt_buf, *tsc, offset_end,
+                                                                                                               SUBBUF_INDEX((offset_end-1), ltt_buf));
+               /* Setting this reserve_count will allow the sub-buffer to be delivered by
+                        the last committer. */
+               reserve_count = 
+      atomic_add_return((SUBBUF_OFFSET((offset_end-1), ltt_buf)+1),
+                       &ltt_buf->reserve_count[SUBBUF_INDEX((offset_end-1), ltt_buf)]);
+               if(reserve_count 
+                                       == atomic_read(&ltt_buf->commit_count[SUBBUF_INDEX((offset_end-1),
+                                                                                                                                                                                                                                                       ltt_buf)])) {
+                       ltt_deliver_callback(ltt_buf, SUBBUF_INDEX((offset_end-1), ltt_buf), NULL);
+               }
+       }
+
+       *slot_size = size;
+
+       //BUG_ON(*slot_size != (data_size + *before_hdr_pad + *after_hdr_pad + *header_size));
+       //BUG_ON(*slot_size != (offset_end - offset_begin));
+       
+       return ltt_buf->start + BUFFER_OFFSET(offset_begin, ltt_buf);
+}
+       
+       
+/* ltt_commit_slot
+ *
+ * Atomic unordered slot commit. Increments the commit count in the
+ * specified sub-buffer, and delivers it if necessary.
+ *
+ * Parameters:
+ *
+ * @buf : the buffer to commit to.
+ * @reserved : address of the beginnig of the reserved slot.
+ * @slot_size : size of the reserved slot.
+ *
+ */
+static inline void __attribute__((no_instrument_function)) ltt_commit_slot(
+                                                                                                                       struct ltt_buf *ltt_buf,
+                                                                                                                       void *reserved,
+                                                                                                                       unsigned int slot_size)
+{
+       unsigned int offset_begin = reserved - ltt_buf->start;
+       int commit_count;
+
+       commit_count = atomic_add_return(slot_size,
+                                                                                                       &ltt_buf->commit_count[SUBBUF_INDEX(offset_begin,
+                                                                                                                                                                                                                                                       ltt_buf)]);
+       
+       /* Check if all commits have been done */
+       if(commit_count ==
+       atomic_read(&ltt_buf->reserve_count[SUBBUF_INDEX(offset_begin, ltt_buf)])) {
+               ltt_deliver_callback(ltt_buf, SUBBUF_INDEX(offset_begin, ltt_buf), NULL);
+       }
+}
+       
+
+
+
+
 #endif //_LTT_USERTRACE_FAST_H
index ca528fbd06795ac7433d1685a5b196ee9f162c3d..8ac69eb5d7c1759763f3003a7f31eb2aaf083a8d 100644 (file)
@@ -16,7 +16,7 @@ void *thr1(void *arg)
 
   //while(1) {}
        for(i=0; i<5; i++) {
-               ltt_usertrace_fast_buffer_switch();
+       //      ltt_usertrace_fast_buffer_switch();
                sleep(1);
        }
 
@@ -33,7 +33,7 @@ void *thr2(void *arg)
     sleep(2);
   //}
        for(i=0; i<2; i++) {
-               ltt_usertrace_fast_buffer_switch();
+       //      ltt_usertrace_fast_buffer_switch();
                sleep(3);
        }
 
@@ -58,7 +58,7 @@ int main()
   if(err!=0) exit(1);
 
        for(i=0; i<2; i++) {
-               ltt_usertrace_fast_buffer_switch();
+       //      ltt_usertrace_fast_buffer_switch();
                sleep(3);
        }
 
This page took 0.03111 seconds and 4 git commands to generate.