ok, i experimented around with the latency tracer, trying to capture the 
trace of a full suspend+resume cycle, and it needed the tracer fix below 
(GTOD clocksource suspend/resume would otherwise confuse the tracer and 
you'd get no trace output as a result).

once that tracer bug was fixed, the best method to generate a trace was 
to do this:

   echo 1 > /proc/sys/kernel/stackframe_tracing
   echo 1 > /proc/sys/kernel/syscall_tracing
   ./trace-cmd bash -c "echo mem > /sys/power/state" > trace.txt

        Ingo

---
 kernel/latency_trace.c |   53 +++++++++++++++++++++++++++++++++----------------
 1 file changed, 36 insertions(+), 17 deletions(-)

Index: linux/kernel/latency_trace.c
===================================================================
--- linux.orig/kernel/latency_trace.c
+++ linux/kernel/latency_trace.c
@@ -67,13 +67,6 @@ static unsigned long notrace cycles_to_u
 }
 #endif
 
-static notrace inline cycle_t now(void)
-{
-       if (trace_use_raw_cycles)
-               return get_cycles();
-       return get_monotonic_cycles();
-}
-
 #ifndef irqs_off
 # define irqs_off()                    0
 #endif
@@ -237,6 +230,8 @@ struct cpu_trace {
        unsigned long critical_sequence;
        atomic_t underrun;
        atomic_t overrun;
+       cycle_t cycles;
+       cycle_t last_cycles;
        int early_warning;
        int latency_type;
        int cpu;
@@ -263,6 +258,30 @@ static struct cpu_trace cpu_traces[NR_CP
 #endif
  } };
 
+static notrace cycle_t now(struct cpu_trace *tr, int monotonic)
+{
+       cycles_t now, delta, last = tr->last_cycles;
+
+       if (trace_use_raw_cycles && !monotonic)
+               now = get_cycles();
+       else
+               now = get_monotonic_cycles();
+
+       /*
+        * Protect against time warps:
+        */
+       if (unlikely(now < last))
+               delta = 1;
+       else
+               delta = now - last;
+
+       tr->last_cycles = now;
+       tr->cycles += delta;
+
+       return tr->cycles;
+}
+
+
 #ifdef CONFIG_EVENT_TRACE
 
 int trace_enabled = 0;
@@ -615,7 +634,7 @@ ____trace(int cpu, enum trace_type type,
 again:
        idx = tr->trace_idx;
        idx_next = idx + 1;
-       timestamp = now();
+       timestamp = now(tr, 0);
 
        if (unlikely((trace_freerunning || print_functions || 
atomic_read(&tr->underrun)) &&
                     (idx_next >= MAX_TRACE) && !atomic_read(&tr->overrun))) {
@@ -1743,7 +1762,7 @@ check_critical_timing(int cpu, struct cp
         * as long as possible:
         */
        T0 = tr->preempt_timestamp;
-       T1 = get_monotonic_cycles();
+       T1 = now(tr, 1);
        delta = T1-T0;
 
        local_save_flags(flags);
@@ -1757,7 +1776,7 @@ check_critical_timing(int cpu, struct cp
         * might change it (it can only get larger so the latency
         * is fair to be reported):
         */
-       T2 = get_monotonic_cycles();
+       T2 = now(tr, 1);
 
        delta = T2-T0;
 
@@ -1807,7 +1826,7 @@ check_critical_timing(int cpu, struct cp
        printk(" =>   ended at timestamp %lu: ", t1);
        print_symbol("<%s>\n", tr->critical_end);
        dump_stack();
-       t1 = cycles_to_usecs(get_monotonic_cycles());
+       t1 = cycles_to_usecs(now(tr, 1));
        printk(" =>   dump-end timestamp %lu\n\n", t1);
 #endif
 
@@ -1817,7 +1836,7 @@ check_critical_timing(int cpu, struct cp
 
 out:
        tr->critical_sequence = max_sequence;
-       tr->preempt_timestamp = get_monotonic_cycles();
+       tr->preempt_timestamp = now(tr, 1);
        tr->early_warning = 0;
        reset_trace_idx(cpu, tr);
        _trace_cmdline(cpu, tr);
@@ -1866,7 +1885,7 @@ __start_critical_timing(unsigned long ei
        atomic_inc(&tr->disabled);
 
        tr->critical_sequence = max_sequence;
-       tr->preempt_timestamp = get_monotonic_cycles();
+       tr->preempt_timestamp = now(tr, 1);
        tr->critical_start = eip;
        reset_trace_idx(cpu, tr);
        tr->latency_type = latency_type;
@@ -2150,7 +2169,7 @@ check_wakeup_timing(struct cpu_trace *tr
                goto out;
 
        T0 = tr->preempt_timestamp;
-       T1 = get_monotonic_cycles();
+       T1 = now(tr, 1);
        /*
         * Any wraparound or time warp and we are out:
         */
@@ -2255,7 +2274,7 @@ void __trace_start_sched_wakeup(struct t
 //     if (!atomic_read(&tr->disabled)) {
                atomic_inc(&tr->disabled);
                tr->critical_sequence = max_sequence;
-               tr->preempt_timestamp = get_monotonic_cycles();
+               tr->preempt_timestamp = now(tr, 1);
                tr->latency_type = WAKEUP_LATENCY;
                tr->critical_start = CALLER_ADDR0;
                _trace_cmdline(raw_smp_processor_id(), tr);
@@ -2366,7 +2385,7 @@ long user_trace_start(void)
        reset_trace_idx(cpu, tr);
 
        tr->critical_sequence = max_sequence;
-       tr->preempt_timestamp = get_monotonic_cycles();
+       tr->preempt_timestamp = now(tr, 1);
        tr->critical_start = CALLER_ADDR0;
        _trace_cmdline(cpu, tr);
        mcount();
@@ -2425,7 +2444,7 @@ long user_trace_stop(void)
                unsigned long long tmp0;
 
                T0 = tr->preempt_timestamp;
-               T1 = get_monotonic_cycles();
+               T1 = now(tr, 1);
                tmp0 = preempt_max_latency;
                if (T1 < T0)
                        T0 = T1;
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to