mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 10:40:15 +02:00 
			
		
		
		
	The Tracing > Histogram page contains a typo in the field display modifiers table. Signed-off-by: Gabriele Monaco <gmonaco@redhat.com> Acked-by: Steven Rostedt (Google) <rostedt@goodmis.org> Signed-off-by: Jonathan Corbet <corbet@lwn.net> Link: https://lore.kernel.org/r/20241003122334.44682-2-gmonaco@redhat.com
		
			
				
	
	
		
			3071 lines
		
	
	
	
		
			160 KiB
		
	
	
	
		
			ReStructuredText
		
	
	
	
	
	
			
		
		
	
	
			3071 lines
		
	
	
	
		
			160 KiB
		
	
	
	
		
			ReStructuredText
		
	
	
	
	
	
================
 | 
						|
Event Histograms
 | 
						|
================
 | 
						|
 | 
						|
Documentation written by Tom Zanussi
 | 
						|
 | 
						|
1. Introduction
 | 
						|
===============
 | 
						|
 | 
						|
  Histogram triggers are special event triggers that can be used to
 | 
						|
  aggregate trace event data into histograms.  For information on
 | 
						|
  trace events and event triggers, see Documentation/trace/events.rst.
 | 
						|
 | 
						|
 | 
						|
2. Histogram Trigger Command
 | 
						|
============================
 | 
						|
 | 
						|
  A histogram trigger command is an event trigger command that
 | 
						|
  aggregates event hits into a hash table keyed on one or more trace
 | 
						|
  event format fields (or stacktrace) and a set of running totals
 | 
						|
  derived from one or more trace event format fields and/or event
 | 
						|
  counts (hitcount).
 | 
						|
 | 
						|
  The format of a hist trigger is as follows::
 | 
						|
 | 
						|
        hist:keys=<field1[,field2,...]>[:values=<field1[,field2,...]>]
 | 
						|
          [:sort=<field1[,field2,...]>][:size=#entries][:pause][:continue]
 | 
						|
          [:clear][:name=histname1][:nohitcount][:<handler>.<action>] [if <filter>]
 | 
						|
 | 
						|
  When a matching event is hit, an entry is added to a hash table
 | 
						|
  using the key(s) and value(s) named.  Keys and values correspond to
 | 
						|
  fields in the event's format description.  Values must correspond to
 | 
						|
  numeric fields - on an event hit, the value(s) will be added to a
 | 
						|
  sum kept for that field.  The special string 'hitcount' can be used
 | 
						|
  in place of an explicit value field - this is simply a count of
 | 
						|
  event hits.  If 'values' isn't specified, an implicit 'hitcount'
 | 
						|
  value will be automatically created and used as the only value.
 | 
						|
  Keys can be any field, or the special string 'common_stacktrace', which
 | 
						|
  will use the event's kernel stacktrace as the key.  The keywords
 | 
						|
  'keys' or 'key' can be used to specify keys, and the keywords
 | 
						|
  'values', 'vals', or 'val' can be used to specify values.  Compound
 | 
						|
  keys consisting of up to three fields can be specified by the 'keys'
 | 
						|
  keyword.  Hashing a compound key produces a unique entry in the
 | 
						|
  table for each unique combination of component keys, and can be
 | 
						|
  useful for providing more fine-grained summaries of event data.
 | 
						|
  Additionally, sort keys consisting of up to two fields can be
 | 
						|
  specified by the 'sort' keyword.  If more than one field is
 | 
						|
  specified, the result will be a 'sort within a sort': the first key
 | 
						|
  is taken to be the primary sort key and the second the secondary
 | 
						|
  key.  If a hist trigger is given a name using the 'name' parameter,
 | 
						|
  its histogram data will be shared with other triggers of the same
 | 
						|
  name, and trigger hits will update this common data.  Only triggers
 | 
						|
  with 'compatible' fields can be combined in this way; triggers are
 | 
						|
  'compatible' if the fields named in the trigger share the same
 | 
						|
  number and type of fields and those fields also have the same names.
 | 
						|
  Note that any two events always share the compatible 'hitcount' and
 | 
						|
  'common_stacktrace' fields and can therefore be combined using those
 | 
						|
  fields, however pointless that may be.
 | 
						|
 | 
						|
  'hist' triggers add a 'hist' file to each event's subdirectory.
 | 
						|
  Reading the 'hist' file for the event will dump the hash table in
 | 
						|
  its entirety to stdout.  If there are multiple hist triggers
 | 
						|
  attached to an event, there will be a table for each trigger in the
 | 
						|
  output.  The table displayed for a named trigger will be the same as
 | 
						|
  any other instance having the same name. Each printed hash table
 | 
						|
  entry is a simple list of the keys and values comprising the entry;
 | 
						|
  keys are printed first and are delineated by curly braces, and are
 | 
						|
  followed by the set of value fields for the entry.  By default,
 | 
						|
  numeric fields are displayed as base-10 integers.  This can be
 | 
						|
  modified by appending any of the following modifiers to the field
 | 
						|
  name:
 | 
						|
 | 
						|
	=============  =================================================
 | 
						|
        .hex           display a number as a hex value
 | 
						|
	.sym           display an address as a symbol
 | 
						|
	.sym-offset    display an address as a symbol and offset
 | 
						|
	.syscall       display a syscall id as a system call name
 | 
						|
	.execname      display a common_pid as a program name
 | 
						|
	.log2          display log2 value rather than raw number
 | 
						|
	.buckets=size  display grouping of values rather than raw number
 | 
						|
	.usecs         display a common_timestamp in microseconds
 | 
						|
        .percent       display a number of percentage value
 | 
						|
        .graph         display a bar-graph of a value
 | 
						|
	.stacktrace    display as a stacktrace (must be a long[] type)
 | 
						|
	=============  =================================================
 | 
						|
 | 
						|
  Note that in general the semantics of a given field aren't
 | 
						|
  interpreted when applying a modifier to it, but there are some
 | 
						|
  restrictions to be aware of in this regard:
 | 
						|
 | 
						|
    - only the 'hex' modifier can be used for values (because values
 | 
						|
      are essentially sums, and the other modifiers don't make sense
 | 
						|
      in that context).
 | 
						|
    - the 'execname' modifier can only be used on a 'common_pid'.  The
 | 
						|
      reason for this is that the execname is simply the 'comm' value
 | 
						|
      saved for the 'current' process when an event was triggered,
 | 
						|
      which is the same as the common_pid value saved by the event
 | 
						|
      tracing code.  Trying to apply that comm value to other pid
 | 
						|
      values wouldn't be correct, and typically events that care save
 | 
						|
      pid-specific comm fields in the event itself.
 | 
						|
 | 
						|
  A typical usage scenario would be the following to enable a hist
 | 
						|
  trigger, read its current contents, and then turn it off::
 | 
						|
 | 
						|
    # echo 'hist:keys=skbaddr.hex:vals=len' > \
 | 
						|
      /sys/kernel/tracing/events/net/netif_rx/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/net/netif_rx/hist
 | 
						|
 | 
						|
    # echo '!hist:keys=skbaddr.hex:vals=len' > \
 | 
						|
      /sys/kernel/tracing/events/net/netif_rx/trigger
 | 
						|
 | 
						|
  The trigger file itself can be read to show the details of the
 | 
						|
  currently attached hist trigger.  This information is also displayed
 | 
						|
  at the top of the 'hist' file when read.
 | 
						|
 | 
						|
  By default, the size of the hash table is 2048 entries.  The 'size'
 | 
						|
  parameter can be used to specify more or fewer than that.  The units
 | 
						|
  are in terms of hashtable entries - if a run uses more entries than
 | 
						|
  specified, the results will show the number of 'drops', the number
 | 
						|
  of hits that were ignored.  The size should be a power of 2 between
 | 
						|
  128 and 131072 (any non- power-of-2 number specified will be rounded
 | 
						|
  up).
 | 
						|
 | 
						|
  The 'sort' parameter can be used to specify a value field to sort
 | 
						|
  on.  The default if unspecified is 'hitcount' and the default sort
 | 
						|
  order is 'ascending'.  To sort in the opposite direction, append
 | 
						|
  .descending' to the sort key.
 | 
						|
 | 
						|
  The 'pause' parameter can be used to pause an existing hist trigger
 | 
						|
  or to start a hist trigger but not log any events until told to do
 | 
						|
  so.  'continue' or 'cont' can be used to start or restart a paused
 | 
						|
  hist trigger.
 | 
						|
 | 
						|
  The 'clear' parameter will clear the contents of a running hist
 | 
						|
  trigger and leave its current paused/active state.
 | 
						|
 | 
						|
  Note that the 'pause', 'cont', and 'clear' parameters should be
 | 
						|
  applied using 'append' shell operator ('>>') if applied to an
 | 
						|
  existing trigger, rather than via the '>' operator, which will cause
 | 
						|
  the trigger to be removed through truncation.
 | 
						|
 | 
						|
  The 'nohitcount' (or NOHC) parameter will suppress display of
 | 
						|
  raw hitcount in the histogram. This option requires at least one
 | 
						|
  value field which is not a 'raw hitcount'. For example,
 | 
						|
  'hist:...:vals=hitcount:nohitcount' is rejected, but
 | 
						|
  'hist:...:vals=hitcount.percent:nohitcount' is OK.
 | 
						|
 | 
						|
- enable_hist/disable_hist
 | 
						|
 | 
						|
  The enable_hist and disable_hist triggers can be used to have one
 | 
						|
  event conditionally start and stop another event's already-attached
 | 
						|
  hist trigger.  Any number of enable_hist and disable_hist triggers
 | 
						|
  can be attached to a given event, allowing that event to kick off
 | 
						|
  and stop aggregations on a host of other events.
 | 
						|
 | 
						|
  The format is very similar to the enable/disable_event triggers::
 | 
						|
 | 
						|
      enable_hist:<system>:<event>[:count]
 | 
						|
      disable_hist:<system>:<event>[:count]
 | 
						|
 | 
						|
  Instead of enabling or disabling the tracing of the target event
 | 
						|
  into the trace buffer as the enable/disable_event triggers do, the
 | 
						|
  enable/disable_hist triggers enable or disable the aggregation of
 | 
						|
  the target event into a hash table.
 | 
						|
 | 
						|
  A typical usage scenario for the enable_hist/disable_hist triggers
 | 
						|
  would be to first set up a paused hist trigger on some event,
 | 
						|
  followed by an enable_hist/disable_hist pair that turns the hist
 | 
						|
  aggregation on and off when conditions of interest are hit::
 | 
						|
 | 
						|
   # echo 'hist:keys=skbaddr.hex:vals=len:pause' > \
 | 
						|
      /sys/kernel/tracing/events/net/netif_receive_skb/trigger
 | 
						|
 | 
						|
    # echo 'enable_hist:net:netif_receive_skb if filename==/usr/bin/wget' > \
 | 
						|
      /sys/kernel/tracing/events/sched/sched_process_exec/trigger
 | 
						|
 | 
						|
    # echo 'disable_hist:net:netif_receive_skb if comm==wget' > \
 | 
						|
      /sys/kernel/tracing/events/sched/sched_process_exit/trigger
 | 
						|
 | 
						|
  The above sets up an initially paused hist trigger which is unpaused
 | 
						|
  and starts aggregating events when a given program is executed, and
 | 
						|
  which stops aggregating when the process exits and the hist trigger
 | 
						|
  is paused again.
 | 
						|
 | 
						|
  The examples below provide a more concrete illustration of the
 | 
						|
  concepts and typical usage patterns discussed above.
 | 
						|
 | 
						|
'special' event fields
 | 
						|
------------------------
 | 
						|
 | 
						|
  There are a number of 'special event fields' available for use as
 | 
						|
  keys or values in a hist trigger.  These look like and behave as if
 | 
						|
  they were actual event fields, but aren't really part of the event's
 | 
						|
  field definition or format file.  They are however available for any
 | 
						|
  event, and can be used anywhere an actual event field could be.
 | 
						|
  They are:
 | 
						|
 | 
						|
    ====================== ==== =======================================
 | 
						|
    common_timestamp       u64  timestamp (from ring buffer) associated
 | 
						|
                                with the event, in nanoseconds.  May be
 | 
						|
			        modified by .usecs to have timestamps
 | 
						|
			        interpreted as microseconds.
 | 
						|
    common_cpu             int  the cpu on which the event occurred.
 | 
						|
    ====================== ==== =======================================
 | 
						|
 | 
						|
Extended error information
 | 
						|
--------------------------
 | 
						|
 | 
						|
  For some error conditions encountered when invoking a hist trigger
 | 
						|
  command, extended error information is available via the
 | 
						|
  tracing/error_log file.  See Error Conditions in
 | 
						|
  :file:`Documentation/trace/ftrace.rst` for details.
 | 
						|
 | 
						|
6.2 'hist' trigger examples
 | 
						|
---------------------------
 | 
						|
 | 
						|
  The first set of examples creates aggregations using the kmalloc
 | 
						|
  event.  The fields that can be used for the hist trigger are listed
 | 
						|
  in the kmalloc event's format file::
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/kmem/kmalloc/format
 | 
						|
    name: kmalloc
 | 
						|
    ID: 374
 | 
						|
    format:
 | 
						|
	field:unsigned short common_type;	offset:0;	size:2;	signed:0;
 | 
						|
	field:unsigned char common_flags;	offset:2;	size:1;	signed:0;
 | 
						|
	field:unsigned char common_preempt_count;		offset:3;	size:1;	signed:0;
 | 
						|
	field:int common_pid;					offset:4;	size:4;	signed:1;
 | 
						|
 | 
						|
	field:unsigned long call_site;				offset:8;	size:8;	signed:0;
 | 
						|
	field:const void * ptr;					offset:16;	size:8;	signed:0;
 | 
						|
	field:size_t bytes_req;					offset:24;	size:8;	signed:0;
 | 
						|
	field:size_t bytes_alloc;				offset:32;	size:8;	signed:0;
 | 
						|
	field:gfp_t gfp_flags;					offset:40;	size:4;	signed:0;
 | 
						|
 | 
						|
  We'll start by creating a hist trigger that generates a simple table
 | 
						|
  that lists the total number of bytes requested for each function in
 | 
						|
  the kernel that made one or more calls to kmalloc::
 | 
						|
 | 
						|
    # echo 'hist:key=call_site:val=bytes_req.buckets=32' > \
 | 
						|
            /sys/kernel/tracing/events/kmem/kmalloc/trigger
 | 
						|
 | 
						|
  This tells the tracing system to create a 'hist' trigger using the
 | 
						|
  call_site field of the kmalloc event as the key for the table, which
 | 
						|
  just means that each unique call_site address will have an entry
 | 
						|
  created for it in the table.  The 'val=bytes_req' parameter tells
 | 
						|
  the hist trigger that for each unique entry (call_site) in the
 | 
						|
  table, it should keep a running total of the number of bytes
 | 
						|
  requested by that call_site.
 | 
						|
 | 
						|
  We'll let it run for awhile and then dump the contents of the 'hist'
 | 
						|
  file in the kmalloc event's subdirectory (for readability, a number
 | 
						|
  of entries have been omitted)::
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/kmem/kmalloc/hist
 | 
						|
    # trigger info: hist:keys=call_site:vals=bytes_req:sort=hitcount:size=2048 [active]
 | 
						|
 | 
						|
    { call_site: 18446744072106379007 } hitcount:          1  bytes_req:        176
 | 
						|
    { call_site: 18446744071579557049 } hitcount:          1  bytes_req:       1024
 | 
						|
    { call_site: 18446744071580608289 } hitcount:          1  bytes_req:      16384
 | 
						|
    { call_site: 18446744071581827654 } hitcount:          1  bytes_req:         24
 | 
						|
    { call_site: 18446744071580700980 } hitcount:          1  bytes_req:          8
 | 
						|
    { call_site: 18446744071579359876 } hitcount:          1  bytes_req:        152
 | 
						|
    { call_site: 18446744071580795365 } hitcount:          3  bytes_req:        144
 | 
						|
    { call_site: 18446744071581303129 } hitcount:          3  bytes_req:        144
 | 
						|
    { call_site: 18446744071580713234 } hitcount:          4  bytes_req:       2560
 | 
						|
    { call_site: 18446744071580933750 } hitcount:          4  bytes_req:        736
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { call_site: 18446744072106047046 } hitcount:         69  bytes_req:       5576
 | 
						|
    { call_site: 18446744071582116407 } hitcount:         73  bytes_req:       2336
 | 
						|
    { call_site: 18446744072106054684 } hitcount:        136  bytes_req:     140504
 | 
						|
    { call_site: 18446744072106224230 } hitcount:        136  bytes_req:      19584
 | 
						|
    { call_site: 18446744072106078074 } hitcount:        153  bytes_req:       2448
 | 
						|
    { call_site: 18446744072106062406 } hitcount:        153  bytes_req:      36720
 | 
						|
    { call_site: 18446744071582507929 } hitcount:        153  bytes_req:      37088
 | 
						|
    { call_site: 18446744072102520590 } hitcount:        273  bytes_req:      10920
 | 
						|
    { call_site: 18446744071582143559 } hitcount:        358  bytes_req:        716
 | 
						|
    { call_site: 18446744072106465852 } hitcount:        417  bytes_req:      56712
 | 
						|
    { call_site: 18446744072102523378 } hitcount:        485  bytes_req:      27160
 | 
						|
    { call_site: 18446744072099568646 } hitcount:       1676  bytes_req:      33520
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 4610
 | 
						|
        Entries: 45
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  The output displays a line for each entry, beginning with the key
 | 
						|
  specified in the trigger, followed by the value(s) also specified in
 | 
						|
  the trigger.  At the beginning of the output is a line that displays
 | 
						|
  the trigger info, which can also be displayed by reading the
 | 
						|
  'trigger' file::
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/kmem/kmalloc/trigger
 | 
						|
    hist:keys=call_site:vals=bytes_req:sort=hitcount:size=2048 [active]
 | 
						|
 | 
						|
  At the end of the output are a few lines that display the overall
 | 
						|
  totals for the run.  The 'Hits' field shows the total number of
 | 
						|
  times the event trigger was hit, the 'Entries' field shows the total
 | 
						|
  number of used entries in the hash table, and the 'Dropped' field
 | 
						|
  shows the number of hits that were dropped because the number of
 | 
						|
  used entries for the run exceeded the maximum number of entries
 | 
						|
  allowed for the table (normally 0, but if not a hint that you may
 | 
						|
  want to increase the size of the table using the 'size' parameter).
 | 
						|
 | 
						|
  Notice in the above output that there's an extra field, 'hitcount',
 | 
						|
  which wasn't specified in the trigger.  Also notice that in the
 | 
						|
  trigger info output, there's a parameter, 'sort=hitcount', which
 | 
						|
  wasn't specified in the trigger either.  The reason for that is that
 | 
						|
  every trigger implicitly keeps a count of the total number of hits
 | 
						|
  attributed to a given entry, called the 'hitcount'.  That hitcount
 | 
						|
  information is explicitly displayed in the output, and in the
 | 
						|
  absence of a user-specified sort parameter, is used as the default
 | 
						|
  sort field.
 | 
						|
 | 
						|
  The value 'hitcount' can be used in place of an explicit value in
 | 
						|
  the 'values' parameter if you don't really need to have any
 | 
						|
  particular field summed and are mainly interested in hit
 | 
						|
  frequencies.
 | 
						|
 | 
						|
  To turn the hist trigger off, simply call up the trigger in the
 | 
						|
  command history and re-execute it with a '!' prepended::
 | 
						|
 | 
						|
    # echo '!hist:key=call_site:val=bytes_req' > \
 | 
						|
           /sys/kernel/tracing/events/kmem/kmalloc/trigger
 | 
						|
 | 
						|
  Finally, notice that the call_site as displayed in the output above
 | 
						|
  isn't really very useful.  It's an address, but normally addresses
 | 
						|
  are displayed in hex.  To have a numeric field displayed as a hex
 | 
						|
  value, simply append '.hex' to the field name in the trigger::
 | 
						|
 | 
						|
    # echo 'hist:key=call_site.hex:val=bytes_req' > \
 | 
						|
           /sys/kernel/tracing/events/kmem/kmalloc/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/kmem/kmalloc/hist
 | 
						|
    # trigger info: hist:keys=call_site.hex:vals=bytes_req:sort=hitcount:size=2048 [active]
 | 
						|
 | 
						|
    { call_site: ffffffffa026b291 } hitcount:          1  bytes_req:        433
 | 
						|
    { call_site: ffffffffa07186ff } hitcount:          1  bytes_req:        176
 | 
						|
    { call_site: ffffffff811ae721 } hitcount:          1  bytes_req:      16384
 | 
						|
    { call_site: ffffffff811c5134 } hitcount:          1  bytes_req:          8
 | 
						|
    { call_site: ffffffffa04a9ebb } hitcount:          1  bytes_req:        511
 | 
						|
    { call_site: ffffffff8122e0a6 } hitcount:          1  bytes_req:         12
 | 
						|
    { call_site: ffffffff8107da84 } hitcount:          1  bytes_req:        152
 | 
						|
    { call_site: ffffffff812d8246 } hitcount:          1  bytes_req:         24
 | 
						|
    { call_site: ffffffff811dc1e5 } hitcount:          3  bytes_req:        144
 | 
						|
    { call_site: ffffffffa02515e8 } hitcount:          3  bytes_req:        648
 | 
						|
    { call_site: ffffffff81258159 } hitcount:          3  bytes_req:        144
 | 
						|
    { call_site: ffffffff811c80f4 } hitcount:          4  bytes_req:        544
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { call_site: ffffffffa06c7646 } hitcount:        106  bytes_req:       8024
 | 
						|
    { call_site: ffffffffa06cb246 } hitcount:        132  bytes_req:      31680
 | 
						|
    { call_site: ffffffffa06cef7a } hitcount:        132  bytes_req:       2112
 | 
						|
    { call_site: ffffffff8137e399 } hitcount:        132  bytes_req:      23232
 | 
						|
    { call_site: ffffffffa06c941c } hitcount:        185  bytes_req:     171360
 | 
						|
    { call_site: ffffffffa06f2a66 } hitcount:        185  bytes_req:      26640
 | 
						|
    { call_site: ffffffffa036a70e } hitcount:        265  bytes_req:      10600
 | 
						|
    { call_site: ffffffff81325447 } hitcount:        292  bytes_req:        584
 | 
						|
    { call_site: ffffffffa072da3c } hitcount:        446  bytes_req:      60656
 | 
						|
    { call_site: ffffffffa036b1f2 } hitcount:        526  bytes_req:      29456
 | 
						|
    { call_site: ffffffffa0099c06 } hitcount:       1780  bytes_req:      35600
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 4775
 | 
						|
        Entries: 46
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  Even that's only marginally more useful - while hex values do look
 | 
						|
  more like addresses, what users are typically more interested in
 | 
						|
  when looking at text addresses are the corresponding symbols
 | 
						|
  instead.  To have an address displayed as symbolic value instead,
 | 
						|
  simply append '.sym' or '.sym-offset' to the field name in the
 | 
						|
  trigger::
 | 
						|
 | 
						|
    # echo 'hist:key=call_site.sym:val=bytes_req' > \
 | 
						|
           /sys/kernel/tracing/events/kmem/kmalloc/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/kmem/kmalloc/hist
 | 
						|
    # trigger info: hist:keys=call_site.sym:vals=bytes_req:sort=hitcount:size=2048 [active]
 | 
						|
 | 
						|
    { call_site: [ffffffff810adcb9] syslog_print_all                              } hitcount:          1  bytes_req:       1024
 | 
						|
    { call_site: [ffffffff8154bc62] usb_control_msg                               } hitcount:          1  bytes_req:          8
 | 
						|
    { call_site: [ffffffffa00bf6fe] hidraw_send_report [hid]                      } hitcount:          1  bytes_req:          7
 | 
						|
    { call_site: [ffffffff8154acbe] usb_alloc_urb                                 } hitcount:          1  bytes_req:        192
 | 
						|
    { call_site: [ffffffffa00bf1ca] hidraw_report_event [hid]                     } hitcount:          1  bytes_req:          7
 | 
						|
    { call_site: [ffffffff811e3a25] __seq_open_private                            } hitcount:          1  bytes_req:         40
 | 
						|
    { call_site: [ffffffff8109524a] alloc_fair_sched_group                        } hitcount:          2  bytes_req:        128
 | 
						|
    { call_site: [ffffffff811febd5] fsnotify_alloc_group                          } hitcount:          2  bytes_req:        528
 | 
						|
    { call_site: [ffffffff81440f58] __tty_buffer_request_room                     } hitcount:          2  bytes_req:       2624
 | 
						|
    { call_site: [ffffffff81200ba6] inotify_new_group                             } hitcount:          2  bytes_req:         96
 | 
						|
    { call_site: [ffffffffa05e19af] ieee80211_start_tx_ba_session [mac80211]      } hitcount:          2  bytes_req:        464
 | 
						|
    { call_site: [ffffffff81672406] tcp_get_metrics                               } hitcount:          2  bytes_req:        304
 | 
						|
    { call_site: [ffffffff81097ec2] alloc_rt_sched_group                          } hitcount:          2  bytes_req:        128
 | 
						|
    { call_site: [ffffffff81089b05] sched_create_group                            } hitcount:          2  bytes_req:       1424
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { call_site: [ffffffffa04a580c] intel_crtc_page_flip [i915]                   } hitcount:       1185  bytes_req:     123240
 | 
						|
    { call_site: [ffffffffa0287592] drm_mode_page_flip_ioctl [drm]                } hitcount:       1185  bytes_req:     104280
 | 
						|
    { call_site: [ffffffffa04c4a3c] intel_plane_duplicate_state [i915]            } hitcount:       1402  bytes_req:     190672
 | 
						|
    { call_site: [ffffffff812891ca] ext4_find_extent                              } hitcount:       1518  bytes_req:     146208
 | 
						|
    { call_site: [ffffffffa029070e] drm_vma_node_allow [drm]                      } hitcount:       1746  bytes_req:      69840
 | 
						|
    { call_site: [ffffffffa045e7c4] i915_gem_do_execbuffer.isra.23 [i915]         } hitcount:       2021  bytes_req:     792312
 | 
						|
    { call_site: [ffffffffa02911f2] drm_modeset_lock_crtc [drm]                   } hitcount:       2592  bytes_req:     145152
 | 
						|
    { call_site: [ffffffffa0489a66] intel_ring_begin [i915]                       } hitcount:       2629  bytes_req:     378576
 | 
						|
    { call_site: [ffffffffa046041c] i915_gem_execbuffer2 [i915]                   } hitcount:       2629  bytes_req:    3783248
 | 
						|
    { call_site: [ffffffff81325607] apparmor_file_alloc_security                  } hitcount:       5192  bytes_req:      10384
 | 
						|
    { call_site: [ffffffffa00b7c06] hid_report_raw_event [hid]                    } hitcount:       5529  bytes_req:     110584
 | 
						|
    { call_site: [ffffffff8131ebf7] aa_alloc_task_context                         } hitcount:      21943  bytes_req:     702176
 | 
						|
    { call_site: [ffffffff8125847d] ext4_htree_store_dirent                       } hitcount:      55759  bytes_req:    5074265
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 109928
 | 
						|
        Entries: 71
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  Because the default sort key above is 'hitcount', the above shows a
 | 
						|
  the list of call_sites by increasing hitcount, so that at the bottom
 | 
						|
  we see the functions that made the most kmalloc calls during the
 | 
						|
  run.  If instead we wanted to see the top kmalloc callers in
 | 
						|
  terms of the number of bytes requested rather than the number of
 | 
						|
  calls, and we wanted the top caller to appear at the top, we can use
 | 
						|
  the 'sort' parameter, along with the 'descending' modifier::
 | 
						|
 | 
						|
    # echo 'hist:key=call_site.sym:val=bytes_req:sort=bytes_req.descending' > \
 | 
						|
           /sys/kernel/tracing/events/kmem/kmalloc/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/kmem/kmalloc/hist
 | 
						|
    # trigger info: hist:keys=call_site.sym:vals=bytes_req:sort=bytes_req.descending:size=2048 [active]
 | 
						|
 | 
						|
    { call_site: [ffffffffa046041c] i915_gem_execbuffer2 [i915]                   } hitcount:       2186  bytes_req:    3397464
 | 
						|
    { call_site: [ffffffffa045e7c4] i915_gem_do_execbuffer.isra.23 [i915]         } hitcount:       1790  bytes_req:     712176
 | 
						|
    { call_site: [ffffffff8125847d] ext4_htree_store_dirent                       } hitcount:       8132  bytes_req:     513135
 | 
						|
    { call_site: [ffffffff811e2a1b] seq_buf_alloc                                 } hitcount:        106  bytes_req:     440128
 | 
						|
    { call_site: [ffffffffa0489a66] intel_ring_begin [i915]                       } hitcount:       2186  bytes_req:     314784
 | 
						|
    { call_site: [ffffffff812891ca] ext4_find_extent                              } hitcount:       2174  bytes_req:     208992
 | 
						|
    { call_site: [ffffffff811ae8e1] __kmalloc                                     } hitcount:          8  bytes_req:     131072
 | 
						|
    { call_site: [ffffffffa04c4a3c] intel_plane_duplicate_state [i915]            } hitcount:        859  bytes_req:     116824
 | 
						|
    { call_site: [ffffffffa02911f2] drm_modeset_lock_crtc [drm]                   } hitcount:       1834  bytes_req:     102704
 | 
						|
    { call_site: [ffffffffa04a580c] intel_crtc_page_flip [i915]                   } hitcount:        972  bytes_req:     101088
 | 
						|
    { call_site: [ffffffffa0287592] drm_mode_page_flip_ioctl [drm]                } hitcount:        972  bytes_req:      85536
 | 
						|
    { call_site: [ffffffffa00b7c06] hid_report_raw_event [hid]                    } hitcount:       3333  bytes_req:      66664
 | 
						|
    { call_site: [ffffffff8137e559] sg_kmalloc                                    } hitcount:        209  bytes_req:      61632
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { call_site: [ffffffff81095225] alloc_fair_sched_group                        } hitcount:          2  bytes_req:        128
 | 
						|
    { call_site: [ffffffff81097ec2] alloc_rt_sched_group                          } hitcount:          2  bytes_req:        128
 | 
						|
    { call_site: [ffffffff812d8406] copy_semundo                                  } hitcount:          2  bytes_req:         48
 | 
						|
    { call_site: [ffffffff81200ba6] inotify_new_group                             } hitcount:          1  bytes_req:         48
 | 
						|
    { call_site: [ffffffffa027121a] drm_getmagic [drm]                            } hitcount:          1  bytes_req:         48
 | 
						|
    { call_site: [ffffffff811e3a25] __seq_open_private                            } hitcount:          1  bytes_req:         40
 | 
						|
    { call_site: [ffffffff811c52f4] bprm_change_interp                            } hitcount:          2  bytes_req:         16
 | 
						|
    { call_site: [ffffffff8154bc62] usb_control_msg                               } hitcount:          1  bytes_req:          8
 | 
						|
    { call_site: [ffffffffa00bf1ca] hidraw_report_event [hid]                     } hitcount:          1  bytes_req:          7
 | 
						|
    { call_site: [ffffffffa00bf6fe] hidraw_send_report [hid]                      } hitcount:          1  bytes_req:          7
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 32133
 | 
						|
        Entries: 81
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  To display the offset and size information in addition to the symbol
 | 
						|
  name, just use 'sym-offset' instead::
 | 
						|
 | 
						|
    # echo 'hist:key=call_site.sym-offset:val=bytes_req:sort=bytes_req.descending' > \
 | 
						|
           /sys/kernel/tracing/events/kmem/kmalloc/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/kmem/kmalloc/hist
 | 
						|
    # trigger info: hist:keys=call_site.sym-offset:vals=bytes_req:sort=bytes_req.descending:size=2048 [active]
 | 
						|
 | 
						|
    { call_site: [ffffffffa046041c] i915_gem_execbuffer2+0x6c/0x2c0 [i915]                  } hitcount:       4569  bytes_req:    3163720
 | 
						|
    { call_site: [ffffffffa0489a66] intel_ring_begin+0xc6/0x1f0 [i915]                      } hitcount:       4569  bytes_req:     657936
 | 
						|
    { call_site: [ffffffffa045e7c4] i915_gem_do_execbuffer.isra.23+0x694/0x1020 [i915]      } hitcount:       1519  bytes_req:     472936
 | 
						|
    { call_site: [ffffffffa045e646] i915_gem_do_execbuffer.isra.23+0x516/0x1020 [i915]      } hitcount:       3050  bytes_req:     211832
 | 
						|
    { call_site: [ffffffff811e2a1b] seq_buf_alloc+0x1b/0x50                                 } hitcount:         34  bytes_req:     148384
 | 
						|
    { call_site: [ffffffffa04a580c] intel_crtc_page_flip+0xbc/0x870 [i915]                  } hitcount:       1385  bytes_req:     144040
 | 
						|
    { call_site: [ffffffff811ae8e1] __kmalloc+0x191/0x1b0                                   } hitcount:          8  bytes_req:     131072
 | 
						|
    { call_site: [ffffffffa0287592] drm_mode_page_flip_ioctl+0x282/0x360 [drm]              } hitcount:       1385  bytes_req:     121880
 | 
						|
    { call_site: [ffffffffa02911f2] drm_modeset_lock_crtc+0x32/0x100 [drm]                  } hitcount:       1848  bytes_req:     103488
 | 
						|
    { call_site: [ffffffffa04c4a3c] intel_plane_duplicate_state+0x2c/0xa0 [i915]            } hitcount:        461  bytes_req:      62696
 | 
						|
    { call_site: [ffffffffa029070e] drm_vma_node_allow+0x2e/0xd0 [drm]                      } hitcount:       1541  bytes_req:      61640
 | 
						|
    { call_site: [ffffffff815f8d7b] sk_prot_alloc+0xcb/0x1b0                                } hitcount:         57  bytes_req:      57456
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { call_site: [ffffffff8109524a] alloc_fair_sched_group+0x5a/0x1a0                       } hitcount:          2  bytes_req:        128
 | 
						|
    { call_site: [ffffffffa027b921] drm_vm_open_locked+0x31/0xa0 [drm]                      } hitcount:          3  bytes_req:         96
 | 
						|
    { call_site: [ffffffff8122e266] proc_self_follow_link+0x76/0xb0                         } hitcount:          8  bytes_req:         96
 | 
						|
    { call_site: [ffffffff81213e80] load_elf_binary+0x240/0x1650                            } hitcount:          3  bytes_req:         84
 | 
						|
    { call_site: [ffffffff8154bc62] usb_control_msg+0x42/0x110                              } hitcount:          1  bytes_req:          8
 | 
						|
    { call_site: [ffffffffa00bf6fe] hidraw_send_report+0x7e/0x1a0 [hid]                     } hitcount:          1  bytes_req:          7
 | 
						|
    { call_site: [ffffffffa00bf1ca] hidraw_report_event+0x8a/0x120 [hid]                    } hitcount:          1  bytes_req:          7
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 26098
 | 
						|
        Entries: 64
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  We can also add multiple fields to the 'values' parameter.  For
 | 
						|
  example, we might want to see the total number of bytes allocated
 | 
						|
  alongside bytes requested, and display the result sorted by bytes
 | 
						|
  allocated in a descending order::
 | 
						|
 | 
						|
    # echo 'hist:keys=call_site.sym:values=bytes_req,bytes_alloc:sort=bytes_alloc.descending' > \
 | 
						|
           /sys/kernel/tracing/events/kmem/kmalloc/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/kmem/kmalloc/hist
 | 
						|
    # trigger info: hist:keys=call_site.sym:vals=bytes_req,bytes_alloc:sort=bytes_alloc.descending:size=2048 [active]
 | 
						|
 | 
						|
    { call_site: [ffffffffa046041c] i915_gem_execbuffer2 [i915]                   } hitcount:       7403  bytes_req:    4084360  bytes_alloc:    5958016
 | 
						|
    { call_site: [ffffffff811e2a1b] seq_buf_alloc                                 } hitcount:        541  bytes_req:    2213968  bytes_alloc:    2228224
 | 
						|
    { call_site: [ffffffffa0489a66] intel_ring_begin [i915]                       } hitcount:       7404  bytes_req:    1066176  bytes_alloc:    1421568
 | 
						|
    { call_site: [ffffffffa045e7c4] i915_gem_do_execbuffer.isra.23 [i915]         } hitcount:       1565  bytes_req:     557368  bytes_alloc:    1037760
 | 
						|
    { call_site: [ffffffff8125847d] ext4_htree_store_dirent                       } hitcount:       9557  bytes_req:     595778  bytes_alloc:     695744
 | 
						|
    { call_site: [ffffffffa045e646] i915_gem_do_execbuffer.isra.23 [i915]         } hitcount:       5839  bytes_req:     430680  bytes_alloc:     470400
 | 
						|
    { call_site: [ffffffffa04c4a3c] intel_plane_duplicate_state [i915]            } hitcount:       2388  bytes_req:     324768  bytes_alloc:     458496
 | 
						|
    { call_site: [ffffffffa02911f2] drm_modeset_lock_crtc [drm]                   } hitcount:       3911  bytes_req:     219016  bytes_alloc:     250304
 | 
						|
    { call_site: [ffffffff815f8d7b] sk_prot_alloc                                 } hitcount:        235  bytes_req:     236880  bytes_alloc:     240640
 | 
						|
    { call_site: [ffffffff8137e559] sg_kmalloc                                    } hitcount:        557  bytes_req:     169024  bytes_alloc:     221760
 | 
						|
    { call_site: [ffffffffa00b7c06] hid_report_raw_event [hid]                    } hitcount:       9378  bytes_req:     187548  bytes_alloc:     206312
 | 
						|
    { call_site: [ffffffffa04a580c] intel_crtc_page_flip [i915]                   } hitcount:       1519  bytes_req:     157976  bytes_alloc:     194432
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { call_site: [ffffffff8109bd3b] sched_autogroup_create_attach                 } hitcount:          2  bytes_req:        144  bytes_alloc:        192
 | 
						|
    { call_site: [ffffffff81097ee8] alloc_rt_sched_group                          } hitcount:          2  bytes_req:        128  bytes_alloc:        128
 | 
						|
    { call_site: [ffffffff8109524a] alloc_fair_sched_group                        } hitcount:          2  bytes_req:        128  bytes_alloc:        128
 | 
						|
    { call_site: [ffffffff81095225] alloc_fair_sched_group                        } hitcount:          2  bytes_req:        128  bytes_alloc:        128
 | 
						|
    { call_site: [ffffffff81097ec2] alloc_rt_sched_group                          } hitcount:          2  bytes_req:        128  bytes_alloc:        128
 | 
						|
    { call_site: [ffffffff81213e80] load_elf_binary                               } hitcount:          3  bytes_req:         84  bytes_alloc:         96
 | 
						|
    { call_site: [ffffffff81079a2e] kthread_create_on_node                        } hitcount:          1  bytes_req:         56  bytes_alloc:         64
 | 
						|
    { call_site: [ffffffffa00bf6fe] hidraw_send_report [hid]                      } hitcount:          1  bytes_req:          7  bytes_alloc:          8
 | 
						|
    { call_site: [ffffffff8154bc62] usb_control_msg                               } hitcount:          1  bytes_req:          8  bytes_alloc:          8
 | 
						|
    { call_site: [ffffffffa00bf1ca] hidraw_report_event [hid]                     } hitcount:          1  bytes_req:          7  bytes_alloc:          8
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 66598
 | 
						|
        Entries: 65
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  Finally, to finish off our kmalloc example, instead of simply having
 | 
						|
  the hist trigger display symbolic call_sites, we can have the hist
 | 
						|
  trigger additionally display the complete set of kernel stack traces
 | 
						|
  that led to each call_site.  To do that, we simply use the special
 | 
						|
  value 'common_stacktrace' for the key parameter::
 | 
						|
 | 
						|
    # echo 'hist:keys=common_stacktrace:values=bytes_req,bytes_alloc:sort=bytes_alloc' > \
 | 
						|
           /sys/kernel/tracing/events/kmem/kmalloc/trigger
 | 
						|
 | 
						|
  The above trigger will use the kernel stack trace in effect when an
 | 
						|
  event is triggered as the key for the hash table.  This allows the
 | 
						|
  enumeration of every kernel callpath that led up to a particular
 | 
						|
  event, along with a running total of any of the event fields for
 | 
						|
  that event.  Here we tally bytes requested and bytes allocated for
 | 
						|
  every callpath in the system that led up to a kmalloc (in this case
 | 
						|
  every callpath to a kmalloc for a kernel compile)::
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/kmem/kmalloc/hist
 | 
						|
    # trigger info: hist:keys=common_stacktrace:vals=bytes_req,bytes_alloc:sort=bytes_alloc:size=2048 [active]
 | 
						|
 | 
						|
    { common_stacktrace:
 | 
						|
         __kmalloc_track_caller+0x10b/0x1a0
 | 
						|
         kmemdup+0x20/0x50
 | 
						|
         hidraw_report_event+0x8a/0x120 [hid]
 | 
						|
         hid_report_raw_event+0x3ea/0x440 [hid]
 | 
						|
         hid_input_report+0x112/0x190 [hid]
 | 
						|
         hid_irq_in+0xc2/0x260 [usbhid]
 | 
						|
         __usb_hcd_giveback_urb+0x72/0x120
 | 
						|
         usb_giveback_urb_bh+0x9e/0xe0
 | 
						|
         tasklet_hi_action+0xf8/0x100
 | 
						|
         __do_softirq+0x114/0x2c0
 | 
						|
         irq_exit+0xa5/0xb0
 | 
						|
         do_IRQ+0x5a/0xf0
 | 
						|
         ret_from_intr+0x0/0x30
 | 
						|
         cpuidle_enter+0x17/0x20
 | 
						|
         cpu_startup_entry+0x315/0x3e0
 | 
						|
         rest_init+0x7c/0x80
 | 
						|
    } hitcount:          3  bytes_req:         21  bytes_alloc:         24
 | 
						|
    { common_stacktrace:
 | 
						|
         __kmalloc_track_caller+0x10b/0x1a0
 | 
						|
         kmemdup+0x20/0x50
 | 
						|
         hidraw_report_event+0x8a/0x120 [hid]
 | 
						|
         hid_report_raw_event+0x3ea/0x440 [hid]
 | 
						|
         hid_input_report+0x112/0x190 [hid]
 | 
						|
         hid_irq_in+0xc2/0x260 [usbhid]
 | 
						|
         __usb_hcd_giveback_urb+0x72/0x120
 | 
						|
         usb_giveback_urb_bh+0x9e/0xe0
 | 
						|
         tasklet_hi_action+0xf8/0x100
 | 
						|
         __do_softirq+0x114/0x2c0
 | 
						|
         irq_exit+0xa5/0xb0
 | 
						|
         do_IRQ+0x5a/0xf0
 | 
						|
         ret_from_intr+0x0/0x30
 | 
						|
    } hitcount:          3  bytes_req:         21  bytes_alloc:         24
 | 
						|
    { common_stacktrace:
 | 
						|
         kmem_cache_alloc_trace+0xeb/0x150
 | 
						|
         aa_alloc_task_context+0x27/0x40
 | 
						|
         apparmor_cred_prepare+0x1f/0x50
 | 
						|
         security_prepare_creds+0x16/0x20
 | 
						|
         prepare_creds+0xdf/0x1a0
 | 
						|
         SyS_capset+0xb5/0x200
 | 
						|
         system_call_fastpath+0x12/0x6a
 | 
						|
    } hitcount:          1  bytes_req:         32  bytes_alloc:         32
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { common_stacktrace:
 | 
						|
         __kmalloc+0x11b/0x1b0
 | 
						|
         i915_gem_execbuffer2+0x6c/0x2c0 [i915]
 | 
						|
         drm_ioctl+0x349/0x670 [drm]
 | 
						|
         do_vfs_ioctl+0x2f0/0x4f0
 | 
						|
         SyS_ioctl+0x81/0xa0
 | 
						|
         system_call_fastpath+0x12/0x6a
 | 
						|
    } hitcount:      17726  bytes_req:   13944120  bytes_alloc:   19593808
 | 
						|
    { common_stacktrace:
 | 
						|
         __kmalloc+0x11b/0x1b0
 | 
						|
         load_elf_phdrs+0x76/0xa0
 | 
						|
         load_elf_binary+0x102/0x1650
 | 
						|
         search_binary_handler+0x97/0x1d0
 | 
						|
         do_execveat_common.isra.34+0x551/0x6e0
 | 
						|
         SyS_execve+0x3a/0x50
 | 
						|
         return_from_execve+0x0/0x23
 | 
						|
    } hitcount:      33348  bytes_req:   17152128  bytes_alloc:   20226048
 | 
						|
    { common_stacktrace:
 | 
						|
         kmem_cache_alloc_trace+0xeb/0x150
 | 
						|
         apparmor_file_alloc_security+0x27/0x40
 | 
						|
         security_file_alloc+0x16/0x20
 | 
						|
         get_empty_filp+0x93/0x1c0
 | 
						|
         path_openat+0x31/0x5f0
 | 
						|
         do_filp_open+0x3a/0x90
 | 
						|
         do_sys_open+0x128/0x220
 | 
						|
         SyS_open+0x1e/0x20
 | 
						|
         system_call_fastpath+0x12/0x6a
 | 
						|
    } hitcount:    4766422  bytes_req:    9532844  bytes_alloc:   38131376
 | 
						|
    { common_stacktrace:
 | 
						|
         __kmalloc+0x11b/0x1b0
 | 
						|
         seq_buf_alloc+0x1b/0x50
 | 
						|
         seq_read+0x2cc/0x370
 | 
						|
         proc_reg_read+0x3d/0x80
 | 
						|
         __vfs_read+0x28/0xe0
 | 
						|
         vfs_read+0x86/0x140
 | 
						|
         SyS_read+0x46/0xb0
 | 
						|
         system_call_fastpath+0x12/0x6a
 | 
						|
    } hitcount:      19133  bytes_req:   78368768  bytes_alloc:   78368768
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 6085872
 | 
						|
        Entries: 253
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  If you key a hist trigger on common_pid, in order for example to
 | 
						|
  gather and display sorted totals for each process, you can use the
 | 
						|
  special .execname modifier to display the executable names for the
 | 
						|
  processes in the table rather than raw pids.  The example below
 | 
						|
  keeps a per-process sum of total bytes read::
 | 
						|
 | 
						|
    # echo 'hist:key=common_pid.execname:val=count:sort=count.descending' > \
 | 
						|
           /sys/kernel/tracing/events/syscalls/sys_enter_read/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/syscalls/sys_enter_read/hist
 | 
						|
    # trigger info: hist:keys=common_pid.execname:vals=count:sort=count.descending:size=2048 [active]
 | 
						|
 | 
						|
    { common_pid: gnome-terminal  [      3196] } hitcount:        280  count:    1093512
 | 
						|
    { common_pid: Xorg            [      1309] } hitcount:        525  count:     256640
 | 
						|
    { common_pid: compiz          [      2889] } hitcount:         59  count:     254400
 | 
						|
    { common_pid: bash            [      8710] } hitcount:          3  count:      66369
 | 
						|
    { common_pid: dbus-daemon-lau [      8703] } hitcount:         49  count:      47739
 | 
						|
    { common_pid: irqbalance      [      1252] } hitcount:         27  count:      27648
 | 
						|
    { common_pid: 01ifupdown      [      8705] } hitcount:          3  count:      17216
 | 
						|
    { common_pid: dbus-daemon     [       772] } hitcount:         10  count:      12396
 | 
						|
    { common_pid: Socket Thread   [      8342] } hitcount:         11  count:      11264
 | 
						|
    { common_pid: nm-dhcp-client. [      8701] } hitcount:          6  count:       7424
 | 
						|
    { common_pid: gmain           [      1315] } hitcount:         18  count:       6336
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { common_pid: postgres        [      1892] } hitcount:          2  count:         32
 | 
						|
    { common_pid: postgres        [      1891] } hitcount:          2  count:         32
 | 
						|
    { common_pid: gmain           [      8704] } hitcount:          2  count:         32
 | 
						|
    { common_pid: upstart-dbus-br [      2740] } hitcount:         21  count:         21
 | 
						|
    { common_pid: nm-dispatcher.a [      8696] } hitcount:          1  count:         16
 | 
						|
    { common_pid: indicator-datet [      2904] } hitcount:          1  count:         16
 | 
						|
    { common_pid: gdbus           [      2998] } hitcount:          1  count:         16
 | 
						|
    { common_pid: rtkit-daemon    [      2052] } hitcount:          1  count:          8
 | 
						|
    { common_pid: init            [         1] } hitcount:          2  count:          2
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 2116
 | 
						|
        Entries: 51
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  Similarly, if you key a hist trigger on syscall id, for example to
 | 
						|
  gather and display a list of systemwide syscall hits, you can use
 | 
						|
  the special .syscall modifier to display the syscall names rather
 | 
						|
  than raw ids.  The example below keeps a running total of syscall
 | 
						|
  counts for the system during the run::
 | 
						|
 | 
						|
    # echo 'hist:key=id.syscall:val=hitcount' > \
 | 
						|
           /sys/kernel/tracing/events/raw_syscalls/sys_enter/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/raw_syscalls/sys_enter/hist
 | 
						|
    # trigger info: hist:keys=id.syscall:vals=hitcount:sort=hitcount:size=2048 [active]
 | 
						|
 | 
						|
    { id: sys_fsync                     [ 74] } hitcount:          1
 | 
						|
    { id: sys_newuname                  [ 63] } hitcount:          1
 | 
						|
    { id: sys_prctl                     [157] } hitcount:          1
 | 
						|
    { id: sys_statfs                    [137] } hitcount:          1
 | 
						|
    { id: sys_symlink                   [ 88] } hitcount:          1
 | 
						|
    { id: sys_sendmmsg                  [307] } hitcount:          1
 | 
						|
    { id: sys_semctl                    [ 66] } hitcount:          1
 | 
						|
    { id: sys_readlink                  [ 89] } hitcount:          3
 | 
						|
    { id: sys_bind                      [ 49] } hitcount:          3
 | 
						|
    { id: sys_getsockname               [ 51] } hitcount:          3
 | 
						|
    { id: sys_unlink                    [ 87] } hitcount:          3
 | 
						|
    { id: sys_rename                    [ 82] } hitcount:          4
 | 
						|
    { id: unknown_syscall               [ 58] } hitcount:          4
 | 
						|
    { id: sys_connect                   [ 42] } hitcount:          4
 | 
						|
    { id: sys_getpid                    [ 39] } hitcount:          4
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { id: sys_rt_sigprocmask            [ 14] } hitcount:        952
 | 
						|
    { id: sys_futex                     [202] } hitcount:       1534
 | 
						|
    { id: sys_write                     [  1] } hitcount:       2689
 | 
						|
    { id: sys_setitimer                 [ 38] } hitcount:       2797
 | 
						|
    { id: sys_read                      [  0] } hitcount:       3202
 | 
						|
    { id: sys_select                    [ 23] } hitcount:       3773
 | 
						|
    { id: sys_writev                    [ 20] } hitcount:       4531
 | 
						|
    { id: sys_poll                      [  7] } hitcount:       8314
 | 
						|
    { id: sys_recvmsg                   [ 47] } hitcount:      13738
 | 
						|
    { id: sys_ioctl                     [ 16] } hitcount:      21843
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 67612
 | 
						|
        Entries: 72
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  The syscall counts above provide a rough overall picture of system
 | 
						|
  call activity on the system; we can see for example that the most
 | 
						|
  popular system call on this system was the 'sys_ioctl' system call.
 | 
						|
 | 
						|
  We can use 'compound' keys to refine that number and provide some
 | 
						|
  further insight as to which processes exactly contribute to the
 | 
						|
  overall ioctl count.
 | 
						|
 | 
						|
  The command below keeps a hitcount for every unique combination of
 | 
						|
  system call id and pid - the end result is essentially a table
 | 
						|
  that keeps a per-pid sum of system call hits.  The results are
 | 
						|
  sorted using the system call id as the primary key, and the
 | 
						|
  hitcount sum as the secondary key::
 | 
						|
 | 
						|
    # echo 'hist:key=id.syscall,common_pid.execname:val=hitcount:sort=id,hitcount' > \
 | 
						|
           /sys/kernel/tracing/events/raw_syscalls/sys_enter/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/raw_syscalls/sys_enter/hist
 | 
						|
    # trigger info: hist:keys=id.syscall,common_pid.execname:vals=hitcount:sort=id.syscall,hitcount:size=2048 [active]
 | 
						|
 | 
						|
    { id: sys_read                      [  0], common_pid: rtkit-daemon    [      1877] } hitcount:          1
 | 
						|
    { id: sys_read                      [  0], common_pid: gdbus           [      2976] } hitcount:          1
 | 
						|
    { id: sys_read                      [  0], common_pid: console-kit-dae [      3400] } hitcount:          1
 | 
						|
    { id: sys_read                      [  0], common_pid: postgres        [      1865] } hitcount:          1
 | 
						|
    { id: sys_read                      [  0], common_pid: deja-dup-monito [      3543] } hitcount:          2
 | 
						|
    { id: sys_read                      [  0], common_pid: NetworkManager  [       890] } hitcount:          2
 | 
						|
    { id: sys_read                      [  0], common_pid: evolution-calen [      3048] } hitcount:          2
 | 
						|
    { id: sys_read                      [  0], common_pid: postgres        [      1864] } hitcount:          2
 | 
						|
    { id: sys_read                      [  0], common_pid: nm-applet       [      3022] } hitcount:          2
 | 
						|
    { id: sys_read                      [  0], common_pid: whoopsie        [      1212] } hitcount:          2
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: bash            [      8479] } hitcount:          1
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: bash            [      3472] } hitcount:         12
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: gnome-terminal  [      3199] } hitcount:         16
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: Xorg            [      1267] } hitcount:       1808
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: compiz          [      2994] } hitcount:       5580
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { id: sys_waitid                    [247], common_pid: upstart-dbus-br [      2690] } hitcount:          3
 | 
						|
    { id: sys_waitid                    [247], common_pid: upstart-dbus-br [      2688] } hitcount:         16
 | 
						|
    { id: sys_inotify_add_watch         [254], common_pid: gmain           [       975] } hitcount:          2
 | 
						|
    { id: sys_inotify_add_watch         [254], common_pid: gmain           [      3204] } hitcount:          4
 | 
						|
    { id: sys_inotify_add_watch         [254], common_pid: gmain           [      2888] } hitcount:          4
 | 
						|
    { id: sys_inotify_add_watch         [254], common_pid: gmain           [      3003] } hitcount:          4
 | 
						|
    { id: sys_inotify_add_watch         [254], common_pid: gmain           [      2873] } hitcount:          4
 | 
						|
    { id: sys_inotify_add_watch         [254], common_pid: gmain           [      3196] } hitcount:          6
 | 
						|
    { id: sys_openat                    [257], common_pid: java            [      2623] } hitcount:          2
 | 
						|
    { id: sys_eventfd2                  [290], common_pid: ibus-ui-gtk3    [      2760] } hitcount:          4
 | 
						|
    { id: sys_eventfd2                  [290], common_pid: compiz          [      2994] } hitcount:          6
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 31536
 | 
						|
        Entries: 323
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  The above list does give us a breakdown of the ioctl syscall by
 | 
						|
  pid, but it also gives us quite a bit more than that, which we
 | 
						|
  don't really care about at the moment.  Since we know the syscall
 | 
						|
  id for sys_ioctl (16, displayed next to the sys_ioctl name), we
 | 
						|
  can use that to filter out all the other syscalls::
 | 
						|
 | 
						|
    # echo 'hist:key=id.syscall,common_pid.execname:val=hitcount:sort=id,hitcount if id == 16' > \
 | 
						|
           /sys/kernel/tracing/events/raw_syscalls/sys_enter/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/raw_syscalls/sys_enter/hist
 | 
						|
    # trigger info: hist:keys=id.syscall,common_pid.execname:vals=hitcount:sort=id.syscall,hitcount:size=2048 if id == 16 [active]
 | 
						|
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: gmain           [      2769] } hitcount:          1
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: evolution-addre [      8571] } hitcount:          1
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: gmain           [      3003] } hitcount:          1
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: gmain           [      2781] } hitcount:          1
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: gmain           [      2829] } hitcount:          1
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: bash            [      8726] } hitcount:          1
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: bash            [      8508] } hitcount:          1
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: gmain           [      2970] } hitcount:          1
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: gmain           [      2768] } hitcount:          1
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: pool            [      8559] } hitcount:         45
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: pool            [      8555] } hitcount:         48
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: pool            [      8551] } hitcount:         48
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: avahi-daemon    [       896] } hitcount:         66
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: Xorg            [      1267] } hitcount:      26674
 | 
						|
    { id: sys_ioctl                     [ 16], common_pid: compiz          [      2994] } hitcount:      73443
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 101162
 | 
						|
        Entries: 103
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  The above output shows that 'compiz' and 'Xorg' are far and away
 | 
						|
  the heaviest ioctl callers (which might lead to questions about
 | 
						|
  whether they really need to be making all those calls and to
 | 
						|
  possible avenues for further investigation.)
 | 
						|
 | 
						|
  The compound key examples used a key and a sum value (hitcount) to
 | 
						|
  sort the output, but we can just as easily use two keys instead.
 | 
						|
  Here's an example where we use a compound key composed of the the
 | 
						|
  common_pid and size event fields.  Sorting with pid as the primary
 | 
						|
  key and 'size' as the secondary key allows us to display an
 | 
						|
  ordered summary of the recvfrom sizes, with counts, received by
 | 
						|
  each process::
 | 
						|
 | 
						|
    # echo 'hist:key=common_pid.execname,size:val=hitcount:sort=common_pid,size' > \
 | 
						|
           /sys/kernel/tracing/events/syscalls/sys_enter_recvfrom/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/syscalls/sys_enter_recvfrom/hist
 | 
						|
    # trigger info: hist:keys=common_pid.execname,size:vals=hitcount:sort=common_pid.execname,size:size=2048 [active]
 | 
						|
 | 
						|
    { common_pid: smbd            [       784], size:          4 } hitcount:          1
 | 
						|
    { common_pid: dnsmasq         [      1412], size:       4096 } hitcount:        672
 | 
						|
    { common_pid: postgres        [      1796], size:       1000 } hitcount:          6
 | 
						|
    { common_pid: postgres        [      1867], size:       1000 } hitcount:         10
 | 
						|
    { common_pid: bamfdaemon      [      2787], size:         28 } hitcount:          2
 | 
						|
    { common_pid: bamfdaemon      [      2787], size:      14360 } hitcount:          1
 | 
						|
    { common_pid: compiz          [      2994], size:          8 } hitcount:          1
 | 
						|
    { common_pid: compiz          [      2994], size:         20 } hitcount:         11
 | 
						|
    { common_pid: gnome-terminal  [      3199], size:          4 } hitcount:          2
 | 
						|
    { common_pid: firefox         [      8817], size:          4 } hitcount:          1
 | 
						|
    { common_pid: firefox         [      8817], size:          8 } hitcount:          5
 | 
						|
    { common_pid: firefox         [      8817], size:        588 } hitcount:          2
 | 
						|
    { common_pid: firefox         [      8817], size:        628 } hitcount:          1
 | 
						|
    { common_pid: firefox         [      8817], size:       6944 } hitcount:          1
 | 
						|
    { common_pid: firefox         [      8817], size:     408880 } hitcount:          2
 | 
						|
    { common_pid: firefox         [      8822], size:          8 } hitcount:          2
 | 
						|
    { common_pid: firefox         [      8822], size:        160 } hitcount:          2
 | 
						|
    { common_pid: firefox         [      8822], size:        320 } hitcount:          2
 | 
						|
    { common_pid: firefox         [      8822], size:        352 } hitcount:          1
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { common_pid: pool            [      8923], size:       1960 } hitcount:         10
 | 
						|
    { common_pid: pool            [      8923], size:       2048 } hitcount:         10
 | 
						|
    { common_pid: pool            [      8924], size:       1960 } hitcount:         10
 | 
						|
    { common_pid: pool            [      8924], size:       2048 } hitcount:         10
 | 
						|
    { common_pid: pool            [      8928], size:       1964 } hitcount:          4
 | 
						|
    { common_pid: pool            [      8928], size:       1965 } hitcount:          2
 | 
						|
    { common_pid: pool            [      8928], size:       2048 } hitcount:          6
 | 
						|
    { common_pid: pool            [      8929], size:       1982 } hitcount:          1
 | 
						|
    { common_pid: pool            [      8929], size:       2048 } hitcount:          1
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 2016
 | 
						|
        Entries: 224
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  The above example also illustrates the fact that although a compound
 | 
						|
  key is treated as a single entity for hashing purposes, the sub-keys
 | 
						|
  it's composed of can be accessed independently.
 | 
						|
 | 
						|
  The next example uses a string field as the hash key and
 | 
						|
  demonstrates how you can manually pause and continue a hist trigger.
 | 
						|
  In this example, we'll aggregate fork counts and don't expect a
 | 
						|
  large number of entries in the hash table, so we'll drop it to a
 | 
						|
  much smaller number, say 256::
 | 
						|
 | 
						|
    # echo 'hist:key=child_comm:val=hitcount:size=256' > \
 | 
						|
           /sys/kernel/tracing/events/sched/sched_process_fork/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/sched/sched_process_fork/hist
 | 
						|
    # trigger info: hist:keys=child_comm:vals=hitcount:sort=hitcount:size=256 [active]
 | 
						|
 | 
						|
    { child_comm: dconf worker                        } hitcount:          1
 | 
						|
    { child_comm: ibus-daemon                         } hitcount:          1
 | 
						|
    { child_comm: whoopsie                            } hitcount:          1
 | 
						|
    { child_comm: smbd                                } hitcount:          1
 | 
						|
    { child_comm: gdbus                               } hitcount:          1
 | 
						|
    { child_comm: kthreadd                            } hitcount:          1
 | 
						|
    { child_comm: dconf worker                        } hitcount:          1
 | 
						|
    { child_comm: evolution-alarm                     } hitcount:          2
 | 
						|
    { child_comm: Socket Thread                       } hitcount:          2
 | 
						|
    { child_comm: postgres                            } hitcount:          2
 | 
						|
    { child_comm: bash                                } hitcount:          3
 | 
						|
    { child_comm: compiz                              } hitcount:          3
 | 
						|
    { child_comm: evolution-sourc                     } hitcount:          4
 | 
						|
    { child_comm: dhclient                            } hitcount:          4
 | 
						|
    { child_comm: pool                                } hitcount:          5
 | 
						|
    { child_comm: nm-dispatcher.a                     } hitcount:          8
 | 
						|
    { child_comm: firefox                             } hitcount:          8
 | 
						|
    { child_comm: dbus-daemon                         } hitcount:          8
 | 
						|
    { child_comm: glib-pacrunner                      } hitcount:         10
 | 
						|
    { child_comm: evolution                           } hitcount:         23
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 89
 | 
						|
        Entries: 20
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  If we want to pause the hist trigger, we can simply append :pause to
 | 
						|
  the command that started the trigger.  Notice that the trigger info
 | 
						|
  displays as [paused]::
 | 
						|
 | 
						|
    # echo 'hist:key=child_comm:val=hitcount:size=256:pause' >> \
 | 
						|
           /sys/kernel/tracing/events/sched/sched_process_fork/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/sched/sched_process_fork/hist
 | 
						|
    # trigger info: hist:keys=child_comm:vals=hitcount:sort=hitcount:size=256 [paused]
 | 
						|
 | 
						|
    { child_comm: dconf worker                        } hitcount:          1
 | 
						|
    { child_comm: kthreadd                            } hitcount:          1
 | 
						|
    { child_comm: dconf worker                        } hitcount:          1
 | 
						|
    { child_comm: gdbus                               } hitcount:          1
 | 
						|
    { child_comm: ibus-daemon                         } hitcount:          1
 | 
						|
    { child_comm: Socket Thread                       } hitcount:          2
 | 
						|
    { child_comm: evolution-alarm                     } hitcount:          2
 | 
						|
    { child_comm: smbd                                } hitcount:          2
 | 
						|
    { child_comm: bash                                } hitcount:          3
 | 
						|
    { child_comm: whoopsie                            } hitcount:          3
 | 
						|
    { child_comm: compiz                              } hitcount:          3
 | 
						|
    { child_comm: evolution-sourc                     } hitcount:          4
 | 
						|
    { child_comm: pool                                } hitcount:          5
 | 
						|
    { child_comm: postgres                            } hitcount:          6
 | 
						|
    { child_comm: firefox                             } hitcount:          8
 | 
						|
    { child_comm: dhclient                            } hitcount:         10
 | 
						|
    { child_comm: emacs                               } hitcount:         12
 | 
						|
    { child_comm: dbus-daemon                         } hitcount:         20
 | 
						|
    { child_comm: nm-dispatcher.a                     } hitcount:         20
 | 
						|
    { child_comm: evolution                           } hitcount:         35
 | 
						|
    { child_comm: glib-pacrunner                      } hitcount:         59
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 199
 | 
						|
        Entries: 21
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  To manually continue having the trigger aggregate events, append
 | 
						|
  :cont instead.  Notice that the trigger info displays as [active]
 | 
						|
  again, and the data has changed::
 | 
						|
 | 
						|
    # echo 'hist:key=child_comm:val=hitcount:size=256:cont' >> \
 | 
						|
           /sys/kernel/tracing/events/sched/sched_process_fork/trigger
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/sched/sched_process_fork/hist
 | 
						|
    # trigger info: hist:keys=child_comm:vals=hitcount:sort=hitcount:size=256 [active]
 | 
						|
 | 
						|
    { child_comm: dconf worker                        } hitcount:          1
 | 
						|
    { child_comm: dconf worker                        } hitcount:          1
 | 
						|
    { child_comm: kthreadd                            } hitcount:          1
 | 
						|
    { child_comm: gdbus                               } hitcount:          1
 | 
						|
    { child_comm: ibus-daemon                         } hitcount:          1
 | 
						|
    { child_comm: Socket Thread                       } hitcount:          2
 | 
						|
    { child_comm: evolution-alarm                     } hitcount:          2
 | 
						|
    { child_comm: smbd                                } hitcount:          2
 | 
						|
    { child_comm: whoopsie                            } hitcount:          3
 | 
						|
    { child_comm: compiz                              } hitcount:          3
 | 
						|
    { child_comm: evolution-sourc                     } hitcount:          4
 | 
						|
    { child_comm: bash                                } hitcount:          5
 | 
						|
    { child_comm: pool                                } hitcount:          5
 | 
						|
    { child_comm: postgres                            } hitcount:          6
 | 
						|
    { child_comm: firefox                             } hitcount:          8
 | 
						|
    { child_comm: dhclient                            } hitcount:         11
 | 
						|
    { child_comm: emacs                               } hitcount:         12
 | 
						|
    { child_comm: dbus-daemon                         } hitcount:         22
 | 
						|
    { child_comm: nm-dispatcher.a                     } hitcount:         22
 | 
						|
    { child_comm: evolution                           } hitcount:         35
 | 
						|
    { child_comm: glib-pacrunner                      } hitcount:         59
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 206
 | 
						|
        Entries: 21
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  The previous example showed how to start and stop a hist trigger by
 | 
						|
  appending 'pause' and 'continue' to the hist trigger command.  A
 | 
						|
  hist trigger can also be started in a paused state by initially
 | 
						|
  starting the trigger with ':pause' appended.  This allows you to
 | 
						|
  start the trigger only when you're ready to start collecting data
 | 
						|
  and not before.  For example, you could start the trigger in a
 | 
						|
  paused state, then unpause it and do something you want to measure,
 | 
						|
  then pause the trigger again when done.
 | 
						|
 | 
						|
  Of course, doing this manually can be difficult and error-prone, but
 | 
						|
  it is possible to automatically start and stop a hist trigger based
 | 
						|
  on some condition, via the enable_hist and disable_hist triggers.
 | 
						|
 | 
						|
  For example, suppose we wanted to take a look at the relative
 | 
						|
  weights in terms of skb length for each callpath that leads to a
 | 
						|
  netif_receive_skb event when downloading a decent-sized file using
 | 
						|
  wget.
 | 
						|
 | 
						|
  First we set up an initially paused stacktrace trigger on the
 | 
						|
  netif_receive_skb event::
 | 
						|
 | 
						|
    # echo 'hist:key=common_stacktrace:vals=len:pause' > \
 | 
						|
           /sys/kernel/tracing/events/net/netif_receive_skb/trigger
 | 
						|
 | 
						|
  Next, we set up an 'enable_hist' trigger on the sched_process_exec
 | 
						|
  event, with an 'if filename==/usr/bin/wget' filter.  The effect of
 | 
						|
  this new trigger is that it will 'unpause' the hist trigger we just
 | 
						|
  set up on netif_receive_skb if and only if it sees a
 | 
						|
  sched_process_exec event with a filename of '/usr/bin/wget'.  When
 | 
						|
  that happens, all netif_receive_skb events are aggregated into a
 | 
						|
  hash table keyed on stacktrace::
 | 
						|
 | 
						|
    # echo 'enable_hist:net:netif_receive_skb if filename==/usr/bin/wget' > \
 | 
						|
           /sys/kernel/tracing/events/sched/sched_process_exec/trigger
 | 
						|
 | 
						|
  The aggregation continues until the netif_receive_skb is paused
 | 
						|
  again, which is what the following disable_hist event does by
 | 
						|
  creating a similar setup on the sched_process_exit event, using the
 | 
						|
  filter 'comm==wget'::
 | 
						|
 | 
						|
    # echo 'disable_hist:net:netif_receive_skb if comm==wget' > \
 | 
						|
           /sys/kernel/tracing/events/sched/sched_process_exit/trigger
 | 
						|
 | 
						|
  Whenever a process exits and the comm field of the disable_hist
 | 
						|
  trigger filter matches 'comm==wget', the netif_receive_skb hist
 | 
						|
  trigger is disabled.
 | 
						|
 | 
						|
  The overall effect is that netif_receive_skb events are aggregated
 | 
						|
  into the hash table for only the duration of the wget.  Executing a
 | 
						|
  wget command and then listing the 'hist' file will display the
 | 
						|
  output generated by the wget command::
 | 
						|
 | 
						|
    $ wget https://www.kernel.org/pub/linux/kernel/v3.x/patch-3.19.xz
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/net/netif_receive_skb/hist
 | 
						|
    # trigger info: hist:keys=common_stacktrace:vals=len:sort=hitcount:size=2048 [paused]
 | 
						|
 | 
						|
    { common_stacktrace:
 | 
						|
         __netif_receive_skb_core+0x46d/0x990
 | 
						|
         __netif_receive_skb+0x18/0x60
 | 
						|
         netif_receive_skb_internal+0x23/0x90
 | 
						|
         napi_gro_receive+0xc8/0x100
 | 
						|
         ieee80211_deliver_skb+0xd6/0x270 [mac80211]
 | 
						|
         ieee80211_rx_handlers+0xccf/0x22f0 [mac80211]
 | 
						|
         ieee80211_prepare_and_rx_handle+0x4e7/0xc40 [mac80211]
 | 
						|
         ieee80211_rx+0x31d/0x900 [mac80211]
 | 
						|
         iwlagn_rx_reply_rx+0x3db/0x6f0 [iwldvm]
 | 
						|
         iwl_rx_dispatch+0x8e/0xf0 [iwldvm]
 | 
						|
         iwl_pcie_irq_handler+0xe3c/0x12f0 [iwlwifi]
 | 
						|
         irq_thread_fn+0x20/0x50
 | 
						|
         irq_thread+0x11f/0x150
 | 
						|
         kthread+0xd2/0xf0
 | 
						|
         ret_from_fork+0x42/0x70
 | 
						|
    } hitcount:         85  len:      28884
 | 
						|
    { common_stacktrace:
 | 
						|
         __netif_receive_skb_core+0x46d/0x990
 | 
						|
         __netif_receive_skb+0x18/0x60
 | 
						|
         netif_receive_skb_internal+0x23/0x90
 | 
						|
         napi_gro_complete+0xa4/0xe0
 | 
						|
         dev_gro_receive+0x23a/0x360
 | 
						|
         napi_gro_receive+0x30/0x100
 | 
						|
         ieee80211_deliver_skb+0xd6/0x270 [mac80211]
 | 
						|
         ieee80211_rx_handlers+0xccf/0x22f0 [mac80211]
 | 
						|
         ieee80211_prepare_and_rx_handle+0x4e7/0xc40 [mac80211]
 | 
						|
         ieee80211_rx+0x31d/0x900 [mac80211]
 | 
						|
         iwlagn_rx_reply_rx+0x3db/0x6f0 [iwldvm]
 | 
						|
         iwl_rx_dispatch+0x8e/0xf0 [iwldvm]
 | 
						|
         iwl_pcie_irq_handler+0xe3c/0x12f0 [iwlwifi]
 | 
						|
         irq_thread_fn+0x20/0x50
 | 
						|
         irq_thread+0x11f/0x150
 | 
						|
         kthread+0xd2/0xf0
 | 
						|
    } hitcount:         98  len:     664329
 | 
						|
    { common_stacktrace:
 | 
						|
         __netif_receive_skb_core+0x46d/0x990
 | 
						|
         __netif_receive_skb+0x18/0x60
 | 
						|
         process_backlog+0xa8/0x150
 | 
						|
         net_rx_action+0x15d/0x340
 | 
						|
         __do_softirq+0x114/0x2c0
 | 
						|
         do_softirq_own_stack+0x1c/0x30
 | 
						|
         do_softirq+0x65/0x70
 | 
						|
         __local_bh_enable_ip+0xb5/0xc0
 | 
						|
         ip_finish_output+0x1f4/0x840
 | 
						|
         ip_output+0x6b/0xc0
 | 
						|
         ip_local_out_sk+0x31/0x40
 | 
						|
         ip_send_skb+0x1a/0x50
 | 
						|
         udp_send_skb+0x173/0x2a0
 | 
						|
         udp_sendmsg+0x2bf/0x9f0
 | 
						|
         inet_sendmsg+0x64/0xa0
 | 
						|
         sock_sendmsg+0x3d/0x50
 | 
						|
    } hitcount:        115  len:      13030
 | 
						|
    { common_stacktrace:
 | 
						|
         __netif_receive_skb_core+0x46d/0x990
 | 
						|
         __netif_receive_skb+0x18/0x60
 | 
						|
         netif_receive_skb_internal+0x23/0x90
 | 
						|
         napi_gro_complete+0xa4/0xe0
 | 
						|
         napi_gro_flush+0x6d/0x90
 | 
						|
         iwl_pcie_irq_handler+0x92a/0x12f0 [iwlwifi]
 | 
						|
         irq_thread_fn+0x20/0x50
 | 
						|
         irq_thread+0x11f/0x150
 | 
						|
         kthread+0xd2/0xf0
 | 
						|
         ret_from_fork+0x42/0x70
 | 
						|
    } hitcount:        934  len:    5512212
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 1232
 | 
						|
        Entries: 4
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  The above shows all the netif_receive_skb callpaths and their total
 | 
						|
  lengths for the duration of the wget command.
 | 
						|
 | 
						|
  The 'clear' hist trigger param can be used to clear the hash table.
 | 
						|
  Suppose we wanted to try another run of the previous example but
 | 
						|
  this time also wanted to see the complete list of events that went
 | 
						|
  into the histogram.  In order to avoid having to set everything up
 | 
						|
  again, we can just clear the histogram first::
 | 
						|
 | 
						|
    # echo 'hist:key=common_stacktrace:vals=len:clear' >> \
 | 
						|
           /sys/kernel/tracing/events/net/netif_receive_skb/trigger
 | 
						|
 | 
						|
  Just to verify that it is in fact cleared, here's what we now see in
 | 
						|
  the hist file::
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/net/netif_receive_skb/hist
 | 
						|
    # trigger info: hist:keys=common_stacktrace:vals=len:sort=hitcount:size=2048 [paused]
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 0
 | 
						|
        Entries: 0
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  Since we want to see the detailed list of every netif_receive_skb
 | 
						|
  event occurring during the new run, which are in fact the same
 | 
						|
  events being aggregated into the hash table, we add some additional
 | 
						|
  'enable_event' events to the triggering sched_process_exec and
 | 
						|
  sched_process_exit events as such::
 | 
						|
 | 
						|
    # echo 'enable_event:net:netif_receive_skb if filename==/usr/bin/wget' > \
 | 
						|
           /sys/kernel/tracing/events/sched/sched_process_exec/trigger
 | 
						|
 | 
						|
    # echo 'disable_event:net:netif_receive_skb if comm==wget' > \
 | 
						|
           /sys/kernel/tracing/events/sched/sched_process_exit/trigger
 | 
						|
 | 
						|
  If you read the trigger files for the sched_process_exec and
 | 
						|
  sched_process_exit triggers, you should see two triggers for each:
 | 
						|
  one enabling/disabling the hist aggregation and the other
 | 
						|
  enabling/disabling the logging of events::
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/sched/sched_process_exec/trigger
 | 
						|
    enable_event:net:netif_receive_skb:unlimited if filename==/usr/bin/wget
 | 
						|
    enable_hist:net:netif_receive_skb:unlimited if filename==/usr/bin/wget
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/sched/sched_process_exit/trigger
 | 
						|
    enable_event:net:netif_receive_skb:unlimited if comm==wget
 | 
						|
    disable_hist:net:netif_receive_skb:unlimited if comm==wget
 | 
						|
 | 
						|
  In other words, whenever either of the sched_process_exec or
 | 
						|
  sched_process_exit events is hit and matches 'wget', it enables or
 | 
						|
  disables both the histogram and the event log, and what you end up
 | 
						|
  with is a hash table and set of events just covering the specified
 | 
						|
  duration.  Run the wget command again::
 | 
						|
 | 
						|
    $ wget https://www.kernel.org/pub/linux/kernel/v3.x/patch-3.19.xz
 | 
						|
 | 
						|
  Displaying the 'hist' file should show something similar to what you
 | 
						|
  saw in the last run, but this time you should also see the
 | 
						|
  individual events in the trace file::
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/trace
 | 
						|
 | 
						|
    # tracer: nop
 | 
						|
    #
 | 
						|
    # entries-in-buffer/entries-written: 183/1426   #P:4
 | 
						|
    #
 | 
						|
    #                              _-----=> irqs-off
 | 
						|
    #                             / _----=> need-resched
 | 
						|
    #                            | / _---=> hardirq/softirq
 | 
						|
    #                            || / _--=> preempt-depth
 | 
						|
    #                            ||| /     delay
 | 
						|
    #           TASK-PID   CPU#  ||||    TIMESTAMP  FUNCTION
 | 
						|
    #              | |       |   ||||       |         |
 | 
						|
                wget-15108 [000] ..s1 31769.606929: netif_receive_skb: dev=lo skbaddr=ffff88009c353100 len=60
 | 
						|
                wget-15108 [000] ..s1 31769.606999: netif_receive_skb: dev=lo skbaddr=ffff88009c353200 len=60
 | 
						|
             dnsmasq-1382  [000] ..s1 31769.677652: netif_receive_skb: dev=lo skbaddr=ffff88009c352b00 len=130
 | 
						|
             dnsmasq-1382  [000] ..s1 31769.685917: netif_receive_skb: dev=lo skbaddr=ffff88009c352200 len=138
 | 
						|
    ##### CPU 2 buffer started ####
 | 
						|
      irq/29-iwlwifi-559   [002] ..s. 31772.031529: netif_receive_skb: dev=wlan0 skbaddr=ffff88009d433d00 len=2948
 | 
						|
      irq/29-iwlwifi-559   [002] ..s. 31772.031572: netif_receive_skb: dev=wlan0 skbaddr=ffff88009d432200 len=1500
 | 
						|
      irq/29-iwlwifi-559   [002] ..s. 31772.032196: netif_receive_skb: dev=wlan0 skbaddr=ffff88009d433100 len=2948
 | 
						|
      irq/29-iwlwifi-559   [002] ..s. 31772.032761: netif_receive_skb: dev=wlan0 skbaddr=ffff88009d433000 len=2948
 | 
						|
      irq/29-iwlwifi-559   [002] ..s. 31772.033220: netif_receive_skb: dev=wlan0 skbaddr=ffff88009d432e00 len=1500
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
 | 
						|
  The following example demonstrates how multiple hist triggers can be
 | 
						|
  attached to a given event.  This capability can be useful for
 | 
						|
  creating a set of different summaries derived from the same set of
 | 
						|
  events, or for comparing the effects of different filters, among
 | 
						|
  other things::
 | 
						|
 | 
						|
    # echo 'hist:keys=skbaddr.hex:vals=len if len < 0' >> \
 | 
						|
           /sys/kernel/tracing/events/net/netif_receive_skb/trigger
 | 
						|
    # echo 'hist:keys=skbaddr.hex:vals=len if len > 4096' >> \
 | 
						|
           /sys/kernel/tracing/events/net/netif_receive_skb/trigger
 | 
						|
    # echo 'hist:keys=skbaddr.hex:vals=len if len == 256' >> \
 | 
						|
           /sys/kernel/tracing/events/net/netif_receive_skb/trigger
 | 
						|
    # echo 'hist:keys=skbaddr.hex:vals=len' >> \
 | 
						|
           /sys/kernel/tracing/events/net/netif_receive_skb/trigger
 | 
						|
    # echo 'hist:keys=len:vals=common_preempt_count' >> \
 | 
						|
           /sys/kernel/tracing/events/net/netif_receive_skb/trigger
 | 
						|
 | 
						|
  The above set of commands create four triggers differing only in
 | 
						|
  their filters, along with a completely different though fairly
 | 
						|
  nonsensical trigger.  Note that in order to append multiple hist
 | 
						|
  triggers to the same file, you should use the '>>' operator to
 | 
						|
  append them ('>' will also add the new hist trigger, but will remove
 | 
						|
  any existing hist triggers beforehand).
 | 
						|
 | 
						|
  Displaying the contents of the 'hist' file for the event shows the
 | 
						|
  contents of all five histograms::
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/net/netif_receive_skb/hist
 | 
						|
 | 
						|
    # event histogram
 | 
						|
    #
 | 
						|
    # trigger info: hist:keys=len:vals=hitcount,common_preempt_count:sort=hitcount:size=2048 [active]
 | 
						|
    #
 | 
						|
 | 
						|
    { len:        176 } hitcount:          1  common_preempt_count:          0
 | 
						|
    { len:        223 } hitcount:          1  common_preempt_count:          0
 | 
						|
    { len:       4854 } hitcount:          1  common_preempt_count:          0
 | 
						|
    { len:        395 } hitcount:          1  common_preempt_count:          0
 | 
						|
    { len:        177 } hitcount:          1  common_preempt_count:          0
 | 
						|
    { len:        446 } hitcount:          1  common_preempt_count:          0
 | 
						|
    { len:       1601 } hitcount:          1  common_preempt_count:          0
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { len:       1280 } hitcount:         66  common_preempt_count:          0
 | 
						|
    { len:        116 } hitcount:         81  common_preempt_count:         40
 | 
						|
    { len:        708 } hitcount:        112  common_preempt_count:          0
 | 
						|
    { len:         46 } hitcount:        221  common_preempt_count:          0
 | 
						|
    { len:       1264 } hitcount:        458  common_preempt_count:          0
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 1428
 | 
						|
        Entries: 147
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
 | 
						|
    # event histogram
 | 
						|
    #
 | 
						|
    # trigger info: hist:keys=skbaddr.hex:vals=hitcount,len:sort=hitcount:size=2048 [active]
 | 
						|
    #
 | 
						|
 | 
						|
    { skbaddr: ffff8800baee5e00 } hitcount:          1  len:        130
 | 
						|
    { skbaddr: ffff88005f3d5600 } hitcount:          1  len:       1280
 | 
						|
    { skbaddr: ffff88005f3d4900 } hitcount:          1  len:       1280
 | 
						|
    { skbaddr: ffff88009fed6300 } hitcount:          1  len:        115
 | 
						|
    { skbaddr: ffff88009fe0ad00 } hitcount:          1  len:        115
 | 
						|
    { skbaddr: ffff88008cdb1900 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff880064b5ef00 } hitcount:          1  len:        118
 | 
						|
    { skbaddr: ffff880044e3c700 } hitcount:          1  len:         60
 | 
						|
    { skbaddr: ffff880100065900 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800d46bd500 } hitcount:          1  len:        116
 | 
						|
    { skbaddr: ffff88005f3d5f00 } hitcount:          1  len:       1280
 | 
						|
    { skbaddr: ffff880100064700 } hitcount:          1  len:        365
 | 
						|
    { skbaddr: ffff8800badb6f00 } hitcount:          1  len:         60
 | 
						|
    .
 | 
						|
    .
 | 
						|
    .
 | 
						|
    { skbaddr: ffff88009fe0be00 } hitcount:         27  len:      24677
 | 
						|
    { skbaddr: ffff88009fe0a400 } hitcount:         27  len:      23052
 | 
						|
    { skbaddr: ffff88009fe0b700 } hitcount:         31  len:      25589
 | 
						|
    { skbaddr: ffff88009fe0b600 } hitcount:         32  len:      27326
 | 
						|
    { skbaddr: ffff88006a462800 } hitcount:         68  len:      71678
 | 
						|
    { skbaddr: ffff88006a463700 } hitcount:         70  len:      72678
 | 
						|
    { skbaddr: ffff88006a462b00 } hitcount:         71  len:      77589
 | 
						|
    { skbaddr: ffff88006a463600 } hitcount:         73  len:      71307
 | 
						|
    { skbaddr: ffff88006a462200 } hitcount:         81  len:      81032
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 1451
 | 
						|
        Entries: 318
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
 | 
						|
    # event histogram
 | 
						|
    #
 | 
						|
    # trigger info: hist:keys=skbaddr.hex:vals=hitcount,len:sort=hitcount:size=2048 if len == 256 [active]
 | 
						|
    #
 | 
						|
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 0
 | 
						|
        Entries: 0
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
 | 
						|
    # event histogram
 | 
						|
    #
 | 
						|
    # trigger info: hist:keys=skbaddr.hex:vals=hitcount,len:sort=hitcount:size=2048 if len > 4096 [active]
 | 
						|
    #
 | 
						|
 | 
						|
    { skbaddr: ffff88009fd2c300 } hitcount:          1  len:       7212
 | 
						|
    { skbaddr: ffff8800d2bcce00 } hitcount:          1  len:       7212
 | 
						|
    { skbaddr: ffff8800d2bcd700 } hitcount:          1  len:       7212
 | 
						|
    { skbaddr: ffff8800d2bcda00 } hitcount:          1  len:      21492
 | 
						|
    { skbaddr: ffff8800ae2e2d00 } hitcount:          1  len:       7212
 | 
						|
    { skbaddr: ffff8800d2bcdb00 } hitcount:          1  len:       7212
 | 
						|
    { skbaddr: ffff88006a4df500 } hitcount:          1  len:       4854
 | 
						|
    { skbaddr: ffff88008ce47b00 } hitcount:          1  len:      18636
 | 
						|
    { skbaddr: ffff8800ae2e2200 } hitcount:          1  len:      12924
 | 
						|
    { skbaddr: ffff88005f3e1000 } hitcount:          1  len:       4356
 | 
						|
    { skbaddr: ffff8800d2bcdc00 } hitcount:          2  len:      24420
 | 
						|
    { skbaddr: ffff8800d2bcc200 } hitcount:          2  len:      12996
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 14
 | 
						|
        Entries: 12
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
 | 
						|
    # event histogram
 | 
						|
    #
 | 
						|
    # trigger info: hist:keys=skbaddr.hex:vals=hitcount,len:sort=hitcount:size=2048 if len < 0 [active]
 | 
						|
    #
 | 
						|
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 0
 | 
						|
        Entries: 0
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  Named triggers can be used to have triggers share a common set of
 | 
						|
  histogram data.  This capability is mostly useful for combining the
 | 
						|
  output of events generated by tracepoints contained inside inline
 | 
						|
  functions, but names can be used in a hist trigger on any event.
 | 
						|
  For example, these two triggers when hit will update the same 'len'
 | 
						|
  field in the shared 'foo' histogram data::
 | 
						|
 | 
						|
    # echo 'hist:name=foo:keys=skbaddr.hex:vals=len' > \
 | 
						|
           /sys/kernel/tracing/events/net/netif_receive_skb/trigger
 | 
						|
    # echo 'hist:name=foo:keys=skbaddr.hex:vals=len' > \
 | 
						|
           /sys/kernel/tracing/events/net/netif_rx/trigger
 | 
						|
 | 
						|
  You can see that they're updating common histogram data by reading
 | 
						|
  each event's hist files at the same time::
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/net/netif_receive_skb/hist;
 | 
						|
      cat /sys/kernel/tracing/events/net/netif_rx/hist
 | 
						|
 | 
						|
    # event histogram
 | 
						|
    #
 | 
						|
    # trigger info: hist:name=foo:keys=skbaddr.hex:vals=hitcount,len:sort=hitcount:size=2048 [active]
 | 
						|
    #
 | 
						|
 | 
						|
    { skbaddr: ffff88000ad53500 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800af5a1500 } hitcount:          1  len:         76
 | 
						|
    { skbaddr: ffff8800d62a1900 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800d2bccb00 } hitcount:          1  len:        468
 | 
						|
    { skbaddr: ffff8800d3c69900 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff88009ff09100 } hitcount:          1  len:         52
 | 
						|
    { skbaddr: ffff88010f13ab00 } hitcount:          1  len:        168
 | 
						|
    { skbaddr: ffff88006a54f400 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800d2bcc500 } hitcount:          1  len:        260
 | 
						|
    { skbaddr: ffff880064505000 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800baf24e00 } hitcount:          1  len:         32
 | 
						|
    { skbaddr: ffff88009fe0ad00 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800d3edff00 } hitcount:          1  len:         44
 | 
						|
    { skbaddr: ffff88009fe0b400 } hitcount:          1  len:        168
 | 
						|
    { skbaddr: ffff8800a1c55a00 } hitcount:          1  len:         40
 | 
						|
    { skbaddr: ffff8800d2bcd100 } hitcount:          1  len:         40
 | 
						|
    { skbaddr: ffff880064505f00 } hitcount:          1  len:        174
 | 
						|
    { skbaddr: ffff8800a8bff200 } hitcount:          1  len:        160
 | 
						|
    { skbaddr: ffff880044e3cc00 } hitcount:          1  len:         76
 | 
						|
    { skbaddr: ffff8800a8bfe700 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800d2bcdc00 } hitcount:          1  len:         32
 | 
						|
    { skbaddr: ffff8800a1f64800 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800d2bcde00 } hitcount:          1  len:        988
 | 
						|
    { skbaddr: ffff88006a5dea00 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff88002e37a200 } hitcount:          1  len:         44
 | 
						|
    { skbaddr: ffff8800a1f32c00 } hitcount:          2  len:        676
 | 
						|
    { skbaddr: ffff88000ad52600 } hitcount:          2  len:        107
 | 
						|
    { skbaddr: ffff8800a1f91e00 } hitcount:          2  len:         92
 | 
						|
    { skbaddr: ffff8800af5a0200 } hitcount:          2  len:        142
 | 
						|
    { skbaddr: ffff8800d2bcc600 } hitcount:          2  len:        220
 | 
						|
    { skbaddr: ffff8800ba36f500 } hitcount:          2  len:         92
 | 
						|
    { skbaddr: ffff8800d021f800 } hitcount:          2  len:         92
 | 
						|
    { skbaddr: ffff8800a1f33600 } hitcount:          2  len:        675
 | 
						|
    { skbaddr: ffff8800a8bfff00 } hitcount:          3  len:        138
 | 
						|
    { skbaddr: ffff8800d62a1300 } hitcount:          3  len:        138
 | 
						|
    { skbaddr: ffff88002e37a100 } hitcount:          4  len:        184
 | 
						|
    { skbaddr: ffff880064504400 } hitcount:          4  len:        184
 | 
						|
    { skbaddr: ffff8800a8bfec00 } hitcount:          4  len:        184
 | 
						|
    { skbaddr: ffff88000ad53700 } hitcount:          5  len:        230
 | 
						|
    { skbaddr: ffff8800d2bcdb00 } hitcount:          5  len:        196
 | 
						|
    { skbaddr: ffff8800a1f90000 } hitcount:          6  len:        276
 | 
						|
    { skbaddr: ffff88006a54f900 } hitcount:          6  len:        276
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 81
 | 
						|
        Entries: 42
 | 
						|
        Dropped: 0
 | 
						|
    # event histogram
 | 
						|
    #
 | 
						|
    # trigger info: hist:name=foo:keys=skbaddr.hex:vals=hitcount,len:sort=hitcount:size=2048 [active]
 | 
						|
    #
 | 
						|
 | 
						|
    { skbaddr: ffff88000ad53500 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800af5a1500 } hitcount:          1  len:         76
 | 
						|
    { skbaddr: ffff8800d62a1900 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800d2bccb00 } hitcount:          1  len:        468
 | 
						|
    { skbaddr: ffff8800d3c69900 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff88009ff09100 } hitcount:          1  len:         52
 | 
						|
    { skbaddr: ffff88010f13ab00 } hitcount:          1  len:        168
 | 
						|
    { skbaddr: ffff88006a54f400 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800d2bcc500 } hitcount:          1  len:        260
 | 
						|
    { skbaddr: ffff880064505000 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800baf24e00 } hitcount:          1  len:         32
 | 
						|
    { skbaddr: ffff88009fe0ad00 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800d3edff00 } hitcount:          1  len:         44
 | 
						|
    { skbaddr: ffff88009fe0b400 } hitcount:          1  len:        168
 | 
						|
    { skbaddr: ffff8800a1c55a00 } hitcount:          1  len:         40
 | 
						|
    { skbaddr: ffff8800d2bcd100 } hitcount:          1  len:         40
 | 
						|
    { skbaddr: ffff880064505f00 } hitcount:          1  len:        174
 | 
						|
    { skbaddr: ffff8800a8bff200 } hitcount:          1  len:        160
 | 
						|
    { skbaddr: ffff880044e3cc00 } hitcount:          1  len:         76
 | 
						|
    { skbaddr: ffff8800a8bfe700 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800d2bcdc00 } hitcount:          1  len:         32
 | 
						|
    { skbaddr: ffff8800a1f64800 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff8800d2bcde00 } hitcount:          1  len:        988
 | 
						|
    { skbaddr: ffff88006a5dea00 } hitcount:          1  len:         46
 | 
						|
    { skbaddr: ffff88002e37a200 } hitcount:          1  len:         44
 | 
						|
    { skbaddr: ffff8800a1f32c00 } hitcount:          2  len:        676
 | 
						|
    { skbaddr: ffff88000ad52600 } hitcount:          2  len:        107
 | 
						|
    { skbaddr: ffff8800a1f91e00 } hitcount:          2  len:         92
 | 
						|
    { skbaddr: ffff8800af5a0200 } hitcount:          2  len:        142
 | 
						|
    { skbaddr: ffff8800d2bcc600 } hitcount:          2  len:        220
 | 
						|
    { skbaddr: ffff8800ba36f500 } hitcount:          2  len:         92
 | 
						|
    { skbaddr: ffff8800d021f800 } hitcount:          2  len:         92
 | 
						|
    { skbaddr: ffff8800a1f33600 } hitcount:          2  len:        675
 | 
						|
    { skbaddr: ffff8800a8bfff00 } hitcount:          3  len:        138
 | 
						|
    { skbaddr: ffff8800d62a1300 } hitcount:          3  len:        138
 | 
						|
    { skbaddr: ffff88002e37a100 } hitcount:          4  len:        184
 | 
						|
    { skbaddr: ffff880064504400 } hitcount:          4  len:        184
 | 
						|
    { skbaddr: ffff8800a8bfec00 } hitcount:          4  len:        184
 | 
						|
    { skbaddr: ffff88000ad53700 } hitcount:          5  len:        230
 | 
						|
    { skbaddr: ffff8800d2bcdb00 } hitcount:          5  len:        196
 | 
						|
    { skbaddr: ffff8800a1f90000 } hitcount:          6  len:        276
 | 
						|
    { skbaddr: ffff88006a54f900 } hitcount:          6  len:        276
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 81
 | 
						|
        Entries: 42
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
  And here's an example that shows how to combine histogram data from
 | 
						|
  any two events even if they don't share any 'compatible' fields
 | 
						|
  other than 'hitcount' and 'common_stacktrace'.  These commands create a
 | 
						|
  couple of triggers named 'bar' using those fields::
 | 
						|
 | 
						|
    # echo 'hist:name=bar:key=common_stacktrace:val=hitcount' > \
 | 
						|
           /sys/kernel/tracing/events/sched/sched_process_fork/trigger
 | 
						|
    # echo 'hist:name=bar:key=common_stacktrace:val=hitcount' > \
 | 
						|
          /sys/kernel/tracing/events/net/netif_rx/trigger
 | 
						|
 | 
						|
  And displaying the output of either shows some interesting if
 | 
						|
  somewhat confusing output::
 | 
						|
 | 
						|
    # cat /sys/kernel/tracing/events/sched/sched_process_fork/hist
 | 
						|
    # cat /sys/kernel/tracing/events/net/netif_rx/hist
 | 
						|
 | 
						|
    # event histogram
 | 
						|
    #
 | 
						|
    # trigger info: hist:name=bar:keys=common_stacktrace:vals=hitcount:sort=hitcount:size=2048 [active]
 | 
						|
    #
 | 
						|
 | 
						|
    { common_stacktrace:
 | 
						|
             kernel_clone+0x18e/0x330
 | 
						|
             kernel_thread+0x29/0x30
 | 
						|
             kthreadd+0x154/0x1b0
 | 
						|
             ret_from_fork+0x3f/0x70
 | 
						|
    } hitcount:          1
 | 
						|
    { common_stacktrace:
 | 
						|
             netif_rx_internal+0xb2/0xd0
 | 
						|
             netif_rx_ni+0x20/0x70
 | 
						|
             dev_loopback_xmit+0xaa/0xd0
 | 
						|
             ip_mc_output+0x126/0x240
 | 
						|
             ip_local_out_sk+0x31/0x40
 | 
						|
             igmp_send_report+0x1e9/0x230
 | 
						|
             igmp_timer_expire+0xe9/0x120
 | 
						|
             call_timer_fn+0x39/0xf0
 | 
						|
             run_timer_softirq+0x1e1/0x290
 | 
						|
             __do_softirq+0xfd/0x290
 | 
						|
             irq_exit+0x98/0xb0
 | 
						|
             smp_apic_timer_interrupt+0x4a/0x60
 | 
						|
             apic_timer_interrupt+0x6d/0x80
 | 
						|
             cpuidle_enter+0x17/0x20
 | 
						|
             call_cpuidle+0x3b/0x60
 | 
						|
             cpu_startup_entry+0x22d/0x310
 | 
						|
    } hitcount:          1
 | 
						|
    { common_stacktrace:
 | 
						|
             netif_rx_internal+0xb2/0xd0
 | 
						|
             netif_rx_ni+0x20/0x70
 | 
						|
             dev_loopback_xmit+0xaa/0xd0
 | 
						|
             ip_mc_output+0x17f/0x240
 | 
						|
             ip_local_out_sk+0x31/0x40
 | 
						|
             ip_send_skb+0x1a/0x50
 | 
						|
             udp_send_skb+0x13e/0x270
 | 
						|
             udp_sendmsg+0x2bf/0x980
 | 
						|
             inet_sendmsg+0x67/0xa0
 | 
						|
             sock_sendmsg+0x38/0x50
 | 
						|
             SYSC_sendto+0xef/0x170
 | 
						|
             SyS_sendto+0xe/0x10
 | 
						|
             entry_SYSCALL_64_fastpath+0x12/0x6a
 | 
						|
    } hitcount:          2
 | 
						|
    { common_stacktrace:
 | 
						|
             netif_rx_internal+0xb2/0xd0
 | 
						|
             netif_rx+0x1c/0x60
 | 
						|
             loopback_xmit+0x6c/0xb0
 | 
						|
             dev_hard_start_xmit+0x219/0x3a0
 | 
						|
             __dev_queue_xmit+0x415/0x4f0
 | 
						|
             dev_queue_xmit_sk+0x13/0x20
 | 
						|
             ip_finish_output2+0x237/0x340
 | 
						|
             ip_finish_output+0x113/0x1d0
 | 
						|
             ip_output+0x66/0xc0
 | 
						|
             ip_local_out_sk+0x31/0x40
 | 
						|
             ip_send_skb+0x1a/0x50
 | 
						|
             udp_send_skb+0x16d/0x270
 | 
						|
             udp_sendmsg+0x2bf/0x980
 | 
						|
             inet_sendmsg+0x67/0xa0
 | 
						|
             sock_sendmsg+0x38/0x50
 | 
						|
             ___sys_sendmsg+0x14e/0x270
 | 
						|
    } hitcount:         76
 | 
						|
    { common_stacktrace:
 | 
						|
             netif_rx_internal+0xb2/0xd0
 | 
						|
             netif_rx+0x1c/0x60
 | 
						|
             loopback_xmit+0x6c/0xb0
 | 
						|
             dev_hard_start_xmit+0x219/0x3a0
 | 
						|
             __dev_queue_xmit+0x415/0x4f0
 | 
						|
             dev_queue_xmit_sk+0x13/0x20
 | 
						|
             ip_finish_output2+0x237/0x340
 | 
						|
             ip_finish_output+0x113/0x1d0
 | 
						|
             ip_output+0x66/0xc0
 | 
						|
             ip_local_out_sk+0x31/0x40
 | 
						|
             ip_send_skb+0x1a/0x50
 | 
						|
             udp_send_skb+0x16d/0x270
 | 
						|
             udp_sendmsg+0x2bf/0x980
 | 
						|
             inet_sendmsg+0x67/0xa0
 | 
						|
             sock_sendmsg+0x38/0x50
 | 
						|
             ___sys_sendmsg+0x269/0x270
 | 
						|
    } hitcount:         77
 | 
						|
    { common_stacktrace:
 | 
						|
             netif_rx_internal+0xb2/0xd0
 | 
						|
             netif_rx+0x1c/0x60
 | 
						|
             loopback_xmit+0x6c/0xb0
 | 
						|
             dev_hard_start_xmit+0x219/0x3a0
 | 
						|
             __dev_queue_xmit+0x415/0x4f0
 | 
						|
             dev_queue_xmit_sk+0x13/0x20
 | 
						|
             ip_finish_output2+0x237/0x340
 | 
						|
             ip_finish_output+0x113/0x1d0
 | 
						|
             ip_output+0x66/0xc0
 | 
						|
             ip_local_out_sk+0x31/0x40
 | 
						|
             ip_send_skb+0x1a/0x50
 | 
						|
             udp_send_skb+0x16d/0x270
 | 
						|
             udp_sendmsg+0x2bf/0x980
 | 
						|
             inet_sendmsg+0x67/0xa0
 | 
						|
             sock_sendmsg+0x38/0x50
 | 
						|
             SYSC_sendto+0xef/0x170
 | 
						|
    } hitcount:         88
 | 
						|
    { common_stacktrace:
 | 
						|
             kernel_clone+0x18e/0x330
 | 
						|
             SyS_clone+0x19/0x20
 | 
						|
             entry_SYSCALL_64_fastpath+0x12/0x6a
 | 
						|
    } hitcount:        244
 | 
						|
 | 
						|
    Totals:
 | 
						|
        Hits: 489
 | 
						|
        Entries: 7
 | 
						|
        Dropped: 0
 | 
						|
 | 
						|
2.2 Inter-event hist triggers
 | 
						|
-----------------------------
 | 
						|
 | 
						|
Inter-event hist triggers are hist triggers that combine values from
 | 
						|
one or more other events and create a histogram using that data.  Data
 | 
						|
from an inter-event histogram can in turn become the source for
 | 
						|
further combined histograms, thus providing a chain of related
 | 
						|
histograms, which is important for some applications.
 | 
						|
 | 
						|
The most important example of an inter-event quantity that can be used
 | 
						|
in this manner is latency, which is simply a difference in timestamps
 | 
						|
between two events.  Although latency is the most important
 | 
						|
inter-event quantity, note that because the support is completely
 | 
						|
general across the trace event subsystem, any event field can be used
 | 
						|
in an inter-event quantity.
 | 
						|
 | 
						|
An example of a histogram that combines data from other histograms
 | 
						|
into a useful chain would be a 'wakeupswitch latency' histogram that
 | 
						|
combines a 'wakeup latency' histogram and a 'switch latency'
 | 
						|
histogram.
 | 
						|
 | 
						|
Normally, a hist trigger specification consists of a (possibly
 | 
						|
compound) key along with one or more numeric values, which are
 | 
						|
continually updated sums associated with that key.  A histogram
 | 
						|
specification in this case consists of individual key and value
 | 
						|
specifications that refer to trace event fields associated with a
 | 
						|
single event type.
 | 
						|
 | 
						|
The inter-event hist trigger extension allows fields from multiple
 | 
						|
events to be referenced and combined into a multi-event histogram
 | 
						|
specification.  In support of this overall goal, a few enabling
 | 
						|
features have been added to the hist trigger support:
 | 
						|
 | 
						|
  - In order to compute an inter-event quantity, a value from one
 | 
						|
    event needs to saved and then referenced from another event.  This
 | 
						|
    requires the introduction of support for histogram 'variables'.
 | 
						|
 | 
						|
  - The computation of inter-event quantities and their combination
 | 
						|
    require some minimal amount of support for applying simple
 | 
						|
    expressions to variables (+ and -).
 | 
						|
 | 
						|
  - A histogram consisting of inter-event quantities isn't logically a
 | 
						|
    histogram on either event (so having the 'hist' file for either
 | 
						|
    event host the histogram output doesn't really make sense).  To
 | 
						|
    address the idea that the histogram is associated with a
 | 
						|
    combination of events, support is added allowing the creation of
 | 
						|
    'synthetic' events that are events derived from other events.
 | 
						|
    These synthetic events are full-fledged events just like any other
 | 
						|
    and can be used as such, as for instance to create the
 | 
						|
    'combination' histograms mentioned previously.
 | 
						|
 | 
						|
  - A set of 'actions' can be associated with histogram entries -
 | 
						|
    these can be used to generate the previously mentioned synthetic
 | 
						|
    events, but can also be used for other purposes, such as for
 | 
						|
    example saving context when a 'max' latency has been hit.
 | 
						|
 | 
						|
  - Trace events don't have a 'timestamp' associated with them, but
 | 
						|
    there is an implicit timestamp saved along with an event in the
 | 
						|
    underlying ftrace ring buffer.  This timestamp is now exposed as a
 | 
						|
    a synthetic field named 'common_timestamp' which can be used in
 | 
						|
    histograms as if it were any other event field; it isn't an actual
 | 
						|
    field in the trace format but rather is a synthesized value that
 | 
						|
    nonetheless can be used as if it were an actual field.  By default
 | 
						|
    it is in units of nanoseconds; appending '.usecs' to a
 | 
						|
    common_timestamp field changes the units to microseconds.
 | 
						|
 | 
						|
A note on inter-event timestamps: If common_timestamp is used in a
 | 
						|
histogram, the trace buffer is automatically switched over to using
 | 
						|
absolute timestamps and the "global" trace clock, in order to avoid
 | 
						|
bogus timestamp differences with other clocks that aren't coherent
 | 
						|
across CPUs.  This can be overridden by specifying one of the other
 | 
						|
trace clocks instead, using the "clock=XXX" hist trigger attribute,
 | 
						|
where XXX is any of the clocks listed in the tracing/trace_clock
 | 
						|
pseudo-file.
 | 
						|
 | 
						|
These features are described in more detail in the following sections.
 | 
						|
 | 
						|
2.2.1 Histogram Variables
 | 
						|
-------------------------
 | 
						|
 | 
						|
Variables are simply named locations used for saving and retrieving
 | 
						|
values between matching events.  A 'matching' event is defined as an
 | 
						|
event that has a matching key - if a variable is saved for a histogram
 | 
						|
entry corresponding to that key, any subsequent event with a matching
 | 
						|
key can access that variable.
 | 
						|
 | 
						|
A variable's value is normally available to any subsequent event until
 | 
						|
it is set to something else by a subsequent event.  The one exception
 | 
						|
to that rule is that any variable used in an expression is essentially
 | 
						|
'read-once' - once it's used by an expression in a subsequent event,
 | 
						|
it's reset to its 'unset' state, which means it can't be used again
 | 
						|
unless it's set again.  This ensures not only that an event doesn't
 | 
						|
use an uninitialized variable in a calculation, but that that variable
 | 
						|
is used only once and not for any unrelated subsequent match.
 | 
						|
 | 
						|
The basic syntax for saving a variable is to simply prefix a unique
 | 
						|
variable name not corresponding to any keyword along with an '=' sign
 | 
						|
to any event field.
 | 
						|
 | 
						|
Either keys or values can be saved and retrieved in this way.  This
 | 
						|
creates a variable named 'ts0' for a histogram entry with the key
 | 
						|
'next_pid'::
 | 
						|
 | 
						|
  # echo 'hist:keys=next_pid:vals=$ts0:ts0=common_timestamp ... >> \
 | 
						|
	event/trigger
 | 
						|
 | 
						|
The ts0 variable can be accessed by any subsequent event having the
 | 
						|
same pid as 'next_pid'.
 | 
						|
 | 
						|
Variable references are formed by prepending the variable name with
 | 
						|
the '$' sign.  Thus for example, the ts0 variable above would be
 | 
						|
referenced as '$ts0' in expressions.
 | 
						|
 | 
						|
Because 'vals=' is used, the common_timestamp variable value above
 | 
						|
will also be summed as a normal histogram value would (though for a
 | 
						|
timestamp it makes little sense).
 | 
						|
 | 
						|
The below shows that a key value can also be saved in the same way::
 | 
						|
 | 
						|
  # echo 'hist:timer_pid=common_pid:key=timer_pid ...' >> event/trigger
 | 
						|
 | 
						|
If a variable isn't a key variable or prefixed with 'vals=', the
 | 
						|
associated event field will be saved in a variable but won't be summed
 | 
						|
as a value::
 | 
						|
 | 
						|
  # echo 'hist:keys=next_pid:ts1=common_timestamp ...' >> event/trigger
 | 
						|
 | 
						|
Multiple variables can be assigned at the same time.  The below would
 | 
						|
result in both ts0 and b being created as variables, with both
 | 
						|
common_timestamp and field1 additionally being summed as values::
 | 
						|
 | 
						|
  # echo 'hist:keys=pid:vals=$ts0,$b:ts0=common_timestamp,b=field1 ...' >> \
 | 
						|
	event/trigger
 | 
						|
 | 
						|
Note that variable assignments can appear either preceding or
 | 
						|
following their use.  The command below behaves identically to the
 | 
						|
command above::
 | 
						|
 | 
						|
  # echo 'hist:keys=pid:ts0=common_timestamp,b=field1:vals=$ts0,$b ...' >> \
 | 
						|
	event/trigger
 | 
						|
 | 
						|
Any number of variables not bound to a 'vals=' prefix can also be
 | 
						|
assigned by simply separating them with colons.  Below is the same
 | 
						|
thing but without the values being summed in the histogram::
 | 
						|
 | 
						|
  # echo 'hist:keys=pid:ts0=common_timestamp:b=field1 ...' >> event/trigger
 | 
						|
 | 
						|
Variables set as above can be referenced and used in expressions on
 | 
						|
another event.
 | 
						|
 | 
						|
For example, here's how a latency can be calculated::
 | 
						|
 | 
						|
  # echo 'hist:keys=pid,prio:ts0=common_timestamp ...' >> event1/trigger
 | 
						|
  # echo 'hist:keys=next_pid:wakeup_lat=common_timestamp-$ts0 ...' >> event2/trigger
 | 
						|
 | 
						|
In the first line above, the event's timestamp is saved into the
 | 
						|
variable ts0.  In the next line, ts0 is subtracted from the second
 | 
						|
event's timestamp to produce the latency, which is then assigned into
 | 
						|
yet another variable, 'wakeup_lat'.  The hist trigger below in turn
 | 
						|
makes use of the wakeup_lat variable to compute a combined latency
 | 
						|
using the same key and variable from yet another event::
 | 
						|
 | 
						|
  # echo 'hist:key=pid:wakeupswitch_lat=$wakeup_lat+$switchtime_lat ...' >> event3/trigger
 | 
						|
 | 
						|
Expressions support the use of addition, subtraction, multiplication and
 | 
						|
division operators (+-\*/).
 | 
						|
 | 
						|
Note if division by zero cannot be detected at parse time (i.e. the
 | 
						|
divisor is not a constant), the result will be -1.
 | 
						|
 | 
						|
Numeric constants can also be used directly in an expression::
 | 
						|
 | 
						|
  # echo 'hist:keys=next_pid:timestamp_secs=common_timestamp/1000000 ...' >> event/trigger
 | 
						|
 | 
						|
or assigned to a variable and referenced in a subsequent expression::
 | 
						|
 | 
						|
  # echo 'hist:keys=next_pid:us_per_sec=1000000 ...' >> event/trigger
 | 
						|
  # echo 'hist:keys=next_pid:timestamp_secs=common_timestamp/$us_per_sec ...' >> event/trigger
 | 
						|
 | 
						|
Variables can even hold stacktraces, which are useful with synthetic events.
 | 
						|
 | 
						|
2.2.2 Synthetic Events
 | 
						|
----------------------
 | 
						|
 | 
						|
Synthetic events are user-defined events generated from hist trigger
 | 
						|
variables or fields associated with one or more other events.  Their
 | 
						|
purpose is to provide a mechanism for displaying data spanning
 | 
						|
multiple events consistent with the existing and already familiar
 | 
						|
usage for normal events.
 | 
						|
 | 
						|
To define a synthetic event, the user writes a simple specification
 | 
						|
consisting of the name of the new event along with one or more
 | 
						|
variables and their types, which can be any valid field type,
 | 
						|
separated by semicolons, to the tracing/synthetic_events file.
 | 
						|
 | 
						|
See synth_field_size() for available types.
 | 
						|
 | 
						|
If field_name contains [n], the field is considered to be a static array.
 | 
						|
 | 
						|
If field_names contains[] (no subscript), the field is considered to
 | 
						|
be a dynamic array, which will only take as much space in the event as
 | 
						|
is required to hold the array.
 | 
						|
 | 
						|
A string field can be specified using either the static notation:
 | 
						|
 | 
						|
  char name[32];
 | 
						|
 | 
						|
Or the dynamic:
 | 
						|
 | 
						|
  char name[];
 | 
						|
 | 
						|
The size limit for either is 256.
 | 
						|
 | 
						|
For instance, the following creates a new event named 'wakeup_latency'
 | 
						|
with 3 fields: lat, pid, and prio.  Each of those fields is simply a
 | 
						|
variable reference to a variable on another event::
 | 
						|
 | 
						|
  # echo 'wakeup_latency \
 | 
						|
          u64 lat; \
 | 
						|
          pid_t pid; \
 | 
						|
	  int prio' >> \
 | 
						|
	  /sys/kernel/tracing/synthetic_events
 | 
						|
 | 
						|
Reading the tracing/synthetic_events file lists all the currently
 | 
						|
defined synthetic events, in this case the event defined above::
 | 
						|
 | 
						|
  # cat /sys/kernel/tracing/synthetic_events
 | 
						|
    wakeup_latency u64 lat; pid_t pid; int prio
 | 
						|
 | 
						|
An existing synthetic event definition can be removed by prepending
 | 
						|
the command that defined it with a '!'::
 | 
						|
 | 
						|
  # echo '!wakeup_latency u64 lat pid_t pid int prio' >> \
 | 
						|
    /sys/kernel/tracing/synthetic_events
 | 
						|
 | 
						|
At this point, there isn't yet an actual 'wakeup_latency' event
 | 
						|
instantiated in the event subsystem - for this to happen, a 'hist
 | 
						|
trigger action' needs to be instantiated and bound to actual fields
 | 
						|
and variables defined on other events (see Section 2.2.3 below on
 | 
						|
how that is done using hist trigger 'onmatch' action). Once that is
 | 
						|
done, the 'wakeup_latency' synthetic event instance is created.
 | 
						|
 | 
						|
The new event is created under the tracing/events/synthetic/ directory
 | 
						|
and looks and behaves just like any other event::
 | 
						|
 | 
						|
  # ls /sys/kernel/tracing/events/synthetic/wakeup_latency
 | 
						|
        enable  filter  format  hist  id  trigger
 | 
						|
 | 
						|
A histogram can now be defined for the new synthetic event::
 | 
						|
 | 
						|
  # echo 'hist:keys=pid,prio,lat.log2:sort=lat' >> \
 | 
						|
        /sys/kernel/tracing/events/synthetic/wakeup_latency/trigger
 | 
						|
 | 
						|
The above shows the latency "lat" in a power of 2 grouping.
 | 
						|
 | 
						|
Like any other event, once a histogram is enabled for the event, the
 | 
						|
output can be displayed by reading the event's 'hist' file::
 | 
						|
 | 
						|
  # cat /sys/kernel/tracing/events/synthetic/wakeup_latency/hist
 | 
						|
 | 
						|
  # event histogram
 | 
						|
  #
 | 
						|
  # trigger info: hist:keys=pid,prio,lat.log2:vals=hitcount:sort=lat.log2:size=2048 [active]
 | 
						|
  #
 | 
						|
 | 
						|
  { pid:       2035, prio:          9, lat: ~ 2^2  } hitcount:         43
 | 
						|
  { pid:       2034, prio:          9, lat: ~ 2^2  } hitcount:         60
 | 
						|
  { pid:       2029, prio:          9, lat: ~ 2^2  } hitcount:        965
 | 
						|
  { pid:       2034, prio:        120, lat: ~ 2^2  } hitcount:          9
 | 
						|
  { pid:       2033, prio:        120, lat: ~ 2^2  } hitcount:          5
 | 
						|
  { pid:       2030, prio:          9, lat: ~ 2^2  } hitcount:        335
 | 
						|
  { pid:       2030, prio:        120, lat: ~ 2^2  } hitcount:         10
 | 
						|
  { pid:       2032, prio:        120, lat: ~ 2^2  } hitcount:          1
 | 
						|
  { pid:       2035, prio:        120, lat: ~ 2^2  } hitcount:          2
 | 
						|
  { pid:       2031, prio:          9, lat: ~ 2^2  } hitcount:        176
 | 
						|
  { pid:       2028, prio:        120, lat: ~ 2^2  } hitcount:         15
 | 
						|
  { pid:       2033, prio:          9, lat: ~ 2^2  } hitcount:         91
 | 
						|
  { pid:       2032, prio:          9, lat: ~ 2^2  } hitcount:        125
 | 
						|
  { pid:       2029, prio:        120, lat: ~ 2^2  } hitcount:          4
 | 
						|
  { pid:       2031, prio:        120, lat: ~ 2^2  } hitcount:          3
 | 
						|
  { pid:       2029, prio:        120, lat: ~ 2^3  } hitcount:          2
 | 
						|
  { pid:       2035, prio:          9, lat: ~ 2^3  } hitcount:         41
 | 
						|
  { pid:       2030, prio:        120, lat: ~ 2^3  } hitcount:          1
 | 
						|
  { pid:       2032, prio:          9, lat: ~ 2^3  } hitcount:         32
 | 
						|
  { pid:       2031, prio:          9, lat: ~ 2^3  } hitcount:         44
 | 
						|
  { pid:       2034, prio:          9, lat: ~ 2^3  } hitcount:         40
 | 
						|
  { pid:       2030, prio:          9, lat: ~ 2^3  } hitcount:         29
 | 
						|
  { pid:       2033, prio:          9, lat: ~ 2^3  } hitcount:         31
 | 
						|
  { pid:       2029, prio:          9, lat: ~ 2^3  } hitcount:         31
 | 
						|
  { pid:       2028, prio:        120, lat: ~ 2^3  } hitcount:         18
 | 
						|
  { pid:       2031, prio:        120, lat: ~ 2^3  } hitcount:          2
 | 
						|
  { pid:       2028, prio:        120, lat: ~ 2^4  } hitcount:          1
 | 
						|
  { pid:       2029, prio:          9, lat: ~ 2^4  } hitcount:          4
 | 
						|
  { pid:       2031, prio:        120, lat: ~ 2^7  } hitcount:          1
 | 
						|
  { pid:       2032, prio:        120, lat: ~ 2^7  } hitcount:          1
 | 
						|
 | 
						|
  Totals:
 | 
						|
      Hits: 2122
 | 
						|
      Entries: 30
 | 
						|
      Dropped: 0
 | 
						|
 | 
						|
 | 
						|
The latency values can also be grouped linearly by a given size with
 | 
						|
the ".buckets" modifier and specify a size (in this case groups of 10)::
 | 
						|
 | 
						|
  # echo 'hist:keys=pid,prio,lat.buckets=10:sort=lat' >> \
 | 
						|
        /sys/kernel/tracing/events/synthetic/wakeup_latency/trigger
 | 
						|
 | 
						|
  # event histogram
 | 
						|
  #
 | 
						|
  # trigger info: hist:keys=pid,prio,lat.buckets=10:vals=hitcount:sort=lat.buckets=10:size=2048 [active]
 | 
						|
  #
 | 
						|
 | 
						|
  { pid:       2067, prio:          9, lat: ~ 0-9 } hitcount:        220
 | 
						|
  { pid:       2068, prio:          9, lat: ~ 0-9 } hitcount:        157
 | 
						|
  { pid:       2070, prio:          9, lat: ~ 0-9 } hitcount:        100
 | 
						|
  { pid:       2067, prio:        120, lat: ~ 0-9 } hitcount:          6
 | 
						|
  { pid:       2065, prio:        120, lat: ~ 0-9 } hitcount:          2
 | 
						|
  { pid:       2066, prio:        120, lat: ~ 0-9 } hitcount:          2
 | 
						|
  { pid:       2069, prio:          9, lat: ~ 0-9 } hitcount:        122
 | 
						|
  { pid:       2069, prio:        120, lat: ~ 0-9 } hitcount:          8
 | 
						|
  { pid:       2070, prio:        120, lat: ~ 0-9 } hitcount:          1
 | 
						|
  { pid:       2068, prio:        120, lat: ~ 0-9 } hitcount:          7
 | 
						|
  { pid:       2066, prio:          9, lat: ~ 0-9 } hitcount:        365
 | 
						|
  { pid:       2064, prio:        120, lat: ~ 0-9 } hitcount:         35
 | 
						|
  { pid:       2065, prio:          9, lat: ~ 0-9 } hitcount:        998
 | 
						|
  { pid:       2071, prio:          9, lat: ~ 0-9 } hitcount:         85
 | 
						|
  { pid:       2065, prio:          9, lat: ~ 10-19 } hitcount:          2
 | 
						|
  { pid:       2064, prio:        120, lat: ~ 10-19 } hitcount:          2
 | 
						|
 | 
						|
  Totals:
 | 
						|
      Hits: 2112
 | 
						|
      Entries: 16
 | 
						|
      Dropped: 0
 | 
						|
 | 
						|
To save stacktraces, create a synthetic event with a field of type "unsigned long[]"
 | 
						|
or even just "long[]". For example, to see how long a task is blocked in an
 | 
						|
uninterruptible state::
 | 
						|
 | 
						|
  # cd /sys/kernel/tracing
 | 
						|
  # echo 's:block_lat pid_t pid; u64 delta; unsigned long[] stack;' > dynamic_events
 | 
						|
  # echo 'hist:keys=next_pid:ts=common_timestamp.usecs,st=common_stacktrace  if prev_state == 2' >> events/sched/sched_switch/trigger
 | 
						|
  # echo 'hist:keys=prev_pid:delta=common_timestamp.usecs-$ts,s=$st:onmax($delta).trace(block_lat,prev_pid,$delta,$s)' >> events/sched/sched_switch/trigger
 | 
						|
  # echo 1 > events/synthetic/block_lat/enable
 | 
						|
  # cat trace
 | 
						|
 | 
						|
  # tracer: nop
 | 
						|
  #
 | 
						|
  # entries-in-buffer/entries-written: 2/2   #P:8
 | 
						|
  #
 | 
						|
  #                                _-----=> irqs-off/BH-disabled
 | 
						|
  #                               / _----=> need-resched
 | 
						|
  #                              | / _---=> hardirq/softirq
 | 
						|
  #                              || / _--=> preempt-depth
 | 
						|
  #                              ||| / _-=> migrate-disable
 | 
						|
  #                              |||| /     delay
 | 
						|
  #           TASK-PID     CPU#  |||||  TIMESTAMP  FUNCTION
 | 
						|
  #              | |         |   |||||     |         |
 | 
						|
            <idle>-0       [005] d..4.   521.164922: block_lat: pid=0 delta=8322 stack=STACK:
 | 
						|
  => __schedule+0x448/0x7b0
 | 
						|
  => schedule+0x5a/0xb0
 | 
						|
  => io_schedule+0x42/0x70
 | 
						|
  => bit_wait_io+0xd/0x60
 | 
						|
  => __wait_on_bit+0x4b/0x140
 | 
						|
  => out_of_line_wait_on_bit+0x91/0xb0
 | 
						|
  => jbd2_journal_commit_transaction+0x1679/0x1a70
 | 
						|
  => kjournald2+0xa9/0x280
 | 
						|
  => kthread+0xe9/0x110
 | 
						|
  => ret_from_fork+0x2c/0x50
 | 
						|
 | 
						|
             <...>-2       [004] d..4.   525.184257: block_lat: pid=2 delta=76 stack=STACK:
 | 
						|
  => __schedule+0x448/0x7b0
 | 
						|
  => schedule+0x5a/0xb0
 | 
						|
  => schedule_timeout+0x11a/0x150
 | 
						|
  => wait_for_completion_killable+0x144/0x1f0
 | 
						|
  => __kthread_create_on_node+0xe7/0x1e0
 | 
						|
  => kthread_create_on_node+0x51/0x70
 | 
						|
  => create_worker+0xcc/0x1a0
 | 
						|
  => worker_thread+0x2ad/0x380
 | 
						|
  => kthread+0xe9/0x110
 | 
						|
  => ret_from_fork+0x2c/0x50
 | 
						|
 | 
						|
A synthetic event that has a stacktrace field may use it as a key in
 | 
						|
histogram::
 | 
						|
 | 
						|
  # echo 'hist:keys=delta.buckets=100,stack.stacktrace:sort=delta' > events/synthetic/block_lat/trigger
 | 
						|
  # cat events/synthetic/block_lat/hist
 | 
						|
 | 
						|
  # event histogram
 | 
						|
  #
 | 
						|
  # trigger info: hist:keys=delta.buckets=100,stack.stacktrace:vals=hitcount:sort=delta.buckets=100:size=2048 [active]
 | 
						|
  #
 | 
						|
  { delta: ~ 0-99, stack.stacktrace         __schedule+0xa19/0x1520
 | 
						|
         schedule+0x6b/0x110
 | 
						|
         io_schedule+0x46/0x80
 | 
						|
         bit_wait_io+0x11/0x80
 | 
						|
         __wait_on_bit+0x4e/0x120
 | 
						|
         out_of_line_wait_on_bit+0x8d/0xb0
 | 
						|
         __wait_on_buffer+0x33/0x40
 | 
						|
         jbd2_journal_commit_transaction+0x155a/0x19b0
 | 
						|
         kjournald2+0xab/0x270
 | 
						|
         kthread+0xfa/0x130
 | 
						|
         ret_from_fork+0x29/0x50
 | 
						|
  } hitcount:          1
 | 
						|
  { delta: ~ 0-99, stack.stacktrace         __schedule+0xa19/0x1520
 | 
						|
         schedule+0x6b/0x110
 | 
						|
         io_schedule+0x46/0x80
 | 
						|
         rq_qos_wait+0xd0/0x170
 | 
						|
         wbt_wait+0x9e/0xf0
 | 
						|
         __rq_qos_throttle+0x25/0x40
 | 
						|
         blk_mq_submit_bio+0x2c3/0x5b0
 | 
						|
         __submit_bio+0xff/0x190
 | 
						|
         submit_bio_noacct_nocheck+0x25b/0x2b0
 | 
						|
         submit_bio_noacct+0x20b/0x600
 | 
						|
         submit_bio+0x28/0x90
 | 
						|
         ext4_bio_write_page+0x1e0/0x8c0
 | 
						|
         mpage_submit_page+0x60/0x80
 | 
						|
         mpage_process_page_bufs+0x16c/0x180
 | 
						|
         mpage_prepare_extent_to_map+0x23f/0x530
 | 
						|
  } hitcount:          1
 | 
						|
  { delta: ~ 0-99, stack.stacktrace         __schedule+0xa19/0x1520
 | 
						|
         schedule+0x6b/0x110
 | 
						|
         schedule_hrtimeout_range_clock+0x97/0x110
 | 
						|
         schedule_hrtimeout_range+0x13/0x20
 | 
						|
         usleep_range_state+0x65/0x90
 | 
						|
         __intel_wait_for_register+0x1c1/0x230 [i915]
 | 
						|
         intel_psr_wait_for_idle_locked+0x171/0x2a0 [i915]
 | 
						|
         intel_pipe_update_start+0x169/0x360 [i915]
 | 
						|
         intel_update_crtc+0x112/0x490 [i915]
 | 
						|
         skl_commit_modeset_enables+0x199/0x600 [i915]
 | 
						|
         intel_atomic_commit_tail+0x7c4/0x1080 [i915]
 | 
						|
         intel_atomic_commit_work+0x12/0x20 [i915]
 | 
						|
         process_one_work+0x21c/0x3f0
 | 
						|
         worker_thread+0x50/0x3e0
 | 
						|
         kthread+0xfa/0x130
 | 
						|
  } hitcount:          3
 | 
						|
  { delta: ~ 0-99, stack.stacktrace         __schedule+0xa19/0x1520
 | 
						|
         schedule+0x6b/0x110
 | 
						|
         schedule_timeout+0x11e/0x160
 | 
						|
         __wait_for_common+0x8f/0x190
 | 
						|
         wait_for_completion+0x24/0x30
 | 
						|
         __flush_work.isra.0+0x1cc/0x360
 | 
						|
         flush_work+0xe/0x20
 | 
						|
         drm_mode_rmfb+0x18b/0x1d0 [drm]
 | 
						|
         drm_mode_rmfb_ioctl+0x10/0x20 [drm]
 | 
						|
         drm_ioctl_kernel+0xb8/0x150 [drm]
 | 
						|
         drm_ioctl+0x243/0x560 [drm]
 | 
						|
         __x64_sys_ioctl+0x92/0xd0
 | 
						|
         do_syscall_64+0x59/0x90
 | 
						|
         entry_SYSCALL_64_after_hwframe+0x72/0xdc
 | 
						|
  } hitcount:          1
 | 
						|
  { delta: ~ 0-99, stack.stacktrace         __schedule+0xa19/0x1520
 | 
						|
         schedule+0x6b/0x110
 | 
						|
         schedule_timeout+0x87/0x160
 | 
						|
         __wait_for_common+0x8f/0x190
 | 
						|
         wait_for_completion_timeout+0x1d/0x30
 | 
						|
         drm_atomic_helper_wait_for_flip_done+0x57/0x90 [drm_kms_helper]
 | 
						|
         intel_atomic_commit_tail+0x8ce/0x1080 [i915]
 | 
						|
         intel_atomic_commit_work+0x12/0x20 [i915]
 | 
						|
         process_one_work+0x21c/0x3f0
 | 
						|
         worker_thread+0x50/0x3e0
 | 
						|
         kthread+0xfa/0x130
 | 
						|
         ret_from_fork+0x29/0x50
 | 
						|
  } hitcount:          1
 | 
						|
  { delta: ~ 100-199, stack.stacktrace         __schedule+0xa19/0x1520
 | 
						|
         schedule+0x6b/0x110
 | 
						|
         schedule_hrtimeout_range_clock+0x97/0x110
 | 
						|
         schedule_hrtimeout_range+0x13/0x20
 | 
						|
         usleep_range_state+0x65/0x90
 | 
						|
         pci_set_low_power_state+0x17f/0x1f0
 | 
						|
         pci_set_power_state+0x49/0x250
 | 
						|
         pci_finish_runtime_suspend+0x4a/0x90
 | 
						|
         pci_pm_runtime_suspend+0xcb/0x1b0
 | 
						|
         __rpm_callback+0x48/0x120
 | 
						|
         rpm_callback+0x67/0x70
 | 
						|
         rpm_suspend+0x167/0x780
 | 
						|
         rpm_idle+0x25a/0x380
 | 
						|
         pm_runtime_work+0x93/0xc0
 | 
						|
         process_one_work+0x21c/0x3f0
 | 
						|
  } hitcount:          1
 | 
						|
 | 
						|
  Totals:
 | 
						|
    Hits: 10
 | 
						|
    Entries: 7
 | 
						|
    Dropped: 0
 | 
						|
 | 
						|
2.2.3 Hist trigger 'handlers' and 'actions'
 | 
						|
-------------------------------------------
 | 
						|
 | 
						|
A hist trigger 'action' is a function that's executed (in most cases
 | 
						|
conditionally) whenever a histogram entry is added or updated.
 | 
						|
 | 
						|
When a histogram entry is added or updated, a hist trigger 'handler'
 | 
						|
is what decides whether the corresponding action is actually invoked
 | 
						|
or not.
 | 
						|
 | 
						|
Hist trigger handlers and actions are paired together in the general
 | 
						|
form:
 | 
						|
 | 
						|
  <handler>.<action>
 | 
						|
 | 
						|
To specify a handler.action pair for a given event, simply specify
 | 
						|
that handler.action pair between colons in the hist trigger
 | 
						|
specification.
 | 
						|
 | 
						|
In theory, any handler can be combined with any action, but in
 | 
						|
practice, not every handler.action combination is currently supported;
 | 
						|
if a given handler.action combination isn't supported, the hist
 | 
						|
trigger will fail with -EINVAL;
 | 
						|
 | 
						|
The default 'handler.action' if none is explicitly specified is as it
 | 
						|
always has been, to simply update the set of values associated with an
 | 
						|
entry.  Some applications, however, may want to perform additional
 | 
						|
actions at that point, such as generate another event, or compare and
 | 
						|
save a maximum.
 | 
						|
 | 
						|
The supported handlers and actions are listed below, and each is
 | 
						|
described in more detail in the following paragraphs, in the context
 | 
						|
of descriptions of some common and useful handler.action combinations.
 | 
						|
 | 
						|
The available handlers are:
 | 
						|
 | 
						|
  - onmatch(matching.event)    - invoke action on any addition or update
 | 
						|
  - onmax(var)                 - invoke action if var exceeds current max
 | 
						|
  - onchange(var)              - invoke action if var changes
 | 
						|
 | 
						|
The available actions are:
 | 
						|
 | 
						|
  - trace(<synthetic_event_name>,param list)   - generate synthetic event
 | 
						|
  - save(field,...)                            - save current event fields
 | 
						|
  - snapshot()                                 - snapshot the trace buffer
 | 
						|
 | 
						|
The following commonly-used handler.action pairs are available:
 | 
						|
 | 
						|
  - onmatch(matching.event).trace(<synthetic_event_name>,param list)
 | 
						|
 | 
						|
    The 'onmatch(matching.event).trace(<synthetic_event_name>,param
 | 
						|
    list)' hist trigger action is invoked whenever an event matches
 | 
						|
    and the histogram entry would be added or updated.  It causes the
 | 
						|
    named synthetic event to be generated with the values given in the
 | 
						|
    'param list'.  The result is the generation of a synthetic event
 | 
						|
    that consists of the values contained in those variables at the
 | 
						|
    time the invoking event was hit.  For example, if the synthetic
 | 
						|
    event name is 'wakeup_latency', a wakeup_latency event is
 | 
						|
    generated using onmatch(event).trace(wakeup_latency,arg1,arg2).
 | 
						|
 | 
						|
    There is also an equivalent alternative form available for
 | 
						|
    generating synthetic events.  In this form, the synthetic event
 | 
						|
    name is used as if it were a function name.  For example, using
 | 
						|
    the 'wakeup_latency' synthetic event name again, the
 | 
						|
    wakeup_latency event would be generated by invoking it as if it
 | 
						|
    were a function call, with the event field values passed in as
 | 
						|
    arguments: onmatch(event).wakeup_latency(arg1,arg2).  The syntax
 | 
						|
    for this form is:
 | 
						|
 | 
						|
      onmatch(matching.event).<synthetic_event_name>(param list)
 | 
						|
 | 
						|
    In either case, the 'param list' consists of one or more
 | 
						|
    parameters which may be either variables or fields defined on
 | 
						|
    either the 'matching.event' or the target event.  The variables or
 | 
						|
    fields specified in the param list may be either fully-qualified
 | 
						|
    or unqualified.  If a variable is specified as unqualified, it
 | 
						|
    must be unique between the two events.  A field name used as a
 | 
						|
    param can be unqualified if it refers to the target event, but
 | 
						|
    must be fully qualified if it refers to the matching event.  A
 | 
						|
    fully-qualified name is of the form 'system.event_name.$var_name'
 | 
						|
    or 'system.event_name.field'.
 | 
						|
 | 
						|
    The 'matching.event' specification is simply the fully qualified
 | 
						|
    event name of the event that matches the target event for the
 | 
						|
    onmatch() functionality, in the form 'system.event_name'. Histogram
 | 
						|
    keys of both events are compared to find if events match. In case
 | 
						|
    multiple histogram keys are used, they all must match in the specified
 | 
						|
    order.
 | 
						|
 | 
						|
    Finally, the number and type of variables/fields in the 'param
 | 
						|
    list' must match the number and types of the fields in the
 | 
						|
    synthetic event being generated.
 | 
						|
 | 
						|
    As an example the below defines a simple synthetic event and uses
 | 
						|
    a variable defined on the sched_wakeup_new event as a parameter
 | 
						|
    when invoking the synthetic event.  Here we define the synthetic
 | 
						|
    event::
 | 
						|
 | 
						|
      # echo 'wakeup_new_test pid_t pid' >> \
 | 
						|
             /sys/kernel/tracing/synthetic_events
 | 
						|
 | 
						|
      # cat /sys/kernel/tracing/synthetic_events
 | 
						|
            wakeup_new_test pid_t pid
 | 
						|
 | 
						|
    The following hist trigger both defines the missing testpid
 | 
						|
    variable and specifies an onmatch() action that generates a
 | 
						|
    wakeup_new_test synthetic event whenever a sched_wakeup_new event
 | 
						|
    occurs, which because of the 'if comm == "cyclictest"' filter only
 | 
						|
    happens when the executable is cyclictest::
 | 
						|
 | 
						|
      # echo 'hist:keys=$testpid:testpid=pid:onmatch(sched.sched_wakeup_new).\
 | 
						|
              wakeup_new_test($testpid) if comm=="cyclictest"' >> \
 | 
						|
              /sys/kernel/tracing/events/sched/sched_wakeup_new/trigger
 | 
						|
 | 
						|
    Or, equivalently, using the 'trace' keyword syntax::
 | 
						|
 | 
						|
      # echo 'hist:keys=$testpid:testpid=pid:onmatch(sched.sched_wakeup_new).\
 | 
						|
              trace(wakeup_new_test,$testpid) if comm=="cyclictest"' >> \
 | 
						|
              /sys/kernel/tracing/events/sched/sched_wakeup_new/trigger
 | 
						|
 | 
						|
    Creating and displaying a histogram based on those events is now
 | 
						|
    just a matter of using the fields and new synthetic event in the
 | 
						|
    tracing/events/synthetic directory, as usual::
 | 
						|
 | 
						|
      # echo 'hist:keys=pid:sort=pid' >> \
 | 
						|
             /sys/kernel/tracing/events/synthetic/wakeup_new_test/trigger
 | 
						|
 | 
						|
    Running 'cyclictest' should cause wakeup_new events to generate
 | 
						|
    wakeup_new_test synthetic events which should result in histogram
 | 
						|
    output in the wakeup_new_test event's hist file::
 | 
						|
 | 
						|
      # cat /sys/kernel/tracing/events/synthetic/wakeup_new_test/hist
 | 
						|
 | 
						|
    A more typical usage would be to use two events to calculate a
 | 
						|
    latency.  The following example uses a set of hist triggers to
 | 
						|
    produce a 'wakeup_latency' histogram.
 | 
						|
 | 
						|
    First, we define a 'wakeup_latency' synthetic event::
 | 
						|
 | 
						|
      # echo 'wakeup_latency u64 lat; pid_t pid; int prio' >> \
 | 
						|
              /sys/kernel/tracing/synthetic_events
 | 
						|
 | 
						|
    Next, we specify that whenever we see a sched_waking event for a
 | 
						|
    cyclictest thread, save the timestamp in a 'ts0' variable::
 | 
						|
 | 
						|
      # echo 'hist:keys=$saved_pid:saved_pid=pid:ts0=common_timestamp.usecs \
 | 
						|
              if comm=="cyclictest"' >> \
 | 
						|
	      /sys/kernel/tracing/events/sched/sched_waking/trigger
 | 
						|
 | 
						|
    Then, when the corresponding thread is actually scheduled onto the
 | 
						|
    CPU by a sched_switch event (saved_pid matches next_pid), calculate
 | 
						|
    the latency and use that along with another variable and an event field
 | 
						|
    to generate a wakeup_latency synthetic event::
 | 
						|
 | 
						|
      # echo 'hist:keys=next_pid:wakeup_lat=common_timestamp.usecs-$ts0:\
 | 
						|
              onmatch(sched.sched_waking).wakeup_latency($wakeup_lat,\
 | 
						|
	              $saved_pid,next_prio) if next_comm=="cyclictest"' >> \
 | 
						|
	      /sys/kernel/tracing/events/sched/sched_switch/trigger
 | 
						|
 | 
						|
    We also need to create a histogram on the wakeup_latency synthetic
 | 
						|
    event in order to aggregate the generated synthetic event data::
 | 
						|
 | 
						|
      # echo 'hist:keys=pid,prio,lat:sort=pid,lat' >> \
 | 
						|
              /sys/kernel/tracing/events/synthetic/wakeup_latency/trigger
 | 
						|
 | 
						|
    Finally, once we've run cyclictest to actually generate some
 | 
						|
    events, we can see the output by looking at the wakeup_latency
 | 
						|
    synthetic event's hist file::
 | 
						|
 | 
						|
      # cat /sys/kernel/tracing/events/synthetic/wakeup_latency/hist
 | 
						|
 | 
						|
  - onmax(var).save(field,..	.)
 | 
						|
 | 
						|
    The 'onmax(var).save(field,...)' hist trigger action is invoked
 | 
						|
    whenever the value of 'var' associated with a histogram entry
 | 
						|
    exceeds the current maximum contained in that variable.
 | 
						|
 | 
						|
    The end result is that the trace event fields specified as the
 | 
						|
    onmax.save() params will be saved if 'var' exceeds the current
 | 
						|
    maximum for that hist trigger entry.  This allows context from the
 | 
						|
    event that exhibited the new maximum to be saved for later
 | 
						|
    reference.  When the histogram is displayed, additional fields
 | 
						|
    displaying the saved values will be printed.
 | 
						|
 | 
						|
    As an example the below defines a couple of hist triggers, one for
 | 
						|
    sched_waking and another for sched_switch, keyed on pid.  Whenever
 | 
						|
    a sched_waking occurs, the timestamp is saved in the entry
 | 
						|
    corresponding to the current pid, and when the scheduler switches
 | 
						|
    back to that pid, the timestamp difference is calculated.  If the
 | 
						|
    resulting latency, stored in wakeup_lat, exceeds the current
 | 
						|
    maximum latency, the values specified in the save() fields are
 | 
						|
    recorded::
 | 
						|
 | 
						|
      # echo 'hist:keys=pid:ts0=common_timestamp.usecs \
 | 
						|
              if comm=="cyclictest"' >> \
 | 
						|
              /sys/kernel/tracing/events/sched/sched_waking/trigger
 | 
						|
 | 
						|
      # echo 'hist:keys=next_pid:\
 | 
						|
              wakeup_lat=common_timestamp.usecs-$ts0:\
 | 
						|
              onmax($wakeup_lat).save(next_comm,prev_pid,prev_prio,prev_comm) \
 | 
						|
              if next_comm=="cyclictest"' >> \
 | 
						|
              /sys/kernel/tracing/events/sched/sched_switch/trigger
 | 
						|
 | 
						|
    When the histogram is displayed, the max value and the saved
 | 
						|
    values corresponding to the max are displayed following the rest
 | 
						|
    of the fields::
 | 
						|
 | 
						|
      # cat /sys/kernel/tracing/events/sched/sched_switch/hist
 | 
						|
        { next_pid:       2255 } hitcount:        239
 | 
						|
          common_timestamp-ts0:          0
 | 
						|
          max:         27
 | 
						|
	  next_comm: cyclictest
 | 
						|
          prev_pid:          0  prev_prio:        120  prev_comm: swapper/1
 | 
						|
 | 
						|
        { next_pid:       2256 } hitcount:       2355
 | 
						|
          common_timestamp-ts0: 0
 | 
						|
          max:         49  next_comm: cyclictest
 | 
						|
          prev_pid:          0  prev_prio:        120  prev_comm: swapper/0
 | 
						|
 | 
						|
        Totals:
 | 
						|
            Hits: 12970
 | 
						|
            Entries: 2
 | 
						|
            Dropped: 0
 | 
						|
 | 
						|
  - onmax(var).snapshot()
 | 
						|
 | 
						|
    The 'onmax(var).snapshot()' hist trigger action is invoked
 | 
						|
    whenever the value of 'var' associated with a histogram entry
 | 
						|
    exceeds the current maximum contained in that variable.
 | 
						|
 | 
						|
    The end result is that a global snapshot of the trace buffer will
 | 
						|
    be saved in the tracing/snapshot file if 'var' exceeds the current
 | 
						|
    maximum for any hist trigger entry.
 | 
						|
 | 
						|
    Note that in this case the maximum is a global maximum for the
 | 
						|
    current trace instance, which is the maximum across all buckets of
 | 
						|
    the histogram.  The key of the specific trace event that caused
 | 
						|
    the global maximum and the global maximum itself are displayed,
 | 
						|
    along with a message stating that a snapshot has been taken and
 | 
						|
    where to find it.  The user can use the key information displayed
 | 
						|
    to locate the corresponding bucket in the histogram for even more
 | 
						|
    detail.
 | 
						|
 | 
						|
    As an example the below defines a couple of hist triggers, one for
 | 
						|
    sched_waking and another for sched_switch, keyed on pid.  Whenever
 | 
						|
    a sched_waking event occurs, the timestamp is saved in the entry
 | 
						|
    corresponding to the current pid, and when the scheduler switches
 | 
						|
    back to that pid, the timestamp difference is calculated.  If the
 | 
						|
    resulting latency, stored in wakeup_lat, exceeds the current
 | 
						|
    maximum latency, a snapshot is taken.  As part of the setup, all
 | 
						|
    the scheduler events are also enabled, which are the events that
 | 
						|
    will show up in the snapshot when it is taken at some point::
 | 
						|
 | 
						|
      # echo 1 > /sys/kernel/tracing/events/sched/enable
 | 
						|
 | 
						|
      # echo 'hist:keys=pid:ts0=common_timestamp.usecs \
 | 
						|
              if comm=="cyclictest"' >> \
 | 
						|
              /sys/kernel/tracing/events/sched/sched_waking/trigger
 | 
						|
 | 
						|
      # echo 'hist:keys=next_pid:wakeup_lat=common_timestamp.usecs-$ts0: \
 | 
						|
              onmax($wakeup_lat).save(next_prio,next_comm,prev_pid,prev_prio, \
 | 
						|
	      prev_comm):onmax($wakeup_lat).snapshot() \
 | 
						|
	      if next_comm=="cyclictest"' >> \
 | 
						|
	      /sys/kernel/tracing/events/sched/sched_switch/trigger
 | 
						|
 | 
						|
    When the histogram is displayed, for each bucket the max value
 | 
						|
    and the saved values corresponding to the max are displayed
 | 
						|
    following the rest of the fields.
 | 
						|
 | 
						|
    If a snapshot was taken, there is also a message indicating that,
 | 
						|
    along with the value and event that triggered the global maximum::
 | 
						|
 | 
						|
      # cat /sys/kernel/tracing/events/sched/sched_switch/hist
 | 
						|
        { next_pid:       2101 } hitcount:        200
 | 
						|
	  max:         52  next_prio:        120  next_comm: cyclictest \
 | 
						|
          prev_pid:          0  prev_prio:        120  prev_comm: swapper/6
 | 
						|
 | 
						|
        { next_pid:       2103 } hitcount:       1326
 | 
						|
	  max:        572  next_prio:         19  next_comm: cyclictest \
 | 
						|
          prev_pid:          0  prev_prio:        120  prev_comm: swapper/1
 | 
						|
 | 
						|
        { next_pid:       2102 } hitcount:       1982 \
 | 
						|
	  max:         74  next_prio:         19  next_comm: cyclictest \
 | 
						|
          prev_pid:          0  prev_prio:        120  prev_comm: swapper/5
 | 
						|
 | 
						|
      Snapshot taken (see tracing/snapshot).  Details:
 | 
						|
	  triggering value { onmax($wakeup_lat) }:        572	\
 | 
						|
	  triggered by event with key: { next_pid:       2103 }
 | 
						|
 | 
						|
      Totals:
 | 
						|
          Hits: 3508
 | 
						|
          Entries: 3
 | 
						|
          Dropped: 0
 | 
						|
 | 
						|
    In the above case, the event that triggered the global maximum has
 | 
						|
    the key with next_pid == 2103.  If you look at the bucket that has
 | 
						|
    2103 as the key, you'll find the additional values save()'d along
 | 
						|
    with the local maximum for that bucket, which should be the same
 | 
						|
    as the global maximum (since that was the same value that
 | 
						|
    triggered the global snapshot).
 | 
						|
 | 
						|
    And finally, looking at the snapshot data should show at or near
 | 
						|
    the end the event that triggered the snapshot (in this case you
 | 
						|
    can verify the timestamps between the sched_waking and
 | 
						|
    sched_switch events, which should match the time displayed in the
 | 
						|
    global maximum)::
 | 
						|
 | 
						|
     # cat /sys/kernel/tracing/snapshot
 | 
						|
 | 
						|
         <...>-2103  [005] d..3   309.873125: sched_switch: prev_comm=cyclictest prev_pid=2103 prev_prio=19 prev_state=D ==> next_comm=swapper/5 next_pid=0 next_prio=120
 | 
						|
         <idle>-0     [005] d.h3   309.873611: sched_waking: comm=cyclictest pid=2102 prio=19 target_cpu=005
 | 
						|
         <idle>-0     [005] dNh4   309.873613: sched_wakeup: comm=cyclictest pid=2102 prio=19 target_cpu=005
 | 
						|
         <idle>-0     [005] d..3   309.873616: sched_switch: prev_comm=swapper/5 prev_pid=0 prev_prio=120 prev_state=S ==> next_comm=cyclictest next_pid=2102 next_prio=19
 | 
						|
         <...>-2102  [005] d..3   309.873625: sched_switch: prev_comm=cyclictest prev_pid=2102 prev_prio=19 prev_state=D ==> next_comm=swapper/5 next_pid=0 next_prio=120
 | 
						|
         <idle>-0     [005] d.h3   309.874624: sched_waking: comm=cyclictest pid=2102 prio=19 target_cpu=005
 | 
						|
         <idle>-0     [005] dNh4   309.874626: sched_wakeup: comm=cyclictest pid=2102 prio=19 target_cpu=005
 | 
						|
         <idle>-0     [005] dNh3   309.874628: sched_waking: comm=cyclictest pid=2103 prio=19 target_cpu=005
 | 
						|
         <idle>-0     [005] dNh4   309.874630: sched_wakeup: comm=cyclictest pid=2103 prio=19 target_cpu=005
 | 
						|
         <idle>-0     [005] d..3   309.874633: sched_switch: prev_comm=swapper/5 prev_pid=0 prev_prio=120 prev_state=S ==> next_comm=cyclictest next_pid=2102 next_prio=19
 | 
						|
         <idle>-0     [004] d.h3   309.874757: sched_waking: comm=gnome-terminal- pid=1699 prio=120 target_cpu=004
 | 
						|
         <idle>-0     [004] dNh4   309.874762: sched_wakeup: comm=gnome-terminal- pid=1699 prio=120 target_cpu=004
 | 
						|
         <idle>-0     [004] d..3   309.874766: sched_switch: prev_comm=swapper/4 prev_pid=0 prev_prio=120 prev_state=S ==> next_comm=gnome-terminal- next_pid=1699 next_prio=120
 | 
						|
     gnome-terminal--1699  [004] d.h2   309.874941: sched_stat_runtime: comm=gnome-terminal- pid=1699 runtime=180706 [ns] vruntime=1126870572 [ns]
 | 
						|
         <idle>-0     [003] d.s4   309.874956: sched_waking: comm=rcu_sched pid=9 prio=120 target_cpu=007
 | 
						|
         <idle>-0     [003] d.s5   309.874960: sched_wake_idle_without_ipi: cpu=7
 | 
						|
         <idle>-0     [003] d.s5   309.874961: sched_wakeup: comm=rcu_sched pid=9 prio=120 target_cpu=007
 | 
						|
         <idle>-0     [007] d..3   309.874963: sched_switch: prev_comm=swapper/7 prev_pid=0 prev_prio=120 prev_state=S ==> next_comm=rcu_sched next_pid=9 next_prio=120
 | 
						|
      rcu_sched-9     [007] d..3   309.874973: sched_stat_runtime: comm=rcu_sched pid=9 runtime=13646 [ns] vruntime=22531430286 [ns]
 | 
						|
      rcu_sched-9     [007] d..3   309.874978: sched_switch: prev_comm=rcu_sched prev_pid=9 prev_prio=120 prev_state=R+ ==> next_comm=swapper/7 next_pid=0 next_prio=120
 | 
						|
          <...>-2102  [005] d..4   309.874994: sched_migrate_task: comm=cyclictest pid=2103 prio=19 orig_cpu=5 dest_cpu=1
 | 
						|
          <...>-2102  [005] d..4   309.875185: sched_wake_idle_without_ipi: cpu=1
 | 
						|
         <idle>-0     [001] d..3   309.875200: sched_switch: prev_comm=swapper/1 prev_pid=0 prev_prio=120 prev_state=S ==> next_comm=cyclictest next_pid=2103 next_prio=19
 | 
						|
 | 
						|
  - onchange(var).save(field,..	.)
 | 
						|
 | 
						|
    The 'onchange(var).save(field,...)' hist trigger action is invoked
 | 
						|
    whenever the value of 'var' associated with a histogram entry
 | 
						|
    changes.
 | 
						|
 | 
						|
    The end result is that the trace event fields specified as the
 | 
						|
    onchange.save() params will be saved if 'var' changes for that
 | 
						|
    hist trigger entry.  This allows context from the event that
 | 
						|
    changed the value to be saved for later reference.  When the
 | 
						|
    histogram is displayed, additional fields displaying the saved
 | 
						|
    values will be printed.
 | 
						|
 | 
						|
  - onchange(var).snapshot()
 | 
						|
 | 
						|
    The 'onchange(var).snapshot()' hist trigger action is invoked
 | 
						|
    whenever the value of 'var' associated with a histogram entry
 | 
						|
    changes.
 | 
						|
 | 
						|
    The end result is that a global snapshot of the trace buffer will
 | 
						|
    be saved in the tracing/snapshot file if 'var' changes for any
 | 
						|
    hist trigger entry.
 | 
						|
 | 
						|
    Note that in this case the changed value is a global variable
 | 
						|
    associated with current trace instance.  The key of the specific
 | 
						|
    trace event that caused the value to change and the global value
 | 
						|
    itself are displayed, along with a message stating that a snapshot
 | 
						|
    has been taken and where to find it.  The user can use the key
 | 
						|
    information displayed to locate the corresponding bucket in the
 | 
						|
    histogram for even more detail.
 | 
						|
 | 
						|
    As an example the below defines a hist trigger on the tcp_probe
 | 
						|
    event, keyed on dport.  Whenever a tcp_probe event occurs, the
 | 
						|
    cwnd field is checked against the current value stored in the
 | 
						|
    $cwnd variable.  If the value has changed, a snapshot is taken.
 | 
						|
    As part of the setup, all the scheduler and tcp events are also
 | 
						|
    enabled, which are the events that will show up in the snapshot
 | 
						|
    when it is taken at some point::
 | 
						|
 | 
						|
      # echo 1 > /sys/kernel/tracing/events/sched/enable
 | 
						|
      # echo 1 > /sys/kernel/tracing/events/tcp/enable
 | 
						|
 | 
						|
      # echo 'hist:keys=dport:cwnd=snd_cwnd: \
 | 
						|
              onchange($cwnd).save(snd_wnd,srtt,rcv_wnd): \
 | 
						|
	      onchange($cwnd).snapshot()' >> \
 | 
						|
	      /sys/kernel/tracing/events/tcp/tcp_probe/trigger
 | 
						|
 | 
						|
    When the histogram is displayed, for each bucket the tracked value
 | 
						|
    and the saved values corresponding to that value are displayed
 | 
						|
    following the rest of the fields.
 | 
						|
 | 
						|
    If a snapshot was taken, there is also a message indicating that,
 | 
						|
    along with the value and event that triggered the snapshot::
 | 
						|
 | 
						|
      # cat /sys/kernel/tracing/events/tcp/tcp_probe/hist
 | 
						|
 | 
						|
      { dport:       1521 } hitcount:          8
 | 
						|
	changed:         10  snd_wnd:      35456  srtt:     154262  rcv_wnd:      42112
 | 
						|
 | 
						|
      { dport:         80 } hitcount:         23
 | 
						|
	changed:         10  snd_wnd:      28960  srtt:      19604  rcv_wnd:      29312
 | 
						|
 | 
						|
      { dport:       9001 } hitcount:        172
 | 
						|
	changed:         10  snd_wnd:      48384  srtt:     260444  rcv_wnd:      55168
 | 
						|
 | 
						|
      { dport:        443 } hitcount:        211
 | 
						|
	changed:         10  snd_wnd:      26960  srtt:      17379  rcv_wnd:      28800
 | 
						|
 | 
						|
      Snapshot taken (see tracing/snapshot).  Details:
 | 
						|
 | 
						|
          triggering value { onchange($cwnd) }:         10
 | 
						|
          triggered by event with key: { dport:         80 }
 | 
						|
 | 
						|
      Totals:
 | 
						|
          Hits: 414
 | 
						|
          Entries: 4
 | 
						|
          Dropped: 0
 | 
						|
 | 
						|
    In the above case, the event that triggered the snapshot has the
 | 
						|
    key with dport == 80.  If you look at the bucket that has 80 as
 | 
						|
    the key, you'll find the additional values save()'d along with the
 | 
						|
    changed value for that bucket, which should be the same as the
 | 
						|
    global changed value (since that was the same value that triggered
 | 
						|
    the global snapshot).
 | 
						|
 | 
						|
    And finally, looking at the snapshot data should show at or near
 | 
						|
    the end the event that triggered the snapshot::
 | 
						|
 | 
						|
      # cat /sys/kernel/tracing/snapshot
 | 
						|
 | 
						|
         gnome-shell-1261  [006] dN.3    49.823113: sched_stat_runtime: comm=gnome-shell pid=1261 runtime=49347 [ns] vruntime=1835730389 [ns]
 | 
						|
       kworker/u16:4-773   [003] d..3    49.823114: sched_switch: prev_comm=kworker/u16:4 prev_pid=773 prev_prio=120 prev_state=R+ ==> next_comm=kworker/3:2 next_pid=135 next_prio=120
 | 
						|
         gnome-shell-1261  [006] d..3    49.823114: sched_switch: prev_comm=gnome-shell prev_pid=1261 prev_prio=120 prev_state=R+ ==> next_comm=kworker/6:2 next_pid=387 next_prio=120
 | 
						|
         kworker/3:2-135   [003] d..3    49.823118: sched_stat_runtime: comm=kworker/3:2 pid=135 runtime=5339 [ns] vruntime=17815800388 [ns]
 | 
						|
         kworker/6:2-387   [006] d..3    49.823120: sched_stat_runtime: comm=kworker/6:2 pid=387 runtime=9594 [ns] vruntime=14589605367 [ns]
 | 
						|
         kworker/6:2-387   [006] d..3    49.823122: sched_switch: prev_comm=kworker/6:2 prev_pid=387 prev_prio=120 prev_state=R+ ==> next_comm=gnome-shell next_pid=1261 next_prio=120
 | 
						|
         kworker/3:2-135   [003] d..3    49.823123: sched_switch: prev_comm=kworker/3:2 prev_pid=135 prev_prio=120 prev_state=T ==> next_comm=swapper/3 next_pid=0 next_prio=120
 | 
						|
              <idle>-0     [004] ..s7    49.823798: tcp_probe: src=10.0.0.10:54326 dest=23.215.104.193:80 mark=0x0 length=32 snd_nxt=0xe3ae2ff5 snd_una=0xe3ae2ecd snd_cwnd=10 ssthresh=2147483647 snd_wnd=28960 srtt=19604 rcv_wnd=29312
 | 
						|
 | 
						|
3. User space creating a trigger
 | 
						|
--------------------------------
 | 
						|
 | 
						|
Writing into /sys/kernel/tracing/trace_marker writes into the ftrace
 | 
						|
ring buffer. This can also act like an event, by writing into the trigger
 | 
						|
file located in /sys/kernel/tracing/events/ftrace/print/
 | 
						|
 | 
						|
Modifying cyclictest to write into the trace_marker file before it sleeps
 | 
						|
and after it wakes up, something like this::
 | 
						|
 | 
						|
  static void traceputs(char *str)
 | 
						|
  {
 | 
						|
	/* tracemark_fd is the trace_marker file descriptor */
 | 
						|
	if (tracemark_fd < 0)
 | 
						|
		return;
 | 
						|
	/* write the tracemark message */
 | 
						|
	write(tracemark_fd, str, strlen(str));
 | 
						|
  }
 | 
						|
 | 
						|
And later add something like::
 | 
						|
 | 
						|
	traceputs("start");
 | 
						|
	clock_nanosleep(...);
 | 
						|
	traceputs("end");
 | 
						|
 | 
						|
We can make a histogram from this::
 | 
						|
 | 
						|
 # cd /sys/kernel/tracing
 | 
						|
 # echo 'latency u64 lat' > synthetic_events
 | 
						|
 # echo 'hist:keys=common_pid:ts0=common_timestamp.usecs if buf == "start"' > events/ftrace/print/trigger
 | 
						|
 # echo 'hist:keys=common_pid:lat=common_timestamp.usecs-$ts0:onmatch(ftrace.print).latency($lat) if buf == "end"' >> events/ftrace/print/trigger
 | 
						|
 # echo 'hist:keys=lat,common_pid:sort=lat' > events/synthetic/latency/trigger
 | 
						|
 | 
						|
The above created a synthetic event called "latency" and two histograms
 | 
						|
against the trace_marker, one gets triggered when "start" is written into the
 | 
						|
trace_marker file and the other when "end" is written. If the pids match, then
 | 
						|
it will call the "latency" synthetic event with the calculated latency as its
 | 
						|
parameter. Finally, a histogram is added to the latency synthetic event to
 | 
						|
record the calculated latency along with the pid.
 | 
						|
 | 
						|
Now running cyclictest with::
 | 
						|
 | 
						|
 # ./cyclictest -p80 -d0 -i250 -n -a -t --tracemark -b 1000
 | 
						|
 | 
						|
 -p80  : run threads at priority 80
 | 
						|
 -d0   : have all threads run at the same interval
 | 
						|
 -i250 : start the interval at 250 microseconds (all threads will do this)
 | 
						|
 -n    : sleep with nanosleep
 | 
						|
 -a    : affine all threads to a separate CPU
 | 
						|
 -t    : one thread per available CPU
 | 
						|
 --tracemark : enable trace mark writing
 | 
						|
 -b 1000 : stop if any latency is greater than 1000 microseconds
 | 
						|
 | 
						|
Note, the -b 1000 is used just to make --tracemark available.
 | 
						|
 | 
						|
Then we can see the histogram created by this with::
 | 
						|
 | 
						|
 # cat events/synthetic/latency/hist
 | 
						|
 # event histogram
 | 
						|
 #
 | 
						|
 # trigger info: hist:keys=lat,common_pid:vals=hitcount:sort=lat:size=2048 [active]
 | 
						|
 #
 | 
						|
 | 
						|
 { lat:        107, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        122, common_pid:       2041 } hitcount:          1
 | 
						|
 { lat:        166, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        174, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        194, common_pid:       2041 } hitcount:          1
 | 
						|
 { lat:        196, common_pid:       2036 } hitcount:          1
 | 
						|
 { lat:        197, common_pid:       2038 } hitcount:          1
 | 
						|
 { lat:        198, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        199, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        200, common_pid:       2041 } hitcount:          1
 | 
						|
 { lat:        201, common_pid:       2039 } hitcount:          2
 | 
						|
 { lat:        202, common_pid:       2038 } hitcount:          1
 | 
						|
 { lat:        202, common_pid:       2043 } hitcount:          1
 | 
						|
 { lat:        203, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        203, common_pid:       2036 } hitcount:          1
 | 
						|
 { lat:        203, common_pid:       2041 } hitcount:          1
 | 
						|
 { lat:        206, common_pid:       2038 } hitcount:          2
 | 
						|
 { lat:        207, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        207, common_pid:       2036 } hitcount:          1
 | 
						|
 { lat:        208, common_pid:       2040 } hitcount:          1
 | 
						|
 { lat:        209, common_pid:       2043 } hitcount:          1
 | 
						|
 { lat:        210, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        211, common_pid:       2039 } hitcount:          4
 | 
						|
 { lat:        212, common_pid:       2043 } hitcount:          1
 | 
						|
 { lat:        212, common_pid:       2039 } hitcount:          2
 | 
						|
 { lat:        213, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        214, common_pid:       2038 } hitcount:          1
 | 
						|
 { lat:        214, common_pid:       2039 } hitcount:          2
 | 
						|
 { lat:        214, common_pid:       2042 } hitcount:          1
 | 
						|
 { lat:        215, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        217, common_pid:       2036 } hitcount:          1
 | 
						|
 { lat:        217, common_pid:       2040 } hitcount:          1
 | 
						|
 { lat:        217, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        218, common_pid:       2039 } hitcount:          6
 | 
						|
 { lat:        219, common_pid:       2039 } hitcount:          9
 | 
						|
 { lat:        220, common_pid:       2039 } hitcount:         11
 | 
						|
 { lat:        221, common_pid:       2039 } hitcount:          5
 | 
						|
 { lat:        221, common_pid:       2042 } hitcount:          1
 | 
						|
 { lat:        222, common_pid:       2039 } hitcount:          7
 | 
						|
 { lat:        223, common_pid:       2036 } hitcount:          1
 | 
						|
 { lat:        223, common_pid:       2039 } hitcount:          3
 | 
						|
 { lat:        224, common_pid:       2039 } hitcount:          4
 | 
						|
 { lat:        224, common_pid:       2037 } hitcount:          1
 | 
						|
 { lat:        224, common_pid:       2036 } hitcount:          2
 | 
						|
 { lat:        225, common_pid:       2039 } hitcount:          5
 | 
						|
 { lat:        225, common_pid:       2042 } hitcount:          1
 | 
						|
 { lat:        226, common_pid:       2039 } hitcount:          7
 | 
						|
 { lat:        226, common_pid:       2036 } hitcount:          4
 | 
						|
 { lat:        227, common_pid:       2039 } hitcount:          6
 | 
						|
 { lat:        227, common_pid:       2036 } hitcount:         12
 | 
						|
 { lat:        227, common_pid:       2043 } hitcount:          1
 | 
						|
 { lat:        228, common_pid:       2039 } hitcount:          7
 | 
						|
 { lat:        228, common_pid:       2036 } hitcount:         14
 | 
						|
 { lat:        229, common_pid:       2039 } hitcount:          9
 | 
						|
 { lat:        229, common_pid:       2036 } hitcount:          8
 | 
						|
 { lat:        229, common_pid:       2038 } hitcount:          1
 | 
						|
 { lat:        230, common_pid:       2039 } hitcount:         11
 | 
						|
 { lat:        230, common_pid:       2036 } hitcount:          6
 | 
						|
 { lat:        230, common_pid:       2043 } hitcount:          1
 | 
						|
 { lat:        230, common_pid:       2042 } hitcount:          2
 | 
						|
 { lat:        231, common_pid:       2041 } hitcount:          1
 | 
						|
 { lat:        231, common_pid:       2036 } hitcount:          6
 | 
						|
 { lat:        231, common_pid:       2043 } hitcount:          1
 | 
						|
 { lat:        231, common_pid:       2039 } hitcount:          8
 | 
						|
 { lat:        232, common_pid:       2037 } hitcount:          1
 | 
						|
 { lat:        232, common_pid:       2039 } hitcount:          6
 | 
						|
 { lat:        232, common_pid:       2040 } hitcount:          2
 | 
						|
 { lat:        232, common_pid:       2036 } hitcount:          5
 | 
						|
 { lat:        232, common_pid:       2043 } hitcount:          1
 | 
						|
 { lat:        233, common_pid:       2036 } hitcount:          5
 | 
						|
 { lat:        233, common_pid:       2039 } hitcount:         11
 | 
						|
 { lat:        234, common_pid:       2039 } hitcount:          4
 | 
						|
 { lat:        234, common_pid:       2038 } hitcount:          2
 | 
						|
 { lat:        234, common_pid:       2043 } hitcount:          2
 | 
						|
 { lat:        234, common_pid:       2036 } hitcount:         11
 | 
						|
 { lat:        234, common_pid:       2040 } hitcount:          1
 | 
						|
 { lat:        235, common_pid:       2037 } hitcount:          2
 | 
						|
 { lat:        235, common_pid:       2036 } hitcount:          8
 | 
						|
 { lat:        235, common_pid:       2043 } hitcount:          2
 | 
						|
 { lat:        235, common_pid:       2039 } hitcount:          5
 | 
						|
 { lat:        235, common_pid:       2042 } hitcount:          2
 | 
						|
 { lat:        235, common_pid:       2040 } hitcount:          4
 | 
						|
 { lat:        235, common_pid:       2041 } hitcount:          1
 | 
						|
 { lat:        236, common_pid:       2036 } hitcount:          7
 | 
						|
 { lat:        236, common_pid:       2037 } hitcount:          1
 | 
						|
 { lat:        236, common_pid:       2041 } hitcount:          5
 | 
						|
 { lat:        236, common_pid:       2039 } hitcount:          3
 | 
						|
 { lat:        236, common_pid:       2043 } hitcount:          9
 | 
						|
 { lat:        236, common_pid:       2040 } hitcount:          7
 | 
						|
 { lat:        237, common_pid:       2037 } hitcount:          1
 | 
						|
 { lat:        237, common_pid:       2040 } hitcount:          1
 | 
						|
 { lat:        237, common_pid:       2036 } hitcount:          9
 | 
						|
 { lat:        237, common_pid:       2039 } hitcount:          3
 | 
						|
 { lat:        237, common_pid:       2043 } hitcount:          8
 | 
						|
 { lat:        237, common_pid:       2042 } hitcount:          2
 | 
						|
 { lat:        237, common_pid:       2041 } hitcount:          2
 | 
						|
 { lat:        238, common_pid:       2043 } hitcount:         10
 | 
						|
 { lat:        238, common_pid:       2040 } hitcount:          1
 | 
						|
 { lat:        238, common_pid:       2037 } hitcount:          9
 | 
						|
 { lat:        238, common_pid:       2038 } hitcount:          1
 | 
						|
 { lat:        238, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        238, common_pid:       2042 } hitcount:          3
 | 
						|
 { lat:        238, common_pid:       2036 } hitcount:          7
 | 
						|
 { lat:        239, common_pid:       2041 } hitcount:          1
 | 
						|
 { lat:        239, common_pid:       2043 } hitcount:         11
 | 
						|
 { lat:        239, common_pid:       2037 } hitcount:         11
 | 
						|
 { lat:        239, common_pid:       2038 } hitcount:          6
 | 
						|
 { lat:        239, common_pid:       2036 } hitcount:          7
 | 
						|
 { lat:        239, common_pid:       2040 } hitcount:          1
 | 
						|
 { lat:        239, common_pid:       2042 } hitcount:          9
 | 
						|
 { lat:        240, common_pid:       2037 } hitcount:         29
 | 
						|
 { lat:        240, common_pid:       2043 } hitcount:         15
 | 
						|
 { lat:        240, common_pid:       2040 } hitcount:         44
 | 
						|
 { lat:        240, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        240, common_pid:       2041 } hitcount:          2
 | 
						|
 { lat:        240, common_pid:       2038 } hitcount:          1
 | 
						|
 { lat:        240, common_pid:       2036 } hitcount:         10
 | 
						|
 { lat:        240, common_pid:       2042 } hitcount:         13
 | 
						|
 { lat:        241, common_pid:       2036 } hitcount:         21
 | 
						|
 { lat:        241, common_pid:       2041 } hitcount:         36
 | 
						|
 { lat:        241, common_pid:       2037 } hitcount:         34
 | 
						|
 { lat:        241, common_pid:       2042 } hitcount:         14
 | 
						|
 { lat:        241, common_pid:       2040 } hitcount:         94
 | 
						|
 { lat:        241, common_pid:       2039 } hitcount:         12
 | 
						|
 { lat:        241, common_pid:       2038 } hitcount:          2
 | 
						|
 { lat:        241, common_pid:       2043 } hitcount:         28
 | 
						|
 { lat:        242, common_pid:       2040 } hitcount:        109
 | 
						|
 { lat:        242, common_pid:       2041 } hitcount:        506
 | 
						|
 { lat:        242, common_pid:       2039 } hitcount:        155
 | 
						|
 { lat:        242, common_pid:       2042 } hitcount:         21
 | 
						|
 { lat:        242, common_pid:       2037 } hitcount:         52
 | 
						|
 { lat:        242, common_pid:       2043 } hitcount:         21
 | 
						|
 { lat:        242, common_pid:       2036 } hitcount:         16
 | 
						|
 { lat:        242, common_pid:       2038 } hitcount:        156
 | 
						|
 { lat:        243, common_pid:       2037 } hitcount:         46
 | 
						|
 { lat:        243, common_pid:       2039 } hitcount:         40
 | 
						|
 { lat:        243, common_pid:       2042 } hitcount:        119
 | 
						|
 { lat:        243, common_pid:       2041 } hitcount:        611
 | 
						|
 { lat:        243, common_pid:       2036 } hitcount:         69
 | 
						|
 { lat:        243, common_pid:       2038 } hitcount:        784
 | 
						|
 { lat:        243, common_pid:       2040 } hitcount:        323
 | 
						|
 { lat:        243, common_pid:       2043 } hitcount:         14
 | 
						|
 { lat:        244, common_pid:       2043 } hitcount:         35
 | 
						|
 { lat:        244, common_pid:       2042 } hitcount:        305
 | 
						|
 { lat:        244, common_pid:       2039 } hitcount:          8
 | 
						|
 { lat:        244, common_pid:       2040 } hitcount:       4515
 | 
						|
 { lat:        244, common_pid:       2038 } hitcount:        371
 | 
						|
 { lat:        244, common_pid:       2037 } hitcount:         31
 | 
						|
 { lat:        244, common_pid:       2036 } hitcount:        114
 | 
						|
 { lat:        244, common_pid:       2041 } hitcount:       3396
 | 
						|
 { lat:        245, common_pid:       2036 } hitcount:        700
 | 
						|
 { lat:        245, common_pid:       2041 } hitcount:       2772
 | 
						|
 { lat:        245, common_pid:       2037 } hitcount:        268
 | 
						|
 { lat:        245, common_pid:       2039 } hitcount:        472
 | 
						|
 { lat:        245, common_pid:       2038 } hitcount:       2758
 | 
						|
 { lat:        245, common_pid:       2042 } hitcount:       3833
 | 
						|
 { lat:        245, common_pid:       2040 } hitcount:       3105
 | 
						|
 { lat:        245, common_pid:       2043 } hitcount:        645
 | 
						|
 { lat:        246, common_pid:       2038 } hitcount:       3451
 | 
						|
 { lat:        246, common_pid:       2041 } hitcount:        142
 | 
						|
 { lat:        246, common_pid:       2037 } hitcount:       5101
 | 
						|
 { lat:        246, common_pid:       2040 } hitcount:         68
 | 
						|
 { lat:        246, common_pid:       2043 } hitcount:       5099
 | 
						|
 { lat:        246, common_pid:       2039 } hitcount:       5608
 | 
						|
 { lat:        246, common_pid:       2042 } hitcount:       3723
 | 
						|
 { lat:        246, common_pid:       2036 } hitcount:       4738
 | 
						|
 { lat:        247, common_pid:       2042 } hitcount:        312
 | 
						|
 { lat:        247, common_pid:       2043 } hitcount:       2385
 | 
						|
 { lat:        247, common_pid:       2041 } hitcount:        452
 | 
						|
 { lat:        247, common_pid:       2038 } hitcount:        792
 | 
						|
 { lat:        247, common_pid:       2040 } hitcount:         78
 | 
						|
 { lat:        247, common_pid:       2036 } hitcount:       2375
 | 
						|
 { lat:        247, common_pid:       2039 } hitcount:       1834
 | 
						|
 { lat:        247, common_pid:       2037 } hitcount:       2655
 | 
						|
 { lat:        248, common_pid:       2037 } hitcount:         36
 | 
						|
 { lat:        248, common_pid:       2042 } hitcount:         11
 | 
						|
 { lat:        248, common_pid:       2038 } hitcount:        122
 | 
						|
 { lat:        248, common_pid:       2036 } hitcount:        135
 | 
						|
 { lat:        248, common_pid:       2039 } hitcount:         26
 | 
						|
 { lat:        248, common_pid:       2041 } hitcount:        503
 | 
						|
 { lat:        248, common_pid:       2043 } hitcount:         66
 | 
						|
 { lat:        248, common_pid:       2040 } hitcount:         46
 | 
						|
 { lat:        249, common_pid:       2037 } hitcount:         29
 | 
						|
 { lat:        249, common_pid:       2038 } hitcount:          1
 | 
						|
 { lat:        249, common_pid:       2043 } hitcount:         29
 | 
						|
 { lat:        249, common_pid:       2039 } hitcount:          8
 | 
						|
 { lat:        249, common_pid:       2042 } hitcount:         56
 | 
						|
 { lat:        249, common_pid:       2040 } hitcount:         27
 | 
						|
 { lat:        249, common_pid:       2041 } hitcount:         11
 | 
						|
 { lat:        249, common_pid:       2036 } hitcount:         27
 | 
						|
 { lat:        250, common_pid:       2038 } hitcount:          1
 | 
						|
 { lat:        250, common_pid:       2036 } hitcount:         30
 | 
						|
 { lat:        250, common_pid:       2040 } hitcount:         19
 | 
						|
 { lat:        250, common_pid:       2043 } hitcount:         22
 | 
						|
 { lat:        250, common_pid:       2042 } hitcount:         20
 | 
						|
 { lat:        250, common_pid:       2041 } hitcount:          1
 | 
						|
 { lat:        250, common_pid:       2039 } hitcount:          6
 | 
						|
 { lat:        250, common_pid:       2037 } hitcount:         48
 | 
						|
 { lat:        251, common_pid:       2037 } hitcount:         43
 | 
						|
 { lat:        251, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        251, common_pid:       2036 } hitcount:         12
 | 
						|
 { lat:        251, common_pid:       2042 } hitcount:          2
 | 
						|
 { lat:        251, common_pid:       2041 } hitcount:          1
 | 
						|
 { lat:        251, common_pid:       2043 } hitcount:         15
 | 
						|
 { lat:        251, common_pid:       2040 } hitcount:          3
 | 
						|
 { lat:        252, common_pid:       2040 } hitcount:          1
 | 
						|
 { lat:        252, common_pid:       2036 } hitcount:         12
 | 
						|
 { lat:        252, common_pid:       2037 } hitcount:         21
 | 
						|
 { lat:        252, common_pid:       2043 } hitcount:         14
 | 
						|
 { lat:        253, common_pid:       2037 } hitcount:         21
 | 
						|
 { lat:        253, common_pid:       2039 } hitcount:          2
 | 
						|
 { lat:        253, common_pid:       2036 } hitcount:          9
 | 
						|
 { lat:        253, common_pid:       2043 } hitcount:          6
 | 
						|
 { lat:        253, common_pid:       2040 } hitcount:          1
 | 
						|
 { lat:        254, common_pid:       2036 } hitcount:          8
 | 
						|
 { lat:        254, common_pid:       2043 } hitcount:          3
 | 
						|
 { lat:        254, common_pid:       2041 } hitcount:          1
 | 
						|
 { lat:        254, common_pid:       2042 } hitcount:          1
 | 
						|
 { lat:        254, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        254, common_pid:       2037 } hitcount:         12
 | 
						|
 { lat:        255, common_pid:       2043 } hitcount:          1
 | 
						|
 { lat:        255, common_pid:       2037 } hitcount:          2
 | 
						|
 { lat:        255, common_pid:       2036 } hitcount:          2
 | 
						|
 { lat:        255, common_pid:       2039 } hitcount:          8
 | 
						|
 { lat:        256, common_pid:       2043 } hitcount:          1
 | 
						|
 { lat:        256, common_pid:       2036 } hitcount:          4
 | 
						|
 { lat:        256, common_pid:       2039 } hitcount:          6
 | 
						|
 { lat:        257, common_pid:       2039 } hitcount:          5
 | 
						|
 { lat:        257, common_pid:       2036 } hitcount:          4
 | 
						|
 { lat:        258, common_pid:       2039 } hitcount:          5
 | 
						|
 { lat:        258, common_pid:       2036 } hitcount:          2
 | 
						|
 { lat:        259, common_pid:       2036 } hitcount:          7
 | 
						|
 { lat:        259, common_pid:       2039 } hitcount:          7
 | 
						|
 { lat:        260, common_pid:       2036 } hitcount:          8
 | 
						|
 { lat:        260, common_pid:       2039 } hitcount:          6
 | 
						|
 { lat:        261, common_pid:       2036 } hitcount:          5
 | 
						|
 { lat:        261, common_pid:       2039 } hitcount:          7
 | 
						|
 { lat:        262, common_pid:       2039 } hitcount:          5
 | 
						|
 { lat:        262, common_pid:       2036 } hitcount:          5
 | 
						|
 { lat:        263, common_pid:       2039 } hitcount:          7
 | 
						|
 { lat:        263, common_pid:       2036 } hitcount:          7
 | 
						|
 { lat:        264, common_pid:       2039 } hitcount:          9
 | 
						|
 { lat:        264, common_pid:       2036 } hitcount:          9
 | 
						|
 { lat:        265, common_pid:       2036 } hitcount:          5
 | 
						|
 { lat:        265, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        266, common_pid:       2036 } hitcount:          1
 | 
						|
 { lat:        266, common_pid:       2039 } hitcount:          3
 | 
						|
 { lat:        267, common_pid:       2036 } hitcount:          1
 | 
						|
 { lat:        267, common_pid:       2039 } hitcount:          3
 | 
						|
 { lat:        268, common_pid:       2036 } hitcount:          1
 | 
						|
 { lat:        268, common_pid:       2039 } hitcount:          6
 | 
						|
 { lat:        269, common_pid:       2036 } hitcount:          1
 | 
						|
 { lat:        269, common_pid:       2043 } hitcount:          1
 | 
						|
 { lat:        269, common_pid:       2039 } hitcount:          2
 | 
						|
 { lat:        270, common_pid:       2040 } hitcount:          1
 | 
						|
 { lat:        270, common_pid:       2039 } hitcount:          6
 | 
						|
 { lat:        271, common_pid:       2041 } hitcount:          1
 | 
						|
 { lat:        271, common_pid:       2039 } hitcount:          5
 | 
						|
 { lat:        272, common_pid:       2039 } hitcount:         10
 | 
						|
 { lat:        273, common_pid:       2039 } hitcount:          8
 | 
						|
 { lat:        274, common_pid:       2039 } hitcount:          2
 | 
						|
 { lat:        275, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        276, common_pid:       2039 } hitcount:          2
 | 
						|
 { lat:        276, common_pid:       2037 } hitcount:          1
 | 
						|
 { lat:        276, common_pid:       2038 } hitcount:          1
 | 
						|
 { lat:        277, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        277, common_pid:       2042 } hitcount:          1
 | 
						|
 { lat:        278, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        279, common_pid:       2039 } hitcount:          4
 | 
						|
 { lat:        279, common_pid:       2043 } hitcount:          1
 | 
						|
 { lat:        280, common_pid:       2039 } hitcount:          3
 | 
						|
 { lat:        283, common_pid:       2036 } hitcount:          2
 | 
						|
 { lat:        284, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        284, common_pid:       2043 } hitcount:          1
 | 
						|
 { lat:        288, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        289, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        300, common_pid:       2039 } hitcount:          1
 | 
						|
 { lat:        384, common_pid:       2039 } hitcount:          1
 | 
						|
 | 
						|
 Totals:
 | 
						|
     Hits: 67625
 | 
						|
     Entries: 278
 | 
						|
     Dropped: 0
 | 
						|
 | 
						|
Note, the writes are around the sleep, so ideally they will all be of 250
 | 
						|
microseconds. If you are wondering how there are several that are under
 | 
						|
250 microseconds, that is because the way cyclictest works, is if one
 | 
						|
iteration comes in late, the next one will set the timer to wake up less that
 | 
						|
250. That is, if an iteration came in 50 microseconds late, the next wake up
 | 
						|
will be at 200 microseconds.
 | 
						|
 | 
						|
But this could easily be done in userspace. To make this even more
 | 
						|
interesting, we can mix the histogram between events that happened in the
 | 
						|
kernel with trace_marker::
 | 
						|
 | 
						|
 # cd /sys/kernel/tracing
 | 
						|
 # echo 'latency u64 lat' > synthetic_events
 | 
						|
 # echo 'hist:keys=pid:ts0=common_timestamp.usecs' > events/sched/sched_waking/trigger
 | 
						|
 # echo 'hist:keys=common_pid:lat=common_timestamp.usecs-$ts0:onmatch(sched.sched_waking).latency($lat) if buf == "end"' > events/ftrace/print/trigger
 | 
						|
 # echo 'hist:keys=lat,common_pid:sort=lat' > events/synthetic/latency/trigger
 | 
						|
 | 
						|
The difference this time is that instead of using the trace_marker to start
 | 
						|
the latency, the sched_waking event is used, matching the common_pid for the
 | 
						|
trace_marker write with the pid that is being woken by sched_waking.
 | 
						|
 | 
						|
After running cyclictest again with the same parameters, we now have::
 | 
						|
 | 
						|
 # cat events/synthetic/latency/hist
 | 
						|
 # event histogram
 | 
						|
 #
 | 
						|
 # trigger info: hist:keys=lat,common_pid:vals=hitcount:sort=lat:size=2048 [active]
 | 
						|
 #
 | 
						|
 | 
						|
 { lat:          7, common_pid:       2302 } hitcount:        640
 | 
						|
 { lat:          7, common_pid:       2299 } hitcount:         42
 | 
						|
 { lat:          7, common_pid:       2303 } hitcount:         18
 | 
						|
 { lat:          7, common_pid:       2305 } hitcount:        166
 | 
						|
 { lat:          7, common_pid:       2306 } hitcount:          1
 | 
						|
 { lat:          7, common_pid:       2301 } hitcount:         91
 | 
						|
 { lat:          7, common_pid:       2300 } hitcount:         17
 | 
						|
 { lat:          8, common_pid:       2303 } hitcount:       8296
 | 
						|
 { lat:          8, common_pid:       2304 } hitcount:       6864
 | 
						|
 { lat:          8, common_pid:       2305 } hitcount:       9464
 | 
						|
 { lat:          8, common_pid:       2301 } hitcount:       9213
 | 
						|
 { lat:          8, common_pid:       2306 } hitcount:       6246
 | 
						|
 { lat:          8, common_pid:       2302 } hitcount:       8797
 | 
						|
 { lat:          8, common_pid:       2299 } hitcount:       8771
 | 
						|
 { lat:          8, common_pid:       2300 } hitcount:       8119
 | 
						|
 { lat:          9, common_pid:       2305 } hitcount:       1519
 | 
						|
 { lat:          9, common_pid:       2299 } hitcount:       2346
 | 
						|
 { lat:          9, common_pid:       2303 } hitcount:       2841
 | 
						|
 { lat:          9, common_pid:       2301 } hitcount:       1846
 | 
						|
 { lat:          9, common_pid:       2304 } hitcount:       3861
 | 
						|
 { lat:          9, common_pid:       2302 } hitcount:       1210
 | 
						|
 { lat:          9, common_pid:       2300 } hitcount:       2762
 | 
						|
 { lat:          9, common_pid:       2306 } hitcount:       4247
 | 
						|
 { lat:         10, common_pid:       2299 } hitcount:         16
 | 
						|
 { lat:         10, common_pid:       2306 } hitcount:        333
 | 
						|
 { lat:         10, common_pid:       2303 } hitcount:         16
 | 
						|
 { lat:         10, common_pid:       2304 } hitcount:        168
 | 
						|
 { lat:         10, common_pid:       2302 } hitcount:        240
 | 
						|
 { lat:         10, common_pid:       2301 } hitcount:         28
 | 
						|
 { lat:         10, common_pid:       2300 } hitcount:         95
 | 
						|
 { lat:         10, common_pid:       2305 } hitcount:         18
 | 
						|
 { lat:         11, common_pid:       2303 } hitcount:          5
 | 
						|
 { lat:         11, common_pid:       2305 } hitcount:          8
 | 
						|
 { lat:         11, common_pid:       2306 } hitcount:        221
 | 
						|
 { lat:         11, common_pid:       2302 } hitcount:         76
 | 
						|
 { lat:         11, common_pid:       2304 } hitcount:         26
 | 
						|
 { lat:         11, common_pid:       2300 } hitcount:        125
 | 
						|
 { lat:         11, common_pid:       2299 } hitcount:          2
 | 
						|
 { lat:         12, common_pid:       2305 } hitcount:          3
 | 
						|
 { lat:         12, common_pid:       2300 } hitcount:          6
 | 
						|
 { lat:         12, common_pid:       2306 } hitcount:         90
 | 
						|
 { lat:         12, common_pid:       2302 } hitcount:          4
 | 
						|
 { lat:         12, common_pid:       2303 } hitcount:          1
 | 
						|
 { lat:         12, common_pid:       2304 } hitcount:        122
 | 
						|
 { lat:         13, common_pid:       2300 } hitcount:         12
 | 
						|
 { lat:         13, common_pid:       2301 } hitcount:          1
 | 
						|
 { lat:         13, common_pid:       2306 } hitcount:         32
 | 
						|
 { lat:         13, common_pid:       2302 } hitcount:          5
 | 
						|
 { lat:         13, common_pid:       2305 } hitcount:          1
 | 
						|
 { lat:         13, common_pid:       2303 } hitcount:          1
 | 
						|
 { lat:         13, common_pid:       2304 } hitcount:         61
 | 
						|
 { lat:         14, common_pid:       2303 } hitcount:          4
 | 
						|
 { lat:         14, common_pid:       2306 } hitcount:          5
 | 
						|
 { lat:         14, common_pid:       2305 } hitcount:          4
 | 
						|
 { lat:         14, common_pid:       2304 } hitcount:         62
 | 
						|
 { lat:         14, common_pid:       2302 } hitcount:         19
 | 
						|
 { lat:         14, common_pid:       2300 } hitcount:         33
 | 
						|
 { lat:         14, common_pid:       2299 } hitcount:          1
 | 
						|
 { lat:         14, common_pid:       2301 } hitcount:          4
 | 
						|
 { lat:         15, common_pid:       2305 } hitcount:          1
 | 
						|
 { lat:         15, common_pid:       2302 } hitcount:         25
 | 
						|
 { lat:         15, common_pid:       2300 } hitcount:         11
 | 
						|
 { lat:         15, common_pid:       2299 } hitcount:          5
 | 
						|
 { lat:         15, common_pid:       2301 } hitcount:          1
 | 
						|
 { lat:         15, common_pid:       2304 } hitcount:          8
 | 
						|
 { lat:         15, common_pid:       2303 } hitcount:          1
 | 
						|
 { lat:         15, common_pid:       2306 } hitcount:          6
 | 
						|
 { lat:         16, common_pid:       2302 } hitcount:         31
 | 
						|
 { lat:         16, common_pid:       2306 } hitcount:          3
 | 
						|
 { lat:         16, common_pid:       2300 } hitcount:          5
 | 
						|
 { lat:         17, common_pid:       2302 } hitcount:          6
 | 
						|
 { lat:         17, common_pid:       2303 } hitcount:          1
 | 
						|
 { lat:         18, common_pid:       2304 } hitcount:          1
 | 
						|
 { lat:         18, common_pid:       2302 } hitcount:          8
 | 
						|
 { lat:         18, common_pid:       2299 } hitcount:          1
 | 
						|
 { lat:         18, common_pid:       2301 } hitcount:          1
 | 
						|
 { lat:         19, common_pid:       2303 } hitcount:          4
 | 
						|
 { lat:         19, common_pid:       2304 } hitcount:          5
 | 
						|
 { lat:         19, common_pid:       2302 } hitcount:          4
 | 
						|
 { lat:         19, common_pid:       2299 } hitcount:          3
 | 
						|
 { lat:         19, common_pid:       2306 } hitcount:          1
 | 
						|
 { lat:         19, common_pid:       2300 } hitcount:          4
 | 
						|
 { lat:         19, common_pid:       2305 } hitcount:          5
 | 
						|
 { lat:         20, common_pid:       2299 } hitcount:          2
 | 
						|
 { lat:         20, common_pid:       2302 } hitcount:          3
 | 
						|
 { lat:         20, common_pid:       2305 } hitcount:          1
 | 
						|
 { lat:         20, common_pid:       2300 } hitcount:          2
 | 
						|
 { lat:         20, common_pid:       2301 } hitcount:          2
 | 
						|
 { lat:         20, common_pid:       2303 } hitcount:          3
 | 
						|
 { lat:         21, common_pid:       2305 } hitcount:          1
 | 
						|
 { lat:         21, common_pid:       2299 } hitcount:          5
 | 
						|
 { lat:         21, common_pid:       2303 } hitcount:          4
 | 
						|
 { lat:         21, common_pid:       2302 } hitcount:          7
 | 
						|
 { lat:         21, common_pid:       2300 } hitcount:          1
 | 
						|
 { lat:         21, common_pid:       2301 } hitcount:          5
 | 
						|
 { lat:         21, common_pid:       2304 } hitcount:          2
 | 
						|
 { lat:         22, common_pid:       2302 } hitcount:          5
 | 
						|
 { lat:         22, common_pid:       2303 } hitcount:          1
 | 
						|
 { lat:         22, common_pid:       2306 } hitcount:          3
 | 
						|
 { lat:         22, common_pid:       2301 } hitcount:          2
 | 
						|
 { lat:         22, common_pid:       2300 } hitcount:          1
 | 
						|
 { lat:         22, common_pid:       2299 } hitcount:          1
 | 
						|
 { lat:         22, common_pid:       2305 } hitcount:          1
 | 
						|
 { lat:         22, common_pid:       2304 } hitcount:          1
 | 
						|
 { lat:         23, common_pid:       2299 } hitcount:          1
 | 
						|
 { lat:         23, common_pid:       2306 } hitcount:          2
 | 
						|
 { lat:         23, common_pid:       2302 } hitcount:          6
 | 
						|
 { lat:         24, common_pid:       2302 } hitcount:          3
 | 
						|
 { lat:         24, common_pid:       2300 } hitcount:          1
 | 
						|
 { lat:         24, common_pid:       2306 } hitcount:          2
 | 
						|
 { lat:         24, common_pid:       2305 } hitcount:          1
 | 
						|
 { lat:         24, common_pid:       2299 } hitcount:          1
 | 
						|
 { lat:         25, common_pid:       2300 } hitcount:          1
 | 
						|
 { lat:         25, common_pid:       2302 } hitcount:          4
 | 
						|
 { lat:         26, common_pid:       2302 } hitcount:          2
 | 
						|
 { lat:         27, common_pid:       2305 } hitcount:          1
 | 
						|
 { lat:         27, common_pid:       2300 } hitcount:          1
 | 
						|
 { lat:         27, common_pid:       2302 } hitcount:          3
 | 
						|
 { lat:         28, common_pid:       2306 } hitcount:          1
 | 
						|
 { lat:         28, common_pid:       2302 } hitcount:          4
 | 
						|
 { lat:         29, common_pid:       2302 } hitcount:          1
 | 
						|
 { lat:         29, common_pid:       2300 } hitcount:          2
 | 
						|
 { lat:         29, common_pid:       2306 } hitcount:          1
 | 
						|
 { lat:         29, common_pid:       2304 } hitcount:          1
 | 
						|
 { lat:         30, common_pid:       2302 } hitcount:          4
 | 
						|
 { lat:         31, common_pid:       2302 } hitcount:          6
 | 
						|
 { lat:         32, common_pid:       2302 } hitcount:          1
 | 
						|
 { lat:         33, common_pid:       2299 } hitcount:          1
 | 
						|
 { lat:         33, common_pid:       2302 } hitcount:          3
 | 
						|
 { lat:         34, common_pid:       2302 } hitcount:          2
 | 
						|
 { lat:         35, common_pid:       2302 } hitcount:          1
 | 
						|
 { lat:         35, common_pid:       2304 } hitcount:          1
 | 
						|
 { lat:         36, common_pid:       2302 } hitcount:          4
 | 
						|
 { lat:         37, common_pid:       2302 } hitcount:          6
 | 
						|
 { lat:         38, common_pid:       2302 } hitcount:          2
 | 
						|
 { lat:         39, common_pid:       2302 } hitcount:          2
 | 
						|
 { lat:         39, common_pid:       2304 } hitcount:          1
 | 
						|
 { lat:         40, common_pid:       2304 } hitcount:          2
 | 
						|
 { lat:         40, common_pid:       2302 } hitcount:          5
 | 
						|
 { lat:         41, common_pid:       2304 } hitcount:          1
 | 
						|
 { lat:         41, common_pid:       2302 } hitcount:          8
 | 
						|
 { lat:         42, common_pid:       2302 } hitcount:          6
 | 
						|
 { lat:         42, common_pid:       2304 } hitcount:          1
 | 
						|
 { lat:         43, common_pid:       2302 } hitcount:          3
 | 
						|
 { lat:         43, common_pid:       2304 } hitcount:          4
 | 
						|
 { lat:         44, common_pid:       2302 } hitcount:          6
 | 
						|
 { lat:         45, common_pid:       2302 } hitcount:          5
 | 
						|
 { lat:         46, common_pid:       2302 } hitcount:          5
 | 
						|
 { lat:         47, common_pid:       2302 } hitcount:          7
 | 
						|
 { lat:         48, common_pid:       2301 } hitcount:          1
 | 
						|
 { lat:         48, common_pid:       2302 } hitcount:          9
 | 
						|
 { lat:         49, common_pid:       2302 } hitcount:          3
 | 
						|
 { lat:         50, common_pid:       2302 } hitcount:          1
 | 
						|
 { lat:         50, common_pid:       2301 } hitcount:          1
 | 
						|
 { lat:         51, common_pid:       2302 } hitcount:          2
 | 
						|
 { lat:         51, common_pid:       2301 } hitcount:          1
 | 
						|
 { lat:         61, common_pid:       2302 } hitcount:          1
 | 
						|
 { lat:        110, common_pid:       2302 } hitcount:          1
 | 
						|
 | 
						|
 Totals:
 | 
						|
     Hits: 89565
 | 
						|
     Entries: 158
 | 
						|
     Dropped: 0
 | 
						|
 | 
						|
This doesn't tell us any information about how late cyclictest may have
 | 
						|
woken up, but it does show us a nice histogram of how long it took from
 | 
						|
the time that cyclictest was woken to the time it made it into user space.
 |