# pmu-tools **Repository Path**: certus/pmu-tools ## Basic Information - **Project Name**: pmu-tools - **Description**: No description available - **Primary Language**: Unknown - **License**: GPL-2.0 - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2020-07-07 - **Last Updated**: 2020-12-18 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README ![pmu-tools](http://halobates.de/pmu-tools.png) ![Python linting and testing](https://github.com/andikleen/pmu-tools/workflows/Python%20linting%20and%20testing/badge.svg) ![jevents test](https://github.com/andikleen/pmu-tools/workflows/jevents%20test/badge.svg) pmu tools is a collection of tools for profile collection and performance analysis on Intel CPUs on top of [Linux perf](https://perf.wiki.kernel.org/index.php/Main_Page). This uses performance counters in the CPU. # Recent new features: * Updated to TMA 4.104: Fixed an error in misprediction-related and Power License metrics * toplev now supports the new fixed TMA metrics counters on Icelake. This requires the upcoming 5.9+ kernel. * toplev was updated to Ahmad Yasin's/Anton Hanna's TMA 4.1 New Metrics: - Re-arrange Retiring Level 2 into Light\_Operations & Heavy\_Operations. Light\_Operations replaces the previous Base (or "General Retirement") while Heavy\_Operations is a superset of the Microcode\_Sequencer node (that moves to Level 3) - Mixing\_Vectors: hints on a pitfall when intermixing the newer AVX* with legacy SSE* vectors, a tree node under Core Bound [SKL onwards] Key Enhancements & fixes - Tuning of Level 2 breakdown for Backend\_Bound, Frontend\_Bound (rollback FRONTEND\_RETIRED 2-events use) [SKL onwards] - Improved branch misprediction related metrics to leverage a new PerfMon event [ICL onwards] - Improved CORE\_CLKS & #Retire\_Slots-based metrics [ICL onwards] - Adjusted cost of all nodes using MEM\_LOAD*RETIRED.* in case of shadow L1 d-cache misses - renamed Frontend_ to FetchLatency/Bandwidth [all] - Additional documentation/details to aid automated parsing in ‘For Tool Developers’. - Other fixes including Thresholds, Tagging (e.g. $issueSnoops), Locate-with, Metric Group * toplev can now generate charts in xlsx files with the --xchart option. * interval-normalize has a new --normalize-cpu option to normalize multi CPU CSV files from toplev or perf * toplev now supports an experimental --drilldown option to automatically rerun with the nodes needed to analyze the bottleneck further. * toplev can now output combined output xlsx files with the --xlsx option. This requires installing the xlsxwriter library using 'pip install xlsxwriter' for the correct python version. This is still an experimental feature. * toplev now hides idle CPUs by default, unless the output is CSV. This is configurable using the new --idle-threshold argument. The default is hiding if less than 5% busy. * toplev now gives suggestions on how to get better sampling events, and also got better at selecting them in the first place. In addition it can now suggest how to get more details on the bottleneck by enabling its children nodes. * toplev can now add whole sub trees with --nodes 'nodename*'. Useful to drill down in specific areas while still keeping multiplexing overhead under control. With /NUM it is possible to to specify the max level nodes to add. * With -v, toplev now indicates nodes that are below threshold with < not ? to avoid confusion. Note this is a potentially breaking change for CSV output users. * pmu-tools is now generally python3 clean (but still runs with python2 by default). One exception is parser which would need to be ported to the python3 construct. * New tool utilized.py to remove idle CPUs from toplev output * toplev --import can now directly decompress xz and gz files. -o / --valcsv / --perf-output can now directly compress files if their names are specified with .xz or .gz. * toplev update to Ahmad Yasin's/Anton Hanna's TMA 4.0: New Models ICL: New model for IceLake client processor Note that running on Icelake with HyperThreading enabled requires updating the perf tool to a recent version that supports the "percore" notifier. New Metrics and Info groups - IpFLOP: Instructions per Floating Point (FP) Operation [BDW onwards] - New breakdown for Frontend\_Bandwidth per fetch unit: MITE, DSB & LSD - IO\_{Read|Write}\_BW: Average IO (network or disk) Bandwidth Use for {Reads|Writes} [server models] - LSD\_Coverage: Fraction of Uops delivered by the LSD (Loop Stream Detector; aka Loop Cache) - New Info group: Frontend that hosts LSD\_Coverage, DSB\_Coverage and IpBAClear Key Enhancements & fixes - Tuned/balanced Frontend\_Latency & Frontend\_Bandwidth (Bandwidth exposed as a very short FE latency) [SKL onwards] - Tuned/balanced Memory\_Bound & Core\_Bound in Backend\_Bound breakdown [SKL onwards] - Tuned L2\_Bound Node for high memory BW workloads [SKL onwards] - BpTB, IpL, IpS & IpB renamed to BpTkBranch, IpLoad, IpStore & IpBranch respectively (Inst\_Mix info metrics) - Backporting IpFarBranch to all pre SKL models - Renamed DRAM\_{Read\_Latency|Parallel\_Reads} to MEM\_{Read\_Latency|Parallel\_Reads} - Fixed Count Domain for (Load|Store|ALU)\_Op\_Utilization [SNB onwards] - Removed OCR L3 prefetch Remote HitM events [SKX,CLX] - Fixed descriptions for Ports\_Utilized\_{0|1|2|3m} - Fixed Pause latency [CLX] * toplev can now generate a script with --gen-script to collect toplev data on a different system. The generated data can be then imported with --import * toplev / event_download / ocperf have been ported to python3. They still work with python2, which is so far the default and used by the standard #! shebangs. But on systems that have no python2 they can be run with a python3 interpreter. This feature is still experimental, please report any regressions. * toplev now supports --per-core / --per-socket output in SMT mode, and also a --global mode. This also works with reprocessed data (from --perf-output / --import), so it is possible to slice a single collection. It is also possible to specify them at the same time to get separate summaries. With --split-output -o file the different aggregations are written to different files. * toplev update to Ahmad Yasin's/Anton Hanna's TMA 3.6: - {Load|Store}\_STLB\_(Hit|Miss): new metrics that breakdown DTLB\_{Load|Store} costs - L2\_Evictions\_(Silent|NonSilent)\_PKI: L2 (silent|non silent) evictions rate per Kilo instructios - IpFarBranch - Instructions per Far Branch - Renamed 0/1/2/3m\_Ports\_Utilized - DSB\_Switches is now available - Count Domain changes for multiple nodes. New threshold for IpTB ( Instructions per Taken Branches ) - Re-organized/renamed Metric Group (e.g. Frontend\_Bound => Frontend) * toplev now can run with the NMI watchdog enabled - This may reduce the need for being root to change this setting - It may still require kernel.perf\_event\_paranoid settings <1, unless --single-thread --user is used. Some functionality like uncore monitoring requires root or kernel.perf_event_paranoid < 0. * toplev now supports running in KVM guests - The guest needs to have the PMU enabled (e.g. -cpu host for qemu) - The guest should report the same CPU type as the host (also -cpu host), otherwise the current CPU needs to be overriden with FORCECPU=../EVENTMAP=.. - PEBS sampling, offcore response, and uncore monitoring are not supported # All features: ## Major tools/libraries * The "ocperf" wrapper to "perf" that provides a full core performance counter event list for common Intel CPUs. This allows to use all the Intel events, not just the builtin events of perf. Can be also used as a library from other python programs * The "toplev.py" tool to identify the micro-architectural bottleneck for a workload. This implements the [TopDown](https://sites.google.com/site/analysismethods/yasin-pubs) or [TopDown2](http://software.intel.com/en-us/articles/how-to-tune-applications-using-a-top-down-characterization-of-microarchitectural-issues) methodology. * The "ucevent" tool to manage and compute uncore performance events. Uncore is the part of the CPU that is not core. Supports many metrics for power management, IO, QPI (interconnect), caches, and others. ucevent automatically generates event descriptions for the perf uncore driver and pretty prints the output. It also supports computing higher level metrics derived from multiple events. * A library to resolve named intel events (like INST_RETIRED.ANY) to perf_event_attr ([jevents](http://halobates.de/jevents.html)) jevents also supports self profiling with Linux since Linux 3.3. Self profiling is a program monitoring its own execution, either for controlled benchmarking or to optimize itself. For self-profiling on older kernels you can use [simple-pmu](http://halobates.de/simple-pmu) * Support for Intel "offcore" events on older Linux systems where the kernel perf subsystem does not support them natively. Offcore events allow to categorize memory accesses that go outside the core. * Workarounds for [some issues](http://software.intel.com/en-us/articles/performance-monitoring-on-intel-xeon-processor-e5-family) with offcore events on Sandy Bridge EP (Intel Xeon E5 v1) This is automatically enabled for the respective events with ocperf, and also available as a standalone program or python library. * A variety of tools for plotting and post processing perf stat -I1000 -x, or toplev.py -I1000 -x, interval measurements. --- ## Experimental/minor tools: - An example program for address profiling on Nehalem and later Intel CPUs (addr) - Some utility programs to access pci space or msrs on the command line - A utility program to program the PMU directly from user space (pmumon.py) for counting. This is mainly useful for testing and experimental purposes. - A program to print the currently running events (event-rmap) - Support for analyzing the raw PEBS records with perf. - A pandas/scipy data model for perf.data analytics (work in progress) --- # Help wanted - The plotting tools could use a lot of improvements. Both tl-serve and tl-barplot. If you're good in python or JS plotting any help improving those would be appreciated. --- # Usage: Check out the repository. Run the tools from the directory you checked out (but it does not need to be the current directory) They automatically search for other modules and data files in the same directory the script was located in. You can set the PATH to include the repository to run the tools from other directories: export PATH=/path/to/pmu-tools:$PATH # What tool to use for what? You want to: - understand CPU bottlenecks on the high-level: use toplev. - display toplev output graphically: use tl-server or toplev --graph - know what CPU events to run, but want to use symbolic names: use ocperf. - measure interconnect/caches/memory/power management on Xeon E5+: use ucevent - Use perf events from a C program: use jevents - Query CPU topology or disable HyperThreading: use cputop - Change Model Specific Registers: use msr - Change PCI config space: use PCI The other tools are for more obscure usages. # Dependencies All tools (except for parser/) should work with a python 2.7 standard installation. All need a reasonably recent perf (RHEL5 is too old) ocperf.py should work with python 2.6, or likely 2.5 when the json module is installed. msr.py will also work with 2.6 if argparse is installed (it is enough to copy those files from a python 2.7 installation) Except for the modules in parser/ there are no special dependencies outside a standard python install on a recent Linux system with perf. parser/ needs a scipy stack with pandas and pyelftools. The perf tool should not be too old. toplev has kernel dependencies, please see https://github.com/andikleen/pmu-tools/wiki/toplev-kernel-support simple-pebs and pebs-grabber require a Linux kernel source tree to build. They may not build on some older versions of Linux (patches welcome) # Tools ## ocperf: ocperf is a wrapper to "perf" that provides a full core event list for common Intel CPUs. This allows to use all the Intel defined events, not just the builtin events of perf. A more detailed introduction is in [Andi's blog](http://halobates.de/blog/p/245) Ahmad Yasin's [toplev/TopDown tutorial](http://www.cs.technion.ac.il/~erangi/TMA_using_Linux_perf__Ahmad_Yasin.pdf) ocperf.py list List all the events perf and ocperf supports on the current CPU ocperf.py stat -e eventname ... ocperf.py record -c default -e eventname ... ocperf.py report When an older kernel is used with offcore events (events that count types of memory accesses outside the CPU core) that does not support offcore events natively, ocperf has to run as root and only one such profiling can be active on a machine. When "-c default" is specified for record, the default sampling overflow value will be filled in for the sampling period. This option needs to be specified before the events and is not supported for all CPUs. By default perf uses a dynamic sampling period, which can cause varying (and sometimes large) overhead. The fixed period minimizes this problem. If you have trouble with one of the many acronyms in the event list descriptions, the [Intel optimization manual](http://www.intel.com/content/www/us/en/architecture-and-technology/64-ia-32-architectures-optimization-manual.html) describes many of them. ### ocperf API ocperf.py can be also used as a python module to convert or list events for the current CPU: ```python import ocperf emap = ocperf.find_emap() if not emap: sys.exit("Unknown CPU or cannot find event table") ev = emap.getevent("BR_MISP_EXEC.ANY") if ev: print "name:", ev.output() print "raw form:", ev.output(use_raw=True) print "description:, ev.desc ``` To retrieve data for other CPUs set the EVENTMAP environment variable to the csv file of the CPU before calling find\_emap() ### changing MSRs and PCI config space The msr.py, pci.py, latego.py can be used as standalone programs or python modules to change MSRs, PCI config space or enable/disable the [workarounds](http://software.intel.com/en-us/articles/performance-monitoring-on-intel-xeon-processor-e5-family). For example to set the MSR 0x123 on all CPUs to value 1 use: $ sudo ./msr.py 0x123 1 To read MSR 0x123 on CPU 0 $ sudo ./msr.py 0x123 To read MSR 0x123 on CPU 3: $ sudo python >>> import msr >>> msr.readmsr(0x123, 3) To set bit 0 in MSR 0x123 on all CPUs: $ sudo python >>> import msr >>> msr.writemsr(0x123, msr.readmsr(0x123) | 1) (this assumes the MSR has the same value on all CPUs, otherwise iterate the readmsr over the CPUs) ## toplev.py: Identify the micro-architectural bottleneck of a workload. The bottlenecks are expressed as a tree with different levels (max 5). Each bottleneck is only meaningful if the parent higher level crossed the threshold (it acts similar to a binary search). The tool automatically only prints meaningful ratios, unless -v is specified. This follows the "Top Down" methodology. The best description of the method is in the "A top-down method for performance analysis and counter architecture" paper (ISPASS 2014, available [here](https://sites.google.com/site/analysismethods/yasin-pubs)) I didn't invent it, I'm just implementing it. A more gentle introduction is in [andi's blog](http://halobates.de/blog/p/262) Please also see the [manual and tutorial](https://github.com/andikleen/pmu-tools/wiki/toplev-manual) toplev.py only supports counting, that is it cannot tell you where in the program the problem occurred, just what happened. There is now an experimental --show-sample option to suggest sampling events for specific problems. The new --run-sample option can also automatically sample the program by re-running. Requires Intel CPUs Sandy Bridge (Core 2nd gen, Xeon 5xxx) or newer or Atom Silvermont or newer. Quark or Xeon Phi are not supported. By default the simple high level model is used. The detailed model is selected with -lX, with X being the level. On non-SMT systems only the program is measured by default, while with SMT on the whole system is measured. [IVB model](http://halobates.de/ivb-hierarchy.svg) [Simple model](http://halobates.de/simple-hierarchy.svg) ## Examples: toplev.py -l2 program measure whole system in level 2 while program is running toplev.py -l1 --single-thread program measure single threaded program. system must be idle. toplev.py -l3 --no-desc -I 100 -x, sleep X measure whole system for X seconds every 100ms, outputting in CSV format. toplev.py --all --core C0 taskset -c 0,1 program Measure program running on core 0 with all nodes and metrics enables ## Options: General operation: --interval INTERVAL, -I INTERVAL Measure every ms instead of only once --no-multiplex Do not multiplex, but run the workload multiple times as needed. Requires reproducible workloads. --single-thread, -S Measure workload as single thread. Workload must run single threaded. In SMT mode other thread must be idle. --fast, -F Skip sanity checks to optimize CPU consumption --import _IMPORT Import specified perf stat output file instead of running perf. Must be for same cpu, same arguments, same /proc/cpuinfo, same topology, unless overriden --gen-script Generate script to collect perfmon information for --import later --drilldown Automatically rerun to get more details on bottleneck Measurement filtering: --kernel Only measure kernel code --user Only measure user code --core CORE Limit output to cores. Comma list of Sx-Cx-Tx. All parts optional. --no-aggr, -A Measure every CPU Select events: --level LEVEL, -l LEVEL Measure upto level N (max 6) --metrics, -m Print extra metrics --sw Measure perf Linux metrics --no-util Do not measure CPU utilization --tsx Measure TSX metrics --all Measure everything available --frequency Measure frequency --power Display power metrics --nodes NODES Include or exclude nodes (with + to add, -|^ to remove, comma separated list, wildcards allowed, add * to include all children/siblings, add /level to specify highest level node to match, add ^ to match related siblings and metrics, start with ! to only include specified nodes) --reduced Use reduced server subset of nodes/metrics --metric-group METRIC_GROUP Add (+) or remove (-|^) metric groups of metrics, comma separated list from --list-metric-groups. Query nodes: --list-metrics List all metrics --list-nodes List all nodes --list-metric-groups List metric groups --list-all List every supported node/metric/metricgroup Workarounds: --no-group Dont use groups --force-events Assume kernel supports all events. May give wrong results. --ignore-errata Do not disable events with errata --handle-errata Disable events with errata Output: --per-core Aggregate output per core --per-socket Aggregate output per socket --per-thread Aggregate output per CPU thread --global Aggregate output for all CPUs --no-desc Do not print event descriptions --desc Force event descriptions --verbose, -v Print all results even when below threshold or exceeding boundaries. Note this can result in bogus values, as the TopDown methodology relies on thresholds to correctly characterize workloads. Values not crossing threshold are marked with <. --csv CSV, -x CSV Enable CSV mode with specified delimeter --output OUTPUT, -o OUTPUT Set output file --split-output Generate multiple output files, one for each specified aggregation option (with -o) --graph Automatically graph interval output with tl-barplot.py --graph-cpu GRAPH_CPU CPU to graph using --graph --title TITLE Set title of graph --quiet Avoid unnecessary status output --long-desc Print long descriptions instead of abbreviated ones. --columns Print CPU output in multiple columns for each node --summary Print summary at the end. Only useful with -I --no-area Hide area column --perf-output PERF_OUTPUT Save perf stat output in specified file --no-perf Don't print perf command line --print Only print perf command line. Don't run --idle-threshold IDLE_THRESHOLD Hide idle CPUs (default <5% of busiest if not CSV, specify percent) Environment: --force-cpu {snb,jkt,ivb,ivt,hsw,hsx,slm,bdw,bdx,simple,skl,knl,skx,clx,icl} Force CPU type --force-topology findsysoutput Use specified topology file (find /sys/devices) --force-cpuinfo cpuinfo Use specified cpuinfo file (/proc/cpuinfo) --force-hypervisor Assume running under hypervisor (no uncore, no offcore, no PEBS) --no-uncore Disable uncore events --no-check Do not check that PMU units exist Additional information: --print-group, -g Print event group assignments --raw Print raw values --valcsv VALCSV, -V VALCSV Write raw counter values into CSV file --stats Show statistics on what events counted xlsx output: --xlsx XLSX Generate xlsx spreadsheet output with data for socket/global/thread/core/summary/raw views with 1s interval. Add --single-thread to only get program output. --xnormalize Add extra sheets with normalized data in xlsx files --xchart Chart data in xlsx files --xkeep Keep temporary CSV files Sampling: --show-sample Show command line to rerun workload with sampling --run-sample Automatically rerun workload with sampling --sample-args SAMPLE_ARGS Extra rguments to pass to perf record for sampling. Use + to specify - --sample-repeat SAMPLE_REPEAT Repeat measurement and sampling N times. This interleaves counting and sampling. Useful for background collection with -a sleep X. --sample-basename SAMPLE_BASENAME Base name of sample perf.data files Other perf arguments allowed (see the perf documentation) After -- perf arguments conflicting with toplev can be used. ## Some caveats: toplev defaults to measuring the full system and show data for all CPUs. Use taskset to limit the workload to known CPUs if needed. In some cases (idle system, single threaded workload) --single-thread can also be used. The lower levels of the measurement tree are less reliable than the higher levels. They also rely on counter multi-plexing, and can not run each equation in a single group, which can cause larger measurement errors with non steady state workloads (If you don't understand this terminology; it means measurements in higher levels are less accurate and it works best with programs that primarily do the same thing over and over) If the program is very reproducible -- such as a simple kernel -- it is also possible to use --no-multiplex. In this case the workload is rerun multiple times until all data is collected. Do not use together with sleep. toplev needs a new enough perf tool and has specific requirements on the kernel. See http://github.com/andikleen/pmu-tools/wiki/toplev-kernel-support Other CPUs can be forced with FORCECPU=name This usually requires setting the correct event map with EVENTMAP=... ## ucevent uncore monitoring Please see the [ucevent documentation](http://github.com/andikleen/pmu-tools/tree/master/ucevent/#ucevent-uncore-monitoring) ## interval-plot: interval-plot.py can plot the output of perf stat -I1000 -x Requires matplotlib to be installed. Below is the level 2 toplev measurement of a Linux kernel compile. Note that tl-barplot below is normally better to plot toplev output. ![plot-example](http://halobates.de/interval.plot.l2.png) ## interval-normalize: This converts the output of perf stat -Ixxx -x, / toplev.py -Ixxx -x, to a normalized output (one column for each event). This allows easier plotting and processing with other tools (spreadsheets, R, JMP, gnuplot etc.) ## plot-normalized: Plot an already normalized data file. Requires pyplot to be installed. ## tl-barplot: Plot output from toplev.py -I 1000 -v -x, --output file.csv -lLEVELS toplev outputs percentages, so it's better to use a stacked plot, instead of the absolute values interval-plot displays. tl-barplot implements a stacked barplot plot. Requires matplotlib to be installed. toplev.py --single-thread -l3 --title "GNU grep" --graph grep -r foo /usr/* ![tl-barplot-example](http://halobates.de/grep.3.svg) This assumes the workload is single threaded. tl-barplot can only display a single CPU, if --single-thread is not appropriate then the CPU to plot needs to be specified with --graph-cpu. With a new enough matplotlib you can also enable xkcd mode (install Humor Sans first) ![tl-barplot-xkcd](http://halobates.de/povray.2.png) ## tl-serve: Display toplev.py output in a web browser. Download dygraphs. Only needs to be done once. wget http://dygraphs.com/1.0.1/dygraph-combined.js Run toplev: toplev.py --all -I 100 -o x.csv ... tl-serve.py x.csv Then browse http://localhost:9001/ in your web browser. ## cputop query cpu topology and print all matching cpu numbers cputop "query" ["format"] query is a python expression, using variables: socket, core, thread or "offline" to query all offline cpus format is a printf format with %d. %d will be replaced with the cpu number. format can be offline to offline the cpu or online to online ### Examples: Print all cores on socket 0 cputop "socket == 0" Print all first threads in each core on socket 0 cputop "thread == 0 and socket == 0" Disable all second threads (disable hyper threading) cputop "thread == 1" offline Reenable all second threads (reenable hyper threading) cputop "thread == 0" online ## pebs-grabber: Older perf doesn't export the raw PEBS output, which contains a lot of useful information. PEBS is a sampling format generated by Intel CPUs for some events. pebs-grabber grabs PEBS data from perf. This assumes the perf pebs handler is running, we just also do trace points with the raw data. May need some minor tweaks as kernel interface change, and will also not likely work on very old kernels. This will create two new trace points trace_pebs_v1 and trace_pebs_v2 that log the complete PEBS record. When the CPU supports PEBSv2 (Haswell) the additional fields will be logged in pebs_v2. make [KDIR=/my/kernel/build/dir] insmod pebs-grabber.ko # needs to record as root perf record -e cycles:p,pebs_v1,pebs\_v2 [command, -a for all etc.] perf report perf script to display pebs data # alternatively trace-cmd and kernelshark can be also used to dump # the pebs data See http://download.intel.com/products/processor/manual/253669.pdf 18.10.2 for a description of the PEBS fields. Note this doesn't work with standard FC18 kernels, as they broke trace points in modules. It works with later and earlier kernels. ## event-rmap event-rmap [cpu] prints the currently running events. This provides an easier answer to question Q2j in Vince Weaver's perf events FAQ. # Testing When modifying toplev please run tl-tester. For ocperf run tester. For jevents run jevents/tester. other-tester tests other random tools. The all-tester script runs all test suites. # Mailing list Please post to the linux-perf-users@vger.kernel.org mailing list. For bugs please open an issue on https://github.com/andikleen/pmu-tools/issues # Licenses ocperf, toplev, ucevent, parser are under GPLv2, jevents is under the modified BSD license. Andi Kleen pmu-tools@halobates.de