Commit 5814bc2d authored by Linus Torvalds's avatar Linus Torvalds

Merge tag 'perf-tools-2020-12-24' of git://git.kernel.org/pub/scm/linux/kernel/git/acme/linux

Pull more perf tools updates from Arnaldo Carvalho de Melo:

 - Refactor 'perf stat' per CPU/socket/die/thread aggregation fixing use
   cases in ARM machines.

 - Fix memory leak when synthesizing SDT probes in 'perf probe'.

 - Update kernel header copies related to KVM, epol_pwait. msr-index and
   powerpc and s390 syscall tables.

* tag 'perf-tools-2020-12-24' of git://git.kernel.org/pub/scm/linux/kernel/git/acme/linux: (24 commits)
  perf probe: Fix memory leak when synthesizing SDT probes
  perf stat aggregation: Add separate thread member
  perf stat aggregation: Add separate core member
  perf stat aggregation: Add separate die member
  perf stat aggregation: Add separate socket member
  perf stat aggregation: Add separate node member
  perf stat aggregation: Start using cpu_aggr_id in map
  perf cpumap: Drop in cpu_aggr_map struct
  perf cpumap: Add new map type for aggregation
  perf stat: Replace aggregation ID with a struct
  perf cpumap: Add new struct for cpu aggregation
  perf cpumap: Use existing allocator to avoid using malloc
  perf tests: Improve topology test to check all aggregation types
  perf tools: Update s390's syscall.tbl copy from the kernel sources
  perf tools: Update powerpc's syscall.tbl copy from the kernel sources
  perf s390: Move syscall.tbl check into check-headers.sh
  perf powerpc: Move syscall.tbl check to check-headers.sh
  tools headers UAPI: Synch KVM's svm.h header with the kernel
  tools kvm headers: Update KVM headers from the kernel sources
  tools headers UAPI: Sync KVM's vmx.h header with the kernel sources
  ...
parents 42dc45e8 5149303f
...@@ -156,9 +156,6 @@ struct kvm_sync_regs { ...@@ -156,9 +156,6 @@ struct kvm_sync_regs {
__u64 device_irq_level; __u64 device_irq_level;
}; };
struct kvm_arch_memory_slot {
};
/* /*
* PMU filter structure. Describe a range of events with a particular * PMU filter structure. Describe a range of events with a particular
* action. To be used with KVM_ARM_VCPU_PMU_V3_FILTER. * action. To be used with KVM_ARM_VCPU_PMU_V3_FILTER.
......
...@@ -237,6 +237,7 @@ ...@@ -237,6 +237,7 @@
#define X86_FEATURE_VMCALL ( 8*32+18) /* "" Hypervisor supports the VMCALL instruction */ #define X86_FEATURE_VMCALL ( 8*32+18) /* "" Hypervisor supports the VMCALL instruction */
#define X86_FEATURE_VMW_VMMCALL ( 8*32+19) /* "" VMware prefers VMMCALL hypercall instruction */ #define X86_FEATURE_VMW_VMMCALL ( 8*32+19) /* "" VMware prefers VMMCALL hypercall instruction */
#define X86_FEATURE_SEV_ES ( 8*32+20) /* AMD Secure Encrypted Virtualization - Encrypted State */ #define X86_FEATURE_SEV_ES ( 8*32+20) /* AMD Secure Encrypted Virtualization - Encrypted State */
#define X86_FEATURE_VM_PAGE_FLUSH ( 8*32+21) /* "" VM Page Flush MSR is supported */
/* Intel-defined CPU features, CPUID level 0x00000007:0 (EBX), word 9 */ /* Intel-defined CPU features, CPUID level 0x00000007:0 (EBX), word 9 */
#define X86_FEATURE_FSGSBASE ( 9*32+ 0) /* RDFSBASE, WRFSBASE, RDGSBASE, WRGSBASE instructions*/ #define X86_FEATURE_FSGSBASE ( 9*32+ 0) /* RDFSBASE, WRFSBASE, RDGSBASE, WRGSBASE instructions*/
...@@ -376,6 +377,7 @@ ...@@ -376,6 +377,7 @@
#define X86_FEATURE_TSXLDTRK (18*32+16) /* TSX Suspend Load Address Tracking */ #define X86_FEATURE_TSXLDTRK (18*32+16) /* TSX Suspend Load Address Tracking */
#define X86_FEATURE_PCONFIG (18*32+18) /* Intel PCONFIG */ #define X86_FEATURE_PCONFIG (18*32+18) /* Intel PCONFIG */
#define X86_FEATURE_ARCH_LBR (18*32+19) /* Intel ARCH LBR */ #define X86_FEATURE_ARCH_LBR (18*32+19) /* Intel ARCH LBR */
#define X86_FEATURE_AVX512_FP16 (18*32+23) /* AVX512 FP16 */
#define X86_FEATURE_SPEC_CTRL (18*32+26) /* "" Speculation Control (IBRS + IBPB) */ #define X86_FEATURE_SPEC_CTRL (18*32+26) /* "" Speculation Control (IBRS + IBPB) */
#define X86_FEATURE_INTEL_STIBP (18*32+27) /* "" Single Thread Indirect Branch Predictors */ #define X86_FEATURE_INTEL_STIBP (18*32+27) /* "" Single Thread Indirect Branch Predictors */
#define X86_FEATURE_FLUSH_L1D (18*32+28) /* Flush L1D cache */ #define X86_FEATURE_FLUSH_L1D (18*32+28) /* Flush L1D cache */
......
...@@ -472,6 +472,7 @@ ...@@ -472,6 +472,7 @@
#define MSR_AMD64_ICIBSEXTDCTL 0xc001103c #define MSR_AMD64_ICIBSEXTDCTL 0xc001103c
#define MSR_AMD64_IBSOPDATA4 0xc001103d #define MSR_AMD64_IBSOPDATA4 0xc001103d
#define MSR_AMD64_IBS_REG_COUNT_MAX 8 /* includes MSR_AMD64_IBSBRTARGET */ #define MSR_AMD64_IBS_REG_COUNT_MAX 8 /* includes MSR_AMD64_IBSBRTARGET */
#define MSR_AMD64_VM_PAGE_FLUSH 0xc001011e
#define MSR_AMD64_SEV_ES_GHCB 0xc0010130 #define MSR_AMD64_SEV_ES_GHCB 0xc0010130
#define MSR_AMD64_SEV 0xc0010131 #define MSR_AMD64_SEV 0xc0010131
#define MSR_AMD64_SEV_ENABLED_BIT 0 #define MSR_AMD64_SEV_ENABLED_BIT 0
......
...@@ -12,6 +12,7 @@ ...@@ -12,6 +12,7 @@
#define KVM_PIO_PAGE_OFFSET 1 #define KVM_PIO_PAGE_OFFSET 1
#define KVM_COALESCED_MMIO_PAGE_OFFSET 2 #define KVM_COALESCED_MMIO_PAGE_OFFSET 2
#define KVM_DIRTY_LOG_PAGE_OFFSET 64
#define DE_VECTOR 0 #define DE_VECTOR 0
#define DB_VECTOR 1 #define DB_VECTOR 1
......
...@@ -77,10 +77,28 @@ ...@@ -77,10 +77,28 @@
#define SVM_EXIT_MWAIT_COND 0x08c #define SVM_EXIT_MWAIT_COND 0x08c
#define SVM_EXIT_XSETBV 0x08d #define SVM_EXIT_XSETBV 0x08d
#define SVM_EXIT_RDPRU 0x08e #define SVM_EXIT_RDPRU 0x08e
#define SVM_EXIT_EFER_WRITE_TRAP 0x08f
#define SVM_EXIT_CR0_WRITE_TRAP 0x090
#define SVM_EXIT_CR1_WRITE_TRAP 0x091
#define SVM_EXIT_CR2_WRITE_TRAP 0x092
#define SVM_EXIT_CR3_WRITE_TRAP 0x093
#define SVM_EXIT_CR4_WRITE_TRAP 0x094
#define SVM_EXIT_CR5_WRITE_TRAP 0x095
#define SVM_EXIT_CR6_WRITE_TRAP 0x096
#define SVM_EXIT_CR7_WRITE_TRAP 0x097
#define SVM_EXIT_CR8_WRITE_TRAP 0x098
#define SVM_EXIT_CR9_WRITE_TRAP 0x099
#define SVM_EXIT_CR10_WRITE_TRAP 0x09a
#define SVM_EXIT_CR11_WRITE_TRAP 0x09b
#define SVM_EXIT_CR12_WRITE_TRAP 0x09c
#define SVM_EXIT_CR13_WRITE_TRAP 0x09d
#define SVM_EXIT_CR14_WRITE_TRAP 0x09e
#define SVM_EXIT_CR15_WRITE_TRAP 0x09f
#define SVM_EXIT_INVPCID 0x0a2 #define SVM_EXIT_INVPCID 0x0a2
#define SVM_EXIT_NPF 0x400 #define SVM_EXIT_NPF 0x400
#define SVM_EXIT_AVIC_INCOMPLETE_IPI 0x401 #define SVM_EXIT_AVIC_INCOMPLETE_IPI 0x401
#define SVM_EXIT_AVIC_UNACCELERATED_ACCESS 0x402 #define SVM_EXIT_AVIC_UNACCELERATED_ACCESS 0x402
#define SVM_EXIT_VMGEXIT 0x403
/* SEV-ES software-defined VMGEXIT events */ /* SEV-ES software-defined VMGEXIT events */
#define SVM_VMGEXIT_MMIO_READ 0x80000001 #define SVM_VMGEXIT_MMIO_READ 0x80000001
...@@ -183,10 +201,20 @@ ...@@ -183,10 +201,20 @@
{ SVM_EXIT_MONITOR, "monitor" }, \ { SVM_EXIT_MONITOR, "monitor" }, \
{ SVM_EXIT_MWAIT, "mwait" }, \ { SVM_EXIT_MWAIT, "mwait" }, \
{ SVM_EXIT_XSETBV, "xsetbv" }, \ { SVM_EXIT_XSETBV, "xsetbv" }, \
{ SVM_EXIT_EFER_WRITE_TRAP, "write_efer_trap" }, \
{ SVM_EXIT_CR0_WRITE_TRAP, "write_cr0_trap" }, \
{ SVM_EXIT_CR4_WRITE_TRAP, "write_cr4_trap" }, \
{ SVM_EXIT_CR8_WRITE_TRAP, "write_cr8_trap" }, \
{ SVM_EXIT_INVPCID, "invpcid" }, \ { SVM_EXIT_INVPCID, "invpcid" }, \
{ SVM_EXIT_NPF, "npf" }, \ { SVM_EXIT_NPF, "npf" }, \
{ SVM_EXIT_AVIC_INCOMPLETE_IPI, "avic_incomplete_ipi" }, \ { SVM_EXIT_AVIC_INCOMPLETE_IPI, "avic_incomplete_ipi" }, \
{ SVM_EXIT_AVIC_UNACCELERATED_ACCESS, "avic_unaccelerated_access" }, \ { SVM_EXIT_AVIC_UNACCELERATED_ACCESS, "avic_unaccelerated_access" }, \
{ SVM_EXIT_VMGEXIT, "vmgexit" }, \
{ SVM_VMGEXIT_MMIO_READ, "vmgexit_mmio_read" }, \
{ SVM_VMGEXIT_MMIO_WRITE, "vmgexit_mmio_write" }, \
{ SVM_VMGEXIT_NMI_COMPLETE, "vmgexit_nmi_complete" }, \
{ SVM_VMGEXIT_AP_HLT_LOOP, "vmgexit_ap_hlt_loop" }, \
{ SVM_VMGEXIT_AP_JUMP_TABLE, "vmgexit_ap_jump_table" }, \
{ SVM_EXIT_ERR, "invalid_guest_state" } { SVM_EXIT_ERR, "invalid_guest_state" }
......
...@@ -32,6 +32,7 @@ ...@@ -32,6 +32,7 @@
#define EXIT_REASON_EXTERNAL_INTERRUPT 1 #define EXIT_REASON_EXTERNAL_INTERRUPT 1
#define EXIT_REASON_TRIPLE_FAULT 2 #define EXIT_REASON_TRIPLE_FAULT 2
#define EXIT_REASON_INIT_SIGNAL 3 #define EXIT_REASON_INIT_SIGNAL 3
#define EXIT_REASON_SIPI_SIGNAL 4
#define EXIT_REASON_INTERRUPT_WINDOW 7 #define EXIT_REASON_INTERRUPT_WINDOW 7
#define EXIT_REASON_NMI_WINDOW 8 #define EXIT_REASON_NMI_WINDOW 8
...@@ -94,6 +95,7 @@ ...@@ -94,6 +95,7 @@
{ EXIT_REASON_EXTERNAL_INTERRUPT, "EXTERNAL_INTERRUPT" }, \ { EXIT_REASON_EXTERNAL_INTERRUPT, "EXTERNAL_INTERRUPT" }, \
{ EXIT_REASON_TRIPLE_FAULT, "TRIPLE_FAULT" }, \ { EXIT_REASON_TRIPLE_FAULT, "TRIPLE_FAULT" }, \
{ EXIT_REASON_INIT_SIGNAL, "INIT_SIGNAL" }, \ { EXIT_REASON_INIT_SIGNAL, "INIT_SIGNAL" }, \
{ EXIT_REASON_SIPI_SIGNAL, "SIPI_SIGNAL" }, \
{ EXIT_REASON_INTERRUPT_WINDOW, "INTERRUPT_WINDOW" }, \ { EXIT_REASON_INTERRUPT_WINDOW, "INTERRUPT_WINDOW" }, \
{ EXIT_REASON_NMI_WINDOW, "NMI_WINDOW" }, \ { EXIT_REASON_NMI_WINDOW, "NMI_WINDOW" }, \
{ EXIT_REASON_TASK_SWITCH, "TASK_SWITCH" }, \ { EXIT_REASON_TASK_SWITCH, "TASK_SWITCH" }, \
......
...@@ -859,9 +859,11 @@ __SYSCALL(__NR_pidfd_getfd, sys_pidfd_getfd) ...@@ -859,9 +859,11 @@ __SYSCALL(__NR_pidfd_getfd, sys_pidfd_getfd)
__SYSCALL(__NR_faccessat2, sys_faccessat2) __SYSCALL(__NR_faccessat2, sys_faccessat2)
#define __NR_process_madvise 440 #define __NR_process_madvise 440
__SYSCALL(__NR_process_madvise, sys_process_madvise) __SYSCALL(__NR_process_madvise, sys_process_madvise)
#define __NR_epoll_pwait2 441
__SC_COMP(__NR_epoll_pwait2, sys_epoll_pwait2, compat_sys_epoll_pwait2)
#undef __NR_syscalls #undef __NR_syscalls
#define __NR_syscalls 441 #define __NR_syscalls 442
/* /*
* 32 bit systems traditionally used different * 32 bit systems traditionally used different
......
...@@ -250,6 +250,7 @@ struct kvm_hyperv_exit { ...@@ -250,6 +250,7 @@ struct kvm_hyperv_exit {
#define KVM_EXIT_ARM_NISV 28 #define KVM_EXIT_ARM_NISV 28
#define KVM_EXIT_X86_RDMSR 29 #define KVM_EXIT_X86_RDMSR 29
#define KVM_EXIT_X86_WRMSR 30 #define KVM_EXIT_X86_WRMSR 30
#define KVM_EXIT_DIRTY_RING_FULL 31
/* For KVM_EXIT_INTERNAL_ERROR */ /* For KVM_EXIT_INTERNAL_ERROR */
/* Emulate instruction failed. */ /* Emulate instruction failed. */
...@@ -1053,6 +1054,8 @@ struct kvm_ppc_resize_hpt { ...@@ -1053,6 +1054,8 @@ struct kvm_ppc_resize_hpt {
#define KVM_CAP_X86_USER_SPACE_MSR 188 #define KVM_CAP_X86_USER_SPACE_MSR 188
#define KVM_CAP_X86_MSR_FILTER 189 #define KVM_CAP_X86_MSR_FILTER 189
#define KVM_CAP_ENFORCE_PV_FEATURE_CPUID 190 #define KVM_CAP_ENFORCE_PV_FEATURE_CPUID 190
#define KVM_CAP_SYS_HYPERV_CPUID 191
#define KVM_CAP_DIRTY_LOG_RING 192
#ifdef KVM_CAP_IRQ_ROUTING #ifdef KVM_CAP_IRQ_ROUTING
...@@ -1511,7 +1514,7 @@ struct kvm_enc_region { ...@@ -1511,7 +1514,7 @@ struct kvm_enc_region {
/* Available with KVM_CAP_MANUAL_DIRTY_LOG_PROTECT_2 */ /* Available with KVM_CAP_MANUAL_DIRTY_LOG_PROTECT_2 */
#define KVM_CLEAR_DIRTY_LOG _IOWR(KVMIO, 0xc0, struct kvm_clear_dirty_log) #define KVM_CLEAR_DIRTY_LOG _IOWR(KVMIO, 0xc0, struct kvm_clear_dirty_log)
/* Available with KVM_CAP_HYPERV_CPUID */ /* Available with KVM_CAP_HYPERV_CPUID (vcpu) / KVM_CAP_SYS_HYPERV_CPUID (system) */
#define KVM_GET_SUPPORTED_HV_CPUID _IOWR(KVMIO, 0xc1, struct kvm_cpuid2) #define KVM_GET_SUPPORTED_HV_CPUID _IOWR(KVMIO, 0xc1, struct kvm_cpuid2)
/* Available with KVM_CAP_ARM_SVE */ /* Available with KVM_CAP_ARM_SVE */
...@@ -1557,6 +1560,9 @@ struct kvm_pv_cmd { ...@@ -1557,6 +1560,9 @@ struct kvm_pv_cmd {
/* Available with KVM_CAP_X86_MSR_FILTER */ /* Available with KVM_CAP_X86_MSR_FILTER */
#define KVM_X86_SET_MSR_FILTER _IOW(KVMIO, 0xc6, struct kvm_msr_filter) #define KVM_X86_SET_MSR_FILTER _IOW(KVMIO, 0xc6, struct kvm_msr_filter)
/* Available with KVM_CAP_DIRTY_LOG_RING */
#define KVM_RESET_DIRTY_RINGS _IO(KVMIO, 0xc7)
/* Secure Encrypted Virtualization command */ /* Secure Encrypted Virtualization command */
enum sev_cmd_id { enum sev_cmd_id {
/* Guest initialization commands */ /* Guest initialization commands */
...@@ -1710,4 +1716,52 @@ struct kvm_hyperv_eventfd { ...@@ -1710,4 +1716,52 @@ struct kvm_hyperv_eventfd {
#define KVM_DIRTY_LOG_MANUAL_PROTECT_ENABLE (1 << 0) #define KVM_DIRTY_LOG_MANUAL_PROTECT_ENABLE (1 << 0)
#define KVM_DIRTY_LOG_INITIALLY_SET (1 << 1) #define KVM_DIRTY_LOG_INITIALLY_SET (1 << 1)
/*
* Arch needs to define the macro after implementing the dirty ring
* feature. KVM_DIRTY_LOG_PAGE_OFFSET should be defined as the
* starting page offset of the dirty ring structures.
*/
#ifndef KVM_DIRTY_LOG_PAGE_OFFSET
#define KVM_DIRTY_LOG_PAGE_OFFSET 0
#endif
/*
* KVM dirty GFN flags, defined as:
*
* |---------------+---------------+--------------|
* | bit 1 (reset) | bit 0 (dirty) | Status |
* |---------------+---------------+--------------|
* | 0 | 0 | Invalid GFN |
* | 0 | 1 | Dirty GFN |
* | 1 | X | GFN to reset |
* |---------------+---------------+--------------|
*
* Lifecycle of a dirty GFN goes like:
*
* dirtied harvested reset
* 00 -----------> 01 -------------> 1X -------+
* ^ |
* | |
* +------------------------------------------+
*
* The userspace program is only responsible for the 01->1X state
* conversion after harvesting an entry. Also, it must not skip any
* dirty bits, so that dirty bits are always harvested in sequence.
*/
#define KVM_DIRTY_GFN_F_DIRTY BIT(0)
#define KVM_DIRTY_GFN_F_RESET BIT(1)
#define KVM_DIRTY_GFN_F_MASK 0x3
/*
* KVM dirty rings should be mapped at KVM_DIRTY_LOG_PAGE_OFFSET of
* per-vcpu mmaped regions as an array of struct kvm_dirty_gfn. The
* size of the gfn buffer is decided by the first argument when
* enabling KVM_CAP_DIRTY_LOG_RING.
*/
struct kvm_dirty_gfn {
__u32 flags;
__u32 slot;
__u64 offset;
};
#endif /* __LINUX_KVM_H */ #endif /* __LINUX_KVM_H */
...@@ -14,7 +14,6 @@ PERF_HAVE_JITDUMP := 1 ...@@ -14,7 +14,6 @@ PERF_HAVE_JITDUMP := 1
out := $(OUTPUT)arch/powerpc/include/generated/asm out := $(OUTPUT)arch/powerpc/include/generated/asm
header32 := $(out)/syscalls_32.c header32 := $(out)/syscalls_32.c
header64 := $(out)/syscalls_64.c header64 := $(out)/syscalls_64.c
syskrn := $(srctree)/arch/powerpc/kernel/syscalls/syscall.tbl
sysprf := $(srctree)/tools/perf/arch/powerpc/entry/syscalls sysprf := $(srctree)/tools/perf/arch/powerpc/entry/syscalls
sysdef := $(sysprf)/syscall.tbl sysdef := $(sysprf)/syscall.tbl
systbl := $(sysprf)/mksyscalltbl systbl := $(sysprf)/mksyscalltbl
...@@ -23,15 +22,9 @@ systbl := $(sysprf)/mksyscalltbl ...@@ -23,15 +22,9 @@ systbl := $(sysprf)/mksyscalltbl
_dummy := $(shell [ -d '$(out)' ] || mkdir -p '$(out)') _dummy := $(shell [ -d '$(out)' ] || mkdir -p '$(out)')
$(header64): $(sysdef) $(systbl) $(header64): $(sysdef) $(systbl)
@(test -d ../../kernel -a -d ../../tools -a -d ../perf && ( \
(diff -B $(sysdef) $(syskrn) >/dev/null) \
|| echo "Warning: Kernel ABI header at '$(sysdef)' differs from latest version at '$(syskrn)'" >&2 )) || true
$(Q)$(SHELL) '$(systbl)' '64' $(sysdef) > $@ $(Q)$(SHELL) '$(systbl)' '64' $(sysdef) > $@
$(header32): $(sysdef) $(systbl) $(header32): $(sysdef) $(systbl)
@(test -d ../../kernel -a -d ../../tools -a -d ../perf && ( \
(diff -B $(sysdef) $(syskrn) >/dev/null) \
|| echo "Warning: Kernel ABI header at '$(sysdef)' differs from latest version at '$(syskrn)'" >&2 )) || true
$(Q)$(SHELL) '$(systbl)' '32' $(sysdef) > $@ $(Q)$(SHELL) '$(systbl)' '32' $(sysdef) > $@
clean:: clean::
......
...@@ -9,7 +9,9 @@ ...@@ -9,7 +9,9 @@
# #
0 nospu restart_syscall sys_restart_syscall 0 nospu restart_syscall sys_restart_syscall
1 nospu exit sys_exit 1 nospu exit sys_exit
2 nospu fork ppc_fork 2 32 fork ppc_fork sys_fork
2 64 fork sys_fork
2 spu fork sys_ni_syscall
3 common read sys_read 3 common read sys_read
4 common write sys_write 4 common write sys_write
5 common open sys_open compat_sys_open 5 common open sys_open compat_sys_open
...@@ -158,7 +160,9 @@ ...@@ -158,7 +160,9 @@
119 32 sigreturn sys_sigreturn compat_sys_sigreturn 119 32 sigreturn sys_sigreturn compat_sys_sigreturn
119 64 sigreturn sys_ni_syscall 119 64 sigreturn sys_ni_syscall
119 spu sigreturn sys_ni_syscall 119 spu sigreturn sys_ni_syscall
120 nospu clone ppc_clone 120 32 clone ppc_clone sys_clone
120 64 clone sys_clone
120 spu clone sys_ni_syscall
121 common setdomainname sys_setdomainname 121 common setdomainname sys_setdomainname
122 common uname sys_newuname 122 common uname sys_newuname
123 common modify_ldt sys_ni_syscall 123 common modify_ldt sys_ni_syscall
...@@ -240,7 +244,9 @@ ...@@ -240,7 +244,9 @@
186 spu sendfile sys_sendfile64 186 spu sendfile sys_sendfile64
187 common getpmsg sys_ni_syscall 187 common getpmsg sys_ni_syscall
188 common putpmsg sys_ni_syscall 188 common putpmsg sys_ni_syscall
189 nospu vfork ppc_vfork 189 32 vfork ppc_vfork sys_vfork
189 64 vfork sys_vfork
189 spu vfork sys_ni_syscall
190 common ugetrlimit sys_getrlimit compat_sys_getrlimit 190 common ugetrlimit sys_getrlimit compat_sys_getrlimit
191 common readahead sys_readahead compat_sys_readahead 191 common readahead sys_readahead compat_sys_readahead
192 32 mmap2 sys_mmap2 compat_sys_mmap2 192 32 mmap2 sys_mmap2 compat_sys_mmap2
...@@ -316,8 +322,8 @@ ...@@ -316,8 +322,8 @@
248 32 clock_nanosleep sys_clock_nanosleep_time32 248 32 clock_nanosleep sys_clock_nanosleep_time32
248 64 clock_nanosleep sys_clock_nanosleep 248 64 clock_nanosleep sys_clock_nanosleep
248 spu clock_nanosleep sys_clock_nanosleep 248 spu clock_nanosleep sys_clock_nanosleep
249 32 swapcontext ppc_swapcontext ppc32_swapcontext 249 32 swapcontext ppc_swapcontext compat_sys_swapcontext
249 64 swapcontext ppc64_swapcontext 249 64 swapcontext sys_swapcontext
249 spu swapcontext sys_ni_syscall 249 spu swapcontext sys_ni_syscall
250 common tgkill sys_tgkill 250 common tgkill sys_tgkill
251 32 utimes sys_utimes_time32 251 32 utimes sys_utimes_time32
...@@ -456,7 +462,7 @@ ...@@ -456,7 +462,7 @@
361 common bpf sys_bpf 361 common bpf sys_bpf
362 nospu execveat sys_execveat compat_sys_execveat 362 nospu execveat sys_execveat compat_sys_execveat
363 32 switch_endian sys_ni_syscall 363 32 switch_endian sys_ni_syscall
363 64 switch_endian ppc_switch_endian 363 64 switch_endian sys_switch_endian
363 spu switch_endian sys_ni_syscall 363 spu switch_endian sys_ni_syscall
364 common userfaultfd sys_userfaultfd 364 common userfaultfd sys_userfaultfd
365 common membarrier sys_membarrier 365 common membarrier sys_membarrier
...@@ -516,6 +522,12 @@ ...@@ -516,6 +522,12 @@
432 common fsmount sys_fsmount 432 common fsmount sys_fsmount
433 common fspick sys_fspick 433 common fspick sys_fspick
434 common pidfd_open sys_pidfd_open 434 common pidfd_open sys_pidfd_open
435 nospu clone3 ppc_clone3 435 32 clone3 ppc_clone3 sys_clone3
435 64 clone3 sys_clone3
435 spu clone3 sys_ni_syscall
436 common close_range sys_close_range
437 common openat2 sys_openat2 437 common openat2 sys_openat2
438 common pidfd_getfd sys_pidfd_getfd 438 common pidfd_getfd sys_pidfd_getfd
439 common faccessat2 sys_faccessat2
440 common process_madvise sys_process_madvise
441 common epoll_pwait2 sys_epoll_pwait2 compat_sys_epoll_pwait2
...@@ -12,7 +12,6 @@ PERF_HAVE_JITDUMP := 1 ...@@ -12,7 +12,6 @@ PERF_HAVE_JITDUMP := 1
out := $(OUTPUT)arch/s390/include/generated/asm out := $(OUTPUT)arch/s390/include/generated/asm
header := $(out)/syscalls_64.c header := $(out)/syscalls_64.c
syskrn := $(srctree)/arch/s390/kernel/syscalls/syscall.tbl
sysprf := $(srctree)/tools/perf/arch/s390/entry/syscalls sysprf := $(srctree)/tools/perf/arch/s390/entry/syscalls
sysdef := $(sysprf)/syscall.tbl sysdef := $(sysprf)/syscall.tbl
systbl := $(sysprf)/mksyscalltbl systbl := $(sysprf)/mksyscalltbl
...@@ -21,9 +20,6 @@ systbl := $(sysprf)/mksyscalltbl ...@@ -21,9 +20,6 @@ systbl := $(sysprf)/mksyscalltbl
_dummy := $(shell [ -d '$(out)' ] || mkdir -p '$(out)') _dummy := $(shell [ -d '$(out)' ] || mkdir -p '$(out)')
$(header): $(sysdef) $(systbl) $(header): $(sysdef) $(systbl)
@(test -d ../../kernel -a -d ../../tools -a -d ../perf && ( \
(diff -B $(sysdef) $(syskrn) >/dev/null) \
|| echo "Warning: Kernel ABI header at '$(sysdef)' differs from latest version at '$(syskrn)'" >&2 )) || true
$(Q)$(SHELL) '$(systbl)' $(sysdef) > $@ $(Q)$(SHELL) '$(systbl)' $(sysdef) > $@
clean:: clean::
......
...@@ -362,6 +362,7 @@ ...@@ -362,6 +362,7 @@
438 common pidfd_getfd sys_pidfd_getfd 438 common pidfd_getfd sys_pidfd_getfd
439 common faccessat2 sys_faccessat2 439 common faccessat2 sys_faccessat2
440 common process_madvise sys_process_madvise 440 common process_madvise sys_process_madvise
441 common epoll_pwait2 sys_epoll_pwait2
# #
# Due to a historical design error, certain syscalls are numbered differently # Due to a historical design error, certain syscalls are numbered differently
......
...@@ -1186,65 +1186,67 @@ static struct option stat_options[] = { ...@@ -1186,65 +1186,67 @@ static struct option stat_options[] = {
OPT_END() OPT_END()
}; };
static int perf_stat__get_socket(struct perf_stat_config *config __maybe_unused, static struct aggr_cpu_id perf_stat__get_socket(struct perf_stat_config *config __maybe_unused,
struct perf_cpu_map *map, int cpu) struct perf_cpu_map *map, int cpu)
{ {
return cpu_map__get_socket(map, cpu, NULL); return cpu_map__get_socket(map, cpu, NULL);
} }
static int perf_stat__get_die(struct perf_stat_config *config __maybe_unused, static struct aggr_cpu_id perf_stat__get_die(struct perf_stat_config *config __maybe_unused,
struct perf_cpu_map *map, int cpu) struct perf_cpu_map *map, int cpu)
{ {
return cpu_map__get_die(map, cpu, NULL); return cpu_map__get_die(map, cpu, NULL);
} }
static int perf_stat__get_core(struct perf_stat_config *config __maybe_unused, static struct aggr_cpu_id perf_stat__get_core(struct perf_stat_config *config __maybe_unused,
struct perf_cpu_map *map, int cpu) struct perf_cpu_map *map, int cpu)
{ {
return cpu_map__get_core(map, cpu, NULL); return cpu_map__get_core(map, cpu, NULL);
} }
static int perf_stat__get_node(struct perf_stat_config *config __maybe_unused, static struct aggr_cpu_id perf_stat__get_node(struct perf_stat_config *config __maybe_unused,
struct perf_cpu_map *map, int cpu) struct perf_cpu_map *map, int cpu)
{ {
return cpu_map__get_node(map, cpu, NULL); return cpu_map__get_node(map, cpu, NULL);
} }
static int perf_stat__get_aggr(struct perf_stat_config *config, static struct aggr_cpu_id perf_stat__get_aggr(struct perf_stat_config *config,
aggr_get_id_t get_id, struct perf_cpu_map *map, int idx) aggr_get_id_t get_id, struct perf_cpu_map *map, int idx)
{ {
int cpu; int cpu;
struct aggr_cpu_id id = cpu_map__empty_aggr_cpu_id();
if (idx >= map->nr) if (idx >= map->nr)
return -1; return id;
cpu = map->map[idx]; cpu = map->map[idx];
if (config->cpus_aggr_map->map[cpu] == -1) if (cpu_map__aggr_cpu_id_is_empty(config->cpus_aggr_map->map[cpu]))
config->cpus_aggr_map->map[cpu] = get_id(config, map, idx); config->cpus_aggr_map->map[cpu] = get_id(config, map, idx);
return config->cpus_aggr_map->map[cpu]; id = config->cpus_aggr_map->map[cpu];
return id;
} }
static int perf_stat__get_socket_cached(struct perf_stat_config *config, static struct aggr_cpu_id perf_stat__get_socket_cached(struct perf_stat_config *config,
struct perf_cpu_map *map, int idx) struct perf_cpu_map *map, int idx)
{ {
return perf_stat__get_aggr(config, perf_stat__get_socket, map, idx); return perf_stat__get_aggr(config, perf_stat__get_socket, map, idx);
} }
static int perf_stat__get_die_cached(struct perf_stat_config *config, static struct aggr_cpu_id perf_stat__get_die_cached(struct perf_stat_config *config,
struct perf_cpu_map *map, int idx) struct perf_cpu_map *map, int idx)
{ {
return perf_stat__get_aggr(config, perf_stat__get_die, map, idx); return perf_stat__get_aggr(config, perf_stat__get_die, map, idx);
} }
static int perf_stat__get_core_cached(struct perf_stat_config *config, static struct aggr_cpu_id perf_stat__get_core_cached(struct perf_stat_config *config,
struct perf_cpu_map *map, int idx) struct perf_cpu_map *map, int idx)
{ {
return perf_stat__get_aggr(config, perf_stat__get_core, map, idx); return perf_stat__get_aggr(config, perf_stat__get_core, map, idx);
} }
static int perf_stat__get_node_cached(struct perf_stat_config *config, static struct aggr_cpu_id perf_stat__get_node_cached(struct perf_stat_config *config,
struct perf_cpu_map *map, int idx) struct perf_cpu_map *map, int idx)
{ {
return perf_stat__get_aggr(config, perf_stat__get_node, map, idx); return perf_stat__get_aggr(config, perf_stat__get_node, map, idx);
...@@ -1318,14 +1320,29 @@ static int perf_stat_init_aggr_mode(void) ...@@ -1318,14 +1320,29 @@ static int perf_stat_init_aggr_mode(void)
* the aggregation translate cpumap. * the aggregation translate cpumap.
*/ */
nr = perf_cpu_map__max(evsel_list->core.cpus); nr = perf_cpu_map__max(evsel_list->core.cpus);
stat_config.cpus_aggr_map = perf_cpu_map__empty_new(nr + 1); stat_config.cpus_aggr_map = cpu_aggr_map__empty_new(nr + 1);
return stat_config.cpus_aggr_map ? 0 : -ENOMEM; return stat_config.cpus_aggr_map ? 0 : -ENOMEM;
} }
static void cpu_aggr_map__delete(struct cpu_aggr_map *map)
{
if (map) {
WARN_ONCE(refcount_read(&map->refcnt) != 0,
"cpu_aggr_map refcnt unbalanced\n");
free(map);
}
}
static void cpu_aggr_map__put(struct cpu_aggr_map *map)
{
if (map && refcount_dec_and_test(&map->refcnt))
cpu_aggr_map__delete(map);
}
static void perf_stat__exit_aggr_mode(void) static void perf_stat__exit_aggr_mode(void)
{ {
perf_cpu_map__put(stat_config.aggr_map); cpu_aggr_map__put(stat_config.aggr_map);
perf_cpu_map__put(stat_config.cpus_aggr_map); cpu_aggr_map__put(stat_config.cpus_aggr_map);
stat_config.aggr_map = NULL; stat_config.aggr_map = NULL;
stat_config.cpus_aggr_map = NULL; stat_config.cpus_aggr_map = NULL;
} }
...@@ -1345,117 +1362,108 @@ static inline int perf_env__get_cpu(struct perf_env *env, struct perf_cpu_map *m ...@@ -1345,117 +1362,108 @@ static inline int perf_env__get_cpu(struct perf_env *env, struct perf_cpu_map *m
return cpu; return cpu;
} }
static int perf_env__get_socket(struct perf_cpu_map *map, int idx, void *data) static struct aggr_cpu_id perf_env__get_socket(struct perf_cpu_map *map, int idx, void *data)
{ {
struct perf_env *env = data; struct perf_env *env = data;
int cpu = perf_env__get_cpu(env, map, idx); int cpu = perf_env__get_cpu(env, map, idx);
struct aggr_cpu_id id = cpu_map__empty_aggr_cpu_id();
if (cpu != -1)
id.socket = env->cpu[cpu].socket_id;
return cpu == -1 ? -1 : env->cpu[cpu].socket_id; return id;
} }
static int perf_env__get_die(struct perf_cpu_map *map, int idx, void *data) static struct aggr_cpu_id perf_env__get_die(struct perf_cpu_map *map, int idx, void *data)
{ {
struct perf_env *env = data; struct perf_env *env = data;
int die_id = -1, cpu = perf_env__get_cpu(env, map, idx); struct aggr_cpu_id id = cpu_map__empty_aggr_cpu_id();
int cpu = perf_env__get_cpu(env, map, idx);
if (cpu != -1) { if (cpu != -1) {
/* /*
* Encode socket in bit range 15:8 * die_id is relative to socket, so start
* die_id is relative to socket, * with the socket ID and then add die to
* we need a global id. So we combine * make a unique ID.
* socket + die id
*/ */
if (WARN_ONCE(env->cpu[cpu].socket_id >> 8, "The socket id number is too big.\n")) id.socket = env->cpu[cpu].socket_id;
return -1; id.die = env->cpu[cpu].die_id;
if (WARN_ONCE(env->cpu[cpu].die_id >> 8, "The die id number is too big.\n"))
return -1;
die_id = (env->cpu[cpu].socket_id << 8) | (env->cpu[cpu].die_id & 0xff);
} }
return die_id; return id;
} }
static int perf_env__get_core(struct perf_cpu_map *map, int idx, void *data) static struct aggr_cpu_id perf_env__get_core(struct perf_cpu_map *map, int idx, void *data)
{ {
struct perf_env *env = data; struct perf_env *env = data;
int core = -1, cpu = perf_env__get_cpu(env, map, idx); struct aggr_cpu_id id = cpu_map__empty_aggr_cpu_id();
int cpu = perf_env__get_cpu(env, map, idx);
if (cpu != -1) { if (cpu != -1) {
/* /*
* Encode socket in bit range 31:24
* encode die id in bit range 23:16
* core_id is relative to socket and die, * core_id is relative to socket and die,
* we need a global id. So we combine * we need a global id. So we set
* socket + die id + core id * socket, die id and core id
*/ */
if (WARN_ONCE(env->cpu[cpu].socket_id >> 8, "The socket id number is too big.\n")) id.socket = env->cpu[cpu].socket_id;
return -1; id.die = env->cpu[cpu].die_id;
id.core = env->cpu[cpu].core_id;
if (WARN_ONCE(env->cpu[cpu].die_id >> 8, "The die id number is too big.\n"))
return -1;
if (WARN_ONCE(env->cpu[cpu].core_id >> 16, "The core id number is too big.\n"))
return -1;
core = (env->cpu[cpu].socket_id << 24) |
(env->cpu[cpu].die_id << 16) |
(env->cpu[cpu].core_id & 0xffff);
} }
return core; return id;
} }
static int perf_env__get_node(struct perf_cpu_map *map, int idx, void *data) static struct aggr_cpu_id perf_env__get_node(struct perf_cpu_map *map, int idx, void *data)
{ {
int cpu = perf_env__get_cpu(data, map, idx); int cpu = perf_env__get_cpu(data, map, idx);
struct aggr_cpu_id id = cpu_map__empty_aggr_cpu_id();
return perf_env__numa_node(data, cpu); id.node = perf_env__numa_node(data, cpu);
return id;
} }
static int perf_env__build_socket_map(struct perf_env *env, struct perf_cpu_map *cpus, static int perf_env__build_socket_map(struct perf_env *env, struct perf_cpu_map *cpus,
struct perf_cpu_map **sockp) struct cpu_aggr_map **sockp)
{ {
return cpu_map__build_map(cpus, sockp, perf_env__get_socket, env); return cpu_map__build_map(cpus, sockp, perf_env__get_socket, env);
} }
static int perf_env__build_die_map(struct perf_env *env, struct perf_cpu_map *cpus, static int perf_env__build_die_map(struct perf_env *env, struct perf_cpu_map *cpus,
struct perf_cpu_map **diep) struct cpu_aggr_map **diep)
{ {
return cpu_map__build_map(cpus, diep, perf_env__get_die, env); return cpu_map__build_map(cpus, diep, perf_env__get_die, env);
} }
static int perf_env__build_core_map(struct perf_env *env, struct perf_cpu_map *cpus, static int perf_env__build_core_map(struct perf_env *env, struct perf_cpu_map *cpus,
struct perf_cpu_map **corep) struct cpu_aggr_map **corep)
{ {
return cpu_map__build_map(cpus, corep, perf_env__get_core, env); return cpu_map__build_map(cpus, corep, perf_env__get_core, env);
} }
static int perf_env__build_node_map(struct perf_env *env, struct perf_cpu_map *cpus, static int perf_env__build_node_map(struct perf_env *env, struct perf_cpu_map *cpus,
struct perf_cpu_map **nodep) struct cpu_aggr_map **nodep)
{ {
return cpu_map__build_map(cpus, nodep, perf_env__get_node, env); return cpu_map__build_map(cpus, nodep, perf_env__get_node, env);
} }
static int perf_stat__get_socket_file(struct perf_stat_config *config __maybe_unused, static struct aggr_cpu_id perf_stat__get_socket_file(struct perf_stat_config *config __maybe_unused,
struct perf_cpu_map *map, int idx) struct perf_cpu_map *map, int idx)
{ {
return perf_env__get_socket(map, idx, &perf_stat.session->header.env); return perf_env__get_socket(map, idx, &perf_stat.session->header.env);
} }
static int perf_stat__get_die_file(struct perf_stat_config *config __maybe_unused, static struct aggr_cpu_id perf_stat__get_die_file(struct perf_stat_config *config __maybe_unused,
struct perf_cpu_map *map, int idx) struct perf_cpu_map *map, int idx)
{ {
return perf_env__get_die(map, idx, &perf_stat.session->header.env); return perf_env__get_die(map, idx, &perf_stat.session->header.env);
} }
static int perf_stat__get_core_file(struct perf_stat_config *config __maybe_unused, static struct aggr_cpu_id perf_stat__get_core_file(struct perf_stat_config *config __maybe_unused,
struct perf_cpu_map *map, int idx) struct perf_cpu_map *map, int idx)
{ {
return perf_env__get_core(map, idx, &perf_stat.session->header.env); return perf_env__get_core(map, idx, &perf_stat.session->header.env);
} }
static int perf_stat__get_node_file(struct perf_stat_config *config __maybe_unused, static struct aggr_cpu_id perf_stat__get_node_file(struct perf_stat_config *config __maybe_unused,
struct perf_cpu_map *map, int idx) struct perf_cpu_map *map, int idx)
{ {
return perf_env__get_node(map, idx, &perf_stat.session->header.env); return perf_env__get_node(map, idx, &perf_stat.session->header.env);
......
...@@ -144,6 +144,8 @@ check arch/x86/lib/insn.c '-I "^#include [\"<]\(../include/\)*asm/in ...@@ -144,6 +144,8 @@ check arch/x86/lib/insn.c '-I "^#include [\"<]\(../include/\)*asm/in
# diff non-symmetric files # diff non-symmetric files
check_2 tools/perf/arch/x86/entry/syscalls/syscall_64.tbl arch/x86/entry/syscalls/syscall_64.tbl check_2 tools/perf/arch/x86/entry/syscalls/syscall_64.tbl arch/x86/entry/syscalls/syscall_64.tbl
check_2 tools/perf/arch/powerpc/entry/syscalls/syscall.tbl arch/powerpc/kernel/syscalls/syscall.tbl
check_2 tools/perf/arch/s390/entry/syscalls/syscall.tbl arch/s390/kernel/syscalls/syscall.tbl
for i in $BEAUTY_FILES; do for i in $BEAUTY_FILES; do
beauty_check $i -B beauty_check $i -B
......
...@@ -65,9 +65,11 @@ static int check_cpu_topology(char *path, struct perf_cpu_map *map) ...@@ -65,9 +65,11 @@ static int check_cpu_topology(char *path, struct perf_cpu_map *map)
.mode = PERF_DATA_MODE_READ, .mode = PERF_DATA_MODE_READ,
}; };
int i; int i;
struct aggr_cpu_id id;
session = perf_session__new(&data, false, NULL); session = perf_session__new(&data, false, NULL);
TEST_ASSERT_VAL("can't get session", !IS_ERR(session)); TEST_ASSERT_VAL("can't get session", !IS_ERR(session));
cpu__setup_cpunode_map();
/* On platforms with large numbers of CPUs process_cpu_topology() /* On platforms with large numbers of CPUs process_cpu_topology()
* might issue an error while reading the perf.data file section * might issue an error while reading the perf.data file section
...@@ -85,11 +87,18 @@ static int check_cpu_topology(char *path, struct perf_cpu_map *map) ...@@ -85,11 +87,18 @@ static int check_cpu_topology(char *path, struct perf_cpu_map *map)
* "socket_id number is too big. You may need to upgrade the * "socket_id number is too big. You may need to upgrade the
* perf tool." * perf tool."
* *
* This is the reason why this test might be skipped. * This is the reason why this test might be skipped. aarch64 and
* s390 always write this part of the header, even when the above
* condition is true (see do_core_id_test in header.c). So always
* run this test on those platforms.
*/ */
if (!session->header.env.cpu) if (!session->header.env.cpu
&& strncmp(session->header.env.arch, "s390", 4)
&& strncmp(session->header.env.arch, "aarch64", 7))
return TEST_SKIP; return TEST_SKIP;
TEST_ASSERT_VAL("Session header CPU map not set", session->header.env.cpu);
for (i = 0; i < session->header.env.nr_cpus_avail; i++) { for (i = 0; i < session->header.env.nr_cpus_avail; i++) {
if (!cpu_map__has(map, i)) if (!cpu_map__has(map, i))
continue; continue;
...@@ -98,14 +107,57 @@ static int check_cpu_topology(char *path, struct perf_cpu_map *map) ...@@ -98,14 +107,57 @@ static int check_cpu_topology(char *path, struct perf_cpu_map *map)
session->header.env.cpu[i].socket_id); session->header.env.cpu[i].socket_id);
} }
// Test that core ID contains socket, die and core
for (i = 0; i < map->nr; i++) {
id = cpu_map__get_core(map, i, NULL);
TEST_ASSERT_VAL("Core map - Core ID doesn't match",
session->header.env.cpu[map->map[i]].core_id == id.core);
TEST_ASSERT_VAL("Core map - Socket ID doesn't match",
session->header.env.cpu[map->map[i]].socket_id == id.socket);
TEST_ASSERT_VAL("Core map - Die ID doesn't match",
session->header.env.cpu[map->map[i]].die_id == id.die);
TEST_ASSERT_VAL("Core map - Node ID is set", id.node == -1);
TEST_ASSERT_VAL("Core map - Thread is set", id.thread == -1);
}
// Test that die ID contains socket and die
for (i = 0; i < map->nr; i++) { for (i = 0; i < map->nr; i++) {
TEST_ASSERT_VAL("Core ID doesn't match", id = cpu_map__get_die(map, i, NULL);
(session->header.env.cpu[map->map[i]].core_id == (cpu_map__get_core(map, i, NULL) & 0xffff))); TEST_ASSERT_VAL("Die map - Socket ID doesn't match",
session->header.env.cpu[map->map[i]].socket_id == id.socket);
TEST_ASSERT_VAL("Socket ID doesn't match", TEST_ASSERT_VAL("Die map - Die ID doesn't match",
(session->header.env.cpu[map->map[i]].socket_id == cpu_map__get_socket(map, i, NULL))); session->header.env.cpu[map->map[i]].die_id == id.die);
TEST_ASSERT_VAL("Die map - Node ID is set", id.node == -1);
TEST_ASSERT_VAL("Die map - Core is set", id.core == -1);
TEST_ASSERT_VAL("Die map - Thread is set", id.thread == -1);
}
// Test that socket ID contains only socket
for (i = 0; i < map->nr; i++) {
id = cpu_map__get_socket(map, i, NULL);
TEST_ASSERT_VAL("Socket map - Socket ID doesn't match",
session->header.env.cpu[map->map[i]].socket_id == id.socket);
TEST_ASSERT_VAL("Socket map - Node ID is set", id.node == -1);
TEST_ASSERT_VAL("Socket map - Die ID is set", id.die == -1);
TEST_ASSERT_VAL("Socket map - Core is set", id.core == -1);
TEST_ASSERT_VAL("Socket map - Thread is set", id.thread == -1);
} }
// Test that node ID contains only node
for (i = 0; i < map->nr; i++) {
id = cpu_map__get_node(map, i, NULL);
TEST_ASSERT_VAL("Node map - Node ID doesn't match",
cpu__get_node(map->map[i]) == id.node);
TEST_ASSERT_VAL("Node map - Socket is set", id.socket == -1);
TEST_ASSERT_VAL("Node map - Die ID is set", id.die == -1);
TEST_ASSERT_VAL("Node map - Core is set", id.core == -1);
TEST_ASSERT_VAL("Node map - Thread is set", id.thread == -1);
}
perf_session__delete(session); perf_session__delete(session);
return 0; return 0;
......
...@@ -15,7 +15,7 @@ x86_msr_index=${arch_x86_header_dir}/msr-index.h ...@@ -15,7 +15,7 @@ x86_msr_index=${arch_x86_header_dir}/msr-index.h
printf "static const char *x86_MSRs[] = {\n" printf "static const char *x86_MSRs[] = {\n"
regex='^[[:space:]]*#[[:space:]]*define[[:space:]]+MSR_([[:alnum:]][[:alnum:]_]+)[[:space:]]+(0x00000[[:xdigit:]]+)[[:space:]]*.*' regex='^[[:space:]]*#[[:space:]]*define[[:space:]]+MSR_([[:alnum:]][[:alnum:]_]+)[[:space:]]+(0x00000[[:xdigit:]]+)[[:space:]]*.*'
egrep $regex ${x86_msr_index} | egrep -v 'MSR_(ATOM|P[46]|AMD64|IA32_TSCDEADLINE|IDT_FCR4)' | \ egrep $regex ${x86_msr_index} | egrep -v 'MSR_(ATOM|P[46]|IA32_(TSCDEADLINE|UCODE_REV)|IDT_FCR4)' | \
sed -r "s/$regex/\2 \1/g" | sort -n | \ sed -r "s/$regex/\2 \1/g" | sort -n | \
xargs printf "\t[%s] = \"%s\",\n" xargs printf "\t[%s] = \"%s\",\n"
printf "};\n\n" printf "};\n\n"
......
...@@ -95,6 +95,23 @@ struct perf_cpu_map *perf_cpu_map__empty_new(int nr) ...@@ -95,6 +95,23 @@ struct perf_cpu_map *perf_cpu_map__empty_new(int nr)
return cpus; return cpus;
} }
struct cpu_aggr_map *cpu_aggr_map__empty_new(int nr)
{
struct cpu_aggr_map *cpus = malloc(sizeof(*cpus) + sizeof(struct aggr_cpu_id) * nr);
if (cpus != NULL) {
int i;
cpus->nr = nr;
for (i = 0; i < nr; i++)
cpus->map[i] = cpu_map__empty_aggr_cpu_id();
refcount_set(&cpus->refcnt, 1);
}
return cpus;
}
static int cpu__get_topology_int(int cpu, const char *name, int *value) static int cpu__get_topology_int(int cpu, const char *name, int *value)
{ {
char path[PATH_MAX]; char path[PATH_MAX];
...@@ -111,40 +128,57 @@ int cpu_map__get_socket_id(int cpu) ...@@ -111,40 +128,57 @@ int cpu_map__get_socket_id(int cpu)
return ret ?: value; return ret ?: value;
} }
int cpu_map__get_socket(struct perf_cpu_map *map, int idx, void *data __maybe_unused) struct aggr_cpu_id cpu_map__get_socket(struct perf_cpu_map *map, int idx,
void *data __maybe_unused)
{ {
int cpu; int cpu;
struct aggr_cpu_id id = cpu_map__empty_aggr_cpu_id();
if (idx > map->nr) if (idx > map->nr)
return -1; return id;
cpu = map->map[idx]; cpu = map->map[idx];
return cpu_map__get_socket_id(cpu); id.socket = cpu_map__get_socket_id(cpu);
return id;
} }
static int cmp_ids(const void *a, const void *b) static int cmp_aggr_cpu_id(const void *a_pointer, const void *b_pointer)
{ {
return *(int *)a - *(int *)b; struct aggr_cpu_id *a = (struct aggr_cpu_id *)a_pointer;
struct aggr_cpu_id *b = (struct aggr_cpu_id *)b_pointer;
if (a->node != b->node)
return a->node - b->node;
else if (a->socket != b->socket)
return a->socket - b->socket;
else if (a->die != b->die)
return a->die - b->die;
else if (a->core != b->core)
return a->core - b->core;
else
return a->thread - b->thread;
} }
int cpu_map__build_map(struct perf_cpu_map *cpus, struct perf_cpu_map **res, int cpu_map__build_map(struct perf_cpu_map *cpus, struct cpu_aggr_map **res,
int (*f)(struct perf_cpu_map *map, int cpu, void *data), struct aggr_cpu_id (*f)(struct perf_cpu_map *map, int cpu, void *data),
void *data) void *data)
{ {
struct perf_cpu_map *c;
int nr = cpus->nr; int nr = cpus->nr;
int cpu, s1, s2; struct cpu_aggr_map *c = cpu_aggr_map__empty_new(nr);
int cpu, s2;
struct aggr_cpu_id s1;
/* allocate as much as possible */
c = calloc(1, sizeof(*c) + nr * sizeof(int));
if (!c) if (!c)
return -1; return -1;
/* Reset size as it may only be partially filled */
c->nr = 0;
for (cpu = 0; cpu < nr; cpu++) { for (cpu = 0; cpu < nr; cpu++) {
s1 = f(cpus, cpu, data); s1 = f(cpus, cpu, data);
for (s2 = 0; s2 < c->nr; s2++) { for (s2 = 0; s2 < c->nr; s2++) {
if (s1 == c->map[s2]) if (cpu_map__compare_aggr_cpu_id(s1, c->map[s2]))
break; break;
} }
if (s2 == c->nr) { if (s2 == c->nr) {
...@@ -153,9 +187,8 @@ int cpu_map__build_map(struct perf_cpu_map *cpus, struct perf_cpu_map **res, ...@@ -153,9 +187,8 @@ int cpu_map__build_map(struct perf_cpu_map *cpus, struct perf_cpu_map **res,
} }
} }
/* ensure we process id in increasing order */ /* ensure we process id in increasing order */
qsort(c->map, c->nr, sizeof(int), cmp_ids); qsort(c->map, c->nr, sizeof(struct aggr_cpu_id), cmp_aggr_cpu_id);
refcount_set(&c->refcnt, 1);
*res = c; *res = c;
return 0; return 0;
} }
...@@ -167,37 +200,32 @@ int cpu_map__get_die_id(int cpu) ...@@ -167,37 +200,32 @@ int cpu_map__get_die_id(int cpu)
return ret ?: value; return ret ?: value;
} }
int cpu_map__get_die(struct perf_cpu_map *map, int idx, void *data) struct aggr_cpu_id cpu_map__get_die(struct perf_cpu_map *map, int idx, void *data)
{ {
int cpu, die_id, s; int cpu, die;
struct aggr_cpu_id id = cpu_map__empty_aggr_cpu_id();
if (idx > map->nr) if (idx > map->nr)
return -1; return id;
cpu = map->map[idx]; cpu = map->map[idx];
die_id = cpu_map__get_die_id(cpu); die = cpu_map__get_die_id(cpu);
/* There is no die_id on legacy system. */ /* There is no die_id on legacy system. */
if (die_id == -1) if (die == -1)
die_id = 0; die = 0;
s = cpu_map__get_socket(map, idx, data);
if (s == -1)
return -1;
/* /*
* Encode socket in bit range 15:8 * die_id is relative to socket, so start
* die_id is relative to socket, and * with the socket ID and then add die to
* we need a global id. So we combine * make a unique ID.
* socket + die id
*/ */
if (WARN_ONCE(die_id >> 8, "The die id number is too big.\n")) id = cpu_map__get_socket(map, idx, data);
return -1; if (cpu_map__aggr_cpu_id_is_empty(id))
return id;
if (WARN_ONCE(s >> 8, "The socket id number is too big.\n"))
return -1;
return (s << 8) | (die_id & 0xff); id.die = die;
return id;
} }
int cpu_map__get_core_id(int cpu) int cpu_map__get_core_id(int cpu)
...@@ -211,59 +239,58 @@ int cpu_map__get_node_id(int cpu) ...@@ -211,59 +239,58 @@ int cpu_map__get_node_id(int cpu)
return cpu__get_node(cpu); return cpu__get_node(cpu);
} }
int cpu_map__get_core(struct perf_cpu_map *map, int idx, void *data) struct aggr_cpu_id cpu_map__get_core(struct perf_cpu_map *map, int idx, void *data)
{ {
int cpu, s_die; int cpu;
struct aggr_cpu_id id = cpu_map__empty_aggr_cpu_id();
if (idx > map->nr) if (idx > map->nr)
return -1; return id;
cpu = map->map[idx]; cpu = map->map[idx];
cpu = cpu_map__get_core_id(cpu); cpu = cpu_map__get_core_id(cpu);
/* s_die is the combination of socket + die id */ /* cpu_map__get_die returns a struct with socket and die set*/
s_die = cpu_map__get_die(map, idx, data); id = cpu_map__get_die(map, idx, data);
if (s_die == -1) if (cpu_map__aggr_cpu_id_is_empty(id))
return -1; return id;
/* /*
* encode socket in bit range 31:24 * core_id is relative to socket and die, we need a global id.
* encode die id in bit range 23:16 * So we combine the result from cpu_map__get_die with the core id
* core_id is relative to socket and die,
* we need a global id. So we combine
* socket + die id + core id
*/ */
if (WARN_ONCE(cpu >> 16, "The core id number is too big.\n")) id.core = cpu;
return -1; return id;
return (s_die << 16) | (cpu & 0xffff);
} }
int cpu_map__get_node(struct perf_cpu_map *map, int idx, void *data __maybe_unused) struct aggr_cpu_id cpu_map__get_node(struct perf_cpu_map *map, int idx, void *data __maybe_unused)
{ {
struct aggr_cpu_id id = cpu_map__empty_aggr_cpu_id();
if (idx < 0 || idx >= map->nr) if (idx < 0 || idx >= map->nr)
return -1; return id;
return cpu_map__get_node_id(map->map[idx]); id.node = cpu_map__get_node_id(map->map[idx]);
return id;
} }
int cpu_map__build_socket_map(struct perf_cpu_map *cpus, struct perf_cpu_map **sockp) int cpu_map__build_socket_map(struct perf_cpu_map *cpus, struct cpu_aggr_map **sockp)
{ {
return cpu_map__build_map(cpus, sockp, cpu_map__get_socket, NULL); return cpu_map__build_map(cpus, sockp, cpu_map__get_socket, NULL);
} }
int cpu_map__build_die_map(struct perf_cpu_map *cpus, struct perf_cpu_map **diep) int cpu_map__build_die_map(struct perf_cpu_map *cpus, struct cpu_aggr_map **diep)
{ {
return cpu_map__build_map(cpus, diep, cpu_map__get_die, NULL); return cpu_map__build_map(cpus, diep, cpu_map__get_die, NULL);
} }
int cpu_map__build_core_map(struct perf_cpu_map *cpus, struct perf_cpu_map **corep) int cpu_map__build_core_map(struct perf_cpu_map *cpus, struct cpu_aggr_map **corep)
{ {
return cpu_map__build_map(cpus, corep, cpu_map__get_core, NULL); return cpu_map__build_map(cpus, corep, cpu_map__get_core, NULL);
} }
int cpu_map__build_node_map(struct perf_cpu_map *cpus, struct perf_cpu_map **numap) int cpu_map__build_node_map(struct perf_cpu_map *cpus, struct cpu_aggr_map **numap)
{ {
return cpu_map__build_map(cpus, numap, cpu_map__get_node, NULL); return cpu_map__build_map(cpus, numap, cpu_map__get_node, NULL);
} }
...@@ -586,3 +613,33 @@ const struct perf_cpu_map *cpu_map__online(void) /* thread unsafe */ ...@@ -586,3 +613,33 @@ const struct perf_cpu_map *cpu_map__online(void) /* thread unsafe */
return online; return online;
} }
bool cpu_map__compare_aggr_cpu_id(struct aggr_cpu_id a, struct aggr_cpu_id b)
{
return a.thread == b.thread &&
a.node == b.node &&
a.socket == b.socket &&
a.die == b.die &&
a.core == b.core;
}
bool cpu_map__aggr_cpu_id_is_empty(struct aggr_cpu_id a)
{
return a.thread == -1 &&
a.node == -1 &&
a.socket == -1 &&
a.die == -1 &&
a.core == -1;
}
struct aggr_cpu_id cpu_map__empty_aggr_cpu_id(void)
{
struct aggr_cpu_id ret = {
.thread = -1,
.node = -1,
.socket = -1,
.die = -1,
.core = -1
};
return ret;
}
...@@ -7,25 +7,41 @@ ...@@ -7,25 +7,41 @@
#include <internal/cpumap.h> #include <internal/cpumap.h>
#include <perf/cpumap.h> #include <perf/cpumap.h>
struct aggr_cpu_id {
int thread;
int node;
int socket;
int die;
int core;
};
struct cpu_aggr_map {
refcount_t refcnt;
int nr;
struct aggr_cpu_id map[];
};
struct perf_record_cpu_map_data; struct perf_record_cpu_map_data;
struct perf_cpu_map *perf_cpu_map__empty_new(int nr); struct perf_cpu_map *perf_cpu_map__empty_new(int nr);
struct cpu_aggr_map *cpu_aggr_map__empty_new(int nr);
struct perf_cpu_map *cpu_map__new_data(struct perf_record_cpu_map_data *data); struct perf_cpu_map *cpu_map__new_data(struct perf_record_cpu_map_data *data);
size_t cpu_map__snprint(struct perf_cpu_map *map, char *buf, size_t size); size_t cpu_map__snprint(struct perf_cpu_map *map, char *buf, size_t size);
size_t cpu_map__snprint_mask(struct perf_cpu_map *map, char *buf, size_t size); size_t cpu_map__snprint_mask(struct perf_cpu_map *map, char *buf, size_t size);
size_t cpu_map__fprintf(struct perf_cpu_map *map, FILE *fp); size_t cpu_map__fprintf(struct perf_cpu_map *map, FILE *fp);
int cpu_map__get_socket_id(int cpu); int cpu_map__get_socket_id(int cpu);
int cpu_map__get_socket(struct perf_cpu_map *map, int idx, void *data); struct aggr_cpu_id cpu_map__get_socket(struct perf_cpu_map *map, int idx, void *data);
int cpu_map__get_die_id(int cpu); int cpu_map__get_die_id(int cpu);
int cpu_map__get_die(struct perf_cpu_map *map, int idx, void *data); struct aggr_cpu_id cpu_map__get_die(struct perf_cpu_map *map, int idx, void *data);
int cpu_map__get_core_id(int cpu); int cpu_map__get_core_id(int cpu);
int cpu_map__get_core(struct perf_cpu_map *map, int idx, void *data); struct aggr_cpu_id cpu_map__get_core(struct perf_cpu_map *map, int idx, void *data);
int cpu_map__get_node_id(int cpu); int cpu_map__get_node_id(int cpu);
int cpu_map__get_node(struct perf_cpu_map *map, int idx, void *data); struct aggr_cpu_id cpu_map__get_node(struct perf_cpu_map *map, int idx, void *data);
int cpu_map__build_socket_map(struct perf_cpu_map *cpus, struct perf_cpu_map **sockp); int cpu_map__build_socket_map(struct perf_cpu_map *cpus, struct cpu_aggr_map **sockp);
int cpu_map__build_die_map(struct perf_cpu_map *cpus, struct perf_cpu_map **diep); int cpu_map__build_die_map(struct perf_cpu_map *cpus, struct cpu_aggr_map **diep);
int cpu_map__build_core_map(struct perf_cpu_map *cpus, struct perf_cpu_map **corep); int cpu_map__build_core_map(struct perf_cpu_map *cpus, struct cpu_aggr_map **corep);
int cpu_map__build_node_map(struct perf_cpu_map *cpus, struct perf_cpu_map **nodep); int cpu_map__build_node_map(struct perf_cpu_map *cpus, struct cpu_aggr_map **nodep);
const struct perf_cpu_map *cpu_map__online(void); /* thread unsafe */ const struct perf_cpu_map *cpu_map__online(void); /* thread unsafe */
static inline int cpu_map__socket(struct perf_cpu_map *sock, int s) static inline int cpu_map__socket(struct perf_cpu_map *sock, int s)
...@@ -35,21 +51,6 @@ static inline int cpu_map__socket(struct perf_cpu_map *sock, int s) ...@@ -35,21 +51,6 @@ static inline int cpu_map__socket(struct perf_cpu_map *sock, int s)
return sock->map[s]; return sock->map[s];
} }
static inline int cpu_map__id_to_socket(int id)
{
return id >> 24;
}
static inline int cpu_map__id_to_die(int id)
{
return (id >> 16) & 0xff;
}
static inline int cpu_map__id_to_cpu(int id)
{
return id & 0xffff;
}
int cpu__setup_cpunode_map(void); int cpu__setup_cpunode_map(void);
int cpu__max_node(void); int cpu__max_node(void);
...@@ -57,11 +58,15 @@ int cpu__max_cpu(void); ...@@ -57,11 +58,15 @@ int cpu__max_cpu(void);
int cpu__max_present_cpu(void); int cpu__max_present_cpu(void);
int cpu__get_node(int cpu); int cpu__get_node(int cpu);
int cpu_map__build_map(struct perf_cpu_map *cpus, struct perf_cpu_map **res, int cpu_map__build_map(struct perf_cpu_map *cpus, struct cpu_aggr_map **res,
int (*f)(struct perf_cpu_map *map, int cpu, void *data), struct aggr_cpu_id (*f)(struct perf_cpu_map *map, int cpu, void *data),
void *data); void *data);
int cpu_map__cpu(struct perf_cpu_map *cpus, int idx); int cpu_map__cpu(struct perf_cpu_map *cpus, int idx);
bool cpu_map__has(struct perf_cpu_map *cpus, int cpu); bool cpu_map__has(struct perf_cpu_map *cpus, int cpu);
bool cpu_map__compare_aggr_cpu_id(struct aggr_cpu_id a, struct aggr_cpu_id b);
bool cpu_map__aggr_cpu_id_is_empty(struct aggr_cpu_id a);
struct aggr_cpu_id cpu_map__empty_aggr_cpu_id(void);
#endif /* __PERF_CPUMAP_H */ #endif /* __PERF_CPUMAP_H */
...@@ -791,7 +791,7 @@ static char *synthesize_sdt_probe_command(struct sdt_note *note, ...@@ -791,7 +791,7 @@ static char *synthesize_sdt_probe_command(struct sdt_note *note,
const char *sdtgrp) const char *sdtgrp)
{ {
struct strbuf buf; struct strbuf buf;
char *ret = NULL, **args; char *ret = NULL;
int i, args_count, err; int i, args_count, err;
unsigned long long ref_ctr_offset; unsigned long long ref_ctr_offset;
...@@ -813,12 +813,19 @@ static char *synthesize_sdt_probe_command(struct sdt_note *note, ...@@ -813,12 +813,19 @@ static char *synthesize_sdt_probe_command(struct sdt_note *note,
goto out; goto out;
if (note->args) { if (note->args) {
args = argv_split(note->args, &args_count); char **args = argv_split(note->args, &args_count);
if (args == NULL)
goto error;
for (i = 0; i < args_count; ++i) { for (i = 0; i < args_count; ++i) {
if (synthesize_sdt_probe_arg(&buf, i, args[i]) < 0) if (synthesize_sdt_probe_arg(&buf, i, args[i]) < 0) {
argv_free(args);
goto error; goto error;
}
} }
argv_free(args);
} }
out: out:
......
...@@ -68,15 +68,15 @@ static void print_cgroup(struct perf_stat_config *config, struct evsel *evsel) ...@@ -68,15 +68,15 @@ static void print_cgroup(struct perf_stat_config *config, struct evsel *evsel)
static void aggr_printout(struct perf_stat_config *config, static void aggr_printout(struct perf_stat_config *config,
struct evsel *evsel, int id, int nr) struct evsel *evsel, struct aggr_cpu_id id, int nr)
{ {
switch (config->aggr_mode) { switch (config->aggr_mode) {
case AGGR_CORE: case AGGR_CORE:
fprintf(config->output, "S%d-D%d-C%*d%s%*d%s", fprintf(config->output, "S%d-D%d-C%*d%s%*d%s",
cpu_map__id_to_socket(id), id.socket,
cpu_map__id_to_die(id), id.die,
config->csv_output ? 0 : -8, config->csv_output ? 0 : -8,
cpu_map__id_to_cpu(id), id.core,
config->csv_sep, config->csv_sep,
config->csv_output ? 0 : 4, config->csv_output ? 0 : 4,
nr, nr,
...@@ -84,9 +84,9 @@ static void aggr_printout(struct perf_stat_config *config, ...@@ -84,9 +84,9 @@ static void aggr_printout(struct perf_stat_config *config,
break; break;
case AGGR_DIE: case AGGR_DIE:
fprintf(config->output, "S%d-D%*d%s%*d%s", fprintf(config->output, "S%d-D%*d%s%*d%s",
cpu_map__id_to_socket(id << 16), id.socket,
config->csv_output ? 0 : -8, config->csv_output ? 0 : -8,
cpu_map__id_to_die(id << 16), id.die,
config->csv_sep, config->csv_sep,
config->csv_output ? 0 : 4, config->csv_output ? 0 : 4,
nr, nr,
...@@ -95,7 +95,7 @@ static void aggr_printout(struct perf_stat_config *config, ...@@ -95,7 +95,7 @@ static void aggr_printout(struct perf_stat_config *config,
case AGGR_SOCKET: case AGGR_SOCKET:
fprintf(config->output, "S%*d%s%*d%s", fprintf(config->output, "S%*d%s%*d%s",
config->csv_output ? 0 : -5, config->csv_output ? 0 : -5,
id, id.socket,
config->csv_sep, config->csv_sep,
config->csv_output ? 0 : 4, config->csv_output ? 0 : 4,
nr, nr,
...@@ -104,7 +104,7 @@ static void aggr_printout(struct perf_stat_config *config, ...@@ -104,7 +104,7 @@ static void aggr_printout(struct perf_stat_config *config,
case AGGR_NODE: case AGGR_NODE:
fprintf(config->output, "N%*d%s%*d%s", fprintf(config->output, "N%*d%s%*d%s",
config->csv_output ? 0 : -5, config->csv_output ? 0 : -5,
id, id.node,
config->csv_sep, config->csv_sep,
config->csv_output ? 0 : 4, config->csv_output ? 0 : 4,
nr, nr,
...@@ -113,23 +113,23 @@ static void aggr_printout(struct perf_stat_config *config, ...@@ -113,23 +113,23 @@ static void aggr_printout(struct perf_stat_config *config,
case AGGR_NONE: case AGGR_NONE:
if (evsel->percore && !config->percore_show_thread) { if (evsel->percore && !config->percore_show_thread) {
fprintf(config->output, "S%d-D%d-C%*d%s", fprintf(config->output, "S%d-D%d-C%*d%s",
cpu_map__id_to_socket(id), id.socket,
cpu_map__id_to_die(id), id.die,
config->csv_output ? 0 : -3, config->csv_output ? 0 : -3,
cpu_map__id_to_cpu(id), config->csv_sep); id.core, config->csv_sep);
} else if (id > -1) { } else if (id.core > -1) {
fprintf(config->output, "CPU%*d%s", fprintf(config->output, "CPU%*d%s",
config->csv_output ? 0 : -7, config->csv_output ? 0 : -7,
evsel__cpus(evsel)->map[id], evsel__cpus(evsel)->map[id.core],
config->csv_sep); config->csv_sep);
} }
break; break;
case AGGR_THREAD: case AGGR_THREAD:
fprintf(config->output, "%*s-%*d%s", fprintf(config->output, "%*s-%*d%s",
config->csv_output ? 0 : 16, config->csv_output ? 0 : 16,
perf_thread_map__comm(evsel->core.threads, id), perf_thread_map__comm(evsel->core.threads, id.thread),
config->csv_output ? 0 : -8, config->csv_output ? 0 : -8,
perf_thread_map__pid(evsel->core.threads, id), perf_thread_map__pid(evsel->core.threads, id.thread),
config->csv_sep); config->csv_sep);
break; break;
case AGGR_GLOBAL: case AGGR_GLOBAL:
...@@ -144,7 +144,8 @@ struct outstate { ...@@ -144,7 +144,8 @@ struct outstate {
bool newline; bool newline;
const char *prefix; const char *prefix;
int nfields; int nfields;
int id, nr; int nr;
struct aggr_cpu_id id;
struct evsel *evsel; struct evsel *evsel;
}; };
...@@ -319,13 +320,13 @@ static void print_metric_header(struct perf_stat_config *config, ...@@ -319,13 +320,13 @@ static void print_metric_header(struct perf_stat_config *config,
} }
static int first_shadow_cpu(struct perf_stat_config *config, static int first_shadow_cpu(struct perf_stat_config *config,
struct evsel *evsel, int id) struct evsel *evsel, struct aggr_cpu_id id)
{ {
struct evlist *evlist = evsel->evlist; struct evlist *evlist = evsel->evlist;
int i; int i;
if (config->aggr_mode == AGGR_NONE) if (config->aggr_mode == AGGR_NONE)
return id; return id.core;
if (!config->aggr_get_id) if (!config->aggr_get_id)
return 0; return 0;
...@@ -333,14 +334,17 @@ static int first_shadow_cpu(struct perf_stat_config *config, ...@@ -333,14 +334,17 @@ static int first_shadow_cpu(struct perf_stat_config *config,
for (i = 0; i < evsel__nr_cpus(evsel); i++) { for (i = 0; i < evsel__nr_cpus(evsel); i++) {
int cpu2 = evsel__cpus(evsel)->map[i]; int cpu2 = evsel__cpus(evsel)->map[i];
if (config->aggr_get_id(config, evlist->core.cpus, cpu2) == id) if (cpu_map__compare_aggr_cpu_id(
config->aggr_get_id(config, evlist->core.cpus, cpu2),
id)) {
return cpu2; return cpu2;
}
} }
return 0; return 0;
} }
static void abs_printout(struct perf_stat_config *config, static void abs_printout(struct perf_stat_config *config,
int id, int nr, struct evsel *evsel, double avg) struct aggr_cpu_id id, int nr, struct evsel *evsel, double avg)
{ {
FILE *output = config->output; FILE *output = config->output;
double sc = evsel->scale; double sc = evsel->scale;
...@@ -393,7 +397,7 @@ static bool is_mixed_hw_group(struct evsel *counter) ...@@ -393,7 +397,7 @@ static bool is_mixed_hw_group(struct evsel *counter)
return false; return false;
} }
static void printout(struct perf_stat_config *config, int id, int nr, static void printout(struct perf_stat_config *config, struct aggr_cpu_id id, int nr,
struct evsel *counter, double uval, struct evsel *counter, double uval,
char *prefix, u64 run, u64 ena, double noise, char *prefix, u64 run, u64 ena, double noise,
struct runtime_stat *st) struct runtime_stat *st)
...@@ -496,7 +500,8 @@ static void printout(struct perf_stat_config *config, int id, int nr, ...@@ -496,7 +500,8 @@ static void printout(struct perf_stat_config *config, int id, int nr,
static void aggr_update_shadow(struct perf_stat_config *config, static void aggr_update_shadow(struct perf_stat_config *config,
struct evlist *evlist) struct evlist *evlist)
{ {
int cpu, s2, id, s; int cpu, s;
struct aggr_cpu_id s2, id;
u64 val; u64 val;
struct evsel *counter; struct evsel *counter;
...@@ -506,7 +511,7 @@ static void aggr_update_shadow(struct perf_stat_config *config, ...@@ -506,7 +511,7 @@ static void aggr_update_shadow(struct perf_stat_config *config,
val = 0; val = 0;
for (cpu = 0; cpu < evsel__nr_cpus(counter); cpu++) { for (cpu = 0; cpu < evsel__nr_cpus(counter); cpu++) {
s2 = config->aggr_get_id(config, evlist->core.cpus, cpu); s2 = config->aggr_get_id(config, evlist->core.cpus, cpu);
if (s2 != id) if (!cpu_map__compare_aggr_cpu_id(s2, id))
continue; continue;
val += perf_counts(counter->counts, cpu, 0)->val; val += perf_counts(counter->counts, cpu, 0)->val;
} }
...@@ -584,7 +589,7 @@ static bool collect_data(struct perf_stat_config *config, struct evsel *counter, ...@@ -584,7 +589,7 @@ static bool collect_data(struct perf_stat_config *config, struct evsel *counter,
struct aggr_data { struct aggr_data {
u64 ena, run, val; u64 ena, run, val;
int id; struct aggr_cpu_id id;
int nr; int nr;
int cpu; int cpu;
}; };
...@@ -593,13 +598,14 @@ static void aggr_cb(struct perf_stat_config *config, ...@@ -593,13 +598,14 @@ static void aggr_cb(struct perf_stat_config *config,
struct evsel *counter, void *data, bool first) struct evsel *counter, void *data, bool first)
{ {
struct aggr_data *ad = data; struct aggr_data *ad = data;
int cpu, s2; int cpu;
struct aggr_cpu_id s2;
for (cpu = 0; cpu < evsel__nr_cpus(counter); cpu++) { for (cpu = 0; cpu < evsel__nr_cpus(counter); cpu++) {
struct perf_counts_values *counts; struct perf_counts_values *counts;
s2 = config->aggr_get_id(config, evsel__cpus(counter), cpu); s2 = config->aggr_get_id(config, evsel__cpus(counter), cpu);
if (s2 != ad->id) if (!cpu_map__compare_aggr_cpu_id(s2, ad->id))
continue; continue;
if (first) if (first)
ad->nr++; ad->nr++;
...@@ -628,7 +634,8 @@ static void print_counter_aggrdata(struct perf_stat_config *config, ...@@ -628,7 +634,8 @@ static void print_counter_aggrdata(struct perf_stat_config *config,
struct aggr_data ad; struct aggr_data ad;
FILE *output = config->output; FILE *output = config->output;
u64 ena, run, val; u64 ena, run, val;
int id, nr; int nr;
struct aggr_cpu_id id;
double uval; double uval;
ad.id = id = config->aggr_map->map[s]; ad.id = id = config->aggr_map->map[s];
...@@ -649,8 +656,12 @@ static void print_counter_aggrdata(struct perf_stat_config *config, ...@@ -649,8 +656,12 @@ static void print_counter_aggrdata(struct perf_stat_config *config,
fprintf(output, "%s", prefix); fprintf(output, "%s", prefix);
uval = val * counter->scale; uval = val * counter->scale;
printout(config, cpu != -1 ? cpu : id, nr, counter, uval, prefix, if (cpu != -1) {
run, ena, 1.0, &rt_stat); id = cpu_map__empty_aggr_cpu_id();
id.core = cpu;
}
printout(config, id, nr, counter, uval,
prefix, run, ena, 1.0, &rt_stat);
if (!metric_only) if (!metric_only)
fputc('\n', output); fputc('\n', output);
} }
...@@ -728,7 +739,8 @@ static struct perf_aggr_thread_value *sort_aggr_thread( ...@@ -728,7 +739,8 @@ static struct perf_aggr_thread_value *sort_aggr_thread(
continue; continue;
buf[i].counter = counter; buf[i].counter = counter;
buf[i].id = thread; buf[i].id = cpu_map__empty_aggr_cpu_id();
buf[i].id.thread = thread;
buf[i].uval = uval; buf[i].uval = uval;
buf[i].val = val; buf[i].val = val;
buf[i].run = run; buf[i].run = run;
...@@ -751,7 +763,8 @@ static void print_aggr_thread(struct perf_stat_config *config, ...@@ -751,7 +763,8 @@ static void print_aggr_thread(struct perf_stat_config *config,
FILE *output = config->output; FILE *output = config->output;
int nthreads = perf_thread_map__nr(counter->core.threads); int nthreads = perf_thread_map__nr(counter->core.threads);
int ncpus = perf_cpu_map__nr(counter->core.cpus); int ncpus = perf_cpu_map__nr(counter->core.cpus);
int thread, sorted_threads, id; int thread, sorted_threads;
struct aggr_cpu_id id;
struct perf_aggr_thread_value *buf; struct perf_aggr_thread_value *buf;
buf = sort_aggr_thread(counter, nthreads, ncpus, &sorted_threads, _target); buf = sort_aggr_thread(counter, nthreads, ncpus, &sorted_threads, _target);
...@@ -768,7 +781,7 @@ static void print_aggr_thread(struct perf_stat_config *config, ...@@ -768,7 +781,7 @@ static void print_aggr_thread(struct perf_stat_config *config,
if (config->stats) if (config->stats)
printout(config, id, 0, buf[thread].counter, buf[thread].uval, printout(config, id, 0, buf[thread].counter, buf[thread].uval,
prefix, buf[thread].run, buf[thread].ena, 1.0, prefix, buf[thread].run, buf[thread].ena, 1.0,
&config->stats[id]); &config->stats[id.thread]);
else else
printout(config, id, 0, buf[thread].counter, buf[thread].uval, printout(config, id, 0, buf[thread].counter, buf[thread].uval,
prefix, buf[thread].run, buf[thread].ena, 1.0, prefix, buf[thread].run, buf[thread].ena, 1.0,
...@@ -814,8 +827,8 @@ static void print_counter_aggr(struct perf_stat_config *config, ...@@ -814,8 +827,8 @@ static void print_counter_aggr(struct perf_stat_config *config,
fprintf(output, "%s", prefix); fprintf(output, "%s", prefix);
uval = cd.avg * counter->scale; uval = cd.avg * counter->scale;
printout(config, -1, 0, counter, uval, prefix, cd.avg_running, cd.avg_enabled, printout(config, cpu_map__empty_aggr_cpu_id(), 0, counter, uval, prefix, cd.avg_running,
cd.avg, &rt_stat); cd.avg_enabled, cd.avg, &rt_stat);
if (!metric_only) if (!metric_only)
fprintf(output, "\n"); fprintf(output, "\n");
} }
...@@ -842,6 +855,7 @@ static void print_counter(struct perf_stat_config *config, ...@@ -842,6 +855,7 @@ static void print_counter(struct perf_stat_config *config,
u64 ena, run, val; u64 ena, run, val;
double uval; double uval;
int cpu; int cpu;
struct aggr_cpu_id id;
for (cpu = 0; cpu < evsel__nr_cpus(counter); cpu++) { for (cpu = 0; cpu < evsel__nr_cpus(counter); cpu++) {
struct aggr_data ad = { .cpu = cpu }; struct aggr_data ad = { .cpu = cpu };
...@@ -856,8 +870,10 @@ static void print_counter(struct perf_stat_config *config, ...@@ -856,8 +870,10 @@ static void print_counter(struct perf_stat_config *config,
fprintf(output, "%s", prefix); fprintf(output, "%s", prefix);
uval = val * counter->scale; uval = val * counter->scale;
printout(config, cpu, 0, counter, uval, prefix, run, ena, 1.0, id = cpu_map__empty_aggr_cpu_id();
&rt_stat); id.core = cpu;
printout(config, id, 0, counter, uval, prefix,
run, ena, 1.0, &rt_stat);
fputc('\n', output); fputc('\n', output);
} }
...@@ -872,6 +888,7 @@ static void print_no_aggr_metric(struct perf_stat_config *config, ...@@ -872,6 +888,7 @@ static void print_no_aggr_metric(struct perf_stat_config *config,
struct evsel *counter; struct evsel *counter;
u64 ena, run, val; u64 ena, run, val;
double uval; double uval;
struct aggr_cpu_id id;
nrcpus = evlist->core.cpus->nr; nrcpus = evlist->core.cpus->nr;
for (cpu = 0; cpu < nrcpus; cpu++) { for (cpu = 0; cpu < nrcpus; cpu++) {
...@@ -880,8 +897,10 @@ static void print_no_aggr_metric(struct perf_stat_config *config, ...@@ -880,8 +897,10 @@ static void print_no_aggr_metric(struct perf_stat_config *config,
if (prefix) if (prefix)
fputs(prefix, config->output); fputs(prefix, config->output);
evlist__for_each_entry(evlist, counter) { evlist__for_each_entry(evlist, counter) {
id = cpu_map__empty_aggr_cpu_id();
id.core = cpu;
if (first) { if (first) {
aggr_printout(config, counter, cpu, 0); aggr_printout(config, counter, id, 0);
first = false; first = false;
} }
val = perf_counts(counter->counts, cpu, 0)->val; val = perf_counts(counter->counts, cpu, 0)->val;
...@@ -889,8 +908,8 @@ static void print_no_aggr_metric(struct perf_stat_config *config, ...@@ -889,8 +908,8 @@ static void print_no_aggr_metric(struct perf_stat_config *config,
run = perf_counts(counter->counts, cpu, 0)->run; run = perf_counts(counter->counts, cpu, 0)->run;
uval = val * counter->scale; uval = val * counter->scale;
printout(config, cpu, 0, counter, uval, prefix, run, ena, 1.0, printout(config, id, 0, counter, uval, prefix,
&rt_stat); run, ena, 1.0, &rt_stat);
} }
fputc('\n', config->output); fputc('\n', config->output);
} }
...@@ -1140,14 +1159,15 @@ static void print_footer(struct perf_stat_config *config) ...@@ -1140,14 +1159,15 @@ static void print_footer(struct perf_stat_config *config)
static void print_percore_thread(struct perf_stat_config *config, static void print_percore_thread(struct perf_stat_config *config,
struct evsel *counter, char *prefix) struct evsel *counter, char *prefix)
{ {
int s, s2, id; int s;
struct aggr_cpu_id s2, id;
bool first = true; bool first = true;
for (int i = 0; i < evsel__nr_cpus(counter); i++) { for (int i = 0; i < evsel__nr_cpus(counter); i++) {
s2 = config->aggr_get_id(config, evsel__cpus(counter), i); s2 = config->aggr_get_id(config, evsel__cpus(counter), i);
for (s = 0; s < config->aggr_map->nr; s++) { for (s = 0; s < config->aggr_map->nr; s++) {
id = config->aggr_map->map[s]; id = config->aggr_map->map[s];
if (s2 == id) if (cpu_map__compare_aggr_cpu_id(s2, id))
break; break;
} }
......
...@@ -313,7 +313,7 @@ static int check_per_pkg(struct evsel *counter, ...@@ -313,7 +313,7 @@ static int check_per_pkg(struct evsel *counter,
if (!(vals->run && vals->ena)) if (!(vals->run && vals->ena))
return 0; return 0;
s = cpu_map__get_socket(cpus, cpu, NULL); s = cpu_map__get_socket(cpus, cpu, NULL).socket;
if (s < 0) if (s < 0)
return -1; return -1;
......
...@@ -6,6 +6,7 @@ ...@@ -6,6 +6,7 @@
#include <stdio.h> #include <stdio.h>
#include <sys/types.h> #include <sys/types.h>
#include <sys/resource.h> #include <sys/resource.h>
#include "cpumap.h"
#include "rblist.h" #include "rblist.h"
struct perf_cpu_map; struct perf_cpu_map;
...@@ -99,7 +100,7 @@ struct runtime_stat { ...@@ -99,7 +100,7 @@ struct runtime_stat {
struct rblist value_list; struct rblist value_list;
}; };
typedef int (*aggr_get_id_t)(struct perf_stat_config *config, typedef struct aggr_cpu_id (*aggr_get_id_t)(struct perf_stat_config *config,
struct perf_cpu_map *m, int cpu); struct perf_cpu_map *m, int cpu);
struct perf_stat_config { struct perf_stat_config {
...@@ -138,9 +139,9 @@ struct perf_stat_config { ...@@ -138,9 +139,9 @@ struct perf_stat_config {
const char *csv_sep; const char *csv_sep;
struct stats *walltime_nsecs_stats; struct stats *walltime_nsecs_stats;
struct rusage ru_data; struct rusage ru_data;
struct perf_cpu_map *aggr_map; struct cpu_aggr_map *aggr_map;
aggr_get_id_t aggr_get_id; aggr_get_id_t aggr_get_id;
struct perf_cpu_map *cpus_aggr_map; struct cpu_aggr_map *cpus_aggr_map;
u64 *walltime_run; u64 *walltime_run;
struct rblist metric_events; struct rblist metric_events;
int ctl_fd; int ctl_fd;
...@@ -170,7 +171,7 @@ struct evlist; ...@@ -170,7 +171,7 @@ struct evlist;
struct perf_aggr_thread_value { struct perf_aggr_thread_value {
struct evsel *counter; struct evsel *counter;
int id; struct aggr_cpu_id id;
double uval; double uval;
u64 val; u64 val;
u64 run; u64 run;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment