Commit 03785a69 authored by Linus Torvalds's avatar Linus Torvalds

Merge tag 's390-6.1-1' of git://git.kernel.org/pub/scm/linux/kernel/git/s390/linux

Pull s390 updates from Vasily Gorbik:

 - Make use of the IBM z16 processor activity instrumentation facility
   extension to count neural network processor assist operations: add a
   new PMU device driver so that perf can make use of this.

 - Rework memcpy_real() to avoid DAT-off mode.

 - Rework absolute lowcore access code.

 - Various small fixes and improvements all over the code.

* tag 's390-6.1-1' of git://git.kernel.org/pub/scm/linux/kernel/git/s390/linux:
  s390/pci: remove unused bus_next field from struct zpci_dev
  s390/cio: remove unused ccw_device_force_console() declaration
  s390/pai: Add support for PAI Extension 1 NNPA counters
  s390/mm: fix no previous prototype warnings in maccess.c
  s390/mm: uninline copy_oldmem_kernel() function
  s390/mm,ptdump: add real memory copy page markers
  s390/mm: rework memcpy_real() to avoid DAT-off mode
  s390/dump: save IPL CPU registers once DAT is available
  s390/pci: convert high_memory to physical address
  s390/smp,ptdump: add absolute lowcore markers
  s390/smp: rework absolute lowcore access
  s390/smp: call smp_reinit_ipl_cpu() before scheduler is available
  s390/ptdump: add missing amode31 markers
  s390/mm: split lowcore pages with set_memory_4k()
  s390/mm: remove unused access parameter from do_fault_error()
  s390/delay: sync comment within __delay() with reality
  s390: move from strlcpy with unused retval to strscpy
parents 2e64066d 8fb65e05
...@@ -10,11 +10,14 @@ ...@@ -10,11 +10,14 @@
#include <asm/sclp.h> #include <asm/sclp.h>
#include <asm/diag.h> #include <asm/diag.h>
#include <asm/uv.h> #include <asm/uv.h>
#include <asm/abs_lowcore.h>
#include "decompressor.h" #include "decompressor.h"
#include "boot.h" #include "boot.h"
#include "uv.h" #include "uv.h"
unsigned long __bootdata_preserved(__kaslr_offset); unsigned long __bootdata_preserved(__kaslr_offset);
unsigned long __bootdata_preserved(__abs_lowcore);
unsigned long __bootdata_preserved(__memcpy_real_area);
unsigned long __bootdata(__amode31_base); unsigned long __bootdata(__amode31_base);
unsigned long __bootdata_preserved(VMALLOC_START); unsigned long __bootdata_preserved(VMALLOC_START);
unsigned long __bootdata_preserved(VMALLOC_END); unsigned long __bootdata_preserved(VMALLOC_END);
...@@ -180,7 +183,10 @@ static void setup_kernel_memory_layout(void) ...@@ -180,7 +183,10 @@ static void setup_kernel_memory_layout(void)
/* force vmalloc and modules below kasan shadow */ /* force vmalloc and modules below kasan shadow */
vmax = min(vmax, KASAN_SHADOW_START); vmax = min(vmax, KASAN_SHADOW_START);
#endif #endif
MODULES_END = vmax; __memcpy_real_area = round_down(vmax - PAGE_SIZE, PAGE_SIZE);
__abs_lowcore = round_down(__memcpy_real_area - ABS_LOWCORE_MAP_SIZE,
sizeof(struct lowcore));
MODULES_END = round_down(__abs_lowcore, _SEGMENT_SIZE);
MODULES_VADDR = MODULES_END - MODULES_LEN; MODULES_VADDR = MODULES_END - MODULES_LEN;
VMALLOC_END = MODULES_VADDR; VMALLOC_END = MODULES_VADDR;
......
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef _ASM_S390_ABS_LOWCORE_H
#define _ASM_S390_ABS_LOWCORE_H
#include <asm/lowcore.h>
#define ABS_LOWCORE_MAP_SIZE (NR_CPUS * sizeof(struct lowcore))
extern unsigned long __abs_lowcore;
extern bool abs_lowcore_mapped;
struct lowcore *get_abs_lowcore(unsigned long *flags);
void put_abs_lowcore(struct lowcore *lc, unsigned long flags);
int abs_lowcore_map(int cpu, struct lowcore *lc, bool alloc);
void abs_lowcore_unmap(int cpu);
#endif /* _ASM_S390_ABS_LOWCORE_H */
...@@ -214,7 +214,6 @@ extern struct ccw_device *ccw_device_create_console(struct ccw_driver *); ...@@ -214,7 +214,6 @@ extern struct ccw_device *ccw_device_create_console(struct ccw_driver *);
extern void ccw_device_destroy_console(struct ccw_device *); extern void ccw_device_destroy_console(struct ccw_device *);
extern int ccw_device_enable_console(struct ccw_device *); extern int ccw_device_enable_console(struct ccw_device *);
extern void ccw_device_wait_idle(struct ccw_device *); extern void ccw_device_wait_idle(struct ccw_device *);
extern int ccw_device_force_console(struct ccw_device *);
extern void *ccw_device_dma_zalloc(struct ccw_device *cdev, size_t size); extern void *ccw_device_dma_zalloc(struct ccw_device *cdev, size_t size);
extern void ccw_device_dma_free(struct ccw_device *cdev, extern void ccw_device_dma_free(struct ccw_device *cdev,
......
...@@ -95,7 +95,8 @@ union ctlreg0 { ...@@ -95,7 +95,8 @@ union ctlreg0 {
Interruption-Filtering Override */ Interruption-Filtering Override */
unsigned long : 3; unsigned long : 3;
unsigned long ccc : 1; /* Cryptography counter control */ unsigned long ccc : 1; /* Cryptography counter control */
unsigned long : 18; unsigned long pec : 1; /* PAI extension control */
unsigned long : 17;
unsigned long : 3; unsigned long : 3;
unsigned long lap : 1; /* Low-address-protection control */ unsigned long lap : 1; /* Low-address-protection control */
unsigned long : 4; unsigned long : 4;
......
...@@ -203,7 +203,9 @@ struct lowcore { ...@@ -203,7 +203,9 @@ struct lowcore {
__u8 pad_0x1400[0x1500-0x1400]; /* 0x1400 */ __u8 pad_0x1400[0x1500-0x1400]; /* 0x1400 */
/* Cryptography-counter designation */ /* Cryptography-counter designation */
__u64 ccd; /* 0x1500 */ __u64 ccd; /* 0x1500 */
__u8 pad_0x1508[0x1800-0x1508]; /* 0x1508 */ /* AI-extension counter designation */
__u64 aicd; /* 0x1508 */
__u8 pad_0x1510[0x1800-0x1510]; /* 0x1510 */
/* Transaction abort diagnostic block */ /* Transaction abort diagnostic block */
struct pgm_tdb pgm_tdb; /* 0x1800 */ struct pgm_tdb pgm_tdb; /* 0x1800 */
......
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef __ASM_S390_MACCESS_H
#define __ASM_S390_MACCESS_H
#include <linux/types.h>
struct iov_iter;
extern unsigned long __memcpy_real_area;
void memcpy_real_init(void);
size_t memcpy_real_iter(struct iov_iter *iter, unsigned long src, size_t count);
int memcpy_real(void *dest, unsigned long src, size_t count);
#ifdef CONFIG_CRASH_DUMP
int copy_oldmem_kernel(void *dst, unsigned long src, size_t count);
#endif
#endif /* __ASM_S390_MACCESS_H */
...@@ -41,20 +41,6 @@ u32 os_info_csum(struct os_info *os_info); ...@@ -41,20 +41,6 @@ u32 os_info_csum(struct os_info *os_info);
#ifdef CONFIG_CRASH_DUMP #ifdef CONFIG_CRASH_DUMP
void *os_info_old_entry(int nr, unsigned long *size); void *os_info_old_entry(int nr, unsigned long *size);
size_t copy_oldmem_iter(struct iov_iter *iter, unsigned long src, size_t count);
static inline int copy_oldmem_kernel(void *dst, unsigned long src, size_t count)
{
struct iov_iter iter;
struct kvec kvec;
kvec.iov_base = dst;
kvec.iov_len = count;
iov_iter_kvec(&iter, WRITE, &kvec, 1, count);
if (copy_oldmem_iter(&iter, src, count) < count)
return -EFAULT;
return 0;
}
#else #else
static inline void *os_info_old_entry(int nr, unsigned long *size) static inline void *os_info_old_entry(int nr, unsigned long *size)
{ {
......
...@@ -17,7 +17,9 @@ struct qpaci_info_block { ...@@ -17,7 +17,9 @@ struct qpaci_info_block {
struct { struct {
u64 : 8; u64 : 8;
u64 num_cc : 8; /* # of supported crypto counters */ u64 num_cc : 8; /* # of supported crypto counters */
u64 : 48; u64 : 9;
u64 num_nnpa : 7; /* # of supported NNPA counters */
u64 : 32;
}; };
}; };
...@@ -42,6 +44,8 @@ static inline int qpaci(struct qpaci_info_block *info) ...@@ -42,6 +44,8 @@ static inline int qpaci(struct qpaci_info_block *info)
#define PAI_CRYPTO_BASE 0x1000 /* First event number */ #define PAI_CRYPTO_BASE 0x1000 /* First event number */
#define PAI_CRYPTO_MAXCTR 256 /* Max # of event counters */ #define PAI_CRYPTO_MAXCTR 256 /* Max # of event counters */
#define PAI_CRYPTO_KERNEL_OFFSET 2048 #define PAI_CRYPTO_KERNEL_OFFSET 2048
#define PAI_NNPA_BASE 0x1800 /* First event number */
#define PAI_NNPA_MAXCTR 128 /* Max # of event counters */
DECLARE_STATIC_KEY_FALSE(pai_key); DECLARE_STATIC_KEY_FALSE(pai_key);
......
...@@ -117,7 +117,6 @@ struct zpci_bus { ...@@ -117,7 +117,6 @@ struct zpci_bus {
struct zpci_dev { struct zpci_dev {
struct zpci_bus *zbus; struct zpci_bus *zbus;
struct list_head entry; /* list of all zpci_devices, needed for hotplug, etc. */ struct list_head entry; /* list of all zpci_devices, needed for hotplug, etc. */
struct list_head bus_next;
struct kref kref; struct kref kref;
struct hotplug_slot hotplug_slot; struct hotplug_slot hotplug_slot;
......
...@@ -1777,6 +1777,10 @@ static inline swp_entry_t __swp_entry(unsigned long type, unsigned long offset) ...@@ -1777,6 +1777,10 @@ static inline swp_entry_t __swp_entry(unsigned long type, unsigned long offset)
extern int vmem_add_mapping(unsigned long start, unsigned long size); extern int vmem_add_mapping(unsigned long start, unsigned long size);
extern void vmem_remove_mapping(unsigned long start, unsigned long size); extern void vmem_remove_mapping(unsigned long start, unsigned long size);
extern int __vmem_map_4k_page(unsigned long addr, unsigned long phys, pgprot_t prot, bool alloc);
extern int vmem_map_4k_page(unsigned long addr, unsigned long phys, pgprot_t prot);
extern void vmem_unmap_4k_page(unsigned long addr);
extern pte_t *vmem_get_alloc_pte(unsigned long addr, bool alloc);
extern int s390_enable_sie(void); extern int s390_enable_sie(void);
extern int s390_enable_skey(void); extern int s390_enable_skey(void);
extern void s390_reset_cmma(struct mm_struct *mm); extern void s390_reset_cmma(struct mm_struct *mm);
......
...@@ -306,23 +306,6 @@ static __always_inline void __noreturn disabled_wait(void) ...@@ -306,23 +306,6 @@ static __always_inline void __noreturn disabled_wait(void)
#define ARCH_LOW_ADDRESS_LIMIT 0x7fffffffUL #define ARCH_LOW_ADDRESS_LIMIT 0x7fffffffUL
extern int memcpy_real(void *, unsigned long, size_t);
extern void memcpy_absolute(void *, void *, size_t);
#define put_abs_lowcore(member, x) do { \
unsigned long __abs_address = offsetof(struct lowcore, member); \
__typeof__(((struct lowcore *)0)->member) __tmp = (x); \
\
memcpy_absolute(__va(__abs_address), &__tmp, sizeof(__tmp)); \
} while (0)
#define get_abs_lowcore(x, member) do { \
unsigned long __abs_address = offsetof(struct lowcore, member); \
__typeof__(((struct lowcore *)0)->member) *__ptr = &(x); \
\
memcpy_absolute(__ptr, __va(__abs_address), sizeof(*__ptr)); \
} while (0)
extern int s390_isolate_bp(void); extern int s390_isolate_bp(void);
extern int s390_isolate_bp_guest(void); extern int s390_isolate_bp_guest(void);
......
...@@ -30,7 +30,8 @@ extern void smp_emergency_stop(void); ...@@ -30,7 +30,8 @@ extern void smp_emergency_stop(void);
extern int smp_find_processor_id(u16 address); extern int smp_find_processor_id(u16 address);
extern int smp_store_status(int cpu); extern int smp_store_status(int cpu);
extern void smp_save_dump_cpus(void); extern void smp_save_dump_ipl_cpu(void);
extern void smp_save_dump_secondary_cpus(void);
extern void smp_yield_cpu(int cpu); extern void smp_yield_cpu(int cpu);
extern void smp_cpu_set_polarization(int cpu, int val); extern void smp_cpu_set_polarization(int cpu, int val);
extern int smp_cpu_get_polarization(int cpu); extern int smp_cpu_get_polarization(int cpu);
...@@ -58,6 +59,7 @@ static inline void smp_cpus_done(unsigned int max_cpus) ...@@ -58,6 +59,7 @@ static inline void smp_cpus_done(unsigned int max_cpus)
{ {
} }
extern int smp_reinit_ipl_cpu(void);
extern int smp_rescan_cpus(void); extern int smp_rescan_cpus(void);
extern void __noreturn cpu_die(void); extern void __noreturn cpu_die(void);
extern void __cpu_die(unsigned int cpu); extern void __cpu_die(unsigned int cpu);
......
...@@ -40,7 +40,7 @@ obj-y += sysinfo.o lgr.o os_info.o machine_kexec.o ...@@ -40,7 +40,7 @@ obj-y += sysinfo.o lgr.o os_info.o machine_kexec.o
obj-y += runtime_instr.o cache.o fpu.o dumpstack.o guarded_storage.o sthyi.o obj-y += runtime_instr.o cache.o fpu.o dumpstack.o guarded_storage.o sthyi.o
obj-y += entry.o reipl.o relocate_kernel.o kdebugfs.o alternative.o obj-y += entry.o reipl.o relocate_kernel.o kdebugfs.o alternative.o
obj-y += nospec-branch.o ipl_vmparm.o machine_kexec_reloc.o unwind_bc.o obj-y += nospec-branch.o ipl_vmparm.o machine_kexec_reloc.o unwind_bc.o
obj-y += smp.o text_amode31.o stacktrace.o obj-y += smp.o text_amode31.o stacktrace.o abs_lowcore.o
extra-y += head64.o vmlinux.lds extra-y += head64.o vmlinux.lds
...@@ -72,7 +72,7 @@ obj-$(CONFIG_IMA_SECURE_AND_OR_TRUSTED_BOOT) += ima_arch.o ...@@ -72,7 +72,7 @@ obj-$(CONFIG_IMA_SECURE_AND_OR_TRUSTED_BOOT) += ima_arch.o
obj-$(CONFIG_PERF_EVENTS) += perf_event.o perf_cpum_cf_common.o obj-$(CONFIG_PERF_EVENTS) += perf_event.o perf_cpum_cf_common.o
obj-$(CONFIG_PERF_EVENTS) += perf_cpum_cf.o perf_cpum_sf.o obj-$(CONFIG_PERF_EVENTS) += perf_cpum_cf.o perf_cpum_sf.o
obj-$(CONFIG_PERF_EVENTS) += perf_cpum_cf_events.o perf_regs.o obj-$(CONFIG_PERF_EVENTS) += perf_cpum_cf_events.o perf_regs.o
obj-$(CONFIG_PERF_EVENTS) += perf_pai_crypto.o obj-$(CONFIG_PERF_EVENTS) += perf_pai_crypto.o perf_pai_ext.o
obj-$(CONFIG_TRACEPOINTS) += trace.o obj-$(CONFIG_TRACEPOINTS) += trace.o
obj-$(findstring y, $(CONFIG_PROTECTED_VIRTUALIZATION_GUEST) $(CONFIG_PGSTE)) += uv.o obj-$(findstring y, $(CONFIG_PROTECTED_VIRTUALIZATION_GUEST) $(CONFIG_PGSTE)) += uv.o
......
// SPDX-License-Identifier: GPL-2.0
#include <linux/pgtable.h>
#include <asm/abs_lowcore.h>
#define ABS_LOWCORE_UNMAPPED 1
#define ABS_LOWCORE_LAP_ON 2
#define ABS_LOWCORE_IRQS_ON 4
unsigned long __bootdata_preserved(__abs_lowcore);
bool __ro_after_init abs_lowcore_mapped;
int abs_lowcore_map(int cpu, struct lowcore *lc, bool alloc)
{
unsigned long addr = __abs_lowcore + (cpu * sizeof(struct lowcore));
unsigned long phys = __pa(lc);
int rc, i;
for (i = 0; i < LC_PAGES; i++) {
rc = __vmem_map_4k_page(addr, phys, PAGE_KERNEL, alloc);
if (rc) {
/*
* Do not unmap allocated page tables in case the
* allocation was not requested. In such a case the
* request is expected coming from an atomic context,
* while the unmap attempt might sleep.
*/
if (alloc) {
for (--i; i >= 0; i--) {
addr -= PAGE_SIZE;
vmem_unmap_4k_page(addr);
}
}
return rc;
}
addr += PAGE_SIZE;
phys += PAGE_SIZE;
}
return 0;
}
void abs_lowcore_unmap(int cpu)
{
unsigned long addr = __abs_lowcore + (cpu * sizeof(struct lowcore));
int i;
for (i = 0; i < LC_PAGES; i++) {
vmem_unmap_4k_page(addr);
addr += PAGE_SIZE;
}
}
struct lowcore *get_abs_lowcore(unsigned long *flags)
{
unsigned long irq_flags;
union ctlreg0 cr0;
int cpu;
*flags = 0;
cpu = get_cpu();
if (abs_lowcore_mapped) {
return ((struct lowcore *)__abs_lowcore) + cpu;
} else {
if (cpu != 0)
panic("Invalid unmapped absolute lowcore access\n");
local_irq_save(irq_flags);
if (!irqs_disabled_flags(irq_flags))
*flags |= ABS_LOWCORE_IRQS_ON;
__ctl_store(cr0.val, 0, 0);
if (cr0.lap) {
*flags |= ABS_LOWCORE_LAP_ON;
__ctl_clear_bit(0, 28);
}
*flags |= ABS_LOWCORE_UNMAPPED;
return lowcore_ptr[0];
}
}
void put_abs_lowcore(struct lowcore *lc, unsigned long flags)
{
if (abs_lowcore_mapped) {
if (flags)
panic("Invalid mapped absolute lowcore release\n");
} else {
if (smp_processor_id() != 0)
panic("Invalid mapped absolute lowcore access\n");
if (!(flags & ABS_LOWCORE_UNMAPPED))
panic("Invalid unmapped absolute lowcore release\n");
if (flags & ABS_LOWCORE_LAP_ON)
__ctl_set_bit(0, 28);
if (flags & ABS_LOWCORE_IRQS_ON)
local_irq_enable();
}
put_cpu();
}
...@@ -21,6 +21,7 @@ ...@@ -21,6 +21,7 @@
#include <asm/elf.h> #include <asm/elf.h>
#include <asm/ipl.h> #include <asm/ipl.h>
#include <asm/sclp.h> #include <asm/sclp.h>
#include <asm/maccess.h>
#define PTR_ADD(x, y) (((char *) (x)) + ((unsigned long) (y))) #define PTR_ADD(x, y) (((char *) (x)) + ((unsigned long) (y)))
#define PTR_SUB(x, y) (((char *) (x)) - ((unsigned long) (y))) #define PTR_SUB(x, y) (((char *) (x)) - ((unsigned long) (y)))
...@@ -53,8 +54,6 @@ struct save_area { ...@@ -53,8 +54,6 @@ struct save_area {
}; };
static LIST_HEAD(dump_save_areas); static LIST_HEAD(dump_save_areas);
static DEFINE_MUTEX(memcpy_real_mutex);
static char memcpy_real_buf[PAGE_SIZE];
/* /*
* Allocate a save area * Allocate a save area
...@@ -116,27 +115,7 @@ void __init save_area_add_vxrs(struct save_area *sa, __vector128 *vxrs) ...@@ -116,27 +115,7 @@ void __init save_area_add_vxrs(struct save_area *sa, __vector128 *vxrs)
memcpy(sa->vxrs_high, vxrs + 16, 16 * sizeof(__vector128)); memcpy(sa->vxrs_high, vxrs + 16, 16 * sizeof(__vector128));
} }
static size_t copy_to_iter_real(struct iov_iter *iter, unsigned long src, size_t count) static size_t copy_oldmem_iter(struct iov_iter *iter, unsigned long src, size_t count)
{
size_t len, copied, res = 0;
mutex_lock(&memcpy_real_mutex);
while (count) {
len = min(PAGE_SIZE, count);
if (memcpy_real(memcpy_real_buf, src, len))
break;
copied = copy_to_iter(memcpy_real_buf, len, iter);
count -= copied;
src += copied;
res += copied;
if (copied < len)
break;
}
mutex_unlock(&memcpy_real_mutex);
return res;
}
size_t copy_oldmem_iter(struct iov_iter *iter, unsigned long src, size_t count)
{ {
size_t len, copied, res = 0; size_t len, copied, res = 0;
...@@ -156,7 +135,7 @@ size_t copy_oldmem_iter(struct iov_iter *iter, unsigned long src, size_t count) ...@@ -156,7 +135,7 @@ size_t copy_oldmem_iter(struct iov_iter *iter, unsigned long src, size_t count)
} else { } else {
len = count; len = count;
} }
copied = copy_to_iter_real(iter, src, len); copied = memcpy_real_iter(iter, src, len);
} }
count -= copied; count -= copied;
src += copied; src += copied;
...@@ -167,6 +146,19 @@ size_t copy_oldmem_iter(struct iov_iter *iter, unsigned long src, size_t count) ...@@ -167,6 +146,19 @@ size_t copy_oldmem_iter(struct iov_iter *iter, unsigned long src, size_t count)
return res; return res;
} }
int copy_oldmem_kernel(void *dst, unsigned long src, size_t count)
{
struct iov_iter iter;
struct kvec kvec;
kvec.iov_base = dst;
kvec.iov_len = count;
iov_iter_kvec(&iter, WRITE, &kvec, 1, count);
if (copy_oldmem_iter(&iter, src, count) < count)
return -EFAULT;
return 0;
}
/* /*
* Copy one page from "oldmem" * Copy one page from "oldmem"
*/ */
......
...@@ -250,7 +250,7 @@ static debug_info_t *debug_info_alloc(const char *name, int pages_per_area, ...@@ -250,7 +250,7 @@ static debug_info_t *debug_info_alloc(const char *name, int pages_per_area,
rc->level = level; rc->level = level;
rc->buf_size = buf_size; rc->buf_size = buf_size;
rc->entry_size = sizeof(debug_entry_t) + buf_size; rc->entry_size = sizeof(debug_entry_t) + buf_size;
strlcpy(rc->name, name, sizeof(rc->name)); strscpy(rc->name, name, sizeof(rc->name));
memset(rc->views, 0, DEBUG_MAX_VIEWS * sizeof(struct debug_view *)); memset(rc->views, 0, DEBUG_MAX_VIEWS * sizeof(struct debug_view *));
memset(rc->debugfs_entries, 0, DEBUG_MAX_VIEWS * sizeof(struct dentry *)); memset(rc->debugfs_entries, 0, DEBUG_MAX_VIEWS * sizeof(struct dentry *));
refcount_set(&(rc->ref_count), 0); refcount_set(&(rc->ref_count), 0);
......
...@@ -267,7 +267,7 @@ char __bootdata(early_command_line)[COMMAND_LINE_SIZE]; ...@@ -267,7 +267,7 @@ char __bootdata(early_command_line)[COMMAND_LINE_SIZE];
static void __init setup_boot_command_line(void) static void __init setup_boot_command_line(void)
{ {
/* copy arch command line */ /* copy arch command line */
strlcpy(boot_command_line, early_command_line, COMMAND_LINE_SIZE); strscpy(boot_command_line, early_command_line, COMMAND_LINE_SIZE);
} }
static void __init check_image_bootable(void) static void __init check_image_bootable(void)
......
...@@ -29,6 +29,7 @@ ...@@ -29,6 +29,7 @@
#include <asm/sclp.h> #include <asm/sclp.h>
#include <asm/checksum.h> #include <asm/checksum.h>
#include <asm/debug.h> #include <asm/debug.h>
#include <asm/abs_lowcore.h>
#include <asm/os_info.h> #include <asm/os_info.h>
#include <asm/sections.h> #include <asm/sections.h>
#include <asm/boot_data.h> #include <asm/boot_data.h>
...@@ -1642,12 +1643,16 @@ static struct shutdown_action __refdata dump_action = { ...@@ -1642,12 +1643,16 @@ static struct shutdown_action __refdata dump_action = {
static void dump_reipl_run(struct shutdown_trigger *trigger) static void dump_reipl_run(struct shutdown_trigger *trigger)
{ {
unsigned long ipib = (unsigned long) reipl_block_actual; unsigned long ipib = (unsigned long) reipl_block_actual;
struct lowcore *abs_lc;
unsigned long flags;
unsigned int csum; unsigned int csum;
csum = (__force unsigned int) csum = (__force unsigned int)
csum_partial(reipl_block_actual, reipl_block_actual->hdr.len, 0); csum_partial(reipl_block_actual, reipl_block_actual->hdr.len, 0);
put_abs_lowcore(ipib, ipib); abs_lc = get_abs_lowcore(&flags);
put_abs_lowcore(ipib_checksum, csum); abs_lc->ipib = ipib;
abs_lc->ipib_checksum = csum;
put_abs_lowcore(abs_lc, flags);
dump_run(trigger); dump_run(trigger);
} }
......
...@@ -21,6 +21,7 @@ ...@@ -21,6 +21,7 @@
#include <asm/elf.h> #include <asm/elf.h>
#include <asm/asm-offsets.h> #include <asm/asm-offsets.h>
#include <asm/cacheflush.h> #include <asm/cacheflush.h>
#include <asm/abs_lowcore.h>
#include <asm/os_info.h> #include <asm/os_info.h>
#include <asm/set_memory.h> #include <asm/set_memory.h>
#include <asm/stacktrace.h> #include <asm/stacktrace.h>
...@@ -222,13 +223,18 @@ void machine_kexec_cleanup(struct kimage *image) ...@@ -222,13 +223,18 @@ void machine_kexec_cleanup(struct kimage *image)
void arch_crash_save_vmcoreinfo(void) void arch_crash_save_vmcoreinfo(void)
{ {
struct lowcore *abs_lc;
unsigned long flags;
VMCOREINFO_SYMBOL(lowcore_ptr); VMCOREINFO_SYMBOL(lowcore_ptr);
VMCOREINFO_SYMBOL(high_memory); VMCOREINFO_SYMBOL(high_memory);
VMCOREINFO_LENGTH(lowcore_ptr, NR_CPUS); VMCOREINFO_LENGTH(lowcore_ptr, NR_CPUS);
vmcoreinfo_append_str("SAMODE31=%lx\n", __samode31); vmcoreinfo_append_str("SAMODE31=%lx\n", __samode31);
vmcoreinfo_append_str("EAMODE31=%lx\n", __eamode31); vmcoreinfo_append_str("EAMODE31=%lx\n", __eamode31);
vmcoreinfo_append_str("KERNELOFFSET=%lx\n", kaslr_offset()); vmcoreinfo_append_str("KERNELOFFSET=%lx\n", kaslr_offset());
put_abs_lowcore(vmcore_info, paddr_vmcoreinfo_note()); abs_lc = get_abs_lowcore(&flags);
abs_lc->vmcore_info = paddr_vmcoreinfo_note();
put_abs_lowcore(abs_lc, flags);
} }
void machine_shutdown(void) void machine_shutdown(void)
......
...@@ -13,8 +13,9 @@ ...@@ -13,8 +13,9 @@
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/slab.h> #include <linux/slab.h>
#include <asm/checksum.h> #include <asm/checksum.h>
#include <asm/lowcore.h> #include <asm/abs_lowcore.h>
#include <asm/os_info.h> #include <asm/os_info.h>
#include <asm/maccess.h>
#include <asm/asm-offsets.h> #include <asm/asm-offsets.h>
/* /*
...@@ -57,13 +58,16 @@ void os_info_entry_add(int nr, void *ptr, u64 size) ...@@ -57,13 +58,16 @@ void os_info_entry_add(int nr, void *ptr, u64 size)
*/ */
void __init os_info_init(void) void __init os_info_init(void)
{ {
void *ptr = &os_info; struct lowcore *abs_lc;
unsigned long flags;
os_info.version_major = OS_INFO_VERSION_MAJOR; os_info.version_major = OS_INFO_VERSION_MAJOR;
os_info.version_minor = OS_INFO_VERSION_MINOR; os_info.version_minor = OS_INFO_VERSION_MINOR;
os_info.magic = OS_INFO_MAGIC; os_info.magic = OS_INFO_MAGIC;
os_info.csum = os_info_csum(&os_info); os_info.csum = os_info_csum(&os_info);
put_abs_lowcore(os_info, __pa(ptr)); abs_lc = get_abs_lowcore(&flags);
abs_lc->os_info = __pa(&os_info);
put_abs_lowcore(abs_lc, flags);
} }
#ifdef CONFIG_CRASH_DUMP #ifdef CONFIG_CRASH_DUMP
......
// SPDX-License-Identifier: GPL-2.0
/*
* Performance event support - Processor Activity Instrumentation Extension
* Facility
*
* Copyright IBM Corp. 2022
* Author(s): Thomas Richter <tmricht@linux.ibm.com>
*/
#define KMSG_COMPONENT "pai_ext"
#define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
#include <linux/kernel.h>
#include <linux/kernel_stat.h>
#include <linux/percpu.h>
#include <linux/notifier.h>
#include <linux/init.h>
#include <linux/export.h>
#include <linux/io.h>
#include <asm/cpu_mcf.h>
#include <asm/ctl_reg.h>
#include <asm/pai.h>
#include <asm/debug.h>
#define PAIE1_CB_SZ 0x200 /* Size of PAIE1 control block */
#define PAIE1_CTRBLOCK_SZ 0x400 /* Size of PAIE1 counter blocks */
static debug_info_t *paiext_dbg;
static unsigned int paiext_cnt; /* Extracted with QPACI instruction */
enum paiext_mode {
PAI_MODE_NONE,
PAI_MODE_SAMPLING,
PAI_MODE_COUNTER,
};
struct pai_userdata {
u16 num;
u64 value;
} __packed;
/* Create the PAI extension 1 control block area.
* The PAI extension control block 1 is pointed to by lowcore
* address 0x1508 for each CPU. This control block is 512 bytes in size
* and requires a 512 byte boundary alignment.
*/
struct paiext_cb { /* PAI extension 1 control block */
u64 header; /* Not used */
u64 reserved1;
u64 acc; /* Addr to analytics counter control block */
u8 reserved2[488];
} __packed;
struct paiext_map {
unsigned long *area; /* Area for CPU to store counters */
struct pai_userdata *save; /* Area to store non-zero counters */
enum paiext_mode mode; /* Type of event */
unsigned int active_events; /* # of PAI Extension users */
unsigned int refcnt;
struct perf_event *event; /* Perf event for sampling */
struct paiext_cb *paiext_cb; /* PAI extension control block area */
};
struct paiext_mapptr {
struct paiext_map *mapptr;
};
static struct paiext_root { /* Anchor to per CPU data */
int refcnt; /* Overall active events */
struct paiext_mapptr __percpu *mapptr;
} paiext_root;
/* Free per CPU data when the last event is removed. */
static void paiext_root_free(void)
{
if (!--paiext_root.refcnt) {
free_percpu(paiext_root.mapptr);
paiext_root.mapptr = NULL;
}
}
/* On initialization of first event also allocate per CPU data dynamically.
* Start with an array of pointers, the array size is the maximum number of
* CPUs possible, which might be larger than the number of CPUs currently
* online.
*/
static int paiext_root_alloc(void)
{
if (++paiext_root.refcnt == 1) {
/* The memory is already zeroed. */
paiext_root.mapptr = alloc_percpu(struct paiext_mapptr);
if (!paiext_root.mapptr) {
/* Returing without refcnt adjustment is ok. The
* error code is handled by paiext_alloc() which
* decrements refcnt when an event can not be
* created.
*/
return -ENOMEM;
}
}
return 0;
}
/* Protects against concurrent increment of sampler and counter member
* increments at the same time and prohibits concurrent execution of
* counting and sampling events.
* Ensures that analytics counter block is deallocated only when the
* sampling and counting on that cpu is zero.
* For details see paiext_alloc().
*/
static DEFINE_MUTEX(paiext_reserve_mutex);
/* Free all memory allocated for event counting/sampling setup */
static void paiext_free(struct paiext_mapptr *mp)
{
kfree(mp->mapptr->area);
kfree(mp->mapptr->paiext_cb);
kvfree(mp->mapptr->save);
kfree(mp->mapptr);
mp->mapptr = NULL;
}
/* Release the PMU if event is the last perf event */
static void paiext_event_destroy(struct perf_event *event)
{
struct paiext_mapptr *mp = per_cpu_ptr(paiext_root.mapptr, event->cpu);
struct paiext_map *cpump = mp->mapptr;
mutex_lock(&paiext_reserve_mutex);
cpump->event = NULL;
if (!--cpump->refcnt) /* Last reference gone */
paiext_free(mp);
paiext_root_free();
mutex_unlock(&paiext_reserve_mutex);
debug_sprintf_event(paiext_dbg, 4, "%s cpu %d mapptr %p\n", __func__,
event->cpu, mp->mapptr);
}
/* Used to avoid races in checking concurrent access of counting and
* sampling for pai_extension events.
*
* Only one instance of event pai_ext/NNPA_ALL/ for sampling is
* allowed and when this event is running, no counting event is allowed.
* Several counting events are allowed in parallel, but no sampling event
* is allowed while one (or more) counting events are running.
*
* This function is called in process context and it is safe to block.
* When the event initialization functions fails, no other call back will
* be invoked.
*
* Allocate the memory for the event.
*/
static int paiext_alloc(struct perf_event_attr *a, struct perf_event *event)
{
struct paiext_mapptr *mp;
struct paiext_map *cpump;
int rc;
mutex_lock(&paiext_reserve_mutex);
rc = paiext_root_alloc();
if (rc)
goto unlock;
mp = per_cpu_ptr(paiext_root.mapptr, event->cpu);
cpump = mp->mapptr;
if (!cpump) { /* Paiext_map allocated? */
rc = -ENOMEM;
cpump = kzalloc(sizeof(*cpump), GFP_KERNEL);
if (!cpump)
goto unlock;
/* Allocate memory for counter area and counter extraction.
* These are
* - a 512 byte block and requires 512 byte boundary alignment.
* - a 1KB byte block and requires 1KB boundary alignment.
* Only the first counting event has to allocate the area.
*
* Note: This works with commit 59bb47985c1d by default.
* Backporting this to kernels without this commit might
* need adjustment.
*/
mp->mapptr = cpump;
cpump->area = kzalloc(PAIE1_CTRBLOCK_SZ, GFP_KERNEL);
cpump->paiext_cb = kzalloc(PAIE1_CB_SZ, GFP_KERNEL);
cpump->save = kvmalloc_array(paiext_cnt + 1,
sizeof(struct pai_userdata),
GFP_KERNEL);
if (!cpump->save || !cpump->area || !cpump->paiext_cb) {
paiext_free(mp);
goto unlock;
}
cpump->mode = a->sample_period ? PAI_MODE_SAMPLING
: PAI_MODE_COUNTER;
} else {
/* Multiple invocation, check whats active.
* Supported are multiple counter events or only one sampling
* event concurrently at any one time.
*/
if (cpump->mode == PAI_MODE_SAMPLING ||
(cpump->mode == PAI_MODE_COUNTER && a->sample_period)) {
rc = -EBUSY;
goto unlock;
}
}
rc = 0;
cpump->event = event;
++cpump->refcnt;
unlock:
if (rc) {
/* Error in allocation of event, decrement anchor. Since
* the event in not created, its destroy() function is never
* invoked. Adjust the reference counter for the anchor.
*/
paiext_root_free();
}
mutex_unlock(&paiext_reserve_mutex);
/* If rc is non-zero, no increment of counter/sampler was done. */
return rc;
}
/* The PAI extension 1 control block supports up to 128 entries. Return
* the index within PAIE1_CB given the event number. Also validate event
* number.
*/
static int paiext_event_valid(struct perf_event *event)
{
u64 cfg = event->attr.config;
if (cfg >= PAI_NNPA_BASE && cfg <= PAI_NNPA_BASE + paiext_cnt) {
/* Offset NNPA in paiext_cb */
event->hw.config_base = offsetof(struct paiext_cb, acc);
return 0;
}
return -EINVAL;
}
/* Might be called on different CPU than the one the event is intended for. */
static int paiext_event_init(struct perf_event *event)
{
struct perf_event_attr *a = &event->attr;
int rc;
/* PMU pai_ext registered as PERF_TYPE_RAW, check event type */
if (a->type != PERF_TYPE_RAW && event->pmu->type != a->type)
return -ENOENT;
/* PAI extension event must be valid and in supported range */
rc = paiext_event_valid(event);
if (rc)
return rc;
/* Allow only CPU wide operation, no process context for now. */
if (event->hw.target || event->cpu == -1)
return -ENOENT;
/* Allow only event NNPA_ALL for sampling. */
if (a->sample_period && a->config != PAI_NNPA_BASE)
return -EINVAL;
/* Prohibit exclude_user event selection */
if (a->exclude_user)
return -EINVAL;
rc = paiext_alloc(a, event);
if (rc)
return rc;
event->hw.last_tag = 0;
event->destroy = paiext_event_destroy;
if (a->sample_period) {
a->sample_period = 1;
a->freq = 0;
/* Register for paicrypt_sched_task() to be called */
event->attach_state |= PERF_ATTACH_SCHED_CB;
/* Add raw data which are the memory mapped counters */
a->sample_type |= PERF_SAMPLE_RAW;
/* Turn off inheritance */
a->inherit = 0;
}
return 0;
}
static u64 paiext_getctr(struct paiext_map *cpump, int nr)
{
return cpump->area[nr];
}
/* Read the counter values. Return value from location in buffer. For event
* NNPA_ALL sum up all events.
*/
static u64 paiext_getdata(struct perf_event *event)
{
struct paiext_mapptr *mp = this_cpu_ptr(paiext_root.mapptr);
struct paiext_map *cpump = mp->mapptr;
u64 sum = 0;
int i;
if (event->attr.config != PAI_NNPA_BASE)
return paiext_getctr(cpump, event->attr.config - PAI_NNPA_BASE);
for (i = 1; i <= paiext_cnt; i++)
sum += paiext_getctr(cpump, i);
return sum;
}
static u64 paiext_getall(struct perf_event *event)
{
return paiext_getdata(event);
}
static void paiext_read(struct perf_event *event)
{
u64 prev, new, delta;
prev = local64_read(&event->hw.prev_count);
new = paiext_getall(event);
local64_set(&event->hw.prev_count, new);
delta = new - prev;
local64_add(delta, &event->count);
}
static void paiext_start(struct perf_event *event, int flags)
{
u64 sum;
if (event->hw.last_tag)
return;
event->hw.last_tag = 1;
sum = paiext_getall(event); /* Get current value */
local64_set(&event->hw.prev_count, sum);
local64_set(&event->count, 0);
}
static int paiext_add(struct perf_event *event, int flags)
{
struct paiext_mapptr *mp = this_cpu_ptr(paiext_root.mapptr);
struct paiext_map *cpump = mp->mapptr;
struct paiext_cb *pcb = cpump->paiext_cb;
if (++cpump->active_events == 1) {
S390_lowcore.aicd = virt_to_phys(cpump->paiext_cb);
pcb->acc = virt_to_phys(cpump->area) | 0x1;
/* Enable CPU instruction lookup for PAIE1 control block */
__ctl_set_bit(0, 49);
debug_sprintf_event(paiext_dbg, 4, "%s 1508 %llx acc %llx\n",
__func__, S390_lowcore.aicd, pcb->acc);
}
if (flags & PERF_EF_START && !event->attr.sample_period) {
/* Only counting needs initial counter value */
paiext_start(event, PERF_EF_RELOAD);
}
event->hw.state = 0;
if (event->attr.sample_period) {
cpump->event = event;
perf_sched_cb_inc(event->pmu);
}
return 0;
}
static void paiext_stop(struct perf_event *event, int flags)
{
paiext_read(event);
event->hw.state = PERF_HES_STOPPED;
}
static void paiext_del(struct perf_event *event, int flags)
{
struct paiext_mapptr *mp = this_cpu_ptr(paiext_root.mapptr);
struct paiext_map *cpump = mp->mapptr;
struct paiext_cb *pcb = cpump->paiext_cb;
if (event->attr.sample_period)
perf_sched_cb_dec(event->pmu);
if (!event->attr.sample_period) {
/* Only counting needs to read counter */
paiext_stop(event, PERF_EF_UPDATE);
}
if (--cpump->active_events == 0) {
/* Disable CPU instruction lookup for PAIE1 control block */
__ctl_clear_bit(0, 49);
pcb->acc = 0;
S390_lowcore.aicd = 0;
debug_sprintf_event(paiext_dbg, 4, "%s 1508 %llx acc %llx\n",
__func__, S390_lowcore.aicd, pcb->acc);
}
}
/* Create raw data and save it in buffer. Returns number of bytes copied.
* Saves only positive counter entries of the form
* 2 bytes: Number of counter
* 8 bytes: Value of counter
*/
static size_t paiext_copy(struct paiext_map *cpump)
{
struct pai_userdata *userdata = cpump->save;
int i, outidx = 0;
for (i = 1; i <= paiext_cnt; i++) {
u64 val = paiext_getctr(cpump, i);
if (val) {
userdata[outidx].num = i;
userdata[outidx].value = val;
outidx++;
}
}
return outidx * sizeof(*userdata);
}
/* Write sample when one or more counters values are nonzero.
*
* Note: The function paiext_sched_task() and paiext_push_sample() are not
* invoked after function paiext_del() has been called because of function
* perf_sched_cb_dec().
* The function paiext_sched_task() and paiext_push_sample() are only
* called when sampling is active. Function perf_sched_cb_inc()
* has been invoked to install function paiext_sched_task() as call back
* to run at context switch time (see paiext_add()).
*
* This causes function perf_event_context_sched_out() and
* perf_event_context_sched_in() to check whether the PMU has installed an
* sched_task() callback. That callback is not active after paiext_del()
* returns and has deleted the event on that CPU.
*/
static int paiext_push_sample(void)
{
struct paiext_mapptr *mp = this_cpu_ptr(paiext_root.mapptr);
struct paiext_map *cpump = mp->mapptr;
struct perf_event *event = cpump->event;
struct perf_sample_data data;
struct perf_raw_record raw;
struct pt_regs regs;
size_t rawsize;
int overflow;
rawsize = paiext_copy(cpump);
if (!rawsize) /* No incremented counters */
return 0;
/* Setup perf sample */
memset(&regs, 0, sizeof(regs));
memset(&raw, 0, sizeof(raw));
memset(&data, 0, sizeof(data));
perf_sample_data_init(&data, 0, event->hw.last_period);
if (event->attr.sample_type & PERF_SAMPLE_TID) {
data.tid_entry.pid = task_tgid_nr(current);
data.tid_entry.tid = task_pid_nr(current);
}
if (event->attr.sample_type & PERF_SAMPLE_TIME)
data.time = event->clock();
if (event->attr.sample_type & (PERF_SAMPLE_ID | PERF_SAMPLE_IDENTIFIER))
data.id = event->id;
if (event->attr.sample_type & PERF_SAMPLE_CPU)
data.cpu_entry.cpu = smp_processor_id();
if (event->attr.sample_type & PERF_SAMPLE_RAW) {
raw.frag.size = rawsize;
raw.frag.data = cpump->save;
raw.size = raw.frag.size;
data.raw = &raw;
}
overflow = perf_event_overflow(event, &data, &regs);
perf_event_update_userpage(event);
/* Clear lowcore area after read */
memset(cpump->area, 0, PAIE1_CTRBLOCK_SZ);
return overflow;
}
/* Called on schedule-in and schedule-out. No access to event structure,
* but for sampling only event NNPA_ALL is allowed.
*/
static void paiext_sched_task(struct perf_event_context *ctx, bool sched_in)
{
/* We started with a clean page on event installation. So read out
* results on schedule_out and if page was dirty, clear values.
*/
if (!sched_in)
paiext_push_sample();
}
/* Attribute definitions for pai extension1 interface. As with other CPU
* Measurement Facilities, there is one attribute per mapped counter.
* The number of mapped counters may vary per machine generation. Use
* the QUERY PROCESSOR ACTIVITY COUNTER INFORMATION (QPACI) instruction
* to determine the number of mapped counters. The instructions returns
* a positive number, which is the highest number of supported counters.
* All counters less than this number are also supported, there are no
* holes. A returned number of zero means no support for mapped counters.
*
* The identification of the counter is a unique number. The chosen range
* is 0x1800 + offset in mapped kernel page.
* All CPU Measurement Facility counters identifiers must be unique and
* the numbers from 0 to 496 are already used for the CPU Measurement
* Counter facility. Number 0x1000 to 0x103e are used for PAI cryptography
* counters.
* Numbers 0xb0000, 0xbc000 and 0xbd000 are already
* used for the CPU Measurement Sampling facility.
*/
PMU_FORMAT_ATTR(event, "config:0-63");
static struct attribute *paiext_format_attr[] = {
&format_attr_event.attr,
NULL,
};
static struct attribute_group paiext_events_group = {
.name = "events",
.attrs = NULL, /* Filled in attr_event_init() */
};
static struct attribute_group paiext_format_group = {
.name = "format",
.attrs = paiext_format_attr,
};
static const struct attribute_group *paiext_attr_groups[] = {
&paiext_events_group,
&paiext_format_group,
NULL,
};
/* Performance monitoring unit for mapped counters */
static struct pmu paiext = {
.task_ctx_nr = perf_invalid_context,
.event_init = paiext_event_init,
.add = paiext_add,
.del = paiext_del,
.start = paiext_start,
.stop = paiext_stop,
.read = paiext_read,
.sched_task = paiext_sched_task,
.attr_groups = paiext_attr_groups,
};
/* List of symbolic PAI extension 1 NNPA counter names. */
static const char * const paiext_ctrnames[] = {
[0] = "NNPA_ALL",
[1] = "NNPA_ADD",
[2] = "NNPA_SUB",
[3] = "NNPA_MUL",
[4] = "NNPA_DIV",
[5] = "NNPA_MIN",
[6] = "NNPA_MAX",
[7] = "NNPA_LOG",
[8] = "NNPA_EXP",
[9] = "NNPA_IBM_RESERVED_9",
[10] = "NNPA_RELU",
[11] = "NNPA_TANH",
[12] = "NNPA_SIGMOID",
[13] = "NNPA_SOFTMAX",
[14] = "NNPA_BATCHNORM",
[15] = "NNPA_MAXPOOL2D",
[16] = "NNPA_AVGPOOL2D",
[17] = "NNPA_LSTMACT",
[18] = "NNPA_GRUACT",
[19] = "NNPA_CONVOLUTION",
[20] = "NNPA_MATMUL_OP",
[21] = "NNPA_MATMUL_OP_BCAST23",
[22] = "NNPA_SMALLBATCH",
[23] = "NNPA_LARGEDIM",
[24] = "NNPA_SMALLTENSOR",
[25] = "NNPA_1MFRAME",
[26] = "NNPA_2GFRAME",
[27] = "NNPA_ACCESSEXCEPT",
};
static void __init attr_event_free(struct attribute **attrs, int num)
{
struct perf_pmu_events_attr *pa;
struct device_attribute *dap;
int i;
for (i = 0; i < num; i++) {
dap = container_of(attrs[i], struct device_attribute, attr);
pa = container_of(dap, struct perf_pmu_events_attr, attr);
kfree(pa);
}
kfree(attrs);
}
static int __init attr_event_init_one(struct attribute **attrs, int num)
{
struct perf_pmu_events_attr *pa;
pa = kzalloc(sizeof(*pa), GFP_KERNEL);
if (!pa)
return -ENOMEM;
sysfs_attr_init(&pa->attr.attr);
pa->id = PAI_NNPA_BASE + num;
pa->attr.attr.name = paiext_ctrnames[num];
pa->attr.attr.mode = 0444;
pa->attr.show = cpumf_events_sysfs_show;
pa->attr.store = NULL;
attrs[num] = &pa->attr.attr;
return 0;
}
/* Create PMU sysfs event attributes on the fly. */
static int __init attr_event_init(void)
{
struct attribute **attrs;
int ret, i;
attrs = kmalloc_array(ARRAY_SIZE(paiext_ctrnames) + 1, sizeof(*attrs),
GFP_KERNEL);
if (!attrs)
return -ENOMEM;
for (i = 0; i < ARRAY_SIZE(paiext_ctrnames); i++) {
ret = attr_event_init_one(attrs, i);
if (ret) {
attr_event_free(attrs, i - 1);
return ret;
}
}
attrs[i] = NULL;
paiext_events_group.attrs = attrs;
return 0;
}
static int __init paiext_init(void)
{
struct qpaci_info_block ib;
int rc = -ENOMEM;
if (!test_facility(197))
return 0;
qpaci(&ib);
paiext_cnt = ib.num_nnpa;
if (paiext_cnt >= PAI_NNPA_MAXCTR)
paiext_cnt = PAI_NNPA_MAXCTR;
if (!paiext_cnt)
return 0;
rc = attr_event_init();
if (rc) {
pr_err("Creation of PMU " KMSG_COMPONENT " /sysfs failed\n");
return rc;
}
/* Setup s390dbf facility */
paiext_dbg = debug_register(KMSG_COMPONENT, 2, 256, 128);
if (!paiext_dbg) {
pr_err("Registration of s390dbf " KMSG_COMPONENT " failed\n");
rc = -ENOMEM;
goto out_init;
}
debug_register_view(paiext_dbg, &debug_sprintf_view);
rc = perf_pmu_register(&paiext, KMSG_COMPONENT, -1);
if (rc) {
pr_err("Registration of " KMSG_COMPONENT " PMU failed with "
"rc=%i\n", rc);
goto out_pmu;
}
return 0;
out_pmu:
debug_unregister_view(paiext_dbg, &debug_sprintf_view);
debug_unregister(paiext_dbg);
out_init:
attr_event_free(paiext_events_group.attrs,
ARRAY_SIZE(paiext_ctrnames) + 1);
return rc;
}
device_initcall(paiext_init);
...@@ -58,7 +58,7 @@ ...@@ -58,7 +58,7 @@
#include <asm/smp.h> #include <asm/smp.h>
#include <asm/mmu_context.h> #include <asm/mmu_context.h>
#include <asm/cpcmd.h> #include <asm/cpcmd.h>
#include <asm/lowcore.h> #include <asm/abs_lowcore.h>
#include <asm/nmi.h> #include <asm/nmi.h>
#include <asm/irq.h> #include <asm/irq.h>
#include <asm/page.h> #include <asm/page.h>
...@@ -74,6 +74,7 @@ ...@@ -74,6 +74,7 @@
#include <asm/alternative.h> #include <asm/alternative.h>
#include <asm/nospec-branch.h> #include <asm/nospec-branch.h>
#include <asm/mem_detect.h> #include <asm/mem_detect.h>
#include <asm/maccess.h>
#include <asm/uv.h> #include <asm/uv.h>
#include <asm/asm-offsets.h> #include <asm/asm-offsets.h>
#include "entry.h" #include "entry.h"
...@@ -395,6 +396,7 @@ void __init arch_call_rest_init(void) ...@@ -395,6 +396,7 @@ void __init arch_call_rest_init(void)
{ {
unsigned long stack; unsigned long stack;
smp_reinit_ipl_cpu();
stack = stack_alloc(); stack = stack_alloc();
if (!stack) if (!stack)
panic("Couldn't allocate kernel stack"); panic("Couldn't allocate kernel stack");
...@@ -411,8 +413,9 @@ void __init arch_call_rest_init(void) ...@@ -411,8 +413,9 @@ void __init arch_call_rest_init(void)
static void __init setup_lowcore_dat_off(void) static void __init setup_lowcore_dat_off(void)
{ {
unsigned long int_psw_mask = PSW_KERNEL_BITS; unsigned long int_psw_mask = PSW_KERNEL_BITS;
struct lowcore *abs_lc, *lc;
unsigned long mcck_stack; unsigned long mcck_stack;
struct lowcore *lc; unsigned long flags;
if (IS_ENABLED(CONFIG_KASAN)) if (IS_ENABLED(CONFIG_KASAN))
int_psw_mask |= PSW_MASK_DAT; int_psw_mask |= PSW_MASK_DAT;
...@@ -474,12 +477,14 @@ static void __init setup_lowcore_dat_off(void) ...@@ -474,12 +477,14 @@ static void __init setup_lowcore_dat_off(void)
lc->restart_data = 0; lc->restart_data = 0;
lc->restart_source = -1U; lc->restart_source = -1U;
put_abs_lowcore(restart_stack, lc->restart_stack); abs_lc = get_abs_lowcore(&flags);
put_abs_lowcore(restart_fn, lc->restart_fn); abs_lc->restart_stack = lc->restart_stack;
put_abs_lowcore(restart_data, lc->restart_data); abs_lc->restart_fn = lc->restart_fn;
put_abs_lowcore(restart_source, lc->restart_source); abs_lc->restart_data = lc->restart_data;
put_abs_lowcore(restart_psw, lc->restart_psw); abs_lc->restart_source = lc->restart_source;
put_abs_lowcore(mcesad, lc->mcesad); abs_lc->restart_psw = lc->restart_psw;
abs_lc->mcesad = lc->mcesad;
put_abs_lowcore(abs_lc, flags);
mcck_stack = (unsigned long)memblock_alloc(THREAD_SIZE, THREAD_SIZE); mcck_stack = (unsigned long)memblock_alloc(THREAD_SIZE, THREAD_SIZE);
if (!mcck_stack) if (!mcck_stack)
...@@ -500,8 +505,8 @@ static void __init setup_lowcore_dat_off(void) ...@@ -500,8 +505,8 @@ static void __init setup_lowcore_dat_off(void)
static void __init setup_lowcore_dat_on(void) static void __init setup_lowcore_dat_on(void)
{ {
struct lowcore *lc = lowcore_ptr[0]; struct lowcore *abs_lc;
int cr; unsigned long flags;
__ctl_clear_bit(0, 28); __ctl_clear_bit(0, 28);
S390_lowcore.external_new_psw.mask |= PSW_MASK_DAT; S390_lowcore.external_new_psw.mask |= PSW_MASK_DAT;
...@@ -510,10 +515,15 @@ static void __init setup_lowcore_dat_on(void) ...@@ -510,10 +515,15 @@ static void __init setup_lowcore_dat_on(void)
S390_lowcore.io_new_psw.mask |= PSW_MASK_DAT; S390_lowcore.io_new_psw.mask |= PSW_MASK_DAT;
__ctl_set_bit(0, 28); __ctl_set_bit(0, 28);
__ctl_store(S390_lowcore.cregs_save_area, 0, 15); __ctl_store(S390_lowcore.cregs_save_area, 0, 15);
put_abs_lowcore(restart_flags, RESTART_FLAG_CTLREGS); if (abs_lowcore_map(0, lowcore_ptr[0], true))
put_abs_lowcore(program_new_psw, lc->program_new_psw); panic("Couldn't setup absolute lowcore");
for (cr = 0; cr < ARRAY_SIZE(lc->cregs_save_area); cr++) abs_lowcore_mapped = true;
put_abs_lowcore(cregs_save_area[cr], lc->cregs_save_area[cr]); abs_lc = get_abs_lowcore(&flags);
abs_lc->restart_flags = RESTART_FLAG_CTLREGS;
abs_lc->program_new_psw = S390_lowcore.program_new_psw;
memcpy(abs_lc->cregs_save_area, S390_lowcore.cregs_save_area,
sizeof(abs_lc->cregs_save_area));
put_abs_lowcore(abs_lc, flags);
} }
static struct resource code_resource = { static struct resource code_resource = {
...@@ -1019,10 +1029,10 @@ void __init setup_arch(char **cmdline_p) ...@@ -1019,10 +1029,10 @@ void __init setup_arch(char **cmdline_p)
reserve_crashkernel(); reserve_crashkernel();
#ifdef CONFIG_CRASH_DUMP #ifdef CONFIG_CRASH_DUMP
/* /*
* Be aware that smp_save_dump_cpus() triggers a system reset. * Be aware that smp_save_dump_secondary_cpus() triggers a system reset.
* Therefore CPU and device initialization should be done afterwards. * Therefore CPU and device initialization should be done afterwards.
*/ */
smp_save_dump_cpus(); smp_save_dump_secondary_cpus();
#endif #endif
setup_resources(); setup_resources();
...@@ -1041,12 +1051,15 @@ void __init setup_arch(char **cmdline_p) ...@@ -1041,12 +1051,15 @@ void __init setup_arch(char **cmdline_p)
* Create kernel page tables and switch to virtual addressing. * Create kernel page tables and switch to virtual addressing.
*/ */
paging_init(); paging_init();
memcpy_real_init();
/* /*
* After paging_init created the kernel page table, the new PSWs * After paging_init created the kernel page table, the new PSWs
* in lowcore can now run with DAT enabled. * in lowcore can now run with DAT enabled.
*/ */
setup_lowcore_dat_on(); setup_lowcore_dat_on();
#ifdef CONFIG_CRASH_DUMP
smp_save_dump_ipl_cpu();
#endif
/* Setup default console */ /* Setup default console */
conmode_default(); conmode_default();
......
...@@ -45,7 +45,7 @@ ...@@ -45,7 +45,7 @@
#include <asm/irq.h> #include <asm/irq.h>
#include <asm/tlbflush.h> #include <asm/tlbflush.h>
#include <asm/vtimer.h> #include <asm/vtimer.h>
#include <asm/lowcore.h> #include <asm/abs_lowcore.h>
#include <asm/sclp.h> #include <asm/sclp.h>
#include <asm/debug.h> #include <asm/debug.h>
#include <asm/os_info.h> #include <asm/os_info.h>
...@@ -55,6 +55,7 @@ ...@@ -55,6 +55,7 @@
#include <asm/stacktrace.h> #include <asm/stacktrace.h>
#include <asm/topology.h> #include <asm/topology.h>
#include <asm/vdso.h> #include <asm/vdso.h>
#include <asm/maccess.h>
#include "entry.h" #include "entry.h"
enum { enum {
...@@ -212,10 +213,14 @@ static int pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu) ...@@ -212,10 +213,14 @@ static int pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu)
lc->preempt_count = PREEMPT_DISABLED; lc->preempt_count = PREEMPT_DISABLED;
if (nmi_alloc_mcesa(&lc->mcesad)) if (nmi_alloc_mcesa(&lc->mcesad))
goto out; goto out;
if (abs_lowcore_map(cpu, lc, true))
goto out_mcesa;
lowcore_ptr[cpu] = lc; lowcore_ptr[cpu] = lc;
pcpu_sigp_retry(pcpu, SIGP_SET_PREFIX, __pa(lc)); pcpu_sigp_retry(pcpu, SIGP_SET_PREFIX, __pa(lc));
return 0; return 0;
out_mcesa:
nmi_free_mcesa(&lc->mcesad);
out: out:
stack_free(mcck_stack); stack_free(mcck_stack);
stack_free(async_stack); stack_free(async_stack);
...@@ -237,6 +242,7 @@ static void pcpu_free_lowcore(struct pcpu *pcpu) ...@@ -237,6 +242,7 @@ static void pcpu_free_lowcore(struct pcpu *pcpu)
mcck_stack = lc->mcck_stack - STACK_INIT_OFFSET; mcck_stack = lc->mcck_stack - STACK_INIT_OFFSET;
pcpu_sigp_retry(pcpu, SIGP_SET_PREFIX, 0); pcpu_sigp_retry(pcpu, SIGP_SET_PREFIX, 0);
lowcore_ptr[cpu] = NULL; lowcore_ptr[cpu] = NULL;
abs_lowcore_unmap(cpu);
nmi_free_mcesa(&lc->mcesad); nmi_free_mcesa(&lc->mcesad);
stack_free(async_stack); stack_free(async_stack);
stack_free(mcck_stack); stack_free(mcck_stack);
...@@ -315,9 +321,12 @@ static void pcpu_delegate(struct pcpu *pcpu, ...@@ -315,9 +321,12 @@ static void pcpu_delegate(struct pcpu *pcpu,
pcpu_delegate_fn *func, pcpu_delegate_fn *func,
void *data, unsigned long stack) void *data, unsigned long stack)
{ {
struct lowcore *lc = lowcore_ptr[pcpu - pcpu_devices]; struct lowcore *lc, *abs_lc;
unsigned int source_cpu = stap(); unsigned int source_cpu;
unsigned long flags;
lc = lowcore_ptr[pcpu - pcpu_devices];
source_cpu = stap();
__load_psw_mask(PSW_KERNEL_BITS | PSW_MASK_DAT); __load_psw_mask(PSW_KERNEL_BITS | PSW_MASK_DAT);
if (pcpu->address == source_cpu) { if (pcpu->address == source_cpu) {
call_on_stack(2, stack, void, __pcpu_delegate, call_on_stack(2, stack, void, __pcpu_delegate,
...@@ -332,10 +341,12 @@ static void pcpu_delegate(struct pcpu *pcpu, ...@@ -332,10 +341,12 @@ static void pcpu_delegate(struct pcpu *pcpu,
lc->restart_data = (unsigned long)data; lc->restart_data = (unsigned long)data;
lc->restart_source = source_cpu; lc->restart_source = source_cpu;
} else { } else {
put_abs_lowcore(restart_stack, stack); abs_lc = get_abs_lowcore(&flags);
put_abs_lowcore(restart_fn, (unsigned long)func); abs_lc->restart_stack = stack;
put_abs_lowcore(restart_data, (unsigned long)data); abs_lc->restart_fn = (unsigned long)func;
put_abs_lowcore(restart_source, source_cpu); abs_lc->restart_data = (unsigned long)data;
abs_lc->restart_source = source_cpu;
put_abs_lowcore(abs_lc, flags);
} }
__bpon(); __bpon();
asm volatile( asm volatile(
...@@ -581,6 +592,8 @@ static DEFINE_SPINLOCK(ctl_lock); ...@@ -581,6 +592,8 @@ static DEFINE_SPINLOCK(ctl_lock);
void smp_ctl_set_clear_bit(int cr, int bit, bool set) void smp_ctl_set_clear_bit(int cr, int bit, bool set)
{ {
struct ec_creg_mask_parms parms = { .cr = cr, }; struct ec_creg_mask_parms parms = { .cr = cr, };
struct lowcore *abs_lc;
unsigned long flags;
u64 ctlreg; u64 ctlreg;
if (set) { if (set) {
...@@ -591,9 +604,11 @@ void smp_ctl_set_clear_bit(int cr, int bit, bool set) ...@@ -591,9 +604,11 @@ void smp_ctl_set_clear_bit(int cr, int bit, bool set)
parms.andval = ~(1UL << bit); parms.andval = ~(1UL << bit);
} }
spin_lock(&ctl_lock); spin_lock(&ctl_lock);
get_abs_lowcore(ctlreg, cregs_save_area[cr]); abs_lc = get_abs_lowcore(&flags);
ctlreg = abs_lc->cregs_save_area[cr];
ctlreg = (ctlreg & parms.andval) | parms.orval; ctlreg = (ctlreg & parms.andval) | parms.orval;
put_abs_lowcore(cregs_save_area[cr], ctlreg); abs_lc->cregs_save_area[cr] = ctlreg;
put_abs_lowcore(abs_lc, flags);
spin_unlock(&ctl_lock); spin_unlock(&ctl_lock);
on_each_cpu(smp_ctl_bit_callback, &parms, 1); on_each_cpu(smp_ctl_bit_callback, &parms, 1);
} }
...@@ -650,35 +665,36 @@ int smp_store_status(int cpu) ...@@ -650,35 +665,36 @@ int smp_store_status(int cpu)
* This case does not exist for s390 anymore, setup_arch explicitly * This case does not exist for s390 anymore, setup_arch explicitly
* deactivates the elfcorehdr= kernel parameter * deactivates the elfcorehdr= kernel parameter
*/ */
static __init void smp_save_cpu_vxrs(struct save_area *sa, u16 addr, static bool dump_available(void)
bool is_boot_cpu, __vector128 *vxrs)
{ {
if (is_boot_cpu) return oldmem_data.start || is_ipl_type_dump();
vxrs = boot_cpu_vector_save_area;
else
__pcpu_sigp_relax(addr, SIGP_STORE_ADDITIONAL_STATUS, __pa(vxrs));
save_area_add_vxrs(sa, vxrs);
} }
static __init void smp_save_cpu_regs(struct save_area *sa, u16 addr, void __init smp_save_dump_ipl_cpu(void)
bool is_boot_cpu, void *regs)
{ {
if (is_boot_cpu) struct save_area *sa;
copy_oldmem_kernel(regs, __LC_FPREGS_SAVE_AREA, 512); void *regs;
else
__pcpu_sigp_relax(addr, SIGP_STORE_STATUS_AT_ADDRESS, __pa(regs)); if (!dump_available())
return;
sa = save_area_alloc(true);
regs = memblock_alloc(512, 8);
if (!sa || !regs)
panic("could not allocate memory for boot CPU save area\n");
copy_oldmem_kernel(regs, __LC_FPREGS_SAVE_AREA, 512);
save_area_add_regs(sa, regs); save_area_add_regs(sa, regs);
memblock_free(regs, 512);
if (MACHINE_HAS_VX)
save_area_add_vxrs(sa, boot_cpu_vector_save_area);
} }
void __init smp_save_dump_cpus(void) void __init smp_save_dump_secondary_cpus(void)
{ {
int addr, boot_cpu_addr, max_cpu_addr; int addr, boot_cpu_addr, max_cpu_addr;
struct save_area *sa; struct save_area *sa;
bool is_boot_cpu;
void *page; void *page;
if (!(oldmem_data.start || is_ipl_type_dump())) if (!dump_available())
/* No previous system present, normal boot. */
return; return;
/* Allocate a page as dumping area for the store status sigps */ /* Allocate a page as dumping area for the store status sigps */
page = memblock_alloc_low(PAGE_SIZE, PAGE_SIZE); page = memblock_alloc_low(PAGE_SIZE, PAGE_SIZE);
...@@ -691,26 +707,20 @@ void __init smp_save_dump_cpus(void) ...@@ -691,26 +707,20 @@ void __init smp_save_dump_cpus(void)
boot_cpu_addr = stap(); boot_cpu_addr = stap();
max_cpu_addr = SCLP_MAX_CORES << sclp.mtid_prev; max_cpu_addr = SCLP_MAX_CORES << sclp.mtid_prev;
for (addr = 0; addr <= max_cpu_addr; addr++) { for (addr = 0; addr <= max_cpu_addr; addr++) {
if (addr == boot_cpu_addr)
continue;
if (__pcpu_sigp_relax(addr, SIGP_SENSE, 0) == if (__pcpu_sigp_relax(addr, SIGP_SENSE, 0) ==
SIGP_CC_NOT_OPERATIONAL) SIGP_CC_NOT_OPERATIONAL)
continue; continue;
is_boot_cpu = (addr == boot_cpu_addr); sa = save_area_alloc(false);
/* Allocate save area */
sa = save_area_alloc(is_boot_cpu);
if (!sa) if (!sa)
panic("could not allocate memory for save area\n"); panic("could not allocate memory for save area\n");
if (MACHINE_HAS_VX) __pcpu_sigp_relax(addr, SIGP_STORE_STATUS_AT_ADDRESS, __pa(page));
/* Get the vector registers */ save_area_add_regs(sa, page);
smp_save_cpu_vxrs(sa, addr, is_boot_cpu, page); if (MACHINE_HAS_VX) {
/* __pcpu_sigp_relax(addr, SIGP_STORE_ADDITIONAL_STATUS, __pa(page));
* For a zfcp/nvme dump OLDMEM_BASE == NULL and the registers save_area_add_vxrs(sa, page);
* of the boot CPU are stored in the HSA. To retrieve }
* these registers an SCLP request is required which is
* done by drivers/s390/char/zcore.c:init_cpu_info()
*/
if (!is_boot_cpu || oldmem_data.start)
/* Get the CPU registers */
smp_save_cpu_regs(sa, addr, is_boot_cpu, page);
} }
memblock_free(page, PAGE_SIZE); memblock_free(page, PAGE_SIZE);
diag_amode31_ops.diag308_reset(); diag_amode31_ops.diag308_reset();
...@@ -1256,7 +1266,7 @@ static __always_inline void set_new_lowcore(struct lowcore *lc) ...@@ -1256,7 +1266,7 @@ static __always_inline void set_new_lowcore(struct lowcore *lc)
: "memory", "cc"); : "memory", "cc");
} }
static int __init smp_reinit_ipl_cpu(void) int __init smp_reinit_ipl_cpu(void)
{ {
unsigned long async_stack, nodat_stack, mcck_stack; unsigned long async_stack, nodat_stack, mcck_stack;
struct lowcore *lc, *lc_ipl; struct lowcore *lc, *lc_ipl;
...@@ -1281,6 +1291,8 @@ static int __init smp_reinit_ipl_cpu(void) ...@@ -1281,6 +1291,8 @@ static int __init smp_reinit_ipl_cpu(void)
__ctl_clear_bit(0, 28); /* disable lowcore protection */ __ctl_clear_bit(0, 28); /* disable lowcore protection */
S390_lowcore.mcesad = mcesad; S390_lowcore.mcesad = mcesad;
__ctl_load(cr0, 0, 0); __ctl_load(cr0, 0, 0);
if (abs_lowcore_map(0, lc, false))
panic("Couldn't remap absolute lowcore");
lowcore_ptr[0] = lc; lowcore_ptr[0] = lc;
local_mcck_enable(); local_mcck_enable();
local_irq_restore(flags); local_irq_restore(flags);
...@@ -1291,4 +1303,3 @@ static int __init smp_reinit_ipl_cpu(void) ...@@ -1291,4 +1303,3 @@ static int __init smp_reinit_ipl_cpu(void)
return 0; return 0;
} }
early_initcall(smp_reinit_ipl_cpu);
...@@ -13,13 +13,10 @@ ...@@ -13,13 +13,10 @@
void __delay(unsigned long loops) void __delay(unsigned long loops)
{ {
/* /*
* To end the bloody studid and useless discussion about the * Loop 'loops' times. Callers must not assume a specific
* BogoMips number I took the liberty to define the __delay * amount of time passes before this function returns.
* function in a way that that resulting BogoMips number will */
* yield the megahertz number of the cpu. The important function
* is udelay and that is done using the tod clock. -- martin.
*/
asm volatile("0: brct %0,0b" : : "d" ((loops/2) + 1)); asm volatile("0: brct %0,0b" : : "d" ((loops/2) + 1));
} }
EXPORT_SYMBOL(__delay); EXPORT_SYMBOL(__delay);
......
...@@ -8,8 +8,10 @@ ...@@ -8,8 +8,10 @@
#include <linux/kasan.h> #include <linux/kasan.h>
#include <asm/ptdump.h> #include <asm/ptdump.h>
#include <asm/kasan.h> #include <asm/kasan.h>
#include <asm/abs_lowcore.h>
#include <asm/nospec-branch.h> #include <asm/nospec-branch.h>
#include <asm/sections.h> #include <asm/sections.h>
#include <asm/maccess.h>
static unsigned long max_addr; static unsigned long max_addr;
...@@ -21,6 +23,8 @@ struct addr_marker { ...@@ -21,6 +23,8 @@ struct addr_marker {
enum address_markers_idx { enum address_markers_idx {
IDENTITY_BEFORE_NR = 0, IDENTITY_BEFORE_NR = 0,
IDENTITY_BEFORE_END_NR, IDENTITY_BEFORE_END_NR,
AMODE31_START_NR,
AMODE31_END_NR,
KERNEL_START_NR, KERNEL_START_NR,
KERNEL_END_NR, KERNEL_END_NR,
#ifdef CONFIG_KFENCE #ifdef CONFIG_KFENCE
...@@ -39,11 +43,17 @@ enum address_markers_idx { ...@@ -39,11 +43,17 @@ enum address_markers_idx {
VMALLOC_END_NR, VMALLOC_END_NR,
MODULES_NR, MODULES_NR,
MODULES_END_NR, MODULES_END_NR,
ABS_LOWCORE_NR,
ABS_LOWCORE_END_NR,
MEMCPY_REAL_NR,
MEMCPY_REAL_END_NR,
}; };
static struct addr_marker address_markers[] = { static struct addr_marker address_markers[] = {
[IDENTITY_BEFORE_NR] = {0, "Identity Mapping Start"}, [IDENTITY_BEFORE_NR] = {0, "Identity Mapping Start"},
[IDENTITY_BEFORE_END_NR] = {(unsigned long)_stext, "Identity Mapping End"}, [IDENTITY_BEFORE_END_NR] = {(unsigned long)_stext, "Identity Mapping End"},
[AMODE31_START_NR] = {0, "Amode31 Area Start"},
[AMODE31_END_NR] = {0, "Amode31 Area End"},
[KERNEL_START_NR] = {(unsigned long)_stext, "Kernel Image Start"}, [KERNEL_START_NR] = {(unsigned long)_stext, "Kernel Image Start"},
[KERNEL_END_NR] = {(unsigned long)_end, "Kernel Image End"}, [KERNEL_END_NR] = {(unsigned long)_end, "Kernel Image End"},
#ifdef CONFIG_KFENCE #ifdef CONFIG_KFENCE
...@@ -62,6 +72,10 @@ static struct addr_marker address_markers[] = { ...@@ -62,6 +72,10 @@ static struct addr_marker address_markers[] = {
[VMALLOC_END_NR] = {0, "vmalloc Area End"}, [VMALLOC_END_NR] = {0, "vmalloc Area End"},
[MODULES_NR] = {0, "Modules Area Start"}, [MODULES_NR] = {0, "Modules Area Start"},
[MODULES_END_NR] = {0, "Modules Area End"}, [MODULES_END_NR] = {0, "Modules Area End"},
[ABS_LOWCORE_NR] = {0, "Lowcore Area Start"},
[ABS_LOWCORE_END_NR] = {0, "Lowcore Area End"},
[MEMCPY_REAL_NR] = {0, "Real Memory Copy Area Start"},
[MEMCPY_REAL_END_NR] = {0, "Real Memory Copy Area End"},
{ -1, NULL } { -1, NULL }
}; };
...@@ -276,8 +290,14 @@ static int pt_dump_init(void) ...@@ -276,8 +290,14 @@ static int pt_dump_init(void)
max_addr = (S390_lowcore.kernel_asce & _REGION_ENTRY_TYPE_MASK) >> 2; max_addr = (S390_lowcore.kernel_asce & _REGION_ENTRY_TYPE_MASK) >> 2;
max_addr = 1UL << (max_addr * 11 + 31); max_addr = 1UL << (max_addr * 11 + 31);
address_markers[IDENTITY_AFTER_END_NR].start_address = ident_map_size; address_markers[IDENTITY_AFTER_END_NR].start_address = ident_map_size;
address_markers[AMODE31_START_NR].start_address = __samode31;
address_markers[AMODE31_END_NR].start_address = __eamode31;
address_markers[MODULES_NR].start_address = MODULES_VADDR; address_markers[MODULES_NR].start_address = MODULES_VADDR;
address_markers[MODULES_END_NR].start_address = MODULES_END; address_markers[MODULES_END_NR].start_address = MODULES_END;
address_markers[ABS_LOWCORE_NR].start_address = __abs_lowcore;
address_markers[ABS_LOWCORE_END_NR].start_address = __abs_lowcore + ABS_LOWCORE_MAP_SIZE;
address_markers[MEMCPY_REAL_NR].start_address = __memcpy_real_area;
address_markers[MEMCPY_REAL_END_NR].start_address = __memcpy_real_area + PAGE_SIZE;
address_markers[VMEMMAP_NR].start_address = (unsigned long) vmemmap; address_markers[VMEMMAP_NR].start_address = (unsigned long) vmemmap;
address_markers[VMEMMAP_END_NR].start_address = (unsigned long)vmemmap + vmemmap_size; address_markers[VMEMMAP_END_NR].start_address = (unsigned long)vmemmap + vmemmap_size;
address_markers[VMALLOC_NR].start_address = VMALLOC_START; address_markers[VMALLOC_NR].start_address = VMALLOC_START;
......
...@@ -268,8 +268,7 @@ static noinline void do_sigbus(struct pt_regs *regs) ...@@ -268,8 +268,7 @@ static noinline void do_sigbus(struct pt_regs *regs)
(void __user *)(regs->int_parm_long & __FAIL_ADDR_MASK)); (void __user *)(regs->int_parm_long & __FAIL_ADDR_MASK));
} }
static noinline void do_fault_error(struct pt_regs *regs, int access, static noinline void do_fault_error(struct pt_regs *regs, vm_fault_t fault)
vm_fault_t fault)
{ {
int si_code; int si_code;
...@@ -516,7 +515,7 @@ void do_protection_exception(struct pt_regs *regs) ...@@ -516,7 +515,7 @@ void do_protection_exception(struct pt_regs *regs)
fault = do_exception(regs, access); fault = do_exception(regs, access);
} }
if (unlikely(fault)) if (unlikely(fault))
do_fault_error(regs, access, fault); do_fault_error(regs, fault);
} }
NOKPROBE_SYMBOL(do_protection_exception); NOKPROBE_SYMBOL(do_protection_exception);
...@@ -528,7 +527,7 @@ void do_dat_exception(struct pt_regs *regs) ...@@ -528,7 +527,7 @@ void do_dat_exception(struct pt_regs *regs)
access = VM_ACCESS_FLAGS; access = VM_ACCESS_FLAGS;
fault = do_exception(regs, access); fault = do_exception(regs, access);
if (unlikely(fault)) if (unlikely(fault))
do_fault_error(regs, access, fault); do_fault_error(regs, fault);
} }
NOKPROBE_SYMBOL(do_dat_exception); NOKPROBE_SYMBOL(do_dat_exception);
...@@ -803,7 +802,7 @@ void do_secure_storage_access(struct pt_regs *regs) ...@@ -803,7 +802,7 @@ void do_secure_storage_access(struct pt_regs *regs)
addr = __gmap_translate(gmap, addr); addr = __gmap_translate(gmap, addr);
mmap_read_unlock(mm); mmap_read_unlock(mm);
if (IS_ERR_VALUE(addr)) { if (IS_ERR_VALUE(addr)) {
do_fault_error(regs, VM_ACCESS_FLAGS, VM_FAULT_BADMAP); do_fault_error(regs, VM_FAULT_BADMAP);
break; break;
} }
fallthrough; fallthrough;
...@@ -813,7 +812,7 @@ void do_secure_storage_access(struct pt_regs *regs) ...@@ -813,7 +812,7 @@ void do_secure_storage_access(struct pt_regs *regs)
vma = find_vma(mm, addr); vma = find_vma(mm, addr);
if (!vma) { if (!vma) {
mmap_read_unlock(mm); mmap_read_unlock(mm);
do_fault_error(regs, VM_READ | VM_WRITE, VM_FAULT_BADMAP); do_fault_error(regs, VM_FAULT_BADMAP);
break; break;
} }
page = follow_page(vma, addr, FOLL_WRITE | FOLL_GET); page = follow_page(vma, addr, FOLL_WRITE | FOLL_GET);
...@@ -836,7 +835,7 @@ void do_secure_storage_access(struct pt_regs *regs) ...@@ -836,7 +835,7 @@ void do_secure_storage_access(struct pt_regs *regs)
BUG(); BUG();
break; break;
default: default:
do_fault_error(regs, VM_READ | VM_WRITE, VM_FAULT_BADMAP); do_fault_error(regs, VM_FAULT_BADMAP);
WARN_ON_ONCE(1); WARN_ON_ONCE(1);
} }
} }
...@@ -848,7 +847,7 @@ void do_non_secure_storage_access(struct pt_regs *regs) ...@@ -848,7 +847,7 @@ void do_non_secure_storage_access(struct pt_regs *regs)
struct gmap *gmap = (struct gmap *)S390_lowcore.gmap; struct gmap *gmap = (struct gmap *)S390_lowcore.gmap;
if (get_fault_type(regs) != GMAP_FAULT) { if (get_fault_type(regs) != GMAP_FAULT) {
do_fault_error(regs, VM_READ | VM_WRITE, VM_FAULT_BADMAP); do_fault_error(regs, VM_FAULT_BADMAP);
WARN_ON_ONCE(1); WARN_ON_ONCE(1);
return; return;
} }
......
...@@ -37,7 +37,7 @@ ...@@ -37,7 +37,7 @@
#include <asm/kfence.h> #include <asm/kfence.h>
#include <asm/ptdump.h> #include <asm/ptdump.h>
#include <asm/dma.h> #include <asm/dma.h>
#include <asm/lowcore.h> #include <asm/abs_lowcore.h>
#include <asm/tlb.h> #include <asm/tlb.h>
#include <asm/tlbflush.h> #include <asm/tlbflush.h>
#include <asm/sections.h> #include <asm/sections.h>
......
...@@ -12,10 +12,17 @@ ...@@ -12,10 +12,17 @@
#include <linux/errno.h> #include <linux/errno.h>
#include <linux/gfp.h> #include <linux/gfp.h>
#include <linux/cpu.h> #include <linux/cpu.h>
#include <linux/uio.h>
#include <asm/asm-extable.h> #include <asm/asm-extable.h>
#include <asm/ctl_reg.h> #include <asm/ctl_reg.h>
#include <asm/io.h> #include <asm/io.h>
#include <asm/abs_lowcore.h>
#include <asm/stacktrace.h> #include <asm/stacktrace.h>
#include <asm/maccess.h>
unsigned long __bootdata_preserved(__memcpy_real_area);
static __ro_after_init pte_t *memcpy_real_ptep;
static DEFINE_MUTEX(memcpy_real_mutex);
static notrace long s390_kernel_write_odd(void *dst, const void *src, size_t size) static notrace long s390_kernel_write_odd(void *dst, const void *src, size_t size)
{ {
...@@ -76,118 +83,72 @@ notrace void *s390_kernel_write(void *dst, const void *src, size_t size) ...@@ -76,118 +83,72 @@ notrace void *s390_kernel_write(void *dst, const void *src, size_t size)
return dst; return dst;
} }
static int __no_sanitize_address __memcpy_real(void *dest, void *src, size_t count) void __init memcpy_real_init(void)
{ {
union register_pair _dst, _src; memcpy_real_ptep = vmem_get_alloc_pte(__memcpy_real_area, true);
int rc = -EFAULT; if (!memcpy_real_ptep)
panic("Couldn't setup memcpy real area");
_dst.even = (unsigned long) dest;
_dst.odd = (unsigned long) count;
_src.even = (unsigned long) src;
_src.odd = (unsigned long) count;
asm volatile (
"0: mvcle %[dst],%[src],0\n"
"1: jo 0b\n"
" lhi %[rc],0\n"
"2:\n"
EX_TABLE(1b,2b)
: [rc] "+&d" (rc), [dst] "+&d" (_dst.pair), [src] "+&d" (_src.pair)
: : "cc", "memory");
return rc;
}
static unsigned long __no_sanitize_address _memcpy_real(unsigned long dest,
unsigned long src,
unsigned long count)
{
int irqs_disabled, rc;
unsigned long flags;
if (!count)
return 0;
flags = arch_local_irq_save();
irqs_disabled = arch_irqs_disabled_flags(flags);
if (!irqs_disabled)
trace_hardirqs_off();
__arch_local_irq_stnsm(0xf8); // disable DAT
rc = __memcpy_real((void *) dest, (void *) src, (size_t) count);
if (flags & PSW_MASK_DAT)
__arch_local_irq_stosm(0x04); // enable DAT
if (!irqs_disabled)
trace_hardirqs_on();
__arch_local_irq_ssm(flags);
return rc;
} }
/* size_t memcpy_real_iter(struct iov_iter *iter, unsigned long src, size_t count)
* Copy memory in real mode (kernel to kernel)
*/
int memcpy_real(void *dest, unsigned long src, size_t count)
{ {
unsigned long _dest = (unsigned long)dest; size_t len, copied, res = 0;
unsigned long _src = (unsigned long)src; unsigned long phys, offset;
unsigned long _count = (unsigned long)count; void *chunk;
int rc; pte_t pte;
if (S390_lowcore.nodat_stack != 0) { while (count) {
preempt_disable(); phys = src & PAGE_MASK;
rc = call_on_stack(3, S390_lowcore.nodat_stack, offset = src & ~PAGE_MASK;
unsigned long, _memcpy_real, chunk = (void *)(__memcpy_real_area + offset);
unsigned long, _dest, len = min(count, PAGE_SIZE - offset);
unsigned long, _src, pte = mk_pte_phys(phys, PAGE_KERNEL_RO);
unsigned long, _count);
preempt_enable(); mutex_lock(&memcpy_real_mutex);
return rc; if (pte_val(pte) != pte_val(*memcpy_real_ptep)) {
__ptep_ipte(__memcpy_real_area, memcpy_real_ptep, 0, 0, IPTE_GLOBAL);
set_pte(memcpy_real_ptep, pte);
}
copied = copy_to_iter(chunk, len, iter);
mutex_unlock(&memcpy_real_mutex);
count -= copied;
src += copied;
res += copied;
if (copied < len)
break;
} }
/* return res;
* This is a really early memcpy_real call, the stacks are
* not set up yet. Just call _memcpy_real on the early boot
* stack
*/
return _memcpy_real(_dest, _src, _count);
} }
/* int memcpy_real(void *dest, unsigned long src, size_t count)
* Copy memory in absolute mode (kernel to kernel)
*/
void memcpy_absolute(void *dest, void *src, size_t count)
{ {
unsigned long cr0, flags, prefix; struct iov_iter iter;
struct kvec kvec;
flags = arch_local_irq_save();
__ctl_store(cr0, 0, 0); kvec.iov_base = dest;
__ctl_clear_bit(0, 28); /* disable lowcore protection */ kvec.iov_len = count;
prefix = store_prefix(); iov_iter_kvec(&iter, WRITE, &kvec, 1, count);
if (prefix) { if (memcpy_real_iter(&iter, src, count) < count)
local_mcck_disable(); return -EFAULT;
set_prefix(0); return 0;
memcpy(dest, src, count);
set_prefix(prefix);
local_mcck_enable();
} else {
memcpy(dest, src, count);
}
__ctl_load(cr0, 0, 0);
arch_local_irq_restore(flags);
} }
/* /*
* Check if physical address is within prefix or zero page * Find CPU that owns swapped prefix page
*/ */
static int is_swapped(phys_addr_t addr) static int get_swapped_owner(phys_addr_t addr)
{ {
phys_addr_t lc; phys_addr_t lc;
int cpu; int cpu;
if (addr < sizeof(struct lowcore))
return 1;
for_each_online_cpu(cpu) { for_each_online_cpu(cpu) {
lc = virt_to_phys(lowcore_ptr[cpu]); lc = virt_to_phys(lowcore_ptr[cpu]);
if (addr > lc + sizeof(struct lowcore) - 1 || addr < lc) if (addr > lc + sizeof(struct lowcore) - 1 || addr < lc)
continue; continue;
return 1; return cpu;
} }
return 0; return -1;
} }
/* /*
...@@ -200,17 +161,35 @@ void *xlate_dev_mem_ptr(phys_addr_t addr) ...@@ -200,17 +161,35 @@ void *xlate_dev_mem_ptr(phys_addr_t addr)
{ {
void *ptr = phys_to_virt(addr); void *ptr = phys_to_virt(addr);
void *bounce = ptr; void *bounce = ptr;
struct lowcore *abs_lc;
unsigned long flags;
unsigned long size; unsigned long size;
int this_cpu, cpu;
cpus_read_lock(); cpus_read_lock();
preempt_disable(); this_cpu = get_cpu();
if (is_swapped(addr)) { if (addr >= sizeof(struct lowcore)) {
size = PAGE_SIZE - (addr & ~PAGE_MASK); cpu = get_swapped_owner(addr);
bounce = (void *) __get_free_page(GFP_ATOMIC); if (cpu < 0)
if (bounce) goto out;
memcpy_absolute(bounce, ptr, size); }
bounce = (void *)__get_free_page(GFP_ATOMIC);
if (!bounce)
goto out;
size = PAGE_SIZE - (addr & ~PAGE_MASK);
if (addr < sizeof(struct lowcore)) {
abs_lc = get_abs_lowcore(&flags);
ptr = (void *)abs_lc + addr;
memcpy(bounce, ptr, size);
put_abs_lowcore(abs_lc, flags);
} else if (cpu == this_cpu) {
ptr = (void *)(addr - virt_to_phys(lowcore_ptr[cpu]));
memcpy(bounce, ptr, size);
} else {
memcpy(bounce, ptr, size);
} }
preempt_enable(); out:
put_cpu();
cpus_read_unlock(); cpus_read_unlock();
return bounce; return bounce;
} }
......
...@@ -240,7 +240,7 @@ static int __ref modify_pmd_table(pud_t *pud, unsigned long addr, ...@@ -240,7 +240,7 @@ static int __ref modify_pmd_table(pud_t *pud, unsigned long addr,
} else if (pmd_none(*pmd)) { } else if (pmd_none(*pmd)) {
if (IS_ALIGNED(addr, PMD_SIZE) && if (IS_ALIGNED(addr, PMD_SIZE) &&
IS_ALIGNED(next, PMD_SIZE) && IS_ALIGNED(next, PMD_SIZE) &&
MACHINE_HAS_EDAT1 && addr && direct && MACHINE_HAS_EDAT1 && direct &&
!debug_pagealloc_enabled()) { !debug_pagealloc_enabled()) {
set_pmd(pmd, __pmd(__pa(addr) | prot)); set_pmd(pmd, __pmd(__pa(addr) | prot));
pages++; pages++;
...@@ -336,7 +336,7 @@ static int modify_pud_table(p4d_t *p4d, unsigned long addr, unsigned long end, ...@@ -336,7 +336,7 @@ static int modify_pud_table(p4d_t *p4d, unsigned long addr, unsigned long end,
} else if (pud_none(*pud)) { } else if (pud_none(*pud)) {
if (IS_ALIGNED(addr, PUD_SIZE) && if (IS_ALIGNED(addr, PUD_SIZE) &&
IS_ALIGNED(next, PUD_SIZE) && IS_ALIGNED(next, PUD_SIZE) &&
MACHINE_HAS_EDAT2 && addr && direct && MACHINE_HAS_EDAT2 && direct &&
!debug_pagealloc_enabled()) { !debug_pagealloc_enabled()) {
set_pud(pud, __pud(__pa(addr) | prot)); set_pud(pud, __pud(__pa(addr) | prot));
pages++; pages++;
...@@ -560,6 +560,103 @@ int vmem_add_mapping(unsigned long start, unsigned long size) ...@@ -560,6 +560,103 @@ int vmem_add_mapping(unsigned long start, unsigned long size)
return ret; return ret;
} }
/*
* Allocate new or return existing page-table entry, but do not map it
* to any physical address. If missing, allocate segment- and region-
* table entries along. Meeting a large segment- or region-table entry
* while traversing is an error, since the function is expected to be
* called against virtual regions reserverd for 4KB mappings only.
*/
pte_t *vmem_get_alloc_pte(unsigned long addr, bool alloc)
{
pte_t *ptep = NULL;
pgd_t *pgd;
p4d_t *p4d;
pud_t *pud;
pmd_t *pmd;
pte_t *pte;
pgd = pgd_offset_k(addr);
if (pgd_none(*pgd)) {
if (!alloc)
goto out;
p4d = vmem_crst_alloc(_REGION2_ENTRY_EMPTY);
if (!p4d)
goto out;
pgd_populate(&init_mm, pgd, p4d);
}
p4d = p4d_offset(pgd, addr);
if (p4d_none(*p4d)) {
if (!alloc)
goto out;
pud = vmem_crst_alloc(_REGION3_ENTRY_EMPTY);
if (!pud)
goto out;
p4d_populate(&init_mm, p4d, pud);
}
pud = pud_offset(p4d, addr);
if (pud_none(*pud)) {
if (!alloc)
goto out;
pmd = vmem_crst_alloc(_SEGMENT_ENTRY_EMPTY);
if (!pmd)
goto out;
pud_populate(&init_mm, pud, pmd);
} else if (WARN_ON_ONCE(pud_large(*pud))) {
goto out;
}
pmd = pmd_offset(pud, addr);
if (pmd_none(*pmd)) {
if (!alloc)
goto out;
pte = vmem_pte_alloc();
if (!pte)
goto out;
pmd_populate(&init_mm, pmd, pte);
} else if (WARN_ON_ONCE(pmd_large(*pmd))) {
goto out;
}
ptep = pte_offset_kernel(pmd, addr);
out:
return ptep;
}
int __vmem_map_4k_page(unsigned long addr, unsigned long phys, pgprot_t prot, bool alloc)
{
pte_t *ptep, pte;
if (!IS_ALIGNED(addr, PAGE_SIZE))
return -EINVAL;
ptep = vmem_get_alloc_pte(addr, alloc);
if (!ptep)
return -ENOMEM;
__ptep_ipte(addr, ptep, 0, 0, IPTE_GLOBAL);
pte = mk_pte_phys(phys, prot);
set_pte(ptep, pte);
return 0;
}
int vmem_map_4k_page(unsigned long addr, unsigned long phys, pgprot_t prot)
{
int rc;
mutex_lock(&vmem_mutex);
rc = __vmem_map_4k_page(addr, phys, prot, true);
mutex_unlock(&vmem_mutex);
return rc;
}
void vmem_unmap_4k_page(unsigned long addr)
{
pte_t *ptep;
mutex_lock(&vmem_mutex);
ptep = virt_to_kpte(addr);
__ptep_ipte(addr, ptep, 0, 0, IPTE_GLOBAL);
pte_clear(&init_mm, addr, ptep);
mutex_unlock(&vmem_mutex);
}
/* /*
* map whole physical memory to virtual memory (identity mapping) * map whole physical memory to virtual memory (identity mapping)
* we reserve enough space in the vmalloc area for vmemmap to hotplug * we reserve enough space in the vmalloc area for vmemmap to hotplug
...@@ -584,6 +681,9 @@ void __init vmem_map_init(void) ...@@ -584,6 +681,9 @@ void __init vmem_map_init(void)
__set_memory(__stext_amode31, (__etext_amode31 - __stext_amode31) >> PAGE_SHIFT, __set_memory(__stext_amode31, (__etext_amode31 - __stext_amode31) >> PAGE_SHIFT,
SET_MEMORY_RO | SET_MEMORY_X); SET_MEMORY_RO | SET_MEMORY_X);
/* lowcore requires 4k mapping for real addresses / prefixing */
set_memory_4k(0, LC_PAGES);
/* lowcore must be executable for LPSWE */ /* lowcore must be executable for LPSWE */
if (!static_key_enabled(&cpu_has_bear)) if (!static_key_enabled(&cpu_has_bear))
set_memory_x(0, 1); set_memory_x(0, 1);
......
...@@ -666,7 +666,7 @@ static int __init dma_alloc_cpu_table_caches(void) ...@@ -666,7 +666,7 @@ static int __init dma_alloc_cpu_table_caches(void)
int __init zpci_dma_init(void) int __init zpci_dma_init(void)
{ {
s390_iommu_aperture = (u64)high_memory; s390_iommu_aperture = (u64)virt_to_phys(high_memory);
if (!s390_iommu_aperture_factor) if (!s390_iommu_aperture_factor)
s390_iommu_aperture = ULONG_MAX; s390_iommu_aperture = ULONG_MAX;
else else
......
...@@ -426,7 +426,7 @@ dasd_add_busid(const char *bus_id, int features) ...@@ -426,7 +426,7 @@ dasd_add_busid(const char *bus_id, int features)
if (!devmap) { if (!devmap) {
/* This bus_id is new. */ /* This bus_id is new. */
new->devindex = dasd_max_devindex++; new->devindex = dasd_max_devindex++;
strlcpy(new->bus_id, bus_id, DASD_BUS_ID_SIZE); strscpy(new->bus_id, bus_id, DASD_BUS_ID_SIZE);
new->features = features; new->features = features;
new->device = NULL; new->device = NULL;
list_add(&new->list, &dasd_hashlists[hash]); list_add(&new->list, &dasd_hashlists[hash]);
......
...@@ -313,7 +313,7 @@ static void dasd_eer_write_standard_trigger(struct dasd_device *device, ...@@ -313,7 +313,7 @@ static void dasd_eer_write_standard_trigger(struct dasd_device *device,
ktime_get_real_ts64(&ts); ktime_get_real_ts64(&ts);
header.tv_sec = ts.tv_sec; header.tv_sec = ts.tv_sec;
header.tv_usec = ts.tv_nsec / NSEC_PER_USEC; header.tv_usec = ts.tv_nsec / NSEC_PER_USEC;
strlcpy(header.busid, dev_name(&device->cdev->dev), strscpy(header.busid, dev_name(&device->cdev->dev),
DASD_EER_BUSID_SIZE); DASD_EER_BUSID_SIZE);
spin_lock_irqsave(&bufferlock, flags); spin_lock_irqsave(&bufferlock, flags);
...@@ -356,7 +356,7 @@ static void dasd_eer_write_snss_trigger(struct dasd_device *device, ...@@ -356,7 +356,7 @@ static void dasd_eer_write_snss_trigger(struct dasd_device *device,
ktime_get_real_ts64(&ts); ktime_get_real_ts64(&ts);
header.tv_sec = ts.tv_sec; header.tv_sec = ts.tv_sec;
header.tv_usec = ts.tv_nsec / NSEC_PER_USEC; header.tv_usec = ts.tv_nsec / NSEC_PER_USEC;
strlcpy(header.busid, dev_name(&device->cdev->dev), strscpy(header.busid, dev_name(&device->cdev->dev),
DASD_EER_BUSID_SIZE); DASD_EER_BUSID_SIZE);
spin_lock_irqsave(&bufferlock, flags); spin_lock_irqsave(&bufferlock, flags);
......
...@@ -614,7 +614,7 @@ dcssblk_add_store(struct device *dev, struct device_attribute *attr, const char ...@@ -614,7 +614,7 @@ dcssblk_add_store(struct device *dev, struct device_attribute *attr, const char
rc = -ENAMETOOLONG; rc = -ENAMETOOLONG;
goto seg_list_del; goto seg_list_del;
} }
strlcpy(local_buf, buf, i + 1); strscpy(local_buf, buf, i + 1);
dev_info->num_of_segments = num_of_segments; dev_info->num_of_segments = num_of_segments;
rc = dcssblk_is_continuous(dev_info); rc = dcssblk_is_continuous(dev_info);
if (rc < 0) if (rc < 0)
......
...@@ -154,7 +154,7 @@ static ssize_t hmcdrv_cache_do(const struct hmcdrv_ftp_cmdspec *ftp, ...@@ -154,7 +154,7 @@ static ssize_t hmcdrv_cache_do(const struct hmcdrv_ftp_cmdspec *ftp,
/* cache some file info (FTP command, file name and file /* cache some file info (FTP command, file name and file
* size) unconditionally * size) unconditionally
*/ */
strlcpy(hmcdrv_cache_file.fname, ftp->fname, strscpy(hmcdrv_cache_file.fname, ftp->fname,
HMCDRV_FTP_FIDENT_MAX); HMCDRV_FTP_FIDENT_MAX);
hmcdrv_cache_file.id = ftp->id; hmcdrv_cache_file.id = ftp->id;
pr_debug("caching cmd %d, file size %zu for '%s'\n", pr_debug("caching cmd %d, file size %zu for '%s'\n",
......
...@@ -54,10 +54,10 @@ struct tape_class_device *register_tape_dev( ...@@ -54,10 +54,10 @@ struct tape_class_device *register_tape_dev(
if (!tcd) if (!tcd)
return ERR_PTR(-ENOMEM); return ERR_PTR(-ENOMEM);
strlcpy(tcd->device_name, device_name, TAPECLASS_NAME_LEN); strscpy(tcd->device_name, device_name, TAPECLASS_NAME_LEN);
for (s = strchr(tcd->device_name, '/'); s; s = strchr(s, '/')) for (s = strchr(tcd->device_name, '/'); s; s = strchr(s, '/'))
*s = '!'; *s = '!';
strlcpy(tcd->mode_name, mode_name, TAPECLASS_NAME_LEN); strscpy(tcd->mode_name, mode_name, TAPECLASS_NAME_LEN);
for (s = strchr(tcd->mode_name, '/'); s; s = strchr(s, '/')) for (s = strchr(tcd->mode_name, '/'); s; s = strchr(s, '/'))
*s = '!'; *s = '!';
......
...@@ -30,6 +30,7 @@ ...@@ -30,6 +30,7 @@
#include <asm/checksum.h> #include <asm/checksum.h>
#include <asm/os_info.h> #include <asm/os_info.h>
#include <asm/switch_to.h> #include <asm/switch_to.h>
#include <asm/maccess.h>
#include "sclp.h" #include "sclp.h"
#define TRACE(x...) debug_sprintf_event(zcore_dbf, 1, x) #define TRACE(x...) debug_sprintf_event(zcore_dbf, 1, x)
......
...@@ -87,7 +87,7 @@ int qdio_allocate_dbf(struct qdio_irq *irq_ptr) ...@@ -87,7 +87,7 @@ int qdio_allocate_dbf(struct qdio_irq *irq_ptr)
debug_unregister(irq_ptr->debug_area); debug_unregister(irq_ptr->debug_area);
return -ENOMEM; return -ENOMEM;
} }
strlcpy(new_entry->dbf_name, text, QDIO_DBF_NAME_LEN); strscpy(new_entry->dbf_name, text, QDIO_DBF_NAME_LEN);
new_entry->dbf_info = irq_ptr->debug_area; new_entry->dbf_info = irq_ptr->debug_area;
mutex_lock(&qdio_dbf_list_mutex); mutex_lock(&qdio_dbf_list_mutex);
list_add(&new_entry->dbf_list, &qdio_dbf_list); list_add(&new_entry->dbf_list, &qdio_dbf_list);
......
...@@ -1566,7 +1566,7 @@ static int ctcm_new_device(struct ccwgroup_device *cgdev) ...@@ -1566,7 +1566,7 @@ static int ctcm_new_device(struct ccwgroup_device *cgdev)
goto out_dev; goto out_dev;
} }
strlcpy(priv->fsm->name, dev->name, sizeof(priv->fsm->name)); strscpy(priv->fsm->name, dev->name, sizeof(priv->fsm->name));
dev_info(&dev->dev, dev_info(&dev->dev,
"setup OK : r/w = %s/%s, protocol : %d\n", "setup OK : r/w = %s/%s, protocol : %d\n",
......
...@@ -28,7 +28,7 @@ init_fsm(char *name, const char **state_names, const char **event_names, int nr_ ...@@ -28,7 +28,7 @@ init_fsm(char *name, const char **state_names, const char **event_names, int nr_
"fsm(%s): init_fsm: Couldn't alloc instance\n", name); "fsm(%s): init_fsm: Couldn't alloc instance\n", name);
return NULL; return NULL;
} }
strlcpy(this->name, name, sizeof(this->name)); strscpy(this->name, name, sizeof(this->name));
init_waitqueue_head(&this->wait_q); init_waitqueue_head(&this->wait_q);
f = kzalloc(sizeof(fsm), order); f = kzalloc(sizeof(fsm), order);
......
...@@ -188,9 +188,9 @@ static void qeth_get_drvinfo(struct net_device *dev, ...@@ -188,9 +188,9 @@ static void qeth_get_drvinfo(struct net_device *dev,
{ {
struct qeth_card *card = dev->ml_priv; struct qeth_card *card = dev->ml_priv;
strlcpy(info->driver, IS_LAYER2(card) ? "qeth_l2" : "qeth_l3", strscpy(info->driver, IS_LAYER2(card) ? "qeth_l2" : "qeth_l3",
sizeof(info->driver)); sizeof(info->driver));
strlcpy(info->fw_version, card->info.mcl_level, strscpy(info->fw_version, card->info.mcl_level,
sizeof(info->fw_version)); sizeof(info->fw_version));
snprintf(info->bus_info, sizeof(info->bus_info), "%s/%s/%s", snprintf(info->bus_info, sizeof(info->bus_info), "%s/%s/%s",
CARD_RDEV_ID(card), CARD_WDEV_ID(card), CARD_DDEV_ID(card)); CARD_RDEV_ID(card), CARD_WDEV_ID(card), CARD_DDEV_ID(card));
......
...@@ -103,7 +103,7 @@ static void __init zfcp_init_device_setup(char *devstr) ...@@ -103,7 +103,7 @@ static void __init zfcp_init_device_setup(char *devstr)
token = strsep(&str, ","); token = strsep(&str, ",");
if (!token || strlen(token) >= ZFCP_BUS_ID_SIZE) if (!token || strlen(token) >= ZFCP_BUS_ID_SIZE)
goto err_out; goto err_out;
strlcpy(busid, token, ZFCP_BUS_ID_SIZE); strscpy(busid, token, ZFCP_BUS_ID_SIZE);
token = strsep(&str, ","); token = strsep(&str, ",");
if (!token || kstrtoull(token, 0, (unsigned long long *) &wwpn)) if (!token || kstrtoull(token, 0, (unsigned long long *) &wwpn))
......
...@@ -885,7 +885,7 @@ static int zfcp_fc_gspn(struct zfcp_adapter *adapter, ...@@ -885,7 +885,7 @@ static int zfcp_fc_gspn(struct zfcp_adapter *adapter,
dev_name(&adapter->ccw_device->dev), dev_name(&adapter->ccw_device->dev),
init_utsname()->nodename); init_utsname()->nodename);
else else
strlcpy(fc_host_symbolic_name(adapter->scsi_host), strscpy(fc_host_symbolic_name(adapter->scsi_host),
gspn_rsp->gspn.fp_name, FC_SYMBOLIC_NAME_SIZE); gspn_rsp->gspn.fp_name, FC_SYMBOLIC_NAME_SIZE);
return 0; return 0;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment