提交 0161b6e0 编写于 作者: L Linus Torvalds

Merge branch 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/s390/linux

Pull more s390 updates from Martin Schwidefsky:
 "There is one larger patch for the AP bus code to make it work with the
  longer reset periods of the latest crypto cards.

  A new default configuration, a naming cleanup for SMP and a few fixes"

* 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/s390/linux:
  s390/kdump: fix compile for !SMP
  s390/kdump: fix nosmt kernel parameter
  s390: new default configuration
  s390/smp: cleanup core vs. cpu in the SCLP interface
  s390/smp: fix sigp cpu detection loop
  s390/zcrypt: Fixed reset and interrupt handling of AP queues
  s390/kdump: fix REGSET_VX_LOW vector register ELF notes
  s390/bpf: Fix backward jumps
...@@ -17,13 +17,15 @@ CONFIG_CGROUP_FREEZER=y ...@@ -17,13 +17,15 @@ CONFIG_CGROUP_FREEZER=y
CONFIG_CGROUP_DEVICE=y CONFIG_CGROUP_DEVICE=y
CONFIG_CPUSETS=y CONFIG_CPUSETS=y
CONFIG_CGROUP_CPUACCT=y CONFIG_CGROUP_CPUACCT=y
CONFIG_RESOURCE_COUNTERS=y
CONFIG_CGROUP_PERF=y CONFIG_CGROUP_PERF=y
CONFIG_CFS_BANDWIDTH=y CONFIG_CFS_BANDWIDTH=y
CONFIG_RT_GROUP_SCHED=y CONFIG_RT_GROUP_SCHED=y
CONFIG_BLK_CGROUP=y CONFIG_BLK_CGROUP=y
CONFIG_NAMESPACES=y
CONFIG_SCHED_AUTOGROUP=y CONFIG_SCHED_AUTOGROUP=y
CONFIG_BLK_DEV_INITRD=y CONFIG_BLK_DEV_INITRD=y
CONFIG_EXPERT=y
CONFIG_BPF_SYSCALL=y
# CONFIG_COMPAT_BRK is not set # CONFIG_COMPAT_BRK is not set
CONFIG_PROFILING=y CONFIG_PROFILING=y
CONFIG_OPROFILE=m CONFIG_OPROFILE=m
...@@ -44,6 +46,7 @@ CONFIG_SOLARIS_X86_PARTITION=y ...@@ -44,6 +46,7 @@ CONFIG_SOLARIS_X86_PARTITION=y
CONFIG_UNIXWARE_DISKLABEL=y CONFIG_UNIXWARE_DISKLABEL=y
CONFIG_CFQ_GROUP_IOSCHED=y CONFIG_CFQ_GROUP_IOSCHED=y
CONFIG_DEFAULT_DEADLINE=y CONFIG_DEFAULT_DEADLINE=y
CONFIG_LIVEPATCH=y
CONFIG_MARCH_Z196=y CONFIG_MARCH_Z196=y
CONFIG_TUNE_ZEC12=y CONFIG_TUNE_ZEC12=y
CONFIG_NR_CPUS=256 CONFIG_NR_CPUS=256
...@@ -242,9 +245,9 @@ CONFIG_NF_CONNTRACK_IPV4=m ...@@ -242,9 +245,9 @@ CONFIG_NF_CONNTRACK_IPV4=m
# CONFIG_NF_CONNTRACK_PROC_COMPAT is not set # CONFIG_NF_CONNTRACK_PROC_COMPAT is not set
CONFIG_NF_TABLES_IPV4=m CONFIG_NF_TABLES_IPV4=m
CONFIG_NFT_CHAIN_ROUTE_IPV4=m CONFIG_NFT_CHAIN_ROUTE_IPV4=m
CONFIG_NFT_CHAIN_NAT_IPV4=m
CONFIG_NF_TABLES_ARP=m CONFIG_NF_TABLES_ARP=m
CONFIG_NF_NAT_IPV4=m CONFIG_NF_NAT_IPV4=m
CONFIG_NFT_CHAIN_NAT_IPV4=m
CONFIG_IP_NF_IPTABLES=m CONFIG_IP_NF_IPTABLES=m
CONFIG_IP_NF_MATCH_AH=m CONFIG_IP_NF_MATCH_AH=m
CONFIG_IP_NF_MATCH_ECN=m CONFIG_IP_NF_MATCH_ECN=m
...@@ -264,8 +267,8 @@ CONFIG_IP_NF_ARP_MANGLE=m ...@@ -264,8 +267,8 @@ CONFIG_IP_NF_ARP_MANGLE=m
CONFIG_NF_CONNTRACK_IPV6=m CONFIG_NF_CONNTRACK_IPV6=m
CONFIG_NF_TABLES_IPV6=m CONFIG_NF_TABLES_IPV6=m
CONFIG_NFT_CHAIN_ROUTE_IPV6=m CONFIG_NFT_CHAIN_ROUTE_IPV6=m
CONFIG_NFT_CHAIN_NAT_IPV6=m
CONFIG_NF_NAT_IPV6=m CONFIG_NF_NAT_IPV6=m
CONFIG_NFT_CHAIN_NAT_IPV6=m
CONFIG_IP6_NF_IPTABLES=m CONFIG_IP6_NF_IPTABLES=m
CONFIG_IP6_NF_MATCH_AH=m CONFIG_IP6_NF_MATCH_AH=m
CONFIG_IP6_NF_MATCH_EUI64=m CONFIG_IP6_NF_MATCH_EUI64=m
...@@ -353,7 +356,6 @@ CONFIG_BLK_DEV_NBD=m ...@@ -353,7 +356,6 @@ CONFIG_BLK_DEV_NBD=m
CONFIG_BLK_DEV_OSD=m CONFIG_BLK_DEV_OSD=m
CONFIG_BLK_DEV_RAM=y CONFIG_BLK_DEV_RAM=y
CONFIG_BLK_DEV_RAM_SIZE=32768 CONFIG_BLK_DEV_RAM_SIZE=32768
CONFIG_BLK_DEV_XIP=y
CONFIG_CDROM_PKTCDVD=m CONFIG_CDROM_PKTCDVD=m
CONFIG_ATA_OVER_ETH=m CONFIG_ATA_OVER_ETH=m
CONFIG_VIRTIO_BLK=y CONFIG_VIRTIO_BLK=y
...@@ -458,7 +460,6 @@ CONFIG_EXT2_FS=y ...@@ -458,7 +460,6 @@ CONFIG_EXT2_FS=y
CONFIG_EXT2_FS_XATTR=y CONFIG_EXT2_FS_XATTR=y
CONFIG_EXT2_FS_POSIX_ACL=y CONFIG_EXT2_FS_POSIX_ACL=y
CONFIG_EXT2_FS_SECURITY=y CONFIG_EXT2_FS_SECURITY=y
CONFIG_EXT2_FS_XIP=y
CONFIG_EXT3_FS=y CONFIG_EXT3_FS=y
# CONFIG_EXT3_DEFAULTS_TO_ORDERED is not set # CONFIG_EXT3_DEFAULTS_TO_ORDERED is not set
CONFIG_EXT3_FS_POSIX_ACL=y CONFIG_EXT3_FS_POSIX_ACL=y
...@@ -544,7 +545,6 @@ CONFIG_FRAME_WARN=1024 ...@@ -544,7 +545,6 @@ CONFIG_FRAME_WARN=1024
CONFIG_READABLE_ASM=y CONFIG_READABLE_ASM=y
CONFIG_UNUSED_SYMBOLS=y CONFIG_UNUSED_SYMBOLS=y
CONFIG_MAGIC_SYSRQ=y CONFIG_MAGIC_SYSRQ=y
CONFIG_DEBUG_KERNEL=y
CONFIG_DEBUG_PAGEALLOC=y CONFIG_DEBUG_PAGEALLOC=y
CONFIG_DEBUG_OBJECTS=y CONFIG_DEBUG_OBJECTS=y
CONFIG_DEBUG_OBJECTS_SELFTEST=y CONFIG_DEBUG_OBJECTS_SELFTEST=y
...@@ -558,6 +558,7 @@ CONFIG_SLUB_STATS=y ...@@ -558,6 +558,7 @@ CONFIG_SLUB_STATS=y
CONFIG_DEBUG_STACK_USAGE=y CONFIG_DEBUG_STACK_USAGE=y
CONFIG_DEBUG_VM=y CONFIG_DEBUG_VM=y
CONFIG_DEBUG_VM_RB=y CONFIG_DEBUG_VM_RB=y
CONFIG_DEBUG_MEMORY_INIT=y
CONFIG_MEMORY_NOTIFIER_ERROR_INJECT=m CONFIG_MEMORY_NOTIFIER_ERROR_INJECT=m
CONFIG_DEBUG_PER_CPU_MAPS=y CONFIG_DEBUG_PER_CPU_MAPS=y
CONFIG_DEBUG_SHIRQ=y CONFIG_DEBUG_SHIRQ=y
...@@ -575,7 +576,6 @@ CONFIG_DEBUG_LIST=y ...@@ -575,7 +576,6 @@ CONFIG_DEBUG_LIST=y
CONFIG_DEBUG_SG=y CONFIG_DEBUG_SG=y
CONFIG_DEBUG_NOTIFIERS=y CONFIG_DEBUG_NOTIFIERS=y
CONFIG_DEBUG_CREDENTIALS=y CONFIG_DEBUG_CREDENTIALS=y
CONFIG_PROVE_RCU=y
CONFIG_RCU_TORTURE_TEST=m CONFIG_RCU_TORTURE_TEST=m
CONFIG_RCU_CPU_STALL_TIMEOUT=300 CONFIG_RCU_CPU_STALL_TIMEOUT=300
CONFIG_NOTIFIER_ERROR_INJECTION=m CONFIG_NOTIFIER_ERROR_INJECTION=m
...@@ -611,7 +611,6 @@ CONFIG_TEST_BPF=m ...@@ -611,7 +611,6 @@ CONFIG_TEST_BPF=m
# CONFIG_STRICT_DEVMEM is not set # CONFIG_STRICT_DEVMEM is not set
CONFIG_S390_PTDUMP=y CONFIG_S390_PTDUMP=y
CONFIG_ENCRYPTED_KEYS=m CONFIG_ENCRYPTED_KEYS=m
CONFIG_KEYS_DEBUG_PROC_KEYS=y
CONFIG_SECURITY=y CONFIG_SECURITY=y
CONFIG_SECURITY_NETWORK=y CONFIG_SECURITY_NETWORK=y
CONFIG_SECURITY_SELINUX=y CONFIG_SECURITY_SELINUX=y
......
...@@ -17,11 +17,13 @@ CONFIG_CGROUP_FREEZER=y ...@@ -17,11 +17,13 @@ CONFIG_CGROUP_FREEZER=y
CONFIG_CGROUP_DEVICE=y CONFIG_CGROUP_DEVICE=y
CONFIG_CPUSETS=y CONFIG_CPUSETS=y
CONFIG_CGROUP_CPUACCT=y CONFIG_CGROUP_CPUACCT=y
CONFIG_RESOURCE_COUNTERS=y
CONFIG_CGROUP_PERF=y CONFIG_CGROUP_PERF=y
CONFIG_BLK_CGROUP=y CONFIG_BLK_CGROUP=y
CONFIG_NAMESPACES=y
CONFIG_SCHED_AUTOGROUP=y CONFIG_SCHED_AUTOGROUP=y
CONFIG_BLK_DEV_INITRD=y CONFIG_BLK_DEV_INITRD=y
CONFIG_EXPERT=y
CONFIG_BPF_SYSCALL=y
# CONFIG_COMPAT_BRK is not set # CONFIG_COMPAT_BRK is not set
CONFIG_PROFILING=y CONFIG_PROFILING=y
CONFIG_OPROFILE=m CONFIG_OPROFILE=m
...@@ -240,9 +242,9 @@ CONFIG_NF_CONNTRACK_IPV4=m ...@@ -240,9 +242,9 @@ CONFIG_NF_CONNTRACK_IPV4=m
# CONFIG_NF_CONNTRACK_PROC_COMPAT is not set # CONFIG_NF_CONNTRACK_PROC_COMPAT is not set
CONFIG_NF_TABLES_IPV4=m CONFIG_NF_TABLES_IPV4=m
CONFIG_NFT_CHAIN_ROUTE_IPV4=m CONFIG_NFT_CHAIN_ROUTE_IPV4=m
CONFIG_NFT_CHAIN_NAT_IPV4=m
CONFIG_NF_TABLES_ARP=m CONFIG_NF_TABLES_ARP=m
CONFIG_NF_NAT_IPV4=m CONFIG_NF_NAT_IPV4=m
CONFIG_NFT_CHAIN_NAT_IPV4=m
CONFIG_IP_NF_IPTABLES=m CONFIG_IP_NF_IPTABLES=m
CONFIG_IP_NF_MATCH_AH=m CONFIG_IP_NF_MATCH_AH=m
CONFIG_IP_NF_MATCH_ECN=m CONFIG_IP_NF_MATCH_ECN=m
...@@ -262,8 +264,8 @@ CONFIG_IP_NF_ARP_MANGLE=m ...@@ -262,8 +264,8 @@ CONFIG_IP_NF_ARP_MANGLE=m
CONFIG_NF_CONNTRACK_IPV6=m CONFIG_NF_CONNTRACK_IPV6=m
CONFIG_NF_TABLES_IPV6=m CONFIG_NF_TABLES_IPV6=m
CONFIG_NFT_CHAIN_ROUTE_IPV6=m CONFIG_NFT_CHAIN_ROUTE_IPV6=m
CONFIG_NFT_CHAIN_NAT_IPV6=m
CONFIG_NF_NAT_IPV6=m CONFIG_NF_NAT_IPV6=m
CONFIG_NFT_CHAIN_NAT_IPV6=m
CONFIG_IP6_NF_IPTABLES=m CONFIG_IP6_NF_IPTABLES=m
CONFIG_IP6_NF_MATCH_AH=m CONFIG_IP6_NF_MATCH_AH=m
CONFIG_IP6_NF_MATCH_EUI64=m CONFIG_IP6_NF_MATCH_EUI64=m
...@@ -350,7 +352,6 @@ CONFIG_BLK_DEV_NBD=m ...@@ -350,7 +352,6 @@ CONFIG_BLK_DEV_NBD=m
CONFIG_BLK_DEV_OSD=m CONFIG_BLK_DEV_OSD=m
CONFIG_BLK_DEV_RAM=y CONFIG_BLK_DEV_RAM=y
CONFIG_BLK_DEV_RAM_SIZE=32768 CONFIG_BLK_DEV_RAM_SIZE=32768
CONFIG_BLK_DEV_XIP=y
CONFIG_CDROM_PKTCDVD=m CONFIG_CDROM_PKTCDVD=m
CONFIG_ATA_OVER_ETH=m CONFIG_ATA_OVER_ETH=m
CONFIG_VIRTIO_BLK=y CONFIG_VIRTIO_BLK=y
...@@ -455,7 +456,6 @@ CONFIG_EXT2_FS=y ...@@ -455,7 +456,6 @@ CONFIG_EXT2_FS=y
CONFIG_EXT2_FS_XATTR=y CONFIG_EXT2_FS_XATTR=y
CONFIG_EXT2_FS_POSIX_ACL=y CONFIG_EXT2_FS_POSIX_ACL=y
CONFIG_EXT2_FS_SECURITY=y CONFIG_EXT2_FS_SECURITY=y
CONFIG_EXT2_FS_XIP=y
CONFIG_EXT3_FS=y CONFIG_EXT3_FS=y
# CONFIG_EXT3_DEFAULTS_TO_ORDERED is not set # CONFIG_EXT3_DEFAULTS_TO_ORDERED is not set
CONFIG_EXT3_FS_POSIX_ACL=y CONFIG_EXT3_FS_POSIX_ACL=y
...@@ -538,7 +538,7 @@ CONFIG_DEBUG_INFO=y ...@@ -538,7 +538,7 @@ CONFIG_DEBUG_INFO=y
CONFIG_FRAME_WARN=1024 CONFIG_FRAME_WARN=1024
CONFIG_UNUSED_SYMBOLS=y CONFIG_UNUSED_SYMBOLS=y
CONFIG_MAGIC_SYSRQ=y CONFIG_MAGIC_SYSRQ=y
CONFIG_DEBUG_KERNEL=y CONFIG_DEBUG_MEMORY_INIT=y
CONFIG_MEMORY_NOTIFIER_ERROR_INJECT=m CONFIG_MEMORY_NOTIFIER_ERROR_INJECT=m
CONFIG_PANIC_ON_OOPS=y CONFIG_PANIC_ON_OOPS=y
CONFIG_TIMER_STATS=y CONFIG_TIMER_STATS=y
...@@ -558,7 +558,6 @@ CONFIG_ATOMIC64_SELFTEST=y ...@@ -558,7 +558,6 @@ CONFIG_ATOMIC64_SELFTEST=y
# CONFIG_STRICT_DEVMEM is not set # CONFIG_STRICT_DEVMEM is not set
CONFIG_S390_PTDUMP=y CONFIG_S390_PTDUMP=y
CONFIG_ENCRYPTED_KEYS=m CONFIG_ENCRYPTED_KEYS=m
CONFIG_KEYS_DEBUG_PROC_KEYS=y
CONFIG_SECURITY=y CONFIG_SECURITY=y
CONFIG_SECURITY_NETWORK=y CONFIG_SECURITY_NETWORK=y
CONFIG_SECURITY_SELINUX=y CONFIG_SECURITY_SELINUX=y
......
...@@ -17,11 +17,13 @@ CONFIG_CGROUP_FREEZER=y ...@@ -17,11 +17,13 @@ CONFIG_CGROUP_FREEZER=y
CONFIG_CGROUP_DEVICE=y CONFIG_CGROUP_DEVICE=y
CONFIG_CPUSETS=y CONFIG_CPUSETS=y
CONFIG_CGROUP_CPUACCT=y CONFIG_CGROUP_CPUACCT=y
CONFIG_RESOURCE_COUNTERS=y
CONFIG_CGROUP_PERF=y CONFIG_CGROUP_PERF=y
CONFIG_BLK_CGROUP=y CONFIG_BLK_CGROUP=y
CONFIG_NAMESPACES=y
CONFIG_SCHED_AUTOGROUP=y CONFIG_SCHED_AUTOGROUP=y
CONFIG_BLK_DEV_INITRD=y CONFIG_BLK_DEV_INITRD=y
CONFIG_EXPERT=y
CONFIG_BPF_SYSCALL=y
# CONFIG_COMPAT_BRK is not set # CONFIG_COMPAT_BRK is not set
CONFIG_PROFILING=y CONFIG_PROFILING=y
CONFIG_OPROFILE=m CONFIG_OPROFILE=m
...@@ -42,9 +44,10 @@ CONFIG_SOLARIS_X86_PARTITION=y ...@@ -42,9 +44,10 @@ CONFIG_SOLARIS_X86_PARTITION=y
CONFIG_UNIXWARE_DISKLABEL=y CONFIG_UNIXWARE_DISKLABEL=y
CONFIG_CFQ_GROUP_IOSCHED=y CONFIG_CFQ_GROUP_IOSCHED=y
CONFIG_DEFAULT_DEADLINE=y CONFIG_DEFAULT_DEADLINE=y
CONFIG_LIVEPATCH=y
CONFIG_MARCH_Z196=y CONFIG_MARCH_Z196=y
CONFIG_TUNE_ZEC12=y CONFIG_TUNE_ZEC12=y
CONFIG_NR_CPUS=256 CONFIG_NR_CPUS=512
CONFIG_HZ_100=y CONFIG_HZ_100=y
CONFIG_MEMORY_HOTPLUG=y CONFIG_MEMORY_HOTPLUG=y
CONFIG_MEMORY_HOTREMOVE=y CONFIG_MEMORY_HOTREMOVE=y
...@@ -238,9 +241,9 @@ CONFIG_NF_CONNTRACK_IPV4=m ...@@ -238,9 +241,9 @@ CONFIG_NF_CONNTRACK_IPV4=m
# CONFIG_NF_CONNTRACK_PROC_COMPAT is not set # CONFIG_NF_CONNTRACK_PROC_COMPAT is not set
CONFIG_NF_TABLES_IPV4=m CONFIG_NF_TABLES_IPV4=m
CONFIG_NFT_CHAIN_ROUTE_IPV4=m CONFIG_NFT_CHAIN_ROUTE_IPV4=m
CONFIG_NFT_CHAIN_NAT_IPV4=m
CONFIG_NF_TABLES_ARP=m CONFIG_NF_TABLES_ARP=m
CONFIG_NF_NAT_IPV4=m CONFIG_NF_NAT_IPV4=m
CONFIG_NFT_CHAIN_NAT_IPV4=m
CONFIG_IP_NF_IPTABLES=m CONFIG_IP_NF_IPTABLES=m
CONFIG_IP_NF_MATCH_AH=m CONFIG_IP_NF_MATCH_AH=m
CONFIG_IP_NF_MATCH_ECN=m CONFIG_IP_NF_MATCH_ECN=m
...@@ -260,8 +263,8 @@ CONFIG_IP_NF_ARP_MANGLE=m ...@@ -260,8 +263,8 @@ CONFIG_IP_NF_ARP_MANGLE=m
CONFIG_NF_CONNTRACK_IPV6=m CONFIG_NF_CONNTRACK_IPV6=m
CONFIG_NF_TABLES_IPV6=m CONFIG_NF_TABLES_IPV6=m
CONFIG_NFT_CHAIN_ROUTE_IPV6=m CONFIG_NFT_CHAIN_ROUTE_IPV6=m
CONFIG_NFT_CHAIN_NAT_IPV6=m
CONFIG_NF_NAT_IPV6=m CONFIG_NF_NAT_IPV6=m
CONFIG_NFT_CHAIN_NAT_IPV6=m
CONFIG_IP6_NF_IPTABLES=m CONFIG_IP6_NF_IPTABLES=m
CONFIG_IP6_NF_MATCH_AH=m CONFIG_IP6_NF_MATCH_AH=m
CONFIG_IP6_NF_MATCH_EUI64=m CONFIG_IP6_NF_MATCH_EUI64=m
...@@ -348,7 +351,6 @@ CONFIG_BLK_DEV_NBD=m ...@@ -348,7 +351,6 @@ CONFIG_BLK_DEV_NBD=m
CONFIG_BLK_DEV_OSD=m CONFIG_BLK_DEV_OSD=m
CONFIG_BLK_DEV_RAM=y CONFIG_BLK_DEV_RAM=y
CONFIG_BLK_DEV_RAM_SIZE=32768 CONFIG_BLK_DEV_RAM_SIZE=32768
CONFIG_BLK_DEV_XIP=y
CONFIG_CDROM_PKTCDVD=m CONFIG_CDROM_PKTCDVD=m
CONFIG_ATA_OVER_ETH=m CONFIG_ATA_OVER_ETH=m
CONFIG_VIRTIO_BLK=y CONFIG_VIRTIO_BLK=y
...@@ -453,7 +455,6 @@ CONFIG_EXT2_FS=y ...@@ -453,7 +455,6 @@ CONFIG_EXT2_FS=y
CONFIG_EXT2_FS_XATTR=y CONFIG_EXT2_FS_XATTR=y
CONFIG_EXT2_FS_POSIX_ACL=y CONFIG_EXT2_FS_POSIX_ACL=y
CONFIG_EXT2_FS_SECURITY=y CONFIG_EXT2_FS_SECURITY=y
CONFIG_EXT2_FS_XIP=y
CONFIG_EXT3_FS=y CONFIG_EXT3_FS=y
# CONFIG_EXT3_DEFAULTS_TO_ORDERED is not set # CONFIG_EXT3_DEFAULTS_TO_ORDERED is not set
CONFIG_EXT3_FS_POSIX_ACL=y CONFIG_EXT3_FS_POSIX_ACL=y
...@@ -536,7 +537,7 @@ CONFIG_DEBUG_INFO=y ...@@ -536,7 +537,7 @@ CONFIG_DEBUG_INFO=y
CONFIG_FRAME_WARN=1024 CONFIG_FRAME_WARN=1024
CONFIG_UNUSED_SYMBOLS=y CONFIG_UNUSED_SYMBOLS=y
CONFIG_MAGIC_SYSRQ=y CONFIG_MAGIC_SYSRQ=y
CONFIG_DEBUG_KERNEL=y CONFIG_DEBUG_MEMORY_INIT=y
CONFIG_PANIC_ON_OOPS=y CONFIG_PANIC_ON_OOPS=y
CONFIG_TIMER_STATS=y CONFIG_TIMER_STATS=y
CONFIG_RCU_TORTURE_TEST=m CONFIG_RCU_TORTURE_TEST=m
...@@ -553,7 +554,6 @@ CONFIG_ATOMIC64_SELFTEST=y ...@@ -553,7 +554,6 @@ CONFIG_ATOMIC64_SELFTEST=y
# CONFIG_STRICT_DEVMEM is not set # CONFIG_STRICT_DEVMEM is not set
CONFIG_S390_PTDUMP=y CONFIG_S390_PTDUMP=y
CONFIG_ENCRYPTED_KEYS=m CONFIG_ENCRYPTED_KEYS=m
CONFIG_KEYS_DEBUG_PROC_KEYS=y
CONFIG_SECURITY=y CONFIG_SECURITY=y
CONFIG_SECURITY_NETWORK=y CONFIG_SECURITY_NETWORK=y
CONFIG_SECURITY_SELINUX=y CONFIG_SECURITY_SELINUX=y
......
...@@ -8,7 +8,6 @@ CONFIG_TASKSTATS=y ...@@ -8,7 +8,6 @@ CONFIG_TASKSTATS=y
CONFIG_TASK_DELAY_ACCT=y CONFIG_TASK_DELAY_ACCT=y
CONFIG_TASK_XACCT=y CONFIG_TASK_XACCT=y
CONFIG_TASK_IO_ACCOUNTING=y CONFIG_TASK_IO_ACCOUNTING=y
CONFIG_RCU_FAST_NO_HZ=y
CONFIG_IKCONFIG=y CONFIG_IKCONFIG=y
CONFIG_IKCONFIG_PROC=y CONFIG_IKCONFIG_PROC=y
CONFIG_CGROUPS=y CONFIG_CGROUPS=y
...@@ -31,9 +30,11 @@ CONFIG_JUMP_LABEL=y ...@@ -31,9 +30,11 @@ CONFIG_JUMP_LABEL=y
CONFIG_MODULES=y CONFIG_MODULES=y
CONFIG_MODULE_UNLOAD=y CONFIG_MODULE_UNLOAD=y
CONFIG_MODVERSIONS=y CONFIG_MODVERSIONS=y
CONFIG_BLK_DEV_INTEGRITY=y
CONFIG_PARTITION_ADVANCED=y CONFIG_PARTITION_ADVANCED=y
CONFIG_IBM_PARTITION=y CONFIG_IBM_PARTITION=y
CONFIG_DEFAULT_DEADLINE=y CONFIG_DEFAULT_DEADLINE=y
CONFIG_LIVEPATCH=y
CONFIG_MARCH_Z196=y CONFIG_MARCH_Z196=y
CONFIG_NR_CPUS=256 CONFIG_NR_CPUS=256
CONFIG_HZ_100=y CONFIG_HZ_100=y
...@@ -41,7 +42,6 @@ CONFIG_MEMORY_HOTPLUG=y ...@@ -41,7 +42,6 @@ CONFIG_MEMORY_HOTPLUG=y
CONFIG_MEMORY_HOTREMOVE=y CONFIG_MEMORY_HOTREMOVE=y
CONFIG_KSM=y CONFIG_KSM=y
CONFIG_TRANSPARENT_HUGEPAGE=y CONFIG_TRANSPARENT_HUGEPAGE=y
CONFIG_CMA=y
CONFIG_CRASH_DUMP=y CONFIG_CRASH_DUMP=y
CONFIG_BINFMT_MISC=m CONFIG_BINFMT_MISC=m
CONFIG_HIBERNATION=y CONFIG_HIBERNATION=y
...@@ -125,6 +125,7 @@ CONFIG_DEBUG_FORCE_WEAK_PER_CPU=y ...@@ -125,6 +125,7 @@ CONFIG_DEBUG_FORCE_WEAK_PER_CPU=y
CONFIG_MAGIC_SYSRQ=y CONFIG_MAGIC_SYSRQ=y
CONFIG_DEBUG_PAGEALLOC=y CONFIG_DEBUG_PAGEALLOC=y
CONFIG_DETECT_HUNG_TASK=y CONFIG_DETECT_HUNG_TASK=y
CONFIG_PANIC_ON_OOPS=y
CONFIG_TIMER_STATS=y CONFIG_TIMER_STATS=y
CONFIG_DEBUG_RT_MUTEXES=y CONFIG_DEBUG_RT_MUTEXES=y
CONFIG_PROVE_LOCKING=y CONFIG_PROVE_LOCKING=y
...@@ -135,12 +136,16 @@ CONFIG_DEBUG_LIST=y ...@@ -135,12 +136,16 @@ CONFIG_DEBUG_LIST=y
CONFIG_DEBUG_PI_LIST=y CONFIG_DEBUG_PI_LIST=y
CONFIG_DEBUG_SG=y CONFIG_DEBUG_SG=y
CONFIG_DEBUG_NOTIFIERS=y CONFIG_DEBUG_NOTIFIERS=y
CONFIG_PROVE_RCU=y
CONFIG_RCU_CPU_STALL_TIMEOUT=60 CONFIG_RCU_CPU_STALL_TIMEOUT=60
# CONFIG_RCU_CPU_STALL_INFO is not set
CONFIG_RCU_TRACE=y CONFIG_RCU_TRACE=y
CONFIG_LATENCYTOP=y CONFIG_LATENCYTOP=y
CONFIG_DEBUG_STRICT_USER_COPY_CHECKS=y CONFIG_DEBUG_STRICT_USER_COPY_CHECKS=y
CONFIG_TRACER_SNAPSHOT=y
CONFIG_TRACER_SNAPSHOT_PER_CPU_SWAP=y
CONFIG_STACK_TRACER=y
CONFIG_BLK_DEV_IO_TRACE=y CONFIG_BLK_DEV_IO_TRACE=y
CONFIG_UPROBE_EVENT=y
CONFIG_KPROBES_SANITY_TEST=y CONFIG_KPROBES_SANITY_TEST=y
# CONFIG_STRICT_DEVMEM is not set # CONFIG_STRICT_DEVMEM is not set
CONFIG_S390_PTDUMP=y CONFIG_S390_PTDUMP=y
...@@ -187,6 +192,7 @@ CONFIG_CRYPTO_ZLIB=m ...@@ -187,6 +192,7 @@ CONFIG_CRYPTO_ZLIB=m
CONFIG_CRYPTO_LZO=m CONFIG_CRYPTO_LZO=m
CONFIG_CRYPTO_LZ4=m CONFIG_CRYPTO_LZ4=m
CONFIG_CRYPTO_LZ4HC=m CONFIG_CRYPTO_LZ4HC=m
CONFIG_CRYPTO_ANSI_CPRNG=m
CONFIG_ZCRYPT=m CONFIG_ZCRYPT=m
CONFIG_CRYPTO_SHA1_S390=m CONFIG_CRYPTO_SHA1_S390=m
CONFIG_CRYPTO_SHA256_S390=m CONFIG_CRYPTO_SHA256_S390=m
......
...@@ -8,8 +8,6 @@ ...@@ -8,8 +8,6 @@
#ifndef _ASM_S390_CPU_H #ifndef _ASM_S390_CPU_H
#define _ASM_S390_CPU_H #define _ASM_S390_CPU_H
#define MAX_CPU_ADDRESS 255
#ifndef __ASSEMBLY__ #ifndef __ASSEMBLY__
#include <linux/types.h> #include <linux/types.h>
......
...@@ -94,7 +94,6 @@ struct dump_save_areas { ...@@ -94,7 +94,6 @@ struct dump_save_areas {
}; };
extern struct dump_save_areas dump_save_areas; extern struct dump_save_areas dump_save_areas;
struct save_area_ext *dump_save_area_create(int cpu);
extern void do_reipl(void); extern void do_reipl(void);
extern void do_halt(void); extern void do_halt(void);
......
...@@ -11,6 +11,7 @@ ...@@ -11,6 +11,7 @@
#include <asm/cpu.h> #include <asm/cpu.h>
#define SCLP_CHP_INFO_MASK_SIZE 32 #define SCLP_CHP_INFO_MASK_SIZE 32
#define SCLP_MAX_CORES 256
struct sclp_chp_info { struct sclp_chp_info {
u8 recognized[SCLP_CHP_INFO_MASK_SIZE]; u8 recognized[SCLP_CHP_INFO_MASK_SIZE];
...@@ -26,7 +27,7 @@ struct sclp_ipl_info { ...@@ -26,7 +27,7 @@ struct sclp_ipl_info {
char loadparm[LOADPARM_LEN]; char loadparm[LOADPARM_LEN];
}; };
struct sclp_cpu_entry { struct sclp_core_entry {
u8 core_id; u8 core_id;
u8 reserved0[2]; u8 reserved0[2];
u8 : 3; u8 : 3;
...@@ -38,12 +39,11 @@ struct sclp_cpu_entry { ...@@ -38,12 +39,11 @@ struct sclp_cpu_entry {
u8 reserved1; u8 reserved1;
} __attribute__((packed)); } __attribute__((packed));
struct sclp_cpu_info { struct sclp_core_info {
unsigned int configured; unsigned int configured;
unsigned int standby; unsigned int standby;
unsigned int combined; unsigned int combined;
int has_cpu_type; struct sclp_core_entry core[SCLP_MAX_CORES];
struct sclp_cpu_entry cpu[MAX_CPU_ADDRESS + 1];
}; };
struct sclp_info { struct sclp_info {
...@@ -51,7 +51,7 @@ struct sclp_info { ...@@ -51,7 +51,7 @@ struct sclp_info {
unsigned char has_vt220 : 1; unsigned char has_vt220 : 1;
unsigned char has_siif : 1; unsigned char has_siif : 1;
unsigned char has_sigpif : 1; unsigned char has_sigpif : 1;
unsigned char has_cpu_type : 1; unsigned char has_core_type : 1;
unsigned char has_sprp : 1; unsigned char has_sprp : 1;
unsigned int ibc; unsigned int ibc;
unsigned int mtid; unsigned int mtid;
...@@ -60,15 +60,15 @@ struct sclp_info { ...@@ -60,15 +60,15 @@ struct sclp_info {
unsigned long long rzm; unsigned long long rzm;
unsigned long long rnmax; unsigned long long rnmax;
unsigned long long hamax; unsigned long long hamax;
unsigned int max_cpu; unsigned int max_cores;
unsigned long hsa_size; unsigned long hsa_size;
unsigned long long facilities; unsigned long long facilities;
}; };
extern struct sclp_info sclp; extern struct sclp_info sclp;
int sclp_get_cpu_info(struct sclp_cpu_info *info); int sclp_get_core_info(struct sclp_core_info *info);
int sclp_cpu_configure(u8 cpu); int sclp_core_configure(u8 core);
int sclp_cpu_deconfigure(u8 cpu); int sclp_core_deconfigure(u8 core);
int sclp_sdias_blk_count(void); int sclp_sdias_blk_count(void);
int sclp_sdias_copy(void *dest, int blk_num, int nr_blks); int sclp_sdias_copy(void *dest, int blk_num, int nr_blks);
int sclp_chp_configure(struct chp_id chpid); int sclp_chp_configure(struct chp_id chpid);
......
...@@ -29,6 +29,7 @@ extern void smp_call_ipl_cpu(void (*func)(void *), void *); ...@@ -29,6 +29,7 @@ extern void smp_call_ipl_cpu(void (*func)(void *), void *);
extern int smp_find_processor_id(u16 address); extern int smp_find_processor_id(u16 address);
extern int smp_store_status(int cpu); extern int smp_store_status(int cpu);
extern void smp_save_dump_cpus(void);
extern int smp_vcpu_scheduled(int cpu); extern int smp_vcpu_scheduled(int cpu);
extern void smp_yield_cpu(int cpu); extern void smp_yield_cpu(int cpu);
extern void smp_cpu_set_polarization(int cpu, int val); extern void smp_cpu_set_polarization(int cpu, int val);
...@@ -54,6 +55,7 @@ static inline int smp_store_status(int cpu) { return 0; } ...@@ -54,6 +55,7 @@ static inline int smp_store_status(int cpu) { return 0; }
static inline int smp_vcpu_scheduled(int cpu) { return 1; } static inline int smp_vcpu_scheduled(int cpu) { return 1; }
static inline void smp_yield_cpu(int cpu) { } static inline void smp_yield_cpu(int cpu) { }
static inline void smp_fill_possible_mask(void) { } static inline void smp_fill_possible_mask(void) { }
static inline void smp_save_dump_cpus(void) { }
#endif /* CONFIG_SMP */ #endif /* CONFIG_SMP */
......
...@@ -78,15 +78,20 @@ s390_base_pgm_handler_fn: ...@@ -78,15 +78,20 @@ s390_base_pgm_handler_fn:
# #
# Calls diag 308 subcode 1 and continues execution # Calls diag 308 subcode 1 and continues execution
# #
# The following conditions must be ensured before calling this function:
# * Prefix register = 0
# * Lowcore protection is disabled
#
ENTRY(diag308_reset) ENTRY(diag308_reset)
larl %r4,.Lctlregs # Save control registers larl %r4,.Lctlregs # Save control registers
stctg %c0,%c15,0(%r4) stctg %c0,%c15,0(%r4)
lg %r2,0(%r4) # Disable lowcore protection
nilh %r2,0xefff
larl %r4,.Lctlreg0
stg %r2,0(%r4)
lctlg %c0,%c0,0(%r4)
larl %r4,.Lfpctl # Floating point control register larl %r4,.Lfpctl # Floating point control register
stfpc 0(%r4) stfpc 0(%r4)
larl %r4,.Lprefix # Save prefix register
stpx 0(%r4)
larl %r4,.Lprefix_zero # Set prefix register to 0
spx 0(%r4)
larl %r4,.Lcontinue_psw # Save PSW flags larl %r4,.Lcontinue_psw # Save PSW flags
epsw %r2,%r3 epsw %r2,%r3
stm %r2,%r3,0(%r4) stm %r2,%r3,0(%r4)
...@@ -106,6 +111,8 @@ ENTRY(diag308_reset) ...@@ -106,6 +111,8 @@ ENTRY(diag308_reset)
lctlg %c0,%c15,0(%r4) lctlg %c0,%c15,0(%r4)
larl %r4,.Lfpctl # Restore floating point ctl register larl %r4,.Lfpctl # Restore floating point ctl register
lfpc 0(%r4) lfpc 0(%r4)
larl %r4,.Lprefix # Restore prefix register
spx 0(%r4)
larl %r4,.Lcontinue_psw # Restore PSW flags larl %r4,.Lcontinue_psw # Restore PSW flags
lpswe 0(%r4) lpswe 0(%r4)
.Lcontinue: .Lcontinue:
...@@ -122,10 +129,16 @@ ENTRY(diag308_reset) ...@@ -122,10 +129,16 @@ ENTRY(diag308_reset)
.section .bss .section .bss
.align 8 .align 8
.Lctlreg0:
.quad 0
.Lctlregs: .Lctlregs:
.rept 16 .rept 16
.quad 0 .quad 0
.endr .endr
.Lfpctl: .Lfpctl:
.long 0 .long 0
.Lprefix:
.long 0
.Lprefix_zero:
.long 0
.previous .previous
...@@ -44,31 +44,6 @@ static struct memblock_type oldmem_type = { ...@@ -44,31 +44,6 @@ static struct memblock_type oldmem_type = {
struct dump_save_areas dump_save_areas; struct dump_save_areas dump_save_areas;
/*
* Allocate and add a save area for a CPU
*/
struct save_area_ext *dump_save_area_create(int cpu)
{
struct save_area_ext **save_areas, *save_area;
save_area = kmalloc(sizeof(*save_area), GFP_KERNEL);
if (!save_area)
return NULL;
if (cpu + 1 > dump_save_areas.count) {
dump_save_areas.count = cpu + 1;
save_areas = krealloc(dump_save_areas.areas,
dump_save_areas.count * sizeof(void *),
GFP_KERNEL | __GFP_ZERO);
if (!save_areas) {
kfree(save_area);
return NULL;
}
dump_save_areas.areas = save_areas;
}
dump_save_areas.areas[cpu] = save_area;
return save_area;
}
/* /*
* Return physical address for virtual address * Return physical address for virtual address
*/ */
...@@ -416,7 +391,7 @@ static void *nt_s390_vx_low(void *ptr, __vector128 *vx_regs) ...@@ -416,7 +391,7 @@ static void *nt_s390_vx_low(void *ptr, __vector128 *vx_regs)
ptr += len; ptr += len;
/* Copy lower halves of SIMD registers 0-15 */ /* Copy lower halves of SIMD registers 0-15 */
for (i = 0; i < 16; i++) { for (i = 0; i < 16; i++) {
memcpy(ptr, &vx_regs[i], 8); memcpy(ptr, &vx_regs[i].u[2], 8);
ptr += 8; ptr += 8;
} }
return ptr; return ptr;
......
...@@ -868,6 +868,11 @@ void __init setup_arch(char **cmdline_p) ...@@ -868,6 +868,11 @@ void __init setup_arch(char **cmdline_p)
check_initrd(); check_initrd();
reserve_crashkernel(); reserve_crashkernel();
/*
* Be aware that smp_save_dump_cpus() triggers a system reset.
* Therefore CPU and device initialization should be done afterwards.
*/
smp_save_dump_cpus();
setup_resources(); setup_resources();
setup_vmcoreinfo(); setup_vmcoreinfo();
......
...@@ -31,6 +31,7 @@ ...@@ -31,6 +31,7 @@
#include <linux/cpu.h> #include <linux/cpu.h>
#include <linux/slab.h> #include <linux/slab.h>
#include <linux/crash_dump.h> #include <linux/crash_dump.h>
#include <linux/memblock.h>
#include <asm/asm-offsets.h> #include <asm/asm-offsets.h>
#include <asm/switch_to.h> #include <asm/switch_to.h>
#include <asm/facility.h> #include <asm/facility.h>
...@@ -69,7 +70,7 @@ struct pcpu { ...@@ -69,7 +70,7 @@ struct pcpu {
u16 address; /* physical cpu address */ u16 address; /* physical cpu address */
}; };
static u8 boot_cpu_type; static u8 boot_core_type;
static struct pcpu pcpu_devices[NR_CPUS]; static struct pcpu pcpu_devices[NR_CPUS];
unsigned int smp_cpu_mt_shift; unsigned int smp_cpu_mt_shift;
...@@ -531,15 +532,12 @@ EXPORT_SYMBOL(smp_ctl_clear_bit); ...@@ -531,15 +532,12 @@ EXPORT_SYMBOL(smp_ctl_clear_bit);
#ifdef CONFIG_CRASH_DUMP #ifdef CONFIG_CRASH_DUMP
static inline void __smp_store_cpu_state(int cpu, u16 address, int is_boot_cpu) static void __smp_store_cpu_state(struct save_area_ext *sa_ext, u16 address,
int is_boot_cpu)
{ {
void *lc = pcpu_devices[0].lowcore; void *lc = (void *)(unsigned long) store_prefix();
struct save_area_ext *sa_ext;
unsigned long vx_sa; unsigned long vx_sa;
sa_ext = dump_save_area_create(cpu);
if (!sa_ext)
panic("could not allocate memory for save area\n");
if (is_boot_cpu) { if (is_boot_cpu) {
/* Copy the registers of the boot CPU. */ /* Copy the registers of the boot CPU. */
copy_oldmem_page(1, (void *) &sa_ext->sa, sizeof(sa_ext->sa), copy_oldmem_page(1, (void *) &sa_ext->sa, sizeof(sa_ext->sa),
...@@ -554,14 +552,33 @@ static inline void __smp_store_cpu_state(int cpu, u16 address, int is_boot_cpu) ...@@ -554,14 +552,33 @@ static inline void __smp_store_cpu_state(int cpu, u16 address, int is_boot_cpu)
if (!MACHINE_HAS_VX) if (!MACHINE_HAS_VX)
return; return;
/* Get the VX registers */ /* Get the VX registers */
vx_sa = __get_free_page(GFP_KERNEL); vx_sa = memblock_alloc(PAGE_SIZE, PAGE_SIZE);
if (!vx_sa) if (!vx_sa)
panic("could not allocate memory for VX save area\n"); panic("could not allocate memory for VX save area\n");
__pcpu_sigp_relax(address, SIGP_STORE_ADDITIONAL_STATUS, vx_sa, NULL); __pcpu_sigp_relax(address, SIGP_STORE_ADDITIONAL_STATUS, vx_sa, NULL);
memcpy(sa_ext->vx_regs, (void *) vx_sa, sizeof(sa_ext->vx_regs)); memcpy(sa_ext->vx_regs, (void *) vx_sa, sizeof(sa_ext->vx_regs));
free_page(vx_sa); memblock_free(vx_sa, PAGE_SIZE);
} }
int smp_store_status(int cpu)
{
unsigned long vx_sa;
struct pcpu *pcpu;
pcpu = pcpu_devices + cpu;
if (__pcpu_sigp_relax(pcpu->address, SIGP_STOP_AND_STORE_STATUS,
0, NULL) != SIGP_CC_ORDER_CODE_ACCEPTED)
return -EIO;
if (!MACHINE_HAS_VX)
return 0;
vx_sa = __pa(pcpu->lowcore->vector_save_area_addr);
__pcpu_sigp_relax(pcpu->address, SIGP_STORE_ADDITIONAL_STATUS,
vx_sa, NULL);
return 0;
}
#endif /* CONFIG_CRASH_DUMP */
/* /*
* Collect CPU state of the previous, crashed system. * Collect CPU state of the previous, crashed system.
* There are four cases: * There are four cases:
...@@ -589,10 +606,12 @@ static inline void __smp_store_cpu_state(int cpu, u16 address, int is_boot_cpu) ...@@ -589,10 +606,12 @@ static inline void __smp_store_cpu_state(int cpu, u16 address, int is_boot_cpu)
* old system. The ELF sections are picked up by the crash_dump code * old system. The ELF sections are picked up by the crash_dump code
* via elfcorehdr_addr. * via elfcorehdr_addr.
*/ */
static void __init smp_store_cpu_states(struct sclp_cpu_info *info) void __init smp_save_dump_cpus(void)
{ {
unsigned int cpu, address, i, j; #ifdef CONFIG_CRASH_DUMP
int is_boot_cpu; int addr, cpu, boot_cpu_addr, max_cpu_addr;
struct save_area_ext *sa_ext;
bool is_boot_cpu;
if (is_kdump_kernel()) if (is_kdump_kernel())
/* Previous system stored the CPU states. Nothing to do. */ /* Previous system stored the CPU states. Nothing to do. */
...@@ -602,42 +621,36 @@ static void __init smp_store_cpu_states(struct sclp_cpu_info *info) ...@@ -602,42 +621,36 @@ static void __init smp_store_cpu_states(struct sclp_cpu_info *info)
return; return;
/* Set multi-threading state to the previous system. */ /* Set multi-threading state to the previous system. */
pcpu_set_smt(sclp.mtid_prev); pcpu_set_smt(sclp.mtid_prev);
/* Collect CPU states. */ max_cpu_addr = SCLP_MAX_CORES << sclp.mtid_prev;
cpu = 0; for (cpu = 0, addr = 0; addr <= max_cpu_addr; addr++) {
for (i = 0; i < info->configured; i++) { if (__pcpu_sigp_relax(addr, SIGP_SENSE, 0, NULL) ==
/* Skip CPUs with different CPU type. */ SIGP_CC_NOT_OPERATIONAL)
if (info->has_cpu_type && info->cpu[i].type != boot_cpu_type)
continue; continue;
for (j = 0; j <= smp_cpu_mtid; j++, cpu++) { cpu += 1;
address = (info->cpu[i].core_id << smp_cpu_mt_shift) + j;
is_boot_cpu = (address == pcpu_devices[0].address);
if (is_boot_cpu && !OLDMEM_BASE)
/* Skip boot CPU for standard zfcp dump. */
continue;
/* Get state for this CPu. */
__smp_store_cpu_state(cpu, address, is_boot_cpu);
}
} }
} dump_save_areas.areas = (void *)memblock_alloc(sizeof(void *) * cpu, 8);
dump_save_areas.count = cpu;
int smp_store_status(int cpu) boot_cpu_addr = stap();
{ for (cpu = 0, addr = 0; addr <= max_cpu_addr; addr++) {
unsigned long vx_sa; if (__pcpu_sigp_relax(addr, SIGP_SENSE, 0, NULL) ==
struct pcpu *pcpu; SIGP_CC_NOT_OPERATIONAL)
continue;
pcpu = pcpu_devices + cpu; sa_ext = (void *) memblock_alloc(sizeof(*sa_ext), 8);
if (__pcpu_sigp_relax(pcpu->address, SIGP_STOP_AND_STORE_STATUS, dump_save_areas.areas[cpu] = sa_ext;
0, NULL) != SIGP_CC_ORDER_CODE_ACCEPTED) if (!sa_ext)
return -EIO; panic("could not allocate memory for save area\n");
if (!MACHINE_HAS_VX) is_boot_cpu = (addr == boot_cpu_addr);
return 0; cpu += 1;
vx_sa = __pa(pcpu->lowcore->vector_save_area_addr); if (is_boot_cpu && !OLDMEM_BASE)
__pcpu_sigp_relax(pcpu->address, SIGP_STORE_ADDITIONAL_STATUS, /* Skip boot CPU for standard zfcp dump. */
vx_sa, NULL); continue;
return 0; /* Get state for this CPU. */
} __smp_store_cpu_state(sa_ext, addr, is_boot_cpu);
}
diag308_reset();
pcpu_set_smt(0);
#endif /* CONFIG_CRASH_DUMP */ #endif /* CONFIG_CRASH_DUMP */
}
void smp_cpu_set_polarization(int cpu, int val) void smp_cpu_set_polarization(int cpu, int val)
{ {
...@@ -649,21 +662,22 @@ int smp_cpu_get_polarization(int cpu) ...@@ -649,21 +662,22 @@ int smp_cpu_get_polarization(int cpu)
return pcpu_devices[cpu].polarization; return pcpu_devices[cpu].polarization;
} }
static struct sclp_cpu_info *smp_get_cpu_info(void) static struct sclp_core_info *smp_get_core_info(void)
{ {
static int use_sigp_detection; static int use_sigp_detection;
struct sclp_cpu_info *info; struct sclp_core_info *info;
int address; int address;
info = kzalloc(sizeof(*info), GFP_KERNEL); info = kzalloc(sizeof(*info), GFP_KERNEL);
if (info && (use_sigp_detection || sclp_get_cpu_info(info))) { if (info && (use_sigp_detection || sclp_get_core_info(info))) {
use_sigp_detection = 1; use_sigp_detection = 1;
for (address = 0; address <= MAX_CPU_ADDRESS; for (address = 0;
address < (SCLP_MAX_CORES << smp_cpu_mt_shift);
address += (1U << smp_cpu_mt_shift)) { address += (1U << smp_cpu_mt_shift)) {
if (__pcpu_sigp_relax(address, SIGP_SENSE, 0, NULL) == if (__pcpu_sigp_relax(address, SIGP_SENSE, 0, NULL) ==
SIGP_CC_NOT_OPERATIONAL) SIGP_CC_NOT_OPERATIONAL)
continue; continue;
info->cpu[info->configured].core_id = info->core[info->configured].core_id =
address >> smp_cpu_mt_shift; address >> smp_cpu_mt_shift;
info->configured++; info->configured++;
} }
...@@ -674,7 +688,7 @@ static struct sclp_cpu_info *smp_get_cpu_info(void) ...@@ -674,7 +688,7 @@ static struct sclp_cpu_info *smp_get_cpu_info(void)
static int smp_add_present_cpu(int cpu); static int smp_add_present_cpu(int cpu);
static int __smp_rescan_cpus(struct sclp_cpu_info *info, int sysfs_add) static int __smp_rescan_cpus(struct sclp_core_info *info, int sysfs_add)
{ {
struct pcpu *pcpu; struct pcpu *pcpu;
cpumask_t avail; cpumask_t avail;
...@@ -685,9 +699,9 @@ static int __smp_rescan_cpus(struct sclp_cpu_info *info, int sysfs_add) ...@@ -685,9 +699,9 @@ static int __smp_rescan_cpus(struct sclp_cpu_info *info, int sysfs_add)
cpumask_xor(&avail, cpu_possible_mask, cpu_present_mask); cpumask_xor(&avail, cpu_possible_mask, cpu_present_mask);
cpu = cpumask_first(&avail); cpu = cpumask_first(&avail);
for (i = 0; (i < info->combined) && (cpu < nr_cpu_ids); i++) { for (i = 0; (i < info->combined) && (cpu < nr_cpu_ids); i++) {
if (info->has_cpu_type && info->cpu[i].type != boot_cpu_type) if (sclp.has_core_type && info->core[i].type != boot_core_type)
continue; continue;
address = info->cpu[i].core_id << smp_cpu_mt_shift; address = info->core[i].core_id << smp_cpu_mt_shift;
for (j = 0; j <= smp_cpu_mtid; j++) { for (j = 0; j <= smp_cpu_mtid; j++) {
if (pcpu_find_address(cpu_present_mask, address + j)) if (pcpu_find_address(cpu_present_mask, address + j))
continue; continue;
...@@ -713,41 +727,37 @@ static int __smp_rescan_cpus(struct sclp_cpu_info *info, int sysfs_add) ...@@ -713,41 +727,37 @@ static int __smp_rescan_cpus(struct sclp_cpu_info *info, int sysfs_add)
static void __init smp_detect_cpus(void) static void __init smp_detect_cpus(void)
{ {
unsigned int cpu, mtid, c_cpus, s_cpus; unsigned int cpu, mtid, c_cpus, s_cpus;
struct sclp_cpu_info *info; struct sclp_core_info *info;
u16 address; u16 address;
/* Get CPU information */ /* Get CPU information */
info = smp_get_cpu_info(); info = smp_get_core_info();
if (!info) if (!info)
panic("smp_detect_cpus failed to allocate memory\n"); panic("smp_detect_cpus failed to allocate memory\n");
/* Find boot CPU type */ /* Find boot CPU type */
if (info->has_cpu_type) { if (sclp.has_core_type) {
address = stap(); address = stap();
for (cpu = 0; cpu < info->combined; cpu++) for (cpu = 0; cpu < info->combined; cpu++)
if (info->cpu[cpu].core_id == address) { if (info->core[cpu].core_id == address) {
/* The boot cpu dictates the cpu type. */ /* The boot cpu dictates the cpu type. */
boot_cpu_type = info->cpu[cpu].type; boot_core_type = info->core[cpu].type;
break; break;
} }
if (cpu >= info->combined) if (cpu >= info->combined)
panic("Could not find boot CPU type"); panic("Could not find boot CPU type");
} }
#ifdef CONFIG_CRASH_DUMP
/* Collect CPU state of previous system */
smp_store_cpu_states(info);
#endif
/* Set multi-threading state for the current system */ /* Set multi-threading state for the current system */
mtid = boot_cpu_type ? sclp.mtid : sclp.mtid_cp; mtid = boot_core_type ? sclp.mtid : sclp.mtid_cp;
mtid = (mtid < smp_max_threads) ? mtid : smp_max_threads - 1; mtid = (mtid < smp_max_threads) ? mtid : smp_max_threads - 1;
pcpu_set_smt(mtid); pcpu_set_smt(mtid);
/* Print number of CPUs */ /* Print number of CPUs */
c_cpus = s_cpus = 0; c_cpus = s_cpus = 0;
for (cpu = 0; cpu < info->combined; cpu++) { for (cpu = 0; cpu < info->combined; cpu++) {
if (info->has_cpu_type && info->cpu[cpu].type != boot_cpu_type) if (sclp.has_core_type &&
info->core[cpu].type != boot_core_type)
continue; continue;
if (cpu < info->configured) if (cpu < info->configured)
c_cpus += smp_cpu_mtid + 1; c_cpus += smp_cpu_mtid + 1;
...@@ -884,7 +894,7 @@ void __init smp_fill_possible_mask(void) ...@@ -884,7 +894,7 @@ void __init smp_fill_possible_mask(void)
sclp_max = max(sclp.mtid, sclp.mtid_cp) + 1; sclp_max = max(sclp.mtid, sclp.mtid_cp) + 1;
sclp_max = min(smp_max_threads, sclp_max); sclp_max = min(smp_max_threads, sclp_max);
sclp_max = sclp.max_cpu * sclp_max ?: nr_cpu_ids; sclp_max = sclp.max_cores * sclp_max ?: nr_cpu_ids;
possible = setup_possible_cpus ?: nr_cpu_ids; possible = setup_possible_cpus ?: nr_cpu_ids;
possible = min(possible, sclp_max); possible = min(possible, sclp_max);
for (cpu = 0; cpu < possible && cpu < nr_cpu_ids; cpu++) for (cpu = 0; cpu < possible && cpu < nr_cpu_ids; cpu++)
...@@ -977,7 +987,7 @@ static ssize_t cpu_configure_store(struct device *dev, ...@@ -977,7 +987,7 @@ static ssize_t cpu_configure_store(struct device *dev,
case 0: case 0:
if (pcpu->state != CPU_STATE_CONFIGURED) if (pcpu->state != CPU_STATE_CONFIGURED)
break; break;
rc = sclp_cpu_deconfigure(pcpu->address >> smp_cpu_mt_shift); rc = sclp_core_deconfigure(pcpu->address >> smp_cpu_mt_shift);
if (rc) if (rc)
break; break;
for (i = 0; i <= smp_cpu_mtid; i++) { for (i = 0; i <= smp_cpu_mtid; i++) {
...@@ -992,7 +1002,7 @@ static ssize_t cpu_configure_store(struct device *dev, ...@@ -992,7 +1002,7 @@ static ssize_t cpu_configure_store(struct device *dev,
case 1: case 1:
if (pcpu->state != CPU_STATE_STANDBY) if (pcpu->state != CPU_STATE_STANDBY)
break; break;
rc = sclp_cpu_configure(pcpu->address >> smp_cpu_mt_shift); rc = sclp_core_configure(pcpu->address >> smp_cpu_mt_shift);
if (rc) if (rc)
break; break;
for (i = 0; i <= smp_cpu_mtid; i++) { for (i = 0; i <= smp_cpu_mtid; i++) {
...@@ -1107,10 +1117,10 @@ static int smp_add_present_cpu(int cpu) ...@@ -1107,10 +1117,10 @@ static int smp_add_present_cpu(int cpu)
int __ref smp_rescan_cpus(void) int __ref smp_rescan_cpus(void)
{ {
struct sclp_cpu_info *info; struct sclp_core_info *info;
int nr; int nr;
info = smp_get_cpu_info(); info = smp_get_core_info();
if (!info) if (!info)
return -ENOMEM; return -ENOMEM;
get_online_cpus(); get_online_cpus();
......
...@@ -250,7 +250,7 @@ static inline void reg_set_seen(struct bpf_jit *jit, u32 b1) ...@@ -250,7 +250,7 @@ static inline void reg_set_seen(struct bpf_jit *jit, u32 b1)
({ \ ({ \
/* Branch instruction needs 6 bytes */ \ /* Branch instruction needs 6 bytes */ \
int rel = (addrs[i + off + 1] - (addrs[i + 1] - 6)) / 2;\ int rel = (addrs[i + off + 1] - (addrs[i + 1] - 6)) / 2;\
_EMIT6(op1 | reg(b1, b2) << 16 | rel, op2 | mask); \ _EMIT6(op1 | reg(b1, b2) << 16 | (rel & 0xffff), op2 | mask); \
REG_SET_SEEN(b1); \ REG_SET_SEEN(b1); \
REG_SET_SEEN(b2); \ REG_SET_SEEN(b2); \
}) })
......
...@@ -92,8 +92,8 @@ struct read_cpu_info_sccb { ...@@ -92,8 +92,8 @@ struct read_cpu_info_sccb {
u8 reserved[4096 - 16]; u8 reserved[4096 - 16];
} __attribute__((packed, aligned(PAGE_SIZE))); } __attribute__((packed, aligned(PAGE_SIZE)));
static void sclp_fill_cpu_info(struct sclp_cpu_info *info, static void sclp_fill_core_info(struct sclp_core_info *info,
struct read_cpu_info_sccb *sccb) struct read_cpu_info_sccb *sccb)
{ {
char *page = (char *) sccb; char *page = (char *) sccb;
...@@ -101,12 +101,11 @@ static void sclp_fill_cpu_info(struct sclp_cpu_info *info, ...@@ -101,12 +101,11 @@ static void sclp_fill_cpu_info(struct sclp_cpu_info *info,
info->configured = sccb->nr_configured; info->configured = sccb->nr_configured;
info->standby = sccb->nr_standby; info->standby = sccb->nr_standby;
info->combined = sccb->nr_configured + sccb->nr_standby; info->combined = sccb->nr_configured + sccb->nr_standby;
info->has_cpu_type = sclp.has_cpu_type; memcpy(&info->core, page + sccb->offset_configured,
memcpy(&info->cpu, page + sccb->offset_configured, info->combined * sizeof(struct sclp_core_entry));
info->combined * sizeof(struct sclp_cpu_entry));
} }
int sclp_get_cpu_info(struct sclp_cpu_info *info) int sclp_get_core_info(struct sclp_core_info *info)
{ {
int rc; int rc;
struct read_cpu_info_sccb *sccb; struct read_cpu_info_sccb *sccb;
...@@ -127,7 +126,7 @@ int sclp_get_cpu_info(struct sclp_cpu_info *info) ...@@ -127,7 +126,7 @@ int sclp_get_cpu_info(struct sclp_cpu_info *info)
rc = -EIO; rc = -EIO;
goto out; goto out;
} }
sclp_fill_cpu_info(info, sccb); sclp_fill_core_info(info, sccb);
out: out:
free_page((unsigned long) sccb); free_page((unsigned long) sccb);
return rc; return rc;
...@@ -137,7 +136,7 @@ struct cpu_configure_sccb { ...@@ -137,7 +136,7 @@ struct cpu_configure_sccb {
struct sccb_header header; struct sccb_header header;
} __attribute__((packed, aligned(8))); } __attribute__((packed, aligned(8)));
static int do_cpu_configure(sclp_cmdw_t cmd) static int do_core_configure(sclp_cmdw_t cmd)
{ {
struct cpu_configure_sccb *sccb; struct cpu_configure_sccb *sccb;
int rc; int rc;
...@@ -171,14 +170,14 @@ static int do_cpu_configure(sclp_cmdw_t cmd) ...@@ -171,14 +170,14 @@ static int do_cpu_configure(sclp_cmdw_t cmd)
return rc; return rc;
} }
int sclp_cpu_configure(u8 cpu) int sclp_core_configure(u8 core)
{ {
return do_cpu_configure(SCLP_CMDW_CONFIGURE_CPU | cpu << 8); return do_core_configure(SCLP_CMDW_CONFIGURE_CPU | core << 8);
} }
int sclp_cpu_deconfigure(u8 cpu) int sclp_core_deconfigure(u8 core)
{ {
return do_cpu_configure(SCLP_CMDW_DECONFIGURE_CPU | cpu << 8); return do_core_configure(SCLP_CMDW_DECONFIGURE_CPU | core << 8);
} }
#ifdef CONFIG_MEMORY_HOTPLUG #ifdef CONFIG_MEMORY_HOTPLUG
......
...@@ -98,7 +98,7 @@ static int __init sclp_read_info_early(struct read_info_sccb *sccb) ...@@ -98,7 +98,7 @@ static int __init sclp_read_info_early(struct read_info_sccb *sccb)
static void __init sclp_facilities_detect(struct read_info_sccb *sccb) static void __init sclp_facilities_detect(struct read_info_sccb *sccb)
{ {
struct sclp_cpu_entry *cpue; struct sclp_core_entry *cpue;
u16 boot_cpu_address, cpu; u16 boot_cpu_address, cpu;
if (sclp_read_info_early(sccb)) if (sclp_read_info_early(sccb))
...@@ -106,7 +106,7 @@ static void __init sclp_facilities_detect(struct read_info_sccb *sccb) ...@@ -106,7 +106,7 @@ static void __init sclp_facilities_detect(struct read_info_sccb *sccb)
sclp.facilities = sccb->facilities; sclp.facilities = sccb->facilities;
sclp.has_sprp = !!(sccb->fac84 & 0x02); sclp.has_sprp = !!(sccb->fac84 & 0x02);
sclp.has_cpu_type = !!(sccb->fac84 & 0x01); sclp.has_core_type = !!(sccb->fac84 & 0x01);
if (sccb->fac85 & 0x02) if (sccb->fac85 & 0x02)
S390_lowcore.machine_flags |= MACHINE_FLAG_ESOP; S390_lowcore.machine_flags |= MACHINE_FLAG_ESOP;
sclp.rnmax = sccb->rnmax ? sccb->rnmax : sccb->rnmax2; sclp.rnmax = sccb->rnmax ? sccb->rnmax : sccb->rnmax2;
...@@ -116,11 +116,11 @@ static void __init sclp_facilities_detect(struct read_info_sccb *sccb) ...@@ -116,11 +116,11 @@ static void __init sclp_facilities_detect(struct read_info_sccb *sccb)
if (!sccb->hcpua) { if (!sccb->hcpua) {
if (MACHINE_IS_VM) if (MACHINE_IS_VM)
sclp.max_cpu = 64; sclp.max_cores = 64;
else else
sclp.max_cpu = sccb->ncpurl; sclp.max_cores = sccb->ncpurl;
} else { } else {
sclp.max_cpu = sccb->hcpua + 1; sclp.max_cores = sccb->hcpua + 1;
} }
boot_cpu_address = stap(); boot_cpu_address = stap();
......
...@@ -154,7 +154,7 @@ static int __init init_cpu_info(enum arch_id arch) ...@@ -154,7 +154,7 @@ static int __init init_cpu_info(enum arch_id arch)
/* get info for boot cpu from lowcore, stored in the HSA */ /* get info for boot cpu from lowcore, stored in the HSA */
sa_ext = dump_save_area_create(0); sa_ext = dump_save_areas.areas[0];
if (!sa_ext) if (!sa_ext)
return -ENOMEM; return -ENOMEM;
if (memcpy_hsa_kernel(&sa_ext->sa, sys_info.sa_base, if (memcpy_hsa_kernel(&sa_ext->sa, sys_info.sa_base,
......
...@@ -60,7 +60,7 @@ static int __ap_poll_device(struct ap_device *ap_dev, unsigned long *flags); ...@@ -60,7 +60,7 @@ static int __ap_poll_device(struct ap_device *ap_dev, unsigned long *flags);
static int ap_device_remove(struct device *dev); static int ap_device_remove(struct device *dev);
static int ap_device_probe(struct device *dev); static int ap_device_probe(struct device *dev);
static void ap_interrupt_handler(struct airq_struct *airq); static void ap_interrupt_handler(struct airq_struct *airq);
static void ap_reset(struct ap_device *ap_dev); static void ap_reset(struct ap_device *ap_dev, unsigned long *flags);
static void ap_config_timeout(unsigned long ptr); static void ap_config_timeout(unsigned long ptr);
static int ap_select_domain(void); static int ap_select_domain(void);
static void ap_query_configuration(void); static void ap_query_configuration(void);
...@@ -310,35 +310,26 @@ static inline int __ap_query_configuration(struct ap_config_info *config) ...@@ -310,35 +310,26 @@ static inline int __ap_query_configuration(struct ap_config_info *config)
static int ap_query_functions(ap_qid_t qid, unsigned int *functions) static int ap_query_functions(ap_qid_t qid, unsigned int *functions)
{ {
struct ap_queue_status status; struct ap_queue_status status;
int i;
status = __ap_query_functions(qid, functions); status = __ap_query_functions(qid, functions);
for (i = 0; i < AP_MAX_RESET; i++) { if (ap_queue_status_invalid_test(&status))
if (ap_queue_status_invalid_test(&status)) return -ENODEV;
return -ENODEV;
switch (status.response_code) { switch (status.response_code) {
case AP_RESPONSE_NORMAL: case AP_RESPONSE_NORMAL:
return 0; return 0;
case AP_RESPONSE_RESET_IN_PROGRESS: case AP_RESPONSE_Q_NOT_AVAIL:
case AP_RESPONSE_BUSY: case AP_RESPONSE_DECONFIGURED:
break; case AP_RESPONSE_CHECKSTOPPED:
case AP_RESPONSE_Q_NOT_AVAIL: case AP_RESPONSE_INVALID_ADDRESS:
case AP_RESPONSE_DECONFIGURED: return -ENODEV;
case AP_RESPONSE_CHECKSTOPPED: case AP_RESPONSE_RESET_IN_PROGRESS:
case AP_RESPONSE_INVALID_ADDRESS: case AP_RESPONSE_BUSY:
return -ENODEV; case AP_RESPONSE_OTHERWISE_CHANGED:
case AP_RESPONSE_OTHERWISE_CHANGED: default:
break; return -EBUSY;
default:
break;
}
if (i < AP_MAX_RESET - 1) {
udelay(5);
status = __ap_query_functions(qid, functions);
}
} }
return -EBUSY;
} }
/** /**
...@@ -350,47 +341,25 @@ static int ap_query_functions(ap_qid_t qid, unsigned int *functions) ...@@ -350,47 +341,25 @@ static int ap_query_functions(ap_qid_t qid, unsigned int *functions)
* on the return value it waits a while and tests the AP queue if interrupts * on the return value it waits a while and tests the AP queue if interrupts
* have been switched on using ap_test_queue(). * have been switched on using ap_test_queue().
*/ */
static int ap_queue_enable_interruption(ap_qid_t qid, void *ind) static int ap_queue_enable_interruption(struct ap_device *ap_dev, void *ind)
{ {
struct ap_queue_status status; struct ap_queue_status status;
int t_depth, t_device_type, rc, i;
rc = -EBUSY; status = ap_queue_interruption_control(ap_dev->qid, ind);
status = ap_queue_interruption_control(qid, ind); switch (status.response_code) {
case AP_RESPONSE_NORMAL:
for (i = 0; i < AP_MAX_RESET; i++) { case AP_RESPONSE_OTHERWISE_CHANGED:
switch (status.response_code) { return 0;
case AP_RESPONSE_NORMAL: case AP_RESPONSE_Q_NOT_AVAIL:
if (status.int_enabled) case AP_RESPONSE_DECONFIGURED:
return 0; case AP_RESPONSE_CHECKSTOPPED:
break; case AP_RESPONSE_INVALID_ADDRESS:
case AP_RESPONSE_RESET_IN_PROGRESS: return -ENODEV;
case AP_RESPONSE_BUSY: case AP_RESPONSE_RESET_IN_PROGRESS:
if (i < AP_MAX_RESET - 1) { case AP_RESPONSE_BUSY:
udelay(5); default:
status = ap_queue_interruption_control(qid, return -EBUSY;
ind);
continue;
}
break;
case AP_RESPONSE_Q_NOT_AVAIL:
case AP_RESPONSE_DECONFIGURED:
case AP_RESPONSE_CHECKSTOPPED:
case AP_RESPONSE_INVALID_ADDRESS:
return -ENODEV;
case AP_RESPONSE_OTHERWISE_CHANGED:
if (status.int_enabled)
return 0;
break;
default:
break;
}
if (i < AP_MAX_RESET - 1) {
udelay(5);
status = ap_test_queue(qid, &t_depth, &t_device_type);
}
} }
return rc;
} }
/** /**
...@@ -510,110 +479,95 @@ int ap_recv(ap_qid_t qid, unsigned long long *psmid, void *msg, size_t length) ...@@ -510,110 +479,95 @@ int ap_recv(ap_qid_t qid, unsigned long long *psmid, void *msg, size_t length)
} }
EXPORT_SYMBOL(ap_recv); EXPORT_SYMBOL(ap_recv);
/**
* __ap_schedule_poll_timer(): Schedule poll timer.
*
* Set up the timer to run the poll tasklet
*/
static inline void __ap_schedule_poll_timer(void)
{
ktime_t hr_time;
spin_lock_bh(&ap_poll_timer_lock);
if (!hrtimer_is_queued(&ap_poll_timer) && !ap_suspend_flag) {
hr_time = ktime_set(0, poll_timeout);
hrtimer_forward_now(&ap_poll_timer, hr_time);
hrtimer_restart(&ap_poll_timer);
}
spin_unlock_bh(&ap_poll_timer_lock);
}
/**
* ap_schedule_poll_timer(): Schedule poll timer.
*
* Set up the timer to run the poll tasklet
*/
static inline void ap_schedule_poll_timer(void)
{
if (ap_using_interrupts())
return;
__ap_schedule_poll_timer();
}
/** /**
* ap_query_queue(): Check if an AP queue is available. * ap_query_queue(): Check if an AP queue is available.
* @qid: The AP queue number * @qid: The AP queue number
* @queue_depth: Pointer to queue depth value * @queue_depth: Pointer to queue depth value
* @device_type: Pointer to device type value * @device_type: Pointer to device type value
*
* The test is repeated for AP_MAX_RESET times.
*/ */
static int ap_query_queue(ap_qid_t qid, int *queue_depth, int *device_type) static int ap_query_queue(ap_qid_t qid, int *queue_depth, int *device_type)
{ {
struct ap_queue_status status; struct ap_queue_status status;
int t_depth, t_device_type, rc, i; int t_depth, t_device_type;
rc = -EBUSY; status = ap_test_queue(qid, &t_depth, &t_device_type);
for (i = 0; i < AP_MAX_RESET; i++) { switch (status.response_code) {
status = ap_test_queue(qid, &t_depth, &t_device_type); case AP_RESPONSE_NORMAL:
switch (status.response_code) { *queue_depth = t_depth + 1;
case AP_RESPONSE_NORMAL: *device_type = t_device_type;
*queue_depth = t_depth + 1; return 0;
*device_type = t_device_type; case AP_RESPONSE_Q_NOT_AVAIL:
rc = 0; case AP_RESPONSE_DECONFIGURED:
break; case AP_RESPONSE_CHECKSTOPPED:
case AP_RESPONSE_Q_NOT_AVAIL: case AP_RESPONSE_INVALID_ADDRESS:
rc = -ENODEV; return -ENODEV;
break; case AP_RESPONSE_RESET_IN_PROGRESS:
case AP_RESPONSE_RESET_IN_PROGRESS: case AP_RESPONSE_OTHERWISE_CHANGED:
break; case AP_RESPONSE_BUSY:
case AP_RESPONSE_DECONFIGURED: return -EBUSY;
rc = -ENODEV; default:
break; BUG();
case AP_RESPONSE_CHECKSTOPPED:
rc = -ENODEV;
break;
case AP_RESPONSE_INVALID_ADDRESS:
rc = -ENODEV;
break;
case AP_RESPONSE_OTHERWISE_CHANGED:
break;
case AP_RESPONSE_BUSY:
break;
default:
BUG();
}
if (rc != -EBUSY)
break;
if (i < AP_MAX_RESET - 1)
udelay(5);
} }
return rc;
} }
/** /**
* ap_init_queue(): Reset an AP queue. * ap_init_queue(): Reset an AP queue.
* @qid: The AP queue number * @qid: The AP queue number
* *
* Reset an AP queue and wait for it to become available again. * Submit the Reset command to an AP queue.
* Since the reset is asynchron set the state to 'RESET_IN_PROGRESS'
* and check later via ap_poll_queue() if the reset is done.
*/ */
static int ap_init_queue(ap_qid_t qid) static int ap_init_queue(struct ap_device *ap_dev)
{ {
struct ap_queue_status status; struct ap_queue_status status;
int rc, dummy, i;
rc = -ENODEV; status = ap_reset_queue(ap_dev->qid);
status = ap_reset_queue(qid); switch (status.response_code) {
for (i = 0; i < AP_MAX_RESET; i++) { case AP_RESPONSE_NORMAL:
switch (status.response_code) { ap_dev->interrupt = AP_INTR_DISABLED;
case AP_RESPONSE_NORMAL: ap_dev->reset = AP_RESET_IN_PROGRESS;
if (status.queue_empty) return 0;
rc = 0; case AP_RESPONSE_RESET_IN_PROGRESS:
break; case AP_RESPONSE_BUSY:
case AP_RESPONSE_Q_NOT_AVAIL: return -EBUSY;
case AP_RESPONSE_DECONFIGURED: case AP_RESPONSE_Q_NOT_AVAIL:
case AP_RESPONSE_CHECKSTOPPED: case AP_RESPONSE_DECONFIGURED:
i = AP_MAX_RESET; /* return with -ENODEV */ case AP_RESPONSE_CHECKSTOPPED:
break; default:
case AP_RESPONSE_RESET_IN_PROGRESS: return -ENODEV;
rc = -EBUSY;
case AP_RESPONSE_BUSY:
default:
break;
}
if (rc != -ENODEV && rc != -EBUSY)
break;
if (i < AP_MAX_RESET - 1) {
/* Time we are waiting until we give up (0.7sec * 90).
* Since the actual request (in progress) will not
* interrupted immediately for the reset command,
* we have to be patient. In worst case we have to
* wait 60sec + reset time (some msec).
*/
schedule_timeout(AP_RESET_TIMEOUT);
status = ap_test_queue(qid, &dummy, &dummy);
}
}
if (rc == 0 && ap_using_interrupts()) {
rc = ap_queue_enable_interruption(qid, ap_airq.lsi_ptr);
/* If interruption mode is supported by the machine,
* but an AP can not be enabled for interruption then
* the AP will be discarded. */
if (rc)
pr_err("Registering adapter interrupts for "
"AP %d failed\n", AP_QID_DEVICE(qid));
} }
return rc;
} }
/** /**
...@@ -729,10 +683,63 @@ static ssize_t ap_pendingq_count_show(struct device *dev, ...@@ -729,10 +683,63 @@ static ssize_t ap_pendingq_count_show(struct device *dev,
static DEVICE_ATTR(pendingq_count, 0444, ap_pendingq_count_show, NULL); static DEVICE_ATTR(pendingq_count, 0444, ap_pendingq_count_show, NULL);
static ssize_t ap_reset_show(struct device *dev,
struct device_attribute *attr, char *buf)
{
struct ap_device *ap_dev = to_ap_dev(dev);
int rc = 0;
spin_lock_bh(&ap_dev->lock);
switch (ap_dev->reset) {
case AP_RESET_IGNORE:
rc = snprintf(buf, PAGE_SIZE, "No Reset Timer set.\n");
break;
case AP_RESET_ARMED:
rc = snprintf(buf, PAGE_SIZE, "Reset Timer armed.\n");
break;
case AP_RESET_DO:
rc = snprintf(buf, PAGE_SIZE, "Reset Timer expired.\n");
break;
case AP_RESET_IN_PROGRESS:
rc = snprintf(buf, PAGE_SIZE, "Reset in progress.\n");
break;
default:
break;
}
spin_unlock_bh(&ap_dev->lock);
return rc;
}
static DEVICE_ATTR(reset, 0444, ap_reset_show, NULL);
static ssize_t ap_interrupt_show(struct device *dev,
struct device_attribute *attr, char *buf)
{
struct ap_device *ap_dev = to_ap_dev(dev);
int rc = 0;
spin_lock_bh(&ap_dev->lock);
switch (ap_dev->interrupt) {
case AP_INTR_DISABLED:
rc = snprintf(buf, PAGE_SIZE, "Interrupts disabled.\n");
break;
case AP_INTR_ENABLED:
rc = snprintf(buf, PAGE_SIZE, "Interrupts enabled.\n");
break;
case AP_INTR_IN_PROGRESS:
rc = snprintf(buf, PAGE_SIZE, "Enable Interrupt pending.\n");
break;
}
spin_unlock_bh(&ap_dev->lock);
return rc;
}
static DEVICE_ATTR(interrupt, 0444, ap_interrupt_show, NULL);
static ssize_t ap_modalias_show(struct device *dev, static ssize_t ap_modalias_show(struct device *dev,
struct device_attribute *attr, char *buf) struct device_attribute *attr, char *buf)
{ {
return sprintf(buf, "ap:t%02X", to_ap_dev(dev)->device_type); return sprintf(buf, "ap:t%02X\n", to_ap_dev(dev)->device_type);
} }
static DEVICE_ATTR(modalias, 0444, ap_modalias_show, NULL); static DEVICE_ATTR(modalias, 0444, ap_modalias_show, NULL);
...@@ -753,6 +760,8 @@ static struct attribute *ap_dev_attrs[] = { ...@@ -753,6 +760,8 @@ static struct attribute *ap_dev_attrs[] = {
&dev_attr_request_count.attr, &dev_attr_request_count.attr,
&dev_attr_requestq_count.attr, &dev_attr_requestq_count.attr,
&dev_attr_pendingq_count.attr, &dev_attr_pendingq_count.attr,
&dev_attr_reset.attr,
&dev_attr_interrupt.attr,
&dev_attr_modalias.attr, &dev_attr_modalias.attr,
&dev_attr_ap_functions.attr, &dev_attr_ap_functions.attr,
NULL NULL
...@@ -926,6 +935,10 @@ static int ap_device_probe(struct device *dev) ...@@ -926,6 +935,10 @@ static int ap_device_probe(struct device *dev)
spin_lock_bh(&ap_device_list_lock); spin_lock_bh(&ap_device_list_lock);
list_del_init(&ap_dev->list); list_del_init(&ap_dev->list);
spin_unlock_bh(&ap_device_list_lock); spin_unlock_bh(&ap_device_list_lock);
} else {
if (ap_dev->reset == AP_RESET_IN_PROGRESS ||
ap_dev->interrupt == AP_INTR_IN_PROGRESS)
__ap_schedule_poll_timer();
} }
return rc; return rc;
} }
...@@ -1411,7 +1424,7 @@ static void ap_scan_bus(struct work_struct *unused) ...@@ -1411,7 +1424,7 @@ static void ap_scan_bus(struct work_struct *unused)
struct ap_device *ap_dev; struct ap_device *ap_dev;
struct device *dev; struct device *dev;
ap_qid_t qid; ap_qid_t qid;
int queue_depth, device_type; int queue_depth = 0, device_type = 0;
unsigned int device_functions; unsigned int device_functions;
int rc, i; int rc, i;
...@@ -1429,15 +1442,9 @@ static void ap_scan_bus(struct work_struct *unused) ...@@ -1429,15 +1442,9 @@ static void ap_scan_bus(struct work_struct *unused)
else else
rc = -ENODEV; rc = -ENODEV;
if (dev) { if (dev) {
if (rc == -EBUSY) {
set_current_state(TASK_UNINTERRUPTIBLE);
schedule_timeout(AP_RESET_TIMEOUT);
rc = ap_query_queue(qid, &queue_depth,
&device_type);
}
ap_dev = to_ap_dev(dev); ap_dev = to_ap_dev(dev);
spin_lock_bh(&ap_dev->lock); spin_lock_bh(&ap_dev->lock);
if (rc || ap_dev->unregistered) { if (rc == -ENODEV || ap_dev->unregistered) {
spin_unlock_bh(&ap_dev->lock); spin_unlock_bh(&ap_dev->lock);
if (ap_dev->unregistered) if (ap_dev->unregistered)
i--; i--;
...@@ -1449,15 +1456,17 @@ static void ap_scan_bus(struct work_struct *unused) ...@@ -1449,15 +1456,17 @@ static void ap_scan_bus(struct work_struct *unused)
put_device(dev); put_device(dev);
continue; continue;
} }
if (rc)
continue;
rc = ap_init_queue(qid);
if (rc) if (rc)
continue; continue;
ap_dev = kzalloc(sizeof(*ap_dev), GFP_KERNEL); ap_dev = kzalloc(sizeof(*ap_dev), GFP_KERNEL);
if (!ap_dev) if (!ap_dev)
break; break;
ap_dev->qid = qid; ap_dev->qid = qid;
rc = ap_init_queue(ap_dev);
if ((rc != 0) && (rc != -EBUSY)) {
kfree(ap_dev);
continue;
}
ap_dev->queue_depth = queue_depth; ap_dev->queue_depth = queue_depth;
ap_dev->unregistered = 1; ap_dev->unregistered = 1;
spin_lock_init(&ap_dev->lock); spin_lock_init(&ap_dev->lock);
...@@ -1519,36 +1528,6 @@ ap_config_timeout(unsigned long ptr) ...@@ -1519,36 +1528,6 @@ ap_config_timeout(unsigned long ptr)
add_timer(&ap_config_timer); add_timer(&ap_config_timer);
} }
/**
* __ap_schedule_poll_timer(): Schedule poll timer.
*
* Set up the timer to run the poll tasklet
*/
static inline void __ap_schedule_poll_timer(void)
{
ktime_t hr_time;
spin_lock_bh(&ap_poll_timer_lock);
if (!hrtimer_is_queued(&ap_poll_timer) && !ap_suspend_flag) {
hr_time = ktime_set(0, poll_timeout);
hrtimer_forward_now(&ap_poll_timer, hr_time);
hrtimer_restart(&ap_poll_timer);
}
spin_unlock_bh(&ap_poll_timer_lock);
}
/**
* ap_schedule_poll_timer(): Schedule poll timer.
*
* Set up the timer to run the poll tasklet
*/
static inline void ap_schedule_poll_timer(void)
{
if (ap_using_interrupts())
return;
__ap_schedule_poll_timer();
}
/** /**
* ap_poll_read(): Receive pending reply messages from an AP device. * ap_poll_read(): Receive pending reply messages from an AP device.
* @ap_dev: pointer to the AP device * @ap_dev: pointer to the AP device
...@@ -1568,6 +1547,7 @@ static int ap_poll_read(struct ap_device *ap_dev, unsigned long *flags) ...@@ -1568,6 +1547,7 @@ static int ap_poll_read(struct ap_device *ap_dev, unsigned long *flags)
ap_dev->reply->message, ap_dev->reply->length); ap_dev->reply->message, ap_dev->reply->length);
switch (status.response_code) { switch (status.response_code) {
case AP_RESPONSE_NORMAL: case AP_RESPONSE_NORMAL:
ap_dev->interrupt = status.int_enabled;
atomic_dec(&ap_poll_requests); atomic_dec(&ap_poll_requests);
ap_decrease_queue_count(ap_dev); ap_decrease_queue_count(ap_dev);
list_for_each_entry(ap_msg, &ap_dev->pendingq, list) { list_for_each_entry(ap_msg, &ap_dev->pendingq, list) {
...@@ -1582,6 +1562,7 @@ static int ap_poll_read(struct ap_device *ap_dev, unsigned long *flags) ...@@ -1582,6 +1562,7 @@ static int ap_poll_read(struct ap_device *ap_dev, unsigned long *flags)
*flags |= 1; *flags |= 1;
break; break;
case AP_RESPONSE_NO_PENDING_REPLY: case AP_RESPONSE_NO_PENDING_REPLY:
ap_dev->interrupt = status.int_enabled;
if (status.queue_empty) { if (status.queue_empty) {
/* The card shouldn't forget requests but who knows. */ /* The card shouldn't forget requests but who knows. */
atomic_sub(ap_dev->queue_count, &ap_poll_requests); atomic_sub(ap_dev->queue_count, &ap_poll_requests);
...@@ -1612,7 +1593,8 @@ static int ap_poll_write(struct ap_device *ap_dev, unsigned long *flags) ...@@ -1612,7 +1593,8 @@ static int ap_poll_write(struct ap_device *ap_dev, unsigned long *flags)
struct ap_message *ap_msg; struct ap_message *ap_msg;
if (ap_dev->requestq_count <= 0 || if (ap_dev->requestq_count <= 0 ||
ap_dev->queue_count >= ap_dev->queue_depth) (ap_dev->queue_count >= ap_dev->queue_depth) ||
(ap_dev->reset == AP_RESET_IN_PROGRESS))
return 0; return 0;
/* Start the next request on the queue. */ /* Start the next request on the queue. */
ap_msg = list_entry(ap_dev->requestq.next, struct ap_message, list); ap_msg = list_entry(ap_dev->requestq.next, struct ap_message, list);
...@@ -1646,6 +1628,8 @@ static int ap_poll_write(struct ap_device *ap_dev, unsigned long *flags) ...@@ -1646,6 +1628,8 @@ static int ap_poll_write(struct ap_device *ap_dev, unsigned long *flags)
/** /**
* ap_poll_queue(): Poll AP device for pending replies and send new messages. * ap_poll_queue(): Poll AP device for pending replies and send new messages.
* Check if the queue has a pending reset. In case it's done re-enable
* interrupts, otherwise reschedule the poll_timer for another attempt.
* @ap_dev: pointer to the bus device * @ap_dev: pointer to the bus device
* @flags: pointer to control flags, bit 2^0 is set if another poll is * @flags: pointer to control flags, bit 2^0 is set if another poll is
* required, bit 2^1 is set if the poll timer needs to get armed * required, bit 2^1 is set if the poll timer needs to get armed
...@@ -1656,7 +1640,51 @@ static int ap_poll_write(struct ap_device *ap_dev, unsigned long *flags) ...@@ -1656,7 +1640,51 @@ static int ap_poll_write(struct ap_device *ap_dev, unsigned long *flags)
*/ */
static inline int ap_poll_queue(struct ap_device *ap_dev, unsigned long *flags) static inline int ap_poll_queue(struct ap_device *ap_dev, unsigned long *flags)
{ {
int rc; int rc, depth, type;
struct ap_queue_status status;
if (ap_dev->reset == AP_RESET_IN_PROGRESS) {
status = ap_test_queue(ap_dev->qid, &depth, &type);
switch (status.response_code) {
case AP_RESPONSE_NORMAL:
ap_dev->reset = AP_RESET_IGNORE;
if (ap_using_interrupts()) {
rc = ap_queue_enable_interruption(
ap_dev, ap_airq.lsi_ptr);
if (!rc)
ap_dev->interrupt = AP_INTR_IN_PROGRESS;
else if (rc == -ENODEV) {
pr_err("Registering adapter interrupts for "
"AP %d failed\n", AP_QID_DEVICE(ap_dev->qid));
return rc;
}
}
/* fall through */
case AP_RESPONSE_BUSY:
case AP_RESPONSE_RESET_IN_PROGRESS:
*flags |= AP_POLL_AFTER_TIMEOUT;
break;
case AP_RESPONSE_Q_NOT_AVAIL:
case AP_RESPONSE_DECONFIGURED:
case AP_RESPONSE_CHECKSTOPPED:
return -ENODEV;
default:
break;
}
}
if ((ap_dev->reset != AP_RESET_IN_PROGRESS) &&
(ap_dev->interrupt == AP_INTR_IN_PROGRESS)) {
status = ap_test_queue(ap_dev->qid, &depth, &type);
if (ap_using_interrupts()) {
if (status.int_enabled == 1)
ap_dev->interrupt = AP_INTR_ENABLED;
else
*flags |= AP_POLL_AFTER_TIMEOUT;
} else
ap_dev->interrupt = AP_INTR_DISABLED;
}
rc = ap_poll_read(ap_dev, flags); rc = ap_poll_read(ap_dev, flags);
if (rc) if (rc)
...@@ -1676,7 +1704,8 @@ static int __ap_queue_message(struct ap_device *ap_dev, struct ap_message *ap_ms ...@@ -1676,7 +1704,8 @@ static int __ap_queue_message(struct ap_device *ap_dev, struct ap_message *ap_ms
struct ap_queue_status status; struct ap_queue_status status;
if (list_empty(&ap_dev->requestq) && if (list_empty(&ap_dev->requestq) &&
ap_dev->queue_count < ap_dev->queue_depth) { (ap_dev->queue_count < ap_dev->queue_depth) &&
(ap_dev->reset != AP_RESET_IN_PROGRESS)) {
status = __ap_send(ap_dev->qid, ap_msg->psmid, status = __ap_send(ap_dev->qid, ap_msg->psmid,
ap_msg->message, ap_msg->length, ap_msg->message, ap_msg->length,
ap_msg->special); ap_msg->special);
...@@ -1789,21 +1818,20 @@ static enum hrtimer_restart ap_poll_timeout(struct hrtimer *unused) ...@@ -1789,21 +1818,20 @@ static enum hrtimer_restart ap_poll_timeout(struct hrtimer *unused)
* Reset a not responding AP device and move all requests from the * Reset a not responding AP device and move all requests from the
* pending queue to the request queue. * pending queue to the request queue.
*/ */
static void ap_reset(struct ap_device *ap_dev) static void ap_reset(struct ap_device *ap_dev, unsigned long *flags)
{ {
int rc; int rc;
ap_dev->reset = AP_RESET_IGNORE;
atomic_sub(ap_dev->queue_count, &ap_poll_requests); atomic_sub(ap_dev->queue_count, &ap_poll_requests);
ap_dev->queue_count = 0; ap_dev->queue_count = 0;
list_splice_init(&ap_dev->pendingq, &ap_dev->requestq); list_splice_init(&ap_dev->pendingq, &ap_dev->requestq);
ap_dev->requestq_count += ap_dev->pendingq_count; ap_dev->requestq_count += ap_dev->pendingq_count;
ap_dev->pendingq_count = 0; ap_dev->pendingq_count = 0;
rc = ap_init_queue(ap_dev->qid); rc = ap_init_queue(ap_dev);
if (rc == -ENODEV) if (rc == -ENODEV)
ap_dev->unregistered = 1; ap_dev->unregistered = 1;
else else
__ap_schedule_poll_timer(); *flags |= AP_POLL_AFTER_TIMEOUT;
} }
static int __ap_poll_device(struct ap_device *ap_dev, unsigned long *flags) static int __ap_poll_device(struct ap_device *ap_dev, unsigned long *flags)
...@@ -1812,7 +1840,7 @@ static int __ap_poll_device(struct ap_device *ap_dev, unsigned long *flags) ...@@ -1812,7 +1840,7 @@ static int __ap_poll_device(struct ap_device *ap_dev, unsigned long *flags)
if (ap_poll_queue(ap_dev, flags)) if (ap_poll_queue(ap_dev, flags))
ap_dev->unregistered = 1; ap_dev->unregistered = 1;
if (ap_dev->reset == AP_RESET_DO) if (ap_dev->reset == AP_RESET_DO)
ap_reset(ap_dev); ap_reset(ap_dev, flags);
} }
return 0; return 0;
} }
...@@ -1845,9 +1873,9 @@ static void ap_poll_all(unsigned long dummy) ...@@ -1845,9 +1873,9 @@ static void ap_poll_all(unsigned long dummy)
spin_unlock(&ap_dev->lock); spin_unlock(&ap_dev->lock);
} }
spin_unlock(&ap_device_list_lock); spin_unlock(&ap_device_list_lock);
} while (flags & 1); } while (flags & AP_POLL_IMMEDIATELY);
if (flags & 2) if (flags & AP_POLL_AFTER_TIMEOUT)
ap_schedule_poll_timer(); __ap_schedule_poll_timer();
} }
/** /**
......
...@@ -32,11 +32,13 @@ ...@@ -32,11 +32,13 @@
#define AP_DEVICES 64 /* Number of AP devices. */ #define AP_DEVICES 64 /* Number of AP devices. */
#define AP_DOMAINS 256 /* Number of AP domains. */ #define AP_DOMAINS 256 /* Number of AP domains. */
#define AP_MAX_RESET 90 /* Maximum number of resets. */
#define AP_RESET_TIMEOUT (HZ*0.7) /* Time in ticks for reset timeouts. */ #define AP_RESET_TIMEOUT (HZ*0.7) /* Time in ticks for reset timeouts. */
#define AP_CONFIG_TIME 30 /* Time in seconds between AP bus rescans. */ #define AP_CONFIG_TIME 30 /* Time in seconds between AP bus rescans. */
#define AP_POLL_TIME 1 /* Time in ticks between receive polls. */ #define AP_POLL_TIME 1 /* Time in ticks between receive polls. */
#define AP_POLL_IMMEDIATELY 1 /* continue running poll tasklet */
#define AP_POLL_AFTER_TIMEOUT 2 /* run poll tasklet again after timout */
extern int ap_domain_index; extern int ap_domain_index;
/** /**
...@@ -135,6 +137,14 @@ static inline int ap_test_bit(unsigned int *ptr, unsigned int nr) ...@@ -135,6 +137,14 @@ static inline int ap_test_bit(unsigned int *ptr, unsigned int nr)
#define AP_RESET_IGNORE 0 /* request timeout will be ignored */ #define AP_RESET_IGNORE 0 /* request timeout will be ignored */
#define AP_RESET_ARMED 1 /* request timeout timer is active */ #define AP_RESET_ARMED 1 /* request timeout timer is active */
#define AP_RESET_DO 2 /* AP reset required */ #define AP_RESET_DO 2 /* AP reset required */
#define AP_RESET_IN_PROGRESS 3 /* AP reset in progress */
/*
* AP interrupt states
*/
#define AP_INTR_DISABLED 0 /* AP interrupt disabled */
#define AP_INTR_ENABLED 1 /* AP interrupt enabled */
#define AP_INTR_IN_PROGRESS 3 /* AP interrupt in progress */
struct ap_device; struct ap_device;
struct ap_message; struct ap_message;
...@@ -168,6 +178,7 @@ struct ap_device { ...@@ -168,6 +178,7 @@ struct ap_device {
struct timer_list timeout; /* Timer for request timeouts. */ struct timer_list timeout; /* Timer for request timeouts. */
int reset; /* Reset required after req. timeout. */ int reset; /* Reset required after req. timeout. */
int interrupt; /* indicate if interrupts are enabled */
int queue_count; /* # messages currently on AP queue. */ int queue_count; /* # messages currently on AP queue. */
struct list_head pendingq; /* List of message sent to AP queue. */ struct list_head pendingq; /* List of message sent to AP queue. */
......
...@@ -39,7 +39,7 @@ ...@@ -39,7 +39,7 @@
* But the maximum time limit managed by the stomper code is set to 60sec. * But the maximum time limit managed by the stomper code is set to 60sec.
* Hence we have to wait at least that time period. * Hence we have to wait at least that time period.
*/ */
#define CEX4_CLEANUP_TIME (61*HZ) #define CEX4_CLEANUP_TIME (900*HZ)
static struct ap_device_id zcrypt_cex4_ids[] = { static struct ap_device_id zcrypt_cex4_ids[] = {
{ AP_DEVICE(AP_DEVICE_TYPE_CEX4) }, { AP_DEVICE(AP_DEVICE_TYPE_CEX4) },
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册