b.liu | e958203 | 2025-04-17 19:18:16 +0800 | [diff] [blame^] | 1 | /* SPDX-License-Identifier: GPL-2.0 */ |
| 2 | /* |
| 3 | * S390 version |
| 4 | * Copyright IBM Corp. 1999 |
| 5 | * Author(s): Hartmut Penner (hp@de.ibm.com), |
| 6 | * Martin Schwidefsky (schwidefsky@de.ibm.com) |
| 7 | * |
| 8 | * Derived from "include/asm-i386/processor.h" |
| 9 | * Copyright (C) 1994, Linus Torvalds |
| 10 | */ |
| 11 | |
| 12 | #ifndef __ASM_S390_PROCESSOR_H |
| 13 | #define __ASM_S390_PROCESSOR_H |
| 14 | |
| 15 | #include <linux/bits.h> |
| 16 | |
| 17 | #define CIF_MCCK_PENDING 0 /* machine check handling is pending */ |
| 18 | #define CIF_ASCE_PRIMARY 1 /* primary asce needs fixup / uaccess */ |
| 19 | #define CIF_ASCE_SECONDARY 2 /* secondary asce needs fixup / uaccess */ |
| 20 | #define CIF_NOHZ_DELAY 3 /* delay HZ disable for a tick */ |
| 21 | #define CIF_FPU 4 /* restore FPU registers */ |
| 22 | #define CIF_IGNORE_IRQ 5 /* ignore interrupt (for udelay) */ |
| 23 | #define CIF_ENABLED_WAIT 6 /* in enabled wait state */ |
| 24 | #define CIF_MCCK_GUEST 7 /* machine check happening in guest */ |
| 25 | #define CIF_DEDICATED_CPU 8 /* this CPU is dedicated */ |
| 26 | |
| 27 | #define _CIF_MCCK_PENDING BIT(CIF_MCCK_PENDING) |
| 28 | #define _CIF_ASCE_PRIMARY BIT(CIF_ASCE_PRIMARY) |
| 29 | #define _CIF_ASCE_SECONDARY BIT(CIF_ASCE_SECONDARY) |
| 30 | #define _CIF_NOHZ_DELAY BIT(CIF_NOHZ_DELAY) |
| 31 | #define _CIF_FPU BIT(CIF_FPU) |
| 32 | #define _CIF_IGNORE_IRQ BIT(CIF_IGNORE_IRQ) |
| 33 | #define _CIF_ENABLED_WAIT BIT(CIF_ENABLED_WAIT) |
| 34 | #define _CIF_MCCK_GUEST BIT(CIF_MCCK_GUEST) |
| 35 | #define _CIF_DEDICATED_CPU BIT(CIF_DEDICATED_CPU) |
| 36 | |
| 37 | #ifndef __ASSEMBLY__ |
| 38 | |
| 39 | #include <linux/cpumask.h> |
| 40 | #include <linux/linkage.h> |
| 41 | #include <linux/irqflags.h> |
| 42 | #include <asm/cpu.h> |
| 43 | #include <asm/page.h> |
| 44 | #include <asm/ptrace.h> |
| 45 | #include <asm/setup.h> |
| 46 | #include <asm/runtime_instr.h> |
| 47 | #include <asm/fpu/types.h> |
| 48 | #include <asm/fpu/internal.h> |
| 49 | |
| 50 | static inline void set_cpu_flag(int flag) |
| 51 | { |
| 52 | S390_lowcore.cpu_flags |= (1UL << flag); |
| 53 | } |
| 54 | |
| 55 | static inline void clear_cpu_flag(int flag) |
| 56 | { |
| 57 | S390_lowcore.cpu_flags &= ~(1UL << flag); |
| 58 | } |
| 59 | |
| 60 | static inline int test_cpu_flag(int flag) |
| 61 | { |
| 62 | return !!(S390_lowcore.cpu_flags & (1UL << flag)); |
| 63 | } |
| 64 | |
| 65 | /* |
| 66 | * Test CIF flag of another CPU. The caller needs to ensure that |
| 67 | * CPU hotplug can not happen, e.g. by disabling preemption. |
| 68 | */ |
| 69 | static inline int test_cpu_flag_of(int flag, int cpu) |
| 70 | { |
| 71 | struct lowcore *lc = lowcore_ptr[cpu]; |
| 72 | return !!(lc->cpu_flags & (1UL << flag)); |
| 73 | } |
| 74 | |
| 75 | #define arch_needs_cpu() test_cpu_flag(CIF_NOHZ_DELAY) |
| 76 | |
| 77 | static inline void get_cpu_id(struct cpuid *ptr) |
| 78 | { |
| 79 | asm volatile("stidp %0" : "=Q" (*ptr)); |
| 80 | } |
| 81 | |
| 82 | void s390_adjust_jiffies(void); |
| 83 | void s390_update_cpu_mhz(void); |
| 84 | void cpu_detect_mhz_feature(void); |
| 85 | |
| 86 | extern const struct seq_operations cpuinfo_op; |
| 87 | extern int sysctl_ieee_emulation_warnings; |
| 88 | extern void execve_tail(void); |
| 89 | extern void __bpon(void); |
| 90 | |
| 91 | /* |
| 92 | * User space process size: 2GB for 31 bit, 4TB or 8PT for 64 bit. |
| 93 | */ |
| 94 | |
| 95 | #define TASK_SIZE_OF(tsk) (test_tsk_thread_flag(tsk, TIF_31BIT) ? \ |
| 96 | (1UL << 31) : -PAGE_SIZE) |
| 97 | #define TASK_UNMAPPED_BASE (test_thread_flag(TIF_31BIT) ? \ |
| 98 | (1UL << 30) : (1UL << 41)) |
| 99 | #define TASK_SIZE TASK_SIZE_OF(current) |
| 100 | #define TASK_SIZE_MAX (-PAGE_SIZE) |
| 101 | |
| 102 | #define STACK_TOP (test_thread_flag(TIF_31BIT) ? \ |
| 103 | (1UL << 31) : (1UL << 42)) |
| 104 | #define STACK_TOP_MAX (1UL << 42) |
| 105 | |
| 106 | #define HAVE_ARCH_PICK_MMAP_LAYOUT |
| 107 | |
| 108 | typedef unsigned int mm_segment_t; |
| 109 | |
| 110 | /* |
| 111 | * Thread structure |
| 112 | */ |
| 113 | struct thread_struct { |
| 114 | unsigned int acrs[NUM_ACRS]; |
| 115 | unsigned long ksp; /* kernel stack pointer */ |
| 116 | unsigned long user_timer; /* task cputime in user space */ |
| 117 | unsigned long guest_timer; /* task cputime in kvm guest */ |
| 118 | unsigned long system_timer; /* task cputime in kernel space */ |
| 119 | unsigned long hardirq_timer; /* task cputime in hardirq context */ |
| 120 | unsigned long softirq_timer; /* task cputime in softirq context */ |
| 121 | unsigned long sys_call_table; /* system call table address */ |
| 122 | mm_segment_t mm_segment; |
| 123 | unsigned long gmap_addr; /* address of last gmap fault. */ |
| 124 | unsigned int gmap_write_flag; /* gmap fault write indication */ |
| 125 | unsigned int gmap_int_code; /* int code of last gmap fault */ |
| 126 | unsigned int gmap_pfault; /* signal of a pending guest pfault */ |
| 127 | /* Per-thread information related to debugging */ |
| 128 | struct per_regs per_user; /* User specified PER registers */ |
| 129 | struct per_event per_event; /* Cause of the last PER trap */ |
| 130 | unsigned long per_flags; /* Flags to control debug behavior */ |
| 131 | unsigned int system_call; /* system call number in signal */ |
| 132 | unsigned long last_break; /* last breaking-event-address. */ |
| 133 | /* pfault_wait is used to block the process on a pfault event */ |
| 134 | unsigned long pfault_wait; |
| 135 | struct list_head list; |
| 136 | /* cpu runtime instrumentation */ |
| 137 | struct runtime_instr_cb *ri_cb; |
| 138 | struct gs_cb *gs_cb; /* Current guarded storage cb */ |
| 139 | struct gs_cb *gs_bc_cb; /* Broadcast guarded storage cb */ |
| 140 | unsigned char trap_tdb[256]; /* Transaction abort diagnose block */ |
| 141 | /* |
| 142 | * Warning: 'fpu' is dynamically-sized. It *MUST* be at |
| 143 | * the end. |
| 144 | */ |
| 145 | struct fpu fpu; /* FP and VX register save area */ |
| 146 | }; |
| 147 | |
| 148 | /* Flag to disable transactions. */ |
| 149 | #define PER_FLAG_NO_TE 1UL |
| 150 | /* Flag to enable random transaction aborts. */ |
| 151 | #define PER_FLAG_TE_ABORT_RAND 2UL |
| 152 | /* Flag to specify random transaction abort mode: |
| 153 | * - abort each transaction at a random instruction before TEND if set. |
| 154 | * - abort random transactions at a random instruction if cleared. |
| 155 | */ |
| 156 | #define PER_FLAG_TE_ABORT_RAND_TEND 4UL |
| 157 | |
| 158 | typedef struct thread_struct thread_struct; |
| 159 | |
| 160 | #define ARCH_MIN_TASKALIGN 8 |
| 161 | |
| 162 | #define INIT_THREAD { \ |
| 163 | .ksp = sizeof(init_stack) + (unsigned long) &init_stack, \ |
| 164 | .fpu.regs = (void *) init_task.thread.fpu.fprs, \ |
| 165 | .last_break = 1, \ |
| 166 | } |
| 167 | |
| 168 | /* |
| 169 | * Do necessary setup to start up a new thread. |
| 170 | */ |
| 171 | #define start_thread(regs, new_psw, new_stackp) do { \ |
| 172 | regs->psw.mask = PSW_USER_BITS | PSW_MASK_EA | PSW_MASK_BA; \ |
| 173 | regs->psw.addr = new_psw; \ |
| 174 | regs->gprs[15] = new_stackp; \ |
| 175 | execve_tail(); \ |
| 176 | } while (0) |
| 177 | |
| 178 | #define start_thread31(regs, new_psw, new_stackp) do { \ |
| 179 | regs->psw.mask = PSW_USER_BITS | PSW_MASK_BA; \ |
| 180 | regs->psw.addr = new_psw; \ |
| 181 | regs->gprs[15] = new_stackp; \ |
| 182 | crst_table_downgrade(current->mm); \ |
| 183 | execve_tail(); \ |
| 184 | } while (0) |
| 185 | |
| 186 | /* Forward declaration, a strange C thing */ |
| 187 | struct task_struct; |
| 188 | struct mm_struct; |
| 189 | struct seq_file; |
| 190 | struct pt_regs; |
| 191 | |
| 192 | void show_registers(struct pt_regs *regs); |
| 193 | void show_cacheinfo(struct seq_file *m); |
| 194 | |
| 195 | /* Free all resources held by a thread. */ |
| 196 | static inline void release_thread(struct task_struct *tsk) { } |
| 197 | |
| 198 | /* Free guarded storage control block */ |
| 199 | void guarded_storage_release(struct task_struct *tsk); |
| 200 | |
| 201 | unsigned long get_wchan(struct task_struct *p); |
| 202 | #define task_pt_regs(tsk) ((struct pt_regs *) \ |
| 203 | (task_stack_page(tsk) + THREAD_SIZE) - 1) |
| 204 | #define KSTK_EIP(tsk) (task_pt_regs(tsk)->psw.addr) |
| 205 | #define KSTK_ESP(tsk) (task_pt_regs(tsk)->gprs[15]) |
| 206 | |
| 207 | /* Has task runtime instrumentation enabled ? */ |
| 208 | #define is_ri_task(tsk) (!!(tsk)->thread.ri_cb) |
| 209 | |
| 210 | static inline unsigned long current_stack_pointer(void) |
| 211 | { |
| 212 | unsigned long sp; |
| 213 | |
| 214 | asm volatile("la %0,0(15)" : "=a" (sp)); |
| 215 | return sp; |
| 216 | } |
| 217 | |
| 218 | static __always_inline unsigned short stap(void) |
| 219 | { |
| 220 | unsigned short cpu_address; |
| 221 | |
| 222 | asm volatile("stap %0" : "=Q" (cpu_address)); |
| 223 | return cpu_address; |
| 224 | } |
| 225 | |
| 226 | #define cpu_relax() barrier() |
| 227 | |
| 228 | #define ECAG_CACHE_ATTRIBUTE 0 |
| 229 | #define ECAG_CPU_ATTRIBUTE 1 |
| 230 | |
| 231 | static inline unsigned long __ecag(unsigned int asi, unsigned char parm) |
| 232 | { |
| 233 | unsigned long val; |
| 234 | |
| 235 | asm volatile(".insn rsy,0xeb000000004c,%0,0,0(%1)" /* ecag */ |
| 236 | : "=d" (val) : "a" (asi << 8 | parm)); |
| 237 | return val; |
| 238 | } |
| 239 | |
| 240 | static inline void psw_set_key(unsigned int key) |
| 241 | { |
| 242 | asm volatile("spka 0(%0)" : : "d" (key)); |
| 243 | } |
| 244 | |
| 245 | /* |
| 246 | * Set PSW to specified value. |
| 247 | */ |
| 248 | static inline void __load_psw(psw_t psw) |
| 249 | { |
| 250 | asm volatile("lpswe %0" : : "Q" (psw) : "cc"); |
| 251 | } |
| 252 | |
| 253 | /* |
| 254 | * Set PSW mask to specified value, while leaving the |
| 255 | * PSW addr pointing to the next instruction. |
| 256 | */ |
| 257 | static __always_inline void __load_psw_mask(unsigned long mask) |
| 258 | { |
| 259 | psw_t psw __uninitialized; |
| 260 | unsigned long addr; |
| 261 | |
| 262 | psw.mask = mask; |
| 263 | |
| 264 | asm volatile( |
| 265 | " larl %0,1f\n" |
| 266 | " stg %0,%1\n" |
| 267 | " lpswe %2\n" |
| 268 | "1:" |
| 269 | : "=&d" (addr), "=Q" (psw.addr) : "Q" (psw) : "memory", "cc"); |
| 270 | } |
| 271 | |
| 272 | /* |
| 273 | * Extract current PSW mask |
| 274 | */ |
| 275 | static inline unsigned long __extract_psw(void) |
| 276 | { |
| 277 | unsigned int reg1, reg2; |
| 278 | |
| 279 | asm volatile("epsw %0,%1" : "=d" (reg1), "=a" (reg2)); |
| 280 | return (((unsigned long) reg1) << 32) | ((unsigned long) reg2); |
| 281 | } |
| 282 | |
| 283 | static inline void local_mcck_enable(void) |
| 284 | { |
| 285 | __load_psw_mask(__extract_psw() | PSW_MASK_MCHECK); |
| 286 | } |
| 287 | |
| 288 | static inline void local_mcck_disable(void) |
| 289 | { |
| 290 | __load_psw_mask(__extract_psw() & ~PSW_MASK_MCHECK); |
| 291 | } |
| 292 | |
| 293 | /* |
| 294 | * Rewind PSW instruction address by specified number of bytes. |
| 295 | */ |
| 296 | static inline unsigned long __rewind_psw(psw_t psw, unsigned long ilc) |
| 297 | { |
| 298 | unsigned long mask; |
| 299 | |
| 300 | mask = (psw.mask & PSW_MASK_EA) ? -1UL : |
| 301 | (psw.mask & PSW_MASK_BA) ? (1UL << 31) - 1 : |
| 302 | (1UL << 24) - 1; |
| 303 | return (psw.addr - ilc) & mask; |
| 304 | } |
| 305 | |
| 306 | /* |
| 307 | * Function to stop a processor until the next interrupt occurs |
| 308 | */ |
| 309 | void enabled_wait(void); |
| 310 | |
| 311 | /* |
| 312 | * Function to drop a processor into disabled wait state |
| 313 | */ |
| 314 | static inline void __noreturn disabled_wait(void) |
| 315 | { |
| 316 | psw_t psw; |
| 317 | |
| 318 | psw.mask = PSW_MASK_BASE | PSW_MASK_WAIT | PSW_MASK_BA | PSW_MASK_EA; |
| 319 | psw.addr = _THIS_IP_; |
| 320 | __load_psw(psw); |
| 321 | while (1); |
| 322 | } |
| 323 | |
| 324 | /* |
| 325 | * Basic Machine Check/Program Check Handler. |
| 326 | */ |
| 327 | |
| 328 | extern void s390_base_pgm_handler(void); |
| 329 | extern void s390_base_ext_handler(void); |
| 330 | |
| 331 | extern void (*s390_base_pgm_handler_fn)(void); |
| 332 | extern void (*s390_base_ext_handler_fn)(void); |
| 333 | |
| 334 | #define ARCH_LOW_ADDRESS_LIMIT 0x7fffffffUL |
| 335 | |
| 336 | extern int memcpy_real(void *, void *, size_t); |
| 337 | extern void memcpy_absolute(void *, void *, size_t); |
| 338 | |
| 339 | #define mem_assign_absolute(dest, val) do { \ |
| 340 | __typeof__(dest) __tmp = (val); \ |
| 341 | \ |
| 342 | BUILD_BUG_ON(sizeof(__tmp) != sizeof(val)); \ |
| 343 | memcpy_absolute(&(dest), &__tmp, sizeof(__tmp)); \ |
| 344 | } while (0) |
| 345 | |
| 346 | extern int s390_isolate_bp(void); |
| 347 | extern int s390_isolate_bp_guest(void); |
| 348 | |
| 349 | #endif /* __ASSEMBLY__ */ |
| 350 | |
| 351 | #endif /* __ASM_S390_PROCESSOR_H */ |