Thomas Gleixner | 2874c5f | 2019-05-27 08:55:01 +0200 | [diff] [blame] | 1 | /* SPDX-License-Identifier: GPL-2.0-or-later */ |
Benjamin Herrenschmidt | 243e251 | 2017-04-05 17:54:50 +1000 | [diff] [blame] | 2 | /* |
| 3 | * Copyright 2016,2017 IBM Corporation. |
Benjamin Herrenschmidt | 243e251 | 2017-04-05 17:54:50 +1000 | [diff] [blame] | 4 | */ |
| 5 | #ifndef _ASM_POWERPC_XIVE_REGS_H |
| 6 | #define _ASM_POWERPC_XIVE_REGS_H |
| 7 | |
| 8 | /* |
Benjamin Herrenschmidt | 12c1f33 | 2018-01-12 13:39:27 +1100 | [diff] [blame] | 9 | * "magic" Event State Buffer (ESB) MMIO offsets. |
| 10 | * |
| 11 | * Each interrupt source has a 2-bit state machine called ESB |
| 12 | * which can be controlled by MMIO. It's made of 2 bits, P and |
| 13 | * Q. P indicates that an interrupt is pending (has been sent |
| 14 | * to a queue and is waiting for an EOI). Q indicates that the |
| 15 | * interrupt has been triggered while pending. |
| 16 | * |
| 17 | * This acts as a coalescing mechanism in order to guarantee |
| 18 | * that a given interrupt only occurs at most once in a queue. |
| 19 | * |
| 20 | * When doing an EOI, the Q bit will indicate if the interrupt |
| 21 | * needs to be re-triggered. |
| 22 | * |
| 23 | * The following offsets into the ESB MMIO allow to read or |
| 24 | * manipulate the PQ bits. They must be used with an 8-bytes |
| 25 | * load instruction. They all return the previous state of the |
| 26 | * interrupt (atomically). |
| 27 | * |
| 28 | * Additionally, some ESB pages support doing an EOI via a |
| 29 | * store at 0 and some ESBs support doing a trigger via a |
| 30 | * separate trigger page. |
| 31 | */ |
| 32 | #define XIVE_ESB_STORE_EOI 0x400 /* Store */ |
| 33 | #define XIVE_ESB_LOAD_EOI 0x000 /* Load */ |
| 34 | #define XIVE_ESB_GET 0x800 /* Load */ |
| 35 | #define XIVE_ESB_SET_PQ_00 0xc00 /* Load */ |
| 36 | #define XIVE_ESB_SET_PQ_01 0xd00 /* Load */ |
| 37 | #define XIVE_ESB_SET_PQ_10 0xe00 /* Load */ |
| 38 | #define XIVE_ESB_SET_PQ_11 0xf00 /* Load */ |
| 39 | |
| 40 | #define XIVE_ESB_VAL_P 0x2 |
| 41 | #define XIVE_ESB_VAL_Q 0x1 |
| 42 | |
| 43 | /* |
Benjamin Herrenschmidt | 243e251 | 2017-04-05 17:54:50 +1000 | [diff] [blame] | 44 | * Thread Management (aka "TM") registers |
| 45 | */ |
| 46 | |
| 47 | /* TM register offsets */ |
| 48 | #define TM_QW0_USER 0x000 /* All rings */ |
| 49 | #define TM_QW1_OS 0x010 /* Ring 0..2 */ |
| 50 | #define TM_QW2_HV_POOL 0x020 /* Ring 0..1 */ |
| 51 | #define TM_QW3_HV_PHYS 0x030 /* Ring 0..1 */ |
| 52 | |
| 53 | /* Byte offsets inside a QW QW0 QW1 QW2 QW3 */ |
| 54 | #define TM_NSR 0x0 /* + + - + */ |
| 55 | #define TM_CPPR 0x1 /* - + - + */ |
| 56 | #define TM_IPB 0x2 /* - + + + */ |
| 57 | #define TM_LSMFB 0x3 /* - + + + */ |
| 58 | #define TM_ACK_CNT 0x4 /* - + - - */ |
| 59 | #define TM_INC 0x5 /* - + - + */ |
| 60 | #define TM_AGE 0x6 /* - + - + */ |
| 61 | #define TM_PIPR 0x7 /* - + - + */ |
| 62 | |
| 63 | #define TM_WORD0 0x0 |
| 64 | #define TM_WORD1 0x4 |
| 65 | |
| 66 | /* |
| 67 | * QW word 2 contains the valid bit at the top and other fields |
| 68 | * depending on the QW. |
| 69 | */ |
| 70 | #define TM_WORD2 0x8 |
| 71 | #define TM_QW0W2_VU PPC_BIT32(0) |
| 72 | #define TM_QW0W2_LOGIC_SERV PPC_BITMASK32(1,31) // XX 2,31 ? |
| 73 | #define TM_QW1W2_VO PPC_BIT32(0) |
| 74 | #define TM_QW1W2_OS_CAM PPC_BITMASK32(8,31) |
| 75 | #define TM_QW2W2_VP PPC_BIT32(0) |
| 76 | #define TM_QW2W2_POOL_CAM PPC_BITMASK32(8,31) |
| 77 | #define TM_QW3W2_VT PPC_BIT32(0) |
| 78 | #define TM_QW3W2_LP PPC_BIT32(6) |
| 79 | #define TM_QW3W2_LE PPC_BIT32(7) |
| 80 | #define TM_QW3W2_T PPC_BIT32(31) |
| 81 | |
| 82 | /* |
| 83 | * In addition to normal loads to "peek" and writes (only when invalid) |
| 84 | * using 4 and 8 bytes accesses, the above registers support these |
| 85 | * "special" byte operations: |
| 86 | * |
| 87 | * - Byte load from QW0[NSR] - User level NSR (EBB) |
| 88 | * - Byte store to QW0[NSR] - User level NSR (EBB) |
| 89 | * - Byte load/store to QW1[CPPR] and QW3[CPPR] - CPPR access |
| 90 | * - Byte load from QW3[TM_WORD2] - Read VT||00000||LP||LE on thrd 0 |
| 91 | * otherwise VT||0000000 |
| 92 | * - Byte store to QW3[TM_WORD2] - Set VT bit (and LP/LE if present) |
| 93 | * |
| 94 | * Then we have all these "special" CI ops at these offset that trigger |
| 95 | * all sorts of side effects: |
| 96 | */ |
| 97 | #define TM_SPC_ACK_EBB 0x800 /* Load8 ack EBB to reg*/ |
| 98 | #define TM_SPC_ACK_OS_REG 0x810 /* Load16 ack OS irq to reg */ |
| 99 | #define TM_SPC_PUSH_USR_CTX 0x808 /* Store32 Push/Validate user context */ |
| 100 | #define TM_SPC_PULL_USR_CTX 0x808 /* Load32 Pull/Invalidate user context */ |
| 101 | #define TM_SPC_SET_OS_PENDING 0x812 /* Store8 Set OS irq pending bit */ |
| 102 | #define TM_SPC_PULL_OS_CTX 0x818 /* Load32/Load64 Pull/Invalidate OS context to reg */ |
| 103 | #define TM_SPC_PULL_POOL_CTX 0x828 /* Load32/Load64 Pull/Invalidate Pool context to reg*/ |
| 104 | #define TM_SPC_ACK_HV_REG 0x830 /* Load16 ack HV irq to reg */ |
| 105 | #define TM_SPC_PULL_USR_CTX_OL 0xc08 /* Store8 Pull/Inval usr ctx to odd line */ |
| 106 | #define TM_SPC_ACK_OS_EL 0xc10 /* Store8 ack OS irq to even line */ |
| 107 | #define TM_SPC_ACK_HV_POOL_EL 0xc20 /* Store8 ack HV evt pool to even line */ |
| 108 | #define TM_SPC_ACK_HV_EL 0xc30 /* Store8 ack HV irq to even line */ |
| 109 | /* XXX more... */ |
| 110 | |
| 111 | /* NSR fields for the various QW ack types */ |
| 112 | #define TM_QW0_NSR_EB PPC_BIT8(0) |
| 113 | #define TM_QW1_NSR_EO PPC_BIT8(0) |
| 114 | #define TM_QW3_NSR_HE PPC_BITMASK8(0,1) |
| 115 | #define TM_QW3_NSR_HE_NONE 0 |
| 116 | #define TM_QW3_NSR_HE_POOL 1 |
| 117 | #define TM_QW3_NSR_HE_PHYS 2 |
| 118 | #define TM_QW3_NSR_HE_LSI 3 |
| 119 | #define TM_QW3_NSR_I PPC_BIT8(2) |
| 120 | #define TM_QW3_NSR_GRP_LVL PPC_BIT8(3,7) |
| 121 | |
Benjamin Herrenschmidt | 243e251 | 2017-04-05 17:54:50 +1000 | [diff] [blame] | 122 | #endif /* _ASM_POWERPC_XIVE_REGS_H */ |