 32cad1ffb8
			
		
	
	
		32cad1ffb8
		
	
	
	
	
		
			
			Headers in include/sysemu/ are not only related to system *emulation*, they are also used by virtualization. Rename as system/ which is clearer. Files renamed manually then mechanical change using sed tool. Signed-off-by: Philippe Mathieu-Daudé <philmd@linaro.org> Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Tested-by: Lei Yang <leiyang@redhat.com> Message-Id: <20241203172445.28576-1-philmd@linaro.org>
		
			
				
	
	
		
			1086 lines
		
	
	
		
			38 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1086 lines
		
	
	
		
			38 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator
 | |
|  *
 | |
|  * RTAS events handling
 | |
|  *
 | |
|  * Copyright (c) 2012 David Gibson, IBM Corporation.
 | |
|  *
 | |
|  * Permission is hereby granted, free of charge, to any person obtaining a copy
 | |
|  * of this software and associated documentation files (the "Software"), to deal
 | |
|  * in the Software without restriction, including without limitation the rights
 | |
|  * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
 | |
|  * copies of the Software, and to permit persons to whom the Software is
 | |
|  * furnished to do so, subject to the following conditions:
 | |
|  *
 | |
|  * The above copyright notice and this permission notice shall be included in
 | |
|  * all copies or substantial portions of the Software.
 | |
|  *
 | |
|  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 | |
|  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 | |
|  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
 | |
|  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 | |
|  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
 | |
|  * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
 | |
|  * THE SOFTWARE.
 | |
|  *
 | |
|  */
 | |
| 
 | |
| #include "qemu/osdep.h"
 | |
| #include "qapi/error.h"
 | |
| #include "system/device_tree.h"
 | |
| #include "system/runstate.h"
 | |
| 
 | |
| #include "hw/ppc/fdt.h"
 | |
| #include "hw/ppc/spapr.h"
 | |
| #include "hw/ppc/spapr_vio.h"
 | |
| #include "hw/pci/pci.h"
 | |
| #include "hw/irq.h"
 | |
| #include "hw/pci-host/spapr.h"
 | |
| #include "hw/ppc/spapr_drc.h"
 | |
| #include "qemu/help_option.h"
 | |
| #include "qemu/bcd.h"
 | |
| #include "qemu/main-loop.h"
 | |
| #include "hw/ppc/spapr_ovec.h"
 | |
| #include <libfdt.h>
 | |
| #include "migration/blocker.h"
 | |
| 
 | |
| #define RTAS_LOG_VERSION_MASK                   0xff000000
 | |
| #define   RTAS_LOG_VERSION_6                    0x06000000
 | |
| #define RTAS_LOG_SEVERITY_MASK                  0x00e00000
 | |
| #define   RTAS_LOG_SEVERITY_ALREADY_REPORTED    0x00c00000
 | |
| #define   RTAS_LOG_SEVERITY_FATAL               0x00a00000
 | |
| #define   RTAS_LOG_SEVERITY_ERROR               0x00800000
 | |
| #define   RTAS_LOG_SEVERITY_ERROR_SYNC          0x00600000
 | |
| #define   RTAS_LOG_SEVERITY_WARNING             0x00400000
 | |
| #define   RTAS_LOG_SEVERITY_EVENT               0x00200000
 | |
| #define   RTAS_LOG_SEVERITY_NO_ERROR            0x00000000
 | |
| #define RTAS_LOG_DISPOSITION_MASK               0x00180000
 | |
| #define   RTAS_LOG_DISPOSITION_FULLY_RECOVERED  0x00000000
 | |
| #define   RTAS_LOG_DISPOSITION_LIMITED_RECOVERY 0x00080000
 | |
| #define   RTAS_LOG_DISPOSITION_NOT_RECOVERED    0x00100000
 | |
| #define RTAS_LOG_OPTIONAL_PART_PRESENT          0x00040000
 | |
| #define RTAS_LOG_INITIATOR_MASK                 0x0000f000
 | |
| #define   RTAS_LOG_INITIATOR_UNKNOWN            0x00000000
 | |
| #define   RTAS_LOG_INITIATOR_CPU                0x00001000
 | |
| #define   RTAS_LOG_INITIATOR_PCI                0x00002000
 | |
| #define   RTAS_LOG_INITIATOR_MEMORY             0x00004000
 | |
| #define   RTAS_LOG_INITIATOR_HOTPLUG            0x00006000
 | |
| #define RTAS_LOG_TARGET_MASK                    0x00000f00
 | |
| #define   RTAS_LOG_TARGET_UNKNOWN               0x00000000
 | |
| #define   RTAS_LOG_TARGET_CPU                   0x00000100
 | |
| #define   RTAS_LOG_TARGET_PCI                   0x00000200
 | |
| #define   RTAS_LOG_TARGET_MEMORY                0x00000400
 | |
| #define   RTAS_LOG_TARGET_HOTPLUG               0x00000600
 | |
| #define RTAS_LOG_TYPE_MASK                      0x000000ff
 | |
| #define   RTAS_LOG_TYPE_OTHER                   0x00000000
 | |
| #define   RTAS_LOG_TYPE_RETRY                   0x00000001
 | |
| #define   RTAS_LOG_TYPE_TCE_ERR                 0x00000002
 | |
| #define   RTAS_LOG_TYPE_INTERN_DEV_FAIL         0x00000003
 | |
| #define   RTAS_LOG_TYPE_TIMEOUT                 0x00000004
 | |
| #define   RTAS_LOG_TYPE_DATA_PARITY             0x00000005
 | |
| #define   RTAS_LOG_TYPE_ADDR_PARITY             0x00000006
 | |
| #define   RTAS_LOG_TYPE_CACHE_PARITY            0x00000007
 | |
| #define   RTAS_LOG_TYPE_ADDR_INVALID            0x00000008
 | |
| #define   RTAS_LOG_TYPE_ECC_UNCORR              0x00000009
 | |
| #define   RTAS_LOG_TYPE_ECC_CORR                0x0000000a
 | |
| #define   RTAS_LOG_TYPE_EPOW                    0x00000040
 | |
| #define   RTAS_LOG_TYPE_HOTPLUG                 0x000000e5
 | |
| 
 | |
| struct rtas_error_log {
 | |
|     uint32_t summary;
 | |
|     uint32_t extended_length;
 | |
| } QEMU_PACKED;
 | |
| 
 | |
| struct rtas_event_log_v6 {
 | |
|     uint8_t b0;
 | |
| #define RTAS_LOG_V6_B0_VALID                          0x80
 | |
| #define RTAS_LOG_V6_B0_UNRECOVERABLE_ERROR            0x40
 | |
| #define RTAS_LOG_V6_B0_RECOVERABLE_ERROR              0x20
 | |
| #define RTAS_LOG_V6_B0_DEGRADED_OPERATION             0x10
 | |
| #define RTAS_LOG_V6_B0_PREDICTIVE_ERROR               0x08
 | |
| #define RTAS_LOG_V6_B0_NEW_LOG                        0x04
 | |
| #define RTAS_LOG_V6_B0_BIGENDIAN                      0x02
 | |
|     uint8_t _resv1;
 | |
|     uint8_t b2;
 | |
| #define RTAS_LOG_V6_B2_POWERPC_FORMAT                 0x80
 | |
| #define RTAS_LOG_V6_B2_LOG_FORMAT_MASK                0x0f
 | |
| #define   RTAS_LOG_V6_B2_LOG_FORMAT_PLATFORM_EVENT    0x0e
 | |
|     uint8_t _resv2[9];
 | |
|     uint32_t company;
 | |
| #define RTAS_LOG_V6_COMPANY_IBM                 0x49424d00 /* IBM<null> */
 | |
| } QEMU_PACKED;
 | |
| 
 | |
| struct rtas_event_log_v6_section_header {
 | |
|     uint16_t section_id;
 | |
|     uint16_t section_length;
 | |
|     uint8_t section_version;
 | |
|     uint8_t section_subtype;
 | |
|     uint16_t creator_component_id;
 | |
| } QEMU_PACKED;
 | |
| 
 | |
| struct rtas_event_log_v6_maina {
 | |
| #define RTAS_LOG_V6_SECTION_ID_MAINA                0x5048 /* PH */
 | |
|     struct rtas_event_log_v6_section_header hdr;
 | |
|     uint32_t creation_date; /* BCD: YYYYMMDD */
 | |
|     uint32_t creation_time; /* BCD: HHMMSS00 */
 | |
|     uint8_t _platform1[8];
 | |
|     char creator_id;
 | |
|     uint8_t _resv1[2];
 | |
|     uint8_t section_count;
 | |
|     uint8_t _resv2[4];
 | |
|     uint8_t _platform2[8];
 | |
|     uint32_t plid;
 | |
|     uint8_t _platform3[4];
 | |
| } QEMU_PACKED;
 | |
| 
 | |
| struct rtas_event_log_v6_mainb {
 | |
| #define RTAS_LOG_V6_SECTION_ID_MAINB                0x5548 /* UH */
 | |
|     struct rtas_event_log_v6_section_header hdr;
 | |
|     uint8_t subsystem_id;
 | |
|     uint8_t _platform1;
 | |
|     uint8_t event_severity;
 | |
|     uint8_t event_subtype;
 | |
|     uint8_t _platform2[4];
 | |
|     uint8_t _resv1[2];
 | |
|     uint16_t action_flags;
 | |
|     uint8_t _resv2[4];
 | |
| } QEMU_PACKED;
 | |
| 
 | |
| struct rtas_event_log_v6_epow {
 | |
| #define RTAS_LOG_V6_SECTION_ID_EPOW                 0x4550 /* EP */
 | |
|     struct rtas_event_log_v6_section_header hdr;
 | |
|     uint8_t sensor_value;
 | |
| #define RTAS_LOG_V6_EPOW_ACTION_RESET                    0
 | |
| #define RTAS_LOG_V6_EPOW_ACTION_WARN_COOLING             1
 | |
| #define RTAS_LOG_V6_EPOW_ACTION_WARN_POWER               2
 | |
| #define RTAS_LOG_V6_EPOW_ACTION_SYSTEM_SHUTDOWN          3
 | |
| #define RTAS_LOG_V6_EPOW_ACTION_SYSTEM_HALT              4
 | |
| #define RTAS_LOG_V6_EPOW_ACTION_MAIN_ENCLOSURE           5
 | |
| #define RTAS_LOG_V6_EPOW_ACTION_POWER_OFF                7
 | |
|     uint8_t event_modifier;
 | |
| #define RTAS_LOG_V6_EPOW_MODIFIER_NORMAL                 1
 | |
| #define RTAS_LOG_V6_EPOW_MODIFIER_ON_UPS                 2
 | |
| #define RTAS_LOG_V6_EPOW_MODIFIER_CRITICAL               3
 | |
| #define RTAS_LOG_V6_EPOW_MODIFIER_TEMPERATURE            4
 | |
|     uint8_t extended_modifier;
 | |
| #define RTAS_LOG_V6_EPOW_XMODIFIER_SYSTEM_WIDE           0
 | |
| #define RTAS_LOG_V6_EPOW_XMODIFIER_PARTITION_SPECIFIC    1
 | |
|     uint8_t _resv;
 | |
|     uint64_t reason_code;
 | |
| } QEMU_PACKED;
 | |
| 
 | |
| struct epow_extended_log {
 | |
|     struct rtas_event_log_v6 v6hdr;
 | |
|     struct rtas_event_log_v6_maina maina;
 | |
|     struct rtas_event_log_v6_mainb mainb;
 | |
|     struct rtas_event_log_v6_epow epow;
 | |
| } QEMU_PACKED;
 | |
| 
 | |
| union drc_identifier {
 | |
|     uint32_t index;
 | |
|     uint32_t count;
 | |
|     struct {
 | |
|         uint32_t count;
 | |
|         uint32_t index;
 | |
|     } count_indexed;
 | |
|     char name[1];
 | |
| } QEMU_PACKED;
 | |
| 
 | |
| struct rtas_event_log_v6_hp {
 | |
| #define RTAS_LOG_V6_SECTION_ID_HOTPLUG              0x4850 /* HP */
 | |
|     struct rtas_event_log_v6_section_header hdr;
 | |
|     uint8_t hotplug_type;
 | |
| #define RTAS_LOG_V6_HP_TYPE_CPU                          1
 | |
| #define RTAS_LOG_V6_HP_TYPE_MEMORY                       2
 | |
| #define RTAS_LOG_V6_HP_TYPE_SLOT                         3
 | |
| #define RTAS_LOG_V6_HP_TYPE_PHB                          4
 | |
| #define RTAS_LOG_V6_HP_TYPE_PCI                          5
 | |
| #define RTAS_LOG_V6_HP_TYPE_PMEM                         6
 | |
|     uint8_t hotplug_action;
 | |
| #define RTAS_LOG_V6_HP_ACTION_ADD                        1
 | |
| #define RTAS_LOG_V6_HP_ACTION_REMOVE                     2
 | |
|     uint8_t hotplug_identifier;
 | |
| #define RTAS_LOG_V6_HP_ID_DRC_NAME                       1
 | |
| #define RTAS_LOG_V6_HP_ID_DRC_INDEX                      2
 | |
| #define RTAS_LOG_V6_HP_ID_DRC_COUNT                      3
 | |
| #define RTAS_LOG_V6_HP_ID_DRC_COUNT_INDEXED              4
 | |
|     uint8_t reserved;
 | |
|     union drc_identifier drc_id;
 | |
| } QEMU_PACKED;
 | |
| 
 | |
| struct hp_extended_log {
 | |
|     struct rtas_event_log_v6 v6hdr;
 | |
|     struct rtas_event_log_v6_maina maina;
 | |
|     struct rtas_event_log_v6_mainb mainb;
 | |
|     struct rtas_event_log_v6_hp hp;
 | |
| } QEMU_PACKED;
 | |
| 
 | |
| struct rtas_event_log_v6_mc {
 | |
| #define RTAS_LOG_V6_SECTION_ID_MC                   0x4D43 /* MC */
 | |
|     struct rtas_event_log_v6_section_header hdr;
 | |
|     uint32_t fru_id;
 | |
|     uint32_t proc_id;
 | |
|     uint8_t error_type;
 | |
| #define RTAS_LOG_V6_MC_TYPE_UE                           0
 | |
| #define RTAS_LOG_V6_MC_TYPE_SLB                          1
 | |
| #define RTAS_LOG_V6_MC_TYPE_ERAT                         2
 | |
| #define RTAS_LOG_V6_MC_TYPE_TLB                          4
 | |
| #define RTAS_LOG_V6_MC_TYPE_D_CACHE                      5
 | |
| #define RTAS_LOG_V6_MC_TYPE_I_CACHE                      7
 | |
|     uint8_t sub_err_type;
 | |
| #define RTAS_LOG_V6_MC_UE_INDETERMINATE                  0
 | |
| #define RTAS_LOG_V6_MC_UE_IFETCH                         1
 | |
| #define RTAS_LOG_V6_MC_UE_PAGE_TABLE_WALK_IFETCH         2
 | |
| #define RTAS_LOG_V6_MC_UE_LOAD_STORE                     3
 | |
| #define RTAS_LOG_V6_MC_UE_PAGE_TABLE_WALK_LOAD_STORE     4
 | |
| #define RTAS_LOG_V6_MC_SLB_PARITY                        0
 | |
| #define RTAS_LOG_V6_MC_SLB_MULTIHIT                      1
 | |
| #define RTAS_LOG_V6_MC_SLB_INDETERMINATE                 2
 | |
| #define RTAS_LOG_V6_MC_ERAT_PARITY                       1
 | |
| #define RTAS_LOG_V6_MC_ERAT_MULTIHIT                     2
 | |
| #define RTAS_LOG_V6_MC_ERAT_INDETERMINATE                3
 | |
| #define RTAS_LOG_V6_MC_TLB_PARITY                        1
 | |
| #define RTAS_LOG_V6_MC_TLB_MULTIHIT                      2
 | |
| #define RTAS_LOG_V6_MC_TLB_INDETERMINATE                 3
 | |
| /*
 | |
|  * Per PAPR,
 | |
|  * For UE error type, set bit 1 of sub_err_type to indicate effective addr is
 | |
|  * provided. For other error types (SLB/ERAT/TLB), set bit 0 to indicate
 | |
|  * same.
 | |
|  */
 | |
| #define RTAS_LOG_V6_MC_UE_EA_ADDR_PROVIDED               0x40
 | |
| #define RTAS_LOG_V6_MC_EA_ADDR_PROVIDED                  0x80
 | |
|     uint8_t reserved_1[6];
 | |
|     uint64_t effective_address;
 | |
|     uint64_t logical_address;
 | |
| } QEMU_PACKED;
 | |
| 
 | |
| struct mc_extended_log {
 | |
|     struct rtas_event_log_v6 v6hdr;
 | |
|     struct rtas_event_log_v6_mc mc;
 | |
| } QEMU_PACKED;
 | |
| 
 | |
| struct MC_ierror_table {
 | |
|     unsigned long srr1_mask;
 | |
|     unsigned long srr1_value;
 | |
|     bool nip_valid; /* nip is a valid indicator of faulting address */
 | |
|     uint8_t error_type;
 | |
|     uint8_t error_subtype;
 | |
|     unsigned int initiator;
 | |
|     unsigned int severity;
 | |
| };
 | |
| 
 | |
| static const struct MC_ierror_table mc_ierror_table[] = {
 | |
| { 0x00000000081c0000, 0x0000000000040000, true,
 | |
|   RTAS_LOG_V6_MC_TYPE_UE, RTAS_LOG_V6_MC_UE_IFETCH,
 | |
|   RTAS_LOG_INITIATOR_CPU, RTAS_LOG_SEVERITY_ERROR_SYNC, },
 | |
| { 0x00000000081c0000, 0x0000000000080000, true,
 | |
|   RTAS_LOG_V6_MC_TYPE_SLB, RTAS_LOG_V6_MC_SLB_PARITY,
 | |
|   RTAS_LOG_INITIATOR_CPU, RTAS_LOG_SEVERITY_ERROR_SYNC, },
 | |
| { 0x00000000081c0000, 0x00000000000c0000, true,
 | |
|   RTAS_LOG_V6_MC_TYPE_SLB, RTAS_LOG_V6_MC_SLB_MULTIHIT,
 | |
|   RTAS_LOG_INITIATOR_CPU, RTAS_LOG_SEVERITY_ERROR_SYNC, },
 | |
| { 0x00000000081c0000, 0x0000000000100000, true,
 | |
|   RTAS_LOG_V6_MC_TYPE_ERAT, RTAS_LOG_V6_MC_ERAT_MULTIHIT,
 | |
|   RTAS_LOG_INITIATOR_CPU, RTAS_LOG_SEVERITY_ERROR_SYNC, },
 | |
| { 0x00000000081c0000, 0x0000000000140000, true,
 | |
|   RTAS_LOG_V6_MC_TYPE_TLB, RTAS_LOG_V6_MC_TLB_MULTIHIT,
 | |
|   RTAS_LOG_INITIATOR_CPU, RTAS_LOG_SEVERITY_ERROR_SYNC, },
 | |
| { 0x00000000081c0000, 0x0000000000180000, true,
 | |
|   RTAS_LOG_V6_MC_TYPE_UE, RTAS_LOG_V6_MC_UE_PAGE_TABLE_WALK_IFETCH,
 | |
|   RTAS_LOG_INITIATOR_CPU, RTAS_LOG_SEVERITY_ERROR_SYNC, } };
 | |
| 
 | |
| struct MC_derror_table {
 | |
|     unsigned long dsisr_value;
 | |
|     bool dar_valid; /* dar is a valid indicator of faulting address */
 | |
|     uint8_t error_type;
 | |
|     uint8_t error_subtype;
 | |
|     unsigned int initiator;
 | |
|     unsigned int severity;
 | |
| };
 | |
| 
 | |
| static const struct MC_derror_table mc_derror_table[] = {
 | |
| { 0x00008000, false,
 | |
|   RTAS_LOG_V6_MC_TYPE_UE, RTAS_LOG_V6_MC_UE_LOAD_STORE,
 | |
|   RTAS_LOG_INITIATOR_CPU, RTAS_LOG_SEVERITY_ERROR_SYNC, },
 | |
| { 0x00004000, true,
 | |
|   RTAS_LOG_V6_MC_TYPE_UE, RTAS_LOG_V6_MC_UE_PAGE_TABLE_WALK_LOAD_STORE,
 | |
|   RTAS_LOG_INITIATOR_CPU, RTAS_LOG_SEVERITY_ERROR_SYNC, },
 | |
| { 0x00000800, true,
 | |
|   RTAS_LOG_V6_MC_TYPE_ERAT, RTAS_LOG_V6_MC_ERAT_MULTIHIT,
 | |
|   RTAS_LOG_INITIATOR_CPU, RTAS_LOG_SEVERITY_ERROR_SYNC, },
 | |
| { 0x00000400, true,
 | |
|   RTAS_LOG_V6_MC_TYPE_TLB, RTAS_LOG_V6_MC_TLB_MULTIHIT,
 | |
|   RTAS_LOG_INITIATOR_CPU, RTAS_LOG_SEVERITY_ERROR_SYNC, },
 | |
| { 0x00000080, true,
 | |
|   RTAS_LOG_V6_MC_TYPE_SLB, RTAS_LOG_V6_MC_SLB_MULTIHIT,  /* Before PARITY */
 | |
|   RTAS_LOG_INITIATOR_CPU, RTAS_LOG_SEVERITY_ERROR_SYNC, },
 | |
| { 0x00000100, true,
 | |
|   RTAS_LOG_V6_MC_TYPE_SLB, RTAS_LOG_V6_MC_SLB_PARITY,
 | |
|   RTAS_LOG_INITIATOR_CPU, RTAS_LOG_SEVERITY_ERROR_SYNC, } };
 | |
| 
 | |
| #define SRR1_MC_LOADSTORE(srr1) ((srr1) & PPC_BIT(42))
 | |
| 
 | |
| typedef enum EventClass {
 | |
|     EVENT_CLASS_INTERNAL_ERRORS     = 0,
 | |
|     EVENT_CLASS_EPOW                = 1,
 | |
|     EVENT_CLASS_RESERVED            = 2,
 | |
|     EVENT_CLASS_HOT_PLUG            = 3,
 | |
|     EVENT_CLASS_IO                  = 4,
 | |
|     EVENT_CLASS_MAX
 | |
| } EventClassIndex;
 | |
| #define EVENT_CLASS_MASK(index) (1 << (31 - index))
 | |
| 
 | |
| static const char * const event_names[EVENT_CLASS_MAX] = {
 | |
|     [EVENT_CLASS_INTERNAL_ERRORS]       = "internal-errors",
 | |
|     [EVENT_CLASS_EPOW]                  = "epow-events",
 | |
|     [EVENT_CLASS_HOT_PLUG]              = "hot-plug-events",
 | |
|     [EVENT_CLASS_IO]                    = "ibm,io-events",
 | |
| };
 | |
| 
 | |
| struct SpaprEventSource {
 | |
|     int irq;
 | |
|     uint32_t mask;
 | |
|     bool enabled;
 | |
| };
 | |
| 
 | |
| static SpaprEventSource *spapr_event_sources_new(void)
 | |
| {
 | |
|     return g_new0(SpaprEventSource, EVENT_CLASS_MAX);
 | |
| }
 | |
| 
 | |
| static void spapr_event_sources_register(SpaprEventSource *event_sources,
 | |
|                                         EventClassIndex index, int irq)
 | |
| {
 | |
|     /* we only support 1 irq per event class at the moment */
 | |
|     g_assert(event_sources);
 | |
|     g_assert(!event_sources[index].enabled);
 | |
|     event_sources[index].irq = irq;
 | |
|     event_sources[index].mask = EVENT_CLASS_MASK(index);
 | |
|     event_sources[index].enabled = true;
 | |
| }
 | |
| 
 | |
| static const SpaprEventSource *
 | |
| spapr_event_sources_get_source(SpaprEventSource *event_sources,
 | |
|                                EventClassIndex index)
 | |
| {
 | |
|     g_assert(index < EVENT_CLASS_MAX);
 | |
|     g_assert(event_sources);
 | |
| 
 | |
|     return &event_sources[index];
 | |
| }
 | |
| 
 | |
| void spapr_dt_events(SpaprMachineState *spapr, void *fdt)
 | |
| {
 | |
|     uint32_t irq_ranges[EVENT_CLASS_MAX * 2];
 | |
|     int i, count = 0, event_sources;
 | |
|     SpaprEventSource *events = spapr->event_sources;
 | |
| 
 | |
|     g_assert(events);
 | |
| 
 | |
|     _FDT(event_sources = fdt_add_subnode(fdt, 0, "event-sources"));
 | |
| 
 | |
|     for (i = 0, count = 0; i < EVENT_CLASS_MAX; i++) {
 | |
|         int node_offset;
 | |
|         uint32_t interrupts[2];
 | |
|         const SpaprEventSource *source =
 | |
|             spapr_event_sources_get_source(events, i);
 | |
|         const char *source_name = event_names[i];
 | |
| 
 | |
|         if (!source->enabled) {
 | |
|             continue;
 | |
|         }
 | |
| 
 | |
|         spapr_dt_irq(interrupts, source->irq, false);
 | |
| 
 | |
|         _FDT(node_offset = fdt_add_subnode(fdt, event_sources, source_name));
 | |
|         _FDT(fdt_setprop(fdt, node_offset, "interrupts", interrupts,
 | |
|                          sizeof(interrupts)));
 | |
| 
 | |
|         irq_ranges[count++] = interrupts[0];
 | |
|         irq_ranges[count++] = cpu_to_be32(1);
 | |
|     }
 | |
| 
 | |
|     _FDT((fdt_setprop(fdt, event_sources, "interrupt-controller", NULL, 0)));
 | |
|     _FDT((fdt_setprop_cell(fdt, event_sources, "#interrupt-cells", 2)));
 | |
|     _FDT((fdt_setprop(fdt, event_sources, "interrupt-ranges",
 | |
|                       irq_ranges, count * sizeof(uint32_t))));
 | |
| }
 | |
| 
 | |
| static const SpaprEventSource *
 | |
| rtas_event_log_to_source(SpaprMachineState *spapr, int log_type)
 | |
| {
 | |
|     const SpaprEventSource *source;
 | |
| 
 | |
|     g_assert(spapr->event_sources);
 | |
| 
 | |
|     switch (log_type) {
 | |
|     case RTAS_LOG_TYPE_HOTPLUG:
 | |
|         source = spapr_event_sources_get_source(spapr->event_sources,
 | |
|                                                 EVENT_CLASS_HOT_PLUG);
 | |
|         if (spapr_ovec_test(spapr->ov5_cas, OV5_HP_EVT)) {
 | |
|             g_assert(source->enabled);
 | |
|             break;
 | |
|         }
 | |
|         /* fall through back to epow for legacy hotplug interrupt source */
 | |
|     case RTAS_LOG_TYPE_EPOW:
 | |
|         source = spapr_event_sources_get_source(spapr->event_sources,
 | |
|                                                 EVENT_CLASS_EPOW);
 | |
|         break;
 | |
|     default:
 | |
|         source = NULL;
 | |
|     }
 | |
| 
 | |
|     return source;
 | |
| }
 | |
| 
 | |
| static int rtas_event_log_to_irq(SpaprMachineState *spapr, int log_type)
 | |
| {
 | |
|     const SpaprEventSource *source;
 | |
| 
 | |
|     source = rtas_event_log_to_source(spapr, log_type);
 | |
|     g_assert(source);
 | |
|     g_assert(source->enabled);
 | |
| 
 | |
|     return source->irq;
 | |
| }
 | |
| 
 | |
| static uint32_t spapr_event_log_entry_type(SpaprEventLogEntry *entry)
 | |
| {
 | |
|     return entry->summary & RTAS_LOG_TYPE_MASK;
 | |
| }
 | |
| 
 | |
| static void rtas_event_log_queue(SpaprMachineState *spapr,
 | |
|                                  SpaprEventLogEntry *entry)
 | |
| {
 | |
|     QTAILQ_INSERT_TAIL(&spapr->pending_events, entry, next);
 | |
| }
 | |
| 
 | |
| static SpaprEventLogEntry *rtas_event_log_dequeue(SpaprMachineState *spapr,
 | |
|                                                   uint32_t event_mask)
 | |
| {
 | |
|     SpaprEventLogEntry *entry = NULL;
 | |
| 
 | |
|     QTAILQ_FOREACH(entry, &spapr->pending_events, next) {
 | |
|         const SpaprEventSource *source =
 | |
|             rtas_event_log_to_source(spapr,
 | |
|                                      spapr_event_log_entry_type(entry));
 | |
| 
 | |
|         g_assert(source);
 | |
|         if (source->mask & event_mask) {
 | |
|             break;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     if (entry) {
 | |
|         QTAILQ_REMOVE(&spapr->pending_events, entry, next);
 | |
|     }
 | |
| 
 | |
|     return entry;
 | |
| }
 | |
| 
 | |
| static bool rtas_event_log_contains(SpaprMachineState *spapr, uint32_t event_mask)
 | |
| {
 | |
|     SpaprEventLogEntry *entry = NULL;
 | |
| 
 | |
|     QTAILQ_FOREACH(entry, &spapr->pending_events, next) {
 | |
|         const SpaprEventSource *source =
 | |
|             rtas_event_log_to_source(spapr,
 | |
|                                      spapr_event_log_entry_type(entry));
 | |
| 
 | |
|         if (source->mask & event_mask) {
 | |
|             return true;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return false;
 | |
| }
 | |
| 
 | |
| static uint32_t next_plid;
 | |
| 
 | |
| static void spapr_init_v6hdr(struct rtas_event_log_v6 *v6hdr)
 | |
| {
 | |
|     v6hdr->b0 = RTAS_LOG_V6_B0_VALID | RTAS_LOG_V6_B0_NEW_LOG
 | |
|         | RTAS_LOG_V6_B0_BIGENDIAN;
 | |
|     v6hdr->b2 = RTAS_LOG_V6_B2_POWERPC_FORMAT
 | |
|         | RTAS_LOG_V6_B2_LOG_FORMAT_PLATFORM_EVENT;
 | |
|     v6hdr->company = cpu_to_be32(RTAS_LOG_V6_COMPANY_IBM);
 | |
| }
 | |
| 
 | |
| static void spapr_init_maina(SpaprMachineState *spapr,
 | |
|                              struct rtas_event_log_v6_maina *maina,
 | |
|                              int section_count)
 | |
| {
 | |
|     struct tm tm;
 | |
|     int year;
 | |
| 
 | |
|     maina->hdr.section_id = cpu_to_be16(RTAS_LOG_V6_SECTION_ID_MAINA);
 | |
|     maina->hdr.section_length = cpu_to_be16(sizeof(*maina));
 | |
|     /* FIXME: section version, subtype and creator id? */
 | |
|     spapr_rtc_read(&spapr->rtc, &tm, NULL);
 | |
|     year = tm.tm_year + 1900;
 | |
|     maina->creation_date = cpu_to_be32((to_bcd(year / 100) << 24)
 | |
|                                        | (to_bcd(year % 100) << 16)
 | |
|                                        | (to_bcd(tm.tm_mon + 1) << 8)
 | |
|                                        | to_bcd(tm.tm_mday));
 | |
|     maina->creation_time = cpu_to_be32((to_bcd(tm.tm_hour) << 24)
 | |
|                                        | (to_bcd(tm.tm_min) << 16)
 | |
|                                        | (to_bcd(tm.tm_sec) << 8));
 | |
|     maina->creator_id = 'H'; /* Hypervisor */
 | |
|     maina->section_count = section_count;
 | |
|     maina->plid = next_plid++;
 | |
| }
 | |
| 
 | |
| static void spapr_powerdown_req(Notifier *n, void *opaque)
 | |
| {
 | |
|     SpaprMachineState *spapr = SPAPR_MACHINE(qdev_get_machine());
 | |
|     SpaprEventLogEntry *entry;
 | |
|     struct rtas_event_log_v6 *v6hdr;
 | |
|     struct rtas_event_log_v6_maina *maina;
 | |
|     struct rtas_event_log_v6_mainb *mainb;
 | |
|     struct rtas_event_log_v6_epow *epow;
 | |
|     struct epow_extended_log *new_epow;
 | |
| 
 | |
|     entry = g_new(SpaprEventLogEntry, 1);
 | |
|     new_epow = g_malloc0(sizeof(*new_epow));
 | |
|     entry->extended_log = new_epow;
 | |
| 
 | |
|     v6hdr = &new_epow->v6hdr;
 | |
|     maina = &new_epow->maina;
 | |
|     mainb = &new_epow->mainb;
 | |
|     epow = &new_epow->epow;
 | |
| 
 | |
|     entry->summary = RTAS_LOG_VERSION_6
 | |
|                        | RTAS_LOG_SEVERITY_EVENT
 | |
|                        | RTAS_LOG_DISPOSITION_NOT_RECOVERED
 | |
|                        | RTAS_LOG_OPTIONAL_PART_PRESENT
 | |
|                        | RTAS_LOG_TYPE_EPOW;
 | |
|     entry->extended_length = sizeof(*new_epow);
 | |
| 
 | |
|     spapr_init_v6hdr(v6hdr);
 | |
|     spapr_init_maina(spapr, maina, 3 /* Main-A, Main-B and EPOW */);
 | |
| 
 | |
|     mainb->hdr.section_id = cpu_to_be16(RTAS_LOG_V6_SECTION_ID_MAINB);
 | |
|     mainb->hdr.section_length = cpu_to_be16(sizeof(*mainb));
 | |
|     /* FIXME: section version, subtype and creator id? */
 | |
|     mainb->subsystem_id = 0xa0; /* External environment */
 | |
|     mainb->event_severity = 0x00; /* Informational / non-error */
 | |
|     mainb->event_subtype = 0xd0; /* Normal shutdown */
 | |
| 
 | |
|     epow->hdr.section_id = cpu_to_be16(RTAS_LOG_V6_SECTION_ID_EPOW);
 | |
|     epow->hdr.section_length = cpu_to_be16(sizeof(*epow));
 | |
|     epow->hdr.section_version = 2; /* includes extended modifier */
 | |
|     /* FIXME: section subtype and creator id? */
 | |
|     epow->sensor_value = RTAS_LOG_V6_EPOW_ACTION_SYSTEM_SHUTDOWN;
 | |
|     epow->event_modifier = RTAS_LOG_V6_EPOW_MODIFIER_NORMAL;
 | |
|     epow->extended_modifier = RTAS_LOG_V6_EPOW_XMODIFIER_PARTITION_SPECIFIC;
 | |
| 
 | |
|     rtas_event_log_queue(spapr, entry);
 | |
| 
 | |
|     qemu_irq_pulse(spapr_qirq(spapr,
 | |
|                    rtas_event_log_to_irq(spapr, RTAS_LOG_TYPE_EPOW)));
 | |
| }
 | |
| 
 | |
| static void spapr_hotplug_req_event(uint8_t hp_id, uint8_t hp_action,
 | |
|                                     SpaprDrcType drc_type,
 | |
|                                     union drc_identifier *drc_id)
 | |
| {
 | |
|     SpaprMachineState *spapr = SPAPR_MACHINE(qdev_get_machine());
 | |
|     SpaprEventLogEntry *entry;
 | |
|     struct hp_extended_log *new_hp;
 | |
|     struct rtas_event_log_v6 *v6hdr;
 | |
|     struct rtas_event_log_v6_maina *maina;
 | |
|     struct rtas_event_log_v6_mainb *mainb;
 | |
|     struct rtas_event_log_v6_hp *hp;
 | |
| 
 | |
|     entry = g_new(SpaprEventLogEntry, 1);
 | |
|     new_hp = g_new0(struct hp_extended_log, 1);
 | |
|     entry->extended_log = new_hp;
 | |
| 
 | |
|     v6hdr = &new_hp->v6hdr;
 | |
|     maina = &new_hp->maina;
 | |
|     mainb = &new_hp->mainb;
 | |
|     hp = &new_hp->hp;
 | |
| 
 | |
|     entry->summary = RTAS_LOG_VERSION_6
 | |
|         | RTAS_LOG_SEVERITY_EVENT
 | |
|         | RTAS_LOG_DISPOSITION_NOT_RECOVERED
 | |
|         | RTAS_LOG_OPTIONAL_PART_PRESENT
 | |
|         | RTAS_LOG_INITIATOR_HOTPLUG
 | |
|         | RTAS_LOG_TYPE_HOTPLUG;
 | |
|     entry->extended_length = sizeof(*new_hp);
 | |
| 
 | |
|     spapr_init_v6hdr(v6hdr);
 | |
|     spapr_init_maina(spapr, maina, 3 /* Main-A, Main-B, HP */);
 | |
| 
 | |
|     mainb->hdr.section_id = cpu_to_be16(RTAS_LOG_V6_SECTION_ID_MAINB);
 | |
|     mainb->hdr.section_length = cpu_to_be16(sizeof(*mainb));
 | |
|     mainb->subsystem_id = 0x80; /* External environment */
 | |
|     mainb->event_severity = 0x00; /* Informational / non-error */
 | |
|     mainb->event_subtype = 0x00; /* Normal shutdown */
 | |
| 
 | |
|     hp->hdr.section_id = cpu_to_be16(RTAS_LOG_V6_SECTION_ID_HOTPLUG);
 | |
|     hp->hdr.section_length = cpu_to_be16(sizeof(*hp));
 | |
|     hp->hdr.section_version = 1; /* includes extended modifier */
 | |
|     hp->hotplug_action = hp_action;
 | |
|     hp->hotplug_identifier = hp_id;
 | |
| 
 | |
|     switch (drc_type) {
 | |
|     case SPAPR_DR_CONNECTOR_TYPE_PCI:
 | |
|         hp->hotplug_type = RTAS_LOG_V6_HP_TYPE_PCI;
 | |
|         break;
 | |
|     case SPAPR_DR_CONNECTOR_TYPE_LMB:
 | |
|         hp->hotplug_type = RTAS_LOG_V6_HP_TYPE_MEMORY;
 | |
|         break;
 | |
|     case SPAPR_DR_CONNECTOR_TYPE_CPU:
 | |
|         hp->hotplug_type = RTAS_LOG_V6_HP_TYPE_CPU;
 | |
|         break;
 | |
|     case SPAPR_DR_CONNECTOR_TYPE_PHB:
 | |
|         hp->hotplug_type = RTAS_LOG_V6_HP_TYPE_PHB;
 | |
|         break;
 | |
|     case SPAPR_DR_CONNECTOR_TYPE_PMEM:
 | |
|         hp->hotplug_type = RTAS_LOG_V6_HP_TYPE_PMEM;
 | |
|         break;
 | |
|     default:
 | |
|         /* we shouldn't be signaling hotplug events for resources
 | |
|          * that don't support them
 | |
|          */
 | |
|         g_assert_not_reached();
 | |
|     }
 | |
| 
 | |
|     if (hp_id == RTAS_LOG_V6_HP_ID_DRC_COUNT) {
 | |
|         hp->drc_id.count = cpu_to_be32(drc_id->count);
 | |
|     } else if (hp_id == RTAS_LOG_V6_HP_ID_DRC_INDEX) {
 | |
|         hp->drc_id.index = cpu_to_be32(drc_id->index);
 | |
|     } else if (hp_id == RTAS_LOG_V6_HP_ID_DRC_COUNT_INDEXED) {
 | |
|         /* we should not be using count_indexed value unless the guest
 | |
|          * supports dedicated hotplug event source
 | |
|          */
 | |
|         g_assert(spapr_memory_hot_unplug_supported(spapr));
 | |
|         hp->drc_id.count_indexed.count =
 | |
|             cpu_to_be32(drc_id->count_indexed.count);
 | |
|         hp->drc_id.count_indexed.index =
 | |
|             cpu_to_be32(drc_id->count_indexed.index);
 | |
|     }
 | |
| 
 | |
|     rtas_event_log_queue(spapr, entry);
 | |
| 
 | |
|     qemu_irq_pulse(spapr_qirq(spapr,
 | |
|                    rtas_event_log_to_irq(spapr, RTAS_LOG_TYPE_HOTPLUG)));
 | |
| }
 | |
| 
 | |
| void spapr_hotplug_req_add_by_index(SpaprDrc *drc)
 | |
| {
 | |
|     SpaprDrcType drc_type = spapr_drc_type(drc);
 | |
|     union drc_identifier drc_id;
 | |
| 
 | |
|     drc_id.index = spapr_drc_index(drc);
 | |
|     spapr_hotplug_req_event(RTAS_LOG_V6_HP_ID_DRC_INDEX,
 | |
|                             RTAS_LOG_V6_HP_ACTION_ADD, drc_type, &drc_id);
 | |
| }
 | |
| 
 | |
| void spapr_hotplug_req_remove_by_index(SpaprDrc *drc)
 | |
| {
 | |
|     SpaprDrcType drc_type = spapr_drc_type(drc);
 | |
|     union drc_identifier drc_id;
 | |
| 
 | |
|     drc_id.index = spapr_drc_index(drc);
 | |
|     spapr_hotplug_req_event(RTAS_LOG_V6_HP_ID_DRC_INDEX,
 | |
|                             RTAS_LOG_V6_HP_ACTION_REMOVE, drc_type, &drc_id);
 | |
| }
 | |
| 
 | |
| void spapr_hotplug_req_add_by_count(SpaprDrcType drc_type,
 | |
|                                        uint32_t count)
 | |
| {
 | |
|     union drc_identifier drc_id;
 | |
| 
 | |
|     drc_id.count = count;
 | |
|     spapr_hotplug_req_event(RTAS_LOG_V6_HP_ID_DRC_COUNT,
 | |
|                             RTAS_LOG_V6_HP_ACTION_ADD, drc_type, &drc_id);
 | |
| }
 | |
| 
 | |
| void spapr_hotplug_req_remove_by_count(SpaprDrcType drc_type,
 | |
|                                           uint32_t count)
 | |
| {
 | |
|     union drc_identifier drc_id;
 | |
| 
 | |
|     drc_id.count = count;
 | |
|     spapr_hotplug_req_event(RTAS_LOG_V6_HP_ID_DRC_COUNT,
 | |
|                             RTAS_LOG_V6_HP_ACTION_REMOVE, drc_type, &drc_id);
 | |
| }
 | |
| 
 | |
| void spapr_hotplug_req_add_by_count_indexed(SpaprDrcType drc_type,
 | |
|                                             uint32_t count, uint32_t index)
 | |
| {
 | |
|     union drc_identifier drc_id;
 | |
| 
 | |
|     drc_id.count_indexed.count = count;
 | |
|     drc_id.count_indexed.index = index;
 | |
|     spapr_hotplug_req_event(RTAS_LOG_V6_HP_ID_DRC_COUNT_INDEXED,
 | |
|                             RTAS_LOG_V6_HP_ACTION_ADD, drc_type, &drc_id);
 | |
| }
 | |
| 
 | |
| void spapr_hotplug_req_remove_by_count_indexed(SpaprDrcType drc_type,
 | |
|                                                uint32_t count, uint32_t index)
 | |
| {
 | |
|     union drc_identifier drc_id;
 | |
| 
 | |
|     drc_id.count_indexed.count = count;
 | |
|     drc_id.count_indexed.index = index;
 | |
|     spapr_hotplug_req_event(RTAS_LOG_V6_HP_ID_DRC_COUNT_INDEXED,
 | |
|                             RTAS_LOG_V6_HP_ACTION_REMOVE, drc_type, &drc_id);
 | |
| }
 | |
| 
 | |
| static void spapr_mc_set_ea_provided_flag(struct mc_extended_log *ext_elog)
 | |
| {
 | |
|     switch (ext_elog->mc.error_type) {
 | |
|     case RTAS_LOG_V6_MC_TYPE_UE:
 | |
|         ext_elog->mc.sub_err_type |= RTAS_LOG_V6_MC_UE_EA_ADDR_PROVIDED;
 | |
|         break;
 | |
|     case RTAS_LOG_V6_MC_TYPE_SLB:
 | |
|     case RTAS_LOG_V6_MC_TYPE_ERAT:
 | |
|     case RTAS_LOG_V6_MC_TYPE_TLB:
 | |
|         ext_elog->mc.sub_err_type |= RTAS_LOG_V6_MC_EA_ADDR_PROVIDED;
 | |
|         break;
 | |
|     default:
 | |
|         break;
 | |
|     }
 | |
| }
 | |
| 
 | |
| static uint32_t spapr_mce_get_elog_type(PowerPCCPU *cpu, bool recovered,
 | |
|                                         struct mc_extended_log *ext_elog)
 | |
| {
 | |
|     int i;
 | |
|     CPUPPCState *env = &cpu->env;
 | |
|     uint32_t summary;
 | |
|     uint64_t dsisr = env->spr[SPR_DSISR];
 | |
| 
 | |
|     summary = RTAS_LOG_VERSION_6 | RTAS_LOG_OPTIONAL_PART_PRESENT;
 | |
|     if (recovered) {
 | |
|         summary |= RTAS_LOG_DISPOSITION_FULLY_RECOVERED;
 | |
|     } else {
 | |
|         summary |= RTAS_LOG_DISPOSITION_NOT_RECOVERED;
 | |
|     }
 | |
| 
 | |
|     if (SRR1_MC_LOADSTORE(env->spr[SPR_SRR1])) {
 | |
|         for (i = 0; i < ARRAY_SIZE(mc_derror_table); i++) {
 | |
|             if (!(dsisr & mc_derror_table[i].dsisr_value)) {
 | |
|                 continue;
 | |
|             }
 | |
| 
 | |
|             ext_elog->mc.error_type = mc_derror_table[i].error_type;
 | |
|             ext_elog->mc.sub_err_type = mc_derror_table[i].error_subtype;
 | |
|             if (mc_derror_table[i].dar_valid) {
 | |
|                 ext_elog->mc.effective_address = cpu_to_be64(env->spr[SPR_DAR]);
 | |
|                 spapr_mc_set_ea_provided_flag(ext_elog);
 | |
|             }
 | |
| 
 | |
|             summary |= mc_derror_table[i].initiator
 | |
|                         | mc_derror_table[i].severity;
 | |
| 
 | |
|             return summary;
 | |
|         }
 | |
|     } else {
 | |
|         for (i = 0; i < ARRAY_SIZE(mc_ierror_table); i++) {
 | |
|             if ((env->spr[SPR_SRR1] & mc_ierror_table[i].srr1_mask) !=
 | |
|                     mc_ierror_table[i].srr1_value) {
 | |
|                 continue;
 | |
|             }
 | |
| 
 | |
|             ext_elog->mc.error_type = mc_ierror_table[i].error_type;
 | |
|             ext_elog->mc.sub_err_type = mc_ierror_table[i].error_subtype;
 | |
|             if (mc_ierror_table[i].nip_valid) {
 | |
|                 ext_elog->mc.effective_address = cpu_to_be64(env->nip);
 | |
|                 spapr_mc_set_ea_provided_flag(ext_elog);
 | |
|             }
 | |
| 
 | |
|             summary |= mc_ierror_table[i].initiator
 | |
|                         | mc_ierror_table[i].severity;
 | |
| 
 | |
|             return summary;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     summary |= RTAS_LOG_INITIATOR_CPU;
 | |
|     return summary;
 | |
| }
 | |
| 
 | |
| static void spapr_mce_dispatch_elog(SpaprMachineState *spapr, PowerPCCPU *cpu,
 | |
|                                     bool recovered)
 | |
| {
 | |
|     CPUState *cs = CPU(cpu);
 | |
|     CPUPPCState *env = &cpu->env;
 | |
|     uint64_t rtas_addr;
 | |
|     struct rtas_error_log log;
 | |
|     struct mc_extended_log *ext_elog;
 | |
|     uint32_t summary;
 | |
| 
 | |
|     ext_elog = g_malloc0(sizeof(*ext_elog));
 | |
|     summary = spapr_mce_get_elog_type(cpu, recovered, ext_elog);
 | |
| 
 | |
|     log.summary = cpu_to_be32(summary);
 | |
|     log.extended_length = cpu_to_be32(sizeof(*ext_elog));
 | |
| 
 | |
|     spapr_init_v6hdr(&ext_elog->v6hdr);
 | |
|     ext_elog->mc.hdr.section_id = cpu_to_be16(RTAS_LOG_V6_SECTION_ID_MC);
 | |
|     ext_elog->mc.hdr.section_length =
 | |
|                     cpu_to_be16(sizeof(struct rtas_event_log_v6_mc));
 | |
|     ext_elog->mc.hdr.section_version = 1;
 | |
| 
 | |
|     /* get rtas addr from fdt */
 | |
|     rtas_addr = spapr_get_rtas_addr();
 | |
|     if (!rtas_addr) {
 | |
|         if (!recovered) {
 | |
|             error_report(
 | |
| "FWNMI: Unable to deliver machine check to guest: rtas_addr not found.");
 | |
|             qemu_system_guest_panicked(NULL);
 | |
|         } else {
 | |
|             warn_report(
 | |
| "FWNMI: Unable to deliver machine check to guest: rtas_addr not found. "
 | |
| "Machine check recovered.");
 | |
|         }
 | |
|         g_free(ext_elog);
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     /*
 | |
|      * By taking the interlock, we assume that the MCE will be
 | |
|      * delivered to the guest. CAUTION: don't add anything that could
 | |
|      * prevent the MCE to be delivered after this line, otherwise the
 | |
|      * guest won't be able to release the interlock and ultimately
 | |
|      * hang/crash?
 | |
|      */
 | |
|     spapr->fwnmi_machine_check_interlock = cpu->vcpu_id;
 | |
| 
 | |
|     stq_be_phys(&address_space_memory, rtas_addr + RTAS_ERROR_LOG_OFFSET,
 | |
|                 env->gpr[3]);
 | |
|     cpu_physical_memory_write(rtas_addr + RTAS_ERROR_LOG_OFFSET +
 | |
|                               sizeof(env->gpr[3]), &log, sizeof(log));
 | |
|     cpu_physical_memory_write(rtas_addr + RTAS_ERROR_LOG_OFFSET +
 | |
|                               sizeof(env->gpr[3]) + sizeof(log), ext_elog,
 | |
|                               sizeof(*ext_elog));
 | |
|     g_free(ext_elog);
 | |
| 
 | |
|     env->gpr[3] = rtas_addr + RTAS_ERROR_LOG_OFFSET;
 | |
| 
 | |
|     ppc_cpu_do_fwnmi_machine_check(cs, spapr->fwnmi_machine_check_addr);
 | |
| }
 | |
| 
 | |
| void spapr_mce_req_event(PowerPCCPU *cpu, bool recovered)
 | |
| {
 | |
|     SpaprMachineState *spapr = SPAPR_MACHINE(qdev_get_machine());
 | |
|     CPUState *cs = CPU(cpu);
 | |
|     int ret;
 | |
| 
 | |
|     if (spapr->fwnmi_machine_check_addr == -1) {
 | |
|         /* Non-FWNMI case, deliver it like an architected CPU interrupt. */
 | |
|         cs->exception_index = POWERPC_EXCP_MCHECK;
 | |
|         ppc_cpu_do_interrupt(cs);
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     /* Wait for FWNMI interlock. */
 | |
|     while (spapr->fwnmi_machine_check_interlock != -1) {
 | |
|         /*
 | |
|          * Check whether the same CPU got machine check error
 | |
|          * while still handling the mc error (i.e., before
 | |
|          * that CPU called "ibm,nmi-interlock")
 | |
|          */
 | |
|         if (spapr->fwnmi_machine_check_interlock == cpu->vcpu_id) {
 | |
|             if (!recovered) {
 | |
|                 error_report(
 | |
| "FWNMI: Unable to deliver machine check to guest: nested machine check.");
 | |
|                 qemu_system_guest_panicked(NULL);
 | |
|             } else {
 | |
|                 warn_report(
 | |
| "FWNMI: Unable to deliver machine check to guest: nested machine check. "
 | |
| "Machine check recovered.");
 | |
|             }
 | |
|             return;
 | |
|         }
 | |
|         qemu_cond_wait_bql(&spapr->fwnmi_machine_check_interlock_cond);
 | |
|         if (spapr->fwnmi_machine_check_addr == -1) {
 | |
|             /*
 | |
|              * If the machine was reset while waiting for the interlock,
 | |
|              * abort the delivery. The machine check applies to a context
 | |
|              * that no longer exists, so it wouldn't make sense to deliver
 | |
|              * it now.
 | |
|              */
 | |
|             return;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     /*
 | |
|      * Try to block migration while FWNMI is being handled, so the
 | |
|      * machine check handler runs where the information passed to it
 | |
|      * actually makes sense.  This shouldn't actually block migration,
 | |
|      * only delay it slightly, assuming migration is retried.  If the
 | |
|      * attempt to block fails, carry on.  Unfortunately, it always
 | |
|      * fails when running with -only-migrate.  A proper interface to
 | |
|      * delay migration completion for a bit could avoid that.
 | |
|      */
 | |
|     error_setg(&spapr->fwnmi_migration_blocker,
 | |
|         "A machine check is being handled during migration. The handler"
 | |
|         "may run and log hardware error on the destination");
 | |
| 
 | |
|     ret = migrate_add_blocker(&spapr->fwnmi_migration_blocker, NULL);
 | |
|     if (ret == -EBUSY) {
 | |
|         warn_report("Received a fwnmi while migration was in progress");
 | |
|     }
 | |
| 
 | |
|     spapr_mce_dispatch_elog(spapr, cpu, recovered);
 | |
| }
 | |
| 
 | |
| static void check_exception(PowerPCCPU *cpu, SpaprMachineState *spapr,
 | |
|                             uint32_t token, uint32_t nargs,
 | |
|                             target_ulong args,
 | |
|                             uint32_t nret, target_ulong rets)
 | |
| {
 | |
|     uint32_t mask, buf, len, event_len;
 | |
|     SpaprEventLogEntry *event;
 | |
|     struct rtas_error_log header;
 | |
|     int i;
 | |
| 
 | |
|     if ((nargs < 6) || (nargs > 7) || nret != 1) {
 | |
|         rtas_st(rets, 0, RTAS_OUT_PARAM_ERROR);
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     mask = rtas_ld(args, 2);
 | |
|     buf = rtas_ld(args, 4);
 | |
|     len = rtas_ld(args, 5);
 | |
| 
 | |
|     event = rtas_event_log_dequeue(spapr, mask);
 | |
|     if (!event) {
 | |
|         goto out_no_events;
 | |
|     }
 | |
| 
 | |
|     event_len = event->extended_length + sizeof(header);
 | |
| 
 | |
|     if (event_len < len) {
 | |
|         len = event_len;
 | |
|     }
 | |
| 
 | |
|     header.summary = cpu_to_be32(event->summary);
 | |
|     header.extended_length = cpu_to_be32(event->extended_length);
 | |
|     cpu_physical_memory_write(buf, &header, sizeof(header));
 | |
|     cpu_physical_memory_write(buf + sizeof(header), event->extended_log,
 | |
|                               event->extended_length);
 | |
|     rtas_st(rets, 0, RTAS_OUT_SUCCESS);
 | |
|     g_free(event->extended_log);
 | |
|     g_free(event);
 | |
| 
 | |
|     /* according to PAPR+, the IRQ must be left asserted, or re-asserted, if
 | |
|      * there are still pending events to be fetched via check-exception. We
 | |
|      * do the latter here, since our code relies on edge-triggered
 | |
|      * interrupts.
 | |
|      */
 | |
|     for (i = 0; i < EVENT_CLASS_MAX; i++) {
 | |
|         if (rtas_event_log_contains(spapr, EVENT_CLASS_MASK(i))) {
 | |
|             const SpaprEventSource *source =
 | |
|                 spapr_event_sources_get_source(spapr->event_sources, i);
 | |
| 
 | |
|             g_assert(source->enabled);
 | |
|             qemu_irq_pulse(spapr_qirq(spapr, source->irq));
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return;
 | |
| 
 | |
| out_no_events:
 | |
|     rtas_st(rets, 0, RTAS_OUT_NO_ERRORS_FOUND);
 | |
| }
 | |
| 
 | |
| static void event_scan(PowerPCCPU *cpu, SpaprMachineState *spapr,
 | |
|                        uint32_t token, uint32_t nargs,
 | |
|                        target_ulong args,
 | |
|                        uint32_t nret, target_ulong rets)
 | |
| {
 | |
|     int i;
 | |
|     if (nargs != 4 || nret != 1) {
 | |
|         rtas_st(rets, 0, RTAS_OUT_PARAM_ERROR);
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     for (i = 0; i < EVENT_CLASS_MAX; i++) {
 | |
|         if (rtas_event_log_contains(spapr, EVENT_CLASS_MASK(i))) {
 | |
|             const SpaprEventSource *source =
 | |
|                 spapr_event_sources_get_source(spapr->event_sources, i);
 | |
| 
 | |
|             g_assert(source->enabled);
 | |
|             qemu_irq_pulse(spapr_qirq(spapr, source->irq));
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     rtas_st(rets, 0, RTAS_OUT_NO_ERRORS_FOUND);
 | |
| }
 | |
| 
 | |
| void spapr_clear_pending_events(SpaprMachineState *spapr)
 | |
| {
 | |
|     SpaprEventLogEntry *entry = NULL, *next_entry;
 | |
| 
 | |
|     QTAILQ_FOREACH_SAFE(entry, &spapr->pending_events, next, next_entry) {
 | |
|         QTAILQ_REMOVE(&spapr->pending_events, entry, next);
 | |
|         g_free(entry->extended_log);
 | |
|         g_free(entry);
 | |
|     }
 | |
| }
 | |
| 
 | |
| void spapr_clear_pending_hotplug_events(SpaprMachineState *spapr)
 | |
| {
 | |
|     SpaprEventLogEntry *entry = NULL, *next_entry;
 | |
| 
 | |
|     QTAILQ_FOREACH_SAFE(entry, &spapr->pending_events, next, next_entry) {
 | |
|         if (spapr_event_log_entry_type(entry) == RTAS_LOG_TYPE_HOTPLUG) {
 | |
|             QTAILQ_REMOVE(&spapr->pending_events, entry, next);
 | |
|             g_free(entry->extended_log);
 | |
|             g_free(entry);
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| void spapr_events_init(SpaprMachineState *spapr)
 | |
| {
 | |
|     int epow_irq = SPAPR_IRQ_EPOW;
 | |
| 
 | |
|     if (SPAPR_MACHINE_GET_CLASS(spapr)->legacy_irq_allocation) {
 | |
|         epow_irq = spapr_irq_findone(spapr, &error_fatal);
 | |
|     }
 | |
| 
 | |
|     spapr_irq_claim(spapr, epow_irq, false, &error_fatal);
 | |
| 
 | |
|     QTAILQ_INIT(&spapr->pending_events);
 | |
| 
 | |
|     spapr->event_sources = spapr_event_sources_new();
 | |
| 
 | |
|     spapr_event_sources_register(spapr->event_sources, EVENT_CLASS_EPOW,
 | |
|                                  epow_irq);
 | |
| 
 | |
|     /* NOTE: if machine supports modern/dedicated hotplug event source,
 | |
|      * we add it to the device-tree unconditionally. This means we may
 | |
|      * have cases where the source is enabled in QEMU, but unused by the
 | |
|      * guest because it does not support modern hotplug events, so we
 | |
|      * take care to rely on checking for negotiation of OV5_HP_EVT option
 | |
|      * before attempting to use it to signal events, rather than simply
 | |
|      * checking that it's enabled.
 | |
|      */
 | |
|     if (spapr->use_hotplug_event_source) {
 | |
|         int hp_irq = SPAPR_IRQ_HOTPLUG;
 | |
| 
 | |
|         if (SPAPR_MACHINE_GET_CLASS(spapr)->legacy_irq_allocation) {
 | |
|             hp_irq = spapr_irq_findone(spapr, &error_fatal);
 | |
|         }
 | |
| 
 | |
|         spapr_irq_claim(spapr, hp_irq, false, &error_fatal);
 | |
| 
 | |
|         spapr_event_sources_register(spapr->event_sources, EVENT_CLASS_HOT_PLUG,
 | |
|                                      hp_irq);
 | |
|     }
 | |
| 
 | |
|     spapr->epow_notifier.notify = spapr_powerdown_req;
 | |
|     qemu_register_powerdown_notifier(&spapr->epow_notifier);
 | |
|     spapr_rtas_register(RTAS_CHECK_EXCEPTION, "check-exception",
 | |
|                         check_exception);
 | |
|     spapr_rtas_register(RTAS_EVENT_SCAN, "event-scan", event_scan);
 | |
| }
 |