1 // SPDX-License-Identifier: GPL-2.0 2 /* Copyright(c) 2022 Intel Corporation. All rights reserved. */ 3 #undef TRACE_SYSTEM 4 #define TRACE_SYSTEM cxl 5 6 #if !defined(_CXL_EVENTS_H) || defined(TRACE_HEADER_MULTI_READ) 7 #define _CXL_EVENTS_H 8 9 #include <linux/tracepoint.h> 10 #include <linux/pci.h> 11 #include <linux/unaligned.h> 12 13 #include <cxl.h> 14 #include <cxlmem.h> 15 #include "core.h" 16 17 #define CXL_RAS_UC_CACHE_DATA_PARITY BIT(0) 18 #define CXL_RAS_UC_CACHE_ADDR_PARITY BIT(1) 19 #define CXL_RAS_UC_CACHE_BE_PARITY BIT(2) 20 #define CXL_RAS_UC_CACHE_DATA_ECC BIT(3) 21 #define CXL_RAS_UC_MEM_DATA_PARITY BIT(4) 22 #define CXL_RAS_UC_MEM_ADDR_PARITY BIT(5) 23 #define CXL_RAS_UC_MEM_BE_PARITY BIT(6) 24 #define CXL_RAS_UC_MEM_DATA_ECC BIT(7) 25 #define CXL_RAS_UC_REINIT_THRESH BIT(8) 26 #define CXL_RAS_UC_RSVD_ENCODE BIT(9) 27 #define CXL_RAS_UC_POISON BIT(10) 28 #define CXL_RAS_UC_RECV_OVERFLOW BIT(11) 29 #define CXL_RAS_UC_INTERNAL_ERR BIT(14) 30 #define CXL_RAS_UC_IDE_TX_ERR BIT(15) 31 #define CXL_RAS_UC_IDE_RX_ERR BIT(16) 32 33 #define show_uc_errs(status) __print_flags(status, " | ", \ 34 { CXL_RAS_UC_CACHE_DATA_PARITY, "Cache Data Parity Error" }, \ 35 { CXL_RAS_UC_CACHE_ADDR_PARITY, "Cache Address Parity Error" }, \ 36 { CXL_RAS_UC_CACHE_BE_PARITY, "Cache Byte Enable Parity Error" }, \ 37 { CXL_RAS_UC_CACHE_DATA_ECC, "Cache Data ECC Error" }, \ 38 { CXL_RAS_UC_MEM_DATA_PARITY, "Memory Data Parity Error" }, \ 39 { CXL_RAS_UC_MEM_ADDR_PARITY, "Memory Address Parity Error" }, \ 40 { CXL_RAS_UC_MEM_BE_PARITY, "Memory Byte Enable Parity Error" }, \ 41 { CXL_RAS_UC_MEM_DATA_ECC, "Memory Data ECC Error" }, \ 42 { CXL_RAS_UC_REINIT_THRESH, "REINIT Threshold Hit" }, \ 43 { CXL_RAS_UC_RSVD_ENCODE, "Received Unrecognized Encoding" }, \ 44 { CXL_RAS_UC_POISON, "Received Poison From Peer" }, \ 45 { CXL_RAS_UC_RECV_OVERFLOW, "Receiver Overflow" }, \ 46 { CXL_RAS_UC_INTERNAL_ERR, "Component Specific Error" }, \ 47 { CXL_RAS_UC_IDE_TX_ERR, "IDE Tx Error" }, \ 48 { CXL_RAS_UC_IDE_RX_ERR, "IDE Rx Error" } \ 49 ) 50 51 TRACE_EVENT(cxl_port_aer_uncorrectable_error, 52 TP_PROTO(struct device *dev, u32 status, u32 fe, u32 *hl), 53 TP_ARGS(dev, status, fe, hl), 54 TP_STRUCT__entry( 55 __string(device, dev_name(dev)) 56 __string(host, dev_name(dev->parent)) 57 __field(u32, status) 58 __field(u32, first_error) 59 __array(u32, header_log, CXL_HEADERLOG_SIZE_U32) 60 ), 61 TP_fast_assign( 62 __assign_str(device); 63 __assign_str(host); 64 __entry->status = status; 65 __entry->first_error = fe; 66 /* 67 * Embed the 512B headerlog data for user app retrieval and 68 * parsing, but no need to print this in the trace buffer. 69 */ 70 memcpy(__entry->header_log, hl, CXL_HEADERLOG_SIZE); 71 ), 72 TP_printk("device=%s host=%s status: '%s' first_error: '%s'", 73 __get_str(device), __get_str(host), 74 show_uc_errs(__entry->status), 75 show_uc_errs(__entry->first_error) 76 ) 77 ); 78 79 TRACE_EVENT(cxl_aer_uncorrectable_error, 80 TP_PROTO(const struct cxl_memdev *cxlmd, u32 status, u32 fe, u32 *hl), 81 TP_ARGS(cxlmd, status, fe, hl), 82 TP_STRUCT__entry( 83 __string(memdev, dev_name(&cxlmd->dev)) 84 __string(host, dev_name(cxlmd->dev.parent)) 85 __field(u64, serial) 86 __field(u32, status) 87 __field(u32, first_error) 88 __array(u32, header_log, CXL_HEADERLOG_SIZE_U32) 89 ), 90 TP_fast_assign( 91 __assign_str(memdev); 92 __assign_str(host); 93 __entry->serial = cxlmd->cxlds->serial; 94 __entry->status = status; 95 __entry->first_error = fe; 96 /* 97 * Embed the 512B headerlog data for user app retrieval and 98 * parsing, but no need to print this in the trace buffer. 99 */ 100 memcpy(__entry->header_log, hl, CXL_HEADERLOG_SIZE); 101 ), 102 TP_printk("memdev=%s host=%s serial=%lld: status: '%s' first_error: '%s'", 103 __get_str(memdev), __get_str(host), __entry->serial, 104 show_uc_errs(__entry->status), 105 show_uc_errs(__entry->first_error) 106 ) 107 ); 108 109 #define CXL_RAS_CE_CACHE_DATA_ECC BIT(0) 110 #define CXL_RAS_CE_MEM_DATA_ECC BIT(1) 111 #define CXL_RAS_CE_CRC_THRESH BIT(2) 112 #define CLX_RAS_CE_RETRY_THRESH BIT(3) 113 #define CXL_RAS_CE_CACHE_POISON BIT(4) 114 #define CXL_RAS_CE_MEM_POISON BIT(5) 115 #define CXL_RAS_CE_PHYS_LAYER_ERR BIT(6) 116 117 #define show_ce_errs(status) __print_flags(status, " | ", \ 118 { CXL_RAS_CE_CACHE_DATA_ECC, "Cache Data ECC Error" }, \ 119 { CXL_RAS_CE_MEM_DATA_ECC, "Memory Data ECC Error" }, \ 120 { CXL_RAS_CE_CRC_THRESH, "CRC Threshold Hit" }, \ 121 { CLX_RAS_CE_RETRY_THRESH, "Retry Threshold" }, \ 122 { CXL_RAS_CE_CACHE_POISON, "Received Cache Poison From Peer" }, \ 123 { CXL_RAS_CE_MEM_POISON, "Received Memory Poison From Peer" }, \ 124 { CXL_RAS_CE_PHYS_LAYER_ERR, "Received Error From Physical Layer" } \ 125 ) 126 127 TRACE_EVENT(cxl_port_aer_correctable_error, 128 TP_PROTO(struct device *dev, u32 status), 129 TP_ARGS(dev, status), 130 TP_STRUCT__entry( 131 __string(device, dev_name(dev)) 132 __string(host, dev_name(dev->parent)) 133 __field(u32, status) 134 ), 135 TP_fast_assign( 136 __assign_str(device); 137 __assign_str(host); 138 __entry->status = status; 139 ), 140 TP_printk("device=%s host=%s status='%s'", 141 __get_str(device), __get_str(host), 142 show_ce_errs(__entry->status) 143 ) 144 ); 145 146 TRACE_EVENT(cxl_aer_correctable_error, 147 TP_PROTO(const struct cxl_memdev *cxlmd, u32 status), 148 TP_ARGS(cxlmd, status), 149 TP_STRUCT__entry( 150 __string(memdev, dev_name(&cxlmd->dev)) 151 __string(host, dev_name(cxlmd->dev.parent)) 152 __field(u64, serial) 153 __field(u32, status) 154 ), 155 TP_fast_assign( 156 __assign_str(memdev); 157 __assign_str(host); 158 __entry->serial = cxlmd->cxlds->serial; 159 __entry->status = status; 160 ), 161 TP_printk("memdev=%s host=%s serial=%lld: status: '%s'", 162 __get_str(memdev), __get_str(host), __entry->serial, 163 show_ce_errs(__entry->status) 164 ) 165 ); 166 167 #define cxl_event_log_type_str(type) \ 168 __print_symbolic(type, \ 169 { CXL_EVENT_TYPE_INFO, "Informational" }, \ 170 { CXL_EVENT_TYPE_WARN, "Warning" }, \ 171 { CXL_EVENT_TYPE_FAIL, "Failure" }, \ 172 { CXL_EVENT_TYPE_FATAL, "Fatal" }) 173 174 TRACE_EVENT(cxl_overflow, 175 176 TP_PROTO(const struct cxl_memdev *cxlmd, enum cxl_event_log_type log, 177 struct cxl_get_event_payload *payload), 178 179 TP_ARGS(cxlmd, log, payload), 180 181 TP_STRUCT__entry( 182 __string(memdev, dev_name(&cxlmd->dev)) 183 __string(host, dev_name(cxlmd->dev.parent)) 184 __field(int, log) 185 __field(u64, serial) 186 __field(u64, first_ts) 187 __field(u64, last_ts) 188 __field(u16, count) 189 ), 190 191 TP_fast_assign( 192 __assign_str(memdev); 193 __assign_str(host); 194 __entry->serial = cxlmd->cxlds->serial; 195 __entry->log = log; 196 __entry->count = le16_to_cpu(payload->overflow_err_count); 197 __entry->first_ts = le64_to_cpu(payload->first_overflow_timestamp); 198 __entry->last_ts = le64_to_cpu(payload->last_overflow_timestamp); 199 ), 200 201 TP_printk("memdev=%s host=%s serial=%lld: log=%s : %u records from %llu to %llu", 202 __get_str(memdev), __get_str(host), __entry->serial, 203 cxl_event_log_type_str(__entry->log), __entry->count, 204 __entry->first_ts, __entry->last_ts) 205 206 ); 207 208 /* 209 * Common Event Record Format 210 * CXL 3.0 section 8.2.9.2.1; Table 8-42 211 */ 212 #define CXL_EVENT_RECORD_FLAG_PERMANENT BIT(2) 213 #define CXL_EVENT_RECORD_FLAG_MAINT_NEEDED BIT(3) 214 #define CXL_EVENT_RECORD_FLAG_PERF_DEGRADED BIT(4) 215 #define CXL_EVENT_RECORD_FLAG_HW_REPLACE BIT(5) 216 #define CXL_EVENT_RECORD_FLAG_MAINT_OP_SUB_CLASS_VALID BIT(6) 217 #define CXL_EVENT_RECORD_FLAG_LD_ID_VALID BIT(7) 218 #define CXL_EVENT_RECORD_FLAG_HEAD_ID_VALID BIT(8) 219 #define show_hdr_flags(flags) __print_flags(flags, " | ", \ 220 { CXL_EVENT_RECORD_FLAG_PERMANENT, "PERMANENT_CONDITION" }, \ 221 { CXL_EVENT_RECORD_FLAG_MAINT_NEEDED, "MAINTENANCE_NEEDED" }, \ 222 { CXL_EVENT_RECORD_FLAG_PERF_DEGRADED, "PERFORMANCE_DEGRADED" }, \ 223 { CXL_EVENT_RECORD_FLAG_HW_REPLACE, "HARDWARE_REPLACEMENT_NEEDED" }, \ 224 { CXL_EVENT_RECORD_FLAG_MAINT_OP_SUB_CLASS_VALID, "MAINT_OP_SUB_CLASS_VALID" }, \ 225 { CXL_EVENT_RECORD_FLAG_LD_ID_VALID, "LD_ID_VALID" }, \ 226 { CXL_EVENT_RECORD_FLAG_HEAD_ID_VALID, "HEAD_ID_VALID" } \ 227 ) 228 229 /* 230 * Define macros for the common header of each CXL event. 231 * 232 * Tracepoints using these macros must do 3 things: 233 * 234 * 1) Add CXL_EVT_TP_entry to TP_STRUCT__entry 235 * 2) Use CXL_EVT_TP_fast_assign within TP_fast_assign; 236 * pass the dev, log, and CXL event header 237 * NOTE: The uuid must be assigned by the specific trace event 238 * 3) Use CXL_EVT_TP_printk() instead of TP_printk() 239 * 240 * See the generic_event tracepoint as an example. 241 */ 242 #define CXL_EVT_TP_entry \ 243 __string(memdev, dev_name(&cxlmd->dev)) \ 244 __string(host, dev_name(cxlmd->dev.parent)) \ 245 __field(int, log) \ 246 __field_struct(uuid_t, hdr_uuid) \ 247 __field(u64, serial) \ 248 __field(u32, hdr_flags) \ 249 __field(u16, hdr_handle) \ 250 __field(u16, hdr_related_handle) \ 251 __field(u64, hdr_timestamp) \ 252 __field(u8, hdr_length) \ 253 __field(u8, hdr_maint_op_class) \ 254 __field(u8, hdr_maint_op_sub_class) \ 255 __field(u16, hdr_ld_id) \ 256 __field(u8, hdr_head_id) 257 258 #define CXL_EVT_TP_fast_assign(cxlmd, l, hdr) \ 259 __assign_str(memdev); \ 260 __assign_str(host); \ 261 __entry->log = (l); \ 262 __entry->serial = (cxlmd)->cxlds->serial; \ 263 __entry->hdr_length = (hdr).length; \ 264 __entry->hdr_flags = get_unaligned_le24((hdr).flags); \ 265 __entry->hdr_handle = le16_to_cpu((hdr).handle); \ 266 __entry->hdr_related_handle = le16_to_cpu((hdr).related_handle); \ 267 __entry->hdr_timestamp = le64_to_cpu((hdr).timestamp); \ 268 __entry->hdr_maint_op_class = (hdr).maint_op_class; \ 269 __entry->hdr_maint_op_sub_class = (hdr).maint_op_sub_class; \ 270 __entry->hdr_ld_id = le16_to_cpu((hdr).ld_id); \ 271 __entry->hdr_head_id = (hdr).head_id 272 273 #define CXL_EVT_TP_printk(fmt, ...) \ 274 TP_printk("memdev=%s host=%s serial=%lld log=%s : time=%llu uuid=%pUb " \ 275 "len=%d flags='%s' handle=%x related_handle=%x " \ 276 "maint_op_class=%u maint_op_sub_class=%u " \ 277 "ld_id=%x head_id=%x : " fmt, \ 278 __get_str(memdev), __get_str(host), __entry->serial, \ 279 cxl_event_log_type_str(__entry->log), \ 280 __entry->hdr_timestamp, &__entry->hdr_uuid, __entry->hdr_length,\ 281 show_hdr_flags(__entry->hdr_flags), __entry->hdr_handle, \ 282 __entry->hdr_related_handle, __entry->hdr_maint_op_class, \ 283 __entry->hdr_maint_op_sub_class, \ 284 __entry->hdr_ld_id, __entry->hdr_head_id, \ 285 ##__VA_ARGS__) 286 287 TRACE_EVENT(cxl_generic_event, 288 289 TP_PROTO(const struct cxl_memdev *cxlmd, enum cxl_event_log_type log, 290 const uuid_t *uuid, struct cxl_event_generic *gen_rec), 291 292 TP_ARGS(cxlmd, log, uuid, gen_rec), 293 294 TP_STRUCT__entry( 295 CXL_EVT_TP_entry 296 __array(u8, data, CXL_EVENT_RECORD_DATA_LENGTH) 297 ), 298 299 TP_fast_assign( 300 CXL_EVT_TP_fast_assign(cxlmd, log, gen_rec->hdr); 301 memcpy(&__entry->hdr_uuid, uuid, sizeof(uuid_t)); 302 memcpy(__entry->data, gen_rec->data, CXL_EVENT_RECORD_DATA_LENGTH); 303 ), 304 305 CXL_EVT_TP_printk("%s", 306 __print_hex(__entry->data, CXL_EVENT_RECORD_DATA_LENGTH)) 307 ); 308 309 /* 310 * Physical Address field masks 311 * 312 * General Media Event Record 313 * CXL rev 3.0 Section 8.2.9.2.1.1; Table 8-43 314 * 315 * DRAM Event Record 316 * CXL rev 3.0 section 8.2.9.2.1.2; Table 8-44 317 */ 318 #define CXL_DPA_FLAGS_MASK GENMASK(1, 0) 319 #define CXL_DPA_MASK GENMASK_ULL(63, 6) 320 321 #define CXL_DPA_VOLATILE BIT(0) 322 #define CXL_DPA_NOT_REPAIRABLE BIT(1) 323 #define show_dpa_flags(flags) __print_flags(flags, "|", \ 324 { CXL_DPA_VOLATILE, "VOLATILE" }, \ 325 { CXL_DPA_NOT_REPAIRABLE, "NOT_REPAIRABLE" } \ 326 ) 327 328 /* 329 * Component ID Format 330 * CXL 3.1 section 8.2.9.2.1; Table 8-44 331 */ 332 #define CXL_PLDM_COMPONENT_ID_ENTITY_VALID BIT(0) 333 #define CXL_PLDM_COMPONENT_ID_RES_VALID BIT(1) 334 335 #define show_comp_id_pldm_flags(flags) __print_flags(flags, " | ", \ 336 { CXL_PLDM_COMPONENT_ID_ENTITY_VALID, "PLDM Entity ID" }, \ 337 { CXL_PLDM_COMPONENT_ID_RES_VALID, "Resource ID" } \ 338 ) 339 340 #define show_pldm_entity_id(flags, valid_comp_id, valid_id_format, comp_id) \ 341 (flags & valid_comp_id && flags & valid_id_format) ? \ 342 (comp_id[0] & CXL_PLDM_COMPONENT_ID_ENTITY_VALID) ? \ 343 __print_hex(&comp_id[1], 6) : "0x00" : "0x00" 344 345 #define show_pldm_resource_id(flags, valid_comp_id, valid_id_format, comp_id) \ 346 (flags & valid_comp_id && flags & valid_id_format) ? \ 347 (comp_id[0] & CXL_PLDM_COMPONENT_ID_RES_VALID) ? \ 348 __print_hex(&comp_id[7], 4) : "0x00" : "0x00" 349 350 /* 351 * General Media Event Record - GMER 352 * CXL rev 3.1 Section 8.2.9.2.1.1; Table 8-45 353 */ 354 #define CXL_GMER_EVT_DESC_UNCORECTABLE_EVENT BIT(0) 355 #define CXL_GMER_EVT_DESC_THRESHOLD_EVENT BIT(1) 356 #define CXL_GMER_EVT_DESC_POISON_LIST_OVERFLOW BIT(2) 357 #define show_event_desc_flags(flags) __print_flags(flags, "|", \ 358 { CXL_GMER_EVT_DESC_UNCORECTABLE_EVENT, "UNCORRECTABLE_EVENT" }, \ 359 { CXL_GMER_EVT_DESC_THRESHOLD_EVENT, "THRESHOLD_EVENT" }, \ 360 { CXL_GMER_EVT_DESC_POISON_LIST_OVERFLOW, "POISON_LIST_OVERFLOW" } \ 361 ) 362 363 #define CXL_GMER_MEM_EVT_TYPE_ECC_ERROR 0x00 364 #define CXL_GMER_MEM_EVT_TYPE_INV_ADDR 0x01 365 #define CXL_GMER_MEM_EVT_TYPE_DATA_PATH_ERROR 0x02 366 #define CXL_GMER_MEM_EVT_TYPE_TE_STATE_VIOLATION 0x03 367 #define CXL_GMER_MEM_EVT_TYPE_SCRUB_MEDIA_ECC_ERROR 0x04 368 #define CXL_GMER_MEM_EVT_TYPE_AP_CME_COUNTER_EXPIRE 0x05 369 #define CXL_GMER_MEM_EVT_TYPE_CKID_VIOLATION 0x06 370 #define show_gmer_mem_event_type(type) __print_symbolic(type, \ 371 { CXL_GMER_MEM_EVT_TYPE_ECC_ERROR, "ECC Error" }, \ 372 { CXL_GMER_MEM_EVT_TYPE_INV_ADDR, "Invalid Address" }, \ 373 { CXL_GMER_MEM_EVT_TYPE_DATA_PATH_ERROR, "Data Path Error" }, \ 374 { CXL_GMER_MEM_EVT_TYPE_TE_STATE_VIOLATION, "TE State Violation" }, \ 375 { CXL_GMER_MEM_EVT_TYPE_SCRUB_MEDIA_ECC_ERROR, "Scrub Media ECC Error" }, \ 376 { CXL_GMER_MEM_EVT_TYPE_AP_CME_COUNTER_EXPIRE, "Adv Prog CME Counter Expiration" }, \ 377 { CXL_GMER_MEM_EVT_TYPE_CKID_VIOLATION, "CKID Violation" } \ 378 ) 379 380 #define CXL_GMER_TRANS_UNKNOWN 0x00 381 #define CXL_GMER_TRANS_HOST_READ 0x01 382 #define CXL_GMER_TRANS_HOST_WRITE 0x02 383 #define CXL_GMER_TRANS_HOST_SCAN_MEDIA 0x03 384 #define CXL_GMER_TRANS_HOST_INJECT_POISON 0x04 385 #define CXL_GMER_TRANS_INTERNAL_MEDIA_SCRUB 0x05 386 #define CXL_GMER_TRANS_INTERNAL_MEDIA_MANAGEMENT 0x06 387 #define CXL_GMER_TRANS_INTERNAL_MEDIA_ECS 0x07 388 #define CXL_GMER_TRANS_MEDIA_INITIALIZATION 0x08 389 #define show_trans_type(type) __print_symbolic(type, \ 390 { CXL_GMER_TRANS_UNKNOWN, "Unknown" }, \ 391 { CXL_GMER_TRANS_HOST_READ, "Host Read" }, \ 392 { CXL_GMER_TRANS_HOST_WRITE, "Host Write" }, \ 393 { CXL_GMER_TRANS_HOST_SCAN_MEDIA, "Host Scan Media" }, \ 394 { CXL_GMER_TRANS_HOST_INJECT_POISON, "Host Inject Poison" }, \ 395 { CXL_GMER_TRANS_INTERNAL_MEDIA_SCRUB, "Internal Media Scrub" }, \ 396 { CXL_GMER_TRANS_INTERNAL_MEDIA_MANAGEMENT, "Internal Media Management" }, \ 397 { CXL_GMER_TRANS_INTERNAL_MEDIA_ECS, "Internal Media Error Check Scrub" }, \ 398 { CXL_GMER_TRANS_MEDIA_INITIALIZATION, "Media Initialization" } \ 399 ) 400 401 #define CXL_GMER_VALID_CHANNEL BIT(0) 402 #define CXL_GMER_VALID_RANK BIT(1) 403 #define CXL_GMER_VALID_DEVICE BIT(2) 404 #define CXL_GMER_VALID_COMPONENT BIT(3) 405 #define CXL_GMER_VALID_COMPONENT_ID_FORMAT BIT(4) 406 #define show_valid_flags(flags) __print_flags(flags, "|", \ 407 { CXL_GMER_VALID_CHANNEL, "CHANNEL" }, \ 408 { CXL_GMER_VALID_RANK, "RANK" }, \ 409 { CXL_GMER_VALID_DEVICE, "DEVICE" }, \ 410 { CXL_GMER_VALID_COMPONENT, "COMPONENT" }, \ 411 { CXL_GMER_VALID_COMPONENT_ID_FORMAT, "COMPONENT PLDM FORMAT" } \ 412 ) 413 414 #define CXL_GMER_CME_EV_FLAG_CME_MULTIPLE_MEDIA BIT(0) 415 #define CXL_GMER_CME_EV_FLAG_THRESHOLD_EXCEEDED BIT(1) 416 #define show_cme_threshold_ev_flags(flags) __print_flags(flags, "|", \ 417 { \ 418 CXL_GMER_CME_EV_FLAG_CME_MULTIPLE_MEDIA, \ 419 "Corrected Memory Errors in Multiple Media Components" \ 420 }, { \ 421 CXL_GMER_CME_EV_FLAG_THRESHOLD_EXCEEDED, \ 422 "Exceeded Programmable Threshold" \ 423 } \ 424 ) 425 426 #define CXL_GMER_MEM_EVT_SUB_TYPE_NOT_REPORTED 0x00 427 #define CXL_GMER_MEM_EVT_SUB_TYPE_INTERNAL_DATAPATH_ERROR 0x01 428 #define CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_COMMAND_TRAINING_ERROR 0x02 429 #define CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_CONTROL_TRAINING_ERROR 0x03 430 #define CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_DATA_TRAINING_ERROR 0x04 431 #define CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_CRC_ERROR 0x05 432 #define show_mem_event_sub_type(sub_type) __print_symbolic(sub_type, \ 433 { CXL_GMER_MEM_EVT_SUB_TYPE_NOT_REPORTED, "Not Reported" }, \ 434 { CXL_GMER_MEM_EVT_SUB_TYPE_INTERNAL_DATAPATH_ERROR, "Internal Datapath Error" }, \ 435 { \ 436 CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_COMMAND_TRAINING_ERROR, \ 437 "Media Link Command Training Error" \ 438 }, { \ 439 CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_CONTROL_TRAINING_ERROR, \ 440 "Media Link Control Training Error" \ 441 }, { \ 442 CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_DATA_TRAINING_ERROR, \ 443 "Media Link Data Training Error" \ 444 }, { \ 445 CXL_GMER_MEM_EVT_SUB_TYPE_MEDIA_LINK_CRC_ERROR, "Media Link CRC Error" \ 446 } \ 447 ) 448 449 TRACE_EVENT(cxl_general_media, 450 451 TP_PROTO(const struct cxl_memdev *cxlmd, enum cxl_event_log_type log, 452 struct cxl_region *cxlr, u64 hpa, u64 hpa_alias0, 453 struct cxl_event_gen_media *rec), 454 455 TP_ARGS(cxlmd, log, cxlr, hpa, hpa_alias0, rec), 456 457 TP_STRUCT__entry( 458 CXL_EVT_TP_entry 459 /* General Media */ 460 __field(u64, dpa) 461 __field(u8, descriptor) 462 __field(u8, type) 463 __field(u8, transaction_type) 464 __field(u8, channel) 465 __field(u32, device) 466 __array(u8, comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE) 467 /* Following are out of order to pack trace record */ 468 __field(u64, hpa) 469 __field(u64, hpa_alias0) 470 __field_struct(uuid_t, region_uuid) 471 __field(u16, validity_flags) 472 __field(u8, rank) 473 __field(u8, dpa_flags) 474 __field(u32, cme_count) 475 __field(u8, sub_type) 476 __field(u8, cme_threshold_ev_flags) 477 __string(region_name, cxlr ? dev_name(&cxlr->dev) : "") 478 ), 479 480 TP_fast_assign( 481 CXL_EVT_TP_fast_assign(cxlmd, log, rec->media_hdr.hdr); 482 __entry->hdr_uuid = CXL_EVENT_GEN_MEDIA_UUID; 483 484 /* General Media */ 485 __entry->dpa = le64_to_cpu(rec->media_hdr.phys_addr); 486 __entry->dpa_flags = __entry->dpa & CXL_DPA_FLAGS_MASK; 487 /* Mask after flags have been parsed */ 488 __entry->dpa &= CXL_DPA_MASK; 489 __entry->descriptor = rec->media_hdr.descriptor; 490 __entry->type = rec->media_hdr.type; 491 __entry->sub_type = rec->sub_type; 492 __entry->transaction_type = rec->media_hdr.transaction_type; 493 __entry->channel = rec->media_hdr.channel; 494 __entry->rank = rec->media_hdr.rank; 495 __entry->device = get_unaligned_le24(rec->device); 496 memcpy(__entry->comp_id, &rec->component_id, 497 CXL_EVENT_GEN_MED_COMP_ID_SIZE); 498 __entry->validity_flags = get_unaligned_le16(&rec->media_hdr.validity_flags); 499 __entry->hpa = hpa; 500 __entry->hpa_alias0 = hpa_alias0; 501 if (cxlr) { 502 __assign_str(region_name); 503 uuid_copy(&__entry->region_uuid, &cxlr->params.uuid); 504 } else { 505 __assign_str(region_name); 506 uuid_copy(&__entry->region_uuid, &uuid_null); 507 } 508 __entry->cme_threshold_ev_flags = rec->cme_threshold_ev_flags; 509 if (rec->media_hdr.descriptor & CXL_GMER_EVT_DESC_THRESHOLD_EVENT) 510 __entry->cme_count = get_unaligned_le24(rec->cme_count); 511 else 512 __entry->cme_count = 0; 513 ), 514 515 CXL_EVT_TP_printk("dpa=%llx dpa_flags='%s' " \ 516 "descriptor='%s' type='%s' sub_type='%s' " \ 517 "transaction_type='%s' channel=%u rank=%u " \ 518 "device=%x validity_flags='%s' " \ 519 "comp_id=%s comp_id_pldm_valid_flags='%s' " \ 520 "pldm_entity_id=%s pldm_resource_id=%s " \ 521 "hpa=%llx hpa_alias0=%llx region=%s region_uuid=%pUb " \ 522 "cme_threshold_ev_flags='%s' cme_count=%u", 523 __entry->dpa, show_dpa_flags(__entry->dpa_flags), 524 show_event_desc_flags(__entry->descriptor), 525 show_gmer_mem_event_type(__entry->type), 526 show_mem_event_sub_type(__entry->sub_type), 527 show_trans_type(__entry->transaction_type), 528 __entry->channel, __entry->rank, __entry->device, 529 show_valid_flags(__entry->validity_flags), 530 __print_hex(__entry->comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE), 531 show_comp_id_pldm_flags(__entry->comp_id[0]), 532 show_pldm_entity_id(__entry->validity_flags, CXL_GMER_VALID_COMPONENT, 533 CXL_GMER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id), 534 show_pldm_resource_id(__entry->validity_flags, CXL_GMER_VALID_COMPONENT, 535 CXL_GMER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id), 536 __entry->hpa, __entry->hpa_alias0, __get_str(region_name), &__entry->region_uuid, 537 show_cme_threshold_ev_flags(__entry->cme_threshold_ev_flags), __entry->cme_count 538 ) 539 ); 540 541 /* 542 * DRAM Event Record - DER 543 * 544 * CXL rev 3.1 section 8.2.9.2.1.2; Table 8-46 545 */ 546 /* 547 * DRAM Event Record defines many fields the same as the General Media Event 548 * Record. Reuse those definitions as appropriate. 549 */ 550 #define CXL_DER_MEM_EVT_TYPE_ECC_ERROR 0x00 551 #define CXL_DER_MEM_EVT_TYPE_SCRUB_MEDIA_ECC_ERROR 0x01 552 #define CXL_DER_MEM_EVT_TYPE_INV_ADDR 0x02 553 #define CXL_DER_MEM_EVT_TYPE_DATA_PATH_ERROR 0x03 554 #define CXL_DER_MEM_EVT_TYPE_TE_STATE_VIOLATION 0x04 555 #define CXL_DER_MEM_EVT_TYPE_AP_CME_COUNTER_EXPIRE 0x05 556 #define CXL_DER_MEM_EVT_TYPE_CKID_VIOLATION 0x06 557 #define show_dram_mem_event_type(type) __print_symbolic(type, \ 558 { CXL_DER_MEM_EVT_TYPE_ECC_ERROR, "ECC Error" }, \ 559 { CXL_DER_MEM_EVT_TYPE_SCRUB_MEDIA_ECC_ERROR, "Scrub Media ECC Error" }, \ 560 { CXL_DER_MEM_EVT_TYPE_INV_ADDR, "Invalid Address" }, \ 561 { CXL_DER_MEM_EVT_TYPE_DATA_PATH_ERROR, "Data Path Error" }, \ 562 { CXL_DER_MEM_EVT_TYPE_TE_STATE_VIOLATION, "TE State Violation" }, \ 563 { CXL_DER_MEM_EVT_TYPE_AP_CME_COUNTER_EXPIRE, "Adv Prog CME Counter Expiration" }, \ 564 { CXL_DER_MEM_EVT_TYPE_CKID_VIOLATION, "CKID Violation" } \ 565 ) 566 567 #define CXL_DER_VALID_CHANNEL BIT(0) 568 #define CXL_DER_VALID_RANK BIT(1) 569 #define CXL_DER_VALID_NIBBLE BIT(2) 570 #define CXL_DER_VALID_BANK_GROUP BIT(3) 571 #define CXL_DER_VALID_BANK BIT(4) 572 #define CXL_DER_VALID_ROW BIT(5) 573 #define CXL_DER_VALID_COLUMN BIT(6) 574 #define CXL_DER_VALID_CORRECTION_MASK BIT(7) 575 #define CXL_DER_VALID_COMPONENT BIT(8) 576 #define CXL_DER_VALID_COMPONENT_ID_FORMAT BIT(9) 577 #define CXL_DER_VALID_SUB_CHANNEL BIT(10) 578 #define show_dram_valid_flags(flags) __print_flags(flags, "|", \ 579 { CXL_DER_VALID_CHANNEL, "CHANNEL" }, \ 580 { CXL_DER_VALID_RANK, "RANK" }, \ 581 { CXL_DER_VALID_NIBBLE, "NIBBLE" }, \ 582 { CXL_DER_VALID_BANK_GROUP, "BANK GROUP" }, \ 583 { CXL_DER_VALID_BANK, "BANK" }, \ 584 { CXL_DER_VALID_ROW, "ROW" }, \ 585 { CXL_DER_VALID_COLUMN, "COLUMN" }, \ 586 { CXL_DER_VALID_CORRECTION_MASK, "CORRECTION MASK" }, \ 587 { CXL_DER_VALID_COMPONENT, "COMPONENT" }, \ 588 { CXL_DER_VALID_COMPONENT_ID_FORMAT, "COMPONENT PLDM FORMAT" }, \ 589 { CXL_DER_VALID_SUB_CHANNEL, "SUB CHANNEL" } \ 590 ) 591 592 TRACE_EVENT(cxl_dram, 593 594 TP_PROTO(const struct cxl_memdev *cxlmd, enum cxl_event_log_type log, 595 struct cxl_region *cxlr, u64 hpa, u64 hpa_alias0, 596 struct cxl_event_dram *rec), 597 598 TP_ARGS(cxlmd, log, cxlr, hpa, hpa_alias0, rec), 599 600 TP_STRUCT__entry( 601 CXL_EVT_TP_entry 602 /* DRAM */ 603 __field(u64, dpa) 604 __field(u8, descriptor) 605 __field(u8, type) 606 __field(u8, transaction_type) 607 __field(u8, channel) 608 __field(u16, validity_flags) 609 __field(u16, column) /* Out of order to pack trace record */ 610 __field(u32, nibble_mask) 611 __field(u32, row) 612 __array(u8, cor_mask, CXL_EVENT_DER_CORRECTION_MASK_SIZE) 613 __field(u64, hpa) 614 __field(u64, hpa_alias0) 615 __field_struct(uuid_t, region_uuid) 616 __field(u8, rank) /* Out of order to pack trace record */ 617 __field(u8, bank_group) /* Out of order to pack trace record */ 618 __field(u8, bank) /* Out of order to pack trace record */ 619 __field(u8, dpa_flags) /* Out of order to pack trace record */ 620 /* Following are out of order to pack trace record */ 621 __array(u8, comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE) 622 __field(u32, cvme_count) 623 __field(u8, sub_type) 624 __field(u8, sub_channel) 625 __field(u8, cme_threshold_ev_flags) 626 __string(region_name, cxlr ? dev_name(&cxlr->dev) : "") 627 ), 628 629 TP_fast_assign( 630 CXL_EVT_TP_fast_assign(cxlmd, log, rec->media_hdr.hdr); 631 __entry->hdr_uuid = CXL_EVENT_DRAM_UUID; 632 633 /* DRAM */ 634 __entry->dpa = le64_to_cpu(rec->media_hdr.phys_addr); 635 __entry->dpa_flags = __entry->dpa & CXL_DPA_FLAGS_MASK; 636 __entry->dpa &= CXL_DPA_MASK; 637 __entry->descriptor = rec->media_hdr.descriptor; 638 __entry->type = rec->media_hdr.type; 639 __entry->sub_type = rec->sub_type; 640 __entry->transaction_type = rec->media_hdr.transaction_type; 641 __entry->validity_flags = get_unaligned_le16(rec->media_hdr.validity_flags); 642 __entry->channel = rec->media_hdr.channel; 643 __entry->rank = rec->media_hdr.rank; 644 __entry->nibble_mask = get_unaligned_le24(rec->nibble_mask); 645 __entry->bank_group = rec->bank_group; 646 __entry->bank = rec->bank; 647 __entry->row = get_unaligned_le24(rec->row); 648 __entry->column = get_unaligned_le16(rec->column); 649 memcpy(__entry->cor_mask, &rec->correction_mask, 650 CXL_EVENT_DER_CORRECTION_MASK_SIZE); 651 __entry->hpa = hpa; 652 __entry->hpa_alias0 = hpa_alias0; 653 if (cxlr) { 654 __assign_str(region_name); 655 uuid_copy(&__entry->region_uuid, &cxlr->params.uuid); 656 } else { 657 __assign_str(region_name); 658 uuid_copy(&__entry->region_uuid, &uuid_null); 659 } 660 memcpy(__entry->comp_id, &rec->component_id, 661 CXL_EVENT_GEN_MED_COMP_ID_SIZE); 662 __entry->sub_channel = rec->sub_channel; 663 __entry->cme_threshold_ev_flags = rec->cme_threshold_ev_flags; 664 if (rec->media_hdr.descriptor & CXL_GMER_EVT_DESC_THRESHOLD_EVENT) 665 __entry->cvme_count = get_unaligned_le24(rec->cvme_count); 666 else 667 __entry->cvme_count = 0; 668 ), 669 670 CXL_EVT_TP_printk("dpa=%llx dpa_flags='%s' descriptor='%s' type='%s' sub_type='%s' " \ 671 "transaction_type='%s' channel=%u rank=%u nibble_mask=%x " \ 672 "bank_group=%u bank=%u row=%u column=%u cor_mask=%s " \ 673 "validity_flags='%s' " \ 674 "comp_id=%s comp_id_pldm_valid_flags='%s' " \ 675 "pldm_entity_id=%s pldm_resource_id=%s " \ 676 "hpa=%llx hpa_alias0=%llx region=%s region_uuid=%pUb " \ 677 "sub_channel=%u cme_threshold_ev_flags='%s' cvme_count=%u", 678 __entry->dpa, show_dpa_flags(__entry->dpa_flags), 679 show_event_desc_flags(__entry->descriptor), 680 show_dram_mem_event_type(__entry->type), 681 show_mem_event_sub_type(__entry->sub_type), 682 show_trans_type(__entry->transaction_type), 683 __entry->channel, __entry->rank, __entry->nibble_mask, 684 __entry->bank_group, __entry->bank, 685 __entry->row, __entry->column, 686 __print_hex(__entry->cor_mask, CXL_EVENT_DER_CORRECTION_MASK_SIZE), 687 show_dram_valid_flags(__entry->validity_flags), 688 __print_hex(__entry->comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE), 689 show_comp_id_pldm_flags(__entry->comp_id[0]), 690 show_pldm_entity_id(__entry->validity_flags, CXL_DER_VALID_COMPONENT, 691 CXL_DER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id), 692 show_pldm_resource_id(__entry->validity_flags, CXL_DER_VALID_COMPONENT, 693 CXL_DER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id), 694 __entry->hpa, __entry->hpa_alias0, __get_str(region_name), &__entry->region_uuid, 695 __entry->sub_channel, show_cme_threshold_ev_flags(__entry->cme_threshold_ev_flags), 696 __entry->cvme_count 697 ) 698 ); 699 700 /* 701 * Memory Module Event Record - MMER 702 * 703 * CXL res 3.1 section 8.2.9.2.1.3; Table 8-47 704 */ 705 #define CXL_MMER_HEALTH_STATUS_CHANGE 0x00 706 #define CXL_MMER_MEDIA_STATUS_CHANGE 0x01 707 #define CXL_MMER_LIFE_USED_CHANGE 0x02 708 #define CXL_MMER_TEMP_CHANGE 0x03 709 #define CXL_MMER_DATA_PATH_ERROR 0x04 710 #define CXL_MMER_LSA_ERROR 0x05 711 #define CXL_MMER_UNRECOV_SIDEBAND_BUS_ERROR 0x06 712 #define CXL_MMER_MEMORY_MEDIA_FRU_ERROR 0x07 713 #define CXL_MMER_POWER_MANAGEMENT_FAULT 0x08 714 #define show_dev_evt_type(type) __print_symbolic(type, \ 715 { CXL_MMER_HEALTH_STATUS_CHANGE, "Health Status Change" }, \ 716 { CXL_MMER_MEDIA_STATUS_CHANGE, "Media Status Change" }, \ 717 { CXL_MMER_LIFE_USED_CHANGE, "Life Used Change" }, \ 718 { CXL_MMER_TEMP_CHANGE, "Temperature Change" }, \ 719 { CXL_MMER_DATA_PATH_ERROR, "Data Path Error" }, \ 720 { CXL_MMER_LSA_ERROR, "LSA Error" }, \ 721 { CXL_MMER_UNRECOV_SIDEBAND_BUS_ERROR, "Unrecoverable Internal Sideband Bus Error" }, \ 722 { CXL_MMER_MEMORY_MEDIA_FRU_ERROR, "Memory Media FRU Error" }, \ 723 { CXL_MMER_POWER_MANAGEMENT_FAULT, "Power Management Fault" } \ 724 ) 725 726 /* 727 * Device Health Information - DHI 728 * 729 * CXL res 3.1 section 8.2.9.9.3.1; Table 8-133 730 */ 731 #define CXL_DHI_HS_MAINTENANCE_NEEDED BIT(0) 732 #define CXL_DHI_HS_PERFORMANCE_DEGRADED BIT(1) 733 #define CXL_DHI_HS_HW_REPLACEMENT_NEEDED BIT(2) 734 #define CXL_DHI_HS_MEM_CAPACITY_DEGRADED BIT(3) 735 #define show_health_status_flags(flags) __print_flags(flags, "|", \ 736 { CXL_DHI_HS_MAINTENANCE_NEEDED, "MAINTENANCE_NEEDED" }, \ 737 { CXL_DHI_HS_PERFORMANCE_DEGRADED, "PERFORMANCE_DEGRADED" }, \ 738 { CXL_DHI_HS_HW_REPLACEMENT_NEEDED, "REPLACEMENT_NEEDED" }, \ 739 { CXL_DHI_HS_MEM_CAPACITY_DEGRADED, "MEM_CAPACITY_DEGRADED" } \ 740 ) 741 742 #define CXL_DHI_MS_NORMAL 0x00 743 #define CXL_DHI_MS_NOT_READY 0x01 744 #define CXL_DHI_MS_WRITE_PERSISTENCY_LOST 0x02 745 #define CXL_DHI_MS_ALL_DATA_LOST 0x03 746 #define CXL_DHI_MS_WRITE_PERSISTENCY_LOSS_EVENT_POWER_LOSS 0x04 747 #define CXL_DHI_MS_WRITE_PERSISTENCY_LOSS_EVENT_SHUTDOWN 0x05 748 #define CXL_DHI_MS_WRITE_PERSISTENCY_LOSS_IMMINENT 0x06 749 #define CXL_DHI_MS_WRITE_ALL_DATA_LOSS_EVENT_POWER_LOSS 0x07 750 #define CXL_DHI_MS_WRITE_ALL_DATA_LOSS_EVENT_SHUTDOWN 0x08 751 #define CXL_DHI_MS_WRITE_ALL_DATA_LOSS_IMMINENT 0x09 752 #define show_media_status(ms) __print_symbolic(ms, \ 753 { CXL_DHI_MS_NORMAL, \ 754 "Normal" }, \ 755 { CXL_DHI_MS_NOT_READY, \ 756 "Not Ready" }, \ 757 { CXL_DHI_MS_WRITE_PERSISTENCY_LOST, \ 758 "Write Persistency Lost" }, \ 759 { CXL_DHI_MS_ALL_DATA_LOST, \ 760 "All Data Lost" }, \ 761 { CXL_DHI_MS_WRITE_PERSISTENCY_LOSS_EVENT_POWER_LOSS, \ 762 "Write Persistency Loss in the Event of Power Loss" }, \ 763 { CXL_DHI_MS_WRITE_PERSISTENCY_LOSS_EVENT_SHUTDOWN, \ 764 "Write Persistency Loss in Event of Shutdown" }, \ 765 { CXL_DHI_MS_WRITE_PERSISTENCY_LOSS_IMMINENT, \ 766 "Write Persistency Loss Imminent" }, \ 767 { CXL_DHI_MS_WRITE_ALL_DATA_LOSS_EVENT_POWER_LOSS, \ 768 "All Data Loss in Event of Power Loss" }, \ 769 { CXL_DHI_MS_WRITE_ALL_DATA_LOSS_EVENT_SHUTDOWN, \ 770 "All Data loss in the Event of Shutdown" }, \ 771 { CXL_DHI_MS_WRITE_ALL_DATA_LOSS_IMMINENT, \ 772 "All Data Loss Imminent" } \ 773 ) 774 775 #define CXL_DHI_AS_NORMAL 0x0 776 #define CXL_DHI_AS_WARNING 0x1 777 #define CXL_DHI_AS_CRITICAL 0x2 778 #define show_two_bit_status(as) __print_symbolic(as, \ 779 { CXL_DHI_AS_NORMAL, "Normal" }, \ 780 { CXL_DHI_AS_WARNING, "Warning" }, \ 781 { CXL_DHI_AS_CRITICAL, "Critical" } \ 782 ) 783 #define show_one_bit_status(as) __print_symbolic(as, \ 784 { CXL_DHI_AS_NORMAL, "Normal" }, \ 785 { CXL_DHI_AS_WARNING, "Warning" } \ 786 ) 787 788 #define CXL_DHI_AS_LIFE_USED(as) (as & 0x3) 789 #define CXL_DHI_AS_DEV_TEMP(as) ((as & 0xC) >> 2) 790 #define CXL_DHI_AS_COR_VOL_ERR_CNT(as) ((as & 0x10) >> 4) 791 #define CXL_DHI_AS_COR_PER_ERR_CNT(as) ((as & 0x20) >> 5) 792 793 #define CXL_MMER_VALID_COMPONENT BIT(0) 794 #define CXL_MMER_VALID_COMPONENT_ID_FORMAT BIT(1) 795 #define show_mem_module_valid_flags(flags) __print_flags(flags, "|", \ 796 { CXL_MMER_VALID_COMPONENT, "COMPONENT" }, \ 797 { CXL_MMER_VALID_COMPONENT_ID_FORMAT, "COMPONENT PLDM FORMAT" } \ 798 ) 799 #define CXL_MMER_DEV_EVT_SUB_TYPE_NOT_REPORTED 0x00 800 #define CXL_MMER_DEV_EVT_SUB_TYPE_INVALID_CONFIG_DATA 0x01 801 #define CXL_MMER_DEV_EVT_SUB_TYPE_UNSUPP_CONFIG_DATA 0x02 802 #define CXL_MMER_DEV_EVT_SUB_TYPE_UNSUPP_MEM_MEDIA_FRU 0x03 803 #define show_dev_event_sub_type(sub_type) __print_symbolic(sub_type, \ 804 { CXL_MMER_DEV_EVT_SUB_TYPE_NOT_REPORTED, "Not Reported" }, \ 805 { CXL_MMER_DEV_EVT_SUB_TYPE_INVALID_CONFIG_DATA, "Invalid Config Data" }, \ 806 { CXL_MMER_DEV_EVT_SUB_TYPE_UNSUPP_CONFIG_DATA, "Unsupported Config Data" }, \ 807 { \ 808 CXL_MMER_DEV_EVT_SUB_TYPE_UNSUPP_MEM_MEDIA_FRU, \ 809 "Unsupported Memory Media FRU" \ 810 } \ 811 ) 812 813 TRACE_EVENT(cxl_memory_module, 814 815 TP_PROTO(const struct cxl_memdev *cxlmd, enum cxl_event_log_type log, 816 struct cxl_event_mem_module *rec), 817 818 TP_ARGS(cxlmd, log, rec), 819 820 TP_STRUCT__entry( 821 CXL_EVT_TP_entry 822 823 /* Memory Module Event */ 824 __field(u8, event_type) 825 826 /* Device Health Info */ 827 __field(u8, health_status) 828 __field(u8, media_status) 829 __field(u8, life_used) 830 __field(u32, dirty_shutdown_cnt) 831 __field(u32, cor_vol_err_cnt) 832 __field(u32, cor_per_err_cnt) 833 __field(s16, device_temp) 834 __field(u8, add_status) 835 __field(u8, event_sub_type) 836 __array(u8, comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE) 837 __field(u16, validity_flags) 838 ), 839 840 TP_fast_assign( 841 CXL_EVT_TP_fast_assign(cxlmd, log, rec->hdr); 842 __entry->hdr_uuid = CXL_EVENT_MEM_MODULE_UUID; 843 844 /* Memory Module Event */ 845 __entry->event_type = rec->event_type; 846 __entry->event_sub_type = rec->event_sub_type; 847 848 /* Device Health Info */ 849 __entry->health_status = rec->info.health_status; 850 __entry->media_status = rec->info.media_status; 851 __entry->life_used = rec->info.life_used; 852 __entry->dirty_shutdown_cnt = get_unaligned_le32(rec->info.dirty_shutdown_cnt); 853 __entry->cor_vol_err_cnt = get_unaligned_le32(rec->info.cor_vol_err_cnt); 854 __entry->cor_per_err_cnt = get_unaligned_le32(rec->info.cor_per_err_cnt); 855 __entry->device_temp = get_unaligned_le16(rec->info.device_temp); 856 __entry->add_status = rec->info.add_status; 857 __entry->validity_flags = get_unaligned_le16(rec->validity_flags); 858 memcpy(__entry->comp_id, &rec->component_id, 859 CXL_EVENT_GEN_MED_COMP_ID_SIZE); 860 ), 861 862 CXL_EVT_TP_printk("event_type='%s' event_sub_type='%s' health_status='%s' " \ 863 "media_status='%s' as_life_used=%s as_dev_temp=%s as_cor_vol_err_cnt=%s " \ 864 "as_cor_per_err_cnt=%s life_used=%u device_temp=%d " \ 865 "dirty_shutdown_cnt=%u cor_vol_err_cnt=%u cor_per_err_cnt=%u " \ 866 "validity_flags='%s' " \ 867 "comp_id=%s comp_id_pldm_valid_flags='%s' " \ 868 "pldm_entity_id=%s pldm_resource_id=%s", 869 show_dev_evt_type(__entry->event_type), 870 show_dev_event_sub_type(__entry->event_sub_type), 871 show_health_status_flags(__entry->health_status), 872 show_media_status(__entry->media_status), 873 show_two_bit_status(CXL_DHI_AS_LIFE_USED(__entry->add_status)), 874 show_two_bit_status(CXL_DHI_AS_DEV_TEMP(__entry->add_status)), 875 show_one_bit_status(CXL_DHI_AS_COR_VOL_ERR_CNT(__entry->add_status)), 876 show_one_bit_status(CXL_DHI_AS_COR_PER_ERR_CNT(__entry->add_status)), 877 __entry->life_used, __entry->device_temp, 878 __entry->dirty_shutdown_cnt, __entry->cor_vol_err_cnt, 879 __entry->cor_per_err_cnt, 880 show_mem_module_valid_flags(__entry->validity_flags), 881 __print_hex(__entry->comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE), 882 show_comp_id_pldm_flags(__entry->comp_id[0]), 883 show_pldm_entity_id(__entry->validity_flags, CXL_MMER_VALID_COMPONENT, 884 CXL_MMER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id), 885 show_pldm_resource_id(__entry->validity_flags, CXL_MMER_VALID_COMPONENT, 886 CXL_MMER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id) 887 ) 888 ); 889 890 /* 891 * Memory Sparing Event Record - MSER 892 * 893 * CXL rev 3.2 section 8.2.10.2.1.4; Table 8-60 894 */ 895 #define CXL_MSER_QUERY_RESOURCE_FLAG BIT(0) 896 #define CXL_MSER_HARD_SPARING_FLAG BIT(1) 897 #define CXL_MSER_DEV_INITED_FLAG BIT(2) 898 #define show_mem_sparing_flags(flags) __print_flags(flags, "|", \ 899 { CXL_MSER_QUERY_RESOURCE_FLAG, "Query Resources" }, \ 900 { CXL_MSER_HARD_SPARING_FLAG, "Hard Sparing" }, \ 901 { CXL_MSER_DEV_INITED_FLAG, "Device Initiated Sparing" } \ 902 ) 903 904 #define CXL_MSER_VALID_CHANNEL BIT(0) 905 #define CXL_MSER_VALID_RANK BIT(1) 906 #define CXL_MSER_VALID_NIBBLE BIT(2) 907 #define CXL_MSER_VALID_BANK_GROUP BIT(3) 908 #define CXL_MSER_VALID_BANK BIT(4) 909 #define CXL_MSER_VALID_ROW BIT(5) 910 #define CXL_MSER_VALID_COLUMN BIT(6) 911 #define CXL_MSER_VALID_COMPONENT_ID BIT(7) 912 #define CXL_MSER_VALID_COMPONENT_ID_FORMAT BIT(8) 913 #define CXL_MSER_VALID_SUB_CHANNEL BIT(9) 914 #define show_mem_sparing_valid_flags(flags) __print_flags(flags, "|", \ 915 { CXL_MSER_VALID_CHANNEL, "CHANNEL" }, \ 916 { CXL_MSER_VALID_RANK, "RANK" }, \ 917 { CXL_MSER_VALID_NIBBLE, "NIBBLE" }, \ 918 { CXL_MSER_VALID_BANK_GROUP, "BANK GROUP" }, \ 919 { CXL_MSER_VALID_BANK, "BANK" }, \ 920 { CXL_MSER_VALID_ROW, "ROW" }, \ 921 { CXL_MSER_VALID_COLUMN, "COLUMN" }, \ 922 { CXL_MSER_VALID_COMPONENT_ID, "COMPONENT ID" }, \ 923 { CXL_MSER_VALID_COMPONENT_ID_FORMAT, "COMPONENT ID PLDM FORMAT" }, \ 924 { CXL_MSER_VALID_SUB_CHANNEL, "SUB CHANNEL" } \ 925 ) 926 927 TRACE_EVENT(cxl_memory_sparing, 928 929 TP_PROTO(const struct cxl_memdev *cxlmd, enum cxl_event_log_type log, 930 struct cxl_event_mem_sparing *rec), 931 932 TP_ARGS(cxlmd, log, rec), 933 934 TP_STRUCT__entry( 935 CXL_EVT_TP_entry 936 937 /* Memory Sparing Event */ 938 __field(u8, flags) 939 __field(u8, result) 940 __field(u16, validity_flags) 941 __field(u16, res_avail) 942 __field(u8, channel) 943 __field(u8, rank) 944 __field(u32, nibble_mask) 945 __field(u8, bank_group) 946 __field(u8, bank) 947 __field(u32, row) 948 __field(u16, column) 949 __field(u8, sub_channel) 950 __array(u8, comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE) 951 ), 952 953 TP_fast_assign( 954 CXL_EVT_TP_fast_assign(cxlmd, log, rec->hdr); 955 __entry->hdr_uuid = CXL_EVENT_MEM_SPARING_UUID; 956 957 /* Memory Sparing Event */ 958 __entry->flags = rec->flags; 959 __entry->result = rec->result; 960 __entry->validity_flags = le16_to_cpu(rec->validity_flags); 961 __entry->res_avail = le16_to_cpu(rec->res_avail); 962 __entry->channel = rec->channel; 963 __entry->rank = rec->rank; 964 __entry->nibble_mask = get_unaligned_le24(rec->nibble_mask); 965 __entry->bank_group = rec->bank_group; 966 __entry->bank = rec->bank; 967 __entry->row = get_unaligned_le24(rec->row); 968 __entry->column = le16_to_cpu(rec->column); 969 __entry->sub_channel = rec->sub_channel; 970 memcpy(__entry->comp_id, &rec->component_id, 971 CXL_EVENT_GEN_MED_COMP_ID_SIZE); 972 ), 973 974 CXL_EVT_TP_printk("flags='%s' result=%u validity_flags='%s' " \ 975 "spare resource avail=%u channel=%u rank=%u " \ 976 "nibble_mask=%x bank_group=%u bank=%u " \ 977 "row=%u column=%u sub_channel=%u " \ 978 "comp_id=%s comp_id_pldm_valid_flags='%s' " \ 979 "pldm_entity_id=%s pldm_resource_id=%s", 980 show_mem_sparing_flags(__entry->flags), 981 __entry->result, 982 show_mem_sparing_valid_flags(__entry->validity_flags), 983 __entry->res_avail, __entry->channel, __entry->rank, 984 __entry->nibble_mask, __entry->bank_group, __entry->bank, 985 __entry->row, __entry->column, __entry->sub_channel, 986 __print_hex(__entry->comp_id, CXL_EVENT_GEN_MED_COMP_ID_SIZE), 987 show_comp_id_pldm_flags(__entry->comp_id[0]), 988 show_pldm_entity_id(__entry->validity_flags, CXL_MSER_VALID_COMPONENT_ID, 989 CXL_MSER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id), 990 show_pldm_resource_id(__entry->validity_flags, CXL_MSER_VALID_COMPONENT_ID, 991 CXL_MSER_VALID_COMPONENT_ID_FORMAT, __entry->comp_id) 992 ) 993 ); 994 995 #define show_poison_trace_type(type) \ 996 __print_symbolic(type, \ 997 { CXL_POISON_TRACE_LIST, "List" }, \ 998 { CXL_POISON_TRACE_INJECT, "Inject" }, \ 999 { CXL_POISON_TRACE_CLEAR, "Clear" }) 1000 1001 #define __show_poison_source(source) \ 1002 __print_symbolic(source, \ 1003 { CXL_POISON_SOURCE_UNKNOWN, "Unknown" }, \ 1004 { CXL_POISON_SOURCE_EXTERNAL, "External" }, \ 1005 { CXL_POISON_SOURCE_INTERNAL, "Internal" }, \ 1006 { CXL_POISON_SOURCE_INJECTED, "Injected" }, \ 1007 { CXL_POISON_SOURCE_VENDOR, "Vendor" }) 1008 1009 #define show_poison_source(source) \ 1010 (((source > CXL_POISON_SOURCE_INJECTED) && \ 1011 (source != CXL_POISON_SOURCE_VENDOR)) ? "Reserved" \ 1012 : __show_poison_source(source)) 1013 1014 #define show_poison_flags(flags) \ 1015 __print_flags(flags, "|", \ 1016 { CXL_POISON_FLAG_MORE, "More" }, \ 1017 { CXL_POISON_FLAG_OVERFLOW, "Overflow" }, \ 1018 { CXL_POISON_FLAG_SCANNING, "Scanning" }) 1019 1020 #define __cxl_poison_addr(record) \ 1021 (le64_to_cpu(record->address)) 1022 #define cxl_poison_record_dpa(record) \ 1023 (__cxl_poison_addr(record) & CXL_POISON_START_MASK) 1024 #define cxl_poison_record_source(record) \ 1025 (__cxl_poison_addr(record) & CXL_POISON_SOURCE_MASK) 1026 #define cxl_poison_record_dpa_length(record) \ 1027 (le32_to_cpu(record->length) * CXL_POISON_LEN_MULT) 1028 #define cxl_poison_overflow(flags, time) \ 1029 (flags & CXL_POISON_FLAG_OVERFLOW ? le64_to_cpu(time) : 0) 1030 1031 TRACE_EVENT(cxl_poison, 1032 1033 TP_PROTO(struct cxl_memdev *cxlmd, struct cxl_region *cxlr, 1034 const struct cxl_poison_record *record, u8 flags, 1035 __le64 overflow_ts, enum cxl_poison_trace_type trace_type), 1036 1037 TP_ARGS(cxlmd, cxlr, record, flags, overflow_ts, trace_type), 1038 1039 TP_STRUCT__entry( 1040 __string(memdev, dev_name(&cxlmd->dev)) 1041 __string(host, dev_name(cxlmd->dev.parent)) 1042 __field(u64, serial) 1043 __field(u8, trace_type) 1044 __string(region, cxlr ? dev_name(&cxlr->dev) : "") 1045 __field(u64, overflow_ts) 1046 __field(u64, hpa) 1047 __field(u64, hpa_alias0) 1048 __field(u64, dpa) 1049 __field(u32, dpa_length) 1050 __array(char, uuid, 16) 1051 __field(u8, source) 1052 __field(u8, flags) 1053 ), 1054 1055 TP_fast_assign( 1056 __assign_str(memdev); 1057 __assign_str(host); 1058 __entry->serial = cxlmd->cxlds->serial; 1059 __entry->overflow_ts = cxl_poison_overflow(flags, overflow_ts); 1060 __entry->dpa = cxl_poison_record_dpa(record); 1061 __entry->dpa_length = cxl_poison_record_dpa_length(record); 1062 __entry->source = cxl_poison_record_source(record); 1063 __entry->trace_type = trace_type; 1064 __entry->flags = flags; 1065 if (cxlr) { 1066 __assign_str(region); 1067 memcpy(__entry->uuid, &cxlr->params.uuid, 16); 1068 __entry->hpa = cxl_dpa_to_hpa(cxlr, cxlmd, 1069 __entry->dpa); 1070 if (__entry->hpa != ULLONG_MAX && cxlr->params.cache_size) 1071 __entry->hpa_alias0 = __entry->hpa + 1072 cxlr->params.cache_size; 1073 else 1074 __entry->hpa_alias0 = ULLONG_MAX; 1075 } else { 1076 __assign_str(region); 1077 memset(__entry->uuid, 0, 16); 1078 __entry->hpa = ULLONG_MAX; 1079 __entry->hpa_alias0 = ULLONG_MAX; 1080 } 1081 ), 1082 1083 TP_printk("memdev=%s host=%s serial=%lld trace_type=%s region=%s " \ 1084 "region_uuid=%pU hpa=0x%llx hpa_alias0=0x%llx dpa=0x%llx " \ 1085 "dpa_length=0x%x source=%s flags=%s overflow_time=%llu", 1086 __get_str(memdev), 1087 __get_str(host), 1088 __entry->serial, 1089 show_poison_trace_type(__entry->trace_type), 1090 __get_str(region), 1091 __entry->uuid, 1092 __entry->hpa, 1093 __entry->hpa_alias0, 1094 __entry->dpa, 1095 __entry->dpa_length, 1096 show_poison_source(__entry->source), 1097 show_poison_flags(__entry->flags), 1098 __entry->overflow_ts 1099 ) 1100 ); 1101 1102 #endif /* _CXL_EVENTS_H */ 1103 1104 #define TRACE_INCLUDE_FILE trace 1105 #include <trace/define_trace.h> 1106