On 5/2/25 1:45 AM, shiju.jose@xxxxxxxxxx wrote: > From: Shiju Jose <shiju.jose@xxxxxxxxxx> > > Memory sparing is defined as a repair function that replaces a portion of > memory with a portion of functional memory at that same DPA. The subclasses > for this operation vary in terms of the scope of the sparing being > performed. The cacheline sparing subclass refers to a sparing action that > can replace a full cacheline. Row sparing is provided as an alternative to > PPR sparing functions and its scope is that of a single DDR row. > As per CXL r3.2 Table 8-125 foot note 1. Memory sparing is preferred over > PPR when possible. > Bank sparing allows an entire bank to be replaced. Rank sparing is defined > as an operation in which an entire DDR rank is replaced. > > Memory sparing maintenance operations may be supported by CXL devices > that implement CXL.mem protocol. A sparing maintenance operation requests > the CXL device to perform a repair operation on its media. > For example, a CXL device with DRAM components that support memory sparing > features may implement sparing maintenance operations. > > The host may issue a query command by setting query resources flag in the > input payload (CXL spec 3.2 Table 8-120) to determine availability of > sparing resources for a given address. In response to a query request, > the device shall report the resource availability by producing the memory > sparing event record (CXL spec 3.2 Table 8-60) in which the Channel, Rank, > Nibble Mask, Bank Group, Bank, Row, Column, Sub-Channel fields are a copy > of the values specified in the request. > > During the execution of a sparing maintenance operation, a CXL memory > device: > - may not retain data > - may not be able to process CXL.mem requests correctly. > These CXL memory device capabilities are specified by restriction flags > in the memory sparing feature readable attributes. > > When a CXL device identifies error on a memory component, the device > may inform the host about the need for a memory sparing maintenance > operation by using DRAM event record, where the 'maintenance needed' flag > may set. The event record contains some of the DPA, Channel, Rank, > Nibble Mask, Bank Group, Bank, Row, Column, Sub-Channel fields that > should be repaired. The userspace tool requests for maintenance operation > if the 'maintenance needed' flag set in the CXL DRAM error record. > > CXL spec 3.2 section 8.2.10.7.1.4 describes the device's memory sparing > maintenance operation feature. > > CXL spec 3.2 section 8.2.10.7.2.3 describes the memory sparing feature > discovery and configuration. > > Add support for controlling CXL memory device memory sparing feature. > Register with EDAC driver, which gets the memory repair attr descriptors > from the EDAC memory repair driver and exposes sysfs repair control > attributes for memory sparing to the userspace. For example CXL memory > sparing control for the CXL mem0 device is exposed in > /sys/bus/edac/devices/cxl_mem0/mem_repairX/ > > Use case > ======== > 1. CXL device identifies a failure in a memory component, report to > userspace in a CXL DRAM trace event with DPA and other attributes of > memory to repair such as channel, rank, nibble mask, bank Group, > bank, row, column, sub-channel. > > 2. Rasdaemon process the trace event and may issue query request in sysfs > check resources available for memory sparing if either of the following > conditions met. > - 'maintenance needed' flag set in the event record. > - 'threshold event' flag set for CVME threshold feature. > - When the number of corrected error reported on a CXL.mem media to the > userspace exceeds the threshold value for corrected error count defined > by the userspace policy. > > 3. Rasdaemon process the memory sparing trace event and issue repair > request for memory sparing. > > Kernel CXL driver shall report memory sparing event record to the userspace > with the resource availability in order rasdaemon to process the event > record and issue a repair request in sysfs for the memory sparing operation > in the CXL device. > > Note: Based on the feedbacks from the community 'query' sysfs attribute is > removed and reporting memory sparing error record to the userspace are not > supported. Instead userspace issues sparing operation and kernel does the > same to the CXL memory device, when 'maintenance needed' flag set in the > DRAM event record. > > Add checks to ensure the memory to be repaired is offline and if online, > then originates from a CXL DRAM error record reported in the current boot > before requesting a memory sparing operation on the device. > > Note: Tested for memory sparing control feature with > "hw/cxl: Add memory sparing control feature" > Repository: "https://gitlab.com/shiju.jose/qemu.git" > Branch: cxl-ras-features-2024-10-24 > > Reviewed-by: Jonathan Cameron <Jonathan.Cameron@xxxxxxxxxx> > Signed-off-by: Shiju Jose <shiju.jose@xxxxxxxxxx> Reviewed-by: Dave Jiang <dave.jiang@xxxxxxxxx> > --- > Documentation/edac/memory_repair.rst | 21 + > drivers/cxl/core/edac.c | 548 ++++++++++++++++++++++++++- > drivers/edac/mem_repair.c | 9 + > include/linux/edac.h | 7 + > 4 files changed, 583 insertions(+), 2 deletions(-) > > diff --git a/Documentation/edac/memory_repair.rst b/Documentation/edac/memory_repair.rst > index 52162a422864..5b1cd8297442 100644 > --- a/Documentation/edac/memory_repair.rst > +++ b/Documentation/edac/memory_repair.rst > @@ -119,3 +119,24 @@ sysfs > > Sysfs files are documented in > `Documentation/ABI/testing/sysfs-edac-memory-repair`. > + > +Examples > +-------- > + > +The memory repair usage takes the form shown in this example: > + > +1. CXL memory sparing > + > +Memory sparing is defined as a repair function that replaces a portion of > +memory with a portion of functional memory at that same DPA. The subclass > +for this operation, cacheline/row/bank/rank sparing, vary in terms of the > +scope of the sparing being performed. > + > +Memory sparing maintenance operations may be supported by CXL devices that > +implement CXL.mem protocol. A sparing maintenance operation requests the > +CXL device to perform a repair operation on its media. For example, a CXL > +device with DRAM components that support memory sparing features may > +implement sparing maintenance operations. > + > +Sysfs files for memory repair are documented in > +`Documentation/ABI/testing/sysfs-edac-memory-repair` > diff --git a/drivers/cxl/core/edac.c b/drivers/cxl/core/edac.c > index 02bd4c675871..19a2e8d09b3d 100644 > --- a/drivers/cxl/core/edac.c > +++ b/drivers/cxl/core/edac.c > @@ -21,7 +21,17 @@ > #include "core.h" > #include "trace.h" > > -#define CXL_NR_EDAC_DEV_FEATURES 2 > +#define CXL_NR_EDAC_DEV_FEATURES 6 > + > +static bool cxl_is_memdev_memory_online(const struct cxl_memdev *cxlmd) > +{ > + struct cxl_port *port = cxlmd->endpoint; > + > + if (port && cxl_num_decoders_committed(port)) > + return true; > + > + return false; > +} > > #ifdef CONFIG_CXL_EDAC_SCRUB > struct cxl_patrol_scrub_context { > @@ -1116,13 +1126,534 @@ int cxl_store_rec_dram(struct cxl_memdev *cxlmd, union cxl_event *evt) > return 0; > } > EXPORT_SYMBOL_NS_GPL(cxl_store_rec_dram, "CXL"); > + > +/* > + * CXL memory sparing control > + */ > +enum cxl_mem_sparing_granularity { > + CXL_MEM_SPARING_CACHELINE, > + CXL_MEM_SPARING_ROW, > + CXL_MEM_SPARING_BANK, > + CXL_MEM_SPARING_RANK, > + CXL_MEM_SPARING_MAX > +}; > + > +struct cxl_mem_sparing_context { > + struct cxl_memdev *cxlmd; > + uuid_t repair_uuid; > + u16 get_feat_size; > + u16 set_feat_size; > + u16 effects; > + u8 instance; > + u8 get_version; > + u8 set_version; > + u8 op_class; > + u8 op_subclass; > + bool cap_safe_when_in_use; > + bool cap_hard_sparing; > + bool cap_soft_sparing; > + u8 channel; > + u8 rank; > + u8 bank_group; > + u32 nibble_mask; > + u64 dpa; > + u32 row; > + u16 column; > + u8 bank; > + u8 sub_channel; > + enum edac_mem_repair_type repair_type; > + bool persist_mode; > +}; > + > +#define CXL_SPARING_RD_CAP_SAFE_IN_USE_MASK BIT(0) > +#define CXL_SPARING_RD_CAP_HARD_SPARING_MASK BIT(1) > +#define CXL_SPARING_RD_CAP_SOFT_SPARING_MASK BIT(2) > + > +#define CXL_SPARING_WR_DEVICE_INITIATED_MASK BIT(0) > + > +#define CXL_SPARING_QUERY_RESOURCE_FLAG BIT(0) > +#define CXL_SET_HARD_SPARING_FLAG BIT(1) > +#define CXL_SPARING_SUB_CHNL_VALID_FLAG BIT(2) > +#define CXL_SPARING_NIB_MASK_VALID_FLAG BIT(3) > + > +#define CXL_GET_SPARING_SAFE_IN_USE(flags) \ > + (FIELD_GET(CXL_SPARING_RD_CAP_SAFE_IN_USE_MASK, \ > + flags) ^ 1) > +#define CXL_GET_CAP_HARD_SPARING(flags) \ > + FIELD_GET(CXL_SPARING_RD_CAP_HARD_SPARING_MASK, \ > + flags) > +#define CXL_GET_CAP_SOFT_SPARING(flags) \ > + FIELD_GET(CXL_SPARING_RD_CAP_SOFT_SPARING_MASK, \ > + flags) > + > +#define CXL_SET_SPARING_QUERY_RESOURCE(val) \ > + FIELD_PREP(CXL_SPARING_QUERY_RESOURCE_FLAG, val) > +#define CXL_SET_HARD_SPARING(val) \ > + FIELD_PREP(CXL_SET_HARD_SPARING_FLAG, val) > +#define CXL_SET_SPARING_SUB_CHNL_VALID(val) \ > + FIELD_PREP(CXL_SPARING_SUB_CHNL_VALID_FLAG, val) > +#define CXL_SET_SPARING_NIB_MASK_VALID(val) \ > + FIELD_PREP(CXL_SPARING_NIB_MASK_VALID_FLAG, val) > + > +/* > + * See CXL spec rev 3.2 @8.2.10.7.2.3 Table 8-134 Memory Sparing Feature > + * Readable Attributes. > + */ > +struct cxl_memdev_repair_rd_attrbs_hdr { > + u8 max_op_latency; > + __le16 op_cap; > + __le16 op_mode; > + u8 op_class; > + u8 op_subclass; > + u8 rsvd[9]; > +} __packed; > + > +struct cxl_memdev_sparing_rd_attrbs { > + struct cxl_memdev_repair_rd_attrbs_hdr hdr; > + u8 rsvd; > + __le16 restriction_flags; > +} __packed; > + > +/* > + * See CXL spec rev 3.2 @8.2.10.7.1.4 Table 8-120 Memory Sparing Input Payload. > + */ > +struct cxl_memdev_sparing_in_payload { > + u8 flags; > + u8 channel; > + u8 rank; > + u8 nibble_mask[3]; > + u8 bank_group; > + u8 bank; > + u8 row[3]; > + __le16 column; > + u8 sub_channel; > +} __packed; > + > +static int > +cxl_mem_sparing_get_attrbs(struct cxl_mem_sparing_context *cxl_sparing_ctx) > +{ > + size_t rd_data_size = sizeof(struct cxl_memdev_sparing_rd_attrbs); > + struct cxl_memdev *cxlmd = cxl_sparing_ctx->cxlmd; > + struct cxl_mailbox *cxl_mbox = &cxlmd->cxlds->cxl_mbox; > + u16 restriction_flags; > + size_t data_size; > + u16 return_code; > + struct cxl_memdev_sparing_rd_attrbs *rd_attrbs __free(kfree) = > + kzalloc(rd_data_size, GFP_KERNEL); > + if (!rd_attrbs) > + return -ENOMEM; > + > + data_size = cxl_get_feature(cxl_mbox, &cxl_sparing_ctx->repair_uuid, > + CXL_GET_FEAT_SEL_CURRENT_VALUE, rd_attrbs, > + rd_data_size, 0, &return_code); > + if (!data_size) > + return -EIO; > + > + cxl_sparing_ctx->op_class = rd_attrbs->hdr.op_class; > + cxl_sparing_ctx->op_subclass = rd_attrbs->hdr.op_subclass; > + restriction_flags = le16_to_cpu(rd_attrbs->restriction_flags); > + cxl_sparing_ctx->cap_safe_when_in_use = > + CXL_GET_SPARING_SAFE_IN_USE(restriction_flags); > + cxl_sparing_ctx->cap_hard_sparing = > + CXL_GET_CAP_HARD_SPARING(restriction_flags); > + cxl_sparing_ctx->cap_soft_sparing = > + CXL_GET_CAP_SOFT_SPARING(restriction_flags); > + > + return 0; > +} > + > +static struct cxl_event_dram * > +cxl_mem_get_rec_dram(struct cxl_memdev *cxlmd, > + struct cxl_mem_sparing_context *ctx) > +{ > + struct cxl_mem_repair_attrbs attrbs = { 0 }; > + > + attrbs.dpa = ctx->dpa; > + attrbs.channel = ctx->channel; > + attrbs.rank = ctx->rank; > + attrbs.nibble_mask = ctx->nibble_mask; > + switch (ctx->repair_type) { > + case EDAC_REPAIR_CACHELINE_SPARING: > + attrbs.repair_type = CXL_CACHELINE_SPARING; > + attrbs.bank_group = ctx->bank_group; > + attrbs.bank = ctx->bank; > + attrbs.row = ctx->row; > + attrbs.column = ctx->column; > + attrbs.sub_channel = ctx->sub_channel; > + break; > + case EDAC_REPAIR_ROW_SPARING: > + attrbs.repair_type = CXL_ROW_SPARING; > + attrbs.bank_group = ctx->bank_group; > + attrbs.bank = ctx->bank; > + attrbs.row = ctx->row; > + break; > + case EDAC_REPAIR_BANK_SPARING: > + attrbs.repair_type = CXL_BANK_SPARING; > + attrbs.bank_group = ctx->bank_group; > + attrbs.bank = ctx->bank; > + break; > + case EDAC_REPAIR_RANK_SPARING: > + attrbs.repair_type = CXL_BANK_SPARING; > + break; > + default: > + return NULL; > + } > + > + return cxl_find_rec_dram(cxlmd, &attrbs); > +} > + > +static int > +cxl_mem_perform_sparing(struct device *dev, > + struct cxl_mem_sparing_context *cxl_sparing_ctx) > +{ > + struct cxl_memdev *cxlmd = cxl_sparing_ctx->cxlmd; > + struct cxl_memdev_sparing_in_payload sparing_pi; > + struct cxl_event_dram *rec = NULL; > + u16 validity_flags = 0; > + > + struct rw_semaphore *region_lock __free(rwsem_read_release) = > + rwsem_read_intr_acquire(&cxl_region_rwsem); > + if (!region_lock) > + return -EINTR; > + > + struct rw_semaphore *dpa_lock __free(rwsem_read_release) = > + rwsem_read_intr_acquire(&cxl_dpa_rwsem); > + if (!dpa_lock) > + return -EINTR; > + > + if (!cxl_sparing_ctx->cap_safe_when_in_use) { > + /* Memory to repair must be offline */ > + if (cxl_is_memdev_memory_online(cxlmd)) > + return -EBUSY; > + } else { > + if (cxl_is_memdev_memory_online(cxlmd)) { > + rec = cxl_mem_get_rec_dram(cxlmd, cxl_sparing_ctx); > + if (!rec) > + return -EINVAL; > + > + if (!rec->media_hdr.validity_flags) > + return -EINVAL; > + } > + } > + > + memset(&sparing_pi, 0, sizeof(sparing_pi)); > + sparing_pi.flags = CXL_SET_SPARING_QUERY_RESOURCE(0); > + if (cxl_sparing_ctx->persist_mode) > + sparing_pi.flags |= CXL_SET_HARD_SPARING(1); > + > + if (rec) > + validity_flags = get_unaligned_le16(rec->media_hdr.validity_flags); > + > + switch (cxl_sparing_ctx->repair_type) { > + case EDAC_REPAIR_CACHELINE_SPARING: > + sparing_pi.column = cpu_to_le16(cxl_sparing_ctx->column); > + if (!rec || (validity_flags & CXL_DER_VALID_SUB_CHANNEL)) { > + sparing_pi.flags |= CXL_SET_SPARING_SUB_CHNL_VALID(1); > + sparing_pi.sub_channel = cxl_sparing_ctx->sub_channel; > + } > + fallthrough; > + case EDAC_REPAIR_ROW_SPARING: > + put_unaligned_le24(cxl_sparing_ctx->row, sparing_pi.row); > + fallthrough; > + case EDAC_REPAIR_BANK_SPARING: > + sparing_pi.bank_group = cxl_sparing_ctx->bank_group; > + sparing_pi.bank = cxl_sparing_ctx->bank; > + fallthrough; > + case EDAC_REPAIR_RANK_SPARING: > + sparing_pi.rank = cxl_sparing_ctx->rank; > + fallthrough; > + default: > + sparing_pi.channel = cxl_sparing_ctx->channel; > + if ((rec && (validity_flags & CXL_DER_VALID_NIBBLE)) || > + (!rec && (!cxl_sparing_ctx->nibble_mask || > + (cxl_sparing_ctx->nibble_mask & 0xFFFFFF)))) { > + sparing_pi.flags |= CXL_SET_SPARING_NIB_MASK_VALID(1); > + put_unaligned_le24(cxl_sparing_ctx->nibble_mask, > + sparing_pi.nibble_mask); > + } > + break; > + } > + > + return cxl_perform_maintenance(&cxlmd->cxlds->cxl_mbox, > + cxl_sparing_ctx->op_class, > + cxl_sparing_ctx->op_subclass, > + &sparing_pi, sizeof(sparing_pi)); > +} > + > +static int cxl_mem_sparing_get_repair_type(struct device *dev, void *drv_data, > + const char **repair_type) > +{ > + struct cxl_mem_sparing_context *ctx = drv_data; > + > + switch (ctx->repair_type) { > + case EDAC_REPAIR_CACHELINE_SPARING: > + case EDAC_REPAIR_ROW_SPARING: > + case EDAC_REPAIR_BANK_SPARING: > + case EDAC_REPAIR_RANK_SPARING: > + *repair_type = edac_repair_type[ctx->repair_type]; > + break; > + default: > + return -EINVAL; > + } > + > + return 0; > +} > + > +#define CXL_SPARING_GET_ATTR(attrb, data_type) \ > + static int cxl_mem_sparing_get_##attrb( \ > + struct device *dev, void *drv_data, data_type *val) \ > + { \ > + struct cxl_mem_sparing_context *ctx = drv_data; \ > + \ > + *val = ctx->attrb; \ > + \ > + return 0; \ > + } > +CXL_SPARING_GET_ATTR(persist_mode, bool) > +CXL_SPARING_GET_ATTR(dpa, u64) > +CXL_SPARING_GET_ATTR(nibble_mask, u32) > +CXL_SPARING_GET_ATTR(bank_group, u32) > +CXL_SPARING_GET_ATTR(bank, u32) > +CXL_SPARING_GET_ATTR(rank, u32) > +CXL_SPARING_GET_ATTR(row, u32) > +CXL_SPARING_GET_ATTR(column, u32) > +CXL_SPARING_GET_ATTR(channel, u32) > +CXL_SPARING_GET_ATTR(sub_channel, u32) > + > +#define CXL_SPARING_SET_ATTR(attrb, data_type) \ > + static int cxl_mem_sparing_set_##attrb(struct device *dev, \ > + void *drv_data, data_type val) \ > + { \ > + struct cxl_mem_sparing_context *ctx = drv_data; \ > + \ > + ctx->attrb = val; \ > + \ > + return 0; \ > + } > +CXL_SPARING_SET_ATTR(nibble_mask, u32) > +CXL_SPARING_SET_ATTR(bank_group, u32) > +CXL_SPARING_SET_ATTR(bank, u32) > +CXL_SPARING_SET_ATTR(rank, u32) > +CXL_SPARING_SET_ATTR(row, u32) > +CXL_SPARING_SET_ATTR(column, u32) > +CXL_SPARING_SET_ATTR(channel, u32) > +CXL_SPARING_SET_ATTR(sub_channel, u32) > + > +static int cxl_mem_sparing_set_persist_mode(struct device *dev, void *drv_data, > + bool persist_mode) > +{ > + struct cxl_mem_sparing_context *ctx = drv_data; > + > + if ((persist_mode && ctx->cap_hard_sparing) || > + (!persist_mode && ctx->cap_soft_sparing)) > + ctx->persist_mode = persist_mode; > + else > + return -EOPNOTSUPP; > + > + return 0; > +} > + > +static int cxl_get_mem_sparing_safe_when_in_use(struct device *dev, > + void *drv_data, bool *safe) > +{ > + struct cxl_mem_sparing_context *ctx = drv_data; > + > + *safe = ctx->cap_safe_when_in_use; > + > + return 0; > +} > + > +static int cxl_mem_sparing_get_min_dpa(struct device *dev, void *drv_data, > + u64 *min_dpa) > +{ > + struct cxl_mem_sparing_context *ctx = drv_data; > + struct cxl_memdev *cxlmd = ctx->cxlmd; > + struct cxl_dev_state *cxlds = cxlmd->cxlds; > + > + *min_dpa = cxlds->dpa_res.start; > + > + return 0; > +} > + > +static int cxl_mem_sparing_get_max_dpa(struct device *dev, void *drv_data, > + u64 *max_dpa) > +{ > + struct cxl_mem_sparing_context *ctx = drv_data; > + struct cxl_memdev *cxlmd = ctx->cxlmd; > + struct cxl_dev_state *cxlds = cxlmd->cxlds; > + > + *max_dpa = cxlds->dpa_res.end; > + > + return 0; > +} > + > +static int cxl_mem_sparing_set_dpa(struct device *dev, void *drv_data, u64 dpa) > +{ > + struct cxl_mem_sparing_context *ctx = drv_data; > + struct cxl_memdev *cxlmd = ctx->cxlmd; > + struct cxl_dev_state *cxlds = cxlmd->cxlds; > + > + if (dpa < cxlds->dpa_res.start || dpa > cxlds->dpa_res.end) > + return -EINVAL; > + > + ctx->dpa = dpa; > + > + return 0; > +} > + > +static int cxl_do_mem_sparing(struct device *dev, void *drv_data, u32 val) > +{ > + struct cxl_mem_sparing_context *ctx = drv_data; > + > + if (val != EDAC_DO_MEM_REPAIR) > + return -EINVAL; > + > + return cxl_mem_perform_sparing(dev, ctx); > +} > + > +#define RANK_OPS \ > + .get_repair_type = cxl_mem_sparing_get_repair_type, \ > + .get_persist_mode = cxl_mem_sparing_get_persist_mode, \ > + .set_persist_mode = cxl_mem_sparing_set_persist_mode, \ > + .get_repair_safe_when_in_use = cxl_get_mem_sparing_safe_when_in_use, \ > + .get_min_dpa = cxl_mem_sparing_get_min_dpa, \ > + .get_max_dpa = cxl_mem_sparing_get_max_dpa, \ > + .get_dpa = cxl_mem_sparing_get_dpa, \ > + .set_dpa = cxl_mem_sparing_set_dpa, \ > + .get_nibble_mask = cxl_mem_sparing_get_nibble_mask, \ > + .set_nibble_mask = cxl_mem_sparing_set_nibble_mask, \ > + .get_rank = cxl_mem_sparing_get_rank, \ > + .set_rank = cxl_mem_sparing_set_rank, \ > + .get_channel = cxl_mem_sparing_get_channel, \ > + .set_channel = cxl_mem_sparing_set_channel, \ > + .do_repair = cxl_do_mem_sparing > + > +#define BANK_OPS \ > + RANK_OPS, .get_bank_group = cxl_mem_sparing_get_bank_group, \ > + .set_bank_group = cxl_mem_sparing_set_bank_group, \ > + .get_bank = cxl_mem_sparing_get_bank, \ > + .set_bank = cxl_mem_sparing_set_bank > + > +#define ROW_OPS \ > + BANK_OPS, .get_row = cxl_mem_sparing_get_row, \ > + .set_row = cxl_mem_sparing_set_row > + > +#define CACHELINE_OPS \ > + ROW_OPS, .get_column = cxl_mem_sparing_get_column, \ > + .set_column = cxl_mem_sparing_set_column, \ > + .get_sub_channel = cxl_mem_sparing_get_sub_channel, \ > + .set_sub_channel = cxl_mem_sparing_set_sub_channel > + > +static const struct edac_mem_repair_ops cxl_rank_sparing_ops = { > + RANK_OPS, > +}; > + > +static const struct edac_mem_repair_ops cxl_bank_sparing_ops = { > + BANK_OPS, > +}; > + > +static const struct edac_mem_repair_ops cxl_row_sparing_ops = { > + ROW_OPS, > +}; > + > +static const struct edac_mem_repair_ops cxl_cacheline_sparing_ops = { > + CACHELINE_OPS, > +}; > + > +struct cxl_mem_sparing_desc { > + const uuid_t repair_uuid; > + enum edac_mem_repair_type repair_type; > + const struct edac_mem_repair_ops *repair_ops; > +}; > + > +static const struct cxl_mem_sparing_desc mem_sparing_desc[] = { > + { > + .repair_uuid = CXL_FEAT_CACHELINE_SPARING_UUID, > + .repair_type = EDAC_REPAIR_CACHELINE_SPARING, > + .repair_ops = &cxl_cacheline_sparing_ops, > + }, > + { > + .repair_uuid = CXL_FEAT_ROW_SPARING_UUID, > + .repair_type = EDAC_REPAIR_ROW_SPARING, > + .repair_ops = &cxl_row_sparing_ops, > + }, > + { > + .repair_uuid = CXL_FEAT_BANK_SPARING_UUID, > + .repair_type = EDAC_REPAIR_BANK_SPARING, > + .repair_ops = &cxl_bank_sparing_ops, > + }, > + { > + .repair_uuid = CXL_FEAT_RANK_SPARING_UUID, > + .repair_type = EDAC_REPAIR_RANK_SPARING, > + .repair_ops = &cxl_rank_sparing_ops, > + }, > +}; > + > +static int cxl_memdev_sparing_init(struct cxl_memdev *cxlmd, > + struct edac_dev_feature *ras_feature, > + const struct cxl_mem_sparing_desc *desc, > + u8 repair_inst) > +{ > + struct cxl_mem_sparing_context *cxl_sparing_ctx; > + struct cxl_feat_entry *feat_entry; > + int ret; > + > + feat_entry = cxl_feature_info(to_cxlfs(cxlmd->cxlds), > + &desc->repair_uuid); > + if (!feat_entry) > + return -EOPNOTSUPP; > + > + if (!(le32_to_cpu(feat_entry->flags) & CXL_FEATURE_F_CHANGEABLE)) > + return -EOPNOTSUPP; > + > + cxl_sparing_ctx = devm_kzalloc(&cxlmd->dev, sizeof(*cxl_sparing_ctx), > + GFP_KERNEL); > + if (!cxl_sparing_ctx) > + return -ENOMEM; > + > + *cxl_sparing_ctx = (struct cxl_mem_sparing_context){ > + .get_feat_size = le16_to_cpu(feat_entry->get_feat_size), > + .set_feat_size = le16_to_cpu(feat_entry->set_feat_size), > + .get_version = feat_entry->get_feat_ver, > + .set_version = feat_entry->set_feat_ver, > + .effects = le16_to_cpu(feat_entry->effects), > + .cxlmd = cxlmd, > + .repair_type = desc->repair_type, > + .instance = repair_inst++, > + }; > + uuid_copy(&cxl_sparing_ctx->repair_uuid, &desc->repair_uuid); > + > + ret = cxl_mem_sparing_get_attrbs(cxl_sparing_ctx); > + if (ret) > + return ret; > + > + if ((cxl_sparing_ctx->cap_soft_sparing && > + cxl_sparing_ctx->cap_hard_sparing) || > + cxl_sparing_ctx->cap_soft_sparing) > + cxl_sparing_ctx->persist_mode = 0; > + else if (cxl_sparing_ctx->cap_hard_sparing) > + cxl_sparing_ctx->persist_mode = 1; > + else > + return -EOPNOTSUPP; > + > + ras_feature->ft_type = RAS_FEAT_MEM_REPAIR; > + ras_feature->instance = cxl_sparing_ctx->instance; > + ras_feature->mem_repair_ops = desc->repair_ops; > + ras_feature->ctx = cxl_sparing_ctx; > + > + return 0; > +} > #endif /* CONFIG_CXL_EDAC_MEM_REPAIR */ > > int devm_cxl_memdev_edac_register(struct cxl_memdev *cxlmd) > { > struct edac_dev_feature ras_features[CXL_NR_EDAC_DEV_FEATURES]; > int num_ras_features = 0; > -#if defined(CONFIG_CXL_EDAC_SCRUB) || defined(CONFIG_CXL_EDAC_ECS) > +#ifdef CONFIG_CXL_EDAC_MEM_REPAIR > + u8 repair_inst = 0; > +#endif > +#if defined(CONFIG_CXL_EDAC_SCRUB) || defined(CONFIG_CXL_EDAC_ECS) || \ > + defined(CXL_EDAC_MEM_REPAIR) > int rc; > #endif > > @@ -1153,6 +1684,19 @@ int devm_cxl_memdev_edac_register(struct cxl_memdev *cxlmd) > cxlmd->array_err_rec = array_rec; > xa_init(&array_rec->rec_gen_media); > xa_init(&array_rec->rec_dram); > + > + for (int i = 0; i < CXL_MEM_SPARING_MAX; i++) { > + rc = cxl_memdev_sparing_init(cxlmd, > + &ras_features[num_ras_features], > + &mem_sparing_desc[i], repair_inst); > + if (rc == -EOPNOTSUPP) > + continue; > + if (rc < 0) > + return rc; > + > + repair_inst++; > + num_ras_features++; > + } > #endif > char *cxl_dev_name __free(kfree) = > kasprintf(GFP_KERNEL, "cxl_%s", dev_name(&cxlmd->dev)); > diff --git a/drivers/edac/mem_repair.c b/drivers/edac/mem_repair.c > index 3b1a845457b0..d1a8caa85369 100755 > --- a/drivers/edac/mem_repair.c > +++ b/drivers/edac/mem_repair.c > @@ -45,6 +45,15 @@ struct edac_mem_repair_context { > struct attribute_group group; > }; > > +const char * const edac_repair_type[] = { > + [EDAC_REPAIR_PPR] = "ppr", > + [EDAC_REPAIR_CACHELINE_SPARING] = "cacheline-sparing", > + [EDAC_REPAIR_ROW_SPARING] = "row-sparing", > + [EDAC_REPAIR_BANK_SPARING] = "bank-sparing", > + [EDAC_REPAIR_RANK_SPARING] = "rank-sparing", > +}; > +EXPORT_SYMBOL_GPL(edac_repair_type); > + > #define TO_MR_DEV_ATTR(_dev_attr) \ > container_of(_dev_attr, struct edac_mem_repair_dev_attr, dev_attr) > > diff --git a/include/linux/edac.h b/include/linux/edac.h > index 451f9c152c99..fa32f2aca22f 100644 > --- a/include/linux/edac.h > +++ b/include/linux/edac.h > @@ -745,9 +745,16 @@ static inline int edac_ecs_get_desc(struct device *ecs_dev, > #endif /* CONFIG_EDAC_ECS */ > > enum edac_mem_repair_type { > + EDAC_REPAIR_PPR, > + EDAC_REPAIR_CACHELINE_SPARING, > + EDAC_REPAIR_ROW_SPARING, > + EDAC_REPAIR_BANK_SPARING, > + EDAC_REPAIR_RANK_SPARING, > EDAC_REPAIR_MAX > }; > > +extern const char * const edac_repair_type[]; > + > enum edac_mem_repair_cmd { > EDAC_DO_MEM_REPAIR = 1, > };