cxl/mbox: Move mailbox related driver state to its own data structure

'struct cxl_dev_state' makes too many assumptions about the capabilities
of a CXL device. In particular it assumes a CXL device has a mailbox and
all of the infrastructure and state that comes along with that.

In preparation for supporting accelerator / Type-2 devices that may not
have a mailbox and in general maintain a minimal core context structure,
make mailbox functionality a super-set of  'struct cxl_dev_state' with
'struct cxl_memdev_state'.

With this reorganization it allows for CXL devices that support HDM
decoder mapping, but not other general-expander / Type-3 capabilities,
to only enable that subset without the rest of the mailbox
infrastructure coming along for the ride.

Reviewed-by: Jonathan Cameron <Jonathan.Cameron@huawei.com>
Reviewed-by: Dave Jiang <dave.jiang@intel.com>
Link: https://lore.kernel.org/r/168679260240.3436160.15520641540463704524.stgit@dwillia2-xfh.jf.intel.com
Signed-off-by: Dan Williams <dan.j.williams@intel.com>
This commit is contained in:
Dan Williams 2023-06-14 18:30:02 -07:00
parent 3fe7feb0f3
commit 59f8d15107
8 changed files with 336 additions and 290 deletions

View File

@ -182,7 +182,7 @@ static const char *cxl_mem_opcode_to_name(u16 opcode)
/** /**
* cxl_internal_send_cmd() - Kernel internal interface to send a mailbox command * cxl_internal_send_cmd() - Kernel internal interface to send a mailbox command
* @cxlds: The device data for the operation * @mds: The driver data for the operation
* @mbox_cmd: initialized command to execute * @mbox_cmd: initialized command to execute
* *
* Context: Any context. * Context: Any context.
@ -198,19 +198,19 @@ static const char *cxl_mem_opcode_to_name(u16 opcode)
* error. While this distinction can be useful for commands from userspace, the * error. While this distinction can be useful for commands from userspace, the
* kernel will only be able to use results when both are successful. * kernel will only be able to use results when both are successful.
*/ */
int cxl_internal_send_cmd(struct cxl_dev_state *cxlds, int cxl_internal_send_cmd(struct cxl_memdev_state *mds,
struct cxl_mbox_cmd *mbox_cmd) struct cxl_mbox_cmd *mbox_cmd)
{ {
size_t out_size, min_out; size_t out_size, min_out;
int rc; int rc;
if (mbox_cmd->size_in > cxlds->payload_size || if (mbox_cmd->size_in > mds->payload_size ||
mbox_cmd->size_out > cxlds->payload_size) mbox_cmd->size_out > mds->payload_size)
return -E2BIG; return -E2BIG;
out_size = mbox_cmd->size_out; out_size = mbox_cmd->size_out;
min_out = mbox_cmd->min_out; min_out = mbox_cmd->min_out;
rc = cxlds->mbox_send(cxlds, mbox_cmd); rc = mds->mbox_send(mds, mbox_cmd);
/* /*
* EIO is reserved for a payload size mismatch and mbox_send() * EIO is reserved for a payload size mismatch and mbox_send()
* may not return this error. * may not return this error.
@ -297,7 +297,7 @@ static bool cxl_payload_from_user_allowed(u16 opcode, void *payload_in)
} }
static int cxl_mbox_cmd_ctor(struct cxl_mbox_cmd *mbox, static int cxl_mbox_cmd_ctor(struct cxl_mbox_cmd *mbox,
struct cxl_dev_state *cxlds, u16 opcode, struct cxl_memdev_state *mds, u16 opcode,
size_t in_size, size_t out_size, u64 in_payload) size_t in_size, size_t out_size, u64 in_payload)
{ {
*mbox = (struct cxl_mbox_cmd) { *mbox = (struct cxl_mbox_cmd) {
@ -312,7 +312,7 @@ static int cxl_mbox_cmd_ctor(struct cxl_mbox_cmd *mbox,
return PTR_ERR(mbox->payload_in); return PTR_ERR(mbox->payload_in);
if (!cxl_payload_from_user_allowed(opcode, mbox->payload_in)) { if (!cxl_payload_from_user_allowed(opcode, mbox->payload_in)) {
dev_dbg(cxlds->dev, "%s: input payload not allowed\n", dev_dbg(mds->cxlds.dev, "%s: input payload not allowed\n",
cxl_mem_opcode_to_name(opcode)); cxl_mem_opcode_to_name(opcode));
kvfree(mbox->payload_in); kvfree(mbox->payload_in);
return -EBUSY; return -EBUSY;
@ -321,7 +321,7 @@ static int cxl_mbox_cmd_ctor(struct cxl_mbox_cmd *mbox,
/* Prepare to handle a full payload for variable sized output */ /* Prepare to handle a full payload for variable sized output */
if (out_size == CXL_VARIABLE_PAYLOAD) if (out_size == CXL_VARIABLE_PAYLOAD)
mbox->size_out = cxlds->payload_size; mbox->size_out = mds->payload_size;
else else
mbox->size_out = out_size; mbox->size_out = out_size;
@ -343,7 +343,7 @@ static void cxl_mbox_cmd_dtor(struct cxl_mbox_cmd *mbox)
static int cxl_to_mem_cmd_raw(struct cxl_mem_command *mem_cmd, static int cxl_to_mem_cmd_raw(struct cxl_mem_command *mem_cmd,
const struct cxl_send_command *send_cmd, const struct cxl_send_command *send_cmd,
struct cxl_dev_state *cxlds) struct cxl_memdev_state *mds)
{ {
if (send_cmd->raw.rsvd) if (send_cmd->raw.rsvd)
return -EINVAL; return -EINVAL;
@ -353,13 +353,13 @@ static int cxl_to_mem_cmd_raw(struct cxl_mem_command *mem_cmd,
* gets passed along without further checking, so it must be * gets passed along without further checking, so it must be
* validated here. * validated here.
*/ */
if (send_cmd->out.size > cxlds->payload_size) if (send_cmd->out.size > mds->payload_size)
return -EINVAL; return -EINVAL;
if (!cxl_mem_raw_command_allowed(send_cmd->raw.opcode)) if (!cxl_mem_raw_command_allowed(send_cmd->raw.opcode))
return -EPERM; return -EPERM;
dev_WARN_ONCE(cxlds->dev, true, "raw command path used\n"); dev_WARN_ONCE(mds->cxlds.dev, true, "raw command path used\n");
*mem_cmd = (struct cxl_mem_command) { *mem_cmd = (struct cxl_mem_command) {
.info = { .info = {
@ -375,7 +375,7 @@ static int cxl_to_mem_cmd_raw(struct cxl_mem_command *mem_cmd,
static int cxl_to_mem_cmd(struct cxl_mem_command *mem_cmd, static int cxl_to_mem_cmd(struct cxl_mem_command *mem_cmd,
const struct cxl_send_command *send_cmd, const struct cxl_send_command *send_cmd,
struct cxl_dev_state *cxlds) struct cxl_memdev_state *mds)
{ {
struct cxl_mem_command *c = &cxl_mem_commands[send_cmd->id]; struct cxl_mem_command *c = &cxl_mem_commands[send_cmd->id];
const struct cxl_command_info *info = &c->info; const struct cxl_command_info *info = &c->info;
@ -390,11 +390,11 @@ static int cxl_to_mem_cmd(struct cxl_mem_command *mem_cmd,
return -EINVAL; return -EINVAL;
/* Check that the command is enabled for hardware */ /* Check that the command is enabled for hardware */
if (!test_bit(info->id, cxlds->enabled_cmds)) if (!test_bit(info->id, mds->enabled_cmds))
return -ENOTTY; return -ENOTTY;
/* Check that the command is not claimed for exclusive kernel use */ /* Check that the command is not claimed for exclusive kernel use */
if (test_bit(info->id, cxlds->exclusive_cmds)) if (test_bit(info->id, mds->exclusive_cmds))
return -EBUSY; return -EBUSY;
/* Check the input buffer is the expected size */ /* Check the input buffer is the expected size */
@ -423,7 +423,7 @@ static int cxl_to_mem_cmd(struct cxl_mem_command *mem_cmd,
/** /**
* cxl_validate_cmd_from_user() - Check fields for CXL_MEM_SEND_COMMAND. * cxl_validate_cmd_from_user() - Check fields for CXL_MEM_SEND_COMMAND.
* @mbox_cmd: Sanitized and populated &struct cxl_mbox_cmd. * @mbox_cmd: Sanitized and populated &struct cxl_mbox_cmd.
* @cxlds: The device data for the operation * @mds: The driver data for the operation
* @send_cmd: &struct cxl_send_command copied in from userspace. * @send_cmd: &struct cxl_send_command copied in from userspace.
* *
* Return: * Return:
@ -438,7 +438,7 @@ static int cxl_to_mem_cmd(struct cxl_mem_command *mem_cmd,
* safe to send to the hardware. * safe to send to the hardware.
*/ */
static int cxl_validate_cmd_from_user(struct cxl_mbox_cmd *mbox_cmd, static int cxl_validate_cmd_from_user(struct cxl_mbox_cmd *mbox_cmd,
struct cxl_dev_state *cxlds, struct cxl_memdev_state *mds,
const struct cxl_send_command *send_cmd) const struct cxl_send_command *send_cmd)
{ {
struct cxl_mem_command mem_cmd; struct cxl_mem_command mem_cmd;
@ -452,20 +452,20 @@ static int cxl_validate_cmd_from_user(struct cxl_mbox_cmd *mbox_cmd,
* supports, but output can be arbitrarily large (simply write out as * supports, but output can be arbitrarily large (simply write out as
* much data as the hardware provides). * much data as the hardware provides).
*/ */
if (send_cmd->in.size > cxlds->payload_size) if (send_cmd->in.size > mds->payload_size)
return -EINVAL; return -EINVAL;
/* Sanitize and construct a cxl_mem_command */ /* Sanitize and construct a cxl_mem_command */
if (send_cmd->id == CXL_MEM_COMMAND_ID_RAW) if (send_cmd->id == CXL_MEM_COMMAND_ID_RAW)
rc = cxl_to_mem_cmd_raw(&mem_cmd, send_cmd, cxlds); rc = cxl_to_mem_cmd_raw(&mem_cmd, send_cmd, mds);
else else
rc = cxl_to_mem_cmd(&mem_cmd, send_cmd, cxlds); rc = cxl_to_mem_cmd(&mem_cmd, send_cmd, mds);
if (rc) if (rc)
return rc; return rc;
/* Sanitize and construct a cxl_mbox_cmd */ /* Sanitize and construct a cxl_mbox_cmd */
return cxl_mbox_cmd_ctor(mbox_cmd, cxlds, mem_cmd.opcode, return cxl_mbox_cmd_ctor(mbox_cmd, mds, mem_cmd.opcode,
mem_cmd.info.size_in, mem_cmd.info.size_out, mem_cmd.info.size_in, mem_cmd.info.size_out,
send_cmd->in.payload); send_cmd->in.payload);
} }
@ -473,6 +473,7 @@ static int cxl_validate_cmd_from_user(struct cxl_mbox_cmd *mbox_cmd,
int cxl_query_cmd(struct cxl_memdev *cxlmd, int cxl_query_cmd(struct cxl_memdev *cxlmd,
struct cxl_mem_query_commands __user *q) struct cxl_mem_query_commands __user *q)
{ {
struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
struct device *dev = &cxlmd->dev; struct device *dev = &cxlmd->dev;
struct cxl_mem_command *cmd; struct cxl_mem_command *cmd;
u32 n_commands; u32 n_commands;
@ -494,9 +495,9 @@ int cxl_query_cmd(struct cxl_memdev *cxlmd,
cxl_for_each_cmd(cmd) { cxl_for_each_cmd(cmd) {
struct cxl_command_info info = cmd->info; struct cxl_command_info info = cmd->info;
if (test_bit(info.id, cxlmd->cxlds->enabled_cmds)) if (test_bit(info.id, mds->enabled_cmds))
info.flags |= CXL_MEM_COMMAND_FLAG_ENABLED; info.flags |= CXL_MEM_COMMAND_FLAG_ENABLED;
if (test_bit(info.id, cxlmd->cxlds->exclusive_cmds)) if (test_bit(info.id, mds->exclusive_cmds))
info.flags |= CXL_MEM_COMMAND_FLAG_EXCLUSIVE; info.flags |= CXL_MEM_COMMAND_FLAG_EXCLUSIVE;
if (copy_to_user(&q->commands[j++], &info, sizeof(info))) if (copy_to_user(&q->commands[j++], &info, sizeof(info)))
@ -511,7 +512,7 @@ int cxl_query_cmd(struct cxl_memdev *cxlmd,
/** /**
* handle_mailbox_cmd_from_user() - Dispatch a mailbox command for userspace. * handle_mailbox_cmd_from_user() - Dispatch a mailbox command for userspace.
* @cxlds: The device data for the operation * @mds: The driver data for the operation
* @mbox_cmd: The validated mailbox command. * @mbox_cmd: The validated mailbox command.
* @out_payload: Pointer to userspace's output payload. * @out_payload: Pointer to userspace's output payload.
* @size_out: (Input) Max payload size to copy out. * @size_out: (Input) Max payload size to copy out.
@ -532,12 +533,12 @@ int cxl_query_cmd(struct cxl_memdev *cxlmd,
* *
* See cxl_send_cmd(). * See cxl_send_cmd().
*/ */
static int handle_mailbox_cmd_from_user(struct cxl_dev_state *cxlds, static int handle_mailbox_cmd_from_user(struct cxl_memdev_state *mds,
struct cxl_mbox_cmd *mbox_cmd, struct cxl_mbox_cmd *mbox_cmd,
u64 out_payload, s32 *size_out, u64 out_payload, s32 *size_out,
u32 *retval) u32 *retval)
{ {
struct device *dev = cxlds->dev; struct device *dev = mds->cxlds.dev;
int rc; int rc;
dev_dbg(dev, dev_dbg(dev,
@ -547,7 +548,7 @@ static int handle_mailbox_cmd_from_user(struct cxl_dev_state *cxlds,
cxl_mem_opcode_to_name(mbox_cmd->opcode), cxl_mem_opcode_to_name(mbox_cmd->opcode),
mbox_cmd->opcode, mbox_cmd->size_in); mbox_cmd->opcode, mbox_cmd->size_in);
rc = cxlds->mbox_send(cxlds, mbox_cmd); rc = mds->mbox_send(mds, mbox_cmd);
if (rc) if (rc)
goto out; goto out;
@ -576,7 +577,7 @@ out:
int cxl_send_cmd(struct cxl_memdev *cxlmd, struct cxl_send_command __user *s) int cxl_send_cmd(struct cxl_memdev *cxlmd, struct cxl_send_command __user *s)
{ {
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
struct device *dev = &cxlmd->dev; struct device *dev = &cxlmd->dev;
struct cxl_send_command send; struct cxl_send_command send;
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
@ -587,11 +588,11 @@ int cxl_send_cmd(struct cxl_memdev *cxlmd, struct cxl_send_command __user *s)
if (copy_from_user(&send, s, sizeof(send))) if (copy_from_user(&send, s, sizeof(send)))
return -EFAULT; return -EFAULT;
rc = cxl_validate_cmd_from_user(&mbox_cmd, cxlmd->cxlds, &send); rc = cxl_validate_cmd_from_user(&mbox_cmd, mds, &send);
if (rc) if (rc)
return rc; return rc;
rc = handle_mailbox_cmd_from_user(cxlds, &mbox_cmd, send.out.payload, rc = handle_mailbox_cmd_from_user(mds, &mbox_cmd, send.out.payload,
&send.out.size, &send.retval); &send.out.size, &send.retval);
if (rc) if (rc)
return rc; return rc;
@ -602,13 +603,14 @@ int cxl_send_cmd(struct cxl_memdev *cxlmd, struct cxl_send_command __user *s)
return 0; return 0;
} }
static int cxl_xfer_log(struct cxl_dev_state *cxlds, uuid_t *uuid, u32 *size, u8 *out) static int cxl_xfer_log(struct cxl_memdev_state *mds, uuid_t *uuid,
u32 *size, u8 *out)
{ {
u32 remaining = *size; u32 remaining = *size;
u32 offset = 0; u32 offset = 0;
while (remaining) { while (remaining) {
u32 xfer_size = min_t(u32, remaining, cxlds->payload_size); u32 xfer_size = min_t(u32, remaining, mds->payload_size);
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
struct cxl_mbox_get_log log; struct cxl_mbox_get_log log;
int rc; int rc;
@ -627,7 +629,7 @@ static int cxl_xfer_log(struct cxl_dev_state *cxlds, uuid_t *uuid, u32 *size, u8
.payload_out = out, .payload_out = out,
}; };
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
/* /*
* The output payload length that indicates the number * The output payload length that indicates the number
@ -654,17 +656,18 @@ static int cxl_xfer_log(struct cxl_dev_state *cxlds, uuid_t *uuid, u32 *size, u8
/** /**
* cxl_walk_cel() - Walk through the Command Effects Log. * cxl_walk_cel() - Walk through the Command Effects Log.
* @cxlds: The device data for the operation * @mds: The driver data for the operation
* @size: Length of the Command Effects Log. * @size: Length of the Command Effects Log.
* @cel: CEL * @cel: CEL
* *
* Iterate over each entry in the CEL and determine if the driver supports the * Iterate over each entry in the CEL and determine if the driver supports the
* command. If so, the command is enabled for the device and can be used later. * command. If so, the command is enabled for the device and can be used later.
*/ */
static void cxl_walk_cel(struct cxl_dev_state *cxlds, size_t size, u8 *cel) static void cxl_walk_cel(struct cxl_memdev_state *mds, size_t size, u8 *cel)
{ {
struct cxl_cel_entry *cel_entry; struct cxl_cel_entry *cel_entry;
const int cel_entries = size / sizeof(*cel_entry); const int cel_entries = size / sizeof(*cel_entry);
struct device *dev = mds->cxlds.dev;
int i; int i;
cel_entry = (struct cxl_cel_entry *) cel; cel_entry = (struct cxl_cel_entry *) cel;
@ -674,39 +677,39 @@ static void cxl_walk_cel(struct cxl_dev_state *cxlds, size_t size, u8 *cel)
struct cxl_mem_command *cmd = cxl_mem_find_command(opcode); struct cxl_mem_command *cmd = cxl_mem_find_command(opcode);
if (!cmd && !cxl_is_poison_command(opcode)) { if (!cmd && !cxl_is_poison_command(opcode)) {
dev_dbg(cxlds->dev, dev_dbg(dev,
"Opcode 0x%04x unsupported by driver\n", opcode); "Opcode 0x%04x unsupported by driver\n", opcode);
continue; continue;
} }
if (cmd) if (cmd)
set_bit(cmd->info.id, cxlds->enabled_cmds); set_bit(cmd->info.id, mds->enabled_cmds);
if (cxl_is_poison_command(opcode)) if (cxl_is_poison_command(opcode))
cxl_set_poison_cmd_enabled(&cxlds->poison, opcode); cxl_set_poison_cmd_enabled(&mds->poison, opcode);
dev_dbg(cxlds->dev, "Opcode 0x%04x enabled\n", opcode); dev_dbg(dev, "Opcode 0x%04x enabled\n", opcode);
} }
} }
static struct cxl_mbox_get_supported_logs *cxl_get_gsl(struct cxl_dev_state *cxlds) static struct cxl_mbox_get_supported_logs *cxl_get_gsl(struct cxl_memdev_state *mds)
{ {
struct cxl_mbox_get_supported_logs *ret; struct cxl_mbox_get_supported_logs *ret;
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
int rc; int rc;
ret = kvmalloc(cxlds->payload_size, GFP_KERNEL); ret = kvmalloc(mds->payload_size, GFP_KERNEL);
if (!ret) if (!ret)
return ERR_PTR(-ENOMEM); return ERR_PTR(-ENOMEM);
mbox_cmd = (struct cxl_mbox_cmd) { mbox_cmd = (struct cxl_mbox_cmd) {
.opcode = CXL_MBOX_OP_GET_SUPPORTED_LOGS, .opcode = CXL_MBOX_OP_GET_SUPPORTED_LOGS,
.size_out = cxlds->payload_size, .size_out = mds->payload_size,
.payload_out = ret, .payload_out = ret,
/* At least the record number field must be valid */ /* At least the record number field must be valid */
.min_out = 2, .min_out = 2,
}; };
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc < 0) { if (rc < 0) {
kvfree(ret); kvfree(ret);
return ERR_PTR(rc); return ERR_PTR(rc);
@ -729,22 +732,22 @@ static const uuid_t log_uuid[] = {
/** /**
* cxl_enumerate_cmds() - Enumerate commands for a device. * cxl_enumerate_cmds() - Enumerate commands for a device.
* @cxlds: The device data for the operation * @mds: The driver data for the operation
* *
* Returns 0 if enumerate completed successfully. * Returns 0 if enumerate completed successfully.
* *
* CXL devices have optional support for certain commands. This function will * CXL devices have optional support for certain commands. This function will
* determine the set of supported commands for the hardware and update the * determine the set of supported commands for the hardware and update the
* enabled_cmds bitmap in the @cxlds. * enabled_cmds bitmap in the @mds.
*/ */
int cxl_enumerate_cmds(struct cxl_dev_state *cxlds) int cxl_enumerate_cmds(struct cxl_memdev_state *mds)
{ {
struct cxl_mbox_get_supported_logs *gsl; struct cxl_mbox_get_supported_logs *gsl;
struct device *dev = cxlds->dev; struct device *dev = mds->cxlds.dev;
struct cxl_mem_command *cmd; struct cxl_mem_command *cmd;
int i, rc; int i, rc;
gsl = cxl_get_gsl(cxlds); gsl = cxl_get_gsl(mds);
if (IS_ERR(gsl)) if (IS_ERR(gsl))
return PTR_ERR(gsl); return PTR_ERR(gsl);
@ -765,19 +768,19 @@ int cxl_enumerate_cmds(struct cxl_dev_state *cxlds)
goto out; goto out;
} }
rc = cxl_xfer_log(cxlds, &uuid, &size, log); rc = cxl_xfer_log(mds, &uuid, &size, log);
if (rc) { if (rc) {
kvfree(log); kvfree(log);
goto out; goto out;
} }
cxl_walk_cel(cxlds, size, log); cxl_walk_cel(mds, size, log);
kvfree(log); kvfree(log);
/* In case CEL was bogus, enable some default commands. */ /* In case CEL was bogus, enable some default commands. */
cxl_for_each_cmd(cmd) cxl_for_each_cmd(cmd)
if (cmd->flags & CXL_CMD_FLAG_FORCE_ENABLE) if (cmd->flags & CXL_CMD_FLAG_FORCE_ENABLE)
set_bit(cmd->info.id, cxlds->enabled_cmds); set_bit(cmd->info.id, mds->enabled_cmds);
/* Found the required CEL */ /* Found the required CEL */
rc = 0; rc = 0;
@ -838,7 +841,7 @@ static void cxl_event_trace_record(const struct cxl_memdev *cxlmd,
} }
} }
static int cxl_clear_event_record(struct cxl_dev_state *cxlds, static int cxl_clear_event_record(struct cxl_memdev_state *mds,
enum cxl_event_log_type log, enum cxl_event_log_type log,
struct cxl_get_event_payload *get_pl) struct cxl_get_event_payload *get_pl)
{ {
@ -852,9 +855,9 @@ static int cxl_clear_event_record(struct cxl_dev_state *cxlds,
int i; int i;
/* Payload size may limit the max handles */ /* Payload size may limit the max handles */
if (pl_size > cxlds->payload_size) { if (pl_size > mds->payload_size) {
max_handles = (cxlds->payload_size - sizeof(*payload)) / max_handles = (mds->payload_size - sizeof(*payload)) /
sizeof(__le16); sizeof(__le16);
pl_size = struct_size(payload, handles, max_handles); pl_size = struct_size(payload, handles, max_handles);
} }
@ -879,12 +882,12 @@ static int cxl_clear_event_record(struct cxl_dev_state *cxlds,
i = 0; i = 0;
for (cnt = 0; cnt < total; cnt++) { for (cnt = 0; cnt < total; cnt++) {
payload->handles[i++] = get_pl->records[cnt].hdr.handle; payload->handles[i++] = get_pl->records[cnt].hdr.handle;
dev_dbg(cxlds->dev, "Event log '%d': Clearing %u\n", dev_dbg(mds->cxlds.dev, "Event log '%d': Clearing %u\n", log,
log, le16_to_cpu(payload->handles[i])); le16_to_cpu(payload->handles[i]));
if (i == max_handles) { if (i == max_handles) {
payload->nr_recs = i; payload->nr_recs = i;
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc) if (rc)
goto free_pl; goto free_pl;
i = 0; i = 0;
@ -895,7 +898,7 @@ static int cxl_clear_event_record(struct cxl_dev_state *cxlds,
if (i) { if (i) {
payload->nr_recs = i; payload->nr_recs = i;
mbox_cmd.size_in = struct_size(payload, handles, i); mbox_cmd.size_in = struct_size(payload, handles, i);
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc) if (rc)
goto free_pl; goto free_pl;
} }
@ -905,32 +908,34 @@ free_pl:
return rc; return rc;
} }
static void cxl_mem_get_records_log(struct cxl_dev_state *cxlds, static void cxl_mem_get_records_log(struct cxl_memdev_state *mds,
enum cxl_event_log_type type) enum cxl_event_log_type type)
{ {
struct cxl_memdev *cxlmd = mds->cxlds.cxlmd;
struct device *dev = mds->cxlds.dev;
struct cxl_get_event_payload *payload; struct cxl_get_event_payload *payload;
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
u8 log_type = type; u8 log_type = type;
u16 nr_rec; u16 nr_rec;
mutex_lock(&cxlds->event.log_lock); mutex_lock(&mds->event.log_lock);
payload = cxlds->event.buf; payload = mds->event.buf;
mbox_cmd = (struct cxl_mbox_cmd) { mbox_cmd = (struct cxl_mbox_cmd) {
.opcode = CXL_MBOX_OP_GET_EVENT_RECORD, .opcode = CXL_MBOX_OP_GET_EVENT_RECORD,
.payload_in = &log_type, .payload_in = &log_type,
.size_in = sizeof(log_type), .size_in = sizeof(log_type),
.payload_out = payload, .payload_out = payload,
.size_out = cxlds->payload_size, .size_out = mds->payload_size,
.min_out = struct_size(payload, records, 0), .min_out = struct_size(payload, records, 0),
}; };
do { do {
int rc, i; int rc, i;
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc) { if (rc) {
dev_err_ratelimited(cxlds->dev, dev_err_ratelimited(dev,
"Event log '%d': Failed to query event records : %d", "Event log '%d': Failed to query event records : %d",
type, rc); type, rc);
break; break;
@ -941,27 +946,27 @@ static void cxl_mem_get_records_log(struct cxl_dev_state *cxlds,
break; break;
for (i = 0; i < nr_rec; i++) for (i = 0; i < nr_rec; i++)
cxl_event_trace_record(cxlds->cxlmd, type, cxl_event_trace_record(cxlmd, type,
&payload->records[i]); &payload->records[i]);
if (payload->flags & CXL_GET_EVENT_FLAG_OVERFLOW) if (payload->flags & CXL_GET_EVENT_FLAG_OVERFLOW)
trace_cxl_overflow(cxlds->cxlmd, type, payload); trace_cxl_overflow(cxlmd, type, payload);
rc = cxl_clear_event_record(cxlds, type, payload); rc = cxl_clear_event_record(mds, type, payload);
if (rc) { if (rc) {
dev_err_ratelimited(cxlds->dev, dev_err_ratelimited(dev,
"Event log '%d': Failed to clear events : %d", "Event log '%d': Failed to clear events : %d",
type, rc); type, rc);
break; break;
} }
} while (nr_rec); } while (nr_rec);
mutex_unlock(&cxlds->event.log_lock); mutex_unlock(&mds->event.log_lock);
} }
/** /**
* cxl_mem_get_event_records - Get Event Records from the device * cxl_mem_get_event_records - Get Event Records from the device
* @cxlds: The device data for the operation * @mds: The driver data for the operation
* @status: Event Status register value identifying which events are available. * @status: Event Status register value identifying which events are available.
* *
* Retrieve all event records available on the device, report them as trace * Retrieve all event records available on the device, report them as trace
@ -970,24 +975,24 @@ static void cxl_mem_get_records_log(struct cxl_dev_state *cxlds,
* See CXL rev 3.0 @8.2.9.2.2 Get Event Records * See CXL rev 3.0 @8.2.9.2.2 Get Event Records
* See CXL rev 3.0 @8.2.9.2.3 Clear Event Records * See CXL rev 3.0 @8.2.9.2.3 Clear Event Records
*/ */
void cxl_mem_get_event_records(struct cxl_dev_state *cxlds, u32 status) void cxl_mem_get_event_records(struct cxl_memdev_state *mds, u32 status)
{ {
dev_dbg(cxlds->dev, "Reading event logs: %x\n", status); dev_dbg(mds->cxlds.dev, "Reading event logs: %x\n", status);
if (status & CXLDEV_EVENT_STATUS_FATAL) if (status & CXLDEV_EVENT_STATUS_FATAL)
cxl_mem_get_records_log(cxlds, CXL_EVENT_TYPE_FATAL); cxl_mem_get_records_log(mds, CXL_EVENT_TYPE_FATAL);
if (status & CXLDEV_EVENT_STATUS_FAIL) if (status & CXLDEV_EVENT_STATUS_FAIL)
cxl_mem_get_records_log(cxlds, CXL_EVENT_TYPE_FAIL); cxl_mem_get_records_log(mds, CXL_EVENT_TYPE_FAIL);
if (status & CXLDEV_EVENT_STATUS_WARN) if (status & CXLDEV_EVENT_STATUS_WARN)
cxl_mem_get_records_log(cxlds, CXL_EVENT_TYPE_WARN); cxl_mem_get_records_log(mds, CXL_EVENT_TYPE_WARN);
if (status & CXLDEV_EVENT_STATUS_INFO) if (status & CXLDEV_EVENT_STATUS_INFO)
cxl_mem_get_records_log(cxlds, CXL_EVENT_TYPE_INFO); cxl_mem_get_records_log(mds, CXL_EVENT_TYPE_INFO);
} }
EXPORT_SYMBOL_NS_GPL(cxl_mem_get_event_records, CXL); EXPORT_SYMBOL_NS_GPL(cxl_mem_get_event_records, CXL);
/** /**
* cxl_mem_get_partition_info - Get partition info * cxl_mem_get_partition_info - Get partition info
* @cxlds: The device data for the operation * @mds: The driver data for the operation
* *
* Retrieve the current partition info for the device specified. The active * Retrieve the current partition info for the device specified. The active
* values are the current capacity in bytes. If not 0, the 'next' values are * values are the current capacity in bytes. If not 0, the 'next' values are
@ -997,7 +1002,7 @@ EXPORT_SYMBOL_NS_GPL(cxl_mem_get_event_records, CXL);
* *
* See CXL @8.2.9.5.2.1 Get Partition Info * See CXL @8.2.9.5.2.1 Get Partition Info
*/ */
static int cxl_mem_get_partition_info(struct cxl_dev_state *cxlds) static int cxl_mem_get_partition_info(struct cxl_memdev_state *mds)
{ {
struct cxl_mbox_get_partition_info pi; struct cxl_mbox_get_partition_info pi;
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
@ -1008,17 +1013,17 @@ static int cxl_mem_get_partition_info(struct cxl_dev_state *cxlds)
.size_out = sizeof(pi), .size_out = sizeof(pi),
.payload_out = &pi, .payload_out = &pi,
}; };
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc) if (rc)
return rc; return rc;
cxlds->active_volatile_bytes = mds->active_volatile_bytes =
le64_to_cpu(pi.active_volatile_cap) * CXL_CAPACITY_MULTIPLIER; le64_to_cpu(pi.active_volatile_cap) * CXL_CAPACITY_MULTIPLIER;
cxlds->active_persistent_bytes = mds->active_persistent_bytes =
le64_to_cpu(pi.active_persistent_cap) * CXL_CAPACITY_MULTIPLIER; le64_to_cpu(pi.active_persistent_cap) * CXL_CAPACITY_MULTIPLIER;
cxlds->next_volatile_bytes = mds->next_volatile_bytes =
le64_to_cpu(pi.next_volatile_cap) * CXL_CAPACITY_MULTIPLIER; le64_to_cpu(pi.next_volatile_cap) * CXL_CAPACITY_MULTIPLIER;
cxlds->next_persistent_bytes = mds->next_persistent_bytes =
le64_to_cpu(pi.next_volatile_cap) * CXL_CAPACITY_MULTIPLIER; le64_to_cpu(pi.next_volatile_cap) * CXL_CAPACITY_MULTIPLIER;
return 0; return 0;
@ -1026,14 +1031,14 @@ static int cxl_mem_get_partition_info(struct cxl_dev_state *cxlds)
/** /**
* cxl_dev_state_identify() - Send the IDENTIFY command to the device. * cxl_dev_state_identify() - Send the IDENTIFY command to the device.
* @cxlds: The device data for the operation * @mds: The driver data for the operation
* *
* Return: 0 if identify was executed successfully or media not ready. * Return: 0 if identify was executed successfully or media not ready.
* *
* This will dispatch the identify command to the device and on success populate * This will dispatch the identify command to the device and on success populate
* structures to be exported to sysfs. * structures to be exported to sysfs.
*/ */
int cxl_dev_state_identify(struct cxl_dev_state *cxlds) int cxl_dev_state_identify(struct cxl_memdev_state *mds)
{ {
/* See CXL 2.0 Table 175 Identify Memory Device Output Payload */ /* See CXL 2.0 Table 175 Identify Memory Device Output Payload */
struct cxl_mbox_identify id; struct cxl_mbox_identify id;
@ -1041,7 +1046,7 @@ int cxl_dev_state_identify(struct cxl_dev_state *cxlds)
u32 val; u32 val;
int rc; int rc;
if (!cxlds->media_ready) if (!mds->cxlds.media_ready)
return 0; return 0;
mbox_cmd = (struct cxl_mbox_cmd) { mbox_cmd = (struct cxl_mbox_cmd) {
@ -1049,25 +1054,26 @@ int cxl_dev_state_identify(struct cxl_dev_state *cxlds)
.size_out = sizeof(id), .size_out = sizeof(id),
.payload_out = &id, .payload_out = &id,
}; };
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc < 0) if (rc < 0)
return rc; return rc;
cxlds->total_bytes = mds->total_bytes =
le64_to_cpu(id.total_capacity) * CXL_CAPACITY_MULTIPLIER; le64_to_cpu(id.total_capacity) * CXL_CAPACITY_MULTIPLIER;
cxlds->volatile_only_bytes = mds->volatile_only_bytes =
le64_to_cpu(id.volatile_capacity) * CXL_CAPACITY_MULTIPLIER; le64_to_cpu(id.volatile_capacity) * CXL_CAPACITY_MULTIPLIER;
cxlds->persistent_only_bytes = mds->persistent_only_bytes =
le64_to_cpu(id.persistent_capacity) * CXL_CAPACITY_MULTIPLIER; le64_to_cpu(id.persistent_capacity) * CXL_CAPACITY_MULTIPLIER;
cxlds->partition_align_bytes = mds->partition_align_bytes =
le64_to_cpu(id.partition_align) * CXL_CAPACITY_MULTIPLIER; le64_to_cpu(id.partition_align) * CXL_CAPACITY_MULTIPLIER;
cxlds->lsa_size = le32_to_cpu(id.lsa_size); mds->lsa_size = le32_to_cpu(id.lsa_size);
memcpy(cxlds->firmware_version, id.fw_revision, sizeof(id.fw_revision)); memcpy(mds->firmware_version, id.fw_revision,
sizeof(id.fw_revision));
if (test_bit(CXL_POISON_ENABLED_LIST, cxlds->poison.enabled_cmds)) { if (test_bit(CXL_POISON_ENABLED_LIST, mds->poison.enabled_cmds)) {
val = get_unaligned_le24(id.poison_list_max_mer); val = get_unaligned_le24(id.poison_list_max_mer);
cxlds->poison.max_errors = min_t(u32, val, CXL_POISON_LIST_MAX); mds->poison.max_errors = min_t(u32, val, CXL_POISON_LIST_MAX);
} }
return 0; return 0;
@ -1100,8 +1106,9 @@ static int add_dpa_res(struct device *dev, struct resource *parent,
return 0; return 0;
} }
int cxl_mem_create_range_info(struct cxl_dev_state *cxlds) int cxl_mem_create_range_info(struct cxl_memdev_state *mds)
{ {
struct cxl_dev_state *cxlds = &mds->cxlds;
struct device *dev = cxlds->dev; struct device *dev = cxlds->dev;
int rc; int rc;
@ -1113,35 +1120,35 @@ int cxl_mem_create_range_info(struct cxl_dev_state *cxlds)
} }
cxlds->dpa_res = cxlds->dpa_res =
(struct resource)DEFINE_RES_MEM(0, cxlds->total_bytes); (struct resource)DEFINE_RES_MEM(0, mds->total_bytes);
if (cxlds->partition_align_bytes == 0) { if (mds->partition_align_bytes == 0) {
rc = add_dpa_res(dev, &cxlds->dpa_res, &cxlds->ram_res, 0, rc = add_dpa_res(dev, &cxlds->dpa_res, &cxlds->ram_res, 0,
cxlds->volatile_only_bytes, "ram"); mds->volatile_only_bytes, "ram");
if (rc) if (rc)
return rc; return rc;
return add_dpa_res(dev, &cxlds->dpa_res, &cxlds->pmem_res, return add_dpa_res(dev, &cxlds->dpa_res, &cxlds->pmem_res,
cxlds->volatile_only_bytes, mds->volatile_only_bytes,
cxlds->persistent_only_bytes, "pmem"); mds->persistent_only_bytes, "pmem");
} }
rc = cxl_mem_get_partition_info(cxlds); rc = cxl_mem_get_partition_info(mds);
if (rc) { if (rc) {
dev_err(dev, "Failed to query partition information\n"); dev_err(dev, "Failed to query partition information\n");
return rc; return rc;
} }
rc = add_dpa_res(dev, &cxlds->dpa_res, &cxlds->ram_res, 0, rc = add_dpa_res(dev, &cxlds->dpa_res, &cxlds->ram_res, 0,
cxlds->active_volatile_bytes, "ram"); mds->active_volatile_bytes, "ram");
if (rc) if (rc)
return rc; return rc;
return add_dpa_res(dev, &cxlds->dpa_res, &cxlds->pmem_res, return add_dpa_res(dev, &cxlds->dpa_res, &cxlds->pmem_res,
cxlds->active_volatile_bytes, mds->active_volatile_bytes,
cxlds->active_persistent_bytes, "pmem"); mds->active_persistent_bytes, "pmem");
} }
EXPORT_SYMBOL_NS_GPL(cxl_mem_create_range_info, CXL); EXPORT_SYMBOL_NS_GPL(cxl_mem_create_range_info, CXL);
int cxl_set_timestamp(struct cxl_dev_state *cxlds) int cxl_set_timestamp(struct cxl_memdev_state *mds)
{ {
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
struct cxl_mbox_set_timestamp_in pi; struct cxl_mbox_set_timestamp_in pi;
@ -1154,7 +1161,7 @@ int cxl_set_timestamp(struct cxl_dev_state *cxlds)
.payload_in = &pi, .payload_in = &pi,
}; };
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
/* /*
* Command is optional. Devices may have another way of providing * Command is optional. Devices may have another way of providing
* a timestamp, or may return all 0s in timestamp fields. * a timestamp, or may return all 0s in timestamp fields.
@ -1170,18 +1177,18 @@ EXPORT_SYMBOL_NS_GPL(cxl_set_timestamp, CXL);
int cxl_mem_get_poison(struct cxl_memdev *cxlmd, u64 offset, u64 len, int cxl_mem_get_poison(struct cxl_memdev *cxlmd, u64 offset, u64 len,
struct cxl_region *cxlr) struct cxl_region *cxlr)
{ {
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
struct cxl_mbox_poison_out *po; struct cxl_mbox_poison_out *po;
struct cxl_mbox_poison_in pi; struct cxl_mbox_poison_in pi;
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
int nr_records = 0; int nr_records = 0;
int rc; int rc;
rc = mutex_lock_interruptible(&cxlds->poison.lock); rc = mutex_lock_interruptible(&mds->poison.lock);
if (rc) if (rc)
return rc; return rc;
po = cxlds->poison.list_out; po = mds->poison.list_out;
pi.offset = cpu_to_le64(offset); pi.offset = cpu_to_le64(offset);
pi.length = cpu_to_le64(len / CXL_POISON_LEN_MULT); pi.length = cpu_to_le64(len / CXL_POISON_LEN_MULT);
@ -1189,13 +1196,13 @@ int cxl_mem_get_poison(struct cxl_memdev *cxlmd, u64 offset, u64 len,
.opcode = CXL_MBOX_OP_GET_POISON, .opcode = CXL_MBOX_OP_GET_POISON,
.size_in = sizeof(pi), .size_in = sizeof(pi),
.payload_in = &pi, .payload_in = &pi,
.size_out = cxlds->payload_size, .size_out = mds->payload_size,
.payload_out = po, .payload_out = po,
.min_out = struct_size(po, record, 0), .min_out = struct_size(po, record, 0),
}; };
do { do {
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc) if (rc)
break; break;
@ -1206,14 +1213,14 @@ int cxl_mem_get_poison(struct cxl_memdev *cxlmd, u64 offset, u64 len,
/* Protect against an uncleared _FLAG_MORE */ /* Protect against an uncleared _FLAG_MORE */
nr_records = nr_records + le16_to_cpu(po->count); nr_records = nr_records + le16_to_cpu(po->count);
if (nr_records >= cxlds->poison.max_errors) { if (nr_records >= mds->poison.max_errors) {
dev_dbg(&cxlmd->dev, "Max Error Records reached: %d\n", dev_dbg(&cxlmd->dev, "Max Error Records reached: %d\n",
nr_records); nr_records);
break; break;
} }
} while (po->flags & CXL_POISON_FLAG_MORE); } while (po->flags & CXL_POISON_FLAG_MORE);
mutex_unlock(&cxlds->poison.lock); mutex_unlock(&mds->poison.lock);
return rc; return rc;
} }
EXPORT_SYMBOL_NS_GPL(cxl_mem_get_poison, CXL); EXPORT_SYMBOL_NS_GPL(cxl_mem_get_poison, CXL);
@ -1223,52 +1230,52 @@ static void free_poison_buf(void *buf)
kvfree(buf); kvfree(buf);
} }
/* Get Poison List output buffer is protected by cxlds->poison.lock */ /* Get Poison List output buffer is protected by mds->poison.lock */
static int cxl_poison_alloc_buf(struct cxl_dev_state *cxlds) static int cxl_poison_alloc_buf(struct cxl_memdev_state *mds)
{ {
cxlds->poison.list_out = kvmalloc(cxlds->payload_size, GFP_KERNEL); mds->poison.list_out = kvmalloc(mds->payload_size, GFP_KERNEL);
if (!cxlds->poison.list_out) if (!mds->poison.list_out)
return -ENOMEM; return -ENOMEM;
return devm_add_action_or_reset(cxlds->dev, free_poison_buf, return devm_add_action_or_reset(mds->cxlds.dev, free_poison_buf,
cxlds->poison.list_out); mds->poison.list_out);
} }
int cxl_poison_state_init(struct cxl_dev_state *cxlds) int cxl_poison_state_init(struct cxl_memdev_state *mds)
{ {
int rc; int rc;
if (!test_bit(CXL_POISON_ENABLED_LIST, cxlds->poison.enabled_cmds)) if (!test_bit(CXL_POISON_ENABLED_LIST, mds->poison.enabled_cmds))
return 0; return 0;
rc = cxl_poison_alloc_buf(cxlds); rc = cxl_poison_alloc_buf(mds);
if (rc) { if (rc) {
clear_bit(CXL_POISON_ENABLED_LIST, cxlds->poison.enabled_cmds); clear_bit(CXL_POISON_ENABLED_LIST, mds->poison.enabled_cmds);
return rc; return rc;
} }
mutex_init(&cxlds->poison.lock); mutex_init(&mds->poison.lock);
return 0; return 0;
} }
EXPORT_SYMBOL_NS_GPL(cxl_poison_state_init, CXL); EXPORT_SYMBOL_NS_GPL(cxl_poison_state_init, CXL);
struct cxl_dev_state *cxl_dev_state_create(struct device *dev) struct cxl_memdev_state *cxl_memdev_state_create(struct device *dev)
{ {
struct cxl_dev_state *cxlds; struct cxl_memdev_state *mds;
cxlds = devm_kzalloc(dev, sizeof(*cxlds), GFP_KERNEL); mds = devm_kzalloc(dev, sizeof(*mds), GFP_KERNEL);
if (!cxlds) { if (!mds) {
dev_err(dev, "No memory available\n"); dev_err(dev, "No memory available\n");
return ERR_PTR(-ENOMEM); return ERR_PTR(-ENOMEM);
} }
mutex_init(&cxlds->mbox_mutex); mutex_init(&mds->mbox_mutex);
mutex_init(&cxlds->event.log_lock); mutex_init(&mds->event.log_lock);
cxlds->dev = dev; mds->cxlds.dev = dev;
return cxlds; return mds;
} }
EXPORT_SYMBOL_NS_GPL(cxl_dev_state_create, CXL); EXPORT_SYMBOL_NS_GPL(cxl_memdev_state_create, CXL);
void __init cxl_mbox_init(void) void __init cxl_mbox_init(void)
{ {

View File

@ -39,8 +39,9 @@ static ssize_t firmware_version_show(struct device *dev,
{ {
struct cxl_memdev *cxlmd = to_cxl_memdev(dev); struct cxl_memdev *cxlmd = to_cxl_memdev(dev);
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_dev_state *cxlds = cxlmd->cxlds;
struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlds);
return sysfs_emit(buf, "%.16s\n", cxlds->firmware_version); return sysfs_emit(buf, "%.16s\n", mds->firmware_version);
} }
static DEVICE_ATTR_RO(firmware_version); static DEVICE_ATTR_RO(firmware_version);
@ -49,8 +50,9 @@ static ssize_t payload_max_show(struct device *dev,
{ {
struct cxl_memdev *cxlmd = to_cxl_memdev(dev); struct cxl_memdev *cxlmd = to_cxl_memdev(dev);
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_dev_state *cxlds = cxlmd->cxlds;
struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlds);
return sysfs_emit(buf, "%zu\n", cxlds->payload_size); return sysfs_emit(buf, "%zu\n", mds->payload_size);
} }
static DEVICE_ATTR_RO(payload_max); static DEVICE_ATTR_RO(payload_max);
@ -59,8 +61,9 @@ static ssize_t label_storage_size_show(struct device *dev,
{ {
struct cxl_memdev *cxlmd = to_cxl_memdev(dev); struct cxl_memdev *cxlmd = to_cxl_memdev(dev);
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_dev_state *cxlds = cxlmd->cxlds;
struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlds);
return sysfs_emit(buf, "%zu\n", cxlds->lsa_size); return sysfs_emit(buf, "%zu\n", mds->lsa_size);
} }
static DEVICE_ATTR_RO(label_storage_size); static DEVICE_ATTR_RO(label_storage_size);
@ -231,7 +234,7 @@ static int cxl_validate_poison_dpa(struct cxl_memdev *cxlmd, u64 dpa)
int cxl_inject_poison(struct cxl_memdev *cxlmd, u64 dpa) int cxl_inject_poison(struct cxl_memdev *cxlmd, u64 dpa)
{ {
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
struct cxl_mbox_inject_poison inject; struct cxl_mbox_inject_poison inject;
struct cxl_poison_record record; struct cxl_poison_record record;
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
@ -255,13 +258,13 @@ int cxl_inject_poison(struct cxl_memdev *cxlmd, u64 dpa)
.size_in = sizeof(inject), .size_in = sizeof(inject),
.payload_in = &inject, .payload_in = &inject,
}; };
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc) if (rc)
goto out; goto out;
cxlr = cxl_dpa_to_region(cxlmd, dpa); cxlr = cxl_dpa_to_region(cxlmd, dpa);
if (cxlr) if (cxlr)
dev_warn_once(cxlds->dev, dev_warn_once(mds->cxlds.dev,
"poison inject dpa:%#llx region: %s\n", dpa, "poison inject dpa:%#llx region: %s\n", dpa,
dev_name(&cxlr->dev)); dev_name(&cxlr->dev));
@ -279,7 +282,7 @@ EXPORT_SYMBOL_NS_GPL(cxl_inject_poison, CXL);
int cxl_clear_poison(struct cxl_memdev *cxlmd, u64 dpa) int cxl_clear_poison(struct cxl_memdev *cxlmd, u64 dpa)
{ {
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
struct cxl_mbox_clear_poison clear; struct cxl_mbox_clear_poison clear;
struct cxl_poison_record record; struct cxl_poison_record record;
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
@ -312,14 +315,15 @@ int cxl_clear_poison(struct cxl_memdev *cxlmd, u64 dpa)
.payload_in = &clear, .payload_in = &clear,
}; };
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc) if (rc)
goto out; goto out;
cxlr = cxl_dpa_to_region(cxlmd, dpa); cxlr = cxl_dpa_to_region(cxlmd, dpa);
if (cxlr) if (cxlr)
dev_warn_once(cxlds->dev, "poison clear dpa:%#llx region: %s\n", dev_warn_once(mds->cxlds.dev,
dpa, dev_name(&cxlr->dev)); "poison clear dpa:%#llx region: %s\n", dpa,
dev_name(&cxlr->dev));
record = (struct cxl_poison_record) { record = (struct cxl_poison_record) {
.address = cpu_to_le64(dpa), .address = cpu_to_le64(dpa),
@ -397,17 +401,18 @@ EXPORT_SYMBOL_NS_GPL(is_cxl_memdev, CXL);
/** /**
* set_exclusive_cxl_commands() - atomically disable user cxl commands * set_exclusive_cxl_commands() - atomically disable user cxl commands
* @cxlds: The device state to operate on * @mds: The device state to operate on
* @cmds: bitmap of commands to mark exclusive * @cmds: bitmap of commands to mark exclusive
* *
* Grab the cxl_memdev_rwsem in write mode to flush in-flight * Grab the cxl_memdev_rwsem in write mode to flush in-flight
* invocations of the ioctl path and then disable future execution of * invocations of the ioctl path and then disable future execution of
* commands with the command ids set in @cmds. * commands with the command ids set in @cmds.
*/ */
void set_exclusive_cxl_commands(struct cxl_dev_state *cxlds, unsigned long *cmds) void set_exclusive_cxl_commands(struct cxl_memdev_state *mds,
unsigned long *cmds)
{ {
down_write(&cxl_memdev_rwsem); down_write(&cxl_memdev_rwsem);
bitmap_or(cxlds->exclusive_cmds, cxlds->exclusive_cmds, cmds, bitmap_or(mds->exclusive_cmds, mds->exclusive_cmds, cmds,
CXL_MEM_COMMAND_ID_MAX); CXL_MEM_COMMAND_ID_MAX);
up_write(&cxl_memdev_rwsem); up_write(&cxl_memdev_rwsem);
} }
@ -415,13 +420,14 @@ EXPORT_SYMBOL_NS_GPL(set_exclusive_cxl_commands, CXL);
/** /**
* clear_exclusive_cxl_commands() - atomically enable user cxl commands * clear_exclusive_cxl_commands() - atomically enable user cxl commands
* @cxlds: The device state to modify * @mds: The device state to modify
* @cmds: bitmap of commands to mark available for userspace * @cmds: bitmap of commands to mark available for userspace
*/ */
void clear_exclusive_cxl_commands(struct cxl_dev_state *cxlds, unsigned long *cmds) void clear_exclusive_cxl_commands(struct cxl_memdev_state *mds,
unsigned long *cmds)
{ {
down_write(&cxl_memdev_rwsem); down_write(&cxl_memdev_rwsem);
bitmap_andnot(cxlds->exclusive_cmds, cxlds->exclusive_cmds, cmds, bitmap_andnot(mds->exclusive_cmds, mds->exclusive_cmds, cmds,
CXL_MEM_COMMAND_ID_MAX); CXL_MEM_COMMAND_ID_MAX);
up_write(&cxl_memdev_rwsem); up_write(&cxl_memdev_rwsem);
} }

View File

@ -267,6 +267,34 @@ struct cxl_poison_state {
* @cxl_dvsec: Offset to the PCIe device DVSEC * @cxl_dvsec: Offset to the PCIe device DVSEC
* @rcd: operating in RCD mode (CXL 3.0 9.11.8 CXL Devices Attached to an RCH) * @rcd: operating in RCD mode (CXL 3.0 9.11.8 CXL Devices Attached to an RCH)
* @media_ready: Indicate whether the device media is usable * @media_ready: Indicate whether the device media is usable
* @dpa_res: Overall DPA resource tree for the device
* @pmem_res: Active Persistent memory capacity configuration
* @ram_res: Active Volatile memory capacity configuration
* @component_reg_phys: register base of component registers
* @serial: PCIe Device Serial Number
*/
struct cxl_dev_state {
struct device *dev;
struct cxl_memdev *cxlmd;
struct cxl_regs regs;
int cxl_dvsec;
bool rcd;
bool media_ready;
struct resource dpa_res;
struct resource pmem_res;
struct resource ram_res;
resource_size_t component_reg_phys;
u64 serial;
};
/**
* struct cxl_memdev_state - Generic Type-3 Memory Device Class driver data
*
* CXL 8.1.12.1 PCI Header - Class Code Register Memory Device defines
* common memory device functionality like the presence of a mailbox and
* the functionality related to that like Identify Memory Device and Get
* Partition Info
* @cxlds: Core driver state common across Type-2 and Type-3 devices
* @payload_size: Size of space for payload * @payload_size: Size of space for payload
* (CXL 2.0 8.2.8.4.3 Mailbox Capabilities Register) * (CXL 2.0 8.2.8.4.3 Mailbox Capabilities Register)
* @lsa_size: Size of Label Storage Area * @lsa_size: Size of Label Storage Area
@ -275,9 +303,6 @@ struct cxl_poison_state {
* @firmware_version: Firmware version for the memory device. * @firmware_version: Firmware version for the memory device.
* @enabled_cmds: Hardware commands found enabled in CEL. * @enabled_cmds: Hardware commands found enabled in CEL.
* @exclusive_cmds: Commands that are kernel-internal only * @exclusive_cmds: Commands that are kernel-internal only
* @dpa_res: Overall DPA resource tree for the device
* @pmem_res: Active Persistent memory capacity configuration
* @ram_res: Active Volatile memory capacity configuration
* @total_bytes: sum of all possible capacities * @total_bytes: sum of all possible capacities
* @volatile_only_bytes: hard volatile capacity * @volatile_only_bytes: hard volatile capacity
* @persistent_only_bytes: hard persistent capacity * @persistent_only_bytes: hard persistent capacity
@ -286,53 +311,41 @@ struct cxl_poison_state {
* @active_persistent_bytes: sum of hard + soft persistent * @active_persistent_bytes: sum of hard + soft persistent
* @next_volatile_bytes: volatile capacity change pending device reset * @next_volatile_bytes: volatile capacity change pending device reset
* @next_persistent_bytes: persistent capacity change pending device reset * @next_persistent_bytes: persistent capacity change pending device reset
* @component_reg_phys: register base of component registers
* @serial: PCIe Device Serial Number
* @event: event log driver state * @event: event log driver state
* @poison: poison driver state info * @poison: poison driver state info
* @mbox_send: @dev specific transport for transmitting mailbox commands * @mbox_send: @dev specific transport for transmitting mailbox commands
* *
* See section 8.2.9.5.2 Capacity Configuration and Label Storage for * See CXL 3.0 8.2.9.8.2 Capacity Configuration and Label Storage for
* details on capacity parameters. * details on capacity parameters.
*/ */
struct cxl_dev_state { struct cxl_memdev_state {
struct device *dev; struct cxl_dev_state cxlds;
struct cxl_memdev *cxlmd;
struct cxl_regs regs;
int cxl_dvsec;
bool rcd;
bool media_ready;
size_t payload_size; size_t payload_size;
size_t lsa_size; size_t lsa_size;
struct mutex mbox_mutex; /* Protects device mailbox and firmware */ struct mutex mbox_mutex; /* Protects device mailbox and firmware */
char firmware_version[0x10]; char firmware_version[0x10];
DECLARE_BITMAP(enabled_cmds, CXL_MEM_COMMAND_ID_MAX); DECLARE_BITMAP(enabled_cmds, CXL_MEM_COMMAND_ID_MAX);
DECLARE_BITMAP(exclusive_cmds, CXL_MEM_COMMAND_ID_MAX); DECLARE_BITMAP(exclusive_cmds, CXL_MEM_COMMAND_ID_MAX);
struct resource dpa_res;
struct resource pmem_res;
struct resource ram_res;
u64 total_bytes; u64 total_bytes;
u64 volatile_only_bytes; u64 volatile_only_bytes;
u64 persistent_only_bytes; u64 persistent_only_bytes;
u64 partition_align_bytes; u64 partition_align_bytes;
u64 active_volatile_bytes; u64 active_volatile_bytes;
u64 active_persistent_bytes; u64 active_persistent_bytes;
u64 next_volatile_bytes; u64 next_volatile_bytes;
u64 next_persistent_bytes; u64 next_persistent_bytes;
resource_size_t component_reg_phys;
u64 serial;
struct cxl_event_state event; struct cxl_event_state event;
struct cxl_poison_state poison; struct cxl_poison_state poison;
int (*mbox_send)(struct cxl_memdev_state *mds,
int (*mbox_send)(struct cxl_dev_state *cxlds, struct cxl_mbox_cmd *cmd); struct cxl_mbox_cmd *cmd);
}; };
static inline struct cxl_memdev_state *
to_cxl_memdev_state(struct cxl_dev_state *cxlds)
{
return container_of(cxlds, struct cxl_memdev_state, cxlds);
}
enum cxl_opcode { enum cxl_opcode {
CXL_MBOX_OP_INVALID = 0x0000, CXL_MBOX_OP_INVALID = 0x0000,
CXL_MBOX_OP_RAW = CXL_MBOX_OP_INVALID, CXL_MBOX_OP_RAW = CXL_MBOX_OP_INVALID,
@ -691,18 +704,20 @@ enum {
CXL_PMEM_SEC_PASS_USER, CXL_PMEM_SEC_PASS_USER,
}; };
int cxl_internal_send_cmd(struct cxl_dev_state *cxlds, int cxl_internal_send_cmd(struct cxl_memdev_state *mds,
struct cxl_mbox_cmd *cmd); struct cxl_mbox_cmd *cmd);
int cxl_dev_state_identify(struct cxl_dev_state *cxlds); int cxl_dev_state_identify(struct cxl_memdev_state *mds);
int cxl_await_media_ready(struct cxl_dev_state *cxlds); int cxl_await_media_ready(struct cxl_dev_state *cxlds);
int cxl_enumerate_cmds(struct cxl_dev_state *cxlds); int cxl_enumerate_cmds(struct cxl_memdev_state *mds);
int cxl_mem_create_range_info(struct cxl_dev_state *cxlds); int cxl_mem_create_range_info(struct cxl_memdev_state *mds);
struct cxl_dev_state *cxl_dev_state_create(struct device *dev); struct cxl_memdev_state *cxl_memdev_state_create(struct device *dev);
void set_exclusive_cxl_commands(struct cxl_dev_state *cxlds, unsigned long *cmds); void set_exclusive_cxl_commands(struct cxl_memdev_state *mds,
void clear_exclusive_cxl_commands(struct cxl_dev_state *cxlds, unsigned long *cmds); unsigned long *cmds);
void cxl_mem_get_event_records(struct cxl_dev_state *cxlds, u32 status); void clear_exclusive_cxl_commands(struct cxl_memdev_state *mds,
int cxl_set_timestamp(struct cxl_dev_state *cxlds); unsigned long *cmds);
int cxl_poison_state_init(struct cxl_dev_state *cxlds); void cxl_mem_get_event_records(struct cxl_memdev_state *mds, u32 status);
int cxl_set_timestamp(struct cxl_memdev_state *mds);
int cxl_poison_state_init(struct cxl_memdev_state *mds);
int cxl_mem_get_poison(struct cxl_memdev *cxlmd, u64 offset, u64 len, int cxl_mem_get_poison(struct cxl_memdev *cxlmd, u64 offset, u64 len,
struct cxl_region *cxlr); struct cxl_region *cxlr);
int cxl_trigger_poison_list(struct cxl_memdev *cxlmd); int cxl_trigger_poison_list(struct cxl_memdev *cxlmd);

View File

@ -117,6 +117,7 @@ DEFINE_DEBUGFS_ATTRIBUTE(cxl_poison_clear_fops, NULL,
static int cxl_mem_probe(struct device *dev) static int cxl_mem_probe(struct device *dev)
{ {
struct cxl_memdev *cxlmd = to_cxl_memdev(dev); struct cxl_memdev *cxlmd = to_cxl_memdev(dev);
struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_dev_state *cxlds = cxlmd->cxlds;
struct device *endpoint_parent; struct device *endpoint_parent;
struct cxl_port *parent_port; struct cxl_port *parent_port;
@ -141,10 +142,10 @@ static int cxl_mem_probe(struct device *dev)
dentry = cxl_debugfs_create_dir(dev_name(dev)); dentry = cxl_debugfs_create_dir(dev_name(dev));
debugfs_create_devm_seqfile(dev, "dpamem", dentry, cxl_mem_dpa_show); debugfs_create_devm_seqfile(dev, "dpamem", dentry, cxl_mem_dpa_show);
if (test_bit(CXL_POISON_ENABLED_INJECT, cxlds->poison.enabled_cmds)) if (test_bit(CXL_POISON_ENABLED_INJECT, mds->poison.enabled_cmds))
debugfs_create_file("inject_poison", 0200, dentry, cxlmd, debugfs_create_file("inject_poison", 0200, dentry, cxlmd,
&cxl_poison_inject_fops); &cxl_poison_inject_fops);
if (test_bit(CXL_POISON_ENABLED_CLEAR, cxlds->poison.enabled_cmds)) if (test_bit(CXL_POISON_ENABLED_CLEAR, mds->poison.enabled_cmds))
debugfs_create_file("clear_poison", 0200, dentry, cxlmd, debugfs_create_file("clear_poison", 0200, dentry, cxlmd,
&cxl_poison_clear_fops); &cxl_poison_clear_fops);
@ -227,9 +228,12 @@ static umode_t cxl_mem_visible(struct kobject *kobj, struct attribute *a, int n)
{ {
if (a == &dev_attr_trigger_poison_list.attr) { if (a == &dev_attr_trigger_poison_list.attr) {
struct device *dev = kobj_to_dev(kobj); struct device *dev = kobj_to_dev(kobj);
struct cxl_memdev *cxlmd = to_cxl_memdev(dev);
struct cxl_memdev_state *mds =
to_cxl_memdev_state(cxlmd->cxlds);
if (!test_bit(CXL_POISON_ENABLED_LIST, if (!test_bit(CXL_POISON_ENABLED_LIST,
to_cxl_memdev(dev)->cxlds->poison.enabled_cmds)) mds->poison.enabled_cmds))
return 0; return 0;
} }
return a->mode; return a->mode;

View File

@ -86,7 +86,7 @@ static int cxl_pci_mbox_wait_for_doorbell(struct cxl_dev_state *cxlds)
/** /**
* __cxl_pci_mbox_send_cmd() - Execute a mailbox command * __cxl_pci_mbox_send_cmd() - Execute a mailbox command
* @cxlds: The device state to communicate with. * @mds: The memory device driver data
* @mbox_cmd: Command to send to the memory device. * @mbox_cmd: Command to send to the memory device.
* *
* Context: Any context. Expects mbox_mutex to be held. * Context: Any context. Expects mbox_mutex to be held.
@ -106,16 +106,17 @@ static int cxl_pci_mbox_wait_for_doorbell(struct cxl_dev_state *cxlds)
* not need to coordinate with each other. The driver only uses the primary * not need to coordinate with each other. The driver only uses the primary
* mailbox. * mailbox.
*/ */
static int __cxl_pci_mbox_send_cmd(struct cxl_dev_state *cxlds, static int __cxl_pci_mbox_send_cmd(struct cxl_memdev_state *mds,
struct cxl_mbox_cmd *mbox_cmd) struct cxl_mbox_cmd *mbox_cmd)
{ {
struct cxl_dev_state *cxlds = &mds->cxlds;
void __iomem *payload = cxlds->regs.mbox + CXLDEV_MBOX_PAYLOAD_OFFSET; void __iomem *payload = cxlds->regs.mbox + CXLDEV_MBOX_PAYLOAD_OFFSET;
struct device *dev = cxlds->dev; struct device *dev = cxlds->dev;
u64 cmd_reg, status_reg; u64 cmd_reg, status_reg;
size_t out_len; size_t out_len;
int rc; int rc;
lockdep_assert_held(&cxlds->mbox_mutex); lockdep_assert_held(&mds->mbox_mutex);
/* /*
* Here are the steps from 8.2.8.4 of the CXL 2.0 spec. * Here are the steps from 8.2.8.4 of the CXL 2.0 spec.
@ -196,8 +197,9 @@ static int __cxl_pci_mbox_send_cmd(struct cxl_dev_state *cxlds,
* have requested less data than the hardware supplied even * have requested less data than the hardware supplied even
* within spec. * within spec.
*/ */
size_t n = min3(mbox_cmd->size_out, cxlds->payload_size, out_len); size_t n;
n = min3(mbox_cmd->size_out, mds->payload_size, out_len);
memcpy_fromio(mbox_cmd->payload_out, payload, n); memcpy_fromio(mbox_cmd->payload_out, payload, n);
mbox_cmd->size_out = n; mbox_cmd->size_out = n;
} else { } else {
@ -207,20 +209,23 @@ static int __cxl_pci_mbox_send_cmd(struct cxl_dev_state *cxlds,
return 0; return 0;
} }
static int cxl_pci_mbox_send(struct cxl_dev_state *cxlds, struct cxl_mbox_cmd *cmd) static int cxl_pci_mbox_send(struct cxl_memdev_state *mds,
struct cxl_mbox_cmd *cmd)
{ {
int rc; int rc;
mutex_lock_io(&cxlds->mbox_mutex); mutex_lock_io(&mds->mbox_mutex);
rc = __cxl_pci_mbox_send_cmd(cxlds, cmd); rc = __cxl_pci_mbox_send_cmd(mds, cmd);
mutex_unlock(&cxlds->mbox_mutex); mutex_unlock(&mds->mbox_mutex);
return rc; return rc;
} }
static int cxl_pci_setup_mailbox(struct cxl_dev_state *cxlds) static int cxl_pci_setup_mailbox(struct cxl_memdev_state *mds)
{ {
struct cxl_dev_state *cxlds = &mds->cxlds;
const int cap = readl(cxlds->regs.mbox + CXLDEV_MBOX_CAPS_OFFSET); const int cap = readl(cxlds->regs.mbox + CXLDEV_MBOX_CAPS_OFFSET);
struct device *dev = cxlds->dev;
unsigned long timeout; unsigned long timeout;
u64 md_status; u64 md_status;
@ -234,8 +239,7 @@ static int cxl_pci_setup_mailbox(struct cxl_dev_state *cxlds)
} while (!time_after(jiffies, timeout)); } while (!time_after(jiffies, timeout));
if (!(md_status & CXLMDEV_MBOX_IF_READY)) { if (!(md_status & CXLMDEV_MBOX_IF_READY)) {
cxl_err(cxlds->dev, md_status, cxl_err(dev, md_status, "timeout awaiting mailbox ready");
"timeout awaiting mailbox ready");
return -ETIMEDOUT; return -ETIMEDOUT;
} }
@ -246,12 +250,12 @@ static int cxl_pci_setup_mailbox(struct cxl_dev_state *cxlds)
* source for future doorbell busy events. * source for future doorbell busy events.
*/ */
if (cxl_pci_mbox_wait_for_doorbell(cxlds) != 0) { if (cxl_pci_mbox_wait_for_doorbell(cxlds) != 0) {
cxl_err(cxlds->dev, md_status, "timeout awaiting mailbox idle"); cxl_err(dev, md_status, "timeout awaiting mailbox idle");
return -ETIMEDOUT; return -ETIMEDOUT;
} }
cxlds->mbox_send = cxl_pci_mbox_send; mds->mbox_send = cxl_pci_mbox_send;
cxlds->payload_size = mds->payload_size =
1 << FIELD_GET(CXLDEV_MBOX_CAP_PAYLOAD_SIZE_MASK, cap); 1 << FIELD_GET(CXLDEV_MBOX_CAP_PAYLOAD_SIZE_MASK, cap);
/* /*
@ -261,15 +265,14 @@ static int cxl_pci_setup_mailbox(struct cxl_dev_state *cxlds)
* there's no point in going forward. If the size is too large, there's * there's no point in going forward. If the size is too large, there's
* no harm is soft limiting it. * no harm is soft limiting it.
*/ */
cxlds->payload_size = min_t(size_t, cxlds->payload_size, SZ_1M); mds->payload_size = min_t(size_t, mds->payload_size, SZ_1M);
if (cxlds->payload_size < 256) { if (mds->payload_size < 256) {
dev_err(cxlds->dev, "Mailbox is too small (%zub)", dev_err(dev, "Mailbox is too small (%zub)",
cxlds->payload_size); mds->payload_size);
return -ENXIO; return -ENXIO;
} }
dev_dbg(cxlds->dev, "Mailbox payload sized %zu", dev_dbg(dev, "Mailbox payload sized %zu", mds->payload_size);
cxlds->payload_size);
return 0; return 0;
} }
@ -433,18 +436,18 @@ static void free_event_buf(void *buf)
/* /*
* There is a single buffer for reading event logs from the mailbox. All logs * There is a single buffer for reading event logs from the mailbox. All logs
* share this buffer protected by the cxlds->event_log_lock. * share this buffer protected by the mds->event_log_lock.
*/ */
static int cxl_mem_alloc_event_buf(struct cxl_dev_state *cxlds) static int cxl_mem_alloc_event_buf(struct cxl_memdev_state *mds)
{ {
struct cxl_get_event_payload *buf; struct cxl_get_event_payload *buf;
buf = kvmalloc(cxlds->payload_size, GFP_KERNEL); buf = kvmalloc(mds->payload_size, GFP_KERNEL);
if (!buf) if (!buf)
return -ENOMEM; return -ENOMEM;
cxlds->event.buf = buf; mds->event.buf = buf;
return devm_add_action_or_reset(cxlds->dev, free_event_buf, buf); return devm_add_action_or_reset(mds->cxlds.dev, free_event_buf, buf);
} }
static int cxl_alloc_irq_vectors(struct pci_dev *pdev) static int cxl_alloc_irq_vectors(struct pci_dev *pdev)
@ -477,6 +480,7 @@ static irqreturn_t cxl_event_thread(int irq, void *id)
{ {
struct cxl_dev_id *dev_id = id; struct cxl_dev_id *dev_id = id;
struct cxl_dev_state *cxlds = dev_id->cxlds; struct cxl_dev_state *cxlds = dev_id->cxlds;
struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlds);
u32 status; u32 status;
do { do {
@ -489,7 +493,7 @@ static irqreturn_t cxl_event_thread(int irq, void *id)
status &= CXLDEV_EVENT_STATUS_ALL; status &= CXLDEV_EVENT_STATUS_ALL;
if (!status) if (!status)
break; break;
cxl_mem_get_event_records(cxlds, status); cxl_mem_get_event_records(mds, status);
cond_resched(); cond_resched();
} while (status); } while (status);
@ -522,7 +526,7 @@ static int cxl_event_req_irq(struct cxl_dev_state *cxlds, u8 setting)
dev_id); dev_id);
} }
static int cxl_event_get_int_policy(struct cxl_dev_state *cxlds, static int cxl_event_get_int_policy(struct cxl_memdev_state *mds,
struct cxl_event_interrupt_policy *policy) struct cxl_event_interrupt_policy *policy)
{ {
struct cxl_mbox_cmd mbox_cmd = { struct cxl_mbox_cmd mbox_cmd = {
@ -532,15 +536,15 @@ static int cxl_event_get_int_policy(struct cxl_dev_state *cxlds,
}; };
int rc; int rc;
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc < 0) if (rc < 0)
dev_err(cxlds->dev, "Failed to get event interrupt policy : %d", dev_err(mds->cxlds.dev,
rc); "Failed to get event interrupt policy : %d", rc);
return rc; return rc;
} }
static int cxl_event_config_msgnums(struct cxl_dev_state *cxlds, static int cxl_event_config_msgnums(struct cxl_memdev_state *mds,
struct cxl_event_interrupt_policy *policy) struct cxl_event_interrupt_policy *policy)
{ {
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
@ -559,23 +563,24 @@ static int cxl_event_config_msgnums(struct cxl_dev_state *cxlds,
.size_in = sizeof(*policy), .size_in = sizeof(*policy),
}; };
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc < 0) { if (rc < 0) {
dev_err(cxlds->dev, "Failed to set event interrupt policy : %d", dev_err(mds->cxlds.dev, "Failed to set event interrupt policy : %d",
rc); rc);
return rc; return rc;
} }
/* Retrieve final interrupt settings */ /* Retrieve final interrupt settings */
return cxl_event_get_int_policy(cxlds, policy); return cxl_event_get_int_policy(mds, policy);
} }
static int cxl_event_irqsetup(struct cxl_dev_state *cxlds) static int cxl_event_irqsetup(struct cxl_memdev_state *mds)
{ {
struct cxl_dev_state *cxlds = &mds->cxlds;
struct cxl_event_interrupt_policy policy; struct cxl_event_interrupt_policy policy;
int rc; int rc;
rc = cxl_event_config_msgnums(cxlds, &policy); rc = cxl_event_config_msgnums(mds, &policy);
if (rc) if (rc)
return rc; return rc;
@ -614,7 +619,7 @@ static bool cxl_event_int_is_fw(u8 setting)
} }
static int cxl_event_config(struct pci_host_bridge *host_bridge, static int cxl_event_config(struct pci_host_bridge *host_bridge,
struct cxl_dev_state *cxlds) struct cxl_memdev_state *mds)
{ {
struct cxl_event_interrupt_policy policy; struct cxl_event_interrupt_policy policy;
int rc; int rc;
@ -626,11 +631,11 @@ static int cxl_event_config(struct pci_host_bridge *host_bridge,
if (!host_bridge->native_cxl_error) if (!host_bridge->native_cxl_error)
return 0; return 0;
rc = cxl_mem_alloc_event_buf(cxlds); rc = cxl_mem_alloc_event_buf(mds);
if (rc) if (rc)
return rc; return rc;
rc = cxl_event_get_int_policy(cxlds, &policy); rc = cxl_event_get_int_policy(mds, &policy);
if (rc) if (rc)
return rc; return rc;
@ -638,15 +643,16 @@ static int cxl_event_config(struct pci_host_bridge *host_bridge,
cxl_event_int_is_fw(policy.warn_settings) || cxl_event_int_is_fw(policy.warn_settings) ||
cxl_event_int_is_fw(policy.failure_settings) || cxl_event_int_is_fw(policy.failure_settings) ||
cxl_event_int_is_fw(policy.fatal_settings)) { cxl_event_int_is_fw(policy.fatal_settings)) {
dev_err(cxlds->dev, "FW still in control of Event Logs despite _OSC settings\n"); dev_err(mds->cxlds.dev,
"FW still in control of Event Logs despite _OSC settings\n");
return -EBUSY; return -EBUSY;
} }
rc = cxl_event_irqsetup(cxlds); rc = cxl_event_irqsetup(mds);
if (rc) if (rc)
return rc; return rc;
cxl_mem_get_event_records(cxlds, CXLDEV_EVENT_STATUS_ALL); cxl_mem_get_event_records(mds, CXLDEV_EVENT_STATUS_ALL);
return 0; return 0;
} }
@ -654,9 +660,10 @@ static int cxl_event_config(struct pci_host_bridge *host_bridge,
static int cxl_pci_probe(struct pci_dev *pdev, const struct pci_device_id *id) static int cxl_pci_probe(struct pci_dev *pdev, const struct pci_device_id *id)
{ {
struct pci_host_bridge *host_bridge = pci_find_host_bridge(pdev->bus); struct pci_host_bridge *host_bridge = pci_find_host_bridge(pdev->bus);
struct cxl_memdev_state *mds;
struct cxl_dev_state *cxlds;
struct cxl_register_map map; struct cxl_register_map map;
struct cxl_memdev *cxlmd; struct cxl_memdev *cxlmd;
struct cxl_dev_state *cxlds;
int rc; int rc;
/* /*
@ -671,9 +678,10 @@ static int cxl_pci_probe(struct pci_dev *pdev, const struct pci_device_id *id)
return rc; return rc;
pci_set_master(pdev); pci_set_master(pdev);
cxlds = cxl_dev_state_create(&pdev->dev); mds = cxl_memdev_state_create(&pdev->dev);
if (IS_ERR(cxlds)) if (IS_ERR(mds))
return PTR_ERR(cxlds); return PTR_ERR(mds);
cxlds = &mds->cxlds;
pci_set_drvdata(pdev, cxlds); pci_set_drvdata(pdev, cxlds);
cxlds->rcd = is_cxl_restricted(pdev); cxlds->rcd = is_cxl_restricted(pdev);
@ -714,27 +722,27 @@ static int cxl_pci_probe(struct pci_dev *pdev, const struct pci_device_id *id)
else else
dev_warn(&pdev->dev, "Media not active (%d)\n", rc); dev_warn(&pdev->dev, "Media not active (%d)\n", rc);
rc = cxl_pci_setup_mailbox(cxlds); rc = cxl_pci_setup_mailbox(mds);
if (rc) if (rc)
return rc; return rc;
rc = cxl_enumerate_cmds(cxlds); rc = cxl_enumerate_cmds(mds);
if (rc) if (rc)
return rc; return rc;
rc = cxl_set_timestamp(cxlds); rc = cxl_set_timestamp(mds);
if (rc) if (rc)
return rc; return rc;
rc = cxl_poison_state_init(cxlds); rc = cxl_poison_state_init(mds);
if (rc) if (rc)
return rc; return rc;
rc = cxl_dev_state_identify(cxlds); rc = cxl_dev_state_identify(mds);
if (rc) if (rc)
return rc; return rc;
rc = cxl_mem_create_range_info(cxlds); rc = cxl_mem_create_range_info(mds);
if (rc) if (rc)
return rc; return rc;
@ -746,7 +754,7 @@ static int cxl_pci_probe(struct pci_dev *pdev, const struct pci_device_id *id)
if (IS_ERR(cxlmd)) if (IS_ERR(cxlmd))
return PTR_ERR(cxlmd); return PTR_ERR(cxlmd);
rc = cxl_event_config(host_bridge, cxlds); rc = cxl_event_config(host_bridge, mds);
if (rc) if (rc)
return rc; return rc;

View File

@ -15,9 +15,9 @@ extern const struct nvdimm_security_ops *cxl_security_ops;
static __read_mostly DECLARE_BITMAP(exclusive_cmds, CXL_MEM_COMMAND_ID_MAX); static __read_mostly DECLARE_BITMAP(exclusive_cmds, CXL_MEM_COMMAND_ID_MAX);
static void clear_exclusive(void *cxlds) static void clear_exclusive(void *mds)
{ {
clear_exclusive_cxl_commands(cxlds, exclusive_cmds); clear_exclusive_cxl_commands(mds, exclusive_cmds);
} }
static void unregister_nvdimm(void *nvdimm) static void unregister_nvdimm(void *nvdimm)
@ -65,13 +65,13 @@ static int cxl_nvdimm_probe(struct device *dev)
struct cxl_nvdimm *cxl_nvd = to_cxl_nvdimm(dev); struct cxl_nvdimm *cxl_nvd = to_cxl_nvdimm(dev);
struct cxl_memdev *cxlmd = cxl_nvd->cxlmd; struct cxl_memdev *cxlmd = cxl_nvd->cxlmd;
struct cxl_nvdimm_bridge *cxl_nvb = cxlmd->cxl_nvb; struct cxl_nvdimm_bridge *cxl_nvb = cxlmd->cxl_nvb;
struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
unsigned long flags = 0, cmd_mask = 0; unsigned long flags = 0, cmd_mask = 0;
struct cxl_dev_state *cxlds = cxlmd->cxlds;
struct nvdimm *nvdimm; struct nvdimm *nvdimm;
int rc; int rc;
set_exclusive_cxl_commands(cxlds, exclusive_cmds); set_exclusive_cxl_commands(mds, exclusive_cmds);
rc = devm_add_action_or_reset(dev, clear_exclusive, cxlds); rc = devm_add_action_or_reset(dev, clear_exclusive, mds);
if (rc) if (rc)
return rc; return rc;
@ -100,22 +100,23 @@ static struct cxl_driver cxl_nvdimm_driver = {
}, },
}; };
static int cxl_pmem_get_config_size(struct cxl_dev_state *cxlds, static int cxl_pmem_get_config_size(struct cxl_memdev_state *mds,
struct nd_cmd_get_config_size *cmd, struct nd_cmd_get_config_size *cmd,
unsigned int buf_len) unsigned int buf_len)
{ {
if (sizeof(*cmd) > buf_len) if (sizeof(*cmd) > buf_len)
return -EINVAL; return -EINVAL;
*cmd = (struct nd_cmd_get_config_size) { *cmd = (struct nd_cmd_get_config_size){
.config_size = cxlds->lsa_size, .config_size = mds->lsa_size,
.max_xfer = cxlds->payload_size - sizeof(struct cxl_mbox_set_lsa), .max_xfer =
mds->payload_size - sizeof(struct cxl_mbox_set_lsa),
}; };
return 0; return 0;
} }
static int cxl_pmem_get_config_data(struct cxl_dev_state *cxlds, static int cxl_pmem_get_config_data(struct cxl_memdev_state *mds,
struct nd_cmd_get_config_data_hdr *cmd, struct nd_cmd_get_config_data_hdr *cmd,
unsigned int buf_len) unsigned int buf_len)
{ {
@ -140,13 +141,13 @@ static int cxl_pmem_get_config_data(struct cxl_dev_state *cxlds,
.payload_out = cmd->out_buf, .payload_out = cmd->out_buf,
}; };
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
cmd->status = 0; cmd->status = 0;
return rc; return rc;
} }
static int cxl_pmem_set_config_data(struct cxl_dev_state *cxlds, static int cxl_pmem_set_config_data(struct cxl_memdev_state *mds,
struct nd_cmd_set_config_hdr *cmd, struct nd_cmd_set_config_hdr *cmd,
unsigned int buf_len) unsigned int buf_len)
{ {
@ -176,7 +177,7 @@ static int cxl_pmem_set_config_data(struct cxl_dev_state *cxlds,
.size_in = struct_size(set_lsa, data, cmd->in_length), .size_in = struct_size(set_lsa, data, cmd->in_length),
}; };
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
/* /*
* Set "firmware" status (4-packed bytes at the end of the input * Set "firmware" status (4-packed bytes at the end of the input
@ -194,18 +195,18 @@ static int cxl_pmem_nvdimm_ctl(struct nvdimm *nvdimm, unsigned int cmd,
struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm); struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm);
unsigned long cmd_mask = nvdimm_cmd_mask(nvdimm); unsigned long cmd_mask = nvdimm_cmd_mask(nvdimm);
struct cxl_memdev *cxlmd = cxl_nvd->cxlmd; struct cxl_memdev *cxlmd = cxl_nvd->cxlmd;
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
if (!test_bit(cmd, &cmd_mask)) if (!test_bit(cmd, &cmd_mask))
return -ENOTTY; return -ENOTTY;
switch (cmd) { switch (cmd) {
case ND_CMD_GET_CONFIG_SIZE: case ND_CMD_GET_CONFIG_SIZE:
return cxl_pmem_get_config_size(cxlds, buf, buf_len); return cxl_pmem_get_config_size(mds, buf, buf_len);
case ND_CMD_GET_CONFIG_DATA: case ND_CMD_GET_CONFIG_DATA:
return cxl_pmem_get_config_data(cxlds, buf, buf_len); return cxl_pmem_get_config_data(mds, buf, buf_len);
case ND_CMD_SET_CONFIG_DATA: case ND_CMD_SET_CONFIG_DATA:
return cxl_pmem_set_config_data(cxlds, buf, buf_len); return cxl_pmem_set_config_data(mds, buf, buf_len);
default: default:
return -ENOTTY; return -ENOTTY;
} }

View File

@ -14,7 +14,7 @@ static unsigned long cxl_pmem_get_security_flags(struct nvdimm *nvdimm,
{ {
struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm); struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm);
struct cxl_memdev *cxlmd = cxl_nvd->cxlmd; struct cxl_memdev *cxlmd = cxl_nvd->cxlmd;
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
unsigned long security_flags = 0; unsigned long security_flags = 0;
struct cxl_get_security_output { struct cxl_get_security_output {
__le32 flags; __le32 flags;
@ -29,7 +29,7 @@ static unsigned long cxl_pmem_get_security_flags(struct nvdimm *nvdimm,
.payload_out = &out, .payload_out = &out,
}; };
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc < 0) if (rc < 0)
return 0; return 0;
@ -67,7 +67,7 @@ static int cxl_pmem_security_change_key(struct nvdimm *nvdimm,
{ {
struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm); struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm);
struct cxl_memdev *cxlmd = cxl_nvd->cxlmd; struct cxl_memdev *cxlmd = cxl_nvd->cxlmd;
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
struct cxl_set_pass set_pass; struct cxl_set_pass set_pass;
@ -84,7 +84,7 @@ static int cxl_pmem_security_change_key(struct nvdimm *nvdimm,
.payload_in = &set_pass, .payload_in = &set_pass,
}; };
return cxl_internal_send_cmd(cxlds, &mbox_cmd); return cxl_internal_send_cmd(mds, &mbox_cmd);
} }
static int __cxl_pmem_security_disable(struct nvdimm *nvdimm, static int __cxl_pmem_security_disable(struct nvdimm *nvdimm,
@ -93,7 +93,7 @@ static int __cxl_pmem_security_disable(struct nvdimm *nvdimm,
{ {
struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm); struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm);
struct cxl_memdev *cxlmd = cxl_nvd->cxlmd; struct cxl_memdev *cxlmd = cxl_nvd->cxlmd;
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
struct cxl_disable_pass dis_pass; struct cxl_disable_pass dis_pass;
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
@ -109,7 +109,7 @@ static int __cxl_pmem_security_disable(struct nvdimm *nvdimm,
.payload_in = &dis_pass, .payload_in = &dis_pass,
}; };
return cxl_internal_send_cmd(cxlds, &mbox_cmd); return cxl_internal_send_cmd(mds, &mbox_cmd);
} }
static int cxl_pmem_security_disable(struct nvdimm *nvdimm, static int cxl_pmem_security_disable(struct nvdimm *nvdimm,
@ -128,12 +128,12 @@ static int cxl_pmem_security_freeze(struct nvdimm *nvdimm)
{ {
struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm); struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm);
struct cxl_memdev *cxlmd = cxl_nvd->cxlmd; struct cxl_memdev *cxlmd = cxl_nvd->cxlmd;
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
struct cxl_mbox_cmd mbox_cmd = { struct cxl_mbox_cmd mbox_cmd = {
.opcode = CXL_MBOX_OP_FREEZE_SECURITY, .opcode = CXL_MBOX_OP_FREEZE_SECURITY,
}; };
return cxl_internal_send_cmd(cxlds, &mbox_cmd); return cxl_internal_send_cmd(mds, &mbox_cmd);
} }
static int cxl_pmem_security_unlock(struct nvdimm *nvdimm, static int cxl_pmem_security_unlock(struct nvdimm *nvdimm,
@ -141,7 +141,7 @@ static int cxl_pmem_security_unlock(struct nvdimm *nvdimm,
{ {
struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm); struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm);
struct cxl_memdev *cxlmd = cxl_nvd->cxlmd; struct cxl_memdev *cxlmd = cxl_nvd->cxlmd;
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
u8 pass[NVDIMM_PASSPHRASE_LEN]; u8 pass[NVDIMM_PASSPHRASE_LEN];
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
int rc; int rc;
@ -153,7 +153,7 @@ static int cxl_pmem_security_unlock(struct nvdimm *nvdimm,
.payload_in = pass, .payload_in = pass,
}; };
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc < 0) if (rc < 0)
return rc; return rc;
@ -166,7 +166,7 @@ static int cxl_pmem_security_passphrase_erase(struct nvdimm *nvdimm,
{ {
struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm); struct cxl_nvdimm *cxl_nvd = nvdimm_provider_data(nvdimm);
struct cxl_memdev *cxlmd = cxl_nvd->cxlmd; struct cxl_memdev *cxlmd = cxl_nvd->cxlmd;
struct cxl_dev_state *cxlds = cxlmd->cxlds; struct cxl_memdev_state *mds = to_cxl_memdev_state(cxlmd->cxlds);
struct cxl_mbox_cmd mbox_cmd; struct cxl_mbox_cmd mbox_cmd;
struct cxl_pass_erase erase; struct cxl_pass_erase erase;
int rc; int rc;
@ -182,7 +182,7 @@ static int cxl_pmem_security_passphrase_erase(struct nvdimm *nvdimm,
.payload_in = &erase, .payload_in = &erase,
}; };
rc = cxl_internal_send_cmd(cxlds, &mbox_cmd); rc = cxl_internal_send_cmd(mds, &mbox_cmd);
if (rc < 0) if (rc < 0)
return rc; return rc;

View File

@ -102,7 +102,7 @@ struct mock_event_log {
}; };
struct mock_event_store { struct mock_event_store {
struct cxl_dev_state *cxlds; struct cxl_memdev_state *mds;
struct mock_event_log mock_logs[CXL_EVENT_TYPE_MAX]; struct mock_event_log mock_logs[CXL_EVENT_TYPE_MAX];
u32 ev_status; u32 ev_status;
}; };
@ -291,7 +291,7 @@ static void cxl_mock_event_trigger(struct device *dev)
event_reset_log(log); event_reset_log(log);
} }
cxl_mem_get_event_records(mes->cxlds, mes->ev_status); cxl_mem_get_event_records(mes->mds, mes->ev_status);
} }
struct cxl_event_record_raw maint_needed = { struct cxl_event_record_raw maint_needed = {
@ -451,7 +451,7 @@ static int mock_gsl(struct cxl_mbox_cmd *cmd)
return 0; return 0;
} }
static int mock_get_log(struct cxl_dev_state *cxlds, struct cxl_mbox_cmd *cmd) static int mock_get_log(struct cxl_memdev_state *mds, struct cxl_mbox_cmd *cmd)
{ {
struct cxl_mbox_get_log *gl = cmd->payload_in; struct cxl_mbox_get_log *gl = cmd->payload_in;
u32 offset = le32_to_cpu(gl->offset); u32 offset = le32_to_cpu(gl->offset);
@ -461,7 +461,7 @@ static int mock_get_log(struct cxl_dev_state *cxlds, struct cxl_mbox_cmd *cmd)
if (cmd->size_in < sizeof(*gl)) if (cmd->size_in < sizeof(*gl))
return -EINVAL; return -EINVAL;
if (length > cxlds->payload_size) if (length > mds->payload_size)
return -EINVAL; return -EINVAL;
if (offset + length > sizeof(mock_cel)) if (offset + length > sizeof(mock_cel))
return -EINVAL; return -EINVAL;
@ -1105,8 +1105,10 @@ static struct attribute *cxl_mock_mem_core_attrs[] = {
}; };
ATTRIBUTE_GROUPS(cxl_mock_mem_core); ATTRIBUTE_GROUPS(cxl_mock_mem_core);
static int cxl_mock_mbox_send(struct cxl_dev_state *cxlds, struct cxl_mbox_cmd *cmd) static int cxl_mock_mbox_send(struct cxl_memdev_state *mds,
struct cxl_mbox_cmd *cmd)
{ {
struct cxl_dev_state *cxlds = &mds->cxlds;
struct device *dev = cxlds->dev; struct device *dev = cxlds->dev;
struct cxl_mockmem_data *mdata = dev_get_drvdata(dev); struct cxl_mockmem_data *mdata = dev_get_drvdata(dev);
int rc = -EIO; int rc = -EIO;
@ -1119,7 +1121,7 @@ static int cxl_mock_mbox_send(struct cxl_dev_state *cxlds, struct cxl_mbox_cmd *
rc = mock_gsl(cmd); rc = mock_gsl(cmd);
break; break;
case CXL_MBOX_OP_GET_LOG: case CXL_MBOX_OP_GET_LOG:
rc = mock_get_log(cxlds, cmd); rc = mock_get_log(mds, cmd);
break; break;
case CXL_MBOX_OP_IDENTIFY: case CXL_MBOX_OP_IDENTIFY:
if (cxlds->rcd) if (cxlds->rcd)
@ -1207,6 +1209,7 @@ static int cxl_mock_mem_probe(struct platform_device *pdev)
{ {
struct device *dev = &pdev->dev; struct device *dev = &pdev->dev;
struct cxl_memdev *cxlmd; struct cxl_memdev *cxlmd;
struct cxl_memdev_state *mds;
struct cxl_dev_state *cxlds; struct cxl_dev_state *cxlds;
struct cxl_mockmem_data *mdata; struct cxl_mockmem_data *mdata;
int rc; int rc;
@ -1223,48 +1226,50 @@ static int cxl_mock_mem_probe(struct platform_device *pdev)
if (rc) if (rc)
return rc; return rc;
cxlds = cxl_dev_state_create(dev); mds = cxl_memdev_state_create(dev);
if (IS_ERR(cxlds)) if (IS_ERR(mds))
return PTR_ERR(cxlds); return PTR_ERR(mds);
mds->mbox_send = cxl_mock_mbox_send;
mds->payload_size = SZ_4K;
mds->event.buf = (struct cxl_get_event_payload *) mdata->event_buf;
cxlds = &mds->cxlds;
cxlds->serial = pdev->id; cxlds->serial = pdev->id;
cxlds->mbox_send = cxl_mock_mbox_send;
cxlds->payload_size = SZ_4K;
cxlds->event.buf = (struct cxl_get_event_payload *) mdata->event_buf;
if (is_rcd(pdev)) { if (is_rcd(pdev)) {
cxlds->rcd = true; cxlds->rcd = true;
cxlds->component_reg_phys = CXL_RESOURCE_NONE; cxlds->component_reg_phys = CXL_RESOURCE_NONE;
} }
rc = cxl_enumerate_cmds(cxlds); rc = cxl_enumerate_cmds(mds);
if (rc) if (rc)
return rc; return rc;
rc = cxl_poison_state_init(cxlds); rc = cxl_poison_state_init(mds);
if (rc) if (rc)
return rc; return rc;
rc = cxl_set_timestamp(cxlds); rc = cxl_set_timestamp(mds);
if (rc) if (rc)
return rc; return rc;
cxlds->media_ready = true; cxlds->media_ready = true;
rc = cxl_dev_state_identify(cxlds); rc = cxl_dev_state_identify(mds);
if (rc) if (rc)
return rc; return rc;
rc = cxl_mem_create_range_info(cxlds); rc = cxl_mem_create_range_info(mds);
if (rc) if (rc)
return rc; return rc;
mdata->mes.cxlds = cxlds; mdata->mes.mds = mds;
cxl_mock_add_event_logs(&mdata->mes); cxl_mock_add_event_logs(&mdata->mes);
cxlmd = devm_cxl_add_memdev(cxlds); cxlmd = devm_cxl_add_memdev(cxlds);
if (IS_ERR(cxlmd)) if (IS_ERR(cxlmd))
return PTR_ERR(cxlmd); return PTR_ERR(cxlmd);
cxl_mem_get_event_records(cxlds, CXLDEV_EVENT_STATUS_ALL); cxl_mem_get_event_records(mds, CXLDEV_EVENT_STATUS_ALL);
return 0; return 0;
} }