Merge branch '6.3/scsi-fixes' into 6.4/scsi-staging
Pull in the fixes branch to resolve an mpi3mr conflict reported by sfr. Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
This commit is contained in:
commit
f467b865cf
@ -1145,10 +1145,12 @@ static int alua_activate(struct scsi_device *sdev,
|
||||
rcu_read_unlock();
|
||||
mutex_unlock(&h->init_mutex);
|
||||
|
||||
if (alua_rtpg_queue(pg, sdev, qdata, true))
|
||||
if (alua_rtpg_queue(pg, sdev, qdata, true)) {
|
||||
fn = NULL;
|
||||
else
|
||||
} else {
|
||||
kfree(qdata);
|
||||
err = SCSI_DH_DEV_OFFLINED;
|
||||
}
|
||||
kref_put(&pg->kref, release_port_group);
|
||||
out:
|
||||
if (fn)
|
||||
|
@ -2525,8 +2525,7 @@ static int interrupt_preinit_v3_hw(struct hisi_hba *hisi_hba)
|
||||
hisi_hba->cq_nvecs = vectors - BASE_VECTORS_V3_HW - hisi_hba->iopoll_q_cnt;
|
||||
shost->nr_hw_queues = hisi_hba->cq_nvecs + hisi_hba->iopoll_q_cnt;
|
||||
|
||||
devm_add_action(&pdev->dev, hisi_sas_v3_free_vectors, pdev);
|
||||
return 0;
|
||||
return devm_add_action(&pdev->dev, hisi_sas_v3_free_vectors, pdev);
|
||||
}
|
||||
|
||||
static int interrupt_init_v3_hw(struct hisi_hba *hisi_hba)
|
||||
|
@ -341,9 +341,6 @@ static void scsi_host_dev_release(struct device *dev)
|
||||
struct Scsi_Host *shost = dev_to_shost(dev);
|
||||
struct device *parent = dev->parent;
|
||||
|
||||
/* In case scsi_remove_host() has not been called. */
|
||||
scsi_proc_hostdir_rm(shost->hostt);
|
||||
|
||||
/* Wait for functions invoked through call_rcu(&scmd->rcu, ...) */
|
||||
rcu_barrier();
|
||||
|
||||
|
@ -7290,6 +7290,8 @@ lpfc_sli4_cgn_params_read(struct lpfc_hba *phba)
|
||||
/* Find out if the FW has a new set of congestion parameters. */
|
||||
len = sizeof(struct lpfc_cgn_param);
|
||||
pdata = kzalloc(len, GFP_KERNEL);
|
||||
if (!pdata)
|
||||
return -ENOMEM;
|
||||
ret = lpfc_read_object(phba, (char *)LPFC_PORT_CFG_NAME,
|
||||
pdata, len);
|
||||
|
||||
|
@ -21862,20 +21862,20 @@ lpfc_get_io_buf_from_private_pool(struct lpfc_hba *phba,
|
||||
static struct lpfc_io_buf *
|
||||
lpfc_get_io_buf_from_expedite_pool(struct lpfc_hba *phba)
|
||||
{
|
||||
struct lpfc_io_buf *lpfc_ncmd;
|
||||
struct lpfc_io_buf *lpfc_ncmd = NULL, *iter;
|
||||
struct lpfc_io_buf *lpfc_ncmd_next;
|
||||
unsigned long iflag;
|
||||
struct lpfc_epd_pool *epd_pool;
|
||||
|
||||
epd_pool = &phba->epd_pool;
|
||||
lpfc_ncmd = NULL;
|
||||
|
||||
spin_lock_irqsave(&epd_pool->lock, iflag);
|
||||
if (epd_pool->count > 0) {
|
||||
list_for_each_entry_safe(lpfc_ncmd, lpfc_ncmd_next,
|
||||
list_for_each_entry_safe(iter, lpfc_ncmd_next,
|
||||
&epd_pool->list, list) {
|
||||
list_del(&lpfc_ncmd->list);
|
||||
list_del(&iter->list);
|
||||
epd_pool->count--;
|
||||
lpfc_ncmd = iter;
|
||||
break;
|
||||
}
|
||||
}
|
||||
@ -22072,10 +22072,6 @@ lpfc_read_object(struct lpfc_hba *phba, char *rdobject, uint32_t *datap,
|
||||
struct lpfc_dmabuf *pcmd;
|
||||
u32 rd_object_name[LPFC_MBX_OBJECT_NAME_LEN_DW] = {0};
|
||||
|
||||
/* sanity check on queue memory */
|
||||
if (!datap)
|
||||
return -ENODEV;
|
||||
|
||||
mbox = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
|
||||
if (!mbox)
|
||||
return -ENOMEM;
|
||||
|
@ -23,8 +23,8 @@
|
||||
/*
|
||||
* MegaRAID SAS Driver meta data
|
||||
*/
|
||||
#define MEGASAS_VERSION "07.719.03.00-rc1"
|
||||
#define MEGASAS_RELDATE "Sep 29, 2021"
|
||||
#define MEGASAS_VERSION "07.725.01.00-rc1"
|
||||
#define MEGASAS_RELDATE "Mar 2, 2023"
|
||||
|
||||
#define MEGASAS_MSIX_NAME_LEN 32
|
||||
|
||||
@ -1519,6 +1519,8 @@ struct megasas_ctrl_info {
|
||||
#define MEGASAS_MAX_LD_IDS (MEGASAS_MAX_LD_CHANNELS * \
|
||||
MEGASAS_MAX_DEV_PER_CHANNEL)
|
||||
|
||||
#define MEGASAS_MAX_SUPPORTED_LD_IDS 240
|
||||
|
||||
#define MEGASAS_MAX_SECTORS (2*1024)
|
||||
#define MEGASAS_MAX_SECTORS_IEEE (2*128)
|
||||
#define MEGASAS_DBG_LVL 1
|
||||
@ -1758,7 +1760,8 @@ union megasas_sgl_frame {
|
||||
typedef union _MFI_CAPABILITIES {
|
||||
struct {
|
||||
#if defined(__BIG_ENDIAN_BITFIELD)
|
||||
u32 reserved:16;
|
||||
u32 reserved:15;
|
||||
u32 support_memdump:1;
|
||||
u32 support_fw_exposed_dev_list:1;
|
||||
u32 support_nvme_passthru:1;
|
||||
u32 support_64bit_mode:1;
|
||||
@ -1792,7 +1795,8 @@ typedef union _MFI_CAPABILITIES {
|
||||
u32 support_64bit_mode:1;
|
||||
u32 support_nvme_passthru:1;
|
||||
u32 support_fw_exposed_dev_list:1;
|
||||
u32 reserved:16;
|
||||
u32 support_memdump:1;
|
||||
u32 reserved:15;
|
||||
#endif
|
||||
} mfi_capabilities;
|
||||
__le32 reg;
|
||||
|
@ -3298,7 +3298,7 @@ fw_crash_buffer_show(struct device *cdev,
|
||||
|
||||
spin_lock_irqsave(&instance->crashdump_lock, flags);
|
||||
buff_offset = instance->fw_crash_buffer_offset;
|
||||
if (!instance->crash_dump_buf &&
|
||||
if (!instance->crash_dump_buf ||
|
||||
!((instance->fw_crash_state == AVAILABLE) ||
|
||||
(instance->fw_crash_state == COPYING))) {
|
||||
dev_err(&instance->pdev->dev,
|
||||
|
@ -358,7 +358,7 @@ u8 MR_ValidateMapInfo(struct megasas_instance *instance, u64 map_id)
|
||||
ld = MR_TargetIdToLdGet(i, drv_map);
|
||||
|
||||
/* For non existing VDs, iterate to next VD*/
|
||||
if (ld >= (MAX_LOGICAL_DRIVES_EXT - 1))
|
||||
if (ld >= MEGASAS_MAX_SUPPORTED_LD_IDS)
|
||||
continue;
|
||||
|
||||
raid = MR_LdRaidGet(ld, drv_map);
|
||||
|
@ -1201,6 +1201,9 @@ megasas_ioc_init_fusion(struct megasas_instance *instance)
|
||||
drv_ops->mfi_capabilities.support_nvme_passthru = 1;
|
||||
drv_ops->mfi_capabilities.support_fw_exposed_dev_list = 1;
|
||||
|
||||
if (reset_devices)
|
||||
drv_ops->mfi_capabilities.support_memdump = 1;
|
||||
|
||||
if (instance->consistent_mask_64bit)
|
||||
drv_ops->mfi_capabilities.support_64bit_mode = 1;
|
||||
|
||||
@ -4768,7 +4771,7 @@ int megasas_task_abort_fusion(struct scsi_cmnd *scmd)
|
||||
devhandle = megasas_get_tm_devhandle(scmd->device);
|
||||
|
||||
if (devhandle == (u16)ULONG_MAX) {
|
||||
ret = SUCCESS;
|
||||
ret = FAILED;
|
||||
sdev_printk(KERN_INFO, scmd->device,
|
||||
"task abort issued for invalid devhandle\n");
|
||||
mutex_unlock(&instance->reset_mutex);
|
||||
@ -4838,7 +4841,7 @@ int megasas_reset_target_fusion(struct scsi_cmnd *scmd)
|
||||
devhandle = megasas_get_tm_devhandle(scmd->device);
|
||||
|
||||
if (devhandle == (u16)ULONG_MAX) {
|
||||
ret = SUCCESS;
|
||||
ret = FAILED;
|
||||
sdev_printk(KERN_INFO, scmd->device,
|
||||
"target reset issued for invalid devhandle\n");
|
||||
mutex_unlock(&instance->reset_mutex);
|
||||
|
@ -909,6 +909,7 @@ struct scmd_priv {
|
||||
* @admin_reply_ephase:Admin reply queue expected phase
|
||||
* @admin_reply_base: Admin reply queue base virtual address
|
||||
* @admin_reply_dma: Admin reply queue base dma address
|
||||
* @admin_reply_q_in_use: Queue is handled by poll/ISR
|
||||
* @ready_timeout: Controller ready timeout
|
||||
* @intr_info: Interrupt cookie pointer
|
||||
* @intr_info_count: Number of interrupt cookies
|
||||
@ -1062,6 +1063,7 @@ struct mpi3mr_ioc {
|
||||
u8 admin_reply_ephase;
|
||||
void *admin_reply_base;
|
||||
dma_addr_t admin_reply_dma;
|
||||
atomic_t admin_reply_q_in_use;
|
||||
|
||||
u32 ready_timeout;
|
||||
|
||||
@ -1397,4 +1399,7 @@ void mpi3mr_add_event_wait_for_device_refresh(struct mpi3mr_ioc *mrioc);
|
||||
void mpi3mr_flush_drv_cmds(struct mpi3mr_ioc *mrioc);
|
||||
void mpi3mr_flush_cmds_for_unrecovered_controller(struct mpi3mr_ioc *mrioc);
|
||||
void mpi3mr_free_enclosure_list(struct mpi3mr_ioc *mrioc);
|
||||
int mpi3mr_process_admin_reply_q(struct mpi3mr_ioc *mrioc);
|
||||
void mpi3mr_expander_node_remove(struct mpi3mr_ioc *mrioc,
|
||||
struct mpi3mr_sas_node *sas_expander);
|
||||
#endif /*MPI3MR_H_INCLUDED*/
|
||||
|
@ -886,7 +886,7 @@ static int mpi3mr_build_nvme_prp(struct mpi3mr_ioc *mrioc,
|
||||
* each time through the loop.
|
||||
*/
|
||||
*prp_entry = cpu_to_le64(dma_addr);
|
||||
if (*prp1_entry & sgemod_mask) {
|
||||
if (*prp_entry & sgemod_mask) {
|
||||
dprint_bsg_err(mrioc,
|
||||
"%s: PRP address collides with SGE modifier\n",
|
||||
__func__);
|
||||
@ -895,7 +895,7 @@ static int mpi3mr_build_nvme_prp(struct mpi3mr_ioc *mrioc,
|
||||
*prp_entry &= ~sgemod_mask;
|
||||
*prp_entry |= sgemod_val;
|
||||
prp_entry++;
|
||||
prp_entry_dma++;
|
||||
prp_entry_dma += prp_size;
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -415,7 +415,7 @@ out:
|
||||
le64_to_cpu(scsi_reply->sense_data_buffer_address));
|
||||
}
|
||||
|
||||
static int mpi3mr_process_admin_reply_q(struct mpi3mr_ioc *mrioc)
|
||||
int mpi3mr_process_admin_reply_q(struct mpi3mr_ioc *mrioc)
|
||||
{
|
||||
u32 exp_phase = mrioc->admin_reply_ephase;
|
||||
u32 admin_reply_ci = mrioc->admin_reply_ci;
|
||||
@ -423,12 +423,17 @@ static int mpi3mr_process_admin_reply_q(struct mpi3mr_ioc *mrioc)
|
||||
u64 reply_dma = 0;
|
||||
struct mpi3_default_reply_descriptor *reply_desc;
|
||||
|
||||
if (!atomic_add_unless(&mrioc->admin_reply_q_in_use, 1, 1))
|
||||
return 0;
|
||||
|
||||
reply_desc = (struct mpi3_default_reply_descriptor *)mrioc->admin_reply_base +
|
||||
admin_reply_ci;
|
||||
|
||||
if ((le16_to_cpu(reply_desc->reply_flags) &
|
||||
MPI3_REPLY_DESCRIPT_FLAGS_PHASE_MASK) != exp_phase)
|
||||
MPI3_REPLY_DESCRIPT_FLAGS_PHASE_MASK) != exp_phase) {
|
||||
atomic_dec(&mrioc->admin_reply_q_in_use);
|
||||
return 0;
|
||||
}
|
||||
|
||||
do {
|
||||
if (mrioc->unrecoverable)
|
||||
@ -454,6 +459,7 @@ static int mpi3mr_process_admin_reply_q(struct mpi3mr_ioc *mrioc)
|
||||
writel(admin_reply_ci, &mrioc->sysif_regs->admin_reply_queue_ci);
|
||||
mrioc->admin_reply_ci = admin_reply_ci;
|
||||
mrioc->admin_reply_ephase = exp_phase;
|
||||
atomic_dec(&mrioc->admin_reply_q_in_use);
|
||||
|
||||
return num_admin_replies;
|
||||
}
|
||||
@ -1192,7 +1198,7 @@ mpi3mr_revalidate_factsdata(struct mpi3mr_ioc *mrioc)
|
||||
*/
|
||||
static int mpi3mr_bring_ioc_ready(struct mpi3mr_ioc *mrioc)
|
||||
{
|
||||
u32 ioc_config, ioc_status, timeout;
|
||||
u32 ioc_config, ioc_status, timeout, host_diagnostic;
|
||||
int retval = 0;
|
||||
enum mpi3mr_iocstate ioc_state;
|
||||
u64 base_info;
|
||||
@ -1246,6 +1252,23 @@ static int mpi3mr_bring_ioc_ready(struct mpi3mr_ioc *mrioc)
|
||||
retval, mpi3mr_iocstate_name(ioc_state));
|
||||
}
|
||||
if (ioc_state != MRIOC_STATE_RESET) {
|
||||
if (ioc_state == MRIOC_STATE_FAULT) {
|
||||
timeout = MPI3_SYSIF_DIAG_SAVE_TIMEOUT * 10;
|
||||
mpi3mr_print_fault_info(mrioc);
|
||||
do {
|
||||
host_diagnostic =
|
||||
readl(&mrioc->sysif_regs->host_diagnostic);
|
||||
if (!(host_diagnostic &
|
||||
MPI3_SYSIF_HOST_DIAG_SAVE_IN_PROGRESS))
|
||||
break;
|
||||
if (!pci_device_is_present(mrioc->pdev)) {
|
||||
mrioc->unrecoverable = 1;
|
||||
ioc_err(mrioc, "controller is not present at the bringup\n");
|
||||
goto out_device_not_present;
|
||||
}
|
||||
msleep(100);
|
||||
} while (--timeout);
|
||||
}
|
||||
mpi3mr_print_fault_info(mrioc);
|
||||
ioc_info(mrioc, "issuing soft reset to bring to reset state\n");
|
||||
retval = mpi3mr_issue_reset(mrioc,
|
||||
@ -2603,6 +2626,7 @@ static int mpi3mr_setup_admin_qpair(struct mpi3mr_ioc *mrioc)
|
||||
MPI3MR_ADMIN_REPLY_FRAME_SZ;
|
||||
mrioc->admin_reply_ci = 0;
|
||||
mrioc->admin_reply_ephase = 1;
|
||||
atomic_set(&mrioc->admin_reply_q_in_use, 0);
|
||||
|
||||
if (!mrioc->admin_req_base) {
|
||||
mrioc->admin_req_base = dma_alloc_coherent(&mrioc->pdev->dev,
|
||||
@ -3811,27 +3835,34 @@ retry_init:
|
||||
|
||||
mpi3mr_print_ioc_info(mrioc);
|
||||
|
||||
dprint_init(mrioc, "allocating config page buffers\n");
|
||||
mrioc->cfg_page = dma_alloc_coherent(&mrioc->pdev->dev,
|
||||
MPI3MR_DEFAULT_CFG_PAGE_SZ, &mrioc->cfg_page_dma, GFP_KERNEL);
|
||||
if (!mrioc->cfg_page)
|
||||
goto out_failed_noretry;
|
||||
|
||||
mrioc->cfg_page_sz = MPI3MR_DEFAULT_CFG_PAGE_SZ;
|
||||
|
||||
retval = mpi3mr_alloc_reply_sense_bufs(mrioc);
|
||||
if (retval) {
|
||||
ioc_err(mrioc,
|
||||
"%s :Failed to allocated reply sense buffers %d\n",
|
||||
__func__, retval);
|
||||
goto out_failed_noretry;
|
||||
if (!mrioc->cfg_page) {
|
||||
dprint_init(mrioc, "allocating config page buffers\n");
|
||||
mrioc->cfg_page_sz = MPI3MR_DEFAULT_CFG_PAGE_SZ;
|
||||
mrioc->cfg_page = dma_alloc_coherent(&mrioc->pdev->dev,
|
||||
mrioc->cfg_page_sz, &mrioc->cfg_page_dma, GFP_KERNEL);
|
||||
if (!mrioc->cfg_page) {
|
||||
retval = -1;
|
||||
goto out_failed_noretry;
|
||||
}
|
||||
}
|
||||
|
||||
retval = mpi3mr_alloc_chain_bufs(mrioc);
|
||||
if (retval) {
|
||||
ioc_err(mrioc, "Failed to allocated chain buffers %d\n",
|
||||
retval);
|
||||
goto out_failed_noretry;
|
||||
if (!mrioc->init_cmds.reply) {
|
||||
retval = mpi3mr_alloc_reply_sense_bufs(mrioc);
|
||||
if (retval) {
|
||||
ioc_err(mrioc,
|
||||
"%s :Failed to allocated reply sense buffers %d\n",
|
||||
__func__, retval);
|
||||
goto out_failed_noretry;
|
||||
}
|
||||
}
|
||||
|
||||
if (!mrioc->chain_sgl_list) {
|
||||
retval = mpi3mr_alloc_chain_bufs(mrioc);
|
||||
if (retval) {
|
||||
ioc_err(mrioc, "Failed to allocated chain buffers %d\n",
|
||||
retval);
|
||||
goto out_failed_noretry;
|
||||
}
|
||||
}
|
||||
|
||||
retval = mpi3mr_issue_iocinit(mrioc);
|
||||
@ -3877,8 +3908,10 @@ retry_init:
|
||||
dprint_init(mrioc, "allocating memory for throttle groups\n");
|
||||
sz = sizeof(struct mpi3mr_throttle_group_info);
|
||||
mrioc->throttle_groups = kcalloc(mrioc->num_io_throttle_group, sz, GFP_KERNEL);
|
||||
if (!mrioc->throttle_groups)
|
||||
if (!mrioc->throttle_groups) {
|
||||
retval = -1;
|
||||
goto out_failed_noretry;
|
||||
}
|
||||
}
|
||||
|
||||
retval = mpi3mr_enable_events(mrioc);
|
||||
@ -3898,6 +3931,7 @@ out_failed:
|
||||
mpi3mr_memset_buffers(mrioc);
|
||||
goto retry_init;
|
||||
}
|
||||
retval = -1;
|
||||
out_failed_noretry:
|
||||
ioc_err(mrioc, "controller initialization failed\n");
|
||||
mpi3mr_issue_reset(mrioc, MPI3_SYSIF_HOST_DIAG_RESET_ACTION_DIAG_FAULT,
|
||||
@ -4010,6 +4044,7 @@ retry_init:
|
||||
ioc_err(mrioc,
|
||||
"cannot create minimum number of operational queues expected:%d created:%d\n",
|
||||
mrioc->shost->nr_hw_queues, mrioc->num_op_reply_q);
|
||||
retval = -1;
|
||||
goto out_failed_noretry;
|
||||
}
|
||||
|
||||
@ -4076,6 +4111,7 @@ out_failed:
|
||||
mpi3mr_memset_buffers(mrioc);
|
||||
goto retry_init;
|
||||
}
|
||||
retval = -1;
|
||||
out_failed_noretry:
|
||||
ioc_err(mrioc, "controller %s is failed\n",
|
||||
(is_resume)?"resume":"re-initialization");
|
||||
@ -4153,6 +4189,7 @@ void mpi3mr_memset_buffers(struct mpi3mr_ioc *mrioc)
|
||||
memset(mrioc->admin_req_base, 0, mrioc->admin_req_q_sz);
|
||||
if (mrioc->admin_reply_base)
|
||||
memset(mrioc->admin_reply_base, 0, mrioc->admin_reply_q_sz);
|
||||
atomic_set(&mrioc->admin_reply_q_in_use, 0);
|
||||
|
||||
if (mrioc->init_cmds.reply) {
|
||||
memset(mrioc->init_cmds.reply, 0, sizeof(*mrioc->init_cmds.reply));
|
||||
@ -4348,13 +4385,20 @@ void mpi3mr_free_mem(struct mpi3mr_ioc *mrioc)
|
||||
mrioc->admin_req_base, mrioc->admin_req_dma);
|
||||
mrioc->admin_req_base = NULL;
|
||||
}
|
||||
|
||||
if (mrioc->cfg_page) {
|
||||
dma_free_coherent(&mrioc->pdev->dev, mrioc->cfg_page_sz,
|
||||
mrioc->cfg_page, mrioc->cfg_page_dma);
|
||||
mrioc->cfg_page = NULL;
|
||||
}
|
||||
if (mrioc->pel_seqnum_virt) {
|
||||
dma_free_coherent(&mrioc->pdev->dev, mrioc->pel_seqnum_sz,
|
||||
mrioc->pel_seqnum_virt, mrioc->pel_seqnum_dma);
|
||||
mrioc->pel_seqnum_virt = NULL;
|
||||
}
|
||||
|
||||
kfree(mrioc->throttle_groups);
|
||||
mrioc->throttle_groups = NULL;
|
||||
|
||||
kfree(mrioc->logdata_buf);
|
||||
mrioc->logdata_buf = NULL;
|
||||
|
||||
|
@ -3740,6 +3740,7 @@ int mpi3mr_issue_tm(struct mpi3mr_ioc *mrioc, u8 tm_type,
|
||||
mpi3mr_poll_pend_io_completions(mrioc);
|
||||
mpi3mr_ioc_enable_intr(mrioc);
|
||||
mpi3mr_poll_pend_io_completions(mrioc);
|
||||
mpi3mr_process_admin_reply_q(mrioc);
|
||||
}
|
||||
switch (tm_type) {
|
||||
case MPI3_SCSITASKMGMT_TASKTYPE_TARGET_RESET:
|
||||
@ -5115,6 +5116,8 @@ static void mpi3mr_remove(struct pci_dev *pdev)
|
||||
struct workqueue_struct *wq;
|
||||
unsigned long flags;
|
||||
struct mpi3mr_tgt_dev *tgtdev, *tgtdev_next;
|
||||
struct mpi3mr_hba_port *port, *hba_port_next;
|
||||
struct mpi3mr_sas_node *sas_expander, *sas_expander_next;
|
||||
|
||||
if (!shost)
|
||||
return;
|
||||
@ -5154,6 +5157,28 @@ static void mpi3mr_remove(struct pci_dev *pdev)
|
||||
mpi3mr_free_mem(mrioc);
|
||||
mpi3mr_cleanup_resources(mrioc);
|
||||
|
||||
spin_lock_irqsave(&mrioc->sas_node_lock, flags);
|
||||
list_for_each_entry_safe_reverse(sas_expander, sas_expander_next,
|
||||
&mrioc->sas_expander_list, list) {
|
||||
spin_unlock_irqrestore(&mrioc->sas_node_lock, flags);
|
||||
mpi3mr_expander_node_remove(mrioc, sas_expander);
|
||||
spin_lock_irqsave(&mrioc->sas_node_lock, flags);
|
||||
}
|
||||
list_for_each_entry_safe(port, hba_port_next, &mrioc->hba_port_table_list, list) {
|
||||
ioc_info(mrioc,
|
||||
"removing hba_port entry: %p port: %d from hba_port list\n",
|
||||
port, port->port_id);
|
||||
list_del(&port->list);
|
||||
kfree(port);
|
||||
}
|
||||
spin_unlock_irqrestore(&mrioc->sas_node_lock, flags);
|
||||
|
||||
if (mrioc->sas_hba.num_phys) {
|
||||
kfree(mrioc->sas_hba.phy);
|
||||
mrioc->sas_hba.phy = NULL;
|
||||
mrioc->sas_hba.num_phys = 0;
|
||||
}
|
||||
|
||||
spin_lock(&mrioc_list_lock);
|
||||
list_del(&mrioc->list);
|
||||
spin_unlock(&mrioc_list_lock);
|
||||
|
@ -9,9 +9,6 @@
|
||||
|
||||
#include "mpi3mr.h"
|
||||
|
||||
static void mpi3mr_expander_node_remove(struct mpi3mr_ioc *mrioc,
|
||||
struct mpi3mr_sas_node *sas_expander);
|
||||
|
||||
/**
|
||||
* mpi3mr_post_transport_req - Issue transport requests and wait
|
||||
* @mrioc: Adapter instance reference
|
||||
@ -1552,7 +1549,8 @@ static void mpi3mr_sas_port_remove(struct mpi3mr_ioc *mrioc, u64 sas_address,
|
||||
|
||||
list_for_each_entry_safe(mr_sas_phy, next_phy,
|
||||
&mr_sas_port->phy_list, port_siblings) {
|
||||
if ((mrioc->logging_level & MPI3_DEBUG_TRANSPORT_INFO))
|
||||
if ((!mrioc->stop_drv_processing) &&
|
||||
(mrioc->logging_level & MPI3_DEBUG_TRANSPORT_INFO))
|
||||
dev_info(&mr_sas_port->port->dev,
|
||||
"remove: sas_address(0x%016llx), phy(%d)\n",
|
||||
(unsigned long long)
|
||||
@ -2163,7 +2161,7 @@ out_fail:
|
||||
*
|
||||
* Return nothing.
|
||||
*/
|
||||
static void mpi3mr_expander_node_remove(struct mpi3mr_ioc *mrioc,
|
||||
void mpi3mr_expander_node_remove(struct mpi3mr_ioc *mrioc,
|
||||
struct mpi3mr_sas_node *sas_expander)
|
||||
{
|
||||
struct mpi3mr_sas_port *mr_sas_port, *next;
|
||||
@ -2357,15 +2355,16 @@ int mpi3mr_report_tgtdev_to_sas_transport(struct mpi3mr_ioc *mrioc,
|
||||
tgtdev->host_exposed = 1;
|
||||
if (!mpi3mr_sas_port_add(mrioc, tgtdev->dev_handle,
|
||||
sas_address_parent, hba_port)) {
|
||||
tgtdev->host_exposed = 0;
|
||||
retval = -1;
|
||||
} else if ((!tgtdev->starget)) {
|
||||
if (!mrioc->is_driver_loading)
|
||||
} else if ((!tgtdev->starget) && (!mrioc->is_driver_loading)) {
|
||||
mpi3mr_sas_port_remove(mrioc, sas_address,
|
||||
sas_address_parent, hba_port);
|
||||
tgtdev->host_exposed = 0;
|
||||
retval = -1;
|
||||
}
|
||||
if (retval) {
|
||||
tgtdev->dev_spec.sas_sata_inf.hba_port = NULL;
|
||||
tgtdev->host_exposed = 0;
|
||||
}
|
||||
return retval;
|
||||
}
|
||||
|
||||
@ -2394,6 +2393,7 @@ void mpi3mr_remove_tgtdev_from_sas_transport(struct mpi3mr_ioc *mrioc,
|
||||
mpi3mr_sas_port_remove(mrioc, sas_address, sas_address_parent,
|
||||
hba_port);
|
||||
tgtdev->host_exposed = 0;
|
||||
tgtdev->dev_spec.sas_sata_inf.hba_port = NULL;
|
||||
}
|
||||
|
||||
/**
|
||||
@ -2450,7 +2450,7 @@ static u8 mpi3mr_get_port_id_by_rphy(struct mpi3mr_ioc *mrioc, struct sas_rphy *
|
||||
|
||||
tgtdev = __mpi3mr_get_tgtdev_by_addr_and_rphy(mrioc,
|
||||
rphy->identify.sas_address, rphy);
|
||||
if (tgtdev) {
|
||||
if (tgtdev && tgtdev->dev_spec.sas_sata_inf.hba_port) {
|
||||
port_id =
|
||||
tgtdev->dev_spec.sas_sata_inf.hba_port->port_id;
|
||||
mpi3mr_tgtdev_put(tgtdev);
|
||||
|
@ -6611,11 +6611,6 @@ _base_allocate_memory_pools(struct MPT3SAS_ADAPTER *ioc)
|
||||
else if (rc == -EAGAIN)
|
||||
goto try_32bit_dma;
|
||||
total_sz += sense_sz;
|
||||
ioc_info(ioc,
|
||||
"sense pool(0x%p)- dma(0x%llx): depth(%d),"
|
||||
"element_size(%d), pool_size(%d kB)\n",
|
||||
ioc->sense, (unsigned long long)ioc->sense_dma, ioc->scsiio_depth,
|
||||
SCSI_SENSE_BUFFERSIZE, sz / 1024);
|
||||
/* reply pool, 4 byte align */
|
||||
sz = ioc->reply_free_queue_depth * ioc->reply_sz;
|
||||
rc = _base_allocate_reply_pool(ioc, sz);
|
||||
|
@ -785,7 +785,7 @@ mpt3sas_transport_port_add(struct MPT3SAS_ADAPTER *ioc, u16 handle,
|
||||
goto out_fail;
|
||||
}
|
||||
port = sas_port_alloc_num(sas_node->parent_dev);
|
||||
if ((sas_port_add(port))) {
|
||||
if (!port || (sas_port_add(port))) {
|
||||
ioc_err(ioc, "failure at %s:%d/%s()!\n",
|
||||
__FILE__, __LINE__, __func__);
|
||||
goto out_fail;
|
||||
@ -824,6 +824,12 @@ mpt3sas_transport_port_add(struct MPT3SAS_ADAPTER *ioc, u16 handle,
|
||||
mpt3sas_port->remote_identify.sas_address;
|
||||
}
|
||||
|
||||
if (!rphy) {
|
||||
ioc_err(ioc, "failure at %s:%d/%s()!\n",
|
||||
__FILE__, __LINE__, __func__);
|
||||
goto out_delete_port;
|
||||
}
|
||||
|
||||
rphy->identify = mpt3sas_port->remote_identify;
|
||||
|
||||
if ((sas_rphy_add(rphy))) {
|
||||
@ -831,6 +837,7 @@ mpt3sas_transport_port_add(struct MPT3SAS_ADAPTER *ioc, u16 handle,
|
||||
__FILE__, __LINE__, __func__);
|
||||
sas_rphy_free(rphy);
|
||||
rphy = NULL;
|
||||
goto out_delete_port;
|
||||
}
|
||||
|
||||
if (mpt3sas_port->remote_identify.device_type == SAS_END_DEVICE) {
|
||||
@ -857,7 +864,10 @@ mpt3sas_transport_port_add(struct MPT3SAS_ADAPTER *ioc, u16 handle,
|
||||
rphy_to_expander_device(rphy), hba_port->port_id);
|
||||
return mpt3sas_port;
|
||||
|
||||
out_fail:
|
||||
out_delete_port:
|
||||
sas_port_delete(port);
|
||||
|
||||
out_fail:
|
||||
list_for_each_entry_safe(mpt3sas_phy, next, &mpt3sas_port->phy_list,
|
||||
port_siblings)
|
||||
list_del(&mpt3sas_phy->port_siblings);
|
||||
|
@ -192,6 +192,7 @@ extern int ql2xsecenable;
|
||||
extern int ql2xenforce_iocb_limit;
|
||||
extern int ql2xabts_wait_nvme;
|
||||
extern u32 ql2xnvme_queues;
|
||||
extern int ql2xfc2target;
|
||||
|
||||
extern int qla2x00_loop_reset(scsi_qla_host_t *);
|
||||
extern void qla2x00_abort_all_cmds(scsi_qla_host_t *, int);
|
||||
|
@ -1840,7 +1840,8 @@ void qla2x00_handle_rscn(scsi_qla_host_t *vha, struct event_arg *ea)
|
||||
case RSCN_PORT_ADDR:
|
||||
fcport = qla2x00_find_fcport_by_nportid(vha, &ea->id, 1);
|
||||
if (fcport) {
|
||||
if (fcport->flags & FCF_FCP2_DEVICE &&
|
||||
if (ql2xfc2target &&
|
||||
fcport->flags & FCF_FCP2_DEVICE &&
|
||||
atomic_read(&fcport->state) == FCS_ONLINE) {
|
||||
ql_dbg(ql_dbg_disc, vha, 0x2115,
|
||||
"Delaying session delete for FCP2 portid=%06x %8phC ",
|
||||
|
@ -1900,6 +1900,8 @@ qla2x00_get_sp_from_handle(scsi_qla_host_t *vha, const char *func,
|
||||
}
|
||||
|
||||
req->outstanding_cmds[index] = NULL;
|
||||
|
||||
qla_put_fw_resources(sp->qpair, &sp->iores);
|
||||
return sp;
|
||||
}
|
||||
|
||||
@ -3112,7 +3114,6 @@ qla25xx_process_bidir_status_iocb(scsi_qla_host_t *vha, void *pkt,
|
||||
}
|
||||
bsg_reply->reply_payload_rcv_len = 0;
|
||||
|
||||
qla_put_fw_resources(sp->qpair, &sp->iores);
|
||||
done:
|
||||
/* Return the vendor specific reply to API */
|
||||
bsg_reply->reply_data.vendor_reply.vendor_rsp[0] = rval;
|
||||
|
@ -360,6 +360,13 @@ MODULE_PARM_DESC(ql2xnvme_queues,
|
||||
"1 - Minimum number of queues supported\n"
|
||||
"8 - Default value");
|
||||
|
||||
int ql2xfc2target = 1;
|
||||
module_param(ql2xfc2target, int, 0444);
|
||||
MODULE_PARM_DESC(qla2xfc2target,
|
||||
"Enables FC2 Target support. "
|
||||
"0 - FC2 Target support is disabled. "
|
||||
"1 - FC2 Target support is enabled (default).");
|
||||
|
||||
static struct scsi_transport_template *qla2xxx_transport_template = NULL;
|
||||
struct scsi_transport_template *qla2xxx_transport_vport_template = NULL;
|
||||
|
||||
@ -1858,6 +1865,17 @@ __qla2x00_abort_all_cmds(struct qla_qpair *qp, int res)
|
||||
for (cnt = 1; cnt < req->num_outstanding_cmds; cnt++) {
|
||||
sp = req->outstanding_cmds[cnt];
|
||||
if (sp) {
|
||||
/*
|
||||
* perform lockless completion during driver unload
|
||||
*/
|
||||
if (qla2x00_chip_is_down(vha)) {
|
||||
req->outstanding_cmds[cnt] = NULL;
|
||||
spin_unlock_irqrestore(qp->qp_lock_ptr, flags);
|
||||
sp->done(sp, res);
|
||||
spin_lock_irqsave(qp->qp_lock_ptr, flags);
|
||||
continue;
|
||||
}
|
||||
|
||||
switch (sp->cmd_type) {
|
||||
case TYPE_SRB:
|
||||
qla2x00_abort_srb(qp, sp, res, &flags);
|
||||
@ -4080,7 +4098,8 @@ qla2x00_mark_all_devices_lost(scsi_qla_host_t *vha)
|
||||
"Mark all dev lost\n");
|
||||
|
||||
list_for_each_entry(fcport, &vha->vp_fcports, list) {
|
||||
if (fcport->loop_id != FC_NO_LOOP_ID &&
|
||||
if (ql2xfc2target &&
|
||||
fcport->loop_id != FC_NO_LOOP_ID &&
|
||||
(fcport->flags & FCF_FCP2_DEVICE) &&
|
||||
fcport->port_type == FCT_TARGET &&
|
||||
!qla2x00_reset_active(vha)) {
|
||||
|
@ -314,11 +314,18 @@ static int scsi_vpd_inquiry(struct scsi_device *sdev, unsigned char *buffer,
|
||||
if (result)
|
||||
return -EIO;
|
||||
|
||||
/* Sanity check that we got the page back that we asked for */
|
||||
/*
|
||||
* Sanity check that we got the page back that we asked for and that
|
||||
* the page size is not 0.
|
||||
*/
|
||||
if (buffer[1] != page)
|
||||
return -EIO;
|
||||
|
||||
return get_unaligned_be16(&buffer[2]) + 4;
|
||||
result = get_unaligned_be16(&buffer[2]);
|
||||
if (!result)
|
||||
return -EIO;
|
||||
|
||||
return result + 4;
|
||||
}
|
||||
|
||||
static int scsi_get_vpd_size(struct scsi_device *sdev, u8 page)
|
||||
@ -326,6 +333,9 @@ static int scsi_get_vpd_size(struct scsi_device *sdev, u8 page)
|
||||
unsigned char vpd_header[SCSI_VPD_HEADER_SIZE] __aligned(4);
|
||||
int result;
|
||||
|
||||
if (sdev->no_vpd_size)
|
||||
return SCSI_DEFAULT_VPD_LEN;
|
||||
|
||||
/*
|
||||
* Fetch the VPD page header to find out how big the page
|
||||
* is. This is done to prevent problems on legacy devices
|
||||
|
@ -134,7 +134,7 @@ static struct {
|
||||
{"3PARdata", "VV", NULL, BLIST_REPORTLUN2},
|
||||
{"ADAPTEC", "AACRAID", NULL, BLIST_FORCELUN},
|
||||
{"ADAPTEC", "Adaptec 5400S", NULL, BLIST_FORCELUN},
|
||||
{"AIX", "VDASD", NULL, BLIST_TRY_VPD_PAGES},
|
||||
{"AIX", "VDASD", NULL, BLIST_TRY_VPD_PAGES | BLIST_NO_VPD_SIZE},
|
||||
{"AFT PRO", "-IX CF", "0.0>", BLIST_FORCELUN},
|
||||
{"BELKIN", "USB 2 HS-CF", "1.95", BLIST_FORCELUN | BLIST_INQUIRY_36},
|
||||
{"BROWNIE", "1200U3P", NULL, BLIST_NOREPORTLUN},
|
||||
@ -188,6 +188,7 @@ static struct {
|
||||
{"HPE", "OPEN-", "*", BLIST_REPORTLUN2 | BLIST_TRY_VPD_PAGES},
|
||||
{"IBM", "AuSaV1S2", NULL, BLIST_FORCELUN},
|
||||
{"IBM", "ProFibre 4000R", "*", BLIST_SPARSELUN | BLIST_LARGELUN},
|
||||
{"IBM", "2076", NULL, BLIST_NO_VPD_SIZE},
|
||||
{"IBM", "2105", NULL, BLIST_RETRY_HWERROR},
|
||||
{"iomega", "jaz 1GB", "J.86", BLIST_NOTQ | BLIST_NOLUN},
|
||||
{"IOMEGA", "ZIP", NULL, BLIST_NOTQ | BLIST_NOLUN},
|
||||
@ -233,6 +234,7 @@ static struct {
|
||||
{"SGI", "RAID5", "*", BLIST_SPARSELUN},
|
||||
{"SGI", "TP9100", "*", BLIST_REPORTLUN2},
|
||||
{"SGI", "Universal Xport", "*", BLIST_NO_ULD_ATTACH},
|
||||
{"SKhynix", "H28U74301AMR", NULL, BLIST_SKIP_VPD_PAGES},
|
||||
{"IBM", "Universal Xport", "*", BLIST_NO_ULD_ATTACH},
|
||||
{"SUN", "Universal Xport", "*", BLIST_NO_ULD_ATTACH},
|
||||
{"DELL", "Universal Xport", "*", BLIST_NO_ULD_ATTACH},
|
||||
|
@ -1057,6 +1057,9 @@ static int scsi_add_lun(struct scsi_device *sdev, unsigned char *inq_result,
|
||||
else if (*bflags & BLIST_SKIP_VPD_PAGES)
|
||||
sdev->skip_vpd_pages = 1;
|
||||
|
||||
if (*bflags & BLIST_NO_VPD_SIZE)
|
||||
sdev->no_vpd_size = 1;
|
||||
|
||||
transport_configure_device(&sdev->sdev_gendev);
|
||||
|
||||
if (sdev->host->hostt->slave_configure) {
|
||||
|
@ -2988,8 +2988,13 @@ static void sd_read_block_characteristics(struct scsi_disk *sdkp)
|
||||
}
|
||||
|
||||
if (sdkp->device->type == TYPE_ZBC) {
|
||||
/* Host-managed */
|
||||
/*
|
||||
* Host-managed: Per ZBC and ZAC specifications, writes in
|
||||
* sequential write required zones of host-managed devices must
|
||||
* be aligned to the device physical block size.
|
||||
*/
|
||||
disk_set_zoned(sdkp->disk, BLK_ZONED_HM);
|
||||
blk_queue_zone_write_granularity(q, sdkp->physical_block_size);
|
||||
} else {
|
||||
sdkp->zoned = zoned;
|
||||
if (sdkp->zoned == 1) {
|
||||
|
@ -965,14 +965,6 @@ int sd_zbc_read_zones(struct scsi_disk *sdkp, u8 buf[SD_BUF_SIZE])
|
||||
disk_set_max_active_zones(disk, 0);
|
||||
nr_zones = round_up(sdkp->capacity, zone_blocks) >> ilog2(zone_blocks);
|
||||
|
||||
/*
|
||||
* Per ZBC and ZAC specifications, writes in sequential write required
|
||||
* zones of host-managed devices must be aligned to the device physical
|
||||
* block size.
|
||||
*/
|
||||
if (blk_queue_zoned_model(q) == BLK_ZONED_HM)
|
||||
blk_queue_zone_write_granularity(q, sdkp->physical_block_size);
|
||||
|
||||
sdkp->early_zone_info.nr_zones = nr_zones;
|
||||
sdkp->early_zone_info.zone_blocks = zone_blocks;
|
||||
|
||||
|
@ -987,6 +987,22 @@ static void storvsc_handle_error(struct vmscsi_request *vm_srb,
|
||||
goto do_work;
|
||||
}
|
||||
|
||||
/*
|
||||
* Check for "Operating parameters have changed"
|
||||
* due to Hyper-V changing the VHD/VHDX BlockSize
|
||||
* when adding/removing a differencing disk. This
|
||||
* causes discard_granularity to change, so do a
|
||||
* rescan to pick up the new granularity. We don't
|
||||
* want scsi_report_sense() to output a message
|
||||
* that a sysadmin wouldn't know what to do with.
|
||||
*/
|
||||
if ((asc == 0x3f) && (ascq != 0x03) &&
|
||||
(ascq != 0x0e)) {
|
||||
process_err_fn = storvsc_device_scan;
|
||||
set_host_byte(scmnd, DID_REQUEUE);
|
||||
goto do_work;
|
||||
}
|
||||
|
||||
/*
|
||||
* Otherwise, let upper layer deal with the
|
||||
* error when sense message is present
|
||||
|
@ -1262,18 +1262,20 @@ static struct iscsi_param *iscsi_check_key(
|
||||
return param;
|
||||
|
||||
if (!(param->phase & phase)) {
|
||||
pr_err("Key \"%s\" may not be negotiated during ",
|
||||
param->name);
|
||||
char *phase_name;
|
||||
|
||||
switch (phase) {
|
||||
case PHASE_SECURITY:
|
||||
pr_debug("Security phase.\n");
|
||||
phase_name = "Security";
|
||||
break;
|
||||
case PHASE_OPERATIONAL:
|
||||
pr_debug("Operational phase.\n");
|
||||
phase_name = "Operational";
|
||||
break;
|
||||
default:
|
||||
pr_debug("Unknown phase.\n");
|
||||
phase_name = "Unknown";
|
||||
}
|
||||
pr_err("Key \"%s\" may not be negotiated during %s phase.\n",
|
||||
param->name, phase_name);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
|
@ -1527,7 +1527,7 @@ start_window:
|
||||
scaling->window_start_t = curr_t;
|
||||
scaling->tot_busy_t = 0;
|
||||
|
||||
if (hba->outstanding_reqs) {
|
||||
if (scaling->active_reqs) {
|
||||
scaling->busy_start_t = curr_t;
|
||||
scaling->is_busy_started = true;
|
||||
} else {
|
||||
@ -2145,7 +2145,7 @@ static void ufshcd_clk_scaling_update_busy(struct ufs_hba *hba)
|
||||
|
||||
spin_lock_irqsave(hba->host->host_lock, flags);
|
||||
hba->clk_scaling.active_reqs--;
|
||||
if (!hba->outstanding_reqs && scaling->is_busy_started) {
|
||||
if (!scaling->active_reqs && scaling->is_busy_started) {
|
||||
scaling->tot_busy_t += ktime_to_us(ktime_sub(ktime_get(),
|
||||
scaling->busy_start_t));
|
||||
scaling->busy_start_t = 0;
|
||||
@ -10543,4 +10543,5 @@ module_exit(ufshcd_core_exit);
|
||||
MODULE_AUTHOR("Santosh Yaragnavi <santosh.sy@samsung.com>");
|
||||
MODULE_AUTHOR("Vinayak Holikatti <h.vinayak@samsung.com>");
|
||||
MODULE_DESCRIPTION("Generic UFS host controller driver Core");
|
||||
MODULE_SOFTDEP("pre: governor_simpleondemand");
|
||||
MODULE_LICENSE("GPL");
|
||||
|
@ -1177,7 +1177,7 @@ static int ufs_qcom_clk_scale_notify(struct ufs_hba *hba,
|
||||
err = ufs_qcom_clk_scale_down_post_change(hba);
|
||||
|
||||
|
||||
if (err || !dev_req_params) {
|
||||
if (err) {
|
||||
ufshcd_uic_hibern8_exit(hba);
|
||||
return err;
|
||||
}
|
||||
@ -1451,8 +1451,8 @@ static int ufs_qcom_mcq_config_resource(struct ufs_hba *hba)
|
||||
if (IS_ERR(res->base)) {
|
||||
dev_err(hba->dev, "Failed to map res %s, err=%d\n",
|
||||
res->name, (int)PTR_ERR(res->base));
|
||||
res->base = NULL;
|
||||
ret = PTR_ERR(res->base);
|
||||
res->base = NULL;
|
||||
return ret;
|
||||
}
|
||||
}
|
||||
@ -1466,7 +1466,7 @@ static int ufs_qcom_mcq_config_resource(struct ufs_hba *hba)
|
||||
/* Explicitly allocate MCQ resource from ufs_mem */
|
||||
res_mcq = devm_kzalloc(hba->dev, sizeof(*res_mcq), GFP_KERNEL);
|
||||
if (!res_mcq)
|
||||
return ret;
|
||||
return -ENOMEM;
|
||||
|
||||
res_mcq->start = res_mem->start +
|
||||
MCQ_SQATTR_OFFSET(hba->mcq_capabilities);
|
||||
@ -1478,7 +1478,7 @@ static int ufs_qcom_mcq_config_resource(struct ufs_hba *hba)
|
||||
if (ret) {
|
||||
dev_err(hba->dev, "Failed to insert MCQ resource, err=%d\n",
|
||||
ret);
|
||||
goto insert_res_err;
|
||||
return ret;
|
||||
}
|
||||
|
||||
res->base = devm_ioremap_resource(hba->dev, res_mcq);
|
||||
@ -1495,8 +1495,6 @@ out:
|
||||
ioremap_err:
|
||||
res->base = NULL;
|
||||
remove_resource(res_mcq);
|
||||
insert_res_err:
|
||||
devm_kfree(hba->dev, res_mcq);
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
@ -145,6 +145,7 @@ struct scsi_device {
|
||||
const char * model; /* ... after scan; point to static string */
|
||||
const char * rev; /* ... "nullnullnullnull" before scan */
|
||||
|
||||
#define SCSI_DEFAULT_VPD_LEN 255 /* default SCSI VPD page size (max) */
|
||||
struct scsi_vpd __rcu *vpd_pg0;
|
||||
struct scsi_vpd __rcu *vpd_pg83;
|
||||
struct scsi_vpd __rcu *vpd_pg80;
|
||||
@ -215,6 +216,7 @@ struct scsi_device {
|
||||
* creation time */
|
||||
unsigned ignore_media_change:1; /* Ignore MEDIA CHANGE on resume */
|
||||
unsigned silence_suspend:1; /* Do not print runtime PM related messages */
|
||||
unsigned no_vpd_size:1; /* No VPD size reported in header */
|
||||
|
||||
unsigned int queue_stopped; /* request queue is quiesced */
|
||||
bool offline_already; /* Device offline message logged */
|
||||
|
@ -32,7 +32,8 @@
|
||||
#define BLIST_IGN_MEDIA_CHANGE ((__force blist_flags_t)(1ULL << 11))
|
||||
/* do not do automatic start on add */
|
||||
#define BLIST_NOSTARTONADD ((__force blist_flags_t)(1ULL << 12))
|
||||
#define __BLIST_UNUSED_13 ((__force blist_flags_t)(1ULL << 13))
|
||||
/* do not ask for VPD page size first on some broken targets */
|
||||
#define BLIST_NO_VPD_SIZE ((__force blist_flags_t)(1ULL << 13))
|
||||
#define __BLIST_UNUSED_14 ((__force blist_flags_t)(1ULL << 14))
|
||||
#define __BLIST_UNUSED_15 ((__force blist_flags_t)(1ULL << 15))
|
||||
#define __BLIST_UNUSED_16 ((__force blist_flags_t)(1ULL << 16))
|
||||
@ -74,8 +75,7 @@
|
||||
#define __BLIST_HIGH_UNUSED (~(__BLIST_LAST_USED | \
|
||||
(__force blist_flags_t) \
|
||||
((__force __u64)__BLIST_LAST_USED - 1ULL)))
|
||||
#define __BLIST_UNUSED_MASK (__BLIST_UNUSED_13 | \
|
||||
__BLIST_UNUSED_14 | \
|
||||
#define __BLIST_UNUSED_MASK (__BLIST_UNUSED_14 | \
|
||||
__BLIST_UNUSED_15 | \
|
||||
__BLIST_UNUSED_16 | \
|
||||
__BLIST_UNUSED_24 | \
|
||||
|
Loading…
Reference in New Issue
Block a user