2019-05-27 09:55:01 +03:00
// SPDX-License-Identifier: GPL-2.0-or-later
2015-08-14 05:47:43 +03:00
/*
* CXL Flash Device Driver
*
* Written by : Manoj N . Kumar < manoj @ linux . vnet . ibm . com > , IBM Corporation
* Matthew R . Ochs < mrochs @ linux . vnet . ibm . com > , IBM Corporation
*
* Copyright ( C ) 2015 IBM Corporation
*/
# include <linux/delay.h>
# include <linux/file.h>
2018-05-11 22:06:19 +03:00
# include <linux/interrupt.h>
# include <linux/pci.h>
2015-08-14 05:47:43 +03:00
# include <linux/syscalls.h>
# include <asm/unaligned.h>
# include <scsi/scsi.h>
# include <scsi/scsi_host.h>
# include <scsi/scsi_cmnd.h>
# include <scsi/scsi_eh.h>
# include <uapi/scsi/cxlflash_ioctl.h>
# include "sislite.h"
# include "common.h"
2015-08-14 05:47:53 +03:00
# include "vlun.h"
2015-08-14 05:47:43 +03:00
# include "superpipe.h"
struct cxlflash_global global ;
2015-08-14 05:47:53 +03:00
/**
* marshal_rele_to_resize ( ) - translate release to resize structure
2021-03-17 12:12:26 +03:00
* @ release : Source structure from which to translate / copy .
2015-08-14 05:47:53 +03:00
* @ resize : Destination structure for the translate / copy .
*/
static void marshal_rele_to_resize ( struct dk_cxlflash_release * release ,
struct dk_cxlflash_resize * resize )
{
resize - > hdr = release - > hdr ;
resize - > context_id = release - > context_id ;
resize - > rsrc_handle = release - > rsrc_handle ;
}
2015-08-14 05:47:43 +03:00
/**
* marshal_det_to_rele ( ) - translate detach to release structure
* @ detach : Destination structure for the translate / copy .
2021-03-17 12:12:26 +03:00
* @ release : Source structure from which to translate / copy .
2015-08-14 05:47:43 +03:00
*/
static void marshal_det_to_rele ( struct dk_cxlflash_detach * detach ,
struct dk_cxlflash_release * release )
{
release - > hdr = detach - > hdr ;
release - > context_id = detach - > context_id ;
}
2017-06-22 05:14:30 +03:00
/**
* marshal_udir_to_rele ( ) - translate udirect to release structure
* @ udirect : Source structure from which to translate / copy .
* @ release : Destination structure for the translate / copy .
*/
static void marshal_udir_to_rele ( struct dk_cxlflash_udirect * udirect ,
struct dk_cxlflash_release * release )
{
release - > hdr = udirect - > hdr ;
release - > context_id = udirect - > context_id ;
release - > rsrc_handle = udirect - > rsrc_handle ;
}
2015-08-14 05:47:43 +03:00
/**
* cxlflash_free_errpage ( ) - frees resources associated with global error page
*/
void cxlflash_free_errpage ( void )
{
mutex_lock ( & global . mutex ) ;
if ( global . err_page ) {
__free_page ( global . err_page ) ;
global . err_page = NULL ;
}
mutex_unlock ( & global . mutex ) ;
}
/**
* cxlflash_stop_term_user_contexts ( ) - stops / terminates known user contexts
* @ cfg : Internal structure associated with the host .
*
* When the host needs to go down , all users must be quiesced and their
* memory freed . This is accomplished by putting the contexts in error
2015-10-21 23:15:06 +03:00
* state which will notify the user and let them ' drive ' the tear down .
2015-08-14 05:47:43 +03:00
* Meanwhile , this routine camps until all user contexts have been removed .
2017-04-12 22:14:51 +03:00
*
* Note that the main loop in this routine will always execute at least once
* to flush the reset_waitq .
2015-08-14 05:47:43 +03:00
*/
void cxlflash_stop_term_user_contexts ( struct cxlflash_cfg * cfg )
{
struct device * dev = & cfg - > dev - > dev ;
2017-04-12 22:14:51 +03:00
int i , found = true ;
2015-08-14 05:47:43 +03:00
cxlflash_mark_contexts_error ( cfg ) ;
while ( true ) {
for ( i = 0 ; i < MAX_CONTEXT ; i + + )
if ( cfg - > ctx_tbl [ i ] ) {
found = true ;
break ;
}
if ( ! found & & list_empty ( & cfg - > ctx_err_recovery ) )
return ;
dev_dbg ( dev , " %s: Wait for user contexts to quiesce... \n " ,
__func__ ) ;
2015-10-21 23:12:00 +03:00
wake_up_all ( & cfg - > reset_waitq ) ;
2015-08-14 05:47:43 +03:00
ssleep ( 1 ) ;
2017-04-12 22:14:51 +03:00
found = false ;
2015-08-14 05:47:43 +03:00
}
}
/**
* find_error_context ( ) - locates a context by cookie on the error recovery list
* @ cfg : Internal structure associated with the host .
* @ rctxid : Desired context by id .
* @ file : Desired context by file .
*
* Return : Found context on success , NULL on failure
*/
static struct ctx_info * find_error_context ( struct cxlflash_cfg * cfg , u64 rctxid ,
struct file * file )
{
struct ctx_info * ctxi ;
list_for_each_entry ( ctxi , & cfg - > ctx_err_recovery , list )
if ( ( ctxi - > ctxid = = rctxid ) | | ( ctxi - > file = = file ) )
return ctxi ;
return NULL ;
}
/**
* get_context ( ) - obtains a validated and locked context reference
* @ cfg : Internal structure associated with the host .
* @ rctxid : Desired context ( raw , un - decoded format ) .
* @ arg : LUN information or file associated with request .
* @ ctx_ctrl : Control information to ' steer ' desired lookup .
*
* NOTE : despite the name pid , in linux , current - > pid actually refers
* to the lightweight process id ( tid ) and can change if the process is
* multi threaded . The tgid remains constant for the process and only changes
* when the process of fork . For all intents and purposes , think of tgid
* as a pid in the traditional sense .
*
* Return : Validated context on success , NULL on failure
*/
struct ctx_info * get_context ( struct cxlflash_cfg * cfg , u64 rctxid ,
void * arg , enum ctx_ctrl ctx_ctrl )
{
struct device * dev = & cfg - > dev - > dev ;
struct ctx_info * ctxi = NULL ;
struct lun_access * lun_access = NULL ;
struct file * file = NULL ;
struct llun_info * lli = arg ;
u64 ctxid = DECODE_CTXID ( rctxid ) ;
int rc ;
2017-10-26 00:36:29 +03:00
pid_t pid = task_tgid_nr ( current ) , ctxpid = 0 ;
2015-08-14 05:47:43 +03:00
if ( ctx_ctrl & CTX_CTRL_FILE ) {
lli = NULL ;
file = ( struct file * ) arg ;
}
if ( ctx_ctrl & CTX_CTRL_CLONE )
2017-10-26 00:36:29 +03:00
pid = task_ppid_nr ( current ) ;
2015-08-14 05:47:43 +03:00
if ( likely ( ctxid < MAX_CONTEXT ) ) {
while ( true ) {
2015-10-21 23:16:32 +03:00
mutex_lock ( & cfg - > ctx_tbl_list_mutex ) ;
2015-08-14 05:47:43 +03:00
ctxi = cfg - > ctx_tbl [ ctxid ] ;
if ( ctxi )
if ( ( file & & ( ctxi - > file ! = file ) ) | |
( ! file & & ( ctxi - > ctxid ! = rctxid ) ) )
ctxi = NULL ;
if ( ( ctx_ctrl & CTX_CTRL_ERR ) | |
( ! ctxi & & ( ctx_ctrl & CTX_CTRL_ERR_FALLBACK ) ) )
ctxi = find_error_context ( cfg , rctxid , file ) ;
if ( ! ctxi ) {
mutex_unlock ( & cfg - > ctx_tbl_list_mutex ) ;
goto out ;
}
/*
* Need to acquire ownership of the context while still
* under the table / list lock to serialize with a remove
* thread . Use the ' try ' to avoid stalling the
* table / list lock for a single context .
*
* Note that the lock order is :
*
* cfg - > ctx_tbl_list_mutex - > ctxi - > mutex
*
* Therefore release ctx_tbl_list_mutex before retrying .
*/
rc = mutex_trylock ( & ctxi - > mutex ) ;
mutex_unlock ( & cfg - > ctx_tbl_list_mutex ) ;
if ( rc )
break ; /* got the context's lock! */
}
if ( ctxi - > unavail )
goto denied ;
ctxpid = ctxi - > pid ;
if ( likely ( ! ( ctx_ctrl & CTX_CTRL_NOPID ) ) )
if ( pid ! = ctxpid )
goto denied ;
if ( lli ) {
list_for_each_entry ( lun_access , & ctxi - > luns , list )
if ( lun_access - > lli = = lli )
goto out ;
goto denied ;
}
}
out :
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: rctxid=%016llx ctxinfo=%p ctxpid=%u pid=%u "
2015-08-14 05:47:43 +03:00
" ctx_ctrl=%u \n " , __func__ , rctxid , ctxi , ctxpid , pid ,
ctx_ctrl ) ;
return ctxi ;
denied :
mutex_unlock ( & ctxi - > mutex ) ;
ctxi = NULL ;
goto out ;
}
/**
* put_context ( ) - release a context that was retrieved from get_context ( )
* @ ctxi : Context to release .
*
* For now , releasing the context equates to unlocking it ' s mutex .
*/
void put_context ( struct ctx_info * ctxi )
{
mutex_unlock ( & ctxi - > mutex ) ;
}
/**
* afu_attach ( ) - attach a context to the AFU
* @ cfg : Internal structure associated with the host .
* @ ctxi : Context to attach .
*
* Upon setting the context capabilities , they must be confirmed with
* a read back operation as the context might have been closed since
* the mailbox was unlocked . When this occurs , registration is failed .
*
* Return : 0 on success , - errno on failure
*/
static int afu_attach ( struct cxlflash_cfg * cfg , struct ctx_info * ctxi )
{
struct device * dev = & cfg - > dev - > dev ;
struct afu * afu = cfg - > afu ;
2015-10-21 23:14:48 +03:00
struct sisl_ctrl_map __iomem * ctrl_map = ctxi - > ctrl_map ;
2015-08-14 05:47:43 +03:00
int rc = 0 ;
2017-04-12 22:15:42 +03:00
struct hwq * hwq = get_hwq ( afu , PRIMARY_HWQ ) ;
2015-08-14 05:47:43 +03:00
u64 val ;
2018-03-26 19:34:42 +03:00
int i ;
2015-08-14 05:47:43 +03:00
/* Unlock cap and restrict user to read/write cmds in translated mode */
readq_be ( & ctrl_map - > mbox_r ) ;
val = ( SISL_CTX_CAP_READ_CMD | SISL_CTX_CAP_WRITE_CMD ) ;
writeq_be ( val , & ctrl_map - > ctx_cap ) ;
val = readq_be ( & ctrl_map - > ctx_cap ) ;
if ( val ! = ( SISL_CTX_CAP_READ_CMD | SISL_CTX_CAP_WRITE_CMD ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: ctx may be closed val=%016llx \n " ,
2015-08-14 05:47:43 +03:00
__func__ , val ) ;
rc = - EAGAIN ;
goto out ;
}
2018-03-26 19:34:42 +03:00
if ( afu_is_ocxl_lisn ( afu ) ) {
/* Set up the LISN effective address for each interrupt */
for ( i = 0 ; i < ctxi - > irqs ; i + + ) {
val = cfg - > ops - > get_irq_objhndl ( ctxi - > ctx , i ) ;
writeq_be ( val , & ctrl_map - > lisn_ea [ i ] ) ;
}
/* Use primary HWQ PASID as identifier for all interrupts */
val = hwq - > ctx_hndl ;
writeq_be ( SISL_LISN_PASID ( val , val ) , & ctrl_map - > lisn_pasid [ 0 ] ) ;
writeq_be ( SISL_LISN_PASID ( 0UL , val ) , & ctrl_map - > lisn_pasid [ 1 ] ) ;
}
2015-08-14 05:47:43 +03:00
/* Set up MMIO registers pointing to the RHT */
writeq_be ( ( u64 ) ctxi - > rht_start , & ctrl_map - > rht_start ) ;
2017-04-12 22:15:42 +03:00
val = SISL_RHT_CNT_ID ( ( u64 ) MAX_RHT_PER_CONTEXT , ( u64 ) ( hwq - > ctx_hndl ) ) ;
2015-08-14 05:47:43 +03:00
writeq_be ( val , & ctrl_map - > rht_cnt_id ) ;
out :
dev_dbg ( dev , " %s: returning rc=%d \n " , __func__ , rc ) ;
return rc ;
}
/**
* read_cap16 ( ) - issues a SCSI READ_CAP16 command
* @ sdev : SCSI device associated with LUN .
* @ lli : LUN destined for capacity request .
*
cxlflash: Fix to avoid potential deadlock on EEH
Ioctl threads that use scsi_execute() can run for an excessive amount
of time due to the fact that they have lengthy timeouts and retry logic
built in. Under normal operation this is not an issue. However, once EEH
enters the picture, a long execution time coupled with the possibility
that a timeout can trigger entry to the driver via registered reset
callbacks becomes a liability.
In particular, a deadlock can occur when an EEH event is encountered
while in running in scsi_execute(). As part of the recovery, the EEH
handler drains all currently running ioctls, waiting until they have
completed before proceeding with a reset. As the scsi_execute()'s are
situated on the ioctl path, the EEH handler will wait until they (and
the remainder of the ioctl handler they're associated with) have
completed. Normally this would not be much of an issue aside from the
longer recovery period. Unfortunately, the scsi_execute() triggers a
reset when it times out. The reset handler will see that the device is
already being reset and wait until that reset completed. This creates
a condition where the EEH handler becomes stuck, infinitely waiting for
the ioctl thread to complete.
To avoid this behavior, temporarily unmark the scsi_execute() threads
as an ioctl thread by releasing the ioctl read semaphore. This allows
the EEH handler to proceed with a recovery while the thread is still
running. Once the scsi_execute() returns, the ioctl read semaphore is
reacquired and the adapter state is rechecked in case it changed while
inside of scsi_execute(). The state check will wait if the adapter is
still being recovered or returns a failure if the recovery failed. In
the event that the adapter reset failed, the failure is simply returned
as the ioctl would be unable to continue.
Reported-by: Brian King <brking@linux.vnet.ibm.com>
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Signed-off-by: Manoj N. Kumar <manoj@linux.vnet.ibm.com>
Reviewed-by: Brian King <brking@linux.vnet.ibm.com>
Reviewed-by: Daniel Axtens <dja@axtens.net>
Reviewed-by: Tomas Henzl <thenzl@redhat.com>
Signed-off-by: James Bottomley <JBottomley@Odin.com>
2015-10-21 23:15:52 +03:00
* The READ_CAP16 can take quite a while to complete . Should an EEH occur while
* in scsi_execute ( ) , the EEH handler will attempt to recover . As part of the
* recovery , the handler drains all currently running ioctls , waiting until they
* have completed before proceeding with a reset . As this routine is used on the
* ioctl path , this can create a condition where the EEH handler becomes stuck ,
* infinitely waiting for this ioctl thread . To avoid this behavior , temporarily
* unmark this thread as an ioctl thread by releasing the ioctl read semaphore .
* This will allow the EEH handler to proceed with a recovery while this thread
* is still running . Once the scsi_execute ( ) returns , reacquire the ioctl read
* semaphore and check the adapter state in case it changed while inside of
* scsi_execute ( ) . The state check will wait if the adapter is still being
* recovered or return a failure if the recovery failed . In the event that the
* adapter reset failed , simply return the failure as the ioctl would be unable
* to continue .
*
* Note that the above puts a requirement on this routine to only be called on
* an ioctl thread .
*
2015-08-14 05:47:43 +03:00
* Return : 0 on success , - errno on failure
*/
static int read_cap16 ( struct scsi_device * sdev , struct llun_info * lli )
{
2017-01-12 04:19:47 +03:00
struct cxlflash_cfg * cfg = shost_priv ( sdev - > host ) ;
2015-08-14 05:47:43 +03:00
struct device * dev = & cfg - > dev - > dev ;
struct glun_info * gli = lli - > parent ;
2017-02-23 18:02:36 +03:00
struct scsi_sense_hdr sshdr ;
2015-08-14 05:47:43 +03:00
u8 * cmd_buf = NULL ;
u8 * scsi_cmd = NULL ;
int rc = 0 ;
int result = 0 ;
int retry_cnt = 0 ;
2015-10-21 23:11:10 +03:00
u32 to = CMD_TIMEOUT * HZ ;
2015-08-14 05:47:43 +03:00
retry :
cmd_buf = kzalloc ( CMD_BUFSIZE , GFP_KERNEL ) ;
scsi_cmd = kzalloc ( MAX_COMMAND_SIZE , GFP_KERNEL ) ;
2018-07-31 22:51:47 +03:00
if ( unlikely ( ! cmd_buf | | ! scsi_cmd ) ) {
2015-08-14 05:47:43 +03:00
rc = - ENOMEM ;
goto out ;
}
scsi_cmd [ 0 ] = SERVICE_ACTION_IN_16 ; /* read cap(16) */
scsi_cmd [ 1 ] = SAI_READ_CAPACITY_16 ; /* service action */
put_unaligned_be32 ( CMD_BUFSIZE , & scsi_cmd [ 10 ] ) ;
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: %ssending cmd(%02x) \n " , __func__ ,
2015-08-14 05:47:43 +03:00
retry_cnt ? " re " : " " , scsi_cmd [ 0 ] ) ;
cxlflash: Fix to avoid potential deadlock on EEH
Ioctl threads that use scsi_execute() can run for an excessive amount
of time due to the fact that they have lengthy timeouts and retry logic
built in. Under normal operation this is not an issue. However, once EEH
enters the picture, a long execution time coupled with the possibility
that a timeout can trigger entry to the driver via registered reset
callbacks becomes a liability.
In particular, a deadlock can occur when an EEH event is encountered
while in running in scsi_execute(). As part of the recovery, the EEH
handler drains all currently running ioctls, waiting until they have
completed before proceeding with a reset. As the scsi_execute()'s are
situated on the ioctl path, the EEH handler will wait until they (and
the remainder of the ioctl handler they're associated with) have
completed. Normally this would not be much of an issue aside from the
longer recovery period. Unfortunately, the scsi_execute() triggers a
reset when it times out. The reset handler will see that the device is
already being reset and wait until that reset completed. This creates
a condition where the EEH handler becomes stuck, infinitely waiting for
the ioctl thread to complete.
To avoid this behavior, temporarily unmark the scsi_execute() threads
as an ioctl thread by releasing the ioctl read semaphore. This allows
the EEH handler to proceed with a recovery while the thread is still
running. Once the scsi_execute() returns, the ioctl read semaphore is
reacquired and the adapter state is rechecked in case it changed while
inside of scsi_execute(). The state check will wait if the adapter is
still being recovered or returns a failure if the recovery failed. In
the event that the adapter reset failed, the failure is simply returned
as the ioctl would be unable to continue.
Reported-by: Brian King <brking@linux.vnet.ibm.com>
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Signed-off-by: Manoj N. Kumar <manoj@linux.vnet.ibm.com>
Reviewed-by: Brian King <brking@linux.vnet.ibm.com>
Reviewed-by: Daniel Axtens <dja@axtens.net>
Reviewed-by: Tomas Henzl <thenzl@redhat.com>
Signed-off-by: James Bottomley <JBottomley@Odin.com>
2015-10-21 23:15:52 +03:00
/* Drop the ioctl read semahpore across lengthy call */
up_read ( & cfg - > ioctl_rwsem ) ;
2015-08-14 05:47:43 +03:00
result = scsi_execute ( sdev , scsi_cmd , DMA_FROM_DEVICE , cmd_buf ,
2018-07-31 22:51:47 +03:00
CMD_BUFSIZE , NULL , & sshdr , to , CMD_RETRIES ,
2017-02-23 18:02:36 +03:00
0 , 0 , NULL ) ;
cxlflash: Fix to avoid potential deadlock on EEH
Ioctl threads that use scsi_execute() can run for an excessive amount
of time due to the fact that they have lengthy timeouts and retry logic
built in. Under normal operation this is not an issue. However, once EEH
enters the picture, a long execution time coupled with the possibility
that a timeout can trigger entry to the driver via registered reset
callbacks becomes a liability.
In particular, a deadlock can occur when an EEH event is encountered
while in running in scsi_execute(). As part of the recovery, the EEH
handler drains all currently running ioctls, waiting until they have
completed before proceeding with a reset. As the scsi_execute()'s are
situated on the ioctl path, the EEH handler will wait until they (and
the remainder of the ioctl handler they're associated with) have
completed. Normally this would not be much of an issue aside from the
longer recovery period. Unfortunately, the scsi_execute() triggers a
reset when it times out. The reset handler will see that the device is
already being reset and wait until that reset completed. This creates
a condition where the EEH handler becomes stuck, infinitely waiting for
the ioctl thread to complete.
To avoid this behavior, temporarily unmark the scsi_execute() threads
as an ioctl thread by releasing the ioctl read semaphore. This allows
the EEH handler to proceed with a recovery while the thread is still
running. Once the scsi_execute() returns, the ioctl read semaphore is
reacquired and the adapter state is rechecked in case it changed while
inside of scsi_execute(). The state check will wait if the adapter is
still being recovered or returns a failure if the recovery failed. In
the event that the adapter reset failed, the failure is simply returned
as the ioctl would be unable to continue.
Reported-by: Brian King <brking@linux.vnet.ibm.com>
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Signed-off-by: Manoj N. Kumar <manoj@linux.vnet.ibm.com>
Reviewed-by: Brian King <brking@linux.vnet.ibm.com>
Reviewed-by: Daniel Axtens <dja@axtens.net>
Reviewed-by: Tomas Henzl <thenzl@redhat.com>
Signed-off-by: James Bottomley <JBottomley@Odin.com>
2015-10-21 23:15:52 +03:00
down_read ( & cfg - > ioctl_rwsem ) ;
rc = check_state ( cfg ) ;
if ( rc ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Failed state result=%08x \n " ,
cxlflash: Fix to avoid potential deadlock on EEH
Ioctl threads that use scsi_execute() can run for an excessive amount
of time due to the fact that they have lengthy timeouts and retry logic
built in. Under normal operation this is not an issue. However, once EEH
enters the picture, a long execution time coupled with the possibility
that a timeout can trigger entry to the driver via registered reset
callbacks becomes a liability.
In particular, a deadlock can occur when an EEH event is encountered
while in running in scsi_execute(). As part of the recovery, the EEH
handler drains all currently running ioctls, waiting until they have
completed before proceeding with a reset. As the scsi_execute()'s are
situated on the ioctl path, the EEH handler will wait until they (and
the remainder of the ioctl handler they're associated with) have
completed. Normally this would not be much of an issue aside from the
longer recovery period. Unfortunately, the scsi_execute() triggers a
reset when it times out. The reset handler will see that the device is
already being reset and wait until that reset completed. This creates
a condition where the EEH handler becomes stuck, infinitely waiting for
the ioctl thread to complete.
To avoid this behavior, temporarily unmark the scsi_execute() threads
as an ioctl thread by releasing the ioctl read semaphore. This allows
the EEH handler to proceed with a recovery while the thread is still
running. Once the scsi_execute() returns, the ioctl read semaphore is
reacquired and the adapter state is rechecked in case it changed while
inside of scsi_execute(). The state check will wait if the adapter is
still being recovered or returns a failure if the recovery failed. In
the event that the adapter reset failed, the failure is simply returned
as the ioctl would be unable to continue.
Reported-by: Brian King <brking@linux.vnet.ibm.com>
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Signed-off-by: Manoj N. Kumar <manoj@linux.vnet.ibm.com>
Reviewed-by: Brian King <brking@linux.vnet.ibm.com>
Reviewed-by: Daniel Axtens <dja@axtens.net>
Reviewed-by: Tomas Henzl <thenzl@redhat.com>
Signed-off-by: James Bottomley <JBottomley@Odin.com>
2015-10-21 23:15:52 +03:00
__func__ , result ) ;
rc = - ENODEV ;
goto out ;
}
2015-08-14 05:47:43 +03:00
2021-04-27 11:30:15 +03:00
if ( result > 0 & & scsi_sense_valid ( & sshdr ) ) {
2015-08-14 05:47:43 +03:00
if ( result & SAM_STAT_CHECK_CONDITION ) {
switch ( sshdr . sense_key ) {
case NO_SENSE :
case RECOVERED_ERROR :
case NOT_READY :
result & = ~ SAM_STAT_CHECK_CONDITION ;
break ;
case UNIT_ATTENTION :
switch ( sshdr . asc ) {
case 0x29 : /* Power on Reset or Device Reset */
2020-08-24 01:36:59 +03:00
fallthrough ;
2015-08-14 05:47:43 +03:00
case 0x2A : /* Device capacity changed */
case 0x3F : /* Report LUNs changed */
/* Retry the command once more */
if ( retry_cnt + + < 1 ) {
kfree ( cmd_buf ) ;
kfree ( scsi_cmd ) ;
goto retry ;
}
}
break ;
default :
break ;
}
}
}
if ( result ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: command failed, result=%08x \n " ,
2015-08-14 05:47:43 +03:00
__func__ , result ) ;
rc = - EIO ;
goto out ;
}
/*
* Read cap was successful , grab values from the buffer ;
* note that we don ' t need to worry about unaligned access
* as the buffer is allocated on an aligned boundary .
*/
mutex_lock ( & gli - > mutex ) ;
2015-10-21 23:14:48 +03:00
gli - > max_lba = be64_to_cpu ( * ( ( __be64 * ) & cmd_buf [ 0 ] ) ) ;
gli - > blk_len = be32_to_cpu ( * ( ( __be32 * ) & cmd_buf [ 8 ] ) ) ;
2015-08-14 05:47:43 +03:00
mutex_unlock ( & gli - > mutex ) ;
out :
kfree ( cmd_buf ) ;
kfree ( scsi_cmd ) ;
dev_dbg ( dev , " %s: maxlba=%lld blklen=%d rc=%d \n " ,
__func__ , gli - > max_lba , gli - > blk_len , rc ) ;
return rc ;
}
/**
* get_rhte ( ) - obtains validated resource handle table entry reference
* @ ctxi : Context owning the resource handle .
* @ rhndl : Resource handle associated with entry .
* @ lli : LUN associated with request .
*
* Return : Validated RHTE on success , NULL on failure
*/
struct sisl_rht_entry * get_rhte ( struct ctx_info * ctxi , res_hndl_t rhndl ,
struct llun_info * lli )
{
2017-01-12 04:19:47 +03:00
struct cxlflash_cfg * cfg = ctxi - > cfg ;
struct device * dev = & cfg - > dev - > dev ;
2015-08-14 05:47:43 +03:00
struct sisl_rht_entry * rhte = NULL ;
if ( unlikely ( ! ctxi - > rht_start ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Context does not have allocated RHT \n " ,
2015-08-14 05:47:43 +03:00
__func__ ) ;
goto out ;
}
if ( unlikely ( rhndl > = MAX_RHT_PER_CONTEXT ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Bad resource handle rhndl=%d \n " ,
__func__ , rhndl ) ;
2015-08-14 05:47:43 +03:00
goto out ;
}
if ( unlikely ( ctxi - > rht_lun [ rhndl ] ! = lli ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Bad resource handle LUN rhndl=%d \n " ,
__func__ , rhndl ) ;
2015-08-14 05:47:43 +03:00
goto out ;
}
rhte = & ctxi - > rht_start [ rhndl ] ;
if ( unlikely ( rhte - > nmask = = 0 ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Unopened resource handle rhndl=%d \n " ,
__func__ , rhndl ) ;
2015-08-14 05:47:43 +03:00
rhte = NULL ;
goto out ;
}
out :
return rhte ;
}
/**
* rhte_checkout ( ) - obtains free / empty resource handle table entry
* @ ctxi : Context owning the resource handle .
* @ lli : LUN associated with request .
*
* Return : Free RHTE on success , NULL on failure
*/
struct sisl_rht_entry * rhte_checkout ( struct ctx_info * ctxi ,
struct llun_info * lli )
{
2017-01-12 04:19:47 +03:00
struct cxlflash_cfg * cfg = ctxi - > cfg ;
struct device * dev = & cfg - > dev - > dev ;
2015-08-14 05:47:43 +03:00
struct sisl_rht_entry * rhte = NULL ;
int i ;
/* Find a free RHT entry */
for ( i = 0 ; i < MAX_RHT_PER_CONTEXT ; i + + )
if ( ctxi - > rht_start [ i ] . nmask = = 0 ) {
rhte = & ctxi - > rht_start [ i ] ;
ctxi - > rht_out + + ;
break ;
}
if ( likely ( rhte ) )
ctxi - > rht_lun [ i ] = lli ;
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: returning rhte=%p index=%d \n " , __func__ , rhte , i ) ;
2015-08-14 05:47:43 +03:00
return rhte ;
}
/**
* rhte_checkin ( ) - releases a resource handle table entry
* @ ctxi : Context owning the resource handle .
* @ rhte : RHTE to release .
*/
void rhte_checkin ( struct ctx_info * ctxi ,
struct sisl_rht_entry * rhte )
{
u32 rsrc_handle = rhte - ctxi - > rht_start ;
rhte - > nmask = 0 ;
rhte - > fp = 0 ;
ctxi - > rht_out - - ;
ctxi - > rht_lun [ rsrc_handle ] = NULL ;
2015-08-14 05:47:53 +03:00
ctxi - > rht_needs_ws [ rsrc_handle ] = false ;
2015-08-14 05:47:43 +03:00
}
/**
2021-03-17 12:12:26 +03:00
* rht_format1 ( ) - populates a RHTE for format 1
2015-08-14 05:47:43 +03:00
* @ rhte : RHTE to populate .
* @ lun_id : LUN ID of LUN associated with RHTE .
* @ perm : Desired permissions for RHTE .
* @ port_sel : Port selection mask
*/
static void rht_format1 ( struct sisl_rht_entry * rhte , u64 lun_id , u32 perm ,
u32 port_sel )
{
/*
* Populate the Format 1 RHT entry for direct access ( physical
* LUN ) using the synchronization sequence defined in the
* SISLite specification .
*/
struct sisl_rht_entry_f1 dummy = { 0 } ;
struct sisl_rht_entry_f1 * rhte_f1 = ( struct sisl_rht_entry_f1 * ) rhte ;
memset ( rhte_f1 , 0 , sizeof ( * rhte_f1 ) ) ;
rhte_f1 - > fp = SISL_RHT_FP ( 1U , 0 ) ;
dma_wmb ( ) ; /* Make setting of format bit visible */
rhte_f1 - > lun_id = lun_id ;
dma_wmb ( ) ; /* Make setting of LUN id visible */
/*
* Use a dummy RHT Format 1 entry to build the second dword
* of the entry that must be populated in a single write when
* enabled ( valid bit set to TRUE ) .
*/
dummy . valid = 0x80 ;
dummy . fp = SISL_RHT_FP ( 1U , perm ) ;
dummy . port_sel = port_sel ;
rhte_f1 - > dw = dummy . dw ;
dma_wmb ( ) ; /* Make remaining RHT entry fields visible */
}
/**
* cxlflash_lun_attach ( ) - attaches a user to a LUN and manages the LUN ' s mode
* @ gli : LUN to attach .
* @ mode : Desired mode of the LUN .
* @ locked : Mutex status on current thread .
*
* Return : 0 on success , - errno on failure
*/
int cxlflash_lun_attach ( struct glun_info * gli , enum lun_mode mode , bool locked )
{
int rc = 0 ;
if ( ! locked )
mutex_lock ( & gli - > mutex ) ;
if ( gli - > mode = = MODE_NONE )
gli - > mode = mode ;
else if ( gli - > mode ! = mode ) {
2017-01-12 04:19:47 +03:00
pr_debug ( " %s: gli_mode=%d requested_mode=%d \n " ,
2015-08-14 05:47:43 +03:00
__func__ , gli - > mode , mode ) ;
rc = - EINVAL ;
goto out ;
}
gli - > users + + ;
WARN_ON ( gli - > users < = 0 ) ;
out :
pr_debug ( " %s: Returning rc=%d gli->mode=%u gli->users=%u \n " ,
__func__ , rc , gli - > mode , gli - > users ) ;
if ( ! locked )
mutex_unlock ( & gli - > mutex ) ;
return rc ;
}
/**
* cxlflash_lun_detach ( ) - detaches a user from a LUN and resets the LUN ' s mode
* @ gli : LUN to detach .
2015-08-14 05:47:53 +03:00
*
* When resetting the mode , terminate block allocation resources as they
* are no longer required ( service is safe to call even when block allocation
* resources were not present - such as when transitioning from physical mode ) .
* These resources will be reallocated when needed ( subsequent transition to
* virtual mode ) .
2015-08-14 05:47:43 +03:00
*/
void cxlflash_lun_detach ( struct glun_info * gli )
{
mutex_lock ( & gli - > mutex ) ;
WARN_ON ( gli - > mode = = MODE_NONE ) ;
2015-08-14 05:47:53 +03:00
if ( - - gli - > users = = 0 ) {
2015-08-14 05:47:43 +03:00
gli - > mode = MODE_NONE ;
2015-08-14 05:47:53 +03:00
cxlflash_ba_terminate ( & gli - > blka . ba_lun ) ;
}
2015-08-14 05:47:43 +03:00
pr_debug ( " %s: gli->users=%u \n " , __func__ , gli - > users ) ;
WARN_ON ( gli - > users < 0 ) ;
mutex_unlock ( & gli - > mutex ) ;
}
/**
* _cxlflash_disk_release ( ) - releases the specified resource entry
* @ sdev : SCSI device associated with LUN .
* @ ctxi : Context owning resources .
* @ release : Release ioctl data structure .
*
2015-08-14 05:47:53 +03:00
* For LUNs in virtual mode , the virtual LUN associated with the specified
* resource handle is resized to 0 prior to releasing the RHTE . Note that the
* AFU sync should _not_ be performed when the context is sitting on the error
* recovery list . A context on the error recovery list is not known to the AFU
* due to reset . When the context is recovered , it will be reattached and made
* known again to the AFU .
2015-08-14 05:47:43 +03:00
*
* Return : 0 on success , - errno on failure
*/
int _cxlflash_disk_release ( struct scsi_device * sdev ,
struct ctx_info * ctxi ,
struct dk_cxlflash_release * release )
{
2017-01-12 04:19:47 +03:00
struct cxlflash_cfg * cfg = shost_priv ( sdev - > host ) ;
2015-08-14 05:47:43 +03:00
struct device * dev = & cfg - > dev - > dev ;
struct llun_info * lli = sdev - > hostdata ;
struct glun_info * gli = lli - > parent ;
struct afu * afu = cfg - > afu ;
bool put_ctx = false ;
2015-08-14 05:47:53 +03:00
struct dk_cxlflash_resize size ;
2015-08-14 05:47:43 +03:00
res_hndl_t rhndl = release - > rsrc_handle ;
int rc = 0 ;
2017-06-22 05:14:30 +03:00
int rcr = 0 ;
2015-08-14 05:47:43 +03:00
u64 ctxid = DECODE_CTXID ( release - > context_id ) ,
rctxid = release - > context_id ;
struct sisl_rht_entry * rhte ;
struct sisl_rht_entry_f1 * rhte_f1 ;
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: ctxid=%llu rhndl=%llu gli->mode=%u gli->users=%u \n " ,
2015-08-14 05:47:43 +03:00
__func__ , ctxid , release - > rsrc_handle , gli - > mode , gli - > users ) ;
if ( ! ctxi ) {
ctxi = get_context ( cfg , rctxid , lli , CTX_CTRL_ERR_FALLBACK ) ;
if ( unlikely ( ! ctxi ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Bad context ctxid=%llu \n " ,
2015-08-14 05:47:43 +03:00
__func__ , ctxid ) ;
rc = - EINVAL ;
goto out ;
}
put_ctx = true ;
}
rhte = get_rhte ( ctxi , rhndl , lli ) ;
if ( unlikely ( ! rhte ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Bad resource handle rhndl=%d \n " ,
2015-08-14 05:47:43 +03:00
__func__ , rhndl ) ;
rc = - EINVAL ;
goto out ;
}
2015-08-14 05:47:53 +03:00
/*
* Resize to 0 for virtual LUNS by setting the size
* to 0. This will clear LXT_START and LXT_CNT fields
* in the RHT entry and properly sync with the AFU .
*
* Afterwards we clear the remaining fields .
*/
2015-08-14 05:47:43 +03:00
switch ( gli - > mode ) {
2015-08-14 05:47:53 +03:00
case MODE_VIRTUAL :
marshal_rele_to_resize ( release , & size ) ;
size . req_size = 0 ;
rc = _cxlflash_vlun_resize ( sdev , ctxi , & size ) ;
if ( rc ) {
dev_dbg ( dev , " %s: resize failed rc %d \n " , __func__ , rc ) ;
goto out ;
}
break ;
2015-08-14 05:47:43 +03:00
case MODE_PHYSICAL :
/*
* Clear the Format 1 RHT entry for direct access
* ( physical LUN ) using the synchronization sequence
* defined in the SISLite specification .
*/
rhte_f1 = ( struct sisl_rht_entry_f1 * ) rhte ;
rhte_f1 - > valid = 0 ;
dma_wmb ( ) ; /* Make revocation of RHT entry visible */
rhte_f1 - > lun_id = 0 ;
dma_wmb ( ) ; /* Make clearing of LUN id visible */
rhte_f1 - > dw = 0 ;
dma_wmb ( ) ; /* Make RHT entry bottom-half clearing visible */
2017-06-22 05:14:30 +03:00
if ( ! ctxi - > err_recovery_active ) {
rcr = cxlflash_afu_sync ( afu , ctxid , rhndl , AFU_HW_SYNC ) ;
if ( unlikely ( rcr ) )
dev_dbg ( dev , " %s: AFU sync failed rc=%d \n " ,
__func__ , rcr ) ;
}
2015-08-14 05:47:43 +03:00
break ;
default :
WARN ( 1 , " Unsupported LUN mode! " ) ;
goto out ;
}
rhte_checkin ( ctxi , rhte ) ;
cxlflash_lun_detach ( gli ) ;
out :
if ( put_ctx )
put_context ( ctxi ) ;
dev_dbg ( dev , " %s: returning rc=%d \n " , __func__ , rc ) ;
return rc ;
}
int cxlflash_disk_release ( struct scsi_device * sdev ,
struct dk_cxlflash_release * release )
{
return _cxlflash_disk_release ( sdev , NULL , release ) ;
}
/**
* destroy_context ( ) - releases a context
* @ cfg : Internal structure associated with the host .
* @ ctxi : Context to release .
*
2016-08-10 02:39:18 +03:00
* This routine is safe to be called with a a non - initialized context .
* Also note that the routine conditionally checks for the existence
* of the context control map before clearing the RHT registers and
* context capabilities because it is possible to destroy a context
* while the context is in the error state ( previous mapping was
* removed [ so there is no need to worry about clearing ] and context
* is waiting for a new mapping ) .
2015-08-14 05:47:43 +03:00
*/
static void destroy_context ( struct cxlflash_cfg * cfg ,
struct ctx_info * ctxi )
{
struct afu * afu = cfg - > afu ;
cxlflash: Split out context initialization
Presently, context information structures are allocated and
initialized in the same routine, create_context(). This imposes
an ordering restriction such that all pieces of information needed
to initialize a context must be known before the context is even
allocated.
This design point is not flexible when the order of context
creation needs to be modified. Specifically, this can lead to
problems when members of the context information structure are
a part of an ordering dependency (i.e. - the 'work' structure
embedded within the context).
To remedy, the allocation is left as-is, inside of the existing
create_context() routine and the initialization is transitioned
to a new void routine, init_context(). At the same time, in
anticipation of these routines not being called in sequence, a
state boolean is added to the context information structure to
track when the context has been initilized. The context teardown
routine, destroy_context(), is modified to support being called
with a non-initialized context.
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Reviewed-by: Uma Krishnan <ukrishn@linux.vnet.ibm.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2016-03-05 00:55:16 +03:00
if ( ctxi - > initialized ) {
WARN_ON ( ! list_empty ( & ctxi - > luns ) ) ;
2015-08-14 05:47:43 +03:00
cxlflash: Split out context initialization
Presently, context information structures are allocated and
initialized in the same routine, create_context(). This imposes
an ordering restriction such that all pieces of information needed
to initialize a context must be known before the context is even
allocated.
This design point is not flexible when the order of context
creation needs to be modified. Specifically, this can lead to
problems when members of the context information structure are
a part of an ordering dependency (i.e. - the 'work' structure
embedded within the context).
To remedy, the allocation is left as-is, inside of the existing
create_context() routine and the initialization is transitioned
to a new void routine, init_context(). At the same time, in
anticipation of these routines not being called in sequence, a
state boolean is added to the context information structure to
track when the context has been initilized. The context teardown
routine, destroy_context(), is modified to support being called
with a non-initialized context.
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Reviewed-by: Uma Krishnan <ukrishn@linux.vnet.ibm.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2016-03-05 00:55:16 +03:00
/* Clear RHT registers and drop all capabilities for context */
if ( afu - > afu_map & & ctxi - > ctrl_map ) {
writeq_be ( 0 , & ctxi - > ctrl_map - > rht_start ) ;
writeq_be ( 0 , & ctxi - > ctrl_map - > rht_cnt_id ) ;
writeq_be ( 0 , & ctxi - > ctrl_map - > ctx_cap ) ;
}
2015-08-14 05:47:43 +03:00
}
/* Free memory associated with context */
free_page ( ( ulong ) ctxi - > rht_start ) ;
2015-08-14 05:47:53 +03:00
kfree ( ctxi - > rht_needs_ws ) ;
2015-08-14 05:47:43 +03:00
kfree ( ctxi - > rht_lun ) ;
kfree ( ctxi ) ;
}
/**
* create_context ( ) - allocates and initializes a context
* @ cfg : Internal structure associated with the host .
*
* Return : Allocated context on success , NULL on failure
*/
cxlflash: Split out context initialization
Presently, context information structures are allocated and
initialized in the same routine, create_context(). This imposes
an ordering restriction such that all pieces of information needed
to initialize a context must be known before the context is even
allocated.
This design point is not flexible when the order of context
creation needs to be modified. Specifically, this can lead to
problems when members of the context information structure are
a part of an ordering dependency (i.e. - the 'work' structure
embedded within the context).
To remedy, the allocation is left as-is, inside of the existing
create_context() routine and the initialization is transitioned
to a new void routine, init_context(). At the same time, in
anticipation of these routines not being called in sequence, a
state boolean is added to the context information structure to
track when the context has been initilized. The context teardown
routine, destroy_context(), is modified to support being called
with a non-initialized context.
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Reviewed-by: Uma Krishnan <ukrishn@linux.vnet.ibm.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2016-03-05 00:55:16 +03:00
static struct ctx_info * create_context ( struct cxlflash_cfg * cfg )
2015-08-14 05:47:43 +03:00
{
struct device * dev = & cfg - > dev - > dev ;
struct ctx_info * ctxi = NULL ;
struct llun_info * * lli = NULL ;
2015-10-21 23:11:34 +03:00
u8 * ws = NULL ;
2015-08-14 05:47:43 +03:00
struct sisl_rht_entry * rhte ;
ctxi = kzalloc ( sizeof ( * ctxi ) , GFP_KERNEL ) ;
lli = kzalloc ( ( MAX_RHT_PER_CONTEXT * sizeof ( * lli ) ) , GFP_KERNEL ) ;
2015-08-14 05:47:53 +03:00
ws = kzalloc ( ( MAX_RHT_PER_CONTEXT * sizeof ( * ws ) ) , GFP_KERNEL ) ;
if ( unlikely ( ! ctxi | | ! lli | | ! ws ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Unable to allocate context \n " , __func__ ) ;
2015-08-14 05:47:43 +03:00
goto err ;
}
rhte = ( struct sisl_rht_entry * ) get_zeroed_page ( GFP_KERNEL ) ;
if ( unlikely ( ! rhte ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Unable to allocate RHT \n " , __func__ ) ;
2015-08-14 05:47:43 +03:00
goto err ;
}
ctxi - > rht_lun = lli ;
2015-08-14 05:47:53 +03:00
ctxi - > rht_needs_ws = ws ;
2015-08-14 05:47:43 +03:00
ctxi - > rht_start = rhte ;
cxlflash: Split out context initialization
Presently, context information structures are allocated and
initialized in the same routine, create_context(). This imposes
an ordering restriction such that all pieces of information needed
to initialize a context must be known before the context is even
allocated.
This design point is not flexible when the order of context
creation needs to be modified. Specifically, this can lead to
problems when members of the context information structure are
a part of an ordering dependency (i.e. - the 'work' structure
embedded within the context).
To remedy, the allocation is left as-is, inside of the existing
create_context() routine and the initialization is transitioned
to a new void routine, init_context(). At the same time, in
anticipation of these routines not being called in sequence, a
state boolean is added to the context information structure to
track when the context has been initilized. The context teardown
routine, destroy_context(), is modified to support being called
with a non-initialized context.
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Reviewed-by: Uma Krishnan <ukrishn@linux.vnet.ibm.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2016-03-05 00:55:16 +03:00
out :
return ctxi ;
err :
kfree ( ws ) ;
kfree ( lli ) ;
kfree ( ctxi ) ;
ctxi = NULL ;
goto out ;
}
/**
* init_context ( ) - initializes a previously allocated context
* @ ctxi : Previously allocated context
* @ cfg : Internal structure associated with the host .
2018-01-04 01:54:25 +03:00
* @ ctx : Previously obtained context cookie .
cxlflash: Split out context initialization
Presently, context information structures are allocated and
initialized in the same routine, create_context(). This imposes
an ordering restriction such that all pieces of information needed
to initialize a context must be known before the context is even
allocated.
This design point is not flexible when the order of context
creation needs to be modified. Specifically, this can lead to
problems when members of the context information structure are
a part of an ordering dependency (i.e. - the 'work' structure
embedded within the context).
To remedy, the allocation is left as-is, inside of the existing
create_context() routine and the initialization is transitioned
to a new void routine, init_context(). At the same time, in
anticipation of these routines not being called in sequence, a
state boolean is added to the context information structure to
track when the context has been initilized. The context teardown
routine, destroy_context(), is modified to support being called
with a non-initialized context.
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Reviewed-by: Uma Krishnan <ukrishn@linux.vnet.ibm.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2016-03-05 00:55:16 +03:00
* @ ctxid : Previously obtained process element associated with CXL context .
* @ file : Previously obtained file associated with CXL context .
* @ perms : User - specified permissions .
2018-01-04 01:54:37 +03:00
* @ irqs : User - specified number of interrupts .
cxlflash: Split out context initialization
Presently, context information structures are allocated and
initialized in the same routine, create_context(). This imposes
an ordering restriction such that all pieces of information needed
to initialize a context must be known before the context is even
allocated.
This design point is not flexible when the order of context
creation needs to be modified. Specifically, this can lead to
problems when members of the context information structure are
a part of an ordering dependency (i.e. - the 'work' structure
embedded within the context).
To remedy, the allocation is left as-is, inside of the existing
create_context() routine and the initialization is transitioned
to a new void routine, init_context(). At the same time, in
anticipation of these routines not being called in sequence, a
state boolean is added to the context information structure to
track when the context has been initilized. The context teardown
routine, destroy_context(), is modified to support being called
with a non-initialized context.
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Reviewed-by: Uma Krishnan <ukrishn@linux.vnet.ibm.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2016-03-05 00:55:16 +03:00
*/
static void init_context ( struct ctx_info * ctxi , struct cxlflash_cfg * cfg ,
2018-01-04 01:54:37 +03:00
void * ctx , int ctxid , struct file * file , u32 perms ,
u64 irqs )
cxlflash: Split out context initialization
Presently, context information structures are allocated and
initialized in the same routine, create_context(). This imposes
an ordering restriction such that all pieces of information needed
to initialize a context must be known before the context is even
allocated.
This design point is not flexible when the order of context
creation needs to be modified. Specifically, this can lead to
problems when members of the context information structure are
a part of an ordering dependency (i.e. - the 'work' structure
embedded within the context).
To remedy, the allocation is left as-is, inside of the existing
create_context() routine and the initialization is transitioned
to a new void routine, init_context(). At the same time, in
anticipation of these routines not being called in sequence, a
state boolean is added to the context information structure to
track when the context has been initilized. The context teardown
routine, destroy_context(), is modified to support being called
with a non-initialized context.
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Reviewed-by: Uma Krishnan <ukrishn@linux.vnet.ibm.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2016-03-05 00:55:16 +03:00
{
struct afu * afu = cfg - > afu ;
2015-08-14 05:47:43 +03:00
cxlflash: Split out context initialization
Presently, context information structures are allocated and
initialized in the same routine, create_context(). This imposes
an ordering restriction such that all pieces of information needed
to initialize a context must be known before the context is even
allocated.
This design point is not flexible when the order of context
creation needs to be modified. Specifically, this can lead to
problems when members of the context information structure are
a part of an ordering dependency (i.e. - the 'work' structure
embedded within the context).
To remedy, the allocation is left as-is, inside of the existing
create_context() routine and the initialization is transitioned
to a new void routine, init_context(). At the same time, in
anticipation of these routines not being called in sequence, a
state boolean is added to the context information structure to
track when the context has been initilized. The context teardown
routine, destroy_context(), is modified to support being called
with a non-initialized context.
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Reviewed-by: Uma Krishnan <ukrishn@linux.vnet.ibm.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2016-03-05 00:55:16 +03:00
ctxi - > rht_perms = perms ;
2015-08-14 05:47:43 +03:00
ctxi - > ctrl_map = & afu - > afu_map - > ctrls [ ctxid ] . ctrl ;
ctxi - > ctxid = ENCODE_CTXID ( ctxi , ctxid ) ;
2018-01-04 01:54:37 +03:00
ctxi - > irqs = irqs ;
2017-10-26 00:36:29 +03:00
ctxi - > pid = task_tgid_nr ( current ) ; /* tgid = pid */
2015-08-14 05:47:43 +03:00
ctxi - > ctx = ctx ;
2016-08-10 02:39:30 +03:00
ctxi - > cfg = cfg ;
2015-08-14 05:47:43 +03:00
ctxi - > file = file ;
cxlflash: Split out context initialization
Presently, context information structures are allocated and
initialized in the same routine, create_context(). This imposes
an ordering restriction such that all pieces of information needed
to initialize a context must be known before the context is even
allocated.
This design point is not flexible when the order of context
creation needs to be modified. Specifically, this can lead to
problems when members of the context information structure are
a part of an ordering dependency (i.e. - the 'work' structure
embedded within the context).
To remedy, the allocation is left as-is, inside of the existing
create_context() routine and the initialization is transitioned
to a new void routine, init_context(). At the same time, in
anticipation of these routines not being called in sequence, a
state boolean is added to the context information structure to
track when the context has been initilized. The context teardown
routine, destroy_context(), is modified to support being called
with a non-initialized context.
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Reviewed-by: Uma Krishnan <ukrishn@linux.vnet.ibm.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2016-03-05 00:55:16 +03:00
ctxi - > initialized = true ;
2015-08-14 05:47:43 +03:00
mutex_init ( & ctxi - > mutex ) ;
2016-08-10 02:39:42 +03:00
kref_init ( & ctxi - > kref ) ;
2015-08-14 05:47:43 +03:00
INIT_LIST_HEAD ( & ctxi - > luns ) ;
INIT_LIST_HEAD ( & ctxi - > list ) ; /* initialize for list_empty() */
}
2016-08-10 02:39:42 +03:00
/**
* remove_context ( ) - context kref release handler
* @ kref : Kernel reference associated with context to be removed .
*
* When a context no longer has any references it can safely be removed
* from global access and destroyed . Note that it is assumed the thread
* relinquishing access to the context holds its mutex .
*/
static void remove_context ( struct kref * kref )
{
struct ctx_info * ctxi = container_of ( kref , struct ctx_info , kref ) ;
struct cxlflash_cfg * cfg = ctxi - > cfg ;
u64 ctxid = DECODE_CTXID ( ctxi - > ctxid ) ;
/* Remove context from table/error list */
WARN_ON ( ! mutex_is_locked ( & ctxi - > mutex ) ) ;
ctxi - > unavail = true ;
mutex_unlock ( & ctxi - > mutex ) ;
mutex_lock ( & cfg - > ctx_tbl_list_mutex ) ;
mutex_lock ( & ctxi - > mutex ) ;
if ( ! list_empty ( & ctxi - > list ) )
list_del ( & ctxi - > list ) ;
cfg - > ctx_tbl [ ctxid ] = NULL ;
mutex_unlock ( & cfg - > ctx_tbl_list_mutex ) ;
mutex_unlock ( & ctxi - > mutex ) ;
/* Context now completely uncoupled/unreachable */
destroy_context ( cfg , ctxi ) ;
}
2015-08-14 05:47:43 +03:00
/**
* _cxlflash_disk_detach ( ) - detaches a LUN from a context
* @ sdev : SCSI device associated with LUN .
* @ ctxi : Context owning resources .
* @ detach : Detach ioctl data structure .
*
* As part of the detach , all per - context resources associated with the LUN
* are cleaned up . When detaching the last LUN for a context , the context
* itself is cleaned up and released .
*
* Return : 0 on success , - errno on failure
*/
static int _cxlflash_disk_detach ( struct scsi_device * sdev ,
struct ctx_info * ctxi ,
struct dk_cxlflash_detach * detach )
{
2017-01-12 04:19:47 +03:00
struct cxlflash_cfg * cfg = shost_priv ( sdev - > host ) ;
2015-08-14 05:47:43 +03:00
struct device * dev = & cfg - > dev - > dev ;
struct llun_info * lli = sdev - > hostdata ;
struct lun_access * lun_access , * t ;
struct dk_cxlflash_release rel ;
bool put_ctx = false ;
int i ;
int rc = 0 ;
u64 ctxid = DECODE_CTXID ( detach - > context_id ) ,
rctxid = detach - > context_id ;
dev_dbg ( dev , " %s: ctxid=%llu \n " , __func__ , ctxid ) ;
if ( ! ctxi ) {
ctxi = get_context ( cfg , rctxid , lli , CTX_CTRL_ERR_FALLBACK ) ;
if ( unlikely ( ! ctxi ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Bad context ctxid=%llu \n " ,
2015-08-14 05:47:43 +03:00
__func__ , ctxid ) ;
rc = - EINVAL ;
goto out ;
}
put_ctx = true ;
}
/* Cleanup outstanding resources tied to this LUN */
if ( ctxi - > rht_out ) {
marshal_det_to_rele ( detach , & rel ) ;
for ( i = 0 ; i < MAX_RHT_PER_CONTEXT ; i + + ) {
if ( ctxi - > rht_lun [ i ] = = lli ) {
rel . rsrc_handle = i ;
_cxlflash_disk_release ( sdev , ctxi , & rel ) ;
}
/* No need to loop further if we're done */
if ( ctxi - > rht_out = = 0 )
break ;
}
}
/* Take our LUN out of context, free the node */
list_for_each_entry_safe ( lun_access , t , & ctxi - > luns , list )
if ( lun_access - > lli = = lli ) {
list_del ( & lun_access - > list ) ;
kfree ( lun_access ) ;
lun_access = NULL ;
break ;
}
2016-08-10 02:39:42 +03:00
/*
* Release the context reference and the sdev reference that
* bound this LUN to the context .
*/
2016-09-02 23:40:41 +03:00
if ( kref_put ( & ctxi - > kref , remove_context ) )
put_ctx = false ;
2015-10-21 23:11:18 +03:00
scsi_device_put ( sdev ) ;
2015-08-14 05:47:43 +03:00
out :
if ( put_ctx )
put_context ( ctxi ) ;
dev_dbg ( dev , " %s: returning rc=%d \n " , __func__ , rc ) ;
return rc ;
}
static int cxlflash_disk_detach ( struct scsi_device * sdev ,
struct dk_cxlflash_detach * detach )
{
return _cxlflash_disk_detach ( sdev , NULL , detach ) ;
}
/**
* cxlflash_cxl_release ( ) - release handler for adapter file descriptor
* @ inode : File - system inode associated with fd .
* @ file : File installed with adapter file descriptor .
*
* This routine is the release handler for the fops registered with
* the CXL services on an initial attach for a context . It is called
2016-08-10 02:39:52 +03:00
* when a close ( explicity by the user or as part of a process tear
* down ) is performed on the adapter file descriptor returned to the
* user . The user should be aware that explicitly performing a close
* considered catastrophic and subsequent usage of the superpipe API
* with previously saved off tokens will fail .
2015-08-14 05:47:43 +03:00
*
2016-08-10 02:39:52 +03:00
* This routine derives the context reference and calls detach for
* each LUN associated with the context . The final detach operation
* causes the context itself to be freed . With exception to when the
* CXL process element ( context id ) lookup fails ( a case that should
* theoretically never occur ) , every call into this routine results
* in a complete freeing of a context .
2015-08-14 05:47:43 +03:00
*
2018-05-11 22:05:22 +03:00
* Detaching the LUN is typically an ioctl ( ) operation and the underlying
* code assumes that ioctl_rwsem has been acquired as a reader . To support
* that design point , the semaphore is acquired and released around detach .
*
2015-08-14 05:47:43 +03:00
* Return : 0 on success
*/
static int cxlflash_cxl_release ( struct inode * inode , struct file * file )
{
struct cxlflash_cfg * cfg = container_of ( file - > f_op , struct cxlflash_cfg ,
cxl_fops ) ;
2018-01-04 01:55:26 +03:00
void * ctx = cfg - > ops - > fops_get_context ( file ) ;
2015-08-14 05:47:43 +03:00
struct device * dev = & cfg - > dev - > dev ;
struct ctx_info * ctxi = NULL ;
struct dk_cxlflash_detach detach = { { 0 } , 0 } ;
struct lun_access * lun_access , * t ;
enum ctx_ctrl ctrl = CTX_CTRL_ERR_FALLBACK | CTX_CTRL_FILE ;
int ctxid ;
2018-01-04 01:55:26 +03:00
ctxid = cfg - > ops - > process_element ( ctx ) ;
2015-08-14 05:47:43 +03:00
if ( unlikely ( ctxid < 0 ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Context %p was closed ctxid=%d \n " ,
2015-08-14 05:47:43 +03:00
__func__ , ctx , ctxid ) ;
goto out ;
}
ctxi = get_context ( cfg , ctxid , file , ctrl ) ;
if ( unlikely ( ! ctxi ) ) {
ctxi = get_context ( cfg , ctxid , file , ctrl | CTX_CTRL_CLONE ) ;
if ( ! ctxi ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: ctxid=%d already free \n " ,
2015-08-14 05:47:43 +03:00
__func__ , ctxid ) ;
goto out_release ;
}
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Another process owns ctxid=%d \n " ,
2015-08-14 05:47:43 +03:00
__func__ , ctxid ) ;
put_context ( ctxi ) ;
goto out ;
}
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: close for ctxid=%d \n " , __func__ , ctxid ) ;
2015-08-14 05:47:43 +03:00
2018-05-11 22:05:22 +03:00
down_read ( & cfg - > ioctl_rwsem ) ;
2015-08-14 05:47:43 +03:00
detach . context_id = ctxi - > ctxid ;
list_for_each_entry_safe ( lun_access , t , & ctxi - > luns , list )
_cxlflash_disk_detach ( lun_access - > sdev , ctxi , & detach ) ;
2018-05-11 22:05:22 +03:00
up_read ( & cfg - > ioctl_rwsem ) ;
2015-08-14 05:47:43 +03:00
out_release :
2018-01-04 01:55:26 +03:00
cfg - > ops - > fd_release ( inode , file ) ;
2015-08-14 05:47:43 +03:00
out :
dev_dbg ( dev , " %s: returning \n " , __func__ ) ;
return 0 ;
}
/**
* unmap_context ( ) - clears a previously established mapping
* @ ctxi : Context owning the mapping .
*
* This routine is used to switch between the error notification page
* ( dummy page of all 1 ' s ) and the real mapping ( established by the CXL
* fault handler ) .
*/
static void unmap_context ( struct ctx_info * ctxi )
{
unmap_mapping_range ( ctxi - > file - > f_mapping , 0 , 0 , 1 ) ;
}
/**
* get_err_page ( ) - obtains and allocates the error notification page
2017-01-12 04:19:47 +03:00
* @ cfg : Internal structure associated with the host .
2015-08-14 05:47:43 +03:00
*
* Return : error notification page on success , NULL on failure
*/
2017-01-12 04:19:47 +03:00
static struct page * get_err_page ( struct cxlflash_cfg * cfg )
2015-08-14 05:47:43 +03:00
{
struct page * err_page = global . err_page ;
2017-01-12 04:19:47 +03:00
struct device * dev = & cfg - > dev - > dev ;
2015-08-14 05:47:43 +03:00
if ( unlikely ( ! err_page ) ) {
err_page = alloc_page ( GFP_KERNEL ) ;
if ( unlikely ( ! err_page ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Unable to allocate err_page \n " ,
__func__ ) ;
2015-08-14 05:47:43 +03:00
goto out ;
}
memset ( page_address ( err_page ) , - 1 , PAGE_SIZE ) ;
/* Serialize update w/ other threads to avoid a leak */
mutex_lock ( & global . mutex ) ;
if ( likely ( ! global . err_page ) )
global . err_page = err_page ;
else {
__free_page ( err_page ) ;
err_page = global . err_page ;
}
mutex_unlock ( & global . mutex ) ;
}
out :
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: returning err_page=%p \n " , __func__ , err_page ) ;
2015-08-14 05:47:43 +03:00
return err_page ;
}
/**
* cxlflash_mmap_fault ( ) - mmap fault handler for adapter file descriptor
* @ vmf : VM fault associated with current fault .
*
* To support error notification via MMIO , faults are ' caught ' by this routine
* that was inserted before passing back the adapter file descriptor on attach .
* When a fault occurs , this routine evaluates if error recovery is active and
* if so , installs the error page to ' notify ' the user about the error state .
* During normal operation , the fault is simply handled by the original fault
* handler that was installed by CXL services as part of initializing the
* adapter file descriptor . The VMA ' s page protection bits are toggled to
* indicate cached / not - cached depending on the memory backing the fault .
*
* Return : 0 on success , VM_FAULT_SIGBUS on failure
*/
2018-06-20 21:25:07 +03:00
static vm_fault_t cxlflash_mmap_fault ( struct vm_fault * vmf )
2015-08-14 05:47:43 +03:00
{
2017-02-25 01:56:41 +03:00
struct vm_area_struct * vma = vmf - > vma ;
2015-08-14 05:47:43 +03:00
struct file * file = vma - > vm_file ;
struct cxlflash_cfg * cfg = container_of ( file - > f_op , struct cxlflash_cfg ,
cxl_fops ) ;
2018-01-04 01:55:26 +03:00
void * ctx = cfg - > ops - > fops_get_context ( file ) ;
2015-08-14 05:47:43 +03:00
struct device * dev = & cfg - > dev - > dev ;
struct ctx_info * ctxi = NULL ;
struct page * err_page = NULL ;
enum ctx_ctrl ctrl = CTX_CTRL_ERR_FALLBACK | CTX_CTRL_FILE ;
2018-06-20 21:25:07 +03:00
vm_fault_t rc = 0 ;
2015-08-14 05:47:43 +03:00
int ctxid ;
2018-01-04 01:55:26 +03:00
ctxid = cfg - > ops - > process_element ( ctx ) ;
2015-08-14 05:47:43 +03:00
if ( unlikely ( ctxid < 0 ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Context %p was closed ctxid=%d \n " ,
2015-08-14 05:47:43 +03:00
__func__ , ctx , ctxid ) ;
goto err ;
}
ctxi = get_context ( cfg , ctxid , file , ctrl ) ;
if ( unlikely ( ! ctxi ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Bad context ctxid=%d \n " , __func__ , ctxid ) ;
2015-08-14 05:47:43 +03:00
goto err ;
}
2016-08-10 02:40:02 +03:00
dev_dbg ( dev , " %s: fault for context %d \n " , __func__ , ctxid ) ;
2015-08-14 05:47:43 +03:00
if ( likely ( ! ctxi - > err_recovery_active ) ) {
vma - > vm_page_prot = pgprot_noncached ( vma - > vm_page_prot ) ;
2017-02-25 01:56:41 +03:00
rc = ctxi - > cxl_mmap_vmops - > fault ( vmf ) ;
2015-08-14 05:47:43 +03:00
} else {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: err recovery active, use err_page \n " ,
2015-08-14 05:47:43 +03:00
__func__ ) ;
2017-01-12 04:19:47 +03:00
err_page = get_err_page ( cfg ) ;
2015-08-14 05:47:43 +03:00
if ( unlikely ( ! err_page ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Could not get err_page \n " , __func__ ) ;
2015-08-14 05:47:43 +03:00
rc = VM_FAULT_RETRY ;
goto out ;
}
get_page ( err_page ) ;
vmf - > page = err_page ;
vma - > vm_page_prot = pgprot_cached ( vma - > vm_page_prot ) ;
}
out :
if ( likely ( ctxi ) )
put_context ( ctxi ) ;
2018-06-20 21:25:07 +03:00
dev_dbg ( dev , " %s: returning rc=%x \n " , __func__ , rc ) ;
2015-08-14 05:47:43 +03:00
return rc ;
err :
rc = VM_FAULT_SIGBUS ;
goto out ;
}
/*
* Local MMAP vmops to ' catch ' faults
*/
static const struct vm_operations_struct cxlflash_mmap_vmops = {
. fault = cxlflash_mmap_fault ,
} ;
/**
* cxlflash_cxl_mmap ( ) - mmap handler for adapter file descriptor
* @ file : File installed with adapter file descriptor .
* @ vma : VM area associated with mapping .
*
* Installs local mmap vmops to ' catch ' faults for error notification support .
*
* Return : 0 on success , - errno on failure
*/
static int cxlflash_cxl_mmap ( struct file * file , struct vm_area_struct * vma )
{
struct cxlflash_cfg * cfg = container_of ( file - > f_op , struct cxlflash_cfg ,
cxl_fops ) ;
2018-01-04 01:55:26 +03:00
void * ctx = cfg - > ops - > fops_get_context ( file ) ;
2015-08-14 05:47:43 +03:00
struct device * dev = & cfg - > dev - > dev ;
struct ctx_info * ctxi = NULL ;
enum ctx_ctrl ctrl = CTX_CTRL_ERR_FALLBACK | CTX_CTRL_FILE ;
int ctxid ;
int rc = 0 ;
2018-01-04 01:55:26 +03:00
ctxid = cfg - > ops - > process_element ( ctx ) ;
2015-08-14 05:47:43 +03:00
if ( unlikely ( ctxid < 0 ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Context %p was closed ctxid=%d \n " ,
2015-08-14 05:47:43 +03:00
__func__ , ctx , ctxid ) ;
rc = - EIO ;
goto out ;
}
ctxi = get_context ( cfg , ctxid , file , ctrl ) ;
if ( unlikely ( ! ctxi ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Bad context ctxid=%d \n " , __func__ , ctxid ) ;
2015-08-14 05:47:43 +03:00
rc = - EIO ;
goto out ;
}
2016-08-10 02:40:02 +03:00
dev_dbg ( dev , " %s: mmap for context %d \n " , __func__ , ctxid ) ;
2015-08-14 05:47:43 +03:00
2018-01-04 01:55:26 +03:00
rc = cfg - > ops - > fd_mmap ( file , vma ) ;
2015-08-14 05:47:43 +03:00
if ( likely ( ! rc ) ) {
/* Insert ourself in the mmap fault handler path */
ctxi - > cxl_mmap_vmops = vma - > vm_ops ;
vma - > vm_ops = & cxlflash_mmap_vmops ;
}
out :
if ( likely ( ctxi ) )
put_context ( ctxi ) ;
return rc ;
}
cxlflash: Fix to avoid corrupting adapter fops
The fops owned by the adapter can be corrupted in certain scenarios,
opening a window where certain fops are temporarily NULLed before being
reset to their proper value. This can potentially lead software to make
incorrect decisions, leaving the user with the inability to function as
intended.
An example of this behavior can be observed when there are a number of
users with a high rate of turn around (attach to LUN, perform an I/O,
detach from LUN, repeat). Every so often a user is given a valid
context and adapter file descriptor, but the file associated with the
descriptor lacks the correct read permission bit (FMODE_CAN_READ) and
thus the read system call bails before calling the valid read fop.
Background:
The fops is stored in the adapter structure to provide the ability to
lookup the adapter structure from within the fop handler. CXL services
use the file's private_data and at present, the CXL context does not
have a private section. In an effort to limit areas of the cxlflash
driver with code specific the superpipe function, a design choice was
made to keep the details of the fops situated away from the legacy
portions of the driver. This drove the behavior that the adapter fops
is set at the beginning of the disk attach ioctl handler when there
are no users present.
The corruption that this fix remedies is due to the fact that the fops
is initially defaulted to values found within a static structure. When
the fops is handed down to the CXL services later in the attach path,
certain services are patched. The fops structure remains correct until
the user count drops to 0 and the fops is reset, triggering the process
to repeat again. The user counts are tightly coupled with the creation
and deletion of the user context. If multiple users perform a disk
attach at the same time, when the user count is currently 0, some users
can be in the middle of obtaining a file descriptor and have not yet
reached the context creation code that [in addition to creating the
context] increments the user count. Subsequent users coming in to
perform the attach see that the user count is still 0, and reinitialize
the fops, temporarily removing the patched fops. The users that are in
the middle obtaining their file descriptor may then receive an invalid
descriptor.
The fix simply removes the user count altogether and moves the fops
initialization to probe time such that it is only performed one time
for the life of the adapter. In the future, if the CXL services adopt
a private member for their context, that could be used to store the
adapter structure reference and cxlflash could revert to a model that
does not require an embedded fops.
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Signed-off-by: Manoj N. Kumar <manoj@linux.vnet.ibm.com>
Reviewed-by: Brian King <brking@linux.vnet.ibm.com>
Reviewed-by: Andrew Donnellan <andrew.donnellan@au1.ibm.com>
Reviewed-by: Daniel Axtens <dja@axtens.net>
Reviewed-by: Tomas Henzl <thenzl@redhat.com>
Signed-off-by: James Bottomley <JBottomley@Odin.com>
2015-10-21 23:15:37 +03:00
const struct file_operations cxlflash_cxl_fops = {
2015-08-14 05:47:43 +03:00
. owner = THIS_MODULE ,
. mmap = cxlflash_cxl_mmap ,
. release = cxlflash_cxl_release ,
} ;
/**
* cxlflash_mark_contexts_error ( ) - move contexts to error state and list
* @ cfg : Internal structure associated with the host .
*
* A context is only moved over to the error list when there are no outstanding
* references to it . This ensures that a running operation has completed .
*
* Return : 0 on success , - errno on failure
*/
int cxlflash_mark_contexts_error ( struct cxlflash_cfg * cfg )
{
int i , rc = 0 ;
struct ctx_info * ctxi = NULL ;
mutex_lock ( & cfg - > ctx_tbl_list_mutex ) ;
for ( i = 0 ; i < MAX_CONTEXT ; i + + ) {
ctxi = cfg - > ctx_tbl [ i ] ;
if ( ctxi ) {
mutex_lock ( & ctxi - > mutex ) ;
cfg - > ctx_tbl [ i ] = NULL ;
list_add ( & ctxi - > list , & cfg - > ctx_err_recovery ) ;
ctxi - > err_recovery_active = true ;
ctxi - > ctrl_map = NULL ;
unmap_context ( ctxi ) ;
mutex_unlock ( & ctxi - > mutex ) ;
}
}
mutex_unlock ( & cfg - > ctx_tbl_list_mutex ) ;
return rc ;
}
/*
* Dummy NULL fops
*/
static const struct file_operations null_fops = {
. owner = THIS_MODULE ,
} ;
2015-10-21 23:11:52 +03:00
/**
* check_state ( ) - checks and responds to the current adapter state
* @ cfg : Internal structure associated with the host .
*
* This routine can block and should only be used on process context .
* It assumes that the caller is an ioctl thread and holding the ioctl
* read semaphore . This is temporarily let up across the wait to allow
* for draining actively running ioctls . Also note that when waking up
* from waiting in reset , the state is unknown and must be checked again
* before proceeding .
*
* Return : 0 on success , - errno on failure
*/
cxlflash: Fix to avoid potential deadlock on EEH
Ioctl threads that use scsi_execute() can run for an excessive amount
of time due to the fact that they have lengthy timeouts and retry logic
built in. Under normal operation this is not an issue. However, once EEH
enters the picture, a long execution time coupled with the possibility
that a timeout can trigger entry to the driver via registered reset
callbacks becomes a liability.
In particular, a deadlock can occur when an EEH event is encountered
while in running in scsi_execute(). As part of the recovery, the EEH
handler drains all currently running ioctls, waiting until they have
completed before proceeding with a reset. As the scsi_execute()'s are
situated on the ioctl path, the EEH handler will wait until they (and
the remainder of the ioctl handler they're associated with) have
completed. Normally this would not be much of an issue aside from the
longer recovery period. Unfortunately, the scsi_execute() triggers a
reset when it times out. The reset handler will see that the device is
already being reset and wait until that reset completed. This creates
a condition where the EEH handler becomes stuck, infinitely waiting for
the ioctl thread to complete.
To avoid this behavior, temporarily unmark the scsi_execute() threads
as an ioctl thread by releasing the ioctl read semaphore. This allows
the EEH handler to proceed with a recovery while the thread is still
running. Once the scsi_execute() returns, the ioctl read semaphore is
reacquired and the adapter state is rechecked in case it changed while
inside of scsi_execute(). The state check will wait if the adapter is
still being recovered or returns a failure if the recovery failed. In
the event that the adapter reset failed, the failure is simply returned
as the ioctl would be unable to continue.
Reported-by: Brian King <brking@linux.vnet.ibm.com>
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Signed-off-by: Manoj N. Kumar <manoj@linux.vnet.ibm.com>
Reviewed-by: Brian King <brking@linux.vnet.ibm.com>
Reviewed-by: Daniel Axtens <dja@axtens.net>
Reviewed-by: Tomas Henzl <thenzl@redhat.com>
Signed-off-by: James Bottomley <JBottomley@Odin.com>
2015-10-21 23:15:52 +03:00
int check_state ( struct cxlflash_cfg * cfg )
2015-10-21 23:11:52 +03:00
{
struct device * dev = & cfg - > dev - > dev ;
int rc = 0 ;
retry :
switch ( cfg - > state ) {
2015-10-21 23:12:00 +03:00
case STATE_RESET :
dev_dbg ( dev , " %s: Reset state, going to wait... \n " , __func__ ) ;
2015-10-21 23:11:52 +03:00
up_read ( & cfg - > ioctl_rwsem ) ;
2015-10-21 23:12:00 +03:00
rc = wait_event_interruptible ( cfg - > reset_waitq ,
cfg - > state ! = STATE_RESET ) ;
2015-10-21 23:11:52 +03:00
down_read ( & cfg - > ioctl_rwsem ) ;
if ( unlikely ( rc ) )
break ;
goto retry ;
case STATE_FAILTERM :
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Failed/Terminating \n " , __func__ ) ;
2015-10-21 23:11:52 +03:00
rc = - ENODEV ;
break ;
default :
break ;
}
return rc ;
}
2015-08-14 05:47:43 +03:00
/**
* cxlflash_disk_attach ( ) - attach a LUN to a context
* @ sdev : SCSI device associated with LUN .
* @ attach : Attach ioctl data structure .
*
* Creates a context and attaches LUN to it . A LUN can only be attached
* one time to a context ( subsequent attaches for the same context / LUN pair
* are not supported ) . Additional LUNs can be attached to a context by
* specifying the ' reuse ' flag defined in the cxlflash_ioctl . h header .
*
* Return : 0 on success , - errno on failure
*/
static int cxlflash_disk_attach ( struct scsi_device * sdev ,
struct dk_cxlflash_attach * attach )
{
2017-01-12 04:19:47 +03:00
struct cxlflash_cfg * cfg = shost_priv ( sdev - > host ) ;
2015-08-14 05:47:43 +03:00
struct device * dev = & cfg - > dev - > dev ;
struct afu * afu = cfg - > afu ;
struct llun_info * lli = sdev - > hostdata ;
struct glun_info * gli = lli - > parent ;
struct ctx_info * ctxi = NULL ;
struct lun_access * lun_access = NULL ;
int rc = 0 ;
u32 perms ;
int ctxid = - 1 ;
2018-01-04 01:54:37 +03:00
u64 irqs = attach - > num_interrupts ;
2017-01-12 04:19:33 +03:00
u64 flags = 0UL ;
2015-08-14 05:47:43 +03:00
u64 rctxid = 0UL ;
2016-03-05 00:55:17 +03:00
struct file * file = NULL ;
2015-08-14 05:47:43 +03:00
2018-01-04 01:54:25 +03:00
void * ctx = NULL ;
2015-08-14 05:47:43 +03:00
int fd = - 1 ;
2018-01-04 01:54:37 +03:00
if ( irqs > 4 ) {
2015-08-14 05:47:43 +03:00
dev_dbg ( dev , " %s: Cannot support this many interrupts %llu \n " ,
2018-01-04 01:54:37 +03:00
__func__ , irqs ) ;
2015-08-14 05:47:43 +03:00
rc = - EINVAL ;
goto out ;
}
if ( gli - > max_lba = = 0 ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: No capacity info for LUN=%016llx \n " ,
2015-08-14 05:47:43 +03:00
__func__ , lli - > lun_id [ sdev - > channel ] ) ;
rc = read_cap16 ( sdev , lli ) ;
if ( rc ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Invalid device rc=%d \n " ,
2015-08-14 05:47:43 +03:00
__func__ , rc ) ;
rc = - ENODEV ;
goto out ;
}
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: LBA = %016llx \n " , __func__ , gli - > max_lba ) ;
dev_dbg ( dev , " %s: BLK_LEN = %08x \n " , __func__ , gli - > blk_len ) ;
2015-08-14 05:47:43 +03:00
}
if ( attach - > hdr . flags & DK_CXLFLASH_ATTACH_REUSE_CONTEXT ) {
rctxid = attach - > context_id ;
ctxi = get_context ( cfg , rctxid , NULL , 0 ) ;
if ( ! ctxi ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Bad context rctxid=%016llx \n " ,
2015-08-14 05:47:43 +03:00
__func__ , rctxid ) ;
rc = - EINVAL ;
goto out ;
}
list_for_each_entry ( lun_access , & ctxi - > luns , list )
if ( lun_access - > lli = = lli ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Already attached \n " ,
2015-08-14 05:47:43 +03:00
__func__ ) ;
rc = - EINVAL ;
goto out ;
}
}
2015-10-21 23:11:18 +03:00
rc = scsi_device_get ( sdev ) ;
if ( unlikely ( rc ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Unable to get sdev reference \n " , __func__ ) ;
2015-10-21 23:11:18 +03:00
goto out ;
}
2015-08-14 05:47:43 +03:00
lun_access = kzalloc ( sizeof ( * lun_access ) , GFP_KERNEL ) ;
if ( unlikely ( ! lun_access ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Unable to allocate lun_access \n " , __func__ ) ;
2015-08-14 05:47:43 +03:00
rc = - ENOMEM ;
2016-03-05 00:55:17 +03:00
goto err ;
2015-08-14 05:47:43 +03:00
}
lun_access - > lli = lli ;
lun_access - > sdev = sdev ;
2016-08-10 02:39:42 +03:00
/* Non-NULL context indicates reuse (another context reference) */
2015-08-14 05:47:43 +03:00
if ( ctxi ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Reusing context for LUN rctxid=%016llx \n " ,
2015-08-14 05:47:43 +03:00
__func__ , rctxid ) ;
2016-08-10 02:39:42 +03:00
kref_get ( & ctxi - > kref ) ;
2015-08-14 05:47:43 +03:00
list_add ( & lun_access - > list , & ctxi - > luns ) ;
goto out_attach ;
}
2016-03-05 00:55:18 +03:00
ctxi = create_context ( cfg ) ;
if ( unlikely ( ! ctxi ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Failed to create context ctxid=%d \n " ,
2016-03-05 00:55:18 +03:00
__func__ , ctxid ) ;
2017-08-15 23:18:14 +03:00
rc = - ENOMEM ;
2016-03-05 00:55:18 +03:00
goto err ;
}
2018-01-04 01:55:26 +03:00
ctx = cfg - > ops - > dev_context_init ( cfg - > dev , cfg - > afu_cookie ) ;
2015-12-04 00:18:49 +03:00
if ( IS_ERR_OR_NULL ( ctx ) ) {
2015-08-14 05:47:43 +03:00
dev_err ( dev , " %s: Could not initialize context %p \n " ,
__func__ , ctx ) ;
rc = - ENODEV ;
2016-03-05 00:55:17 +03:00
goto err ;
2015-08-14 05:47:43 +03:00
}
2018-01-04 01:55:26 +03:00
rc = cfg - > ops - > start_work ( ctx , irqs ) ;
2016-03-05 00:55:18 +03:00
if ( unlikely ( rc ) ) {
dev_dbg ( dev , " %s: Could not start context rc=%d \n " ,
__func__ , rc ) ;
goto err ;
}
2018-01-04 01:55:26 +03:00
ctxid = cfg - > ops - > process_element ( ctx ) ;
2015-12-03 02:26:28 +03:00
if ( unlikely ( ( ctxid > = MAX_CONTEXT ) | | ( ctxid < 0 ) ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: ctxid=%d invalid \n " , __func__ , ctxid ) ;
2015-08-14 05:47:43 +03:00
rc = - EPERM ;
2016-03-05 00:55:17 +03:00
goto err ;
2015-08-14 05:47:43 +03:00
}
2018-01-04 01:55:26 +03:00
file = cfg - > ops - > get_fd ( ctx , & cfg - > cxl_fops , & fd ) ;
2015-08-14 05:47:43 +03:00
if ( unlikely ( fd < 0 ) ) {
rc = - ENODEV ;
dev_err ( dev , " %s: Could not get file descriptor \n " , __func__ ) ;
2016-03-05 00:55:17 +03:00
goto err ;
2015-08-14 05:47:43 +03:00
}
/* Translate read/write O_* flags from fcntl.h to AFU permission bits */
perms = SISL_RHT_PERM ( attach - > hdr . flags + 1 ) ;
cxlflash: Split out context initialization
Presently, context information structures are allocated and
initialized in the same routine, create_context(). This imposes
an ordering restriction such that all pieces of information needed
to initialize a context must be known before the context is even
allocated.
This design point is not flexible when the order of context
creation needs to be modified. Specifically, this can lead to
problems when members of the context information structure are
a part of an ordering dependency (i.e. - the 'work' structure
embedded within the context).
To remedy, the allocation is left as-is, inside of the existing
create_context() routine and the initialization is transitioned
to a new void routine, init_context(). At the same time, in
anticipation of these routines not being called in sequence, a
state boolean is added to the context information structure to
track when the context has been initilized. The context teardown
routine, destroy_context(), is modified to support being called
with a non-initialized context.
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Reviewed-by: Uma Krishnan <ukrishn@linux.vnet.ibm.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2016-03-05 00:55:16 +03:00
/* Context mutex is locked upon return */
2018-01-04 01:54:37 +03:00
init_context ( ctxi , cfg , ctx , ctxid , file , perms , irqs ) ;
cxlflash: Split out context initialization
Presently, context information structures are allocated and
initialized in the same routine, create_context(). This imposes
an ordering restriction such that all pieces of information needed
to initialize a context must be known before the context is even
allocated.
This design point is not flexible when the order of context
creation needs to be modified. Specifically, this can lead to
problems when members of the context information structure are
a part of an ordering dependency (i.e. - the 'work' structure
embedded within the context).
To remedy, the allocation is left as-is, inside of the existing
create_context() routine and the initialization is transitioned
to a new void routine, init_context(). At the same time, in
anticipation of these routines not being called in sequence, a
state boolean is added to the context information structure to
track when the context has been initilized. The context teardown
routine, destroy_context(), is modified to support being called
with a non-initialized context.
Signed-off-by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>
Reviewed-by: Uma Krishnan <ukrishn@linux.vnet.ibm.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
2016-03-05 00:55:16 +03:00
2015-08-14 05:47:43 +03:00
rc = afu_attach ( cfg , ctxi ) ;
if ( unlikely ( rc ) ) {
dev_err ( dev , " %s: Could not attach AFU rc %d \n " , __func__ , rc ) ;
2016-03-05 00:55:17 +03:00
goto err ;
2015-08-14 05:47:43 +03:00
}
/*
* No error paths after this point . Once the fd is installed it ' s
* visible to user space and can ' t be undone safely on this thread .
* There is no need to worry about a deadlock here because no one
* knows about us yet ; we can be the only one holding our mutex .
*/
list_add ( & lun_access - > list , & ctxi - > luns ) ;
mutex_lock ( & cfg - > ctx_tbl_list_mutex ) ;
mutex_lock ( & ctxi - > mutex ) ;
cfg - > ctx_tbl [ ctxid ] = ctxi ;
mutex_unlock ( & cfg - > ctx_tbl_list_mutex ) ;
fd_install ( fd , file ) ;
out_attach :
2016-08-10 02:39:52 +03:00
if ( fd ! = - 1 )
2017-01-12 04:19:33 +03:00
flags | = DK_CXLFLASH_APP_CLOSE_ADAP_FD ;
if ( afu_is_sq_cmd_mode ( afu ) )
flags | = DK_CXLFLASH_CONTEXT_SQ_CMD_MODE ;
2016-08-10 02:39:52 +03:00
2017-01-12 04:19:33 +03:00
attach - > hdr . return_flags = flags ;
2015-08-14 05:47:43 +03:00
attach - > context_id = ctxi - > ctxid ;
attach - > block_size = gli - > blk_len ;
attach - > mmio_size = sizeof ( afu - > afu_map - > hosts [ 0 ] . harea ) ;
attach - > last_lba = gli - > max_lba ;
2015-10-21 23:11:10 +03:00
attach - > max_xfer = sdev - > host - > max_sectors * MAX_SECTOR_UNIT ;
attach - > max_xfer / = gli - > blk_len ;
2015-08-14 05:47:43 +03:00
out :
attach - > adap_fd = fd ;
if ( ctxi )
put_context ( ctxi ) ;
dev_dbg ( dev , " %s: returning ctxid=%d fd=%d bs=%lld rc=%d llba=%lld \n " ,
__func__ , ctxid , fd , attach - > block_size , rc , attach - > last_lba ) ;
return rc ;
2016-03-05 00:55:17 +03:00
err :
/* Cleanup CXL context; okay to 'stop' even if it was not started */
if ( ! IS_ERR_OR_NULL ( ctx ) ) {
2018-01-04 01:55:26 +03:00
cfg - > ops - > stop_context ( ctx ) ;
cfg - > ops - > release_context ( ctx ) ;
2016-03-05 00:55:17 +03:00
ctx = NULL ;
}
2015-08-14 05:47:43 +03:00
/*
* Here , we ' re overriding the fops with a dummy all - NULL fops because
* fput ( ) calls the release fop , which will cause us to mistakenly
* call into the CXL code . Rather than try to add yet more complexity
* to that routine ( cxlflash_cxl_release ) we should try to fix the
* issue here .
*/
2016-03-05 00:55:17 +03:00
if ( fd > 0 ) {
file - > f_op = & null_fops ;
fput ( file ) ;
put_unused_fd ( fd ) ;
fd = - 1 ;
file = NULL ;
}
2016-08-10 02:39:18 +03:00
/* Cleanup our context */
2016-03-05 00:55:17 +03:00
if ( ctxi ) {
destroy_context ( cfg , ctxi ) ;
ctxi = NULL ;
}
2015-08-14 05:47:43 +03:00
kfree ( lun_access ) ;
2015-10-21 23:11:18 +03:00
scsi_device_put ( sdev ) ;
2015-08-14 05:47:43 +03:00
goto out ;
}
/**
* recover_context ( ) - recovers a context in error
* @ cfg : Internal structure associated with the host .
* @ ctxi : Context to release .
2016-08-10 02:40:02 +03:00
* @ adap_fd : Adapter file descriptor associated with new / recovered context .
2015-08-14 05:47:43 +03:00
*
* Restablishes the state for a context - in - error .
*
* Return : 0 on success , - errno on failure
*/
2016-08-10 02:40:02 +03:00
static int recover_context ( struct cxlflash_cfg * cfg ,
struct ctx_info * ctxi ,
int * adap_fd )
2015-08-14 05:47:43 +03:00
{
struct device * dev = & cfg - > dev - > dev ;
int rc = 0 ;
2016-08-10 02:39:52 +03:00
int fd = - 1 ;
2015-08-14 05:47:43 +03:00
int ctxid = - 1 ;
struct file * file ;
2018-01-04 01:54:25 +03:00
void * ctx ;
2015-08-14 05:47:43 +03:00
struct afu * afu = cfg - > afu ;
2018-01-04 01:55:26 +03:00
ctx = cfg - > ops - > dev_context_init ( cfg - > dev , cfg - > afu_cookie ) ;
2015-12-04 00:18:49 +03:00
if ( IS_ERR_OR_NULL ( ctx ) ) {
2015-08-14 05:47:43 +03:00
dev_err ( dev , " %s: Could not initialize context %p \n " ,
__func__ , ctx ) ;
rc = - ENODEV ;
goto out ;
}
2018-01-04 01:55:26 +03:00
rc = cfg - > ops - > start_work ( ctx , ctxi - > irqs ) ;
2016-03-05 00:55:18 +03:00
if ( unlikely ( rc ) ) {
dev_dbg ( dev , " %s: Could not start context rc=%d \n " ,
__func__ , rc ) ;
goto err1 ;
}
2018-01-04 01:55:26 +03:00
ctxid = cfg - > ops - > process_element ( ctx ) ;
2015-12-03 02:26:28 +03:00
if ( unlikely ( ( ctxid > = MAX_CONTEXT ) | | ( ctxid < 0 ) ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: ctxid=%d invalid \n " , __func__ , ctxid ) ;
2015-08-14 05:47:43 +03:00
rc = - EPERM ;
2016-03-05 00:55:18 +03:00
goto err2 ;
2015-08-14 05:47:43 +03:00
}
2018-01-04 01:55:26 +03:00
file = cfg - > ops - > get_fd ( ctx , & cfg - > cxl_fops , & fd ) ;
2015-08-14 05:47:43 +03:00
if ( unlikely ( fd < 0 ) ) {
rc = - ENODEV ;
dev_err ( dev , " %s: Could not get file descriptor \n " , __func__ ) ;
goto err2 ;
}
/* Update with new MMIO area based on updated context id */
ctxi - > ctrl_map = & afu - > afu_map - > ctrls [ ctxid ] . ctrl ;
rc = afu_attach ( cfg , ctxi ) ;
if ( rc ) {
dev_err ( dev , " %s: Could not attach AFU rc %d \n " , __func__ , rc ) ;
goto err3 ;
}
/*
* No error paths after this point . Once the fd is installed it ' s
* visible to user space and can ' t be undone safely on this thread .
*/
ctxi - > ctxid = ENCODE_CTXID ( ctxi , ctxid ) ;
ctxi - > ctx = ctx ;
ctxi - > file = file ;
/*
* Put context back in table ( note the reinit of the context list ) ;
* we must first drop the context ' s mutex and then acquire it in
* order with the table / list mutex to avoid a deadlock - safe to do
* here because no one can find us at this moment in time .
*/
mutex_unlock ( & ctxi - > mutex ) ;
mutex_lock ( & cfg - > ctx_tbl_list_mutex ) ;
mutex_lock ( & ctxi - > mutex ) ;
list_del_init ( & ctxi - > list ) ;
cfg - > ctx_tbl [ ctxid ] = ctxi ;
mutex_unlock ( & cfg - > ctx_tbl_list_mutex ) ;
fd_install ( fd , file ) ;
2016-08-10 02:40:02 +03:00
* adap_fd = fd ;
2015-08-14 05:47:43 +03:00
out :
dev_dbg ( dev , " %s: returning ctxid=%d fd=%d rc=%d \n " ,
__func__ , ctxid , fd , rc ) ;
return rc ;
err3 :
fput ( file ) ;
put_unused_fd ( fd ) ;
2016-03-05 00:55:18 +03:00
err2 :
2018-01-04 01:55:26 +03:00
cfg - > ops - > stop_context ( ctx ) ;
2015-08-14 05:47:43 +03:00
err1 :
2018-01-04 01:55:26 +03:00
cfg - > ops - > release_context ( ctx ) ;
2015-08-14 05:47:43 +03:00
goto out ;
}
/**
* cxlflash_afu_recover ( ) - initiates AFU recovery
* @ sdev : SCSI device associated with LUN .
* @ recover : Recover ioctl data structure .
*
* Only a single recovery is allowed at a time to avoid exhausting CXL
* resources ( leading to recovery failure ) in the event that we ' re up
* against the maximum number of contexts limit . For similar reasons ,
* a context recovery is retried if there are multiple recoveries taking
* place at the same time and the failure was due to CXL services being
* unable to keep up .
*
2016-05-03 19:27:34 +03:00
* As this routine is called on ioctl context , it holds the ioctl r / w
* semaphore that is used to drain ioctls in recovery scenarios . The
* implementation to achieve the pacing described above ( a local mutex )
* requires that the ioctl r / w semaphore be dropped and reacquired to
* avoid a 3 - way deadlock when multiple process recoveries operate in
* parallel .
*
2015-08-14 05:47:43 +03:00
* Because a user can detect an error condition before the kernel , it is
* quite possible for this routine to act as the kernel ' s EEH detection
* source ( MMIO read of mbox_r ) . Because of this , there is a window of
* time where an EEH might have been detected but not yet ' serviced '
2015-10-21 23:12:00 +03:00
* ( callback invoked , causing the device to enter reset state ) . To avoid
2015-08-14 05:47:43 +03:00
* looping in this routine during that window , a 1 second sleep is in place
* between the time the MMIO failure is detected and the time a wait on the
2015-10-21 23:12:00 +03:00
* reset wait queue is attempted via check_state ( ) .
2015-08-14 05:47:43 +03:00
*
* Return : 0 on success , - errno on failure
*/
static int cxlflash_afu_recover ( struct scsi_device * sdev ,
struct dk_cxlflash_recover_afu * recover )
{
2017-01-12 04:19:47 +03:00
struct cxlflash_cfg * cfg = shost_priv ( sdev - > host ) ;
2015-08-14 05:47:43 +03:00
struct device * dev = & cfg - > dev - > dev ;
struct llun_info * lli = sdev - > hostdata ;
struct afu * afu = cfg - > afu ;
struct ctx_info * ctxi = NULL ;
struct mutex * mutex = & cfg - > ctx_recovery_mutex ;
2017-04-12 22:15:42 +03:00
struct hwq * hwq = get_hwq ( afu , PRIMARY_HWQ ) ;
2017-01-12 04:19:33 +03:00
u64 flags ;
2015-08-14 05:47:43 +03:00
u64 ctxid = DECODE_CTXID ( recover - > context_id ) ,
rctxid = recover - > context_id ;
long reg ;
2017-08-26 01:18:03 +03:00
bool locked = true ;
2015-08-14 05:47:43 +03:00
int lretry = 20 ; /* up to 2 seconds */
2016-08-10 02:40:02 +03:00
int new_adap_fd = - 1 ;
2015-08-14 05:47:43 +03:00
int rc = 0 ;
atomic_inc ( & cfg - > recovery_threads ) ;
2016-05-03 19:27:34 +03:00
up_read ( & cfg - > ioctl_rwsem ) ;
2015-08-14 05:47:43 +03:00
rc = mutex_lock_interruptible ( mutex ) ;
2016-05-03 19:27:34 +03:00
down_read ( & cfg - > ioctl_rwsem ) ;
2017-08-26 01:18:03 +03:00
if ( rc ) {
locked = false ;
2015-08-14 05:47:43 +03:00
goto out ;
2017-08-26 01:18:03 +03:00
}
2016-05-03 19:27:34 +03:00
rc = check_state ( cfg ) ;
if ( rc ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Failed state rc=%d \n " , __func__ , rc ) ;
2016-05-03 19:27:34 +03:00
rc = - ENODEV ;
goto out ;
}
2015-08-14 05:47:43 +03:00
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: reason=%016llx rctxid=%016llx \n " ,
2015-08-14 05:47:43 +03:00
__func__ , recover - > reason , rctxid ) ;
retry :
/* Ensure that this process is attached to the context */
ctxi = get_context ( cfg , rctxid , lli , CTX_CTRL_ERR_FALLBACK ) ;
if ( unlikely ( ! ctxi ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Bad context ctxid=%llu \n " , __func__ , ctxid ) ;
2015-08-14 05:47:43 +03:00
rc = - EINVAL ;
goto out ;
}
if ( ctxi - > err_recovery_active ) {
retry_recover :
2016-08-10 02:40:02 +03:00
rc = recover_context ( cfg , ctxi , & new_adap_fd ) ;
2015-08-14 05:47:43 +03:00
if ( unlikely ( rc ) ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Recovery failed ctxid=%llu rc=%d \n " ,
2015-08-14 05:47:43 +03:00
__func__ , ctxid , rc ) ;
if ( ( rc = = - ENODEV ) & &
( ( atomic_read ( & cfg - > recovery_threads ) > 1 ) | |
( lretry - - ) ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Going to try again \n " ,
2015-08-14 05:47:43 +03:00
__func__ ) ;
mutex_unlock ( mutex ) ;
msleep ( 100 ) ;
rc = mutex_lock_interruptible ( mutex ) ;
2017-08-26 01:18:03 +03:00
if ( rc ) {
locked = false ;
2015-08-14 05:47:43 +03:00
goto out ;
2017-08-26 01:18:03 +03:00
}
2015-08-14 05:47:43 +03:00
goto retry_recover ;
}
goto out ;
}
ctxi - > err_recovery_active = false ;
2017-01-12 04:19:33 +03:00
flags = DK_CXLFLASH_APP_CLOSE_ADAP_FD |
DK_CXLFLASH_RECOVER_AFU_CONTEXT_RESET ;
if ( afu_is_sq_cmd_mode ( afu ) )
flags | = DK_CXLFLASH_CONTEXT_SQ_CMD_MODE ;
recover - > hdr . return_flags = flags ;
2015-08-14 05:47:43 +03:00
recover - > context_id = ctxi - > ctxid ;
2016-08-10 02:40:02 +03:00
recover - > adap_fd = new_adap_fd ;
2015-08-14 05:47:43 +03:00
recover - > mmio_size = sizeof ( afu - > afu_map - > hosts [ 0 ] . harea ) ;
goto out ;
}
/* Test if in error state */
2017-04-12 22:15:42 +03:00
reg = readq_be ( & hwq - > ctrl_map - > mbox_r ) ;
2015-08-14 05:47:43 +03:00
if ( reg = = - 1 ) {
2015-10-21 23:11:52 +03:00
dev_dbg ( dev , " %s: MMIO fail, wait for recovery. \n " , __func__ ) ;
/*
* Before checking the state , put back the context obtained with
* get_context ( ) as it is no longer needed and sleep for a short
* period of time ( see prolog notes ) .
*/
put_context ( ctxi ) ;
2015-08-14 05:47:43 +03:00
ctxi = NULL ;
ssleep ( 1 ) ;
rc = check_state ( cfg ) ;
if ( unlikely ( rc ) )
goto out ;
goto retry ;
}
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: MMIO working, no recovery required \n " , __func__ ) ;
2015-08-14 05:47:43 +03:00
out :
if ( likely ( ctxi ) )
put_context ( ctxi ) ;
2017-08-26 01:18:03 +03:00
if ( locked )
mutex_unlock ( mutex ) ;
2015-08-14 05:47:43 +03:00
atomic_dec_if_positive ( & cfg - > recovery_threads ) ;
return rc ;
}
/**
* process_sense ( ) - evaluates and processes sense data
* @ sdev : SCSI device associated with LUN .
* @ verify : Verify ioctl data structure .
*
* Return : 0 on success , - errno on failure
*/
static int process_sense ( struct scsi_device * sdev ,
struct dk_cxlflash_verify * verify )
{
2017-01-12 04:19:47 +03:00
struct cxlflash_cfg * cfg = shost_priv ( sdev - > host ) ;
2015-08-14 05:47:43 +03:00
struct device * dev = & cfg - > dev - > dev ;
struct llun_info * lli = sdev - > hostdata ;
struct glun_info * gli = lli - > parent ;
u64 prev_lba = gli - > max_lba ;
struct scsi_sense_hdr sshdr = { 0 } ;
int rc = 0 ;
rc = scsi_normalize_sense ( ( const u8 * ) & verify - > sense_data ,
DK_CXLFLASH_VERIFY_SENSE_LEN , & sshdr ) ;
if ( ! rc ) {
2017-01-12 04:19:47 +03:00
dev_err ( dev , " %s: Failed to normalize sense data \n " , __func__ ) ;
2015-08-14 05:47:43 +03:00
rc = - EINVAL ;
goto out ;
}
switch ( sshdr . sense_key ) {
case NO_SENSE :
case RECOVERED_ERROR :
case NOT_READY :
break ;
case UNIT_ATTENTION :
switch ( sshdr . asc ) {
case 0x29 : /* Power on Reset or Device Reset */
2020-08-24 01:36:59 +03:00
fallthrough ;
2015-08-14 05:47:43 +03:00
case 0x2A : /* Device settings/capacity changed */
rc = read_cap16 ( sdev , lli ) ;
if ( rc ) {
rc = - ENODEV ;
break ;
}
if ( prev_lba ! = gli - > max_lba )
dev_dbg ( dev , " %s: Capacity changed old=%lld "
" new=%lld \n " , __func__ , prev_lba ,
gli - > max_lba ) ;
break ;
case 0x3F : /* Report LUNs changed, Rescan. */
scsi_scan_host ( cfg - > host ) ;
break ;
default :
rc = - EIO ;
break ;
}
break ;
default :
rc = - EIO ;
break ;
}
out :
dev_dbg ( dev , " %s: sense_key %x asc %x ascq %x rc %d \n " , __func__ ,
sshdr . sense_key , sshdr . asc , sshdr . ascq , rc ) ;
return rc ;
}
/**
* cxlflash_disk_verify ( ) - verifies a LUN is the same and handle size changes
* @ sdev : SCSI device associated with LUN .
* @ verify : Verify ioctl data structure .
*
* Return : 0 on success , - errno on failure
*/
static int cxlflash_disk_verify ( struct scsi_device * sdev ,
struct dk_cxlflash_verify * verify )
{
int rc = 0 ;
struct ctx_info * ctxi = NULL ;
2017-01-12 04:19:47 +03:00
struct cxlflash_cfg * cfg = shost_priv ( sdev - > host ) ;
2015-08-14 05:47:43 +03:00
struct device * dev = & cfg - > dev - > dev ;
struct llun_info * lli = sdev - > hostdata ;
struct glun_info * gli = lli - > parent ;
struct sisl_rht_entry * rhte = NULL ;
res_hndl_t rhndl = verify - > rsrc_handle ;
u64 ctxid = DECODE_CTXID ( verify - > context_id ) ,
rctxid = verify - > context_id ;
u64 last_lba = 0 ;
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: ctxid=%llu rhndl=%016llx, hint=%016llx, "
" flags=%016llx \n " , __func__ , ctxid , verify - > rsrc_handle ,
2015-08-14 05:47:43 +03:00
verify - > hint , verify - > hdr . flags ) ;
ctxi = get_context ( cfg , rctxid , lli , 0 ) ;
if ( unlikely ( ! ctxi ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Bad context ctxid=%llu \n " , __func__ , ctxid ) ;
2015-08-14 05:47:43 +03:00
rc = - EINVAL ;
goto out ;
}
rhte = get_rhte ( ctxi , rhndl , lli ) ;
if ( unlikely ( ! rhte ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Bad resource handle rhndl=%d \n " ,
2015-08-14 05:47:43 +03:00
__func__ , rhndl ) ;
rc = - EINVAL ;
goto out ;
}
/*
* Look at the hint / sense to see if it requires us to redrive
* inquiry ( i . e . the Unit attention is due to the WWN changing ) .
*/
if ( verify - > hint & DK_CXLFLASH_VERIFY_HINT_SENSE ) {
2015-10-21 23:14:56 +03:00
/* Can't hold mutex across process_sense/read_cap16,
* since we could have an intervening EEH event .
*/
ctxi - > unavail = true ;
mutex_unlock ( & ctxi - > mutex ) ;
2015-08-14 05:47:43 +03:00
rc = process_sense ( sdev , verify ) ;
if ( unlikely ( rc ) ) {
dev_err ( dev , " %s: Failed to validate sense data (%d) \n " ,
__func__ , rc ) ;
2015-10-21 23:14:56 +03:00
mutex_lock ( & ctxi - > mutex ) ;
ctxi - > unavail = false ;
2015-08-14 05:47:43 +03:00
goto out ;
}
2015-10-21 23:14:56 +03:00
mutex_lock ( & ctxi - > mutex ) ;
ctxi - > unavail = false ;
2015-08-14 05:47:43 +03:00
}
switch ( gli - > mode ) {
case MODE_PHYSICAL :
last_lba = gli - > max_lba ;
break ;
2015-08-14 05:47:53 +03:00
case MODE_VIRTUAL :
/* Cast lxt_cnt to u64 for multiply to be treated as 64bit op */
last_lba = ( ( u64 ) rhte - > lxt_cnt * MC_CHUNK_SIZE * gli - > blk_len ) ;
last_lba / = CXLFLASH_BLOCK_SIZE ;
last_lba - - ;
break ;
2015-08-14 05:47:43 +03:00
default :
WARN ( 1 , " Unsupported LUN mode! " ) ;
}
verify - > last_lba = last_lba ;
out :
if ( likely ( ctxi ) )
put_context ( ctxi ) ;
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: returning rc=%d llba=%llx \n " ,
2015-08-14 05:47:43 +03:00
__func__ , rc , verify - > last_lba ) ;
return rc ;
}
/**
* decode_ioctl ( ) - translates an encoded ioctl to an easily identifiable string
* @ cmd : The ioctl command to decode .
*
* Return : A string identifying the decoded ioctl .
*/
2019-02-07 19:07:20 +03:00
static char * decode_ioctl ( unsigned int cmd )
2015-08-14 05:47:43 +03:00
{
switch ( cmd ) {
case DK_CXLFLASH_ATTACH :
return __stringify_1 ( DK_CXLFLASH_ATTACH ) ;
case DK_CXLFLASH_USER_DIRECT :
return __stringify_1 ( DK_CXLFLASH_USER_DIRECT ) ;
2015-08-14 05:47:53 +03:00
case DK_CXLFLASH_USER_VIRTUAL :
return __stringify_1 ( DK_CXLFLASH_USER_VIRTUAL ) ;
case DK_CXLFLASH_VLUN_RESIZE :
return __stringify_1 ( DK_CXLFLASH_VLUN_RESIZE ) ;
2015-08-14 05:47:43 +03:00
case DK_CXLFLASH_RELEASE :
return __stringify_1 ( DK_CXLFLASH_RELEASE ) ;
case DK_CXLFLASH_DETACH :
return __stringify_1 ( DK_CXLFLASH_DETACH ) ;
case DK_CXLFLASH_VERIFY :
return __stringify_1 ( DK_CXLFLASH_VERIFY ) ;
2015-08-14 05:47:53 +03:00
case DK_CXLFLASH_VLUN_CLONE :
return __stringify_1 ( DK_CXLFLASH_VLUN_CLONE ) ;
2015-08-14 05:47:43 +03:00
case DK_CXLFLASH_RECOVER_AFU :
return __stringify_1 ( DK_CXLFLASH_RECOVER_AFU ) ;
case DK_CXLFLASH_MANAGE_LUN :
return __stringify_1 ( DK_CXLFLASH_MANAGE_LUN ) ;
}
return " UNKNOWN " ;
}
/**
* cxlflash_disk_direct_open ( ) - opens a direct ( physical ) disk
* @ sdev : SCSI device associated with LUN .
* @ arg : UDirect ioctl data structure .
*
* On successful return , the user is informed of the resource handle
* to be used to identify the direct lun and the size ( in blocks ) of
* the direct lun in last LBA format .
*
* Return : 0 on success , - errno on failure
*/
static int cxlflash_disk_direct_open ( struct scsi_device * sdev , void * arg )
{
2017-01-12 04:19:47 +03:00
struct cxlflash_cfg * cfg = shost_priv ( sdev - > host ) ;
2015-08-14 05:47:43 +03:00
struct device * dev = & cfg - > dev - > dev ;
struct afu * afu = cfg - > afu ;
struct llun_info * lli = sdev - > hostdata ;
struct glun_info * gli = lli - > parent ;
2017-06-22 05:14:30 +03:00
struct dk_cxlflash_release rel = { { 0 } , 0 } ;
2015-08-14 05:47:43 +03:00
struct dk_cxlflash_udirect * pphys = ( struct dk_cxlflash_udirect * ) arg ;
u64 ctxid = DECODE_CTXID ( pphys - > context_id ) ,
rctxid = pphys - > context_id ;
u64 lun_size = 0 ;
u64 last_lba = 0 ;
u64 rsrc_handle = - 1 ;
2017-04-12 22:14:05 +03:00
u32 port = CHAN2PORTMASK ( sdev - > channel ) ;
2015-08-14 05:47:43 +03:00
int rc = 0 ;
struct ctx_info * ctxi = NULL ;
struct sisl_rht_entry * rhte = NULL ;
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: ctxid=%llu ls=%llu \n " , __func__ , ctxid , lun_size ) ;
2015-08-14 05:47:43 +03:00
rc = cxlflash_lun_attach ( gli , MODE_PHYSICAL , false ) ;
if ( unlikely ( rc ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Failed attach to LUN (PHYSICAL) \n " , __func__ ) ;
2015-08-14 05:47:43 +03:00
goto out ;
}
ctxi = get_context ( cfg , rctxid , lli , 0 ) ;
if ( unlikely ( ! ctxi ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Bad context ctxid=%llu \n " , __func__ , ctxid ) ;
2015-08-14 05:47:43 +03:00
rc = - EINVAL ;
goto err1 ;
}
rhte = rhte_checkout ( ctxi , lli ) ;
if ( unlikely ( ! rhte ) ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Too many opens ctxid=%lld \n " ,
__func__ , ctxid ) ;
2015-08-14 05:47:43 +03:00
rc = - EMFILE ; /* too many opens */
goto err1 ;
}
rsrc_handle = ( rhte - ctxi - > rht_start ) ;
rht_format1 ( rhte , lli - > lun_id [ sdev - > channel ] , ctxi - > rht_perms , port ) ;
last_lba = gli - > max_lba ;
pphys - > hdr . return_flags = 0 ;
pphys - > last_lba = last_lba ;
pphys - > rsrc_handle = rsrc_handle ;
2017-06-22 05:14:30 +03:00
rc = cxlflash_afu_sync ( afu , ctxid , rsrc_handle , AFU_LW_SYNC ) ;
if ( unlikely ( rc ) ) {
dev_dbg ( dev , " %s: AFU sync failed rc=%d \n " , __func__ , rc ) ;
goto err2 ;
}
2015-08-14 05:47:43 +03:00
out :
if ( likely ( ctxi ) )
put_context ( ctxi ) ;
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: returning handle=%llu rc=%d llba=%llu \n " ,
2015-08-14 05:47:43 +03:00
__func__ , rsrc_handle , rc , last_lba ) ;
return rc ;
2017-06-22 05:14:30 +03:00
err2 :
marshal_udir_to_rele ( pphys , & rel ) ;
_cxlflash_disk_release ( sdev , ctxi , & rel ) ;
goto out ;
2015-08-14 05:47:43 +03:00
err1 :
cxlflash_lun_detach ( gli ) ;
goto out ;
}
/**
* ioctl_common ( ) - common IOCTL handler for driver
* @ sdev : SCSI device associated with LUN .
* @ cmd : IOCTL command .
*
* Handles common fencing operations that are valid for multiple ioctls . Always
* allow through ioctls that are cleanup oriented in nature , even when operating
* in a failed / terminating state .
*
* Return : 0 on success , - errno on failure
*/
2019-02-07 19:07:20 +03:00
static int ioctl_common ( struct scsi_device * sdev , unsigned int cmd )
2015-08-14 05:47:43 +03:00
{
2017-01-12 04:19:47 +03:00
struct cxlflash_cfg * cfg = shost_priv ( sdev - > host ) ;
2015-08-14 05:47:43 +03:00
struct device * dev = & cfg - > dev - > dev ;
struct llun_info * lli = sdev - > hostdata ;
int rc = 0 ;
if ( unlikely ( ! lli ) ) {
dev_dbg ( dev , " %s: Unknown LUN \n " , __func__ ) ;
rc = - EINVAL ;
goto out ;
}
rc = check_state ( cfg ) ;
if ( unlikely ( rc ) & & ( cfg - > state = = STATE_FAILTERM ) ) {
switch ( cmd ) {
2015-08-14 05:47:53 +03:00
case DK_CXLFLASH_VLUN_RESIZE :
2015-08-14 05:47:43 +03:00
case DK_CXLFLASH_RELEASE :
case DK_CXLFLASH_DETACH :
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Command override rc=%d \n " ,
2015-08-14 05:47:43 +03:00
__func__ , rc ) ;
rc = 0 ;
break ;
}
}
out :
return rc ;
}
/**
* cxlflash_ioctl ( ) - IOCTL handler for driver
* @ sdev : SCSI device associated with LUN .
* @ cmd : IOCTL command .
* @ arg : Userspace ioctl data structure .
*
2015-10-21 23:11:52 +03:00
* A read / write semaphore is used to implement a ' drain ' of currently
* running ioctls . The read semaphore is taken at the beginning of each
* ioctl thread and released upon concluding execution . Additionally the
* semaphore should be released and then reacquired in any ioctl execution
* path which will wait for an event to occur that is outside the scope of
* the ioctl ( i . e . an adapter reset ) . To drain the ioctls currently running ,
* a thread simply needs to acquire the write semaphore .
*
2015-08-14 05:47:43 +03:00
* Return : 0 on success , - errno on failure
*/
2019-02-07 19:07:20 +03:00
int cxlflash_ioctl ( struct scsi_device * sdev , unsigned int cmd , void __user * arg )
2015-08-14 05:47:43 +03:00
{
typedef int ( * sioctl ) ( struct scsi_device * , void * ) ;
2017-01-12 04:19:47 +03:00
struct cxlflash_cfg * cfg = shost_priv ( sdev - > host ) ;
2015-08-14 05:47:43 +03:00
struct device * dev = & cfg - > dev - > dev ;
struct afu * afu = cfg - > afu ;
struct dk_cxlflash_hdr * hdr ;
char buf [ sizeof ( union cxlflash_ioctls ) ] ;
size_t size = 0 ;
bool known_ioctl = false ;
int idx ;
int rc = 0 ;
struct Scsi_Host * shost = sdev - > host ;
sioctl do_ioctl = NULL ;
static const struct {
size_t size ;
sioctl ioctl ;
} ioctl_tbl [ ] = { /* NOTE: order matters here */
{ sizeof ( struct dk_cxlflash_attach ) , ( sioctl ) cxlflash_disk_attach } ,
{ sizeof ( struct dk_cxlflash_udirect ) , cxlflash_disk_direct_open } ,
{ sizeof ( struct dk_cxlflash_release ) , ( sioctl ) cxlflash_disk_release } ,
{ sizeof ( struct dk_cxlflash_detach ) , ( sioctl ) cxlflash_disk_detach } ,
{ sizeof ( struct dk_cxlflash_verify ) , ( sioctl ) cxlflash_disk_verify } ,
{ sizeof ( struct dk_cxlflash_recover_afu ) , ( sioctl ) cxlflash_afu_recover } ,
{ sizeof ( struct dk_cxlflash_manage_lun ) , ( sioctl ) cxlflash_manage_lun } ,
2015-08-14 05:47:53 +03:00
{ sizeof ( struct dk_cxlflash_uvirtual ) , cxlflash_disk_virtual_open } ,
{ sizeof ( struct dk_cxlflash_resize ) , ( sioctl ) cxlflash_vlun_resize } ,
{ sizeof ( struct dk_cxlflash_clone ) , ( sioctl ) cxlflash_disk_clone } ,
2015-08-14 05:47:43 +03:00
} ;
2015-10-21 23:11:52 +03:00
/* Hold read semaphore so we can drain if needed */
down_read ( & cfg - > ioctl_rwsem ) ;
2015-08-14 05:47:43 +03:00
/* Restrict command set to physical support only for internal LUN */
if ( afu - > internal_lun )
switch ( cmd ) {
case DK_CXLFLASH_RELEASE :
2015-08-14 05:47:53 +03:00
case DK_CXLFLASH_USER_VIRTUAL :
case DK_CXLFLASH_VLUN_RESIZE :
case DK_CXLFLASH_VLUN_CLONE :
2015-08-14 05:47:43 +03:00
dev_dbg ( dev , " %s: %s not supported for lun_mode=%d \n " ,
__func__ , decode_ioctl ( cmd ) , afu - > internal_lun ) ;
rc = - EINVAL ;
goto cxlflash_ioctl_exit ;
}
switch ( cmd ) {
case DK_CXLFLASH_ATTACH :
case DK_CXLFLASH_USER_DIRECT :
case DK_CXLFLASH_RELEASE :
case DK_CXLFLASH_DETACH :
case DK_CXLFLASH_VERIFY :
case DK_CXLFLASH_RECOVER_AFU :
2015-08-14 05:47:53 +03:00
case DK_CXLFLASH_USER_VIRTUAL :
case DK_CXLFLASH_VLUN_RESIZE :
case DK_CXLFLASH_VLUN_CLONE :
2015-08-14 05:47:43 +03:00
dev_dbg ( dev , " %s: %s (%08X) on dev(%d/%d/%d/%llu) \n " ,
__func__ , decode_ioctl ( cmd ) , cmd , shost - > host_no ,
sdev - > channel , sdev - > id , sdev - > lun ) ;
rc = ioctl_common ( sdev , cmd ) ;
if ( unlikely ( rc ) )
goto cxlflash_ioctl_exit ;
2020-08-24 01:36:59 +03:00
fallthrough ;
2015-08-14 05:47:43 +03:00
case DK_CXLFLASH_MANAGE_LUN :
known_ioctl = true ;
idx = _IOC_NR ( cmd ) - _IOC_NR ( DK_CXLFLASH_ATTACH ) ;
size = ioctl_tbl [ idx ] . size ;
do_ioctl = ioctl_tbl [ idx ] . ioctl ;
if ( likely ( do_ioctl ) )
break ;
2020-08-24 01:36:59 +03:00
fallthrough ;
2015-08-14 05:47:43 +03:00
default :
rc = - EINVAL ;
goto cxlflash_ioctl_exit ;
}
if ( unlikely ( copy_from_user ( & buf , arg , size ) ) ) {
2019-02-07 19:07:20 +03:00
dev_err ( dev , " %s: copy_from_user() fail size=%lu cmd=%u (%s) arg=%p \n " ,
2015-08-14 05:47:43 +03:00
__func__ , size , cmd , decode_ioctl ( cmd ) , arg ) ;
rc = - EFAULT ;
goto cxlflash_ioctl_exit ;
}
hdr = ( struct dk_cxlflash_hdr * ) & buf ;
if ( hdr - > version ! = DK_CXLFLASH_VERSION_0 ) {
dev_dbg ( dev , " %s: Version %u not supported for %s \n " ,
__func__ , hdr - > version , decode_ioctl ( cmd ) ) ;
rc = - EINVAL ;
goto cxlflash_ioctl_exit ;
}
if ( hdr - > rsvd [ 0 ] | | hdr - > rsvd [ 1 ] | | hdr - > rsvd [ 2 ] | | hdr - > return_flags ) {
2017-01-12 04:19:47 +03:00
dev_dbg ( dev , " %s: Reserved/rflags populated \n " , __func__ ) ;
2015-08-14 05:47:43 +03:00
rc = - EINVAL ;
goto cxlflash_ioctl_exit ;
}
rc = do_ioctl ( sdev , ( void * ) & buf ) ;
if ( likely ( ! rc ) )
if ( unlikely ( copy_to_user ( arg , & buf , size ) ) ) {
2019-02-07 19:07:20 +03:00
dev_err ( dev , " %s: copy_to_user() fail size=%lu cmd=%u (%s) arg=%p \n " ,
2015-08-14 05:47:43 +03:00
__func__ , size , cmd , decode_ioctl ( cmd ) , arg ) ;
rc = - EFAULT ;
}
/* fall through to exit */
cxlflash_ioctl_exit :
2015-10-21 23:11:52 +03:00
up_read ( & cfg - > ioctl_rwsem ) ;
2015-08-14 05:47:43 +03:00
if ( unlikely ( rc & & known_ioctl ) )
dev_err ( dev , " %s: ioctl %s (%08X) on dev(%d/%d/%d/%llu) "
" returned rc %d \n " , __func__ ,
decode_ioctl ( cmd ) , cmd , shost - > host_no ,
sdev - > channel , sdev - > id , sdev - > lun , rc ) ;
else
dev_dbg ( dev , " %s: ioctl %s (%08X) on dev(%d/%d/%d/%llu) "
" returned rc %d \n " , __func__ , decode_ioctl ( cmd ) ,
cmd , shost - > host_no , sdev - > channel , sdev - > id ,
sdev - > lun , rc ) ;
return rc ;
}