2008-01-29 16:51:59 +03:00
/*
* Functions related to sysfs handling
*/
# include <linux/kernel.h>
# include <linux/module.h>
# include <linux/bio.h>
# include <linux/blkdev.h>
# include <linux/blktrace_api.h>
# include "blk.h"
struct queue_sysfs_entry {
struct attribute attr ;
ssize_t ( * show ) ( struct request_queue * , char * ) ;
ssize_t ( * store ) ( struct request_queue * , const char * , size_t ) ;
} ;
static ssize_t
2009-07-17 11:26:26 +04:00
queue_var_show ( unsigned long var , char * page )
2008-01-29 16:51:59 +03:00
{
2009-07-17 11:26:26 +04:00
return sprintf ( page , " %lu \n " , var ) ;
2008-01-29 16:51:59 +03:00
}
static ssize_t
queue_var_store ( unsigned long * var , const char * page , size_t count )
{
char * p = ( char * ) page ;
* var = simple_strtoul ( p , & p , 10 ) ;
return count ;
}
static ssize_t queue_requests_show ( struct request_queue * q , char * page )
{
return queue_var_show ( q - > nr_requests , ( page ) ) ;
}
static ssize_t
queue_requests_store ( struct request_queue * q , const char * page , size_t count )
{
struct request_list * rl = & q - > rq ;
unsigned long nr ;
int ret = queue_var_store ( & nr , page , count ) ;
if ( nr < BLKDEV_MIN_RQ )
nr = BLKDEV_MIN_RQ ;
spin_lock_irq ( q - > queue_lock ) ;
q - > nr_requests = nr ;
blk_queue_congestion_threshold ( q ) ;
2009-04-06 16:48:01 +04:00
if ( rl - > count [ BLK_RW_SYNC ] > = queue_congestion_on_threshold ( q ) )
blk_set_queue_congested ( q , BLK_RW_SYNC ) ;
else if ( rl - > count [ BLK_RW_SYNC ] < queue_congestion_off_threshold ( q ) )
blk_clear_queue_congested ( q , BLK_RW_SYNC ) ;
if ( rl - > count [ BLK_RW_ASYNC ] > = queue_congestion_on_threshold ( q ) )
blk_set_queue_congested ( q , BLK_RW_ASYNC ) ;
else if ( rl - > count [ BLK_RW_ASYNC ] < queue_congestion_off_threshold ( q ) )
blk_clear_queue_congested ( q , BLK_RW_ASYNC ) ;
if ( rl - > count [ BLK_RW_SYNC ] > = q - > nr_requests ) {
blk_set_queue_full ( q , BLK_RW_SYNC ) ;
} else if ( rl - > count [ BLK_RW_SYNC ] + 1 < = q - > nr_requests ) {
blk_clear_queue_full ( q , BLK_RW_SYNC ) ;
wake_up ( & rl - > wait [ BLK_RW_SYNC ] ) ;
2008-01-29 16:51:59 +03:00
}
2009-04-06 16:48:01 +04:00
if ( rl - > count [ BLK_RW_ASYNC ] > = q - > nr_requests ) {
blk_set_queue_full ( q , BLK_RW_ASYNC ) ;
} else if ( rl - > count [ BLK_RW_ASYNC ] + 1 < = q - > nr_requests ) {
blk_clear_queue_full ( q , BLK_RW_ASYNC ) ;
wake_up ( & rl - > wait [ BLK_RW_ASYNC ] ) ;
2008-01-29 16:51:59 +03:00
}
spin_unlock_irq ( q - > queue_lock ) ;
return ret ;
}
static ssize_t queue_ra_show ( struct request_queue * q , char * page )
{
2009-07-17 11:26:26 +04:00
unsigned long ra_kb = q - > backing_dev_info . ra_pages < <
( PAGE_CACHE_SHIFT - 10 ) ;
2008-01-29 16:51:59 +03:00
return queue_var_show ( ra_kb , ( page ) ) ;
}
static ssize_t
queue_ra_store ( struct request_queue * q , const char * page , size_t count )
{
unsigned long ra_kb ;
ssize_t ret = queue_var_store ( & ra_kb , page , count ) ;
q - > backing_dev_info . ra_pages = ra_kb > > ( PAGE_CACHE_SHIFT - 10 ) ;
return ret ;
}
static ssize_t queue_max_sectors_show ( struct request_queue * q , char * page )
{
2009-05-23 01:17:50 +04:00
int max_sectors_kb = queue_max_sectors ( q ) > > 1 ;
2008-01-29 16:51:59 +03:00
return queue_var_show ( max_sectors_kb , ( page ) ) ;
}
2009-05-23 01:17:49 +04:00
static ssize_t queue_logical_block_size_show ( struct request_queue * q , char * page )
2008-01-29 21:14:08 +03:00
{
2009-05-23 01:17:49 +04:00
return queue_var_show ( queue_logical_block_size ( q ) , page ) ;
2008-01-29 21:14:08 +03:00
}
2009-05-23 01:17:53 +04:00
static ssize_t queue_physical_block_size_show ( struct request_queue * q , char * page )
{
return queue_var_show ( queue_physical_block_size ( q ) , page ) ;
}
static ssize_t queue_io_min_show ( struct request_queue * q , char * page )
{
return queue_var_show ( queue_io_min ( q ) , page ) ;
}
static ssize_t queue_io_opt_show ( struct request_queue * q , char * page )
{
return queue_var_show ( queue_io_opt ( q ) , page ) ;
2008-01-29 21:14:08 +03:00
}
2008-01-29 16:51:59 +03:00
static ssize_t
queue_max_sectors_store ( struct request_queue * q , const char * page , size_t count )
{
unsigned long max_sectors_kb ,
2009-05-23 01:17:50 +04:00
max_hw_sectors_kb = queue_max_hw_sectors ( q ) > > 1 ,
2008-01-29 16:51:59 +03:00
page_kb = 1 < < ( PAGE_CACHE_SHIFT - 10 ) ;
ssize_t ret = queue_var_store ( & max_sectors_kb , page , count ) ;
if ( max_sectors_kb > max_hw_sectors_kb | | max_sectors_kb < page_kb )
return - EINVAL ;
2008-11-25 11:08:39 +03:00
2008-01-29 16:51:59 +03:00
spin_lock_irq ( q - > queue_lock ) ;
2009-05-23 01:17:50 +04:00
blk_queue_max_sectors ( q , max_sectors_kb < < 1 ) ;
2008-01-29 16:51:59 +03:00
spin_unlock_irq ( q - > queue_lock ) ;
return ret ;
}
static ssize_t queue_max_hw_sectors_show ( struct request_queue * q , char * page )
{
2009-05-23 01:17:50 +04:00
int max_hw_sectors_kb = queue_max_hw_sectors ( q ) > > 1 ;
2008-01-29 16:51:59 +03:00
return queue_var_show ( max_hw_sectors_kb , ( page ) ) ;
}
2009-01-07 14:22:39 +03:00
static ssize_t queue_nonrot_show ( struct request_queue * q , char * page )
{
return queue_var_show ( ! blk_queue_nonrot ( q ) , page ) ;
}
static ssize_t queue_nonrot_store ( struct request_queue * q , const char * page ,
size_t count )
{
unsigned long nm ;
ssize_t ret = queue_var_store ( & nm , page , count ) ;
spin_lock_irq ( q - > queue_lock ) ;
if ( nm )
queue_flag_clear ( QUEUE_FLAG_NONROT , q ) ;
else
queue_flag_set ( QUEUE_FLAG_NONROT , q ) ;
spin_unlock_irq ( q - > queue_lock ) ;
return ret ;
}
2008-04-29 16:44:19 +04:00
static ssize_t queue_nomerges_show ( struct request_queue * q , char * page )
{
return queue_var_show ( blk_queue_nomerges ( q ) , page ) ;
}
static ssize_t queue_nomerges_store ( struct request_queue * q , const char * page ,
size_t count )
{
unsigned long nm ;
ssize_t ret = queue_var_store ( & nm , page , count ) ;
2008-05-07 11:09:39 +04:00
spin_lock_irq ( q - > queue_lock ) ;
2008-04-29 16:44:19 +04:00
if ( nm )
2008-05-07 11:09:39 +04:00
queue_flag_set ( QUEUE_FLAG_NOMERGES , q ) ;
2008-04-29 16:44:19 +04:00
else
2008-05-07 11:09:39 +04:00
queue_flag_clear ( QUEUE_FLAG_NOMERGES , q ) ;
spin_unlock_irq ( q - > queue_lock ) ;
2009-01-07 14:22:39 +03:00
2008-04-29 16:44:19 +04:00
return ret ;
}
2008-09-13 22:26:01 +04:00
static ssize_t queue_rq_affinity_show ( struct request_queue * q , char * page )
{
2009-07-17 11:26:26 +04:00
bool set = test_bit ( QUEUE_FLAG_SAME_COMP , & q - > queue_flags ) ;
2008-09-13 22:26:01 +04:00
2009-07-17 11:26:26 +04:00
return queue_var_show ( set , page ) ;
2008-09-13 22:26:01 +04:00
}
static ssize_t
queue_rq_affinity_store ( struct request_queue * q , const char * page , size_t count )
{
ssize_t ret = - EINVAL ;
# if defined(CONFIG_USE_GENERIC_SMP_HELPERS)
unsigned long val ;
ret = queue_var_store ( & val , page , count ) ;
spin_lock_irq ( q - > queue_lock ) ;
if ( val )
queue_flag_set ( QUEUE_FLAG_SAME_COMP , q ) ;
else
queue_flag_clear ( QUEUE_FLAG_SAME_COMP , q ) ;
spin_unlock_irq ( q - > queue_lock ) ;
# endif
return ret ;
}
2008-01-29 16:51:59 +03:00
2009-01-23 12:54:44 +03:00
static ssize_t queue_iostats_show ( struct request_queue * q , char * page )
{
return queue_var_show ( blk_queue_io_stat ( q ) , page ) ;
}
static ssize_t queue_iostats_store ( struct request_queue * q , const char * page ,
size_t count )
{
unsigned long stats ;
ssize_t ret = queue_var_store ( & stats , page , count ) ;
spin_lock_irq ( q - > queue_lock ) ;
if ( stats )
queue_flag_set ( QUEUE_FLAG_IO_STAT , q ) ;
else
queue_flag_clear ( QUEUE_FLAG_IO_STAT , q ) ;
spin_unlock_irq ( q - > queue_lock ) ;
return ret ;
}
2008-01-29 16:51:59 +03:00
static struct queue_sysfs_entry queue_requests_entry = {
. attr = { . name = " nr_requests " , . mode = S_IRUGO | S_IWUSR } ,
. show = queue_requests_show ,
. store = queue_requests_store ,
} ;
static struct queue_sysfs_entry queue_ra_entry = {
. attr = { . name = " read_ahead_kb " , . mode = S_IRUGO | S_IWUSR } ,
. show = queue_ra_show ,
. store = queue_ra_store ,
} ;
static struct queue_sysfs_entry queue_max_sectors_entry = {
. attr = { . name = " max_sectors_kb " , . mode = S_IRUGO | S_IWUSR } ,
. show = queue_max_sectors_show ,
. store = queue_max_sectors_store ,
} ;
static struct queue_sysfs_entry queue_max_hw_sectors_entry = {
. attr = { . name = " max_hw_sectors_kb " , . mode = S_IRUGO } ,
. show = queue_max_hw_sectors_show ,
} ;
static struct queue_sysfs_entry queue_iosched_entry = {
. attr = { . name = " scheduler " , . mode = S_IRUGO | S_IWUSR } ,
. show = elv_iosched_show ,
. store = elv_iosched_store ,
} ;
2008-01-29 21:14:08 +03:00
static struct queue_sysfs_entry queue_hw_sector_size_entry = {
. attr = { . name = " hw_sector_size " , . mode = S_IRUGO } ,
2009-05-23 01:17:49 +04:00
. show = queue_logical_block_size_show ,
} ;
static struct queue_sysfs_entry queue_logical_block_size_entry = {
. attr = { . name = " logical_block_size " , . mode = S_IRUGO } ,
. show = queue_logical_block_size_show ,
2008-01-29 21:14:08 +03:00
} ;
2009-05-23 01:17:53 +04:00
static struct queue_sysfs_entry queue_physical_block_size_entry = {
. attr = { . name = " physical_block_size " , . mode = S_IRUGO } ,
. show = queue_physical_block_size_show ,
} ;
static struct queue_sysfs_entry queue_io_min_entry = {
. attr = { . name = " minimum_io_size " , . mode = S_IRUGO } ,
. show = queue_io_min_show ,
} ;
static struct queue_sysfs_entry queue_io_opt_entry = {
. attr = { . name = " optimal_io_size " , . mode = S_IRUGO } ,
. show = queue_io_opt_show ,
2008-01-29 21:14:08 +03:00
} ;
2009-01-07 14:22:39 +03:00
static struct queue_sysfs_entry queue_nonrot_entry = {
. attr = { . name = " rotational " , . mode = S_IRUGO | S_IWUSR } ,
. show = queue_nonrot_show ,
. store = queue_nonrot_store ,
} ;
2008-04-29 16:44:19 +04:00
static struct queue_sysfs_entry queue_nomerges_entry = {
. attr = { . name = " nomerges " , . mode = S_IRUGO | S_IWUSR } ,
. show = queue_nomerges_show ,
. store = queue_nomerges_store ,
} ;
2008-09-13 22:26:01 +04:00
static struct queue_sysfs_entry queue_rq_affinity_entry = {
. attr = { . name = " rq_affinity " , . mode = S_IRUGO | S_IWUSR } ,
. show = queue_rq_affinity_show ,
. store = queue_rq_affinity_store ,
} ;
2009-01-23 12:54:44 +03:00
static struct queue_sysfs_entry queue_iostats_entry = {
. attr = { . name = " iostats " , . mode = S_IRUGO | S_IWUSR } ,
. show = queue_iostats_show ,
. store = queue_iostats_store ,
} ;
2008-01-29 16:51:59 +03:00
static struct attribute * default_attrs [ ] = {
& queue_requests_entry . attr ,
& queue_ra_entry . attr ,
& queue_max_hw_sectors_entry . attr ,
& queue_max_sectors_entry . attr ,
& queue_iosched_entry . attr ,
2008-01-29 21:14:08 +03:00
& queue_hw_sector_size_entry . attr ,
2009-05-23 01:17:49 +04:00
& queue_logical_block_size_entry . attr ,
2009-05-23 01:17:53 +04:00
& queue_physical_block_size_entry . attr ,
& queue_io_min_entry . attr ,
& queue_io_opt_entry . attr ,
2009-01-07 14:22:39 +03:00
& queue_nonrot_entry . attr ,
2008-04-29 16:44:19 +04:00
& queue_nomerges_entry . attr ,
2008-09-13 22:26:01 +04:00
& queue_rq_affinity_entry . attr ,
2009-01-23 12:54:44 +03:00
& queue_iostats_entry . attr ,
2008-01-29 16:51:59 +03:00
NULL ,
} ;
# define to_queue(atr) container_of((atr), struct queue_sysfs_entry, attr)
static ssize_t
queue_attr_show ( struct kobject * kobj , struct attribute * attr , char * page )
{
struct queue_sysfs_entry * entry = to_queue ( attr ) ;
struct request_queue * q =
container_of ( kobj , struct request_queue , kobj ) ;
ssize_t res ;
if ( ! entry - > show )
return - EIO ;
mutex_lock ( & q - > sysfs_lock ) ;
if ( test_bit ( QUEUE_FLAG_DEAD , & q - > queue_flags ) ) {
mutex_unlock ( & q - > sysfs_lock ) ;
return - ENOENT ;
}
res = entry - > show ( q , page ) ;
mutex_unlock ( & q - > sysfs_lock ) ;
return res ;
}
static ssize_t
queue_attr_store ( struct kobject * kobj , struct attribute * attr ,
const char * page , size_t length )
{
struct queue_sysfs_entry * entry = to_queue ( attr ) ;
2008-01-31 15:03:55 +03:00
struct request_queue * q ;
2008-01-29 16:51:59 +03:00
ssize_t res ;
if ( ! entry - > store )
return - EIO ;
2008-01-31 15:03:55 +03:00
q = container_of ( kobj , struct request_queue , kobj ) ;
2008-01-29 16:51:59 +03:00
mutex_lock ( & q - > sysfs_lock ) ;
if ( test_bit ( QUEUE_FLAG_DEAD , & q - > queue_flags ) ) {
mutex_unlock ( & q - > sysfs_lock ) ;
return - ENOENT ;
}
res = entry - > store ( q , page , length ) ;
mutex_unlock ( & q - > sysfs_lock ) ;
return res ;
}
/**
* blk_cleanup_queue : - release a & struct request_queue when it is no longer needed
* @ kobj : the kobj belonging of the request queue to be released
*
* Description :
* blk_cleanup_queue is the pair to blk_init_queue ( ) or
* blk_queue_make_request ( ) . It should be called when a request queue is
* being released ; typically when a block device is being de - registered .
* Currently , its primary task it to free all the & struct request
* structures that were allocated to the queue and the queue itself .
*
* Caveat :
* Hopefully the low level driver will have finished any
* outstanding requests first . . .
* */
static void blk_release_queue ( struct kobject * kobj )
{
struct request_queue * q =
container_of ( kobj , struct request_queue , kobj ) ;
struct request_list * rl = & q - > rq ;
blk_sync_queue ( q ) ;
if ( rl - > rq_pool )
mempool_destroy ( rl - > rq_pool ) ;
if ( q - > queue_tags )
__blk_queue_free_tags ( q ) ;
blk_trace_shutdown ( q ) ;
bdi_destroy ( & q - > backing_dev_info ) ;
kmem_cache_free ( blk_requestq_cachep , q ) ;
}
static struct sysfs_ops queue_sysfs_ops = {
. show = queue_attr_show ,
. store = queue_attr_store ,
} ;
struct kobj_type blk_queue_ktype = {
. sysfs_ops = & queue_sysfs_ops ,
. default_attrs = default_attrs ,
. release = blk_release_queue ,
} ;
int blk_register_queue ( struct gendisk * disk )
{
int ret ;
2009-04-14 10:00:05 +04:00
struct device * dev = disk_to_dev ( disk ) ;
2008-01-29 16:51:59 +03:00
struct request_queue * q = disk - > queue ;
2008-04-21 11:51:06 +04:00
if ( WARN_ON ( ! q ) )
2008-01-29 16:51:59 +03:00
return - ENXIO ;
2009-04-14 10:00:05 +04:00
ret = blk_trace_init_sysfs ( dev ) ;
if ( ret )
return ret ;
2009-06-11 21:52:27 +04:00
ret = kobject_add ( & q - > kobj , kobject_get ( & dev - > kobj ) , " %s " , " queue " ) ;
2008-01-29 16:51:59 +03:00
if ( ret < 0 )
return ret ;
kobject_uevent ( & q - > kobj , KOBJ_ADD ) ;
2009-05-23 01:17:52 +04:00
if ( ! q - > request_fn )
return 0 ;
2008-01-29 16:51:59 +03:00
ret = elv_register_queue ( q ) ;
if ( ret ) {
kobject_uevent ( & q - > kobj , KOBJ_REMOVE ) ;
kobject_del ( & q - > kobj ) ;
return ret ;
}
return 0 ;
}
void blk_unregister_queue ( struct gendisk * disk )
{
struct request_queue * q = disk - > queue ;
2008-04-21 11:51:06 +04:00
if ( WARN_ON ( ! q ) )
return ;
if ( q - > request_fn ) {
2008-01-29 16:51:59 +03:00
elv_unregister_queue ( q ) ;
kobject_uevent ( & q - > kobj , KOBJ_REMOVE ) ;
kobject_del ( & q - > kobj ) ;
2008-08-25 14:56:05 +04:00
kobject_put ( & disk_to_dev ( disk ) - > kobj ) ;
2008-01-29 16:51:59 +03:00
}
}