2009-04-18 05:33:26 +04:00
/*
* Copyright 2008 Cisco Systems , Inc . All rights reserved .
* Copyright 2007 Nuova Systems , Inc . All rights reserved .
*
* This program is free software ; you may redistribute it and / or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation ; version 2 of the License .
*
* THE SOFTWARE IS PROVIDED " AS IS " , WITHOUT WARRANTY OF ANY KIND ,
* EXPRESS OR IMPLIED , INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
* MERCHANTABILITY , FITNESS FOR A PARTICULAR PURPOSE AND
* NONINFRINGEMENT . IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
* BE LIABLE FOR ANY CLAIM , DAMAGES OR OTHER LIABILITY , WHETHER IN AN
* ACTION OF CONTRACT , TORT OR OTHERWISE , ARISING FROM , OUT OF OR IN
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE .
*/
# ifndef _VNIC_WQ_H_
# define _VNIC_WQ_H_
# include <linux/pci.h>
# include "vnic_dev.h"
# include "vnic_cq.h"
/*
* These defines avoid symbol clash between fnic and enic ( Cisco 10 G Eth
* Driver ) when both are built with CONFIG options = y
*/
# define vnic_wq_desc_avail fnic_wq_desc_avail
# define vnic_wq_desc_used fnic_wq_desc_used
# define vnic_wq_next_desc fni_cwq_next_desc
# define vnic_wq_post fnic_wq_post
# define vnic_wq_service fnic_wq_service
# define vnic_wq_free fnic_wq_free
# define vnic_wq_alloc fnic_wq_alloc
2019-01-19 01:51:42 +03:00
# define vnic_wq_devcmd2_alloc fnic_wq_devcmd2_alloc
# define vnic_wq_init_start fnic_wq_init_start
2009-04-18 05:33:26 +04:00
# define vnic_wq_init fnic_wq_init
# define vnic_wq_error_status fnic_wq_error_status
# define vnic_wq_enable fnic_wq_enable
# define vnic_wq_disable fnic_wq_disable
# define vnic_wq_clean fnic_wq_clean
/* Work queue control */
struct vnic_wq_ctrl {
u64 ring_base ; /* 0x00 */
u32 ring_size ; /* 0x08 */
u32 pad0 ;
u32 posted_index ; /* 0x10 */
u32 pad1 ;
u32 cq_index ; /* 0x18 */
u32 pad2 ;
u32 enable ; /* 0x20 */
u32 pad3 ;
u32 running ; /* 0x28 */
u32 pad4 ;
u32 fetch_index ; /* 0x30 */
u32 pad5 ;
u32 dca_value ; /* 0x38 */
u32 pad6 ;
u32 error_interrupt_enable ; /* 0x40 */
u32 pad7 ;
u32 error_interrupt_offset ; /* 0x48 */
u32 pad8 ;
u32 error_status ; /* 0x50 */
u32 pad9 ;
} ;
struct vnic_wq_buf {
struct vnic_wq_buf * next ;
dma_addr_t dma_addr ;
void * os_buf ;
unsigned int len ;
unsigned int index ;
int sop ;
void * desc ;
} ;
/* Break the vnic_wq_buf allocations into blocks of 64 entries */
# define VNIC_WQ_BUF_BLK_ENTRIES 64
# define VNIC_WQ_BUF_BLK_SZ \
( VNIC_WQ_BUF_BLK_ENTRIES * sizeof ( struct vnic_wq_buf ) )
# define VNIC_WQ_BUF_BLKS_NEEDED(entries) \
DIV_ROUND_UP ( entries , VNIC_WQ_BUF_BLK_ENTRIES )
# define VNIC_WQ_BUF_BLKS_MAX VNIC_WQ_BUF_BLKS_NEEDED(4096)
struct vnic_wq {
unsigned int index ;
struct vnic_dev * vdev ;
struct vnic_wq_ctrl __iomem * ctrl ; /* memory-mapped */
struct vnic_dev_ring ring ;
struct vnic_wq_buf * bufs [ VNIC_WQ_BUF_BLKS_MAX ] ;
struct vnic_wq_buf * to_use ;
struct vnic_wq_buf * to_clean ;
unsigned int pkts_outstanding ;
} ;
static inline unsigned int vnic_wq_desc_avail ( struct vnic_wq * wq )
{
/* how many does SW own? */
return wq - > ring . desc_avail ;
}
static inline unsigned int vnic_wq_desc_used ( struct vnic_wq * wq )
{
/* how many does HW own? */
return wq - > ring . desc_count - wq - > ring . desc_avail - 1 ;
}
static inline void * vnic_wq_next_desc ( struct vnic_wq * wq )
{
return wq - > to_use - > desc ;
}
static inline void vnic_wq_post ( struct vnic_wq * wq ,
void * os_buf , dma_addr_t dma_addr ,
unsigned int len , int sop , int eop )
{
struct vnic_wq_buf * buf = wq - > to_use ;
buf - > sop = sop ;
buf - > os_buf = eop ? os_buf : NULL ;
buf - > dma_addr = dma_addr ;
buf - > len = len ;
buf = buf - > next ;
if ( eop ) {
/* Adding write memory barrier prevents compiler and/or CPU
* reordering , thus avoiding descriptor posting before
* descriptor is initialized . Otherwise , hardware can read
* stale descriptor fields .
*/
wmb ( ) ;
iowrite32 ( buf - > index , & wq - > ctrl - > posted_index ) ;
}
wq - > to_use = buf ;
wq - > ring . desc_avail - - ;
}
static inline void vnic_wq_service ( struct vnic_wq * wq ,
struct cq_desc * cq_desc , u16 completed_index ,
void ( * buf_service ) ( struct vnic_wq * wq ,
struct cq_desc * cq_desc , struct vnic_wq_buf * buf , void * opaque ) ,
void * opaque )
{
struct vnic_wq_buf * buf ;
buf = wq - > to_clean ;
while ( 1 ) {
( * buf_service ) ( wq , cq_desc , buf , opaque ) ;
wq - > ring . desc_avail + + ;
wq - > to_clean = buf - > next ;
if ( buf - > index = = completed_index )
break ;
buf = wq - > to_clean ;
}
}
void vnic_wq_free ( struct vnic_wq * wq ) ;
int vnic_wq_alloc ( struct vnic_dev * vdev , struct vnic_wq * wq , unsigned int index ,
unsigned int desc_count , unsigned int desc_size ) ;
2019-01-19 01:51:42 +03:00
int vnic_wq_devcmd2_alloc ( struct vnic_dev * vdev , struct vnic_wq * wq ,
unsigned int desc_count , unsigned int desc_size ) ;
void vnic_wq_init_start ( struct vnic_wq * wq , unsigned int cq_index ,
unsigned int fetch_index , unsigned int posted_index ,
unsigned int error_interrupt_enable ,
unsigned int error_interrupt_offset ) ;
2009-04-18 05:33:26 +04:00
void vnic_wq_init ( struct vnic_wq * wq , unsigned int cq_index ,
unsigned int error_interrupt_enable ,
unsigned int error_interrupt_offset ) ;
unsigned int vnic_wq_error_status ( struct vnic_wq * wq ) ;
void vnic_wq_enable ( struct vnic_wq * wq ) ;
int vnic_wq_disable ( struct vnic_wq * wq ) ;
void vnic_wq_clean ( struct vnic_wq * wq ,
void ( * buf_clean ) ( struct vnic_wq * wq , struct vnic_wq_buf * buf ) ) ;
# endif /* _VNIC_WQ_H_ */