Merge branch 'ucc_geth-improvements'
Rasmus Villemoes says: ==================== ucc_geth improvements This is a resend of some improvements to the ucc_geth driver that was previously sent together with bug fixes, which have by now been applied. v2: rebase to net/master; address minor style issues; don't introduce a use-after-free in patch "don't statically allocate eight ucc_geth_info". ==================== Link: https://lore.kernel.org/r/20210119150802.19997-1-rasmus.villemoes@prevas.dk Signed-off-by: Jakub Kicinski <kuba@kernel.org>
This commit is contained in:
commit
961629bd32
@ -70,9 +70,32 @@ static struct {
|
||||
module_param_named(debug, debug.msg_enable, int, 0);
|
||||
MODULE_PARM_DESC(debug, "Debug verbosity level (0=none, ..., 0xffff=all)");
|
||||
|
||||
static struct ucc_geth_info ugeth_primary_info = {
|
||||
static int ucc_geth_thread_count(enum ucc_geth_num_of_threads idx)
|
||||
{
|
||||
static const u8 count[] = {
|
||||
[UCC_GETH_NUM_OF_THREADS_1] = 1,
|
||||
[UCC_GETH_NUM_OF_THREADS_2] = 2,
|
||||
[UCC_GETH_NUM_OF_THREADS_4] = 4,
|
||||
[UCC_GETH_NUM_OF_THREADS_6] = 6,
|
||||
[UCC_GETH_NUM_OF_THREADS_8] = 8,
|
||||
};
|
||||
if (idx >= ARRAY_SIZE(count))
|
||||
return 0;
|
||||
return count[idx];
|
||||
}
|
||||
|
||||
static inline int ucc_geth_tx_queues(const struct ucc_geth_info *info)
|
||||
{
|
||||
return 1;
|
||||
}
|
||||
|
||||
static inline int ucc_geth_rx_queues(const struct ucc_geth_info *info)
|
||||
{
|
||||
return 1;
|
||||
}
|
||||
|
||||
static const struct ucc_geth_info ugeth_primary_info = {
|
||||
.uf_info = {
|
||||
.bd_mem_part = MEM_PART_SYSTEM,
|
||||
.rtsm = UCC_FAST_SEND_IDLES_BETWEEN_FRAMES,
|
||||
.max_rx_buf_length = 1536,
|
||||
/* adjusted at startup if max-speed 1000 */
|
||||
@ -90,8 +113,6 @@ static struct ucc_geth_info ugeth_primary_info = {
|
||||
.tcrc = UCC_FAST_16_BIT_CRC,
|
||||
.synl = UCC_FAST_SYNC_LEN_NOT_USED,
|
||||
},
|
||||
.numQueuesTx = 1,
|
||||
.numQueuesRx = 1,
|
||||
.extendedFilteringChainPointer = ((uint32_t) NULL),
|
||||
.typeorlen = 3072 /*1536 */ ,
|
||||
.nonBackToBackIfgPart1 = 0x40,
|
||||
@ -157,8 +178,6 @@ static struct ucc_geth_info ugeth_primary_info = {
|
||||
.riscRx = QE_RISC_ALLOCATION_RISC1_AND_RISC2,
|
||||
};
|
||||
|
||||
static struct ucc_geth_info ugeth_info[8];
|
||||
|
||||
#ifdef DEBUG
|
||||
static void mem_disp(u8 *addr, int size)
|
||||
{
|
||||
@ -558,7 +577,7 @@ static void dump_bds(struct ucc_geth_private *ugeth)
|
||||
int i;
|
||||
int length;
|
||||
|
||||
for (i = 0; i < ugeth->ug_info->numQueuesTx; i++) {
|
||||
for (i = 0; i < ucc_geth_tx_queues(ugeth->ug_info); i++) {
|
||||
if (ugeth->p_tx_bd_ring[i]) {
|
||||
length =
|
||||
(ugeth->ug_info->bdRingLenTx[i] *
|
||||
@ -567,7 +586,7 @@ static void dump_bds(struct ucc_geth_private *ugeth)
|
||||
mem_disp(ugeth->p_tx_bd_ring[i], length);
|
||||
}
|
||||
}
|
||||
for (i = 0; i < ugeth->ug_info->numQueuesRx; i++) {
|
||||
for (i = 0; i < ucc_geth_rx_queues(ugeth->ug_info); i++) {
|
||||
if (ugeth->p_rx_bd_ring[i]) {
|
||||
length =
|
||||
(ugeth->ug_info->bdRingLenRx[i] *
|
||||
@ -671,32 +690,12 @@ static void dump_regs(struct ucc_geth_private *ugeth)
|
||||
in_be32(&ugeth->ug_regs->scam));
|
||||
|
||||
if (ugeth->p_thread_data_tx) {
|
||||
int numThreadsTxNumerical;
|
||||
switch (ugeth->ug_info->numThreadsTx) {
|
||||
case UCC_GETH_NUM_OF_THREADS_1:
|
||||
numThreadsTxNumerical = 1;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_2:
|
||||
numThreadsTxNumerical = 2;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_4:
|
||||
numThreadsTxNumerical = 4;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_6:
|
||||
numThreadsTxNumerical = 6;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_8:
|
||||
numThreadsTxNumerical = 8;
|
||||
break;
|
||||
default:
|
||||
numThreadsTxNumerical = 0;
|
||||
break;
|
||||
}
|
||||
int count = ucc_geth_thread_count(ugeth->ug_info->numThreadsTx);
|
||||
|
||||
pr_info("Thread data TXs:\n");
|
||||
pr_info("Base address: 0x%08x\n",
|
||||
(u32)ugeth->p_thread_data_tx);
|
||||
for (i = 0; i < numThreadsTxNumerical; i++) {
|
||||
for (i = 0; i < count; i++) {
|
||||
pr_info("Thread data TX[%d]:\n", i);
|
||||
pr_info("Base address: 0x%08x\n",
|
||||
(u32)&ugeth->p_thread_data_tx[i]);
|
||||
@ -705,32 +704,12 @@ static void dump_regs(struct ucc_geth_private *ugeth)
|
||||
}
|
||||
}
|
||||
if (ugeth->p_thread_data_rx) {
|
||||
int numThreadsRxNumerical;
|
||||
switch (ugeth->ug_info->numThreadsRx) {
|
||||
case UCC_GETH_NUM_OF_THREADS_1:
|
||||
numThreadsRxNumerical = 1;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_2:
|
||||
numThreadsRxNumerical = 2;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_4:
|
||||
numThreadsRxNumerical = 4;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_6:
|
||||
numThreadsRxNumerical = 6;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_8:
|
||||
numThreadsRxNumerical = 8;
|
||||
break;
|
||||
default:
|
||||
numThreadsRxNumerical = 0;
|
||||
break;
|
||||
}
|
||||
int count = ucc_geth_thread_count(ugeth->ug_info->numThreadsRx);
|
||||
|
||||
pr_info("Thread data RX:\n");
|
||||
pr_info("Base address: 0x%08x\n",
|
||||
(u32)ugeth->p_thread_data_rx);
|
||||
for (i = 0; i < numThreadsRxNumerical; i++) {
|
||||
for (i = 0; i < count; i++) {
|
||||
pr_info("Thread data RX[%d]:\n", i);
|
||||
pr_info("Base address: 0x%08x\n",
|
||||
(u32)&ugeth->p_thread_data_rx[i]);
|
||||
@ -905,7 +884,7 @@ static void dump_regs(struct ucc_geth_private *ugeth)
|
||||
if (ugeth->p_send_q_mem_reg) {
|
||||
pr_info("Send Q memory registers:\n");
|
||||
pr_info("Base address: 0x%08x\n", (u32)ugeth->p_send_q_mem_reg);
|
||||
for (i = 0; i < ugeth->ug_info->numQueuesTx; i++) {
|
||||
for (i = 0; i < ucc_geth_tx_queues(ugeth->ug_info); i++) {
|
||||
pr_info("SQQD[%d]:\n", i);
|
||||
pr_info("Base address: 0x%08x\n",
|
||||
(u32)&ugeth->p_send_q_mem_reg->sqqd[i]);
|
||||
@ -937,7 +916,7 @@ static void dump_regs(struct ucc_geth_private *ugeth)
|
||||
pr_info("RX IRQ coalescing tables:\n");
|
||||
pr_info("Base address: 0x%08x\n",
|
||||
(u32)ugeth->p_rx_irq_coalescing_tbl);
|
||||
for (i = 0; i < ugeth->ug_info->numQueuesRx; i++) {
|
||||
for (i = 0; i < ucc_geth_rx_queues(ugeth->ug_info); i++) {
|
||||
pr_info("RX IRQ coalescing table entry[%d]:\n", i);
|
||||
pr_info("Base address: 0x%08x\n",
|
||||
(u32)&ugeth->p_rx_irq_coalescing_tbl->
|
||||
@ -959,7 +938,7 @@ static void dump_regs(struct ucc_geth_private *ugeth)
|
||||
if (ugeth->p_rx_bd_qs_tbl) {
|
||||
pr_info("RX BD QS tables:\n");
|
||||
pr_info("Base address: 0x%08x\n", (u32)ugeth->p_rx_bd_qs_tbl);
|
||||
for (i = 0; i < ugeth->ug_info->numQueuesRx; i++) {
|
||||
for (i = 0; i < ucc_geth_rx_queues(ugeth->ug_info); i++) {
|
||||
pr_info("RX BD QS table[%d]:\n", i);
|
||||
pr_info("Base address: 0x%08x\n",
|
||||
(u32)&ugeth->p_rx_bd_qs_tbl[i]);
|
||||
@ -1835,7 +1814,7 @@ static void ucc_geth_free_rx(struct ucc_geth_private *ugeth)
|
||||
ug_info = ugeth->ug_info;
|
||||
uf_info = &ug_info->uf_info;
|
||||
|
||||
for (i = 0; i < ugeth->ug_info->numQueuesRx; i++) {
|
||||
for (i = 0; i < ucc_geth_rx_queues(ugeth->ug_info); i++) {
|
||||
if (ugeth->p_rx_bd_ring[i]) {
|
||||
/* Return existing data buffers in ring */
|
||||
bd = ugeth->p_rx_bd_ring[i];
|
||||
@ -1856,12 +1835,7 @@ static void ucc_geth_free_rx(struct ucc_geth_private *ugeth)
|
||||
|
||||
kfree(ugeth->rx_skbuff[i]);
|
||||
|
||||
if (ugeth->ug_info->uf_info.bd_mem_part ==
|
||||
MEM_PART_SYSTEM)
|
||||
kfree((void *)ugeth->rx_bd_ring_offset[i]);
|
||||
else if (ugeth->ug_info->uf_info.bd_mem_part ==
|
||||
MEM_PART_MURAM)
|
||||
qe_muram_free(ugeth->rx_bd_ring_offset[i]);
|
||||
kfree(ugeth->p_rx_bd_ring[i]);
|
||||
ugeth->p_rx_bd_ring[i] = NULL;
|
||||
}
|
||||
}
|
||||
@ -1880,7 +1854,7 @@ static void ucc_geth_free_tx(struct ucc_geth_private *ugeth)
|
||||
ug_info = ugeth->ug_info;
|
||||
uf_info = &ug_info->uf_info;
|
||||
|
||||
for (i = 0; i < ugeth->ug_info->numQueuesTx; i++) {
|
||||
for (i = 0; i < ucc_geth_tx_queues(ugeth->ug_info); i++) {
|
||||
bd = ugeth->p_tx_bd_ring[i];
|
||||
if (!bd)
|
||||
continue;
|
||||
@ -1898,15 +1872,8 @@ static void ucc_geth_free_tx(struct ucc_geth_private *ugeth)
|
||||
|
||||
kfree(ugeth->tx_skbuff[i]);
|
||||
|
||||
if (ugeth->p_tx_bd_ring[i]) {
|
||||
if (ugeth->ug_info->uf_info.bd_mem_part ==
|
||||
MEM_PART_SYSTEM)
|
||||
kfree((void *)ugeth->tx_bd_ring_offset[i]);
|
||||
else if (ugeth->ug_info->uf_info.bd_mem_part ==
|
||||
MEM_PART_MURAM)
|
||||
qe_muram_free(ugeth->tx_bd_ring_offset[i]);
|
||||
ugeth->p_tx_bd_ring[i] = NULL;
|
||||
}
|
||||
kfree(ugeth->p_tx_bd_ring[i]);
|
||||
ugeth->p_tx_bd_ring[i] = NULL;
|
||||
}
|
||||
|
||||
}
|
||||
@ -1921,50 +1888,39 @@ static void ucc_geth_memclean(struct ucc_geth_private *ugeth)
|
||||
ugeth->uccf = NULL;
|
||||
}
|
||||
|
||||
if (ugeth->p_thread_data_tx) {
|
||||
qe_muram_free(ugeth->thread_dat_tx_offset);
|
||||
ugeth->p_thread_data_tx = NULL;
|
||||
}
|
||||
if (ugeth->p_thread_data_rx) {
|
||||
qe_muram_free(ugeth->thread_dat_rx_offset);
|
||||
ugeth->p_thread_data_rx = NULL;
|
||||
}
|
||||
if (ugeth->p_exf_glbl_param) {
|
||||
qe_muram_free(ugeth->exf_glbl_param_offset);
|
||||
ugeth->p_exf_glbl_param = NULL;
|
||||
}
|
||||
if (ugeth->p_rx_glbl_pram) {
|
||||
qe_muram_free(ugeth->rx_glbl_pram_offset);
|
||||
ugeth->p_rx_glbl_pram = NULL;
|
||||
}
|
||||
if (ugeth->p_tx_glbl_pram) {
|
||||
qe_muram_free(ugeth->tx_glbl_pram_offset);
|
||||
ugeth->p_tx_glbl_pram = NULL;
|
||||
}
|
||||
if (ugeth->p_send_q_mem_reg) {
|
||||
qe_muram_free(ugeth->send_q_mem_reg_offset);
|
||||
ugeth->p_send_q_mem_reg = NULL;
|
||||
}
|
||||
if (ugeth->p_scheduler) {
|
||||
qe_muram_free(ugeth->scheduler_offset);
|
||||
ugeth->p_scheduler = NULL;
|
||||
}
|
||||
if (ugeth->p_tx_fw_statistics_pram) {
|
||||
qe_muram_free(ugeth->tx_fw_statistics_pram_offset);
|
||||
ugeth->p_tx_fw_statistics_pram = NULL;
|
||||
}
|
||||
if (ugeth->p_rx_fw_statistics_pram) {
|
||||
qe_muram_free(ugeth->rx_fw_statistics_pram_offset);
|
||||
ugeth->p_rx_fw_statistics_pram = NULL;
|
||||
}
|
||||
if (ugeth->p_rx_irq_coalescing_tbl) {
|
||||
qe_muram_free(ugeth->rx_irq_coalescing_tbl_offset);
|
||||
ugeth->p_rx_irq_coalescing_tbl = NULL;
|
||||
}
|
||||
if (ugeth->p_rx_bd_qs_tbl) {
|
||||
qe_muram_free(ugeth->rx_bd_qs_tbl_offset);
|
||||
ugeth->p_rx_bd_qs_tbl = NULL;
|
||||
}
|
||||
qe_muram_free_addr(ugeth->p_thread_data_tx);
|
||||
ugeth->p_thread_data_tx = NULL;
|
||||
|
||||
qe_muram_free_addr(ugeth->p_thread_data_rx);
|
||||
ugeth->p_thread_data_rx = NULL;
|
||||
|
||||
qe_muram_free_addr(ugeth->p_exf_glbl_param);
|
||||
ugeth->p_exf_glbl_param = NULL;
|
||||
|
||||
qe_muram_free_addr(ugeth->p_rx_glbl_pram);
|
||||
ugeth->p_rx_glbl_pram = NULL;
|
||||
|
||||
qe_muram_free_addr(ugeth->p_tx_glbl_pram);
|
||||
ugeth->p_tx_glbl_pram = NULL;
|
||||
|
||||
qe_muram_free_addr(ugeth->p_send_q_mem_reg);
|
||||
ugeth->p_send_q_mem_reg = NULL;
|
||||
|
||||
qe_muram_free_addr(ugeth->p_scheduler);
|
||||
ugeth->p_scheduler = NULL;
|
||||
|
||||
qe_muram_free_addr(ugeth->p_tx_fw_statistics_pram);
|
||||
ugeth->p_tx_fw_statistics_pram = NULL;
|
||||
|
||||
qe_muram_free_addr(ugeth->p_rx_fw_statistics_pram);
|
||||
ugeth->p_rx_fw_statistics_pram = NULL;
|
||||
|
||||
qe_muram_free_addr(ugeth->p_rx_irq_coalescing_tbl);
|
||||
ugeth->p_rx_irq_coalescing_tbl = NULL;
|
||||
|
||||
qe_muram_free_addr(ugeth->p_rx_bd_qs_tbl);
|
||||
ugeth->p_rx_bd_qs_tbl = NULL;
|
||||
|
||||
if (ugeth->p_init_enet_param_shadow) {
|
||||
return_init_enet_entries(ugeth,
|
||||
&(ugeth->p_init_enet_param_shadow->
|
||||
@ -2073,15 +2029,8 @@ static int ucc_struct_init(struct ucc_geth_private *ugeth)
|
||||
ug_info = ugeth->ug_info;
|
||||
uf_info = &ug_info->uf_info;
|
||||
|
||||
if (!((uf_info->bd_mem_part == MEM_PART_SYSTEM) ||
|
||||
(uf_info->bd_mem_part == MEM_PART_MURAM))) {
|
||||
if (netif_msg_probe(ugeth))
|
||||
pr_err("Bad memory partition value\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
/* Rx BD lengths */
|
||||
for (i = 0; i < ug_info->numQueuesRx; i++) {
|
||||
for (i = 0; i < ucc_geth_rx_queues(ug_info); i++) {
|
||||
if ((ug_info->bdRingLenRx[i] < UCC_GETH_RX_BD_RING_SIZE_MIN) ||
|
||||
(ug_info->bdRingLenRx[i] %
|
||||
UCC_GETH_RX_BD_RING_SIZE_ALIGNMENT)) {
|
||||
@ -2092,7 +2041,7 @@ static int ucc_struct_init(struct ucc_geth_private *ugeth)
|
||||
}
|
||||
|
||||
/* Tx BD lengths */
|
||||
for (i = 0; i < ug_info->numQueuesTx; i++) {
|
||||
for (i = 0; i < ucc_geth_tx_queues(ug_info); i++) {
|
||||
if (ug_info->bdRingLenTx[i] < UCC_GETH_TX_BD_RING_SIZE_MIN) {
|
||||
if (netif_msg_probe(ugeth))
|
||||
pr_err("Tx BD ring length must be no smaller than 2\n");
|
||||
@ -2109,14 +2058,14 @@ static int ucc_struct_init(struct ucc_geth_private *ugeth)
|
||||
}
|
||||
|
||||
/* num Tx queues */
|
||||
if (ug_info->numQueuesTx > NUM_TX_QUEUES) {
|
||||
if (ucc_geth_tx_queues(ug_info) > NUM_TX_QUEUES) {
|
||||
if (netif_msg_probe(ugeth))
|
||||
pr_err("number of tx queues too large\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
/* num Rx queues */
|
||||
if (ug_info->numQueuesRx > NUM_RX_QUEUES) {
|
||||
if (ucc_geth_rx_queues(ug_info) > NUM_RX_QUEUES) {
|
||||
if (netif_msg_probe(ugeth))
|
||||
pr_err("number of rx queues too large\n");
|
||||
return -EINVAL;
|
||||
@ -2124,7 +2073,7 @@ static int ucc_struct_init(struct ucc_geth_private *ugeth)
|
||||
|
||||
/* l2qt */
|
||||
for (i = 0; i < UCC_GETH_VLAN_PRIORITY_MAX; i++) {
|
||||
if (ug_info->l2qt[i] >= ug_info->numQueuesRx) {
|
||||
if (ug_info->l2qt[i] >= ucc_geth_rx_queues(ug_info)) {
|
||||
if (netif_msg_probe(ugeth))
|
||||
pr_err("VLAN priority table entry must not be larger than number of Rx queues\n");
|
||||
return -EINVAL;
|
||||
@ -2133,7 +2082,7 @@ static int ucc_struct_init(struct ucc_geth_private *ugeth)
|
||||
|
||||
/* l3qt */
|
||||
for (i = 0; i < UCC_GETH_IP_PRIORITY_MAX; i++) {
|
||||
if (ug_info->l3qt[i] >= ug_info->numQueuesRx) {
|
||||
if (ug_info->l3qt[i] >= ucc_geth_rx_queues(ug_info)) {
|
||||
if (netif_msg_probe(ugeth))
|
||||
pr_err("IP priority table entry must not be larger than number of Rx queues\n");
|
||||
return -EINVAL;
|
||||
@ -2156,10 +2105,10 @@ static int ucc_struct_init(struct ucc_geth_private *ugeth)
|
||||
|
||||
/* Generate uccm_mask for receive */
|
||||
uf_info->uccm_mask = ug_info->eventRegMask & UCCE_OTHER;/* Errors */
|
||||
for (i = 0; i < ug_info->numQueuesRx; i++)
|
||||
for (i = 0; i < ucc_geth_rx_queues(ug_info); i++)
|
||||
uf_info->uccm_mask |= (UCC_GETH_UCCE_RXF0 << i);
|
||||
|
||||
for (i = 0; i < ug_info->numQueuesTx; i++)
|
||||
for (i = 0; i < ucc_geth_tx_queues(ug_info); i++)
|
||||
uf_info->uccm_mask |= (UCC_GETH_UCCE_TXB0 << i);
|
||||
/* Initialize the general fast UCC block. */
|
||||
if (ucc_fast_init(uf_info, &ugeth->uccf)) {
|
||||
@ -2198,53 +2147,32 @@ static int ucc_geth_alloc_tx(struct ucc_geth_private *ugeth)
|
||||
uf_info = &ug_info->uf_info;
|
||||
|
||||
/* Allocate Tx bds */
|
||||
for (j = 0; j < ug_info->numQueuesTx; j++) {
|
||||
/* Allocate in multiple of
|
||||
UCC_GETH_TX_BD_RING_SIZE_MEMORY_ALIGNMENT,
|
||||
according to spec */
|
||||
length = ((ug_info->bdRingLenTx[j] * sizeof(struct qe_bd))
|
||||
/ UCC_GETH_TX_BD_RING_SIZE_MEMORY_ALIGNMENT)
|
||||
* UCC_GETH_TX_BD_RING_SIZE_MEMORY_ALIGNMENT;
|
||||
if ((ug_info->bdRingLenTx[j] * sizeof(struct qe_bd)) %
|
||||
UCC_GETH_TX_BD_RING_SIZE_MEMORY_ALIGNMENT)
|
||||
length += UCC_GETH_TX_BD_RING_SIZE_MEMORY_ALIGNMENT;
|
||||
if (uf_info->bd_mem_part == MEM_PART_SYSTEM) {
|
||||
u32 align = 4;
|
||||
if (UCC_GETH_TX_BD_RING_ALIGNMENT > 4)
|
||||
align = UCC_GETH_TX_BD_RING_ALIGNMENT;
|
||||
ugeth->tx_bd_ring_offset[j] =
|
||||
(u32) kmalloc((u32) (length + align), GFP_KERNEL);
|
||||
for (j = 0; j < ucc_geth_tx_queues(ug_info); j++) {
|
||||
u32 align = max(UCC_GETH_TX_BD_RING_ALIGNMENT,
|
||||
UCC_GETH_TX_BD_RING_SIZE_MEMORY_ALIGNMENT);
|
||||
u32 alloc;
|
||||
|
||||
length = ug_info->bdRingLenTx[j] * sizeof(struct qe_bd);
|
||||
alloc = round_up(length, align);
|
||||
alloc = roundup_pow_of_two(alloc);
|
||||
|
||||
ugeth->p_tx_bd_ring[j] = kmalloc(alloc, GFP_KERNEL);
|
||||
|
||||
if (ugeth->tx_bd_ring_offset[j] != 0)
|
||||
ugeth->p_tx_bd_ring[j] =
|
||||
(u8 __iomem *)((ugeth->tx_bd_ring_offset[j] +
|
||||
align) & ~(align - 1));
|
||||
} else if (uf_info->bd_mem_part == MEM_PART_MURAM) {
|
||||
ugeth->tx_bd_ring_offset[j] =
|
||||
qe_muram_alloc(length,
|
||||
UCC_GETH_TX_BD_RING_ALIGNMENT);
|
||||
if (!IS_ERR_VALUE(ugeth->tx_bd_ring_offset[j]))
|
||||
ugeth->p_tx_bd_ring[j] =
|
||||
(u8 __iomem *) qe_muram_addr(ugeth->
|
||||
tx_bd_ring_offset[j]);
|
||||
}
|
||||
if (!ugeth->p_tx_bd_ring[j]) {
|
||||
if (netif_msg_ifup(ugeth))
|
||||
pr_err("Can not allocate memory for Tx bd rings\n");
|
||||
return -ENOMEM;
|
||||
}
|
||||
/* Zero unused end of bd ring, according to spec */
|
||||
memset_io((void __iomem *)(ugeth->p_tx_bd_ring[j] +
|
||||
ug_info->bdRingLenTx[j] * sizeof(struct qe_bd)), 0,
|
||||
length - ug_info->bdRingLenTx[j] * sizeof(struct qe_bd));
|
||||
memset(ugeth->p_tx_bd_ring[j] + length, 0, alloc - length);
|
||||
}
|
||||
|
||||
/* Init Tx bds */
|
||||
for (j = 0; j < ug_info->numQueuesTx; j++) {
|
||||
for (j = 0; j < ucc_geth_tx_queues(ug_info); j++) {
|
||||
/* Setup the skbuff rings */
|
||||
ugeth->tx_skbuff[j] =
|
||||
kmalloc_array(ugeth->ug_info->bdRingLenTx[j],
|
||||
sizeof(struct sk_buff *), GFP_KERNEL);
|
||||
kcalloc(ugeth->ug_info->bdRingLenTx[j],
|
||||
sizeof(struct sk_buff *), GFP_KERNEL);
|
||||
|
||||
if (ugeth->tx_skbuff[j] == NULL) {
|
||||
if (netif_msg_ifup(ugeth))
|
||||
@ -2252,9 +2180,6 @@ static int ucc_geth_alloc_tx(struct ucc_geth_private *ugeth)
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
for (i = 0; i < ugeth->ug_info->bdRingLenTx[j]; i++)
|
||||
ugeth->tx_skbuff[j][i] = NULL;
|
||||
|
||||
ugeth->skb_curtx[j] = ugeth->skb_dirtytx[j] = 0;
|
||||
bd = ugeth->confBd[j] = ugeth->txBd[j] = ugeth->p_tx_bd_ring[j];
|
||||
for (i = 0; i < ug_info->bdRingLenTx[j]; i++) {
|
||||
@ -2284,27 +2209,15 @@ static int ucc_geth_alloc_rx(struct ucc_geth_private *ugeth)
|
||||
uf_info = &ug_info->uf_info;
|
||||
|
||||
/* Allocate Rx bds */
|
||||
for (j = 0; j < ug_info->numQueuesRx; j++) {
|
||||
for (j = 0; j < ucc_geth_rx_queues(ug_info); j++) {
|
||||
u32 align = UCC_GETH_RX_BD_RING_ALIGNMENT;
|
||||
u32 alloc;
|
||||
|
||||
length = ug_info->bdRingLenRx[j] * sizeof(struct qe_bd);
|
||||
if (uf_info->bd_mem_part == MEM_PART_SYSTEM) {
|
||||
u32 align = 4;
|
||||
if (UCC_GETH_RX_BD_RING_ALIGNMENT > 4)
|
||||
align = UCC_GETH_RX_BD_RING_ALIGNMENT;
|
||||
ugeth->rx_bd_ring_offset[j] =
|
||||
(u32) kmalloc((u32) (length + align), GFP_KERNEL);
|
||||
if (ugeth->rx_bd_ring_offset[j] != 0)
|
||||
ugeth->p_rx_bd_ring[j] =
|
||||
(u8 __iomem *)((ugeth->rx_bd_ring_offset[j] +
|
||||
align) & ~(align - 1));
|
||||
} else if (uf_info->bd_mem_part == MEM_PART_MURAM) {
|
||||
ugeth->rx_bd_ring_offset[j] =
|
||||
qe_muram_alloc(length,
|
||||
UCC_GETH_RX_BD_RING_ALIGNMENT);
|
||||
if (!IS_ERR_VALUE(ugeth->rx_bd_ring_offset[j]))
|
||||
ugeth->p_rx_bd_ring[j] =
|
||||
(u8 __iomem *) qe_muram_addr(ugeth->
|
||||
rx_bd_ring_offset[j]);
|
||||
}
|
||||
alloc = round_up(length, align);
|
||||
alloc = roundup_pow_of_two(alloc);
|
||||
|
||||
ugeth->p_rx_bd_ring[j] = kmalloc(alloc, GFP_KERNEL);
|
||||
if (!ugeth->p_rx_bd_ring[j]) {
|
||||
if (netif_msg_ifup(ugeth))
|
||||
pr_err("Can not allocate memory for Rx bd rings\n");
|
||||
@ -2313,11 +2226,11 @@ static int ucc_geth_alloc_rx(struct ucc_geth_private *ugeth)
|
||||
}
|
||||
|
||||
/* Init Rx bds */
|
||||
for (j = 0; j < ug_info->numQueuesRx; j++) {
|
||||
for (j = 0; j < ucc_geth_rx_queues(ug_info); j++) {
|
||||
/* Setup the skbuff rings */
|
||||
ugeth->rx_skbuff[j] =
|
||||
kmalloc_array(ugeth->ug_info->bdRingLenRx[j],
|
||||
sizeof(struct sk_buff *), GFP_KERNEL);
|
||||
kcalloc(ugeth->ug_info->bdRingLenRx[j],
|
||||
sizeof(struct sk_buff *), GFP_KERNEL);
|
||||
|
||||
if (ugeth->rx_skbuff[j] == NULL) {
|
||||
if (netif_msg_ifup(ugeth))
|
||||
@ -2325,9 +2238,6 @@ static int ucc_geth_alloc_rx(struct ucc_geth_private *ugeth)
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
for (i = 0; i < ugeth->ug_info->bdRingLenRx[j]; i++)
|
||||
ugeth->rx_skbuff[j][i] = NULL;
|
||||
|
||||
ugeth->skb_currx[j] = 0;
|
||||
bd = ugeth->rxBd[j] = ugeth->p_rx_bd_ring[j];
|
||||
for (i = 0; i < ug_info->bdRingLenRx[j]; i++) {
|
||||
@ -2359,10 +2269,10 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
u32 init_enet_pram_offset, cecr_subblock, command;
|
||||
u32 ifstat, i, j, size, l2qt, l3qt;
|
||||
u16 temoder = UCC_GETH_TEMODER_INIT;
|
||||
u16 test;
|
||||
u8 function_code = 0;
|
||||
u8 __iomem *endOfRing;
|
||||
u8 numThreadsRxNumerical, numThreadsTxNumerical;
|
||||
s32 rx_glbl_pram_offset, tx_glbl_pram_offset;
|
||||
|
||||
ugeth_vdbg("%s: IN", __func__);
|
||||
uccf = ugeth->uccf;
|
||||
@ -2371,45 +2281,15 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
uf_regs = uccf->uf_regs;
|
||||
ug_regs = ugeth->ug_regs;
|
||||
|
||||
switch (ug_info->numThreadsRx) {
|
||||
case UCC_GETH_NUM_OF_THREADS_1:
|
||||
numThreadsRxNumerical = 1;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_2:
|
||||
numThreadsRxNumerical = 2;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_4:
|
||||
numThreadsRxNumerical = 4;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_6:
|
||||
numThreadsRxNumerical = 6;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_8:
|
||||
numThreadsRxNumerical = 8;
|
||||
break;
|
||||
default:
|
||||
numThreadsRxNumerical = ucc_geth_thread_count(ug_info->numThreadsRx);
|
||||
if (!numThreadsRxNumerical) {
|
||||
if (netif_msg_ifup(ugeth))
|
||||
pr_err("Bad number of Rx threads value\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
switch (ug_info->numThreadsTx) {
|
||||
case UCC_GETH_NUM_OF_THREADS_1:
|
||||
numThreadsTxNumerical = 1;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_2:
|
||||
numThreadsTxNumerical = 2;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_4:
|
||||
numThreadsTxNumerical = 4;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_6:
|
||||
numThreadsTxNumerical = 6;
|
||||
break;
|
||||
case UCC_GETH_NUM_OF_THREADS_8:
|
||||
numThreadsTxNumerical = 8;
|
||||
break;
|
||||
default:
|
||||
numThreadsTxNumerical = ucc_geth_thread_count(ug_info->numThreadsTx);
|
||||
if (!numThreadsTxNumerical) {
|
||||
if (netif_msg_ifup(ugeth))
|
||||
pr_err("Bad number of Tx threads value\n");
|
||||
return -EINVAL;
|
||||
@ -2507,20 +2387,15 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
*/
|
||||
/* Tx global PRAM */
|
||||
/* Allocate global tx parameter RAM page */
|
||||
ugeth->tx_glbl_pram_offset =
|
||||
tx_glbl_pram_offset =
|
||||
qe_muram_alloc(sizeof(struct ucc_geth_tx_global_pram),
|
||||
UCC_GETH_TX_GLOBAL_PRAM_ALIGNMENT);
|
||||
if (IS_ERR_VALUE(ugeth->tx_glbl_pram_offset)) {
|
||||
if (tx_glbl_pram_offset < 0) {
|
||||
if (netif_msg_ifup(ugeth))
|
||||
pr_err("Can not allocate DPRAM memory for p_tx_glbl_pram\n");
|
||||
return -ENOMEM;
|
||||
}
|
||||
ugeth->p_tx_glbl_pram =
|
||||
(struct ucc_geth_tx_global_pram __iomem *) qe_muram_addr(ugeth->
|
||||
tx_glbl_pram_offset);
|
||||
/* Zero out p_tx_glbl_pram */
|
||||
memset_io((void __iomem *)ugeth->p_tx_glbl_pram, 0, sizeof(struct ucc_geth_tx_global_pram));
|
||||
|
||||
ugeth->p_tx_glbl_pram = qe_muram_addr(tx_glbl_pram_offset);
|
||||
/* Fill global PRAM */
|
||||
|
||||
/* TQPTR */
|
||||
@ -2554,7 +2429,7 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
/* SQPTR */
|
||||
/* Size varies with number of Tx queues */
|
||||
ugeth->send_q_mem_reg_offset =
|
||||
qe_muram_alloc(ug_info->numQueuesTx *
|
||||
qe_muram_alloc(ucc_geth_tx_queues(ug_info) *
|
||||
sizeof(struct ucc_geth_send_queue_qd),
|
||||
UCC_GETH_SEND_QUEUE_QUEUE_DESCRIPTOR_ALIGNMENT);
|
||||
if (IS_ERR_VALUE(ugeth->send_q_mem_reg_offset)) {
|
||||
@ -2570,29 +2445,20 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
|
||||
/* Setup the table */
|
||||
/* Assume BD rings are already established */
|
||||
for (i = 0; i < ug_info->numQueuesTx; i++) {
|
||||
for (i = 0; i < ucc_geth_tx_queues(ug_info); i++) {
|
||||
endOfRing =
|
||||
ugeth->p_tx_bd_ring[i] + (ug_info->bdRingLenTx[i] -
|
||||
1) * sizeof(struct qe_bd);
|
||||
if (ugeth->ug_info->uf_info.bd_mem_part == MEM_PART_SYSTEM) {
|
||||
out_be32(&ugeth->p_send_q_mem_reg->sqqd[i].bd_ring_base,
|
||||
(u32) virt_to_phys(ugeth->p_tx_bd_ring[i]));
|
||||
out_be32(&ugeth->p_send_q_mem_reg->sqqd[i].
|
||||
last_bd_completed_address,
|
||||
(u32) virt_to_phys(endOfRing));
|
||||
} else if (ugeth->ug_info->uf_info.bd_mem_part ==
|
||||
MEM_PART_MURAM) {
|
||||
out_be32(&ugeth->p_send_q_mem_reg->sqqd[i].bd_ring_base,
|
||||
(u32)qe_muram_dma(ugeth->p_tx_bd_ring[i]));
|
||||
out_be32(&ugeth->p_send_q_mem_reg->sqqd[i].
|
||||
last_bd_completed_address,
|
||||
(u32)qe_muram_dma(endOfRing));
|
||||
}
|
||||
out_be32(&ugeth->p_send_q_mem_reg->sqqd[i].bd_ring_base,
|
||||
(u32) virt_to_phys(ugeth->p_tx_bd_ring[i]));
|
||||
out_be32(&ugeth->p_send_q_mem_reg->sqqd[i].
|
||||
last_bd_completed_address,
|
||||
(u32) virt_to_phys(endOfRing));
|
||||
}
|
||||
|
||||
/* schedulerbasepointer */
|
||||
|
||||
if (ug_info->numQueuesTx > 1) {
|
||||
if (ucc_geth_tx_queues(ug_info) > 1) {
|
||||
/* scheduler exists only if more than 1 tx queue */
|
||||
ugeth->scheduler_offset =
|
||||
qe_muram_alloc(sizeof(struct ucc_geth_scheduler),
|
||||
@ -2608,8 +2474,6 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
scheduler_offset);
|
||||
out_be32(&ugeth->p_tx_glbl_pram->schedulerbasepointer,
|
||||
ugeth->scheduler_offset);
|
||||
/* Zero out p_scheduler */
|
||||
memset_io((void __iomem *)ugeth->p_scheduler, 0, sizeof(struct ucc_geth_scheduler));
|
||||
|
||||
/* Set values in scheduler */
|
||||
out_be32(&ugeth->p_scheduler->mblinterval,
|
||||
@ -2652,23 +2516,18 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
ugeth->p_tx_fw_statistics_pram =
|
||||
(struct ucc_geth_tx_firmware_statistics_pram __iomem *)
|
||||
qe_muram_addr(ugeth->tx_fw_statistics_pram_offset);
|
||||
/* Zero out p_tx_fw_statistics_pram */
|
||||
memset_io((void __iomem *)ugeth->p_tx_fw_statistics_pram,
|
||||
0, sizeof(struct ucc_geth_tx_firmware_statistics_pram));
|
||||
}
|
||||
|
||||
/* temoder */
|
||||
/* Already has speed set */
|
||||
|
||||
if (ug_info->numQueuesTx > 1)
|
||||
if (ucc_geth_tx_queues(ug_info) > 1)
|
||||
temoder |= TEMODER_SCHEDULER_ENABLE;
|
||||
if (ug_info->ipCheckSumGenerate)
|
||||
temoder |= TEMODER_IP_CHECKSUM_GENERATE;
|
||||
temoder |= ((ug_info->numQueuesTx - 1) << TEMODER_NUM_OF_QUEUES_SHIFT);
|
||||
temoder |= ((ucc_geth_tx_queues(ug_info) - 1) << TEMODER_NUM_OF_QUEUES_SHIFT);
|
||||
out_be16(&ugeth->p_tx_glbl_pram->temoder, temoder);
|
||||
|
||||
test = in_be16(&ugeth->p_tx_glbl_pram->temoder);
|
||||
|
||||
/* Function code register value to be used later */
|
||||
function_code = UCC_BMR_BO_BE | UCC_BMR_GBL;
|
||||
/* Required for QE */
|
||||
@ -2678,20 +2537,15 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
|
||||
/* Rx global PRAM */
|
||||
/* Allocate global rx parameter RAM page */
|
||||
ugeth->rx_glbl_pram_offset =
|
||||
rx_glbl_pram_offset =
|
||||
qe_muram_alloc(sizeof(struct ucc_geth_rx_global_pram),
|
||||
UCC_GETH_RX_GLOBAL_PRAM_ALIGNMENT);
|
||||
if (IS_ERR_VALUE(ugeth->rx_glbl_pram_offset)) {
|
||||
if (rx_glbl_pram_offset < 0) {
|
||||
if (netif_msg_ifup(ugeth))
|
||||
pr_err("Can not allocate DPRAM memory for p_rx_glbl_pram\n");
|
||||
return -ENOMEM;
|
||||
}
|
||||
ugeth->p_rx_glbl_pram =
|
||||
(struct ucc_geth_rx_global_pram __iomem *) qe_muram_addr(ugeth->
|
||||
rx_glbl_pram_offset);
|
||||
/* Zero out p_rx_glbl_pram */
|
||||
memset_io((void __iomem *)ugeth->p_rx_glbl_pram, 0, sizeof(struct ucc_geth_rx_global_pram));
|
||||
|
||||
ugeth->p_rx_glbl_pram = qe_muram_addr(rx_glbl_pram_offset);
|
||||
/* Fill global PRAM */
|
||||
|
||||
/* RQPTR */
|
||||
@ -2729,16 +2583,13 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
ugeth->p_rx_fw_statistics_pram =
|
||||
(struct ucc_geth_rx_firmware_statistics_pram __iomem *)
|
||||
qe_muram_addr(ugeth->rx_fw_statistics_pram_offset);
|
||||
/* Zero out p_rx_fw_statistics_pram */
|
||||
memset_io((void __iomem *)ugeth->p_rx_fw_statistics_pram, 0,
|
||||
sizeof(struct ucc_geth_rx_firmware_statistics_pram));
|
||||
}
|
||||
|
||||
/* intCoalescingPtr */
|
||||
|
||||
/* Size varies with number of Rx queues */
|
||||
ugeth->rx_irq_coalescing_tbl_offset =
|
||||
qe_muram_alloc(ug_info->numQueuesRx *
|
||||
qe_muram_alloc(ucc_geth_rx_queues(ug_info) *
|
||||
sizeof(struct ucc_geth_rx_interrupt_coalescing_entry)
|
||||
+ 4, UCC_GETH_RX_INTERRUPT_COALESCING_ALIGNMENT);
|
||||
if (IS_ERR_VALUE(ugeth->rx_irq_coalescing_tbl_offset)) {
|
||||
@ -2754,7 +2605,7 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
ugeth->rx_irq_coalescing_tbl_offset);
|
||||
|
||||
/* Fill interrupt coalescing table */
|
||||
for (i = 0; i < ug_info->numQueuesRx; i++) {
|
||||
for (i = 0; i < ucc_geth_rx_queues(ug_info); i++) {
|
||||
out_be32(&ugeth->p_rx_irq_coalescing_tbl->coalescingentry[i].
|
||||
interruptcoalescingmaxvalue,
|
||||
ug_info->interruptcoalescingmaxvalue[i]);
|
||||
@ -2803,7 +2654,7 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
/* RBDQPTR */
|
||||
/* Size varies with number of Rx queues */
|
||||
ugeth->rx_bd_qs_tbl_offset =
|
||||
qe_muram_alloc(ug_info->numQueuesRx *
|
||||
qe_muram_alloc(ucc_geth_rx_queues(ug_info) *
|
||||
(sizeof(struct ucc_geth_rx_bd_queues_entry) +
|
||||
sizeof(struct ucc_geth_rx_prefetched_bds)),
|
||||
UCC_GETH_RX_BD_QUEUES_ALIGNMENT);
|
||||
@ -2817,23 +2668,12 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
(struct ucc_geth_rx_bd_queues_entry __iomem *) qe_muram_addr(ugeth->
|
||||
rx_bd_qs_tbl_offset);
|
||||
out_be32(&ugeth->p_rx_glbl_pram->rbdqptr, ugeth->rx_bd_qs_tbl_offset);
|
||||
/* Zero out p_rx_bd_qs_tbl */
|
||||
memset_io((void __iomem *)ugeth->p_rx_bd_qs_tbl,
|
||||
0,
|
||||
ug_info->numQueuesRx * (sizeof(struct ucc_geth_rx_bd_queues_entry) +
|
||||
sizeof(struct ucc_geth_rx_prefetched_bds)));
|
||||
|
||||
/* Setup the table */
|
||||
/* Assume BD rings are already established */
|
||||
for (i = 0; i < ug_info->numQueuesRx; i++) {
|
||||
if (ugeth->ug_info->uf_info.bd_mem_part == MEM_PART_SYSTEM) {
|
||||
out_be32(&ugeth->p_rx_bd_qs_tbl[i].externalbdbaseptr,
|
||||
(u32) virt_to_phys(ugeth->p_rx_bd_ring[i]));
|
||||
} else if (ugeth->ug_info->uf_info.bd_mem_part ==
|
||||
MEM_PART_MURAM) {
|
||||
out_be32(&ugeth->p_rx_bd_qs_tbl[i].externalbdbaseptr,
|
||||
(u32)qe_muram_dma(ugeth->p_rx_bd_ring[i]));
|
||||
}
|
||||
for (i = 0; i < ucc_geth_rx_queues(ug_info); i++) {
|
||||
out_be32(&ugeth->p_rx_bd_qs_tbl[i].externalbdbaseptr,
|
||||
(u32) virt_to_phys(ugeth->p_rx_bd_ring[i]));
|
||||
/* rest of fields handled by QE */
|
||||
}
|
||||
|
||||
@ -2854,7 +2694,7 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
ug_info->
|
||||
vlanOperationNonTagged << REMODER_VLAN_OPERATION_NON_TAGGED_SHIFT;
|
||||
remoder |= ug_info->rxQoSMode << REMODER_RX_QOS_MODE_SHIFT;
|
||||
remoder |= ((ug_info->numQueuesRx - 1) << REMODER_NUM_OF_QUEUES_SHIFT);
|
||||
remoder |= ((ucc_geth_rx_queues(ug_info) - 1) << REMODER_NUM_OF_QUEUES_SHIFT);
|
||||
if (ug_info->ipCheckSumCheck)
|
||||
remoder |= REMODER_IP_CHECKSUM_CHECK;
|
||||
if (ug_info->ipAddressAlignment)
|
||||
@ -2937,14 +2777,11 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
* allocated resources can be released when the channel is freed.
|
||||
*/
|
||||
if (!(ugeth->p_init_enet_param_shadow =
|
||||
kmalloc(sizeof(struct ucc_geth_init_pram), GFP_KERNEL))) {
|
||||
kzalloc(sizeof(struct ucc_geth_init_pram), GFP_KERNEL))) {
|
||||
if (netif_msg_ifup(ugeth))
|
||||
pr_err("Can not allocate memory for p_UccInitEnetParamShadows\n");
|
||||
return -ENOMEM;
|
||||
}
|
||||
/* Zero out *p_init_enet_param_shadow */
|
||||
memset((char *)ugeth->p_init_enet_param_shadow,
|
||||
0, sizeof(struct ucc_geth_init_pram));
|
||||
|
||||
/* Fill shadow InitEnet command parameter structure */
|
||||
|
||||
@ -2964,7 +2801,7 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
((u32) ug_info->numThreadsTx) << ENET_INIT_PARAM_TGF_SHIFT;
|
||||
|
||||
ugeth->p_init_enet_param_shadow->rgftgfrxglobal |=
|
||||
ugeth->rx_glbl_pram_offset | ug_info->riscRx;
|
||||
rx_glbl_pram_offset | ug_info->riscRx;
|
||||
if ((ug_info->largestexternallookupkeysize !=
|
||||
QE_FLTR_LARGEST_EXTERNAL_TABLE_LOOKUP_KEY_SIZE_NONE) &&
|
||||
(ug_info->largestexternallookupkeysize !=
|
||||
@ -3002,7 +2839,7 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
}
|
||||
|
||||
ugeth->p_init_enet_param_shadow->txglobal =
|
||||
ugeth->tx_glbl_pram_offset | ug_info->riscTx;
|
||||
tx_glbl_pram_offset | ug_info->riscTx;
|
||||
if ((ret_val =
|
||||
fill_init_enet_entries(ugeth,
|
||||
&(ugeth->p_init_enet_param_shadow->
|
||||
@ -3016,7 +2853,7 @@ static int ucc_geth_startup(struct ucc_geth_private *ugeth)
|
||||
}
|
||||
|
||||
/* Load Rx bds with buffers */
|
||||
for (i = 0; i < ug_info->numQueuesRx; i++) {
|
||||
for (i = 0; i < ucc_geth_rx_queues(ug_info); i++) {
|
||||
if ((ret_val = rx_bd_buffer_set(ugeth, (u8) i)) != 0) {
|
||||
if (netif_msg_ifup(ugeth))
|
||||
pr_err("Can not fill Rx bds with buffers\n");
|
||||
@ -3287,12 +3124,12 @@ static int ucc_geth_poll(struct napi_struct *napi, int budget)
|
||||
|
||||
/* Tx event processing */
|
||||
spin_lock(&ugeth->lock);
|
||||
for (i = 0; i < ug_info->numQueuesTx; i++)
|
||||
for (i = 0; i < ucc_geth_tx_queues(ug_info); i++)
|
||||
ucc_geth_tx(ugeth->ndev, i);
|
||||
spin_unlock(&ugeth->lock);
|
||||
|
||||
howmany = 0;
|
||||
for (i = 0; i < ug_info->numQueuesRx; i++)
|
||||
for (i = 0; i < ucc_geth_rx_queues(ug_info); i++)
|
||||
howmany += ucc_geth_rx(ugeth, i, budget - howmany);
|
||||
|
||||
if (howmany < budget) {
|
||||
@ -3685,6 +3522,36 @@ static const struct net_device_ops ucc_geth_netdev_ops = {
|
||||
#endif
|
||||
};
|
||||
|
||||
static int ucc_geth_parse_clock(struct device_node *np, const char *which,
|
||||
enum qe_clock *out)
|
||||
{
|
||||
const char *sprop;
|
||||
char buf[24];
|
||||
|
||||
snprintf(buf, sizeof(buf), "%s-clock-name", which);
|
||||
sprop = of_get_property(np, buf, NULL);
|
||||
if (sprop) {
|
||||
*out = qe_clock_source(sprop);
|
||||
} else {
|
||||
u32 val;
|
||||
|
||||
snprintf(buf, sizeof(buf), "%s-clock", which);
|
||||
if (of_property_read_u32(np, buf, &val)) {
|
||||
/* If both *-clock-name and *-clock are missing,
|
||||
* we want to tell people to use *-clock-name.
|
||||
*/
|
||||
pr_err("missing %s-clock-name property\n", buf);
|
||||
return -EINVAL;
|
||||
}
|
||||
*out = val;
|
||||
}
|
||||
if (*out < QE_CLK_NONE || *out > QE_CLK24) {
|
||||
pr_err("invalid %s property\n", buf);
|
||||
return -EINVAL;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int ucc_geth_probe(struct platform_device* ofdev)
|
||||
{
|
||||
struct device *device = &ofdev->dev;
|
||||
@ -3695,7 +3562,6 @@ static int ucc_geth_probe(struct platform_device* ofdev)
|
||||
struct resource res;
|
||||
int err, ucc_num, max_speed = 0;
|
||||
const unsigned int *prop;
|
||||
const char *sprop;
|
||||
const void *mac_addr;
|
||||
phy_interface_t phy_interface;
|
||||
static const int enet_to_speed[] = {
|
||||
@ -3725,62 +3591,23 @@ static int ucc_geth_probe(struct platform_device* ofdev)
|
||||
if ((ucc_num < 0) || (ucc_num > 7))
|
||||
return -ENODEV;
|
||||
|
||||
ug_info = &ugeth_info[ucc_num];
|
||||
if (ug_info == NULL) {
|
||||
if (netif_msg_probe(&debug))
|
||||
pr_err("[%d] Missing additional data!\n", ucc_num);
|
||||
return -ENODEV;
|
||||
}
|
||||
ug_info = kmalloc(sizeof(*ug_info), GFP_KERNEL);
|
||||
if (ug_info == NULL)
|
||||
return -ENOMEM;
|
||||
memcpy(ug_info, &ugeth_primary_info, sizeof(*ug_info));
|
||||
|
||||
ug_info->uf_info.ucc_num = ucc_num;
|
||||
|
||||
sprop = of_get_property(np, "rx-clock-name", NULL);
|
||||
if (sprop) {
|
||||
ug_info->uf_info.rx_clock = qe_clock_source(sprop);
|
||||
if ((ug_info->uf_info.rx_clock < QE_CLK_NONE) ||
|
||||
(ug_info->uf_info.rx_clock > QE_CLK24)) {
|
||||
pr_err("invalid rx-clock-name property\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
} else {
|
||||
prop = of_get_property(np, "rx-clock", NULL);
|
||||
if (!prop) {
|
||||
/* If both rx-clock-name and rx-clock are missing,
|
||||
we want to tell people to use rx-clock-name. */
|
||||
pr_err("missing rx-clock-name property\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
if ((*prop < QE_CLK_NONE) || (*prop > QE_CLK24)) {
|
||||
pr_err("invalid rx-clock property\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
ug_info->uf_info.rx_clock = *prop;
|
||||
}
|
||||
|
||||
sprop = of_get_property(np, "tx-clock-name", NULL);
|
||||
if (sprop) {
|
||||
ug_info->uf_info.tx_clock = qe_clock_source(sprop);
|
||||
if ((ug_info->uf_info.tx_clock < QE_CLK_NONE) ||
|
||||
(ug_info->uf_info.tx_clock > QE_CLK24)) {
|
||||
pr_err("invalid tx-clock-name property\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
} else {
|
||||
prop = of_get_property(np, "tx-clock", NULL);
|
||||
if (!prop) {
|
||||
pr_err("missing tx-clock-name property\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
if ((*prop < QE_CLK_NONE) || (*prop > QE_CLK24)) {
|
||||
pr_err("invalid tx-clock property\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
ug_info->uf_info.tx_clock = *prop;
|
||||
}
|
||||
err = ucc_geth_parse_clock(np, "rx", &ug_info->uf_info.rx_clock);
|
||||
if (err)
|
||||
goto err_free_info;
|
||||
err = ucc_geth_parse_clock(np, "tx", &ug_info->uf_info.tx_clock);
|
||||
if (err)
|
||||
goto err_free_info;
|
||||
|
||||
err = of_address_to_resource(np, 0, &res);
|
||||
if (err)
|
||||
return -EINVAL;
|
||||
goto err_free_info;
|
||||
|
||||
ug_info->uf_info.regs = res.start;
|
||||
ug_info->uf_info.irq = irq_of_parse_and_map(np, 0);
|
||||
@ -3793,7 +3620,7 @@ static int ucc_geth_probe(struct platform_device* ofdev)
|
||||
*/
|
||||
err = of_phy_register_fixed_link(np);
|
||||
if (err)
|
||||
return err;
|
||||
goto err_free_info;
|
||||
ug_info->phy_node = of_node_get(np);
|
||||
}
|
||||
|
||||
@ -3924,6 +3751,8 @@ err_deregister_fixed_link:
|
||||
of_phy_deregister_fixed_link(np);
|
||||
of_node_put(ug_info->tbi_node);
|
||||
of_node_put(ug_info->phy_node);
|
||||
err_free_info:
|
||||
kfree(ug_info);
|
||||
|
||||
return err;
|
||||
}
|
||||
@ -3940,6 +3769,7 @@ static int ucc_geth_remove(struct platform_device* ofdev)
|
||||
of_phy_deregister_fixed_link(np);
|
||||
of_node_put(ugeth->ug_info->tbi_node);
|
||||
of_node_put(ugeth->ug_info->phy_node);
|
||||
kfree(ugeth->ug_info);
|
||||
free_netdev(dev);
|
||||
|
||||
return 0;
|
||||
@ -3968,17 +3798,10 @@ static struct platform_driver ucc_geth_driver = {
|
||||
|
||||
static int __init ucc_geth_init(void)
|
||||
{
|
||||
int i, ret;
|
||||
|
||||
if (netif_msg_drv(&debug))
|
||||
pr_info(DRV_DESC "\n");
|
||||
for (i = 0; i < 8; i++)
|
||||
memcpy(&(ugeth_info[i]), &ugeth_primary_info,
|
||||
sizeof(ugeth_primary_info));
|
||||
|
||||
ret = platform_driver_register(&ucc_geth_driver);
|
||||
|
||||
return ret;
|
||||
return platform_driver_register(&ucc_geth_driver);
|
||||
}
|
||||
|
||||
static void __exit ucc_geth_exit(void)
|
||||
|
@ -1076,8 +1076,6 @@ struct ucc_geth_tad_params {
|
||||
/* GETH protocol initialization structure */
|
||||
struct ucc_geth_info {
|
||||
struct ucc_fast_info uf_info;
|
||||
u8 numQueuesTx;
|
||||
u8 numQueuesRx;
|
||||
int ipCheckSumCheck;
|
||||
int ipCheckSumGenerate;
|
||||
int rxExtendedFiltering;
|
||||
@ -1165,9 +1163,7 @@ struct ucc_geth_private {
|
||||
struct ucc_geth_exf_global_pram __iomem *p_exf_glbl_param;
|
||||
u32 exf_glbl_param_offset;
|
||||
struct ucc_geth_rx_global_pram __iomem *p_rx_glbl_pram;
|
||||
u32 rx_glbl_pram_offset;
|
||||
struct ucc_geth_tx_global_pram __iomem *p_tx_glbl_pram;
|
||||
u32 tx_glbl_pram_offset;
|
||||
struct ucc_geth_send_queue_mem_region __iomem *p_send_q_mem_reg;
|
||||
u32 send_q_mem_reg_offset;
|
||||
struct ucc_geth_thread_data_tx __iomem *p_thread_data_tx;
|
||||
@ -1185,9 +1181,7 @@ struct ucc_geth_private {
|
||||
struct ucc_geth_rx_bd_queues_entry __iomem *p_rx_bd_qs_tbl;
|
||||
u32 rx_bd_qs_tbl_offset;
|
||||
u8 __iomem *p_tx_bd_ring[NUM_TX_QUEUES];
|
||||
u32 tx_bd_ring_offset[NUM_TX_QUEUES];
|
||||
u8 __iomem *p_rx_bd_ring[NUM_RX_QUEUES];
|
||||
u32 rx_bd_ring_offset[NUM_RX_QUEUES];
|
||||
u8 __iomem *confBd[NUM_TX_QUEUES];
|
||||
u8 __iomem *txBd[NUM_TX_QUEUES];
|
||||
u8 __iomem *rxBd[NUM_RX_QUEUES];
|
||||
|
@ -27,7 +27,7 @@
|
||||
|
||||
static struct gen_pool *muram_pool;
|
||||
static spinlock_t cpm_muram_lock;
|
||||
static u8 __iomem *muram_vbase;
|
||||
static void __iomem *muram_vbase;
|
||||
static phys_addr_t muram_pbase;
|
||||
|
||||
struct muram_block {
|
||||
@ -223,9 +223,9 @@ void __iomem *cpm_muram_addr(unsigned long offset)
|
||||
}
|
||||
EXPORT_SYMBOL(cpm_muram_addr);
|
||||
|
||||
unsigned long cpm_muram_offset(void __iomem *addr)
|
||||
unsigned long cpm_muram_offset(const void __iomem *addr)
|
||||
{
|
||||
return addr - (void __iomem *)muram_vbase;
|
||||
return addr - muram_vbase;
|
||||
}
|
||||
EXPORT_SYMBOL(cpm_muram_offset);
|
||||
|
||||
@ -235,6 +235,18 @@ EXPORT_SYMBOL(cpm_muram_offset);
|
||||
*/
|
||||
dma_addr_t cpm_muram_dma(void __iomem *addr)
|
||||
{
|
||||
return muram_pbase + ((u8 __iomem *)addr - muram_vbase);
|
||||
return muram_pbase + (addr - muram_vbase);
|
||||
}
|
||||
EXPORT_SYMBOL(cpm_muram_dma);
|
||||
|
||||
/*
|
||||
* As cpm_muram_free, but takes the virtual address rather than the
|
||||
* muram offset.
|
||||
*/
|
||||
void cpm_muram_free_addr(const void __iomem *addr)
|
||||
{
|
||||
if (!addr)
|
||||
return;
|
||||
cpm_muram_free(cpm_muram_offset(addr));
|
||||
}
|
||||
EXPORT_SYMBOL(cpm_muram_free_addr);
|
||||
|
@ -27,12 +27,6 @@
|
||||
#define QE_NUM_OF_BRGS 16
|
||||
#define QE_NUM_OF_PORTS 1024
|
||||
|
||||
/* Memory partitions
|
||||
*/
|
||||
#define MEM_PART_SYSTEM 0
|
||||
#define MEM_PART_SECONDARY 1
|
||||
#define MEM_PART_MURAM 2
|
||||
|
||||
/* Clocks and BRGs */
|
||||
enum qe_clock {
|
||||
QE_CLK_NONE = 0,
|
||||
@ -102,8 +96,9 @@ s32 cpm_muram_alloc(unsigned long size, unsigned long align);
|
||||
void cpm_muram_free(s32 offset);
|
||||
s32 cpm_muram_alloc_fixed(unsigned long offset, unsigned long size);
|
||||
void __iomem *cpm_muram_addr(unsigned long offset);
|
||||
unsigned long cpm_muram_offset(void __iomem *addr);
|
||||
unsigned long cpm_muram_offset(const void __iomem *addr);
|
||||
dma_addr_t cpm_muram_dma(void __iomem *addr);
|
||||
void cpm_muram_free_addr(const void __iomem *addr);
|
||||
#else
|
||||
static inline s32 cpm_muram_alloc(unsigned long size,
|
||||
unsigned long align)
|
||||
@ -126,7 +121,7 @@ static inline void __iomem *cpm_muram_addr(unsigned long offset)
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static inline unsigned long cpm_muram_offset(void __iomem *addr)
|
||||
static inline unsigned long cpm_muram_offset(const void __iomem *addr)
|
||||
{
|
||||
return -ENOSYS;
|
||||
}
|
||||
@ -135,6 +130,9 @@ static inline dma_addr_t cpm_muram_dma(void __iomem *addr)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
static inline void cpm_muram_free_addr(const void __iomem *addr)
|
||||
{
|
||||
}
|
||||
#endif /* defined(CONFIG_CPM) || defined(CONFIG_QUICC_ENGINE) */
|
||||
|
||||
/* QE PIO */
|
||||
@ -239,6 +237,7 @@ static inline int qe_alive_during_sleep(void)
|
||||
#define qe_muram_addr cpm_muram_addr
|
||||
#define qe_muram_offset cpm_muram_offset
|
||||
#define qe_muram_dma cpm_muram_dma
|
||||
#define qe_muram_free_addr cpm_muram_free_addr
|
||||
|
||||
#ifdef CONFIG_PPC32
|
||||
#define qe_iowrite8(val, addr) out_8(addr, val)
|
||||
|
@ -146,7 +146,6 @@ struct ucc_fast_info {
|
||||
resource_size_t regs;
|
||||
int irq;
|
||||
u32 uccm_mask;
|
||||
int bd_mem_part;
|
||||
int brkpt_support;
|
||||
int grant_support;
|
||||
int tsa;
|
||||
|
Loading…
x
Reference in New Issue
Block a user