4405 lines
115 KiB
C
4405 lines
115 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/* Copyright (C) 2023 Intel Corporation */
|
|
|
|
#include <net/libeth/rx.h>
|
|
#include <net/libeth/tx.h>
|
|
|
|
#include "idpf.h"
|
|
#include "idpf_virtchnl.h"
|
|
|
|
struct idpf_tx_stash {
|
|
struct hlist_node hlist;
|
|
struct libeth_sqe buf;
|
|
};
|
|
|
|
#define idpf_tx_buf_compl_tag(buf) (*(u32 *)&(buf)->priv)
|
|
LIBETH_SQE_CHECK_PRIV(u32);
|
|
|
|
static bool idpf_chk_linearize(struct sk_buff *skb, unsigned int max_bufs,
|
|
unsigned int count);
|
|
|
|
/**
|
|
* idpf_buf_lifo_push - push a buffer pointer onto stack
|
|
* @stack: pointer to stack struct
|
|
* @buf: pointer to buf to push
|
|
*
|
|
* Returns 0 on success, negative on failure
|
|
**/
|
|
static int idpf_buf_lifo_push(struct idpf_buf_lifo *stack,
|
|
struct idpf_tx_stash *buf)
|
|
{
|
|
if (unlikely(stack->top == stack->size))
|
|
return -ENOSPC;
|
|
|
|
stack->bufs[stack->top++] = buf;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_buf_lifo_pop - pop a buffer pointer from stack
|
|
* @stack: pointer to stack struct
|
|
**/
|
|
static struct idpf_tx_stash *idpf_buf_lifo_pop(struct idpf_buf_lifo *stack)
|
|
{
|
|
if (unlikely(!stack->top))
|
|
return NULL;
|
|
|
|
return stack->bufs[--stack->top];
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_timeout - Respond to a Tx Hang
|
|
* @netdev: network interface device structure
|
|
* @txqueue: TX queue
|
|
*/
|
|
void idpf_tx_timeout(struct net_device *netdev, unsigned int txqueue)
|
|
{
|
|
struct idpf_adapter *adapter = idpf_netdev_to_adapter(netdev);
|
|
|
|
adapter->tx_timeout_count++;
|
|
|
|
netdev_err(netdev, "Detected Tx timeout: Count %d, Queue %d\n",
|
|
adapter->tx_timeout_count, txqueue);
|
|
if (!idpf_is_reset_in_prog(adapter)) {
|
|
set_bit(IDPF_HR_FUNC_RESET, adapter->flags);
|
|
queue_delayed_work(adapter->vc_event_wq,
|
|
&adapter->vc_event_task,
|
|
msecs_to_jiffies(10));
|
|
}
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_buf_rel_all - Free any empty Tx buffers
|
|
* @txq: queue to be cleaned
|
|
*/
|
|
static void idpf_tx_buf_rel_all(struct idpf_tx_queue *txq)
|
|
{
|
|
struct libeth_sq_napi_stats ss = { };
|
|
struct idpf_buf_lifo *buf_stack;
|
|
struct idpf_tx_stash *stash;
|
|
struct libeth_cq_pp cp = {
|
|
.dev = txq->dev,
|
|
.ss = &ss,
|
|
};
|
|
struct hlist_node *tmp;
|
|
u32 i, tag;
|
|
|
|
/* Buffers already cleared, nothing to do */
|
|
if (!txq->tx_buf)
|
|
return;
|
|
|
|
/* Free all the Tx buffer sk_buffs */
|
|
for (i = 0; i < txq->desc_count; i++)
|
|
libeth_tx_complete(&txq->tx_buf[i], &cp);
|
|
|
|
kfree(txq->tx_buf);
|
|
txq->tx_buf = NULL;
|
|
|
|
if (!idpf_queue_has(FLOW_SCH_EN, txq))
|
|
return;
|
|
|
|
buf_stack = &txq->stash->buf_stack;
|
|
if (!buf_stack->bufs)
|
|
return;
|
|
|
|
/*
|
|
* If a Tx timeout occurred, there are potentially still bufs in the
|
|
* hash table, free them here.
|
|
*/
|
|
hash_for_each_safe(txq->stash->sched_buf_hash, tag, tmp, stash,
|
|
hlist) {
|
|
if (!stash)
|
|
continue;
|
|
|
|
libeth_tx_complete(&stash->buf, &cp);
|
|
hash_del(&stash->hlist);
|
|
idpf_buf_lifo_push(buf_stack, stash);
|
|
}
|
|
|
|
for (i = 0; i < buf_stack->size; i++)
|
|
kfree(buf_stack->bufs[i]);
|
|
|
|
kfree(buf_stack->bufs);
|
|
buf_stack->bufs = NULL;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_desc_rel - Free Tx resources per queue
|
|
* @txq: Tx descriptor ring for a specific queue
|
|
*
|
|
* Free all transmit software resources
|
|
*/
|
|
static void idpf_tx_desc_rel(struct idpf_tx_queue *txq)
|
|
{
|
|
idpf_tx_buf_rel_all(txq);
|
|
netdev_tx_reset_subqueue(txq->netdev, txq->idx);
|
|
|
|
if (!txq->desc_ring)
|
|
return;
|
|
|
|
dmam_free_coherent(txq->dev, txq->size, txq->desc_ring, txq->dma);
|
|
txq->desc_ring = NULL;
|
|
txq->next_to_use = 0;
|
|
txq->next_to_clean = 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_compl_desc_rel - Free completion resources per queue
|
|
* @complq: completion queue
|
|
*
|
|
* Free all completion software resources.
|
|
*/
|
|
static void idpf_compl_desc_rel(struct idpf_compl_queue *complq)
|
|
{
|
|
if (!complq->comp)
|
|
return;
|
|
|
|
dma_free_coherent(complq->netdev->dev.parent, complq->size,
|
|
complq->comp, complq->dma);
|
|
complq->comp = NULL;
|
|
complq->next_to_use = 0;
|
|
complq->next_to_clean = 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_desc_rel_all - Free Tx Resources for All Queues
|
|
* @vport: virtual port structure
|
|
*
|
|
* Free all transmit software resources
|
|
*/
|
|
static void idpf_tx_desc_rel_all(struct idpf_vport *vport)
|
|
{
|
|
int i, j;
|
|
|
|
if (!vport->txq_grps)
|
|
return;
|
|
|
|
for (i = 0; i < vport->num_txq_grp; i++) {
|
|
struct idpf_txq_group *txq_grp = &vport->txq_grps[i];
|
|
|
|
for (j = 0; j < txq_grp->num_txq; j++)
|
|
idpf_tx_desc_rel(txq_grp->txqs[j]);
|
|
|
|
if (idpf_is_queue_model_split(vport->txq_model))
|
|
idpf_compl_desc_rel(txq_grp->complq);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_buf_alloc_all - Allocate memory for all buffer resources
|
|
* @tx_q: queue for which the buffers are allocated
|
|
*
|
|
* Returns 0 on success, negative on failure
|
|
*/
|
|
static int idpf_tx_buf_alloc_all(struct idpf_tx_queue *tx_q)
|
|
{
|
|
struct idpf_buf_lifo *buf_stack;
|
|
int buf_size;
|
|
int i;
|
|
|
|
/* Allocate book keeping buffers only. Buffers to be supplied to HW
|
|
* are allocated by kernel network stack and received as part of skb
|
|
*/
|
|
buf_size = sizeof(struct idpf_tx_buf) * tx_q->desc_count;
|
|
tx_q->tx_buf = kzalloc(buf_size, GFP_KERNEL);
|
|
if (!tx_q->tx_buf)
|
|
return -ENOMEM;
|
|
|
|
if (!idpf_queue_has(FLOW_SCH_EN, tx_q))
|
|
return 0;
|
|
|
|
buf_stack = &tx_q->stash->buf_stack;
|
|
|
|
/* Initialize tx buf stack for out-of-order completions if
|
|
* flow scheduling offload is enabled
|
|
*/
|
|
buf_stack->bufs = kcalloc(tx_q->desc_count, sizeof(*buf_stack->bufs),
|
|
GFP_KERNEL);
|
|
if (!buf_stack->bufs)
|
|
return -ENOMEM;
|
|
|
|
buf_stack->size = tx_q->desc_count;
|
|
buf_stack->top = tx_q->desc_count;
|
|
|
|
for (i = 0; i < tx_q->desc_count; i++) {
|
|
buf_stack->bufs[i] = kzalloc(sizeof(*buf_stack->bufs[i]),
|
|
GFP_KERNEL);
|
|
if (!buf_stack->bufs[i])
|
|
return -ENOMEM;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_desc_alloc - Allocate the Tx descriptors
|
|
* @vport: vport to allocate resources for
|
|
* @tx_q: the tx ring to set up
|
|
*
|
|
* Returns 0 on success, negative on failure
|
|
*/
|
|
static int idpf_tx_desc_alloc(const struct idpf_vport *vport,
|
|
struct idpf_tx_queue *tx_q)
|
|
{
|
|
struct device *dev = tx_q->dev;
|
|
int err;
|
|
|
|
err = idpf_tx_buf_alloc_all(tx_q);
|
|
if (err)
|
|
goto err_alloc;
|
|
|
|
tx_q->size = tx_q->desc_count * sizeof(*tx_q->base_tx);
|
|
|
|
/* Allocate descriptors also round up to nearest 4K */
|
|
tx_q->size = ALIGN(tx_q->size, 4096);
|
|
tx_q->desc_ring = dmam_alloc_coherent(dev, tx_q->size, &tx_q->dma,
|
|
GFP_KERNEL);
|
|
if (!tx_q->desc_ring) {
|
|
dev_err(dev, "Unable to allocate memory for the Tx descriptor ring, size=%d\n",
|
|
tx_q->size);
|
|
err = -ENOMEM;
|
|
goto err_alloc;
|
|
}
|
|
|
|
tx_q->next_to_use = 0;
|
|
tx_q->next_to_clean = 0;
|
|
idpf_queue_set(GEN_CHK, tx_q);
|
|
|
|
return 0;
|
|
|
|
err_alloc:
|
|
idpf_tx_desc_rel(tx_q);
|
|
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* idpf_compl_desc_alloc - allocate completion descriptors
|
|
* @vport: vport to allocate resources for
|
|
* @complq: completion queue to set up
|
|
*
|
|
* Return: 0 on success, -errno on failure.
|
|
*/
|
|
static int idpf_compl_desc_alloc(const struct idpf_vport *vport,
|
|
struct idpf_compl_queue *complq)
|
|
{
|
|
complq->size = array_size(complq->desc_count, sizeof(*complq->comp));
|
|
|
|
complq->comp = dma_alloc_coherent(complq->netdev->dev.parent,
|
|
complq->size, &complq->dma,
|
|
GFP_KERNEL);
|
|
if (!complq->comp)
|
|
return -ENOMEM;
|
|
|
|
complq->next_to_use = 0;
|
|
complq->next_to_clean = 0;
|
|
idpf_queue_set(GEN_CHK, complq);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_desc_alloc_all - allocate all queues Tx resources
|
|
* @vport: virtual port private structure
|
|
*
|
|
* Returns 0 on success, negative on failure
|
|
*/
|
|
static int idpf_tx_desc_alloc_all(struct idpf_vport *vport)
|
|
{
|
|
int err = 0;
|
|
int i, j;
|
|
|
|
/* Setup buffer queues. In single queue model buffer queues and
|
|
* completion queues will be same
|
|
*/
|
|
for (i = 0; i < vport->num_txq_grp; i++) {
|
|
for (j = 0; j < vport->txq_grps[i].num_txq; j++) {
|
|
struct idpf_tx_queue *txq = vport->txq_grps[i].txqs[j];
|
|
u8 gen_bits = 0;
|
|
u16 bufidx_mask;
|
|
|
|
err = idpf_tx_desc_alloc(vport, txq);
|
|
if (err) {
|
|
pci_err(vport->adapter->pdev,
|
|
"Allocation for Tx Queue %u failed\n",
|
|
i);
|
|
goto err_out;
|
|
}
|
|
|
|
if (!idpf_is_queue_model_split(vport->txq_model))
|
|
continue;
|
|
|
|
txq->compl_tag_cur_gen = 0;
|
|
|
|
/* Determine the number of bits in the bufid
|
|
* mask and add one to get the start of the
|
|
* generation bits
|
|
*/
|
|
bufidx_mask = txq->desc_count - 1;
|
|
while (bufidx_mask >> 1) {
|
|
txq->compl_tag_gen_s++;
|
|
bufidx_mask = bufidx_mask >> 1;
|
|
}
|
|
txq->compl_tag_gen_s++;
|
|
|
|
gen_bits = IDPF_TX_SPLITQ_COMPL_TAG_WIDTH -
|
|
txq->compl_tag_gen_s;
|
|
txq->compl_tag_gen_max = GETMAXVAL(gen_bits);
|
|
|
|
/* Set bufid mask based on location of first
|
|
* gen bit; it cannot simply be the descriptor
|
|
* ring size-1 since we can have size values
|
|
* where not all of those bits are set.
|
|
*/
|
|
txq->compl_tag_bufid_m =
|
|
GETMAXVAL(txq->compl_tag_gen_s);
|
|
}
|
|
|
|
if (!idpf_is_queue_model_split(vport->txq_model))
|
|
continue;
|
|
|
|
/* Setup completion queues */
|
|
err = idpf_compl_desc_alloc(vport, vport->txq_grps[i].complq);
|
|
if (err) {
|
|
pci_err(vport->adapter->pdev,
|
|
"Allocation for Tx Completion Queue %u failed\n",
|
|
i);
|
|
goto err_out;
|
|
}
|
|
}
|
|
|
|
err_out:
|
|
if (err)
|
|
idpf_tx_desc_rel_all(vport);
|
|
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_page_rel - Release an rx buffer page
|
|
* @rx_buf: the buffer to free
|
|
*/
|
|
static void idpf_rx_page_rel(struct libeth_fqe *rx_buf)
|
|
{
|
|
if (unlikely(!rx_buf->page))
|
|
return;
|
|
|
|
page_pool_put_full_page(rx_buf->page->pp, rx_buf->page, false);
|
|
|
|
rx_buf->page = NULL;
|
|
rx_buf->offset = 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_hdr_buf_rel_all - Release header buffer memory
|
|
* @bufq: queue to use
|
|
*/
|
|
static void idpf_rx_hdr_buf_rel_all(struct idpf_buf_queue *bufq)
|
|
{
|
|
struct libeth_fq fq = {
|
|
.fqes = bufq->hdr_buf,
|
|
.pp = bufq->hdr_pp,
|
|
};
|
|
|
|
for (u32 i = 0; i < bufq->desc_count; i++)
|
|
idpf_rx_page_rel(&bufq->hdr_buf[i]);
|
|
|
|
libeth_rx_fq_destroy(&fq);
|
|
bufq->hdr_buf = NULL;
|
|
bufq->hdr_pp = NULL;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_buf_rel_bufq - Free all Rx buffer resources for a buffer queue
|
|
* @bufq: queue to be cleaned
|
|
*/
|
|
static void idpf_rx_buf_rel_bufq(struct idpf_buf_queue *bufq)
|
|
{
|
|
struct libeth_fq fq = {
|
|
.fqes = bufq->buf,
|
|
.pp = bufq->pp,
|
|
};
|
|
|
|
/* queue already cleared, nothing to do */
|
|
if (!bufq->buf)
|
|
return;
|
|
|
|
/* Free all the bufs allocated and given to hw on Rx queue */
|
|
for (u32 i = 0; i < bufq->desc_count; i++)
|
|
idpf_rx_page_rel(&bufq->buf[i]);
|
|
|
|
if (idpf_queue_has(HSPLIT_EN, bufq))
|
|
idpf_rx_hdr_buf_rel_all(bufq);
|
|
|
|
libeth_rx_fq_destroy(&fq);
|
|
bufq->buf = NULL;
|
|
bufq->pp = NULL;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_buf_rel_all - Free all Rx buffer resources for a receive queue
|
|
* @rxq: queue to be cleaned
|
|
*/
|
|
static void idpf_rx_buf_rel_all(struct idpf_rx_queue *rxq)
|
|
{
|
|
struct libeth_fq fq = {
|
|
.fqes = rxq->rx_buf,
|
|
.pp = rxq->pp,
|
|
};
|
|
|
|
if (!rxq->rx_buf)
|
|
return;
|
|
|
|
for (u32 i = 0; i < rxq->desc_count; i++)
|
|
idpf_rx_page_rel(&rxq->rx_buf[i]);
|
|
|
|
libeth_rx_fq_destroy(&fq);
|
|
rxq->rx_buf = NULL;
|
|
rxq->pp = NULL;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_desc_rel - Free a specific Rx q resources
|
|
* @rxq: queue to clean the resources from
|
|
* @dev: device to free DMA memory
|
|
* @model: single or split queue model
|
|
*
|
|
* Free a specific rx queue resources
|
|
*/
|
|
static void idpf_rx_desc_rel(struct idpf_rx_queue *rxq, struct device *dev,
|
|
u32 model)
|
|
{
|
|
if (!rxq)
|
|
return;
|
|
|
|
if (rxq->skb) {
|
|
dev_kfree_skb_any(rxq->skb);
|
|
rxq->skb = NULL;
|
|
}
|
|
|
|
if (!idpf_is_queue_model_split(model))
|
|
idpf_rx_buf_rel_all(rxq);
|
|
|
|
rxq->next_to_alloc = 0;
|
|
rxq->next_to_clean = 0;
|
|
rxq->next_to_use = 0;
|
|
if (!rxq->desc_ring)
|
|
return;
|
|
|
|
dmam_free_coherent(dev, rxq->size, rxq->desc_ring, rxq->dma);
|
|
rxq->desc_ring = NULL;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_desc_rel_bufq - free buffer queue resources
|
|
* @bufq: buffer queue to clean the resources from
|
|
* @dev: device to free DMA memory
|
|
*/
|
|
static void idpf_rx_desc_rel_bufq(struct idpf_buf_queue *bufq,
|
|
struct device *dev)
|
|
{
|
|
if (!bufq)
|
|
return;
|
|
|
|
idpf_rx_buf_rel_bufq(bufq);
|
|
|
|
bufq->next_to_alloc = 0;
|
|
bufq->next_to_clean = 0;
|
|
bufq->next_to_use = 0;
|
|
|
|
if (!bufq->split_buf)
|
|
return;
|
|
|
|
dma_free_coherent(dev, bufq->size, bufq->split_buf, bufq->dma);
|
|
bufq->split_buf = NULL;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_desc_rel_all - Free Rx Resources for All Queues
|
|
* @vport: virtual port structure
|
|
*
|
|
* Free all rx queues resources
|
|
*/
|
|
static void idpf_rx_desc_rel_all(struct idpf_vport *vport)
|
|
{
|
|
struct device *dev = &vport->adapter->pdev->dev;
|
|
struct idpf_rxq_group *rx_qgrp;
|
|
u16 num_rxq;
|
|
int i, j;
|
|
|
|
if (!vport->rxq_grps)
|
|
return;
|
|
|
|
for (i = 0; i < vport->num_rxq_grp; i++) {
|
|
rx_qgrp = &vport->rxq_grps[i];
|
|
|
|
if (!idpf_is_queue_model_split(vport->rxq_model)) {
|
|
for (j = 0; j < rx_qgrp->singleq.num_rxq; j++)
|
|
idpf_rx_desc_rel(rx_qgrp->singleq.rxqs[j], dev,
|
|
VIRTCHNL2_QUEUE_MODEL_SINGLE);
|
|
continue;
|
|
}
|
|
|
|
num_rxq = rx_qgrp->splitq.num_rxq_sets;
|
|
for (j = 0; j < num_rxq; j++)
|
|
idpf_rx_desc_rel(&rx_qgrp->splitq.rxq_sets[j]->rxq,
|
|
dev, VIRTCHNL2_QUEUE_MODEL_SPLIT);
|
|
|
|
if (!rx_qgrp->splitq.bufq_sets)
|
|
continue;
|
|
|
|
for (j = 0; j < vport->num_bufqs_per_qgrp; j++) {
|
|
struct idpf_bufq_set *bufq_set =
|
|
&rx_qgrp->splitq.bufq_sets[j];
|
|
|
|
idpf_rx_desc_rel_bufq(&bufq_set->bufq, dev);
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_buf_hw_update - Store the new tail and head values
|
|
* @bufq: queue to bump
|
|
* @val: new head index
|
|
*/
|
|
static void idpf_rx_buf_hw_update(struct idpf_buf_queue *bufq, u32 val)
|
|
{
|
|
bufq->next_to_use = val;
|
|
|
|
if (unlikely(!bufq->tail))
|
|
return;
|
|
|
|
/* writel has an implicit memory barrier */
|
|
writel(val, bufq->tail);
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_hdr_buf_alloc_all - Allocate memory for header buffers
|
|
* @bufq: ring to use
|
|
*
|
|
* Returns 0 on success, negative on failure.
|
|
*/
|
|
static int idpf_rx_hdr_buf_alloc_all(struct idpf_buf_queue *bufq)
|
|
{
|
|
struct libeth_fq fq = {
|
|
.count = bufq->desc_count,
|
|
.type = LIBETH_FQE_HDR,
|
|
.nid = idpf_q_vector_to_mem(bufq->q_vector),
|
|
};
|
|
int ret;
|
|
|
|
ret = libeth_rx_fq_create(&fq, &bufq->q_vector->napi);
|
|
if (ret)
|
|
return ret;
|
|
|
|
bufq->hdr_pp = fq.pp;
|
|
bufq->hdr_buf = fq.fqes;
|
|
bufq->hdr_truesize = fq.truesize;
|
|
bufq->rx_hbuf_size = fq.buf_len;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_post_buf_refill - Post buffer id to refill queue
|
|
* @refillq: refill queue to post to
|
|
* @buf_id: buffer id to post
|
|
*/
|
|
static void idpf_rx_post_buf_refill(struct idpf_sw_queue *refillq, u16 buf_id)
|
|
{
|
|
u32 nta = refillq->next_to_use;
|
|
|
|
/* store the buffer ID and the SW maintained GEN bit to the refillq */
|
|
refillq->ring[nta] =
|
|
FIELD_PREP(IDPF_RX_BI_BUFID_M, buf_id) |
|
|
FIELD_PREP(IDPF_RX_BI_GEN_M,
|
|
idpf_queue_has(GEN_CHK, refillq));
|
|
|
|
if (unlikely(++nta == refillq->desc_count)) {
|
|
nta = 0;
|
|
idpf_queue_change(GEN_CHK, refillq);
|
|
}
|
|
|
|
refillq->next_to_use = nta;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_post_buf_desc - Post buffer to bufq descriptor ring
|
|
* @bufq: buffer queue to post to
|
|
* @buf_id: buffer id to post
|
|
*
|
|
* Returns false if buffer could not be allocated, true otherwise.
|
|
*/
|
|
static bool idpf_rx_post_buf_desc(struct idpf_buf_queue *bufq, u16 buf_id)
|
|
{
|
|
struct virtchnl2_splitq_rx_buf_desc *splitq_rx_desc = NULL;
|
|
struct libeth_fq_fp fq = {
|
|
.count = bufq->desc_count,
|
|
};
|
|
u16 nta = bufq->next_to_alloc;
|
|
dma_addr_t addr;
|
|
|
|
splitq_rx_desc = &bufq->split_buf[nta];
|
|
|
|
if (idpf_queue_has(HSPLIT_EN, bufq)) {
|
|
fq.pp = bufq->hdr_pp;
|
|
fq.fqes = bufq->hdr_buf;
|
|
fq.truesize = bufq->hdr_truesize;
|
|
|
|
addr = libeth_rx_alloc(&fq, buf_id);
|
|
if (addr == DMA_MAPPING_ERROR)
|
|
return false;
|
|
|
|
splitq_rx_desc->hdr_addr = cpu_to_le64(addr);
|
|
}
|
|
|
|
fq.pp = bufq->pp;
|
|
fq.fqes = bufq->buf;
|
|
fq.truesize = bufq->truesize;
|
|
|
|
addr = libeth_rx_alloc(&fq, buf_id);
|
|
if (addr == DMA_MAPPING_ERROR)
|
|
return false;
|
|
|
|
splitq_rx_desc->pkt_addr = cpu_to_le64(addr);
|
|
splitq_rx_desc->qword0.buf_id = cpu_to_le16(buf_id);
|
|
|
|
nta++;
|
|
if (unlikely(nta == bufq->desc_count))
|
|
nta = 0;
|
|
bufq->next_to_alloc = nta;
|
|
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_post_init_bufs - Post initial buffers to bufq
|
|
* @bufq: buffer queue to post working set to
|
|
* @working_set: number of buffers to put in working set
|
|
*
|
|
* Returns true if @working_set bufs were posted successfully, false otherwise.
|
|
*/
|
|
static bool idpf_rx_post_init_bufs(struct idpf_buf_queue *bufq,
|
|
u16 working_set)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < working_set; i++) {
|
|
if (!idpf_rx_post_buf_desc(bufq, i))
|
|
return false;
|
|
}
|
|
|
|
idpf_rx_buf_hw_update(bufq, ALIGN_DOWN(bufq->next_to_alloc,
|
|
IDPF_RX_BUF_STRIDE));
|
|
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_buf_alloc_singleq - Allocate memory for all buffer resources
|
|
* @rxq: queue for which the buffers are allocated
|
|
*
|
|
* Return: 0 on success, -ENOMEM on failure.
|
|
*/
|
|
static int idpf_rx_buf_alloc_singleq(struct idpf_rx_queue *rxq)
|
|
{
|
|
if (idpf_rx_singleq_buf_hw_alloc_all(rxq, rxq->desc_count - 1))
|
|
goto err;
|
|
|
|
return 0;
|
|
|
|
err:
|
|
idpf_rx_buf_rel_all(rxq);
|
|
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_bufs_init_singleq - Initialize page pool and allocate Rx bufs
|
|
* @rxq: buffer queue to create page pool for
|
|
*
|
|
* Return: 0 on success, -errno on failure.
|
|
*/
|
|
static int idpf_rx_bufs_init_singleq(struct idpf_rx_queue *rxq)
|
|
{
|
|
struct libeth_fq fq = {
|
|
.count = rxq->desc_count,
|
|
.type = LIBETH_FQE_MTU,
|
|
.nid = idpf_q_vector_to_mem(rxq->q_vector),
|
|
};
|
|
int ret;
|
|
|
|
ret = libeth_rx_fq_create(&fq, &rxq->q_vector->napi);
|
|
if (ret)
|
|
return ret;
|
|
|
|
rxq->pp = fq.pp;
|
|
rxq->rx_buf = fq.fqes;
|
|
rxq->truesize = fq.truesize;
|
|
rxq->rx_buf_size = fq.buf_len;
|
|
|
|
return idpf_rx_buf_alloc_singleq(rxq);
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_buf_alloc_all - Allocate memory for all buffer resources
|
|
* @rxbufq: queue for which the buffers are allocated
|
|
*
|
|
* Returns 0 on success, negative on failure
|
|
*/
|
|
static int idpf_rx_buf_alloc_all(struct idpf_buf_queue *rxbufq)
|
|
{
|
|
int err = 0;
|
|
|
|
if (idpf_queue_has(HSPLIT_EN, rxbufq)) {
|
|
err = idpf_rx_hdr_buf_alloc_all(rxbufq);
|
|
if (err)
|
|
goto rx_buf_alloc_all_out;
|
|
}
|
|
|
|
/* Allocate buffers to be given to HW. */
|
|
if (!idpf_rx_post_init_bufs(rxbufq, IDPF_RX_BUFQ_WORKING_SET(rxbufq)))
|
|
err = -ENOMEM;
|
|
|
|
rx_buf_alloc_all_out:
|
|
if (err)
|
|
idpf_rx_buf_rel_bufq(rxbufq);
|
|
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_bufs_init - Initialize page pool, allocate rx bufs, and post to HW
|
|
* @bufq: buffer queue to create page pool for
|
|
* @type: type of Rx buffers to allocate
|
|
*
|
|
* Returns 0 on success, negative on failure
|
|
*/
|
|
static int idpf_rx_bufs_init(struct idpf_buf_queue *bufq,
|
|
enum libeth_fqe_type type)
|
|
{
|
|
struct libeth_fq fq = {
|
|
.truesize = bufq->truesize,
|
|
.count = bufq->desc_count,
|
|
.type = type,
|
|
.hsplit = idpf_queue_has(HSPLIT_EN, bufq),
|
|
.nid = idpf_q_vector_to_mem(bufq->q_vector),
|
|
};
|
|
int ret;
|
|
|
|
ret = libeth_rx_fq_create(&fq, &bufq->q_vector->napi);
|
|
if (ret)
|
|
return ret;
|
|
|
|
bufq->pp = fq.pp;
|
|
bufq->buf = fq.fqes;
|
|
bufq->truesize = fq.truesize;
|
|
bufq->rx_buf_size = fq.buf_len;
|
|
|
|
return idpf_rx_buf_alloc_all(bufq);
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_bufs_init_all - Initialize all RX bufs
|
|
* @vport: virtual port struct
|
|
*
|
|
* Returns 0 on success, negative on failure
|
|
*/
|
|
int idpf_rx_bufs_init_all(struct idpf_vport *vport)
|
|
{
|
|
bool split = idpf_is_queue_model_split(vport->rxq_model);
|
|
int i, j, err;
|
|
|
|
for (i = 0; i < vport->num_rxq_grp; i++) {
|
|
struct idpf_rxq_group *rx_qgrp = &vport->rxq_grps[i];
|
|
u32 truesize = 0;
|
|
|
|
/* Allocate bufs for the rxq itself in singleq */
|
|
if (!split) {
|
|
int num_rxq = rx_qgrp->singleq.num_rxq;
|
|
|
|
for (j = 0; j < num_rxq; j++) {
|
|
struct idpf_rx_queue *q;
|
|
|
|
q = rx_qgrp->singleq.rxqs[j];
|
|
err = idpf_rx_bufs_init_singleq(q);
|
|
if (err)
|
|
return err;
|
|
}
|
|
|
|
continue;
|
|
}
|
|
|
|
/* Otherwise, allocate bufs for the buffer queues */
|
|
for (j = 0; j < vport->num_bufqs_per_qgrp; j++) {
|
|
enum libeth_fqe_type type;
|
|
struct idpf_buf_queue *q;
|
|
|
|
q = &rx_qgrp->splitq.bufq_sets[j].bufq;
|
|
q->truesize = truesize;
|
|
|
|
type = truesize ? LIBETH_FQE_SHORT : LIBETH_FQE_MTU;
|
|
|
|
err = idpf_rx_bufs_init(q, type);
|
|
if (err)
|
|
return err;
|
|
|
|
truesize = q->truesize >> 1;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_desc_alloc - Allocate queue Rx resources
|
|
* @vport: vport to allocate resources for
|
|
* @rxq: Rx queue for which the resources are setup
|
|
*
|
|
* Returns 0 on success, negative on failure
|
|
*/
|
|
static int idpf_rx_desc_alloc(const struct idpf_vport *vport,
|
|
struct idpf_rx_queue *rxq)
|
|
{
|
|
struct device *dev = &vport->adapter->pdev->dev;
|
|
|
|
rxq->size = rxq->desc_count * sizeof(union virtchnl2_rx_desc);
|
|
|
|
/* Allocate descriptors and also round up to nearest 4K */
|
|
rxq->size = ALIGN(rxq->size, 4096);
|
|
rxq->desc_ring = dmam_alloc_coherent(dev, rxq->size,
|
|
&rxq->dma, GFP_KERNEL);
|
|
if (!rxq->desc_ring) {
|
|
dev_err(dev, "Unable to allocate memory for the Rx descriptor ring, size=%d\n",
|
|
rxq->size);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
rxq->next_to_alloc = 0;
|
|
rxq->next_to_clean = 0;
|
|
rxq->next_to_use = 0;
|
|
idpf_queue_set(GEN_CHK, rxq);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_bufq_desc_alloc - Allocate buffer queue descriptor ring
|
|
* @vport: vport to allocate resources for
|
|
* @bufq: buffer queue for which the resources are set up
|
|
*
|
|
* Return: 0 on success, -ENOMEM on failure.
|
|
*/
|
|
static int idpf_bufq_desc_alloc(const struct idpf_vport *vport,
|
|
struct idpf_buf_queue *bufq)
|
|
{
|
|
struct device *dev = &vport->adapter->pdev->dev;
|
|
|
|
bufq->size = array_size(bufq->desc_count, sizeof(*bufq->split_buf));
|
|
|
|
bufq->split_buf = dma_alloc_coherent(dev, bufq->size, &bufq->dma,
|
|
GFP_KERNEL);
|
|
if (!bufq->split_buf)
|
|
return -ENOMEM;
|
|
|
|
bufq->next_to_alloc = 0;
|
|
bufq->next_to_clean = 0;
|
|
bufq->next_to_use = 0;
|
|
|
|
idpf_queue_set(GEN_CHK, bufq);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_desc_alloc_all - allocate all RX queues resources
|
|
* @vport: virtual port structure
|
|
*
|
|
* Returns 0 on success, negative on failure
|
|
*/
|
|
static int idpf_rx_desc_alloc_all(struct idpf_vport *vport)
|
|
{
|
|
struct idpf_rxq_group *rx_qgrp;
|
|
int i, j, err;
|
|
u16 num_rxq;
|
|
|
|
for (i = 0; i < vport->num_rxq_grp; i++) {
|
|
rx_qgrp = &vport->rxq_grps[i];
|
|
if (idpf_is_queue_model_split(vport->rxq_model))
|
|
num_rxq = rx_qgrp->splitq.num_rxq_sets;
|
|
else
|
|
num_rxq = rx_qgrp->singleq.num_rxq;
|
|
|
|
for (j = 0; j < num_rxq; j++) {
|
|
struct idpf_rx_queue *q;
|
|
|
|
if (idpf_is_queue_model_split(vport->rxq_model))
|
|
q = &rx_qgrp->splitq.rxq_sets[j]->rxq;
|
|
else
|
|
q = rx_qgrp->singleq.rxqs[j];
|
|
|
|
err = idpf_rx_desc_alloc(vport, q);
|
|
if (err) {
|
|
pci_err(vport->adapter->pdev,
|
|
"Memory allocation for Rx Queue %u failed\n",
|
|
i);
|
|
goto err_out;
|
|
}
|
|
}
|
|
|
|
if (!idpf_is_queue_model_split(vport->rxq_model))
|
|
continue;
|
|
|
|
for (j = 0; j < vport->num_bufqs_per_qgrp; j++) {
|
|
struct idpf_buf_queue *q;
|
|
|
|
q = &rx_qgrp->splitq.bufq_sets[j].bufq;
|
|
|
|
err = idpf_bufq_desc_alloc(vport, q);
|
|
if (err) {
|
|
pci_err(vport->adapter->pdev,
|
|
"Memory allocation for Rx Buffer Queue %u failed\n",
|
|
i);
|
|
goto err_out;
|
|
}
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_out:
|
|
idpf_rx_desc_rel_all(vport);
|
|
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* idpf_txq_group_rel - Release all resources for txq groups
|
|
* @vport: vport to release txq groups on
|
|
*/
|
|
static void idpf_txq_group_rel(struct idpf_vport *vport)
|
|
{
|
|
bool split, flow_sch_en;
|
|
int i, j;
|
|
|
|
if (!vport->txq_grps)
|
|
return;
|
|
|
|
split = idpf_is_queue_model_split(vport->txq_model);
|
|
flow_sch_en = !idpf_is_cap_ena(vport->adapter, IDPF_OTHER_CAPS,
|
|
VIRTCHNL2_CAP_SPLITQ_QSCHED);
|
|
|
|
for (i = 0; i < vport->num_txq_grp; i++) {
|
|
struct idpf_txq_group *txq_grp = &vport->txq_grps[i];
|
|
|
|
for (j = 0; j < txq_grp->num_txq; j++) {
|
|
kfree(txq_grp->txqs[j]);
|
|
txq_grp->txqs[j] = NULL;
|
|
}
|
|
|
|
if (!split)
|
|
continue;
|
|
|
|
kfree(txq_grp->complq);
|
|
txq_grp->complq = NULL;
|
|
|
|
if (flow_sch_en)
|
|
kfree(txq_grp->stashes);
|
|
}
|
|
kfree(vport->txq_grps);
|
|
vport->txq_grps = NULL;
|
|
}
|
|
|
|
/**
|
|
* idpf_rxq_sw_queue_rel - Release software queue resources
|
|
* @rx_qgrp: rx queue group with software queues
|
|
*/
|
|
static void idpf_rxq_sw_queue_rel(struct idpf_rxq_group *rx_qgrp)
|
|
{
|
|
int i, j;
|
|
|
|
for (i = 0; i < rx_qgrp->vport->num_bufqs_per_qgrp; i++) {
|
|
struct idpf_bufq_set *bufq_set = &rx_qgrp->splitq.bufq_sets[i];
|
|
|
|
for (j = 0; j < bufq_set->num_refillqs; j++) {
|
|
kfree(bufq_set->refillqs[j].ring);
|
|
bufq_set->refillqs[j].ring = NULL;
|
|
}
|
|
kfree(bufq_set->refillqs);
|
|
bufq_set->refillqs = NULL;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* idpf_rxq_group_rel - Release all resources for rxq groups
|
|
* @vport: vport to release rxq groups on
|
|
*/
|
|
static void idpf_rxq_group_rel(struct idpf_vport *vport)
|
|
{
|
|
int i;
|
|
|
|
if (!vport->rxq_grps)
|
|
return;
|
|
|
|
for (i = 0; i < vport->num_rxq_grp; i++) {
|
|
struct idpf_rxq_group *rx_qgrp = &vport->rxq_grps[i];
|
|
u16 num_rxq;
|
|
int j;
|
|
|
|
if (idpf_is_queue_model_split(vport->rxq_model)) {
|
|
num_rxq = rx_qgrp->splitq.num_rxq_sets;
|
|
for (j = 0; j < num_rxq; j++) {
|
|
kfree(rx_qgrp->splitq.rxq_sets[j]);
|
|
rx_qgrp->splitq.rxq_sets[j] = NULL;
|
|
}
|
|
|
|
idpf_rxq_sw_queue_rel(rx_qgrp);
|
|
kfree(rx_qgrp->splitq.bufq_sets);
|
|
rx_qgrp->splitq.bufq_sets = NULL;
|
|
} else {
|
|
num_rxq = rx_qgrp->singleq.num_rxq;
|
|
for (j = 0; j < num_rxq; j++) {
|
|
kfree(rx_qgrp->singleq.rxqs[j]);
|
|
rx_qgrp->singleq.rxqs[j] = NULL;
|
|
}
|
|
}
|
|
}
|
|
kfree(vport->rxq_grps);
|
|
vport->rxq_grps = NULL;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_queue_grp_rel_all - Release all queue groups
|
|
* @vport: vport to release queue groups for
|
|
*/
|
|
static void idpf_vport_queue_grp_rel_all(struct idpf_vport *vport)
|
|
{
|
|
idpf_txq_group_rel(vport);
|
|
idpf_rxq_group_rel(vport);
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_queues_rel - Free memory for all queues
|
|
* @vport: virtual port
|
|
*
|
|
* Free the memory allocated for queues associated to a vport
|
|
*/
|
|
void idpf_vport_queues_rel(struct idpf_vport *vport)
|
|
{
|
|
idpf_tx_desc_rel_all(vport);
|
|
idpf_rx_desc_rel_all(vport);
|
|
idpf_vport_queue_grp_rel_all(vport);
|
|
|
|
kfree(vport->txqs);
|
|
vport->txqs = NULL;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_init_fast_path_txqs - Initialize fast path txq array
|
|
* @vport: vport to init txqs on
|
|
*
|
|
* We get a queue index from skb->queue_mapping and we need a fast way to
|
|
* dereference the queue from queue groups. This allows us to quickly pull a
|
|
* txq based on a queue index.
|
|
*
|
|
* Returns 0 on success, negative on failure
|
|
*/
|
|
static int idpf_vport_init_fast_path_txqs(struct idpf_vport *vport)
|
|
{
|
|
int i, j, k = 0;
|
|
|
|
vport->txqs = kcalloc(vport->num_txq, sizeof(*vport->txqs),
|
|
GFP_KERNEL);
|
|
|
|
if (!vport->txqs)
|
|
return -ENOMEM;
|
|
|
|
for (i = 0; i < vport->num_txq_grp; i++) {
|
|
struct idpf_txq_group *tx_grp = &vport->txq_grps[i];
|
|
|
|
for (j = 0; j < tx_grp->num_txq; j++, k++) {
|
|
vport->txqs[k] = tx_grp->txqs[j];
|
|
vport->txqs[k]->idx = k;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_init_num_qs - Initialize number of queues
|
|
* @vport: vport to initialize queues
|
|
* @vport_msg: data to be filled into vport
|
|
*/
|
|
void idpf_vport_init_num_qs(struct idpf_vport *vport,
|
|
struct virtchnl2_create_vport *vport_msg)
|
|
{
|
|
struct idpf_vport_user_config_data *config_data;
|
|
u16 idx = vport->idx;
|
|
|
|
config_data = &vport->adapter->vport_config[idx]->user_config;
|
|
vport->num_txq = le16_to_cpu(vport_msg->num_tx_q);
|
|
vport->num_rxq = le16_to_cpu(vport_msg->num_rx_q);
|
|
/* number of txqs and rxqs in config data will be zeros only in the
|
|
* driver load path and we dont update them there after
|
|
*/
|
|
if (!config_data->num_req_tx_qs && !config_data->num_req_rx_qs) {
|
|
config_data->num_req_tx_qs = le16_to_cpu(vport_msg->num_tx_q);
|
|
config_data->num_req_rx_qs = le16_to_cpu(vport_msg->num_rx_q);
|
|
}
|
|
|
|
if (idpf_is_queue_model_split(vport->txq_model))
|
|
vport->num_complq = le16_to_cpu(vport_msg->num_tx_complq);
|
|
if (idpf_is_queue_model_split(vport->rxq_model))
|
|
vport->num_bufq = le16_to_cpu(vport_msg->num_rx_bufq);
|
|
|
|
/* Adjust number of buffer queues per Rx queue group. */
|
|
if (!idpf_is_queue_model_split(vport->rxq_model)) {
|
|
vport->num_bufqs_per_qgrp = 0;
|
|
|
|
return;
|
|
}
|
|
|
|
vport->num_bufqs_per_qgrp = IDPF_MAX_BUFQS_PER_RXQ_GRP;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_calc_num_q_desc - Calculate number of queue groups
|
|
* @vport: vport to calculate q groups for
|
|
*/
|
|
void idpf_vport_calc_num_q_desc(struct idpf_vport *vport)
|
|
{
|
|
struct idpf_vport_user_config_data *config_data;
|
|
int num_bufqs = vport->num_bufqs_per_qgrp;
|
|
u32 num_req_txq_desc, num_req_rxq_desc;
|
|
u16 idx = vport->idx;
|
|
int i;
|
|
|
|
config_data = &vport->adapter->vport_config[idx]->user_config;
|
|
num_req_txq_desc = config_data->num_req_txq_desc;
|
|
num_req_rxq_desc = config_data->num_req_rxq_desc;
|
|
|
|
vport->complq_desc_count = 0;
|
|
if (num_req_txq_desc) {
|
|
vport->txq_desc_count = num_req_txq_desc;
|
|
if (idpf_is_queue_model_split(vport->txq_model)) {
|
|
vport->complq_desc_count = num_req_txq_desc;
|
|
if (vport->complq_desc_count < IDPF_MIN_TXQ_COMPLQ_DESC)
|
|
vport->complq_desc_count =
|
|
IDPF_MIN_TXQ_COMPLQ_DESC;
|
|
}
|
|
} else {
|
|
vport->txq_desc_count = IDPF_DFLT_TX_Q_DESC_COUNT;
|
|
if (idpf_is_queue_model_split(vport->txq_model))
|
|
vport->complq_desc_count =
|
|
IDPF_DFLT_TX_COMPLQ_DESC_COUNT;
|
|
}
|
|
|
|
if (num_req_rxq_desc)
|
|
vport->rxq_desc_count = num_req_rxq_desc;
|
|
else
|
|
vport->rxq_desc_count = IDPF_DFLT_RX_Q_DESC_COUNT;
|
|
|
|
for (i = 0; i < num_bufqs; i++) {
|
|
if (!vport->bufq_desc_count[i])
|
|
vport->bufq_desc_count[i] =
|
|
IDPF_RX_BUFQ_DESC_COUNT(vport->rxq_desc_count,
|
|
num_bufqs);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_calc_total_qs - Calculate total number of queues
|
|
* @adapter: private data struct
|
|
* @vport_idx: vport idx to retrieve vport pointer
|
|
* @vport_msg: message to fill with data
|
|
* @max_q: vport max queue info
|
|
*
|
|
* Return 0 on success, error value on failure.
|
|
*/
|
|
int idpf_vport_calc_total_qs(struct idpf_adapter *adapter, u16 vport_idx,
|
|
struct virtchnl2_create_vport *vport_msg,
|
|
struct idpf_vport_max_q *max_q)
|
|
{
|
|
int dflt_splitq_txq_grps = 0, dflt_singleq_txqs = 0;
|
|
int dflt_splitq_rxq_grps = 0, dflt_singleq_rxqs = 0;
|
|
u16 num_req_tx_qs = 0, num_req_rx_qs = 0;
|
|
struct idpf_vport_config *vport_config;
|
|
u16 num_txq_grps, num_rxq_grps;
|
|
u32 num_qs;
|
|
|
|
vport_config = adapter->vport_config[vport_idx];
|
|
if (vport_config) {
|
|
num_req_tx_qs = vport_config->user_config.num_req_tx_qs;
|
|
num_req_rx_qs = vport_config->user_config.num_req_rx_qs;
|
|
} else {
|
|
int num_cpus;
|
|
|
|
/* Restrict num of queues to cpus online as a default
|
|
* configuration to give best performance. User can always
|
|
* override to a max number of queues via ethtool.
|
|
*/
|
|
num_cpus = num_online_cpus();
|
|
|
|
dflt_splitq_txq_grps = min_t(int, max_q->max_txq, num_cpus);
|
|
dflt_singleq_txqs = min_t(int, max_q->max_txq, num_cpus);
|
|
dflt_splitq_rxq_grps = min_t(int, max_q->max_rxq, num_cpus);
|
|
dflt_singleq_rxqs = min_t(int, max_q->max_rxq, num_cpus);
|
|
}
|
|
|
|
if (idpf_is_queue_model_split(le16_to_cpu(vport_msg->txq_model))) {
|
|
num_txq_grps = num_req_tx_qs ? num_req_tx_qs : dflt_splitq_txq_grps;
|
|
vport_msg->num_tx_complq = cpu_to_le16(num_txq_grps *
|
|
IDPF_COMPLQ_PER_GROUP);
|
|
vport_msg->num_tx_q = cpu_to_le16(num_txq_grps *
|
|
IDPF_DFLT_SPLITQ_TXQ_PER_GROUP);
|
|
} else {
|
|
num_txq_grps = IDPF_DFLT_SINGLEQ_TX_Q_GROUPS;
|
|
num_qs = num_txq_grps * (num_req_tx_qs ? num_req_tx_qs :
|
|
dflt_singleq_txqs);
|
|
vport_msg->num_tx_q = cpu_to_le16(num_qs);
|
|
vport_msg->num_tx_complq = 0;
|
|
}
|
|
if (idpf_is_queue_model_split(le16_to_cpu(vport_msg->rxq_model))) {
|
|
num_rxq_grps = num_req_rx_qs ? num_req_rx_qs : dflt_splitq_rxq_grps;
|
|
vport_msg->num_rx_bufq = cpu_to_le16(num_rxq_grps *
|
|
IDPF_MAX_BUFQS_PER_RXQ_GRP);
|
|
vport_msg->num_rx_q = cpu_to_le16(num_rxq_grps *
|
|
IDPF_DFLT_SPLITQ_RXQ_PER_GROUP);
|
|
} else {
|
|
num_rxq_grps = IDPF_DFLT_SINGLEQ_RX_Q_GROUPS;
|
|
num_qs = num_rxq_grps * (num_req_rx_qs ? num_req_rx_qs :
|
|
dflt_singleq_rxqs);
|
|
vport_msg->num_rx_q = cpu_to_le16(num_qs);
|
|
vport_msg->num_rx_bufq = 0;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_calc_num_q_groups - Calculate number of queue groups
|
|
* @vport: vport to calculate q groups for
|
|
*/
|
|
void idpf_vport_calc_num_q_groups(struct idpf_vport *vport)
|
|
{
|
|
if (idpf_is_queue_model_split(vport->txq_model))
|
|
vport->num_txq_grp = vport->num_txq;
|
|
else
|
|
vport->num_txq_grp = IDPF_DFLT_SINGLEQ_TX_Q_GROUPS;
|
|
|
|
if (idpf_is_queue_model_split(vport->rxq_model))
|
|
vport->num_rxq_grp = vport->num_rxq;
|
|
else
|
|
vport->num_rxq_grp = IDPF_DFLT_SINGLEQ_RX_Q_GROUPS;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_calc_numq_per_grp - Calculate number of queues per group
|
|
* @vport: vport to calculate queues for
|
|
* @num_txq: return parameter for number of TX queues
|
|
* @num_rxq: return parameter for number of RX queues
|
|
*/
|
|
static void idpf_vport_calc_numq_per_grp(struct idpf_vport *vport,
|
|
u16 *num_txq, u16 *num_rxq)
|
|
{
|
|
if (idpf_is_queue_model_split(vport->txq_model))
|
|
*num_txq = IDPF_DFLT_SPLITQ_TXQ_PER_GROUP;
|
|
else
|
|
*num_txq = vport->num_txq;
|
|
|
|
if (idpf_is_queue_model_split(vport->rxq_model))
|
|
*num_rxq = IDPF_DFLT_SPLITQ_RXQ_PER_GROUP;
|
|
else
|
|
*num_rxq = vport->num_rxq;
|
|
}
|
|
|
|
/**
|
|
* idpf_rxq_set_descids - set the descids supported by this queue
|
|
* @vport: virtual port data structure
|
|
* @q: rx queue for which descids are set
|
|
*
|
|
*/
|
|
static void idpf_rxq_set_descids(const struct idpf_vport *vport,
|
|
struct idpf_rx_queue *q)
|
|
{
|
|
if (idpf_is_queue_model_split(vport->rxq_model)) {
|
|
q->rxdids = VIRTCHNL2_RXDID_2_FLEX_SPLITQ_M;
|
|
} else {
|
|
if (vport->base_rxd)
|
|
q->rxdids = VIRTCHNL2_RXDID_1_32B_BASE_M;
|
|
else
|
|
q->rxdids = VIRTCHNL2_RXDID_2_FLEX_SQ_NIC_M;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* idpf_txq_group_alloc - Allocate all txq group resources
|
|
* @vport: vport to allocate txq groups for
|
|
* @num_txq: number of txqs to allocate for each group
|
|
*
|
|
* Returns 0 on success, negative on failure
|
|
*/
|
|
static int idpf_txq_group_alloc(struct idpf_vport *vport, u16 num_txq)
|
|
{
|
|
bool split, flow_sch_en;
|
|
int i;
|
|
|
|
vport->txq_grps = kcalloc(vport->num_txq_grp,
|
|
sizeof(*vport->txq_grps), GFP_KERNEL);
|
|
if (!vport->txq_grps)
|
|
return -ENOMEM;
|
|
|
|
split = idpf_is_queue_model_split(vport->txq_model);
|
|
flow_sch_en = !idpf_is_cap_ena(vport->adapter, IDPF_OTHER_CAPS,
|
|
VIRTCHNL2_CAP_SPLITQ_QSCHED);
|
|
|
|
for (i = 0; i < vport->num_txq_grp; i++) {
|
|
struct idpf_txq_group *tx_qgrp = &vport->txq_grps[i];
|
|
struct idpf_adapter *adapter = vport->adapter;
|
|
struct idpf_txq_stash *stashes;
|
|
int j;
|
|
|
|
tx_qgrp->vport = vport;
|
|
tx_qgrp->num_txq = num_txq;
|
|
|
|
for (j = 0; j < tx_qgrp->num_txq; j++) {
|
|
tx_qgrp->txqs[j] = kzalloc(sizeof(*tx_qgrp->txqs[j]),
|
|
GFP_KERNEL);
|
|
if (!tx_qgrp->txqs[j])
|
|
goto err_alloc;
|
|
}
|
|
|
|
if (split && flow_sch_en) {
|
|
stashes = kcalloc(num_txq, sizeof(*stashes),
|
|
GFP_KERNEL);
|
|
if (!stashes)
|
|
goto err_alloc;
|
|
|
|
tx_qgrp->stashes = stashes;
|
|
}
|
|
|
|
for (j = 0; j < tx_qgrp->num_txq; j++) {
|
|
struct idpf_tx_queue *q = tx_qgrp->txqs[j];
|
|
|
|
q->dev = &adapter->pdev->dev;
|
|
q->desc_count = vport->txq_desc_count;
|
|
q->tx_max_bufs = idpf_get_max_tx_bufs(adapter);
|
|
q->tx_min_pkt_len = idpf_get_min_tx_pkt_len(adapter);
|
|
q->netdev = vport->netdev;
|
|
q->txq_grp = tx_qgrp;
|
|
|
|
if (!split) {
|
|
q->clean_budget = vport->compln_clean_budget;
|
|
idpf_queue_assign(CRC_EN, q,
|
|
vport->crc_enable);
|
|
}
|
|
|
|
if (!flow_sch_en)
|
|
continue;
|
|
|
|
if (split) {
|
|
q->stash = &stashes[j];
|
|
hash_init(q->stash->sched_buf_hash);
|
|
}
|
|
|
|
idpf_queue_set(FLOW_SCH_EN, q);
|
|
}
|
|
|
|
if (!split)
|
|
continue;
|
|
|
|
tx_qgrp->complq = kcalloc(IDPF_COMPLQ_PER_GROUP,
|
|
sizeof(*tx_qgrp->complq),
|
|
GFP_KERNEL);
|
|
if (!tx_qgrp->complq)
|
|
goto err_alloc;
|
|
|
|
tx_qgrp->complq->desc_count = vport->complq_desc_count;
|
|
tx_qgrp->complq->txq_grp = tx_qgrp;
|
|
tx_qgrp->complq->netdev = vport->netdev;
|
|
tx_qgrp->complq->clean_budget = vport->compln_clean_budget;
|
|
|
|
if (flow_sch_en)
|
|
idpf_queue_set(FLOW_SCH_EN, tx_qgrp->complq);
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_alloc:
|
|
idpf_txq_group_rel(vport);
|
|
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/**
|
|
* idpf_rxq_group_alloc - Allocate all rxq group resources
|
|
* @vport: vport to allocate rxq groups for
|
|
* @num_rxq: number of rxqs to allocate for each group
|
|
*
|
|
* Returns 0 on success, negative on failure
|
|
*/
|
|
static int idpf_rxq_group_alloc(struct idpf_vport *vport, u16 num_rxq)
|
|
{
|
|
int i, k, err = 0;
|
|
bool hs;
|
|
|
|
vport->rxq_grps = kcalloc(vport->num_rxq_grp,
|
|
sizeof(struct idpf_rxq_group), GFP_KERNEL);
|
|
if (!vport->rxq_grps)
|
|
return -ENOMEM;
|
|
|
|
hs = idpf_vport_get_hsplit(vport) == ETHTOOL_TCP_DATA_SPLIT_ENABLED;
|
|
|
|
for (i = 0; i < vport->num_rxq_grp; i++) {
|
|
struct idpf_rxq_group *rx_qgrp = &vport->rxq_grps[i];
|
|
int j;
|
|
|
|
rx_qgrp->vport = vport;
|
|
if (!idpf_is_queue_model_split(vport->rxq_model)) {
|
|
rx_qgrp->singleq.num_rxq = num_rxq;
|
|
for (j = 0; j < num_rxq; j++) {
|
|
rx_qgrp->singleq.rxqs[j] =
|
|
kzalloc(sizeof(*rx_qgrp->singleq.rxqs[j]),
|
|
GFP_KERNEL);
|
|
if (!rx_qgrp->singleq.rxqs[j]) {
|
|
err = -ENOMEM;
|
|
goto err_alloc;
|
|
}
|
|
}
|
|
goto skip_splitq_rx_init;
|
|
}
|
|
rx_qgrp->splitq.num_rxq_sets = num_rxq;
|
|
|
|
for (j = 0; j < num_rxq; j++) {
|
|
rx_qgrp->splitq.rxq_sets[j] =
|
|
kzalloc(sizeof(struct idpf_rxq_set),
|
|
GFP_KERNEL);
|
|
if (!rx_qgrp->splitq.rxq_sets[j]) {
|
|
err = -ENOMEM;
|
|
goto err_alloc;
|
|
}
|
|
}
|
|
|
|
rx_qgrp->splitq.bufq_sets = kcalloc(vport->num_bufqs_per_qgrp,
|
|
sizeof(struct idpf_bufq_set),
|
|
GFP_KERNEL);
|
|
if (!rx_qgrp->splitq.bufq_sets) {
|
|
err = -ENOMEM;
|
|
goto err_alloc;
|
|
}
|
|
|
|
for (j = 0; j < vport->num_bufqs_per_qgrp; j++) {
|
|
struct idpf_bufq_set *bufq_set =
|
|
&rx_qgrp->splitq.bufq_sets[j];
|
|
int swq_size = sizeof(struct idpf_sw_queue);
|
|
struct idpf_buf_queue *q;
|
|
|
|
q = &rx_qgrp->splitq.bufq_sets[j].bufq;
|
|
q->desc_count = vport->bufq_desc_count[j];
|
|
q->rx_buffer_low_watermark = IDPF_LOW_WATERMARK;
|
|
|
|
idpf_queue_assign(HSPLIT_EN, q, hs);
|
|
|
|
bufq_set->num_refillqs = num_rxq;
|
|
bufq_set->refillqs = kcalloc(num_rxq, swq_size,
|
|
GFP_KERNEL);
|
|
if (!bufq_set->refillqs) {
|
|
err = -ENOMEM;
|
|
goto err_alloc;
|
|
}
|
|
for (k = 0; k < bufq_set->num_refillqs; k++) {
|
|
struct idpf_sw_queue *refillq =
|
|
&bufq_set->refillqs[k];
|
|
|
|
refillq->desc_count =
|
|
vport->bufq_desc_count[j];
|
|
idpf_queue_set(GEN_CHK, refillq);
|
|
idpf_queue_set(RFL_GEN_CHK, refillq);
|
|
refillq->ring = kcalloc(refillq->desc_count,
|
|
sizeof(*refillq->ring),
|
|
GFP_KERNEL);
|
|
if (!refillq->ring) {
|
|
err = -ENOMEM;
|
|
goto err_alloc;
|
|
}
|
|
}
|
|
}
|
|
|
|
skip_splitq_rx_init:
|
|
for (j = 0; j < num_rxq; j++) {
|
|
struct idpf_rx_queue *q;
|
|
|
|
if (!idpf_is_queue_model_split(vport->rxq_model)) {
|
|
q = rx_qgrp->singleq.rxqs[j];
|
|
goto setup_rxq;
|
|
}
|
|
q = &rx_qgrp->splitq.rxq_sets[j]->rxq;
|
|
rx_qgrp->splitq.rxq_sets[j]->refillq[0] =
|
|
&rx_qgrp->splitq.bufq_sets[0].refillqs[j];
|
|
if (vport->num_bufqs_per_qgrp > IDPF_SINGLE_BUFQ_PER_RXQ_GRP)
|
|
rx_qgrp->splitq.rxq_sets[j]->refillq[1] =
|
|
&rx_qgrp->splitq.bufq_sets[1].refillqs[j];
|
|
|
|
idpf_queue_assign(HSPLIT_EN, q, hs);
|
|
|
|
setup_rxq:
|
|
q->desc_count = vport->rxq_desc_count;
|
|
q->rx_ptype_lkup = vport->rx_ptype_lkup;
|
|
q->netdev = vport->netdev;
|
|
q->bufq_sets = rx_qgrp->splitq.bufq_sets;
|
|
q->idx = (i * num_rxq) + j;
|
|
q->rx_buffer_low_watermark = IDPF_LOW_WATERMARK;
|
|
q->rx_max_pkt_size = vport->netdev->mtu +
|
|
LIBETH_RX_LL_LEN;
|
|
idpf_rxq_set_descids(vport, q);
|
|
}
|
|
}
|
|
|
|
err_alloc:
|
|
if (err)
|
|
idpf_rxq_group_rel(vport);
|
|
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_queue_grp_alloc_all - Allocate all queue groups/resources
|
|
* @vport: vport with qgrps to allocate
|
|
*
|
|
* Returns 0 on success, negative on failure
|
|
*/
|
|
static int idpf_vport_queue_grp_alloc_all(struct idpf_vport *vport)
|
|
{
|
|
u16 num_txq, num_rxq;
|
|
int err;
|
|
|
|
idpf_vport_calc_numq_per_grp(vport, &num_txq, &num_rxq);
|
|
|
|
err = idpf_txq_group_alloc(vport, num_txq);
|
|
if (err)
|
|
goto err_out;
|
|
|
|
err = idpf_rxq_group_alloc(vport, num_rxq);
|
|
if (err)
|
|
goto err_out;
|
|
|
|
return 0;
|
|
|
|
err_out:
|
|
idpf_vport_queue_grp_rel_all(vport);
|
|
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_queues_alloc - Allocate memory for all queues
|
|
* @vport: virtual port
|
|
*
|
|
* Allocate memory for queues associated with a vport. Returns 0 on success,
|
|
* negative on failure.
|
|
*/
|
|
int idpf_vport_queues_alloc(struct idpf_vport *vport)
|
|
{
|
|
int err;
|
|
|
|
err = idpf_vport_queue_grp_alloc_all(vport);
|
|
if (err)
|
|
goto err_out;
|
|
|
|
err = idpf_tx_desc_alloc_all(vport);
|
|
if (err)
|
|
goto err_out;
|
|
|
|
err = idpf_rx_desc_alloc_all(vport);
|
|
if (err)
|
|
goto err_out;
|
|
|
|
err = idpf_vport_init_fast_path_txqs(vport);
|
|
if (err)
|
|
goto err_out;
|
|
|
|
return 0;
|
|
|
|
err_out:
|
|
idpf_vport_queues_rel(vport);
|
|
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_handle_sw_marker - Handle queue marker packet
|
|
* @tx_q: tx queue to handle software marker
|
|
*/
|
|
static void idpf_tx_handle_sw_marker(struct idpf_tx_queue *tx_q)
|
|
{
|
|
struct idpf_netdev_priv *priv = netdev_priv(tx_q->netdev);
|
|
struct idpf_vport *vport = priv->vport;
|
|
int i;
|
|
|
|
idpf_queue_clear(SW_MARKER, tx_q);
|
|
/* Hardware must write marker packets to all queues associated with
|
|
* completion queues. So check if all queues received marker packets
|
|
*/
|
|
for (i = 0; i < vport->num_txq; i++)
|
|
/* If we're still waiting on any other TXQ marker completions,
|
|
* just return now since we cannot wake up the marker_wq yet.
|
|
*/
|
|
if (idpf_queue_has(SW_MARKER, vport->txqs[i]))
|
|
return;
|
|
|
|
/* Drain complete */
|
|
set_bit(IDPF_VPORT_SW_MARKER, vport->flags);
|
|
wake_up(&vport->sw_marker_wq);
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_clean_stashed_bufs - clean bufs that were stored for
|
|
* out of order completions
|
|
* @txq: queue to clean
|
|
* @compl_tag: completion tag of packet to clean (from completion descriptor)
|
|
* @cleaned: pointer to stats struct to track cleaned packets/bytes
|
|
* @budget: Used to determine if we are in netpoll
|
|
*/
|
|
static void idpf_tx_clean_stashed_bufs(struct idpf_tx_queue *txq,
|
|
u16 compl_tag,
|
|
struct libeth_sq_napi_stats *cleaned,
|
|
int budget)
|
|
{
|
|
struct idpf_tx_stash *stash;
|
|
struct hlist_node *tmp_buf;
|
|
struct libeth_cq_pp cp = {
|
|
.dev = txq->dev,
|
|
.ss = cleaned,
|
|
.napi = budget,
|
|
};
|
|
|
|
/* Buffer completion */
|
|
hash_for_each_possible_safe(txq->stash->sched_buf_hash, stash, tmp_buf,
|
|
hlist, compl_tag) {
|
|
if (unlikely(idpf_tx_buf_compl_tag(&stash->buf) != compl_tag))
|
|
continue;
|
|
|
|
hash_del(&stash->hlist);
|
|
libeth_tx_complete(&stash->buf, &cp);
|
|
|
|
/* Push shadow buf back onto stack */
|
|
idpf_buf_lifo_push(&txq->stash->buf_stack, stash);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* idpf_stash_flow_sch_buffers - store buffer parameters info to be freed at a
|
|
* later time (only relevant for flow scheduling mode)
|
|
* @txq: Tx queue to clean
|
|
* @tx_buf: buffer to store
|
|
*/
|
|
static int idpf_stash_flow_sch_buffers(struct idpf_tx_queue *txq,
|
|
struct idpf_tx_buf *tx_buf)
|
|
{
|
|
struct idpf_tx_stash *stash;
|
|
|
|
if (unlikely(tx_buf->type <= LIBETH_SQE_CTX))
|
|
return 0;
|
|
|
|
stash = idpf_buf_lifo_pop(&txq->stash->buf_stack);
|
|
if (unlikely(!stash)) {
|
|
net_err_ratelimited("%s: No out-of-order TX buffers left!\n",
|
|
netdev_name(txq->netdev));
|
|
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* Store buffer params in shadow buffer */
|
|
stash->buf.skb = tx_buf->skb;
|
|
stash->buf.bytes = tx_buf->bytes;
|
|
stash->buf.packets = tx_buf->packets;
|
|
stash->buf.type = tx_buf->type;
|
|
stash->buf.nr_frags = tx_buf->nr_frags;
|
|
dma_unmap_addr_set(&stash->buf, dma, dma_unmap_addr(tx_buf, dma));
|
|
dma_unmap_len_set(&stash->buf, len, dma_unmap_len(tx_buf, len));
|
|
idpf_tx_buf_compl_tag(&stash->buf) = idpf_tx_buf_compl_tag(tx_buf);
|
|
|
|
/* Add buffer to buf_hash table to be freed later */
|
|
hash_add(txq->stash->sched_buf_hash, &stash->hlist,
|
|
idpf_tx_buf_compl_tag(&stash->buf));
|
|
|
|
tx_buf->type = LIBETH_SQE_EMPTY;
|
|
|
|
return 0;
|
|
}
|
|
|
|
#define idpf_tx_splitq_clean_bump_ntc(txq, ntc, desc, buf) \
|
|
do { \
|
|
if (unlikely(++(ntc) == (txq)->desc_count)) { \
|
|
ntc = 0; \
|
|
buf = (txq)->tx_buf; \
|
|
desc = &(txq)->flex_tx[0]; \
|
|
} else { \
|
|
(buf)++; \
|
|
(desc)++; \
|
|
} \
|
|
} while (0)
|
|
|
|
/**
|
|
* idpf_tx_splitq_clean - Reclaim resources from buffer queue
|
|
* @tx_q: Tx queue to clean
|
|
* @end: queue index until which it should be cleaned
|
|
* @napi_budget: Used to determine if we are in netpoll
|
|
* @cleaned: pointer to stats struct to track cleaned packets/bytes
|
|
* @descs_only: true if queue is using flow-based scheduling and should
|
|
* not clean buffers at this time
|
|
*
|
|
* Cleans the queue descriptor ring. If the queue is using queue-based
|
|
* scheduling, the buffers will be cleaned as well. If the queue is using
|
|
* flow-based scheduling, only the descriptors are cleaned at this time.
|
|
* Separate packet completion events will be reported on the completion queue,
|
|
* and the buffers will be cleaned separately. The stats are not updated from
|
|
* this function when using flow-based scheduling.
|
|
*
|
|
* Furthermore, in flow scheduling mode, check to make sure there are enough
|
|
* reserve buffers to stash the packet. If there are not, return early, which
|
|
* will leave next_to_clean pointing to the packet that failed to be stashed.
|
|
*
|
|
* Return: false in the scenario above, true otherwise.
|
|
*/
|
|
static bool idpf_tx_splitq_clean(struct idpf_tx_queue *tx_q, u16 end,
|
|
int napi_budget,
|
|
struct libeth_sq_napi_stats *cleaned,
|
|
bool descs_only)
|
|
{
|
|
union idpf_tx_flex_desc *next_pending_desc = NULL;
|
|
union idpf_tx_flex_desc *tx_desc;
|
|
u32 ntc = tx_q->next_to_clean;
|
|
struct libeth_cq_pp cp = {
|
|
.dev = tx_q->dev,
|
|
.ss = cleaned,
|
|
.napi = napi_budget,
|
|
};
|
|
struct idpf_tx_buf *tx_buf;
|
|
bool clean_complete = true;
|
|
|
|
tx_desc = &tx_q->flex_tx[ntc];
|
|
next_pending_desc = &tx_q->flex_tx[end];
|
|
tx_buf = &tx_q->tx_buf[ntc];
|
|
|
|
while (tx_desc != next_pending_desc) {
|
|
u32 eop_idx;
|
|
|
|
/* If this entry in the ring was used as a context descriptor,
|
|
* it's corresponding entry in the buffer ring is reserved. We
|
|
* can skip this descriptor since there is no buffer to clean.
|
|
*/
|
|
if (tx_buf->type <= LIBETH_SQE_CTX)
|
|
goto fetch_next_txq_desc;
|
|
|
|
if (unlikely(tx_buf->type != LIBETH_SQE_SKB))
|
|
break;
|
|
|
|
eop_idx = tx_buf->rs_idx;
|
|
|
|
if (descs_only) {
|
|
if (IDPF_TX_BUF_RSV_UNUSED(tx_q) < tx_buf->nr_frags) {
|
|
clean_complete = false;
|
|
goto tx_splitq_clean_out;
|
|
}
|
|
|
|
idpf_stash_flow_sch_buffers(tx_q, tx_buf);
|
|
|
|
while (ntc != eop_idx) {
|
|
idpf_tx_splitq_clean_bump_ntc(tx_q, ntc,
|
|
tx_desc, tx_buf);
|
|
idpf_stash_flow_sch_buffers(tx_q, tx_buf);
|
|
}
|
|
} else {
|
|
libeth_tx_complete(tx_buf, &cp);
|
|
|
|
/* unmap remaining buffers */
|
|
while (ntc != eop_idx) {
|
|
idpf_tx_splitq_clean_bump_ntc(tx_q, ntc,
|
|
tx_desc, tx_buf);
|
|
|
|
/* unmap any remaining paged data */
|
|
libeth_tx_complete(tx_buf, &cp);
|
|
}
|
|
}
|
|
|
|
fetch_next_txq_desc:
|
|
idpf_tx_splitq_clean_bump_ntc(tx_q, ntc, tx_desc, tx_buf);
|
|
}
|
|
|
|
tx_splitq_clean_out:
|
|
tx_q->next_to_clean = ntc;
|
|
|
|
return clean_complete;
|
|
}
|
|
|
|
#define idpf_tx_clean_buf_ring_bump_ntc(txq, ntc, buf) \
|
|
do { \
|
|
(buf)++; \
|
|
(ntc)++; \
|
|
if (unlikely((ntc) == (txq)->desc_count)) { \
|
|
buf = (txq)->tx_buf; \
|
|
ntc = 0; \
|
|
} \
|
|
} while (0)
|
|
|
|
/**
|
|
* idpf_tx_clean_buf_ring - clean flow scheduling TX queue buffers
|
|
* @txq: queue to clean
|
|
* @compl_tag: completion tag of packet to clean (from completion descriptor)
|
|
* @cleaned: pointer to stats struct to track cleaned packets/bytes
|
|
* @budget: Used to determine if we are in netpoll
|
|
*
|
|
* Cleans all buffers associated with the input completion tag either from the
|
|
* TX buffer ring or from the hash table if the buffers were previously
|
|
* stashed. Returns the byte/segment count for the cleaned packet associated
|
|
* this completion tag.
|
|
*/
|
|
static bool idpf_tx_clean_buf_ring(struct idpf_tx_queue *txq, u16 compl_tag,
|
|
struct libeth_sq_napi_stats *cleaned,
|
|
int budget)
|
|
{
|
|
u16 idx = compl_tag & txq->compl_tag_bufid_m;
|
|
struct idpf_tx_buf *tx_buf = NULL;
|
|
struct libeth_cq_pp cp = {
|
|
.dev = txq->dev,
|
|
.ss = cleaned,
|
|
.napi = budget,
|
|
};
|
|
u16 ntc, orig_idx = idx;
|
|
|
|
tx_buf = &txq->tx_buf[idx];
|
|
|
|
if (unlikely(tx_buf->type <= LIBETH_SQE_CTX ||
|
|
idpf_tx_buf_compl_tag(tx_buf) != compl_tag))
|
|
return false;
|
|
|
|
if (tx_buf->type == LIBETH_SQE_SKB)
|
|
libeth_tx_complete(tx_buf, &cp);
|
|
|
|
idpf_tx_clean_buf_ring_bump_ntc(txq, idx, tx_buf);
|
|
|
|
while (idpf_tx_buf_compl_tag(tx_buf) == compl_tag) {
|
|
libeth_tx_complete(tx_buf, &cp);
|
|
idpf_tx_clean_buf_ring_bump_ntc(txq, idx, tx_buf);
|
|
}
|
|
|
|
/*
|
|
* It's possible the packet we just cleaned was an out of order
|
|
* completion, which means we can stash the buffers starting from
|
|
* the original next_to_clean and reuse the descriptors. We need
|
|
* to compare the descriptor ring next_to_clean packet's "first" buffer
|
|
* to the "first" buffer of the packet we just cleaned to determine if
|
|
* this is the case. Howevever, next_to_clean can point to either a
|
|
* reserved buffer that corresponds to a context descriptor used for the
|
|
* next_to_clean packet (TSO packet) or the "first" buffer (single
|
|
* packet). The orig_idx from the packet we just cleaned will always
|
|
* point to the "first" buffer. If next_to_clean points to a reserved
|
|
* buffer, let's bump ntc once and start the comparison from there.
|
|
*/
|
|
ntc = txq->next_to_clean;
|
|
tx_buf = &txq->tx_buf[ntc];
|
|
|
|
if (tx_buf->type == LIBETH_SQE_CTX)
|
|
idpf_tx_clean_buf_ring_bump_ntc(txq, ntc, tx_buf);
|
|
|
|
/*
|
|
* If ntc still points to a different "first" buffer, clean the
|
|
* descriptor ring and stash all of the buffers for later cleaning. If
|
|
* we cannot stash all of the buffers, next_to_clean will point to the
|
|
* "first" buffer of the packet that could not be stashed and cleaning
|
|
* will start there next time.
|
|
*/
|
|
if (unlikely(tx_buf != &txq->tx_buf[orig_idx] &&
|
|
!idpf_tx_splitq_clean(txq, orig_idx, budget, cleaned,
|
|
true)))
|
|
return true;
|
|
|
|
/*
|
|
* Otherwise, update next_to_clean to reflect the cleaning that was
|
|
* done above.
|
|
*/
|
|
txq->next_to_clean = idx;
|
|
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_handle_rs_completion - clean a single packet and all of its buffers
|
|
* whether on the buffer ring or in the hash table
|
|
* @txq: Tx ring to clean
|
|
* @desc: pointer to completion queue descriptor to extract completion
|
|
* information from
|
|
* @cleaned: pointer to stats struct to track cleaned packets/bytes
|
|
* @budget: Used to determine if we are in netpoll
|
|
*
|
|
* Returns bytes/packets cleaned
|
|
*/
|
|
static void idpf_tx_handle_rs_completion(struct idpf_tx_queue *txq,
|
|
struct idpf_splitq_tx_compl_desc *desc,
|
|
struct libeth_sq_napi_stats *cleaned,
|
|
int budget)
|
|
{
|
|
u16 compl_tag;
|
|
|
|
if (!idpf_queue_has(FLOW_SCH_EN, txq)) {
|
|
u16 head = le16_to_cpu(desc->q_head_compl_tag.q_head);
|
|
|
|
idpf_tx_splitq_clean(txq, head, budget, cleaned, false);
|
|
return;
|
|
}
|
|
|
|
compl_tag = le16_to_cpu(desc->q_head_compl_tag.compl_tag);
|
|
|
|
/* If we didn't clean anything on the ring, this packet must be
|
|
* in the hash table. Go clean it there.
|
|
*/
|
|
if (!idpf_tx_clean_buf_ring(txq, compl_tag, cleaned, budget))
|
|
idpf_tx_clean_stashed_bufs(txq, compl_tag, cleaned, budget);
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_clean_complq - Reclaim resources on completion queue
|
|
* @complq: Tx ring to clean
|
|
* @budget: Used to determine if we are in netpoll
|
|
* @cleaned: returns number of packets cleaned
|
|
*
|
|
* Returns true if there's any budget left (e.g. the clean is finished)
|
|
*/
|
|
static bool idpf_tx_clean_complq(struct idpf_compl_queue *complq, int budget,
|
|
int *cleaned)
|
|
{
|
|
struct idpf_splitq_tx_compl_desc *tx_desc;
|
|
s16 ntc = complq->next_to_clean;
|
|
struct idpf_netdev_priv *np;
|
|
unsigned int complq_budget;
|
|
bool complq_ok = true;
|
|
int i;
|
|
|
|
complq_budget = complq->clean_budget;
|
|
tx_desc = &complq->comp[ntc];
|
|
ntc -= complq->desc_count;
|
|
|
|
do {
|
|
struct libeth_sq_napi_stats cleaned_stats = { };
|
|
struct idpf_tx_queue *tx_q;
|
|
int rel_tx_qid;
|
|
u16 hw_head;
|
|
u8 ctype; /* completion type */
|
|
u16 gen;
|
|
|
|
/* if the descriptor isn't done, no work yet to do */
|
|
gen = le16_get_bits(tx_desc->qid_comptype_gen,
|
|
IDPF_TXD_COMPLQ_GEN_M);
|
|
if (idpf_queue_has(GEN_CHK, complq) != gen)
|
|
break;
|
|
|
|
/* Find necessary info of TX queue to clean buffers */
|
|
rel_tx_qid = le16_get_bits(tx_desc->qid_comptype_gen,
|
|
IDPF_TXD_COMPLQ_QID_M);
|
|
if (rel_tx_qid >= complq->txq_grp->num_txq ||
|
|
!complq->txq_grp->txqs[rel_tx_qid]) {
|
|
netdev_err(complq->netdev, "TxQ not found\n");
|
|
goto fetch_next_desc;
|
|
}
|
|
tx_q = complq->txq_grp->txqs[rel_tx_qid];
|
|
|
|
/* Determine completion type */
|
|
ctype = le16_get_bits(tx_desc->qid_comptype_gen,
|
|
IDPF_TXD_COMPLQ_COMPL_TYPE_M);
|
|
switch (ctype) {
|
|
case IDPF_TXD_COMPLT_RE:
|
|
hw_head = le16_to_cpu(tx_desc->q_head_compl_tag.q_head);
|
|
|
|
idpf_tx_splitq_clean(tx_q, hw_head, budget,
|
|
&cleaned_stats, true);
|
|
break;
|
|
case IDPF_TXD_COMPLT_RS:
|
|
idpf_tx_handle_rs_completion(tx_q, tx_desc,
|
|
&cleaned_stats, budget);
|
|
break;
|
|
case IDPF_TXD_COMPLT_SW_MARKER:
|
|
idpf_tx_handle_sw_marker(tx_q);
|
|
break;
|
|
default:
|
|
netdev_err(tx_q->netdev,
|
|
"Unknown TX completion type: %d\n", ctype);
|
|
goto fetch_next_desc;
|
|
}
|
|
|
|
u64_stats_update_begin(&tx_q->stats_sync);
|
|
u64_stats_add(&tx_q->q_stats.packets, cleaned_stats.packets);
|
|
u64_stats_add(&tx_q->q_stats.bytes, cleaned_stats.bytes);
|
|
tx_q->cleaned_pkts += cleaned_stats.packets;
|
|
tx_q->cleaned_bytes += cleaned_stats.bytes;
|
|
complq->num_completions++;
|
|
u64_stats_update_end(&tx_q->stats_sync);
|
|
|
|
fetch_next_desc:
|
|
tx_desc++;
|
|
ntc++;
|
|
if (unlikely(!ntc)) {
|
|
ntc -= complq->desc_count;
|
|
tx_desc = &complq->comp[0];
|
|
idpf_queue_change(GEN_CHK, complq);
|
|
}
|
|
|
|
prefetch(tx_desc);
|
|
|
|
/* update budget accounting */
|
|
complq_budget--;
|
|
} while (likely(complq_budget));
|
|
|
|
/* Store the state of the complq to be used later in deciding if a
|
|
* TXQ can be started again
|
|
*/
|
|
if (unlikely(IDPF_TX_COMPLQ_PENDING(complq->txq_grp) >
|
|
IDPF_TX_COMPLQ_OVERFLOW_THRESH(complq)))
|
|
complq_ok = false;
|
|
|
|
np = netdev_priv(complq->netdev);
|
|
for (i = 0; i < complq->txq_grp->num_txq; ++i) {
|
|
struct idpf_tx_queue *tx_q = complq->txq_grp->txqs[i];
|
|
struct netdev_queue *nq;
|
|
bool dont_wake;
|
|
|
|
/* We didn't clean anything on this queue, move along */
|
|
if (!tx_q->cleaned_bytes)
|
|
continue;
|
|
|
|
*cleaned += tx_q->cleaned_pkts;
|
|
|
|
/* Update BQL */
|
|
nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx);
|
|
|
|
dont_wake = !complq_ok || IDPF_TX_BUF_RSV_LOW(tx_q) ||
|
|
np->state != __IDPF_VPORT_UP ||
|
|
!netif_carrier_ok(tx_q->netdev);
|
|
/* Check if the TXQ needs to and can be restarted */
|
|
__netif_txq_completed_wake(nq, tx_q->cleaned_pkts, tx_q->cleaned_bytes,
|
|
IDPF_DESC_UNUSED(tx_q), IDPF_TX_WAKE_THRESH,
|
|
dont_wake);
|
|
|
|
/* Reset cleaned stats for the next time this queue is
|
|
* cleaned
|
|
*/
|
|
tx_q->cleaned_bytes = 0;
|
|
tx_q->cleaned_pkts = 0;
|
|
}
|
|
|
|
ntc += complq->desc_count;
|
|
complq->next_to_clean = ntc;
|
|
|
|
return !!complq_budget;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_splitq_build_ctb - populate command tag and size for queue
|
|
* based scheduling descriptors
|
|
* @desc: descriptor to populate
|
|
* @params: pointer to tx params struct
|
|
* @td_cmd: command to be filled in desc
|
|
* @size: size of buffer
|
|
*/
|
|
void idpf_tx_splitq_build_ctb(union idpf_tx_flex_desc *desc,
|
|
struct idpf_tx_splitq_params *params,
|
|
u16 td_cmd, u16 size)
|
|
{
|
|
desc->q.qw1.cmd_dtype =
|
|
le16_encode_bits(params->dtype, IDPF_FLEX_TXD_QW1_DTYPE_M);
|
|
desc->q.qw1.cmd_dtype |=
|
|
le16_encode_bits(td_cmd, IDPF_FLEX_TXD_QW1_CMD_M);
|
|
desc->q.qw1.buf_size = cpu_to_le16(size);
|
|
desc->q.qw1.l2tags.l2tag1 = cpu_to_le16(params->td_tag);
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_splitq_build_flow_desc - populate command tag and size for flow
|
|
* scheduling descriptors
|
|
* @desc: descriptor to populate
|
|
* @params: pointer to tx params struct
|
|
* @td_cmd: command to be filled in desc
|
|
* @size: size of buffer
|
|
*/
|
|
void idpf_tx_splitq_build_flow_desc(union idpf_tx_flex_desc *desc,
|
|
struct idpf_tx_splitq_params *params,
|
|
u16 td_cmd, u16 size)
|
|
{
|
|
desc->flow.qw1.cmd_dtype = (u16)params->dtype | td_cmd;
|
|
desc->flow.qw1.rxr_bufsize = cpu_to_le16((u16)size);
|
|
desc->flow.qw1.compl_tag = cpu_to_le16(params->compl_tag);
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_maybe_stop_splitq - 1st level check for Tx splitq stop conditions
|
|
* @tx_q: the queue to be checked
|
|
* @descs_needed: number of descriptors required for this packet
|
|
*
|
|
* Returns 0 if stop is not needed
|
|
*/
|
|
static int idpf_tx_maybe_stop_splitq(struct idpf_tx_queue *tx_q,
|
|
unsigned int descs_needed)
|
|
{
|
|
if (idpf_tx_maybe_stop_common(tx_q, descs_needed))
|
|
goto out;
|
|
|
|
/* If there are too many outstanding completions expected on the
|
|
* completion queue, stop the TX queue to give the device some time to
|
|
* catch up
|
|
*/
|
|
if (unlikely(IDPF_TX_COMPLQ_PENDING(tx_q->txq_grp) >
|
|
IDPF_TX_COMPLQ_OVERFLOW_THRESH(tx_q->txq_grp->complq)))
|
|
goto splitq_stop;
|
|
|
|
/* Also check for available book keeping buffers; if we are low, stop
|
|
* the queue to wait for more completions
|
|
*/
|
|
if (unlikely(IDPF_TX_BUF_RSV_LOW(tx_q)))
|
|
goto splitq_stop;
|
|
|
|
return 0;
|
|
|
|
splitq_stop:
|
|
netif_stop_subqueue(tx_q->netdev, tx_q->idx);
|
|
|
|
out:
|
|
u64_stats_update_begin(&tx_q->stats_sync);
|
|
u64_stats_inc(&tx_q->q_stats.q_busy);
|
|
u64_stats_update_end(&tx_q->stats_sync);
|
|
|
|
return -EBUSY;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_buf_hw_update - Store the new tail value
|
|
* @tx_q: queue to bump
|
|
* @val: new tail index
|
|
* @xmit_more: more skb's pending
|
|
*
|
|
* The naming here is special in that 'hw' signals that this function is about
|
|
* to do a register write to update our queue status. We know this can only
|
|
* mean tail here as HW should be owning head for TX.
|
|
*/
|
|
void idpf_tx_buf_hw_update(struct idpf_tx_queue *tx_q, u32 val,
|
|
bool xmit_more)
|
|
{
|
|
struct netdev_queue *nq;
|
|
|
|
nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx);
|
|
tx_q->next_to_use = val;
|
|
|
|
if (idpf_tx_maybe_stop_common(tx_q, IDPF_TX_DESC_NEEDED)) {
|
|
u64_stats_update_begin(&tx_q->stats_sync);
|
|
u64_stats_inc(&tx_q->q_stats.q_busy);
|
|
u64_stats_update_end(&tx_q->stats_sync);
|
|
}
|
|
|
|
/* Force memory writes to complete before letting h/w
|
|
* know there are new descriptors to fetch. (Only
|
|
* applicable for weak-ordered memory model archs,
|
|
* such as IA-64).
|
|
*/
|
|
wmb();
|
|
|
|
/* notify HW of packet */
|
|
if (netif_xmit_stopped(nq) || !xmit_more)
|
|
writel(val, tx_q->tail);
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_desc_count_required - calculate number of Tx descriptors needed
|
|
* @txq: queue to send buffer on
|
|
* @skb: send buffer
|
|
*
|
|
* Returns number of data descriptors needed for this skb.
|
|
*/
|
|
unsigned int idpf_tx_desc_count_required(struct idpf_tx_queue *txq,
|
|
struct sk_buff *skb)
|
|
{
|
|
const struct skb_shared_info *shinfo;
|
|
unsigned int count = 0, i;
|
|
|
|
count += !!skb_headlen(skb);
|
|
|
|
if (!skb_is_nonlinear(skb))
|
|
return count;
|
|
|
|
shinfo = skb_shinfo(skb);
|
|
for (i = 0; i < shinfo->nr_frags; i++) {
|
|
unsigned int size;
|
|
|
|
size = skb_frag_size(&shinfo->frags[i]);
|
|
|
|
/* We only need to use the idpf_size_to_txd_count check if the
|
|
* fragment is going to span multiple descriptors,
|
|
* i.e. size >= 16K.
|
|
*/
|
|
if (size >= SZ_16K)
|
|
count += idpf_size_to_txd_count(size);
|
|
else
|
|
count++;
|
|
}
|
|
|
|
if (idpf_chk_linearize(skb, txq->tx_max_bufs, count)) {
|
|
if (__skb_linearize(skb))
|
|
return 0;
|
|
|
|
count = idpf_size_to_txd_count(skb->len);
|
|
u64_stats_update_begin(&txq->stats_sync);
|
|
u64_stats_inc(&txq->q_stats.linearize);
|
|
u64_stats_update_end(&txq->stats_sync);
|
|
}
|
|
|
|
return count;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_dma_map_error - handle TX DMA map errors
|
|
* @txq: queue to send buffer on
|
|
* @skb: send buffer
|
|
* @first: original first buffer info buffer for packet
|
|
* @idx: starting point on ring to unwind
|
|
*/
|
|
void idpf_tx_dma_map_error(struct idpf_tx_queue *txq, struct sk_buff *skb,
|
|
struct idpf_tx_buf *first, u16 idx)
|
|
{
|
|
struct libeth_sq_napi_stats ss = { };
|
|
struct libeth_cq_pp cp = {
|
|
.dev = txq->dev,
|
|
.ss = &ss,
|
|
};
|
|
|
|
u64_stats_update_begin(&txq->stats_sync);
|
|
u64_stats_inc(&txq->q_stats.dma_map_errs);
|
|
u64_stats_update_end(&txq->stats_sync);
|
|
|
|
/* clear dma mappings for failed tx_buf map */
|
|
for (;;) {
|
|
struct idpf_tx_buf *tx_buf;
|
|
|
|
tx_buf = &txq->tx_buf[idx];
|
|
libeth_tx_complete(tx_buf, &cp);
|
|
if (tx_buf == first)
|
|
break;
|
|
if (idx == 0)
|
|
idx = txq->desc_count;
|
|
idx--;
|
|
}
|
|
|
|
if (skb_is_gso(skb)) {
|
|
union idpf_tx_flex_desc *tx_desc;
|
|
|
|
/* If we failed a DMA mapping for a TSO packet, we will have
|
|
* used one additional descriptor for a context
|
|
* descriptor. Reset that here.
|
|
*/
|
|
tx_desc = &txq->flex_tx[idx];
|
|
memset(tx_desc, 0, sizeof(struct idpf_flex_tx_ctx_desc));
|
|
if (idx == 0)
|
|
idx = txq->desc_count;
|
|
idx--;
|
|
}
|
|
|
|
/* Update tail in case netdev_xmit_more was previously true */
|
|
idpf_tx_buf_hw_update(txq, idx, false);
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_splitq_bump_ntu - adjust NTU and generation
|
|
* @txq: the tx ring to wrap
|
|
* @ntu: ring index to bump
|
|
*/
|
|
static unsigned int idpf_tx_splitq_bump_ntu(struct idpf_tx_queue *txq, u16 ntu)
|
|
{
|
|
ntu++;
|
|
|
|
if (ntu == txq->desc_count) {
|
|
ntu = 0;
|
|
txq->compl_tag_cur_gen = IDPF_TX_ADJ_COMPL_TAG_GEN(txq);
|
|
}
|
|
|
|
return ntu;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_splitq_map - Build the Tx flex descriptor
|
|
* @tx_q: queue to send buffer on
|
|
* @params: pointer to splitq params struct
|
|
* @first: first buffer info buffer to use
|
|
*
|
|
* This function loops over the skb data pointed to by *first
|
|
* and gets a physical address for each memory location and programs
|
|
* it and the length into the transmit flex descriptor.
|
|
*/
|
|
static void idpf_tx_splitq_map(struct idpf_tx_queue *tx_q,
|
|
struct idpf_tx_splitq_params *params,
|
|
struct idpf_tx_buf *first)
|
|
{
|
|
union idpf_tx_flex_desc *tx_desc;
|
|
unsigned int data_len, size;
|
|
struct idpf_tx_buf *tx_buf;
|
|
u16 i = tx_q->next_to_use;
|
|
struct netdev_queue *nq;
|
|
struct sk_buff *skb;
|
|
skb_frag_t *frag;
|
|
u16 td_cmd = 0;
|
|
dma_addr_t dma;
|
|
|
|
skb = first->skb;
|
|
|
|
td_cmd = params->offload.td_cmd;
|
|
|
|
data_len = skb->data_len;
|
|
size = skb_headlen(skb);
|
|
|
|
tx_desc = &tx_q->flex_tx[i];
|
|
|
|
dma = dma_map_single(tx_q->dev, skb->data, size, DMA_TO_DEVICE);
|
|
|
|
tx_buf = first;
|
|
first->nr_frags = 0;
|
|
|
|
params->compl_tag =
|
|
(tx_q->compl_tag_cur_gen << tx_q->compl_tag_gen_s) | i;
|
|
|
|
for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
|
|
unsigned int max_data = IDPF_TX_MAX_DESC_DATA_ALIGNED;
|
|
|
|
if (dma_mapping_error(tx_q->dev, dma))
|
|
return idpf_tx_dma_map_error(tx_q, skb, first, i);
|
|
|
|
first->nr_frags++;
|
|
idpf_tx_buf_compl_tag(tx_buf) = params->compl_tag;
|
|
tx_buf->type = LIBETH_SQE_FRAG;
|
|
|
|
/* record length, and DMA address */
|
|
dma_unmap_len_set(tx_buf, len, size);
|
|
dma_unmap_addr_set(tx_buf, dma, dma);
|
|
|
|
/* buf_addr is in same location for both desc types */
|
|
tx_desc->q.buf_addr = cpu_to_le64(dma);
|
|
|
|
/* The stack can send us fragments that are too large for a
|
|
* single descriptor i.e. frag size > 16K-1. We will need to
|
|
* split the fragment across multiple descriptors in this case.
|
|
* To adhere to HW alignment restrictions, the fragment needs
|
|
* to be split such that the first chunk ends on a 4K boundary
|
|
* and all subsequent chunks start on a 4K boundary. We still
|
|
* want to send as much data as possible though, so our
|
|
* intermediate descriptor chunk size will be 12K.
|
|
*
|
|
* For example, consider a 32K fragment mapped to DMA addr 2600.
|
|
* ------------------------------------------------------------
|
|
* | frag_size = 32K |
|
|
* ------------------------------------------------------------
|
|
* |2600 |16384 |28672
|
|
*
|
|
* 3 descriptors will be used for this fragment. The HW expects
|
|
* the descriptors to contain the following:
|
|
* ------------------------------------------------------------
|
|
* | size = 13784 | size = 12K | size = 6696 |
|
|
* | dma = 2600 | dma = 16384 | dma = 28672 |
|
|
* ------------------------------------------------------------
|
|
*
|
|
* We need to first adjust the max_data for the first chunk so
|
|
* that it ends on a 4K boundary. By negating the value of the
|
|
* DMA address and taking only the low order bits, we're
|
|
* effectively calculating
|
|
* 4K - (DMA addr lower order bits) =
|
|
* bytes to next boundary.
|
|
*
|
|
* Add that to our base aligned max_data (12K) and we have
|
|
* our first chunk size. In the example above,
|
|
* 13784 = 12K + (4096-2600)
|
|
*
|
|
* After guaranteeing the first chunk ends on a 4K boundary, we
|
|
* will give the intermediate descriptors 12K chunks and
|
|
* whatever is left to the final descriptor. This ensures that
|
|
* all descriptors used for the remaining chunks of the
|
|
* fragment start on a 4K boundary and we use as few
|
|
* descriptors as possible.
|
|
*/
|
|
max_data += -dma & (IDPF_TX_MAX_READ_REQ_SIZE - 1);
|
|
while (unlikely(size > IDPF_TX_MAX_DESC_DATA)) {
|
|
idpf_tx_splitq_build_desc(tx_desc, params, td_cmd,
|
|
max_data);
|
|
|
|
if (unlikely(++i == tx_q->desc_count)) {
|
|
tx_buf = tx_q->tx_buf;
|
|
tx_desc = &tx_q->flex_tx[0];
|
|
i = 0;
|
|
tx_q->compl_tag_cur_gen =
|
|
IDPF_TX_ADJ_COMPL_TAG_GEN(tx_q);
|
|
} else {
|
|
tx_buf++;
|
|
tx_desc++;
|
|
}
|
|
|
|
/* Since this packet has a buffer that is going to span
|
|
* multiple descriptors, it's going to leave holes in
|
|
* to the TX buffer ring. To ensure these holes do not
|
|
* cause issues in the cleaning routines, we will clear
|
|
* them of any stale data and assign them the same
|
|
* completion tag as the current packet. Then when the
|
|
* packet is being cleaned, the cleaning routines will
|
|
* simply pass over these holes and finish cleaning the
|
|
* rest of the packet.
|
|
*/
|
|
tx_buf->type = LIBETH_SQE_EMPTY;
|
|
idpf_tx_buf_compl_tag(tx_buf) = params->compl_tag;
|
|
|
|
/* Adjust the DMA offset and the remaining size of the
|
|
* fragment. On the first iteration of this loop,
|
|
* max_data will be >= 12K and <= 16K-1. On any
|
|
* subsequent iteration of this loop, max_data will
|
|
* always be 12K.
|
|
*/
|
|
dma += max_data;
|
|
size -= max_data;
|
|
|
|
/* Reset max_data since remaining chunks will be 12K
|
|
* at most
|
|
*/
|
|
max_data = IDPF_TX_MAX_DESC_DATA_ALIGNED;
|
|
|
|
/* buf_addr is in same location for both desc types */
|
|
tx_desc->q.buf_addr = cpu_to_le64(dma);
|
|
}
|
|
|
|
if (!data_len)
|
|
break;
|
|
|
|
idpf_tx_splitq_build_desc(tx_desc, params, td_cmd, size);
|
|
|
|
if (unlikely(++i == tx_q->desc_count)) {
|
|
tx_buf = tx_q->tx_buf;
|
|
tx_desc = &tx_q->flex_tx[0];
|
|
i = 0;
|
|
tx_q->compl_tag_cur_gen = IDPF_TX_ADJ_COMPL_TAG_GEN(tx_q);
|
|
} else {
|
|
tx_buf++;
|
|
tx_desc++;
|
|
}
|
|
|
|
size = skb_frag_size(frag);
|
|
data_len -= size;
|
|
|
|
dma = skb_frag_dma_map(tx_q->dev, frag, 0, size,
|
|
DMA_TO_DEVICE);
|
|
}
|
|
|
|
/* record SW timestamp if HW timestamp is not available */
|
|
skb_tx_timestamp(skb);
|
|
|
|
first->type = LIBETH_SQE_SKB;
|
|
|
|
/* write last descriptor with RS and EOP bits */
|
|
first->rs_idx = i;
|
|
td_cmd |= params->eop_cmd;
|
|
idpf_tx_splitq_build_desc(tx_desc, params, td_cmd, size);
|
|
i = idpf_tx_splitq_bump_ntu(tx_q, i);
|
|
|
|
tx_q->txq_grp->num_completions_pending++;
|
|
|
|
/* record bytecount for BQL */
|
|
nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx);
|
|
netdev_tx_sent_queue(nq, first->bytes);
|
|
|
|
idpf_tx_buf_hw_update(tx_q, i, netdev_xmit_more());
|
|
}
|
|
|
|
/**
|
|
* idpf_tso - computes mss and TSO length to prepare for TSO
|
|
* @skb: pointer to skb
|
|
* @off: pointer to struct that holds offload parameters
|
|
*
|
|
* Returns error (negative) if TSO was requested but cannot be applied to the
|
|
* given skb, 0 if TSO does not apply to the given skb, or 1 otherwise.
|
|
*/
|
|
int idpf_tso(struct sk_buff *skb, struct idpf_tx_offload_params *off)
|
|
{
|
|
const struct skb_shared_info *shinfo;
|
|
union {
|
|
struct iphdr *v4;
|
|
struct ipv6hdr *v6;
|
|
unsigned char *hdr;
|
|
} ip;
|
|
union {
|
|
struct tcphdr *tcp;
|
|
struct udphdr *udp;
|
|
unsigned char *hdr;
|
|
} l4;
|
|
u32 paylen, l4_start;
|
|
int err;
|
|
|
|
if (!skb_is_gso(skb))
|
|
return 0;
|
|
|
|
err = skb_cow_head(skb, 0);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
shinfo = skb_shinfo(skb);
|
|
|
|
ip.hdr = skb_network_header(skb);
|
|
l4.hdr = skb_transport_header(skb);
|
|
|
|
/* initialize outer IP header fields */
|
|
if (ip.v4->version == 4) {
|
|
ip.v4->tot_len = 0;
|
|
ip.v4->check = 0;
|
|
} else if (ip.v6->version == 6) {
|
|
ip.v6->payload_len = 0;
|
|
}
|
|
|
|
l4_start = skb_transport_offset(skb);
|
|
|
|
/* remove payload length from checksum */
|
|
paylen = skb->len - l4_start;
|
|
|
|
switch (shinfo->gso_type & ~SKB_GSO_DODGY) {
|
|
case SKB_GSO_TCPV4:
|
|
case SKB_GSO_TCPV6:
|
|
csum_replace_by_diff(&l4.tcp->check,
|
|
(__force __wsum)htonl(paylen));
|
|
off->tso_hdr_len = __tcp_hdrlen(l4.tcp) + l4_start;
|
|
break;
|
|
case SKB_GSO_UDP_L4:
|
|
csum_replace_by_diff(&l4.udp->check,
|
|
(__force __wsum)htonl(paylen));
|
|
/* compute length of segmentation header */
|
|
off->tso_hdr_len = sizeof(struct udphdr) + l4_start;
|
|
l4.udp->len = htons(shinfo->gso_size + sizeof(struct udphdr));
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
|
|
off->tso_len = skb->len - off->tso_hdr_len;
|
|
off->mss = shinfo->gso_size;
|
|
off->tso_segs = shinfo->gso_segs;
|
|
|
|
off->tx_flags |= IDPF_TX_FLAGS_TSO;
|
|
|
|
return 1;
|
|
}
|
|
|
|
/**
|
|
* __idpf_chk_linearize - Check skb is not using too many buffers
|
|
* @skb: send buffer
|
|
* @max_bufs: maximum number of buffers
|
|
*
|
|
* For TSO we need to count the TSO header and segment payload separately. As
|
|
* such we need to check cases where we have max_bufs-1 fragments or more as we
|
|
* can potentially require max_bufs+1 DMA transactions, 1 for the TSO header, 1
|
|
* for the segment payload in the first descriptor, and another max_buf-1 for
|
|
* the fragments.
|
|
*/
|
|
static bool __idpf_chk_linearize(struct sk_buff *skb, unsigned int max_bufs)
|
|
{
|
|
const struct skb_shared_info *shinfo = skb_shinfo(skb);
|
|
const skb_frag_t *frag, *stale;
|
|
int nr_frags, sum;
|
|
|
|
/* no need to check if number of frags is less than max_bufs - 1 */
|
|
nr_frags = shinfo->nr_frags;
|
|
if (nr_frags < (max_bufs - 1))
|
|
return false;
|
|
|
|
/* We need to walk through the list and validate that each group
|
|
* of max_bufs-2 fragments totals at least gso_size.
|
|
*/
|
|
nr_frags -= max_bufs - 2;
|
|
frag = &shinfo->frags[0];
|
|
|
|
/* Initialize size to the negative value of gso_size minus 1. We use
|
|
* this as the worst case scenario in which the frag ahead of us only
|
|
* provides one byte which is why we are limited to max_bufs-2
|
|
* descriptors for a single transmit as the header and previous
|
|
* fragment are already consuming 2 descriptors.
|
|
*/
|
|
sum = 1 - shinfo->gso_size;
|
|
|
|
/* Add size of frags 0 through 4 to create our initial sum */
|
|
sum += skb_frag_size(frag++);
|
|
sum += skb_frag_size(frag++);
|
|
sum += skb_frag_size(frag++);
|
|
sum += skb_frag_size(frag++);
|
|
sum += skb_frag_size(frag++);
|
|
|
|
/* Walk through fragments adding latest fragment, testing it, and
|
|
* then removing stale fragments from the sum.
|
|
*/
|
|
for (stale = &shinfo->frags[0];; stale++) {
|
|
int stale_size = skb_frag_size(stale);
|
|
|
|
sum += skb_frag_size(frag++);
|
|
|
|
/* The stale fragment may present us with a smaller
|
|
* descriptor than the actual fragment size. To account
|
|
* for that we need to remove all the data on the front and
|
|
* figure out what the remainder would be in the last
|
|
* descriptor associated with the fragment.
|
|
*/
|
|
if (stale_size > IDPF_TX_MAX_DESC_DATA) {
|
|
int align_pad = -(skb_frag_off(stale)) &
|
|
(IDPF_TX_MAX_READ_REQ_SIZE - 1);
|
|
|
|
sum -= align_pad;
|
|
stale_size -= align_pad;
|
|
|
|
do {
|
|
sum -= IDPF_TX_MAX_DESC_DATA_ALIGNED;
|
|
stale_size -= IDPF_TX_MAX_DESC_DATA_ALIGNED;
|
|
} while (stale_size > IDPF_TX_MAX_DESC_DATA);
|
|
}
|
|
|
|
/* if sum is negative we failed to make sufficient progress */
|
|
if (sum < 0)
|
|
return true;
|
|
|
|
if (!nr_frags--)
|
|
break;
|
|
|
|
sum -= stale_size;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
/**
|
|
* idpf_chk_linearize - Check if skb exceeds max descriptors per packet
|
|
* @skb: send buffer
|
|
* @max_bufs: maximum scatter gather buffers for single packet
|
|
* @count: number of buffers this packet needs
|
|
*
|
|
* Make sure we don't exceed maximum scatter gather buffers for a single
|
|
* packet. We have to do some special checking around the boundary (max_bufs-1)
|
|
* if TSO is on since we need count the TSO header and payload separately.
|
|
* E.g.: a packet with 7 fragments can require 9 DMA transactions; 1 for TSO
|
|
* header, 1 for segment payload, and then 7 for the fragments.
|
|
*/
|
|
static bool idpf_chk_linearize(struct sk_buff *skb, unsigned int max_bufs,
|
|
unsigned int count)
|
|
{
|
|
if (likely(count < max_bufs))
|
|
return false;
|
|
if (skb_is_gso(skb))
|
|
return __idpf_chk_linearize(skb, max_bufs);
|
|
|
|
return count > max_bufs;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_splitq_get_ctx_desc - grab next desc and update buffer ring
|
|
* @txq: queue to put context descriptor on
|
|
*
|
|
* Since the TX buffer rings mimics the descriptor ring, update the tx buffer
|
|
* ring entry to reflect that this index is a context descriptor
|
|
*/
|
|
static struct idpf_flex_tx_ctx_desc *
|
|
idpf_tx_splitq_get_ctx_desc(struct idpf_tx_queue *txq)
|
|
{
|
|
struct idpf_flex_tx_ctx_desc *desc;
|
|
int i = txq->next_to_use;
|
|
|
|
txq->tx_buf[i].type = LIBETH_SQE_CTX;
|
|
|
|
/* grab the next descriptor */
|
|
desc = &txq->flex_ctx[i];
|
|
txq->next_to_use = idpf_tx_splitq_bump_ntu(txq, i);
|
|
|
|
return desc;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_drop_skb - free the SKB and bump tail if necessary
|
|
* @tx_q: queue to send buffer on
|
|
* @skb: pointer to skb
|
|
*/
|
|
netdev_tx_t idpf_tx_drop_skb(struct idpf_tx_queue *tx_q, struct sk_buff *skb)
|
|
{
|
|
u64_stats_update_begin(&tx_q->stats_sync);
|
|
u64_stats_inc(&tx_q->q_stats.skb_drops);
|
|
u64_stats_update_end(&tx_q->stats_sync);
|
|
|
|
idpf_tx_buf_hw_update(tx_q, tx_q->next_to_use, false);
|
|
|
|
dev_kfree_skb(skb);
|
|
|
|
return NETDEV_TX_OK;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_splitq_frame - Sends buffer on Tx ring using flex descriptors
|
|
* @skb: send buffer
|
|
* @tx_q: queue to send buffer on
|
|
*
|
|
* Returns NETDEV_TX_OK if sent, else an error code
|
|
*/
|
|
static netdev_tx_t idpf_tx_splitq_frame(struct sk_buff *skb,
|
|
struct idpf_tx_queue *tx_q)
|
|
{
|
|
struct idpf_tx_splitq_params tx_params = { };
|
|
struct idpf_tx_buf *first;
|
|
unsigned int count;
|
|
int tso;
|
|
|
|
count = idpf_tx_desc_count_required(tx_q, skb);
|
|
if (unlikely(!count))
|
|
return idpf_tx_drop_skb(tx_q, skb);
|
|
|
|
tso = idpf_tso(skb, &tx_params.offload);
|
|
if (unlikely(tso < 0))
|
|
return idpf_tx_drop_skb(tx_q, skb);
|
|
|
|
/* Check for splitq specific TX resources */
|
|
count += (IDPF_TX_DESCS_PER_CACHE_LINE + tso);
|
|
if (idpf_tx_maybe_stop_splitq(tx_q, count)) {
|
|
idpf_tx_buf_hw_update(tx_q, tx_q->next_to_use, false);
|
|
|
|
return NETDEV_TX_BUSY;
|
|
}
|
|
|
|
if (tso) {
|
|
/* If tso is needed, set up context desc */
|
|
struct idpf_flex_tx_ctx_desc *ctx_desc =
|
|
idpf_tx_splitq_get_ctx_desc(tx_q);
|
|
|
|
ctx_desc->tso.qw1.cmd_dtype =
|
|
cpu_to_le16(IDPF_TX_DESC_DTYPE_FLEX_TSO_CTX |
|
|
IDPF_TX_FLEX_CTX_DESC_CMD_TSO);
|
|
ctx_desc->tso.qw0.flex_tlen =
|
|
cpu_to_le32(tx_params.offload.tso_len &
|
|
IDPF_TXD_FLEX_CTX_TLEN_M);
|
|
ctx_desc->tso.qw0.mss_rt =
|
|
cpu_to_le16(tx_params.offload.mss &
|
|
IDPF_TXD_FLEX_CTX_MSS_RT_M);
|
|
ctx_desc->tso.qw0.hdr_len = tx_params.offload.tso_hdr_len;
|
|
|
|
u64_stats_update_begin(&tx_q->stats_sync);
|
|
u64_stats_inc(&tx_q->q_stats.lso_pkts);
|
|
u64_stats_update_end(&tx_q->stats_sync);
|
|
}
|
|
|
|
/* record the location of the first descriptor for this packet */
|
|
first = &tx_q->tx_buf[tx_q->next_to_use];
|
|
first->skb = skb;
|
|
|
|
if (tso) {
|
|
first->packets = tx_params.offload.tso_segs;
|
|
first->bytes = skb->len +
|
|
((first->packets - 1) * tx_params.offload.tso_hdr_len);
|
|
} else {
|
|
first->packets = 1;
|
|
first->bytes = max_t(unsigned int, skb->len, ETH_ZLEN);
|
|
}
|
|
|
|
if (idpf_queue_has(FLOW_SCH_EN, tx_q)) {
|
|
tx_params.dtype = IDPF_TX_DESC_DTYPE_FLEX_FLOW_SCHE;
|
|
tx_params.eop_cmd = IDPF_TXD_FLEX_FLOW_CMD_EOP;
|
|
/* Set the RE bit to catch any packets that may have not been
|
|
* stashed during RS completion cleaning. MIN_GAP is set to
|
|
* MIN_RING size to ensure it will be set at least once each
|
|
* time around the ring.
|
|
*/
|
|
if (!(tx_q->next_to_use % IDPF_TX_SPLITQ_RE_MIN_GAP)) {
|
|
tx_params.eop_cmd |= IDPF_TXD_FLEX_FLOW_CMD_RE;
|
|
tx_q->txq_grp->num_completions_pending++;
|
|
}
|
|
|
|
if (skb->ip_summed == CHECKSUM_PARTIAL)
|
|
tx_params.offload.td_cmd |= IDPF_TXD_FLEX_FLOW_CMD_CS_EN;
|
|
|
|
} else {
|
|
tx_params.dtype = IDPF_TX_DESC_DTYPE_FLEX_L2TAG1_L2TAG2;
|
|
tx_params.eop_cmd = IDPF_TXD_LAST_DESC_CMD;
|
|
|
|
if (skb->ip_summed == CHECKSUM_PARTIAL)
|
|
tx_params.offload.td_cmd |= IDPF_TX_FLEX_DESC_CMD_CS_EN;
|
|
}
|
|
|
|
idpf_tx_splitq_map(tx_q, &tx_params, first);
|
|
|
|
return NETDEV_TX_OK;
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_start - Selects the right Tx queue to send buffer
|
|
* @skb: send buffer
|
|
* @netdev: network interface device structure
|
|
*
|
|
* Returns NETDEV_TX_OK if sent, else an error code
|
|
*/
|
|
netdev_tx_t idpf_tx_start(struct sk_buff *skb, struct net_device *netdev)
|
|
{
|
|
struct idpf_vport *vport = idpf_netdev_to_vport(netdev);
|
|
struct idpf_tx_queue *tx_q;
|
|
|
|
if (unlikely(skb_get_queue_mapping(skb) >= vport->num_txq)) {
|
|
dev_kfree_skb_any(skb);
|
|
|
|
return NETDEV_TX_OK;
|
|
}
|
|
|
|
tx_q = vport->txqs[skb_get_queue_mapping(skb)];
|
|
|
|
/* hardware can't handle really short frames, hardware padding works
|
|
* beyond this point
|
|
*/
|
|
if (skb_put_padto(skb, tx_q->tx_min_pkt_len)) {
|
|
idpf_tx_buf_hw_update(tx_q, tx_q->next_to_use, false);
|
|
|
|
return NETDEV_TX_OK;
|
|
}
|
|
|
|
if (idpf_is_queue_model_split(vport->txq_model))
|
|
return idpf_tx_splitq_frame(skb, tx_q);
|
|
else
|
|
return idpf_tx_singleq_frame(skb, tx_q);
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_hash - set the hash value in the skb
|
|
* @rxq: Rx descriptor ring packet is being transacted on
|
|
* @skb: pointer to current skb being populated
|
|
* @rx_desc: Receive descriptor
|
|
* @decoded: Decoded Rx packet type related fields
|
|
*/
|
|
static void
|
|
idpf_rx_hash(const struct idpf_rx_queue *rxq, struct sk_buff *skb,
|
|
const struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc,
|
|
struct libeth_rx_pt decoded)
|
|
{
|
|
u32 hash;
|
|
|
|
if (!libeth_rx_pt_has_hash(rxq->netdev, decoded))
|
|
return;
|
|
|
|
hash = le16_to_cpu(rx_desc->hash1) |
|
|
(rx_desc->ff2_mirrid_hash2.hash2 << 16) |
|
|
(rx_desc->hash3 << 24);
|
|
|
|
libeth_rx_pt_set_hash(skb, hash, decoded);
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_csum - Indicate in skb if checksum is good
|
|
* @rxq: Rx descriptor ring packet is being transacted on
|
|
* @skb: pointer to current skb being populated
|
|
* @csum_bits: checksum fields extracted from the descriptor
|
|
* @decoded: Decoded Rx packet type related fields
|
|
*
|
|
* skb->protocol must be set before this function is called
|
|
*/
|
|
static void idpf_rx_csum(struct idpf_rx_queue *rxq, struct sk_buff *skb,
|
|
struct idpf_rx_csum_decoded csum_bits,
|
|
struct libeth_rx_pt decoded)
|
|
{
|
|
bool ipv4, ipv6;
|
|
|
|
/* check if Rx checksum is enabled */
|
|
if (!libeth_rx_pt_has_checksum(rxq->netdev, decoded))
|
|
return;
|
|
|
|
/* check if HW has decoded the packet and checksum */
|
|
if (unlikely(!csum_bits.l3l4p))
|
|
return;
|
|
|
|
ipv4 = libeth_rx_pt_get_ip_ver(decoded) == LIBETH_RX_PT_OUTER_IPV4;
|
|
ipv6 = libeth_rx_pt_get_ip_ver(decoded) == LIBETH_RX_PT_OUTER_IPV6;
|
|
|
|
if (unlikely(ipv4 && (csum_bits.ipe || csum_bits.eipe)))
|
|
goto checksum_fail;
|
|
|
|
if (unlikely(ipv6 && csum_bits.ipv6exadd))
|
|
return;
|
|
|
|
/* check for L4 errors and handle packets that were not able to be
|
|
* checksummed
|
|
*/
|
|
if (unlikely(csum_bits.l4e))
|
|
goto checksum_fail;
|
|
|
|
if (csum_bits.raw_csum_inv ||
|
|
decoded.inner_prot == LIBETH_RX_PT_INNER_SCTP) {
|
|
skb->ip_summed = CHECKSUM_UNNECESSARY;
|
|
return;
|
|
}
|
|
|
|
skb->csum = csum_unfold((__force __sum16)~swab16(csum_bits.raw_csum));
|
|
skb->ip_summed = CHECKSUM_COMPLETE;
|
|
|
|
return;
|
|
|
|
checksum_fail:
|
|
u64_stats_update_begin(&rxq->stats_sync);
|
|
u64_stats_inc(&rxq->q_stats.hw_csum_err);
|
|
u64_stats_update_end(&rxq->stats_sync);
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_splitq_extract_csum_bits - Extract checksum bits from descriptor
|
|
* @rx_desc: receive descriptor
|
|
*
|
|
* Return: parsed checksum status.
|
|
**/
|
|
static struct idpf_rx_csum_decoded
|
|
idpf_rx_splitq_extract_csum_bits(const struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc)
|
|
{
|
|
struct idpf_rx_csum_decoded csum = { };
|
|
u8 qword0, qword1;
|
|
|
|
qword0 = rx_desc->status_err0_qw0;
|
|
qword1 = rx_desc->status_err0_qw1;
|
|
|
|
csum.ipe = FIELD_GET(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_IPE_M,
|
|
qword1);
|
|
csum.eipe = FIELD_GET(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EIPE_M,
|
|
qword1);
|
|
csum.l4e = FIELD_GET(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_L4E_M,
|
|
qword1);
|
|
csum.l3l4p = FIELD_GET(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_L3L4P_M,
|
|
qword1);
|
|
csum.ipv6exadd = FIELD_GET(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_IPV6EXADD_M,
|
|
qword0);
|
|
csum.raw_csum_inv =
|
|
le16_get_bits(rx_desc->ptype_err_fflags0,
|
|
VIRTCHNL2_RX_FLEX_DESC_ADV_RAW_CSUM_INV_M);
|
|
csum.raw_csum = le16_to_cpu(rx_desc->misc.raw_cs);
|
|
|
|
return csum;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_rsc - Set the RSC fields in the skb
|
|
* @rxq : Rx descriptor ring packet is being transacted on
|
|
* @skb : pointer to current skb being populated
|
|
* @rx_desc: Receive descriptor
|
|
* @decoded: Decoded Rx packet type related fields
|
|
*
|
|
* Return 0 on success and error code on failure
|
|
*
|
|
* Populate the skb fields with the total number of RSC segments, RSC payload
|
|
* length and packet type.
|
|
*/
|
|
static int idpf_rx_rsc(struct idpf_rx_queue *rxq, struct sk_buff *skb,
|
|
const struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc,
|
|
struct libeth_rx_pt decoded)
|
|
{
|
|
u16 rsc_segments, rsc_seg_len;
|
|
bool ipv4, ipv6;
|
|
int len;
|
|
|
|
if (unlikely(libeth_rx_pt_get_ip_ver(decoded) ==
|
|
LIBETH_RX_PT_OUTER_L2))
|
|
return -EINVAL;
|
|
|
|
rsc_seg_len = le16_to_cpu(rx_desc->misc.rscseglen);
|
|
if (unlikely(!rsc_seg_len))
|
|
return -EINVAL;
|
|
|
|
ipv4 = libeth_rx_pt_get_ip_ver(decoded) == LIBETH_RX_PT_OUTER_IPV4;
|
|
ipv6 = libeth_rx_pt_get_ip_ver(decoded) == LIBETH_RX_PT_OUTER_IPV6;
|
|
|
|
if (unlikely(!(ipv4 ^ ipv6)))
|
|
return -EINVAL;
|
|
|
|
rsc_segments = DIV_ROUND_UP(skb->data_len, rsc_seg_len);
|
|
if (unlikely(rsc_segments == 1))
|
|
return 0;
|
|
|
|
NAPI_GRO_CB(skb)->count = rsc_segments;
|
|
skb_shinfo(skb)->gso_size = rsc_seg_len;
|
|
|
|
skb_reset_network_header(skb);
|
|
len = skb->len - skb_transport_offset(skb);
|
|
|
|
if (ipv4) {
|
|
struct iphdr *ipv4h = ip_hdr(skb);
|
|
|
|
skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4;
|
|
|
|
/* Reset and set transport header offset in skb */
|
|
skb_set_transport_header(skb, sizeof(struct iphdr));
|
|
|
|
/* Compute the TCP pseudo header checksum*/
|
|
tcp_hdr(skb)->check =
|
|
~tcp_v4_check(len, ipv4h->saddr, ipv4h->daddr, 0);
|
|
} else {
|
|
struct ipv6hdr *ipv6h = ipv6_hdr(skb);
|
|
|
|
skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6;
|
|
skb_set_transport_header(skb, sizeof(struct ipv6hdr));
|
|
tcp_hdr(skb)->check =
|
|
~tcp_v6_check(len, &ipv6h->saddr, &ipv6h->daddr, 0);
|
|
}
|
|
|
|
tcp_gro_complete(skb);
|
|
|
|
u64_stats_update_begin(&rxq->stats_sync);
|
|
u64_stats_inc(&rxq->q_stats.rsc_pkts);
|
|
u64_stats_update_end(&rxq->stats_sync);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_process_skb_fields - Populate skb header fields from Rx descriptor
|
|
* @rxq: Rx descriptor ring packet is being transacted on
|
|
* @skb: pointer to current skb being populated
|
|
* @rx_desc: Receive descriptor
|
|
*
|
|
* This function checks the ring, descriptor, and packet information in
|
|
* order to populate the hash, checksum, protocol, and
|
|
* other fields within the skb.
|
|
*/
|
|
static int
|
|
idpf_rx_process_skb_fields(struct idpf_rx_queue *rxq, struct sk_buff *skb,
|
|
const struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc)
|
|
{
|
|
struct idpf_rx_csum_decoded csum_bits;
|
|
struct libeth_rx_pt decoded;
|
|
u16 rx_ptype;
|
|
|
|
rx_ptype = le16_get_bits(rx_desc->ptype_err_fflags0,
|
|
VIRTCHNL2_RX_FLEX_DESC_ADV_PTYPE_M);
|
|
decoded = rxq->rx_ptype_lkup[rx_ptype];
|
|
|
|
/* process RSS/hash */
|
|
idpf_rx_hash(rxq, skb, rx_desc, decoded);
|
|
|
|
skb->protocol = eth_type_trans(skb, rxq->netdev);
|
|
|
|
if (le16_get_bits(rx_desc->hdrlen_flags,
|
|
VIRTCHNL2_RX_FLEX_DESC_ADV_RSC_M))
|
|
return idpf_rx_rsc(rxq, skb, rx_desc, decoded);
|
|
|
|
csum_bits = idpf_rx_splitq_extract_csum_bits(rx_desc);
|
|
idpf_rx_csum(rxq, skb, csum_bits, decoded);
|
|
|
|
skb_record_rx_queue(skb, rxq->idx);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_add_frag - Add contents of Rx buffer to sk_buff as a frag
|
|
* @rx_buf: buffer containing page to add
|
|
* @skb: sk_buff to place the data into
|
|
* @size: packet length from rx_desc
|
|
*
|
|
* This function will add the data contained in rx_buf->page to the skb.
|
|
* It will just attach the page as a frag to the skb.
|
|
* The function will then update the page offset.
|
|
*/
|
|
void idpf_rx_add_frag(struct idpf_rx_buf *rx_buf, struct sk_buff *skb,
|
|
unsigned int size)
|
|
{
|
|
u32 hr = rx_buf->page->pp->p.offset;
|
|
|
|
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, rx_buf->page,
|
|
rx_buf->offset + hr, size, rx_buf->truesize);
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_hsplit_wa - handle header buffer overflows and split errors
|
|
* @hdr: Rx buffer for the headers
|
|
* @buf: Rx buffer for the payload
|
|
* @data_len: number of bytes received to the payload buffer
|
|
*
|
|
* When a header buffer overflow occurs or the HW was unable do parse the
|
|
* packet type to perform header split, the whole frame gets placed to the
|
|
* payload buffer. We can't build a valid skb around a payload buffer when
|
|
* the header split is active since it doesn't reserve any head- or tailroom.
|
|
* In that case, copy either the whole frame when it's short or just the
|
|
* Ethernet header to the header buffer to be able to build an skb and adjust
|
|
* the data offset in the payload buffer, IOW emulate the header split.
|
|
*
|
|
* Return: number of bytes copied to the header buffer.
|
|
*/
|
|
static u32 idpf_rx_hsplit_wa(const struct libeth_fqe *hdr,
|
|
struct libeth_fqe *buf, u32 data_len)
|
|
{
|
|
u32 copy = data_len <= L1_CACHE_BYTES ? data_len : ETH_HLEN;
|
|
const void *src;
|
|
void *dst;
|
|
|
|
if (!libeth_rx_sync_for_cpu(buf, copy))
|
|
return 0;
|
|
|
|
dst = page_address(hdr->page) + hdr->offset + hdr->page->pp->p.offset;
|
|
src = page_address(buf->page) + buf->offset + buf->page->pp->p.offset;
|
|
memcpy(dst, src, LARGEST_ALIGN(copy));
|
|
|
|
buf->offset += copy;
|
|
|
|
return copy;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_build_skb - Allocate skb and populate it from header buffer
|
|
* @buf: Rx buffer to pull data from
|
|
* @size: the length of the packet
|
|
*
|
|
* This function allocates an skb. It then populates it with the page data from
|
|
* the current receive descriptor, taking care to set up the skb correctly.
|
|
*/
|
|
struct sk_buff *idpf_rx_build_skb(const struct libeth_fqe *buf, u32 size)
|
|
{
|
|
u32 hr = buf->page->pp->p.offset;
|
|
struct sk_buff *skb;
|
|
void *va;
|
|
|
|
va = page_address(buf->page) + buf->offset;
|
|
prefetch(va + hr);
|
|
|
|
skb = napi_build_skb(va, buf->truesize);
|
|
if (unlikely(!skb))
|
|
return NULL;
|
|
|
|
skb_mark_for_recycle(skb);
|
|
|
|
skb_reserve(skb, hr);
|
|
__skb_put(skb, size);
|
|
|
|
return skb;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_splitq_test_staterr - tests bits in Rx descriptor
|
|
* status and error fields
|
|
* @stat_err_field: field from descriptor to test bits in
|
|
* @stat_err_bits: value to mask
|
|
*
|
|
*/
|
|
static bool idpf_rx_splitq_test_staterr(const u8 stat_err_field,
|
|
const u8 stat_err_bits)
|
|
{
|
|
return !!(stat_err_field & stat_err_bits);
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_splitq_is_eop - process handling of EOP buffers
|
|
* @rx_desc: Rx descriptor for current buffer
|
|
*
|
|
* If the buffer is an EOP buffer, this function exits returning true,
|
|
* otherwise return false indicating that this is in fact a non-EOP buffer.
|
|
*/
|
|
static bool idpf_rx_splitq_is_eop(struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc)
|
|
{
|
|
/* if we are the last buffer then there is nothing else to do */
|
|
return likely(idpf_rx_splitq_test_staterr(rx_desc->status_err0_qw1,
|
|
IDPF_RXD_EOF_SPLITQ));
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_splitq_clean - Clean completed descriptors from Rx queue
|
|
* @rxq: Rx descriptor queue to retrieve receive buffer queue
|
|
* @budget: Total limit on number of packets to process
|
|
*
|
|
* This function provides a "bounce buffer" approach to Rx interrupt
|
|
* processing. The advantage to this is that on systems that have
|
|
* expensive overhead for IOMMU access this provides a means of avoiding
|
|
* it by maintaining the mapping of the page to the system.
|
|
*
|
|
* Returns amount of work completed
|
|
*/
|
|
static int idpf_rx_splitq_clean(struct idpf_rx_queue *rxq, int budget)
|
|
{
|
|
int total_rx_bytes = 0, total_rx_pkts = 0;
|
|
struct idpf_buf_queue *rx_bufq = NULL;
|
|
struct sk_buff *skb = rxq->skb;
|
|
u16 ntc = rxq->next_to_clean;
|
|
|
|
/* Process Rx packets bounded by budget */
|
|
while (likely(total_rx_pkts < budget)) {
|
|
struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc;
|
|
struct libeth_fqe *hdr, *rx_buf = NULL;
|
|
struct idpf_sw_queue *refillq = NULL;
|
|
struct idpf_rxq_set *rxq_set = NULL;
|
|
unsigned int pkt_len = 0;
|
|
unsigned int hdr_len = 0;
|
|
u16 gen_id, buf_id = 0;
|
|
int bufq_id;
|
|
u8 rxdid;
|
|
|
|
/* get the Rx desc from Rx queue based on 'next_to_clean' */
|
|
rx_desc = &rxq->rx[ntc].flex_adv_nic_3_wb;
|
|
|
|
/* This memory barrier is needed to keep us from reading
|
|
* any other fields out of the rx_desc
|
|
*/
|
|
dma_rmb();
|
|
|
|
/* if the descriptor isn't done, no work yet to do */
|
|
gen_id = le16_get_bits(rx_desc->pktlen_gen_bufq_id,
|
|
VIRTCHNL2_RX_FLEX_DESC_ADV_GEN_M);
|
|
|
|
if (idpf_queue_has(GEN_CHK, rxq) != gen_id)
|
|
break;
|
|
|
|
rxdid = FIELD_GET(VIRTCHNL2_RX_FLEX_DESC_ADV_RXDID_M,
|
|
rx_desc->rxdid_ucast);
|
|
if (rxdid != VIRTCHNL2_RXDID_2_FLEX_SPLITQ) {
|
|
IDPF_RX_BUMP_NTC(rxq, ntc);
|
|
u64_stats_update_begin(&rxq->stats_sync);
|
|
u64_stats_inc(&rxq->q_stats.bad_descs);
|
|
u64_stats_update_end(&rxq->stats_sync);
|
|
continue;
|
|
}
|
|
|
|
pkt_len = le16_get_bits(rx_desc->pktlen_gen_bufq_id,
|
|
VIRTCHNL2_RX_FLEX_DESC_ADV_LEN_PBUF_M);
|
|
|
|
bufq_id = le16_get_bits(rx_desc->pktlen_gen_bufq_id,
|
|
VIRTCHNL2_RX_FLEX_DESC_ADV_BUFQ_ID_M);
|
|
|
|
rxq_set = container_of(rxq, struct idpf_rxq_set, rxq);
|
|
refillq = rxq_set->refillq[bufq_id];
|
|
|
|
/* retrieve buffer from the rxq */
|
|
rx_bufq = &rxq->bufq_sets[bufq_id].bufq;
|
|
|
|
buf_id = le16_to_cpu(rx_desc->buf_id);
|
|
|
|
rx_buf = &rx_bufq->buf[buf_id];
|
|
|
|
if (!rx_bufq->hdr_pp)
|
|
goto payload;
|
|
|
|
#define __HBO_BIT VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_HBO_M
|
|
#define __HDR_LEN_MASK VIRTCHNL2_RX_FLEX_DESC_ADV_LEN_HDR_M
|
|
if (likely(!(rx_desc->status_err0_qw1 & __HBO_BIT)))
|
|
/* If a header buffer overflow, occurs, i.e. header is
|
|
* too large to fit in the header split buffer, HW will
|
|
* put the entire packet, including headers, in the
|
|
* data/payload buffer.
|
|
*/
|
|
hdr_len = le16_get_bits(rx_desc->hdrlen_flags,
|
|
__HDR_LEN_MASK);
|
|
#undef __HDR_LEN_MASK
|
|
#undef __HBO_BIT
|
|
|
|
hdr = &rx_bufq->hdr_buf[buf_id];
|
|
|
|
if (unlikely(!hdr_len && !skb)) {
|
|
hdr_len = idpf_rx_hsplit_wa(hdr, rx_buf, pkt_len);
|
|
pkt_len -= hdr_len;
|
|
|
|
u64_stats_update_begin(&rxq->stats_sync);
|
|
u64_stats_inc(&rxq->q_stats.hsplit_buf_ovf);
|
|
u64_stats_update_end(&rxq->stats_sync);
|
|
}
|
|
|
|
if (libeth_rx_sync_for_cpu(hdr, hdr_len)) {
|
|
skb = idpf_rx_build_skb(hdr, hdr_len);
|
|
if (!skb)
|
|
break;
|
|
|
|
u64_stats_update_begin(&rxq->stats_sync);
|
|
u64_stats_inc(&rxq->q_stats.hsplit_pkts);
|
|
u64_stats_update_end(&rxq->stats_sync);
|
|
}
|
|
|
|
hdr->page = NULL;
|
|
|
|
payload:
|
|
if (!libeth_rx_sync_for_cpu(rx_buf, pkt_len))
|
|
goto skip_data;
|
|
|
|
if (skb)
|
|
idpf_rx_add_frag(rx_buf, skb, pkt_len);
|
|
else
|
|
skb = idpf_rx_build_skb(rx_buf, pkt_len);
|
|
|
|
/* exit if we failed to retrieve a buffer */
|
|
if (!skb)
|
|
break;
|
|
|
|
skip_data:
|
|
rx_buf->page = NULL;
|
|
|
|
idpf_rx_post_buf_refill(refillq, buf_id);
|
|
IDPF_RX_BUMP_NTC(rxq, ntc);
|
|
|
|
/* skip if it is non EOP desc */
|
|
if (!idpf_rx_splitq_is_eop(rx_desc) || unlikely(!skb))
|
|
continue;
|
|
|
|
/* pad skb if needed (to make valid ethernet frame) */
|
|
if (eth_skb_pad(skb)) {
|
|
skb = NULL;
|
|
continue;
|
|
}
|
|
|
|
/* probably a little skewed due to removing CRC */
|
|
total_rx_bytes += skb->len;
|
|
|
|
/* protocol */
|
|
if (unlikely(idpf_rx_process_skb_fields(rxq, skb, rx_desc))) {
|
|
dev_kfree_skb_any(skb);
|
|
skb = NULL;
|
|
continue;
|
|
}
|
|
|
|
/* send completed skb up the stack */
|
|
napi_gro_receive(rxq->napi, skb);
|
|
skb = NULL;
|
|
|
|
/* update budget accounting */
|
|
total_rx_pkts++;
|
|
}
|
|
|
|
rxq->next_to_clean = ntc;
|
|
|
|
rxq->skb = skb;
|
|
u64_stats_update_begin(&rxq->stats_sync);
|
|
u64_stats_add(&rxq->q_stats.packets, total_rx_pkts);
|
|
u64_stats_add(&rxq->q_stats.bytes, total_rx_bytes);
|
|
u64_stats_update_end(&rxq->stats_sync);
|
|
|
|
/* guarantee a trip back through this routine if there was a failure */
|
|
return total_rx_pkts;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_update_bufq_desc - Update buffer queue descriptor
|
|
* @bufq: Pointer to the buffer queue
|
|
* @buf_id: buffer ID
|
|
* @buf_desc: Buffer queue descriptor
|
|
*
|
|
* Return 0 on success and negative on failure.
|
|
*/
|
|
static int idpf_rx_update_bufq_desc(struct idpf_buf_queue *bufq, u32 buf_id,
|
|
struct virtchnl2_splitq_rx_buf_desc *buf_desc)
|
|
{
|
|
struct libeth_fq_fp fq = {
|
|
.pp = bufq->pp,
|
|
.fqes = bufq->buf,
|
|
.truesize = bufq->truesize,
|
|
.count = bufq->desc_count,
|
|
};
|
|
dma_addr_t addr;
|
|
|
|
addr = libeth_rx_alloc(&fq, buf_id);
|
|
if (addr == DMA_MAPPING_ERROR)
|
|
return -ENOMEM;
|
|
|
|
buf_desc->pkt_addr = cpu_to_le64(addr);
|
|
buf_desc->qword0.buf_id = cpu_to_le16(buf_id);
|
|
|
|
if (!idpf_queue_has(HSPLIT_EN, bufq))
|
|
return 0;
|
|
|
|
fq.pp = bufq->hdr_pp;
|
|
fq.fqes = bufq->hdr_buf;
|
|
fq.truesize = bufq->hdr_truesize;
|
|
|
|
addr = libeth_rx_alloc(&fq, buf_id);
|
|
if (addr == DMA_MAPPING_ERROR)
|
|
return -ENOMEM;
|
|
|
|
buf_desc->hdr_addr = cpu_to_le64(addr);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_clean_refillq - Clean refill queue buffers
|
|
* @bufq: buffer queue to post buffers back to
|
|
* @refillq: refill queue to clean
|
|
*
|
|
* This function takes care of the buffer refill management
|
|
*/
|
|
static void idpf_rx_clean_refillq(struct idpf_buf_queue *bufq,
|
|
struct idpf_sw_queue *refillq)
|
|
{
|
|
struct virtchnl2_splitq_rx_buf_desc *buf_desc;
|
|
u16 bufq_nta = bufq->next_to_alloc;
|
|
u16 ntc = refillq->next_to_clean;
|
|
int cleaned = 0;
|
|
|
|
buf_desc = &bufq->split_buf[bufq_nta];
|
|
|
|
/* make sure we stop at ring wrap in the unlikely case ring is full */
|
|
while (likely(cleaned < refillq->desc_count)) {
|
|
u32 buf_id, refill_desc = refillq->ring[ntc];
|
|
bool failure;
|
|
|
|
if (idpf_queue_has(RFL_GEN_CHK, refillq) !=
|
|
!!(refill_desc & IDPF_RX_BI_GEN_M))
|
|
break;
|
|
|
|
buf_id = FIELD_GET(IDPF_RX_BI_BUFID_M, refill_desc);
|
|
failure = idpf_rx_update_bufq_desc(bufq, buf_id, buf_desc);
|
|
if (failure)
|
|
break;
|
|
|
|
if (unlikely(++ntc == refillq->desc_count)) {
|
|
idpf_queue_change(RFL_GEN_CHK, refillq);
|
|
ntc = 0;
|
|
}
|
|
|
|
if (unlikely(++bufq_nta == bufq->desc_count)) {
|
|
buf_desc = &bufq->split_buf[0];
|
|
bufq_nta = 0;
|
|
} else {
|
|
buf_desc++;
|
|
}
|
|
|
|
cleaned++;
|
|
}
|
|
|
|
if (!cleaned)
|
|
return;
|
|
|
|
/* We want to limit how many transactions on the bus we trigger with
|
|
* tail writes so we only do it in strides. It's also important we
|
|
* align the write to a multiple of 8 as required by HW.
|
|
*/
|
|
if (((bufq->next_to_use <= bufq_nta ? 0 : bufq->desc_count) +
|
|
bufq_nta - bufq->next_to_use) >= IDPF_RX_BUF_POST_STRIDE)
|
|
idpf_rx_buf_hw_update(bufq, ALIGN_DOWN(bufq_nta,
|
|
IDPF_RX_BUF_POST_STRIDE));
|
|
|
|
/* update next to alloc since we have filled the ring */
|
|
refillq->next_to_clean = ntc;
|
|
bufq->next_to_alloc = bufq_nta;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_clean_refillq_all - Clean all refill queues
|
|
* @bufq: buffer queue with refill queues
|
|
* @nid: ID of the closest NUMA node with memory
|
|
*
|
|
* Iterates through all refill queues assigned to the buffer queue assigned to
|
|
* this vector. Returns true if clean is complete within budget, false
|
|
* otherwise.
|
|
*/
|
|
static void idpf_rx_clean_refillq_all(struct idpf_buf_queue *bufq, int nid)
|
|
{
|
|
struct idpf_bufq_set *bufq_set;
|
|
int i;
|
|
|
|
page_pool_nid_changed(bufq->pp, nid);
|
|
if (bufq->hdr_pp)
|
|
page_pool_nid_changed(bufq->hdr_pp, nid);
|
|
|
|
bufq_set = container_of(bufq, struct idpf_bufq_set, bufq);
|
|
for (i = 0; i < bufq_set->num_refillqs; i++)
|
|
idpf_rx_clean_refillq(bufq, &bufq_set->refillqs[i]);
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_clean_queues - MSIX mode Interrupt Handler
|
|
* @irq: interrupt number
|
|
* @data: pointer to a q_vector
|
|
*
|
|
*/
|
|
static irqreturn_t idpf_vport_intr_clean_queues(int __always_unused irq,
|
|
void *data)
|
|
{
|
|
struct idpf_q_vector *q_vector = (struct idpf_q_vector *)data;
|
|
|
|
q_vector->total_events++;
|
|
napi_schedule(&q_vector->napi);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_napi_del_all - Unregister napi for all q_vectors in vport
|
|
* @vport: virtual port structure
|
|
*
|
|
*/
|
|
static void idpf_vport_intr_napi_del_all(struct idpf_vport *vport)
|
|
{
|
|
u16 v_idx;
|
|
|
|
for (v_idx = 0; v_idx < vport->num_q_vectors; v_idx++)
|
|
netif_napi_del(&vport->q_vectors[v_idx].napi);
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_napi_dis_all - Disable NAPI for all q_vectors in the vport
|
|
* @vport: main vport structure
|
|
*/
|
|
static void idpf_vport_intr_napi_dis_all(struct idpf_vport *vport)
|
|
{
|
|
int v_idx;
|
|
|
|
for (v_idx = 0; v_idx < vport->num_q_vectors; v_idx++)
|
|
napi_disable(&vport->q_vectors[v_idx].napi);
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_rel - Free memory allocated for interrupt vectors
|
|
* @vport: virtual port
|
|
*
|
|
* Free the memory allocated for interrupt vectors associated to a vport
|
|
*/
|
|
void idpf_vport_intr_rel(struct idpf_vport *vport)
|
|
{
|
|
for (u32 v_idx = 0; v_idx < vport->num_q_vectors; v_idx++) {
|
|
struct idpf_q_vector *q_vector = &vport->q_vectors[v_idx];
|
|
|
|
kfree(q_vector->complq);
|
|
q_vector->complq = NULL;
|
|
kfree(q_vector->bufq);
|
|
q_vector->bufq = NULL;
|
|
kfree(q_vector->tx);
|
|
q_vector->tx = NULL;
|
|
kfree(q_vector->rx);
|
|
q_vector->rx = NULL;
|
|
|
|
free_cpumask_var(q_vector->affinity_mask);
|
|
}
|
|
|
|
kfree(vport->q_vectors);
|
|
vport->q_vectors = NULL;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_rel_irq - Free the IRQ association with the OS
|
|
* @vport: main vport structure
|
|
*/
|
|
static void idpf_vport_intr_rel_irq(struct idpf_vport *vport)
|
|
{
|
|
struct idpf_adapter *adapter = vport->adapter;
|
|
int vector;
|
|
|
|
for (vector = 0; vector < vport->num_q_vectors; vector++) {
|
|
struct idpf_q_vector *q_vector = &vport->q_vectors[vector];
|
|
int irq_num, vidx;
|
|
|
|
/* free only the irqs that were actually requested */
|
|
if (!q_vector)
|
|
continue;
|
|
|
|
vidx = vport->q_vector_idxs[vector];
|
|
irq_num = adapter->msix_entries[vidx].vector;
|
|
|
|
/* clear the affinity_mask in the IRQ descriptor */
|
|
irq_set_affinity_hint(irq_num, NULL);
|
|
kfree(free_irq(irq_num, q_vector));
|
|
}
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_dis_irq_all - Disable all interrupt
|
|
* @vport: main vport structure
|
|
*/
|
|
static void idpf_vport_intr_dis_irq_all(struct idpf_vport *vport)
|
|
{
|
|
struct idpf_q_vector *q_vector = vport->q_vectors;
|
|
int q_idx;
|
|
|
|
for (q_idx = 0; q_idx < vport->num_q_vectors; q_idx++)
|
|
writel(0, q_vector[q_idx].intr_reg.dyn_ctl);
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_buildreg_itr - Enable default interrupt generation settings
|
|
* @q_vector: pointer to q_vector
|
|
*/
|
|
static u32 idpf_vport_intr_buildreg_itr(struct idpf_q_vector *q_vector)
|
|
{
|
|
u32 itr_val = q_vector->intr_reg.dyn_ctl_intena_m;
|
|
int type = IDPF_NO_ITR_UPDATE_IDX;
|
|
u16 itr = 0;
|
|
|
|
if (q_vector->wb_on_itr) {
|
|
/*
|
|
* Trigger a software interrupt when exiting wb_on_itr, to make
|
|
* sure we catch any pending write backs that might have been
|
|
* missed due to interrupt state transition.
|
|
*/
|
|
itr_val |= q_vector->intr_reg.dyn_ctl_swint_trig_m |
|
|
q_vector->intr_reg.dyn_ctl_sw_itridx_ena_m;
|
|
type = IDPF_SW_ITR_UPDATE_IDX;
|
|
itr = IDPF_ITR_20K;
|
|
}
|
|
|
|
itr &= IDPF_ITR_MASK;
|
|
/* Don't clear PBA because that can cause lost interrupts that
|
|
* came in while we were cleaning/polling
|
|
*/
|
|
itr_val |= (type << q_vector->intr_reg.dyn_ctl_itridx_s) |
|
|
(itr << (q_vector->intr_reg.dyn_ctl_intrvl_s - 1));
|
|
|
|
return itr_val;
|
|
}
|
|
|
|
/**
|
|
* idpf_update_dim_sample - Update dim sample with packets and bytes
|
|
* @q_vector: the vector associated with the interrupt
|
|
* @dim_sample: dim sample to update
|
|
* @dim: dim instance structure
|
|
* @packets: total packets
|
|
* @bytes: total bytes
|
|
*
|
|
* Update the dim sample with the packets and bytes which are passed to this
|
|
* function. Set the dim state appropriately if the dim settings gets stale.
|
|
*/
|
|
static void idpf_update_dim_sample(struct idpf_q_vector *q_vector,
|
|
struct dim_sample *dim_sample,
|
|
struct dim *dim, u64 packets, u64 bytes)
|
|
{
|
|
dim_update_sample(q_vector->total_events, packets, bytes, dim_sample);
|
|
dim_sample->comp_ctr = 0;
|
|
|
|
/* if dim settings get stale, like when not updated for 1 second or
|
|
* longer, force it to start again. This addresses the frequent case
|
|
* of an idle queue being switched to by the scheduler.
|
|
*/
|
|
if (ktime_ms_delta(dim_sample->time, dim->start_sample.time) >= HZ)
|
|
dim->state = DIM_START_MEASURE;
|
|
}
|
|
|
|
/**
|
|
* idpf_net_dim - Update net DIM algorithm
|
|
* @q_vector: the vector associated with the interrupt
|
|
*
|
|
* Create a DIM sample and notify net_dim() so that it can possibly decide
|
|
* a new ITR value based on incoming packets, bytes, and interrupts.
|
|
*
|
|
* This function is a no-op if the queue is not configured to dynamic ITR.
|
|
*/
|
|
static void idpf_net_dim(struct idpf_q_vector *q_vector)
|
|
{
|
|
struct dim_sample dim_sample = { };
|
|
u64 packets, bytes;
|
|
u32 i;
|
|
|
|
if (!IDPF_ITR_IS_DYNAMIC(q_vector->tx_intr_mode))
|
|
goto check_rx_itr;
|
|
|
|
for (i = 0, packets = 0, bytes = 0; i < q_vector->num_txq; i++) {
|
|
struct idpf_tx_queue *txq = q_vector->tx[i];
|
|
unsigned int start;
|
|
|
|
do {
|
|
start = u64_stats_fetch_begin(&txq->stats_sync);
|
|
packets += u64_stats_read(&txq->q_stats.packets);
|
|
bytes += u64_stats_read(&txq->q_stats.bytes);
|
|
} while (u64_stats_fetch_retry(&txq->stats_sync, start));
|
|
}
|
|
|
|
idpf_update_dim_sample(q_vector, &dim_sample, &q_vector->tx_dim,
|
|
packets, bytes);
|
|
net_dim(&q_vector->tx_dim, &dim_sample);
|
|
|
|
check_rx_itr:
|
|
if (!IDPF_ITR_IS_DYNAMIC(q_vector->rx_intr_mode))
|
|
return;
|
|
|
|
for (i = 0, packets = 0, bytes = 0; i < q_vector->num_rxq; i++) {
|
|
struct idpf_rx_queue *rxq = q_vector->rx[i];
|
|
unsigned int start;
|
|
|
|
do {
|
|
start = u64_stats_fetch_begin(&rxq->stats_sync);
|
|
packets += u64_stats_read(&rxq->q_stats.packets);
|
|
bytes += u64_stats_read(&rxq->q_stats.bytes);
|
|
} while (u64_stats_fetch_retry(&rxq->stats_sync, start));
|
|
}
|
|
|
|
idpf_update_dim_sample(q_vector, &dim_sample, &q_vector->rx_dim,
|
|
packets, bytes);
|
|
net_dim(&q_vector->rx_dim, &dim_sample);
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_update_itr_ena_irq - Update itr and re-enable MSIX interrupt
|
|
* @q_vector: q_vector for which itr is being updated and interrupt enabled
|
|
*
|
|
* Update the net_dim() algorithm and re-enable the interrupt associated with
|
|
* this vector.
|
|
*/
|
|
void idpf_vport_intr_update_itr_ena_irq(struct idpf_q_vector *q_vector)
|
|
{
|
|
u32 intval;
|
|
|
|
/* net_dim() updates ITR out-of-band using a work item */
|
|
idpf_net_dim(q_vector);
|
|
|
|
intval = idpf_vport_intr_buildreg_itr(q_vector);
|
|
q_vector->wb_on_itr = false;
|
|
|
|
writel(intval, q_vector->intr_reg.dyn_ctl);
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_req_irq - get MSI-X vectors from the OS for the vport
|
|
* @vport: main vport structure
|
|
*/
|
|
static int idpf_vport_intr_req_irq(struct idpf_vport *vport)
|
|
{
|
|
struct idpf_adapter *adapter = vport->adapter;
|
|
const char *drv_name, *if_name, *vec_name;
|
|
int vector, err, irq_num, vidx;
|
|
|
|
drv_name = dev_driver_string(&adapter->pdev->dev);
|
|
if_name = netdev_name(vport->netdev);
|
|
|
|
for (vector = 0; vector < vport->num_q_vectors; vector++) {
|
|
struct idpf_q_vector *q_vector = &vport->q_vectors[vector];
|
|
char *name;
|
|
|
|
vidx = vport->q_vector_idxs[vector];
|
|
irq_num = adapter->msix_entries[vidx].vector;
|
|
|
|
if (q_vector->num_rxq && q_vector->num_txq)
|
|
vec_name = "TxRx";
|
|
else if (q_vector->num_rxq)
|
|
vec_name = "Rx";
|
|
else if (q_vector->num_txq)
|
|
vec_name = "Tx";
|
|
else
|
|
continue;
|
|
|
|
name = kasprintf(GFP_KERNEL, "%s-%s-%s-%d", drv_name, if_name,
|
|
vec_name, vidx);
|
|
|
|
err = request_irq(irq_num, idpf_vport_intr_clean_queues, 0,
|
|
name, q_vector);
|
|
if (err) {
|
|
netdev_err(vport->netdev,
|
|
"Request_irq failed, error: %d\n", err);
|
|
goto free_q_irqs;
|
|
}
|
|
/* assign the mask for this irq */
|
|
irq_set_affinity_hint(irq_num, q_vector->affinity_mask);
|
|
}
|
|
|
|
return 0;
|
|
|
|
free_q_irqs:
|
|
while (--vector >= 0) {
|
|
vidx = vport->q_vector_idxs[vector];
|
|
irq_num = adapter->msix_entries[vidx].vector;
|
|
kfree(free_irq(irq_num, &vport->q_vectors[vector]));
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_write_itr - Write ITR value to the ITR register
|
|
* @q_vector: q_vector structure
|
|
* @itr: Interrupt throttling rate
|
|
* @tx: Tx or Rx ITR
|
|
*/
|
|
void idpf_vport_intr_write_itr(struct idpf_q_vector *q_vector, u16 itr, bool tx)
|
|
{
|
|
struct idpf_intr_reg *intr_reg;
|
|
|
|
if (tx && !q_vector->tx)
|
|
return;
|
|
else if (!tx && !q_vector->rx)
|
|
return;
|
|
|
|
intr_reg = &q_vector->intr_reg;
|
|
writel(ITR_REG_ALIGN(itr) >> IDPF_ITR_GRAN_S,
|
|
tx ? intr_reg->tx_itr : intr_reg->rx_itr);
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_ena_irq_all - Enable IRQ for the given vport
|
|
* @vport: main vport structure
|
|
*/
|
|
static void idpf_vport_intr_ena_irq_all(struct idpf_vport *vport)
|
|
{
|
|
bool dynamic;
|
|
int q_idx;
|
|
u16 itr;
|
|
|
|
for (q_idx = 0; q_idx < vport->num_q_vectors; q_idx++) {
|
|
struct idpf_q_vector *qv = &vport->q_vectors[q_idx];
|
|
|
|
/* Set the initial ITR values */
|
|
if (qv->num_txq) {
|
|
dynamic = IDPF_ITR_IS_DYNAMIC(qv->tx_intr_mode);
|
|
itr = vport->tx_itr_profile[qv->tx_dim.profile_ix];
|
|
idpf_vport_intr_write_itr(qv, dynamic ?
|
|
itr : qv->tx_itr_value,
|
|
true);
|
|
}
|
|
|
|
if (qv->num_rxq) {
|
|
dynamic = IDPF_ITR_IS_DYNAMIC(qv->rx_intr_mode);
|
|
itr = vport->rx_itr_profile[qv->rx_dim.profile_ix];
|
|
idpf_vport_intr_write_itr(qv, dynamic ?
|
|
itr : qv->rx_itr_value,
|
|
false);
|
|
}
|
|
|
|
if (qv->num_txq || qv->num_rxq)
|
|
idpf_vport_intr_update_itr_ena_irq(qv);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_deinit - Release all vector associations for the vport
|
|
* @vport: main vport structure
|
|
*/
|
|
void idpf_vport_intr_deinit(struct idpf_vport *vport)
|
|
{
|
|
idpf_vport_intr_dis_irq_all(vport);
|
|
idpf_vport_intr_napi_dis_all(vport);
|
|
idpf_vport_intr_napi_del_all(vport);
|
|
idpf_vport_intr_rel_irq(vport);
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_dim_work - Call back from the stack
|
|
* @work: work queue structure
|
|
*/
|
|
static void idpf_tx_dim_work(struct work_struct *work)
|
|
{
|
|
struct idpf_q_vector *q_vector;
|
|
struct idpf_vport *vport;
|
|
struct dim *dim;
|
|
u16 itr;
|
|
|
|
dim = container_of(work, struct dim, work);
|
|
q_vector = container_of(dim, struct idpf_q_vector, tx_dim);
|
|
vport = q_vector->vport;
|
|
|
|
if (dim->profile_ix >= ARRAY_SIZE(vport->tx_itr_profile))
|
|
dim->profile_ix = ARRAY_SIZE(vport->tx_itr_profile) - 1;
|
|
|
|
/* look up the values in our local table */
|
|
itr = vport->tx_itr_profile[dim->profile_ix];
|
|
|
|
idpf_vport_intr_write_itr(q_vector, itr, true);
|
|
|
|
dim->state = DIM_START_MEASURE;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_dim_work - Call back from the stack
|
|
* @work: work queue structure
|
|
*/
|
|
static void idpf_rx_dim_work(struct work_struct *work)
|
|
{
|
|
struct idpf_q_vector *q_vector;
|
|
struct idpf_vport *vport;
|
|
struct dim *dim;
|
|
u16 itr;
|
|
|
|
dim = container_of(work, struct dim, work);
|
|
q_vector = container_of(dim, struct idpf_q_vector, rx_dim);
|
|
vport = q_vector->vport;
|
|
|
|
if (dim->profile_ix >= ARRAY_SIZE(vport->rx_itr_profile))
|
|
dim->profile_ix = ARRAY_SIZE(vport->rx_itr_profile) - 1;
|
|
|
|
/* look up the values in our local table */
|
|
itr = vport->rx_itr_profile[dim->profile_ix];
|
|
|
|
idpf_vport_intr_write_itr(q_vector, itr, false);
|
|
|
|
dim->state = DIM_START_MEASURE;
|
|
}
|
|
|
|
/**
|
|
* idpf_init_dim - Set up dynamic interrupt moderation
|
|
* @qv: q_vector structure
|
|
*/
|
|
static void idpf_init_dim(struct idpf_q_vector *qv)
|
|
{
|
|
INIT_WORK(&qv->tx_dim.work, idpf_tx_dim_work);
|
|
qv->tx_dim.mode = DIM_CQ_PERIOD_MODE_START_FROM_EQE;
|
|
qv->tx_dim.profile_ix = IDPF_DIM_DEFAULT_PROFILE_IX;
|
|
|
|
INIT_WORK(&qv->rx_dim.work, idpf_rx_dim_work);
|
|
qv->rx_dim.mode = DIM_CQ_PERIOD_MODE_START_FROM_EQE;
|
|
qv->rx_dim.profile_ix = IDPF_DIM_DEFAULT_PROFILE_IX;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_napi_ena_all - Enable NAPI for all q_vectors in the vport
|
|
* @vport: main vport structure
|
|
*/
|
|
static void idpf_vport_intr_napi_ena_all(struct idpf_vport *vport)
|
|
{
|
|
int q_idx;
|
|
|
|
for (q_idx = 0; q_idx < vport->num_q_vectors; q_idx++) {
|
|
struct idpf_q_vector *q_vector = &vport->q_vectors[q_idx];
|
|
|
|
idpf_init_dim(q_vector);
|
|
napi_enable(&q_vector->napi);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* idpf_tx_splitq_clean_all- Clean completion queues
|
|
* @q_vec: queue vector
|
|
* @budget: Used to determine if we are in netpoll
|
|
* @cleaned: returns number of packets cleaned
|
|
*
|
|
* Returns false if clean is not complete else returns true
|
|
*/
|
|
static bool idpf_tx_splitq_clean_all(struct idpf_q_vector *q_vec,
|
|
int budget, int *cleaned)
|
|
{
|
|
u16 num_complq = q_vec->num_complq;
|
|
bool clean_complete = true;
|
|
int i, budget_per_q;
|
|
|
|
if (unlikely(!num_complq))
|
|
return true;
|
|
|
|
budget_per_q = DIV_ROUND_UP(budget, num_complq);
|
|
|
|
for (i = 0; i < num_complq; i++)
|
|
clean_complete &= idpf_tx_clean_complq(q_vec->complq[i],
|
|
budget_per_q, cleaned);
|
|
|
|
return clean_complete;
|
|
}
|
|
|
|
/**
|
|
* idpf_rx_splitq_clean_all- Clean completion queues
|
|
* @q_vec: queue vector
|
|
* @budget: Used to determine if we are in netpoll
|
|
* @cleaned: returns number of packets cleaned
|
|
*
|
|
* Returns false if clean is not complete else returns true
|
|
*/
|
|
static bool idpf_rx_splitq_clean_all(struct idpf_q_vector *q_vec, int budget,
|
|
int *cleaned)
|
|
{
|
|
u16 num_rxq = q_vec->num_rxq;
|
|
bool clean_complete = true;
|
|
int pkts_cleaned = 0;
|
|
int i, budget_per_q;
|
|
int nid;
|
|
|
|
/* We attempt to distribute budget to each Rx queue fairly, but don't
|
|
* allow the budget to go below 1 because that would exit polling early.
|
|
*/
|
|
budget_per_q = num_rxq ? max(budget / num_rxq, 1) : 0;
|
|
for (i = 0; i < num_rxq; i++) {
|
|
struct idpf_rx_queue *rxq = q_vec->rx[i];
|
|
int pkts_cleaned_per_q;
|
|
|
|
pkts_cleaned_per_q = idpf_rx_splitq_clean(rxq, budget_per_q);
|
|
/* if we clean as many as budgeted, we must not be done */
|
|
if (pkts_cleaned_per_q >= budget_per_q)
|
|
clean_complete = false;
|
|
pkts_cleaned += pkts_cleaned_per_q;
|
|
}
|
|
*cleaned = pkts_cleaned;
|
|
|
|
nid = numa_mem_id();
|
|
|
|
for (i = 0; i < q_vec->num_bufq; i++)
|
|
idpf_rx_clean_refillq_all(q_vec->bufq[i], nid);
|
|
|
|
return clean_complete;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_splitq_napi_poll - NAPI handler
|
|
* @napi: struct from which you get q_vector
|
|
* @budget: budget provided by stack
|
|
*/
|
|
static int idpf_vport_splitq_napi_poll(struct napi_struct *napi, int budget)
|
|
{
|
|
struct idpf_q_vector *q_vector =
|
|
container_of(napi, struct idpf_q_vector, napi);
|
|
bool clean_complete;
|
|
int work_done = 0;
|
|
|
|
/* Handle case where we are called by netpoll with a budget of 0 */
|
|
if (unlikely(!budget)) {
|
|
idpf_tx_splitq_clean_all(q_vector, budget, &work_done);
|
|
|
|
return 0;
|
|
}
|
|
|
|
clean_complete = idpf_rx_splitq_clean_all(q_vector, budget, &work_done);
|
|
clean_complete &= idpf_tx_splitq_clean_all(q_vector, budget, &work_done);
|
|
|
|
/* If work not completed, return budget and polling will return */
|
|
if (!clean_complete) {
|
|
idpf_vport_intr_set_wb_on_itr(q_vector);
|
|
return budget;
|
|
}
|
|
|
|
work_done = min_t(int, work_done, budget - 1);
|
|
|
|
/* Exit the polling mode, but don't re-enable interrupts if stack might
|
|
* poll us due to busy-polling
|
|
*/
|
|
if (likely(napi_complete_done(napi, work_done)))
|
|
idpf_vport_intr_update_itr_ena_irq(q_vector);
|
|
else
|
|
idpf_vport_intr_set_wb_on_itr(q_vector);
|
|
|
|
/* Switch to poll mode in the tear-down path after sending disable
|
|
* queues virtchnl message, as the interrupts will be disabled after
|
|
* that
|
|
*/
|
|
if (unlikely(q_vector->num_txq && idpf_queue_has(POLL_MODE,
|
|
q_vector->tx[0])))
|
|
return budget;
|
|
else
|
|
return work_done;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_map_vector_to_qs - Map vectors to queues
|
|
* @vport: virtual port
|
|
*
|
|
* Mapping for vectors to queues
|
|
*/
|
|
static void idpf_vport_intr_map_vector_to_qs(struct idpf_vport *vport)
|
|
{
|
|
bool split = idpf_is_queue_model_split(vport->rxq_model);
|
|
u16 num_txq_grp = vport->num_txq_grp;
|
|
struct idpf_rxq_group *rx_qgrp;
|
|
struct idpf_txq_group *tx_qgrp;
|
|
u32 i, qv_idx, q_index;
|
|
|
|
for (i = 0, qv_idx = 0; i < vport->num_rxq_grp; i++) {
|
|
u16 num_rxq;
|
|
|
|
if (qv_idx >= vport->num_q_vectors)
|
|
qv_idx = 0;
|
|
|
|
rx_qgrp = &vport->rxq_grps[i];
|
|
if (split)
|
|
num_rxq = rx_qgrp->splitq.num_rxq_sets;
|
|
else
|
|
num_rxq = rx_qgrp->singleq.num_rxq;
|
|
|
|
for (u32 j = 0; j < num_rxq; j++) {
|
|
struct idpf_rx_queue *q;
|
|
|
|
if (split)
|
|
q = &rx_qgrp->splitq.rxq_sets[j]->rxq;
|
|
else
|
|
q = rx_qgrp->singleq.rxqs[j];
|
|
q->q_vector = &vport->q_vectors[qv_idx];
|
|
q_index = q->q_vector->num_rxq;
|
|
q->q_vector->rx[q_index] = q;
|
|
q->q_vector->num_rxq++;
|
|
|
|
if (split)
|
|
q->napi = &q->q_vector->napi;
|
|
}
|
|
|
|
if (split) {
|
|
for (u32 j = 0; j < vport->num_bufqs_per_qgrp; j++) {
|
|
struct idpf_buf_queue *bufq;
|
|
|
|
bufq = &rx_qgrp->splitq.bufq_sets[j].bufq;
|
|
bufq->q_vector = &vport->q_vectors[qv_idx];
|
|
q_index = bufq->q_vector->num_bufq;
|
|
bufq->q_vector->bufq[q_index] = bufq;
|
|
bufq->q_vector->num_bufq++;
|
|
}
|
|
}
|
|
|
|
qv_idx++;
|
|
}
|
|
|
|
split = idpf_is_queue_model_split(vport->txq_model);
|
|
|
|
for (i = 0, qv_idx = 0; i < num_txq_grp; i++) {
|
|
u16 num_txq;
|
|
|
|
if (qv_idx >= vport->num_q_vectors)
|
|
qv_idx = 0;
|
|
|
|
tx_qgrp = &vport->txq_grps[i];
|
|
num_txq = tx_qgrp->num_txq;
|
|
|
|
for (u32 j = 0; j < num_txq; j++) {
|
|
struct idpf_tx_queue *q;
|
|
|
|
q = tx_qgrp->txqs[j];
|
|
q->q_vector = &vport->q_vectors[qv_idx];
|
|
q->q_vector->tx[q->q_vector->num_txq++] = q;
|
|
}
|
|
|
|
if (split) {
|
|
struct idpf_compl_queue *q = tx_qgrp->complq;
|
|
|
|
q->q_vector = &vport->q_vectors[qv_idx];
|
|
q->q_vector->complq[q->q_vector->num_complq++] = q;
|
|
}
|
|
|
|
qv_idx++;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_init_vec_idx - Initialize the vector indexes
|
|
* @vport: virtual port
|
|
*
|
|
* Initialize vector indexes with values returened over mailbox
|
|
*/
|
|
static int idpf_vport_intr_init_vec_idx(struct idpf_vport *vport)
|
|
{
|
|
struct idpf_adapter *adapter = vport->adapter;
|
|
struct virtchnl2_alloc_vectors *ac;
|
|
u16 *vecids, total_vecs;
|
|
int i;
|
|
|
|
ac = adapter->req_vec_chunks;
|
|
if (!ac) {
|
|
for (i = 0; i < vport->num_q_vectors; i++)
|
|
vport->q_vectors[i].v_idx = vport->q_vector_idxs[i];
|
|
|
|
return 0;
|
|
}
|
|
|
|
total_vecs = idpf_get_reserved_vecs(adapter);
|
|
vecids = kcalloc(total_vecs, sizeof(u16), GFP_KERNEL);
|
|
if (!vecids)
|
|
return -ENOMEM;
|
|
|
|
idpf_get_vec_ids(adapter, vecids, total_vecs, &ac->vchunks);
|
|
|
|
for (i = 0; i < vport->num_q_vectors; i++)
|
|
vport->q_vectors[i].v_idx = vecids[vport->q_vector_idxs[i]];
|
|
|
|
kfree(vecids);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_napi_add_all- Register napi handler for all qvectors
|
|
* @vport: virtual port structure
|
|
*/
|
|
static void idpf_vport_intr_napi_add_all(struct idpf_vport *vport)
|
|
{
|
|
int (*napi_poll)(struct napi_struct *napi, int budget);
|
|
u16 v_idx;
|
|
|
|
if (idpf_is_queue_model_split(vport->txq_model))
|
|
napi_poll = idpf_vport_splitq_napi_poll;
|
|
else
|
|
napi_poll = idpf_vport_singleq_napi_poll;
|
|
|
|
for (v_idx = 0; v_idx < vport->num_q_vectors; v_idx++) {
|
|
struct idpf_q_vector *q_vector = &vport->q_vectors[v_idx];
|
|
|
|
netif_napi_add(vport->netdev, &q_vector->napi, napi_poll);
|
|
|
|
/* only set affinity_mask if the CPU is online */
|
|
if (cpu_online(v_idx))
|
|
cpumask_set_cpu(v_idx, q_vector->affinity_mask);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_alloc - Allocate memory for interrupt vectors
|
|
* @vport: virtual port
|
|
*
|
|
* We allocate one q_vector per queue interrupt. If allocation fails we
|
|
* return -ENOMEM.
|
|
*/
|
|
int idpf_vport_intr_alloc(struct idpf_vport *vport)
|
|
{
|
|
u16 txqs_per_vector, rxqs_per_vector, bufqs_per_vector;
|
|
struct idpf_q_vector *q_vector;
|
|
u32 complqs_per_vector, v_idx;
|
|
|
|
vport->q_vectors = kcalloc(vport->num_q_vectors,
|
|
sizeof(struct idpf_q_vector), GFP_KERNEL);
|
|
if (!vport->q_vectors)
|
|
return -ENOMEM;
|
|
|
|
txqs_per_vector = DIV_ROUND_UP(vport->num_txq_grp,
|
|
vport->num_q_vectors);
|
|
rxqs_per_vector = DIV_ROUND_UP(vport->num_rxq_grp,
|
|
vport->num_q_vectors);
|
|
bufqs_per_vector = vport->num_bufqs_per_qgrp *
|
|
DIV_ROUND_UP(vport->num_rxq_grp,
|
|
vport->num_q_vectors);
|
|
complqs_per_vector = DIV_ROUND_UP(vport->num_txq_grp,
|
|
vport->num_q_vectors);
|
|
|
|
for (v_idx = 0; v_idx < vport->num_q_vectors; v_idx++) {
|
|
q_vector = &vport->q_vectors[v_idx];
|
|
q_vector->vport = vport;
|
|
|
|
q_vector->tx_itr_value = IDPF_ITR_TX_DEF;
|
|
q_vector->tx_intr_mode = IDPF_ITR_DYNAMIC;
|
|
q_vector->tx_itr_idx = VIRTCHNL2_ITR_IDX_1;
|
|
|
|
q_vector->rx_itr_value = IDPF_ITR_RX_DEF;
|
|
q_vector->rx_intr_mode = IDPF_ITR_DYNAMIC;
|
|
q_vector->rx_itr_idx = VIRTCHNL2_ITR_IDX_0;
|
|
|
|
if (!zalloc_cpumask_var(&q_vector->affinity_mask, GFP_KERNEL))
|
|
goto error;
|
|
|
|
q_vector->tx = kcalloc(txqs_per_vector, sizeof(*q_vector->tx),
|
|
GFP_KERNEL);
|
|
if (!q_vector->tx)
|
|
goto error;
|
|
|
|
q_vector->rx = kcalloc(rxqs_per_vector, sizeof(*q_vector->rx),
|
|
GFP_KERNEL);
|
|
if (!q_vector->rx)
|
|
goto error;
|
|
|
|
if (!idpf_is_queue_model_split(vport->rxq_model))
|
|
continue;
|
|
|
|
q_vector->bufq = kcalloc(bufqs_per_vector,
|
|
sizeof(*q_vector->bufq),
|
|
GFP_KERNEL);
|
|
if (!q_vector->bufq)
|
|
goto error;
|
|
|
|
q_vector->complq = kcalloc(complqs_per_vector,
|
|
sizeof(*q_vector->complq),
|
|
GFP_KERNEL);
|
|
if (!q_vector->complq)
|
|
goto error;
|
|
}
|
|
|
|
return 0;
|
|
|
|
error:
|
|
idpf_vport_intr_rel(vport);
|
|
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/**
|
|
* idpf_vport_intr_init - Setup all vectors for the given vport
|
|
* @vport: virtual port
|
|
*
|
|
* Returns 0 on success or negative on failure
|
|
*/
|
|
int idpf_vport_intr_init(struct idpf_vport *vport)
|
|
{
|
|
int err;
|
|
|
|
err = idpf_vport_intr_init_vec_idx(vport);
|
|
if (err)
|
|
return err;
|
|
|
|
idpf_vport_intr_map_vector_to_qs(vport);
|
|
idpf_vport_intr_napi_add_all(vport);
|
|
|
|
err = vport->adapter->dev_ops.reg_ops.intr_reg_init(vport);
|
|
if (err)
|
|
goto unroll_vectors_alloc;
|
|
|
|
err = idpf_vport_intr_req_irq(vport);
|
|
if (err)
|
|
goto unroll_vectors_alloc;
|
|
|
|
return 0;
|
|
|
|
unroll_vectors_alloc:
|
|
idpf_vport_intr_napi_del_all(vport);
|
|
|
|
return err;
|
|
}
|
|
|
|
void idpf_vport_intr_ena(struct idpf_vport *vport)
|
|
{
|
|
idpf_vport_intr_napi_ena_all(vport);
|
|
idpf_vport_intr_ena_irq_all(vport);
|
|
}
|
|
|
|
/**
|
|
* idpf_config_rss - Send virtchnl messages to configure RSS
|
|
* @vport: virtual port
|
|
*
|
|
* Return 0 on success, negative on failure
|
|
*/
|
|
int idpf_config_rss(struct idpf_vport *vport)
|
|
{
|
|
int err;
|
|
|
|
err = idpf_send_get_set_rss_key_msg(vport, false);
|
|
if (err)
|
|
return err;
|
|
|
|
return idpf_send_get_set_rss_lut_msg(vport, false);
|
|
}
|
|
|
|
/**
|
|
* idpf_fill_dflt_rss_lut - Fill the indirection table with the default values
|
|
* @vport: virtual port structure
|
|
*/
|
|
static void idpf_fill_dflt_rss_lut(struct idpf_vport *vport)
|
|
{
|
|
struct idpf_adapter *adapter = vport->adapter;
|
|
u16 num_active_rxq = vport->num_rxq;
|
|
struct idpf_rss_data *rss_data;
|
|
int i;
|
|
|
|
rss_data = &adapter->vport_config[vport->idx]->user_config.rss_data;
|
|
|
|
for (i = 0; i < rss_data->rss_lut_size; i++) {
|
|
rss_data->rss_lut[i] = i % num_active_rxq;
|
|
rss_data->cached_lut[i] = rss_data->rss_lut[i];
|
|
}
|
|
}
|
|
|
|
/**
|
|
* idpf_init_rss - Allocate and initialize RSS resources
|
|
* @vport: virtual port
|
|
*
|
|
* Return 0 on success, negative on failure
|
|
*/
|
|
int idpf_init_rss(struct idpf_vport *vport)
|
|
{
|
|
struct idpf_adapter *adapter = vport->adapter;
|
|
struct idpf_rss_data *rss_data;
|
|
u32 lut_size;
|
|
|
|
rss_data = &adapter->vport_config[vport->idx]->user_config.rss_data;
|
|
|
|
lut_size = rss_data->rss_lut_size * sizeof(u32);
|
|
rss_data->rss_lut = kzalloc(lut_size, GFP_KERNEL);
|
|
if (!rss_data->rss_lut)
|
|
return -ENOMEM;
|
|
|
|
rss_data->cached_lut = kzalloc(lut_size, GFP_KERNEL);
|
|
if (!rss_data->cached_lut) {
|
|
kfree(rss_data->rss_lut);
|
|
rss_data->rss_lut = NULL;
|
|
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* Fill the default RSS lut values */
|
|
idpf_fill_dflt_rss_lut(vport);
|
|
|
|
return idpf_config_rss(vport);
|
|
}
|
|
|
|
/**
|
|
* idpf_deinit_rss - Release RSS resources
|
|
* @vport: virtual port
|
|
*/
|
|
void idpf_deinit_rss(struct idpf_vport *vport)
|
|
{
|
|
struct idpf_adapter *adapter = vport->adapter;
|
|
struct idpf_rss_data *rss_data;
|
|
|
|
rss_data = &adapter->vport_config[vport->idx]->user_config.rss_data;
|
|
kfree(rss_data->cached_lut);
|
|
rss_data->cached_lut = NULL;
|
|
kfree(rss_data->rss_lut);
|
|
rss_data->rss_lut = NULL;
|
|
}
|