100 lines
2.5 KiB
C
100 lines
2.5 KiB
C
|
// SPDX-License-Identifier: (GPL-2.0 OR MIT)
|
||
|
/* Google virtual Ethernet (gve) driver
|
||
|
*
|
||
|
* Copyright (C) 2015-2021 Google, Inc.
|
||
|
*/
|
||
|
|
||
|
#include "gve.h"
|
||
|
#include "gve_adminq.h"
|
||
|
#include "gve_utils.h"
|
||
|
|
||
|
void gve_tx_remove_from_block(struct gve_priv *priv, int queue_idx)
|
||
|
{
|
||
|
struct gve_notify_block *block =
|
||
|
&priv->ntfy_blocks[gve_tx_idx_to_ntfy(priv, queue_idx)];
|
||
|
|
||
|
block->tx = NULL;
|
||
|
}
|
||
|
|
||
|
void gve_tx_add_to_block(struct gve_priv *priv, int queue_idx)
|
||
|
{
|
||
|
unsigned int active_cpus = min_t(int, priv->num_ntfy_blks / 2,
|
||
|
num_online_cpus());
|
||
|
int ntfy_idx = gve_tx_idx_to_ntfy(priv, queue_idx);
|
||
|
struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
|
||
|
struct gve_tx_ring *tx = &priv->tx[queue_idx];
|
||
|
|
||
|
block->tx = tx;
|
||
|
tx->ntfy_id = ntfy_idx;
|
||
|
netif_set_xps_queue(priv->dev, get_cpu_mask(ntfy_idx % active_cpus),
|
||
|
queue_idx);
|
||
|
}
|
||
|
|
||
|
void gve_rx_remove_from_block(struct gve_priv *priv, int queue_idx)
|
||
|
{
|
||
|
struct gve_notify_block *block =
|
||
|
&priv->ntfy_blocks[gve_rx_idx_to_ntfy(priv, queue_idx)];
|
||
|
|
||
|
block->rx = NULL;
|
||
|
}
|
||
|
|
||
|
void gve_rx_add_to_block(struct gve_priv *priv, int queue_idx)
|
||
|
{
|
||
|
u32 ntfy_idx = gve_rx_idx_to_ntfy(priv, queue_idx);
|
||
|
struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
|
||
|
struct gve_rx_ring *rx = &priv->rx[queue_idx];
|
||
|
|
||
|
block->rx = rx;
|
||
|
rx->ntfy_id = ntfy_idx;
|
||
|
}
|
||
|
|
||
|
struct sk_buff *gve_rx_copy(struct net_device *dev, struct napi_struct *napi,
|
||
|
struct gve_rx_slot_page_info *page_info, u16 len,
|
||
|
u16 padding, struct gve_rx_ctx *ctx)
|
||
|
{
|
||
|
void *va = page_info->page_address + padding + page_info->page_offset;
|
||
|
int skb_linear_offset = 0;
|
||
|
bool set_protocol = false;
|
||
|
struct sk_buff *skb;
|
||
|
|
||
|
if (ctx) {
|
||
|
if (!ctx->skb_head)
|
||
|
ctx->skb_head = napi_alloc_skb(napi, ctx->total_expected_size);
|
||
|
|
||
|
if (unlikely(!ctx->skb_head))
|
||
|
return NULL;
|
||
|
skb = ctx->skb_head;
|
||
|
skb_linear_offset = skb->len;
|
||
|
set_protocol = ctx->curr_frag_cnt == ctx->expected_frag_cnt - 1;
|
||
|
} else {
|
||
|
skb = napi_alloc_skb(napi, len);
|
||
|
|
||
|
if (unlikely(!skb))
|
||
|
return NULL;
|
||
|
set_protocol = true;
|
||
|
}
|
||
|
__skb_put(skb, len);
|
||
|
skb_copy_to_linear_data_offset(skb, skb_linear_offset, va, len);
|
||
|
|
||
|
if (set_protocol)
|
||
|
skb->protocol = eth_type_trans(skb, dev);
|
||
|
|
||
|
return skb;
|
||
|
}
|
||
|
|
||
|
void gve_dec_pagecnt_bias(struct gve_rx_slot_page_info *page_info)
|
||
|
{
|
||
|
page_info->pagecnt_bias--;
|
||
|
if (page_info->pagecnt_bias == 0) {
|
||
|
int pagecount = page_count(page_info->page);
|
||
|
|
||
|
/* If we have run out of bias - set it back up to INT_MAX
|
||
|
* minus the existing refs.
|
||
|
*/
|
||
|
page_info->pagecnt_bias = INT_MAX - pagecount;
|
||
|
|
||
|
/* Set pagecount back up to max. */
|
||
|
page_ref_add(page_info->page, INT_MAX - pagecount);
|
||
|
}
|
||
|
}
|