aboutsummaryrefslogtreecommitdiffstatshomepage
path: root/net/core/gro_cells.c
diff options
context:
space:
mode:
authorEric Dumazet <edumazet@google.com>2022-03-10 21:14:20 -0800
committerJakub Kicinski <kuba@kernel.org>2022-03-11 23:17:24 -0800
commit625788b5844511cf4c30cffa7fa0bc3a69cebc82 (patch)
tree1b1b091212352474f56d6c4497017bc206c7c785 /net/core/gro_cells.c
parentMerge branch 'nfp-preliminary-support-for-nfp-3800' (diff)
downloadwireguard-linux-625788b5844511cf4c30cffa7fa0bc3a69cebc82.tar.xz
wireguard-linux-625788b5844511cf4c30cffa7fa0bc3a69cebc82.zip
net: add per-cpu storage and net->core_stats
Before adding yet another possibly contended atomic_long_t, it is time to add per-cpu storage for existing ones: dev->tx_dropped, dev->rx_dropped, and dev->rx_nohandler Because many devices do not have to increment such counters, allocate the per-cpu storage on demand, so that dev_get_stats() does not have to spend considerable time folding zero counters. Note that some drivers have abused these counters which were supposed to be only used by core networking stack. v4: should use per_cpu_ptr() in dev_get_stats() (Jakub) v3: added a READ_ONCE() in netdev_core_stats_alloc() (Paolo) v2: add a missing include (reported by kernel test robot <lkp@intel.com>) Change in netdev_core_stats_alloc() (Jakub) Signed-off-by: Eric Dumazet <edumazet@google.com> Cc: jeffreyji <jeffreyji@google.com> Reviewed-by: Brian Vazquez <brianvv@google.com> Reviewed-by: Jakub Kicinski <kuba@kernel.org> Acked-by: Paolo Abeni <pabeni@redhat.com> Link: https://lore.kernel.org/r/20220311051420.2608812-1-eric.dumazet@gmail.com Signed-off-by: Jakub Kicinski <kuba@kernel.org>
Diffstat (limited to 'net/core/gro_cells.c')
-rw-r--r--net/core/gro_cells.c2
1 files changed, 1 insertions, 1 deletions
diff --git a/net/core/gro_cells.c b/net/core/gro_cells.c
index 8462f926ab45..541c7a72a28a 100644
--- a/net/core/gro_cells.c
+++ b/net/core/gro_cells.c
@@ -28,7 +28,7 @@ int gro_cells_receive(struct gro_cells *gcells, struct sk_buff *skb)
if (skb_queue_len(&cell->napi_skbs) > netdev_max_backlog) {
drop:
- atomic_long_inc(&dev->rx_dropped);
+ dev_core_stats_rx_dropped_inc(dev);
kfree_skb(skb);
res = NET_RX_DROP;
goto unlock;