/* SPDX-License-Identifier: GPL-2.0-only */ /* Atlantic Network Driver * * Copyright (C) 2014-2019 aQuantia Corporation * Copyright (C) 2019-2020 Marvell International Ltd. */ /* File aq_ring.h: Declaration of functions for Rx/Tx rings. */ #ifndef AQ_RING_H #define AQ_RING_H #include "aq_common.h" #include "aq_vec.h" #define AQ_XDP_HEADROOM ALIGN(max(NET_SKB_PAD, XDP_PACKET_HEADROOM), 8) #define AQ_XDP_TAILROOM SKB_DATA_ALIGN(sizeof(struct skb_shared_info)) struct page; struct aq_nic_cfg_s; struct aq_rxpage { struct page *page; dma_addr_t daddr; unsigned int order; unsigned int pg_off; }; /* TxC SOP DX EOP * +----------+----------+----------+----------- * 8bytes|len l3,l4 | pa | pa | pa * +----------+----------+----------+----------- * 4/8bytes|len pkt |len pkt | | skb * +----------+----------+----------+----------- * 4/8bytes|is_gso |len,flags |len |len,is_eop * +----------+----------+----------+----------- * * This aq_ring_buff_s doesn't have endianness dependency. * It is __packed for cache line optimizations. */ struct __packed aq_ring_buff_s { union { /* RX/TX */ dma_addr_t pa; /* RX */ struct { u32 rss_hash; u16 next; u8 is_hash_l4; u8 rsvd1; struct aq_rxpage rxdata; u16 vlan_rx_tag; }; /* EOP */ struct { dma_addr_t pa_eop; struct sk_buff *skb; struct xdp_frame *xdpf; }; /* TxC */ struct { u32 mss; u8 len_l2; u8 len_l3; u8 len_l4; u8 is_ipv6:1; u8 rsvd2:7; u32 len_pkt; u16 vlan_tx_tag; }; }; union { struct { u32 len:16; u32 is_ip_cso:1; u32 is_udp_cso:1; u32 is_tcp_cso:1; u32 is_cso_err:1; u32 is_sop:1; u32 is_eop:1; u32 is_gso_tcp:1; u32 is_gso_udp:1; u32 is_mapped:1; u32 is_cleaned:1; u32 is_error:1; u32 is_vlan:1; u32 is_lro:1; u32 rsvd3:3; u16 eop_index; u16 rsvd4; }; u64 flags; }; }; struct aq_ring_stats_rx_s { struct u64_stats_sync syncp; /* must be first */ u64 errors; u64 packets; u64 bytes; u64 lro_packets; u64 jumbo_packets; u64 alloc_fails; u64 skb_alloc_fails; u64 polls; u64 pg_losts; u64 pg_flips; u64 pg_reuses; u64 xdp_aborted; u64 xdp_drop; u64 xdp_pass; u64 xdp_tx; u64 xdp_invalid; u64 xdp_redirect; }; struct aq_ring_stats_tx_s { struct u64_stats_sync syncp; /* must be first */ u64 errors; u64 packets; u64 bytes; u64 queue_restarts; }; union aq_ring_stats_s { struct aq_ring_stats_rx_s rx; struct aq_ring_stats_tx_s tx; }; enum atl_ring_type { ATL_RING_TX, ATL_RING_RX, }; struct aq_ring_s { struct aq_ring_buff_s *buff_ring; u8 *dx_ring; /* descriptors ring, dma shared mem */ struct aq_nic_s *aq_nic; unsigned int idx; /* for HW layer registers operations */ unsigned int hw_head; unsigned int sw_head; unsigned int sw_tail; unsigned int size; /* descriptors number */ unsigned int dx_size; /* TX or RX descriptor size, */ /* stored here for fater math */ u16 page_order; u16 page_offset; u16 frame_max; u16 tail_size; union aq_ring_stats_s stats; dma_addr_t dx_ring_pa; struct bpf_prog *xdp_prog; enum atl_ring_type ring_type; struct xdp_rxq_info xdp_rxq; }; struct aq_ring_param_s { unsigned int vec_idx; unsigned int cpu; cpumask_t affinity_mask; }; static inline void *aq_buf_vaddr(struct aq_rxpage *rxpage) { return page_to_virt(rxpage->page) + rxpage->pg_off; } static inline dma_addr_t aq_buf_daddr(struct aq_rxpage *rxpage) { return rxpage->daddr + rxpage->pg_off; } static inline unsigned int aq_ring_next_dx(struct aq_ring_s *self, unsigned int dx) { return (++dx >= self->size) ? 0U : dx; } static inline unsigned int aq_ring_avail_dx(struct aq_ring_s *self) { return (((self->sw_tail >= self->sw_head)) ? (self->size - 1) - self->sw_tail + self->sw_head : self->sw_head - self->sw_tail - 1); } struct aq_ring_s *aq_ring_tx_alloc(struct aq_ring_s *self, struct aq_nic_s *aq_nic, unsigned int idx, struct aq_nic_cfg_s *aq_nic_cfg); struct aq_ring_s *aq_ring_rx_alloc(struct aq_ring_s *self, struct aq_nic_s *aq_nic, unsigned int idx, struct aq_nic_cfg_s *aq_nic_cfg); int aq_ring_init(struct aq_ring_s *self, const enum atl_ring_type ring_type); void aq_ring_rx_deinit(struct aq_ring_s *self); void aq_ring_free(struct aq_ring_s *self); void aq_ring_update_queue_state(struct aq_ring_s *ring); void aq_ring_queue_wake(struct aq_ring_s *ring); void aq_ring_queue_stop(struct aq_ring_s *ring); bool aq_ring_tx_clean(struct aq_ring_s *self); int aq_xdp_xmit(struct net_device *dev, int num_frames, struct xdp_frame **frames, u32 flags); int aq_ring_rx_clean(struct aq_ring_s *self, struct napi_struct *napi, int *work_done, int budget); int aq_ring_rx_fill(struct aq_ring_s *self); struct aq_ring_s *aq_ring_hwts_rx_alloc(struct aq_ring_s *self, struct aq_nic_s *aq_nic, unsigned int idx, unsigned int size, unsigned int dx_size); void aq_ring_hwts_rx_clean(struct aq_ring_s *self, struct aq_nic_s *aq_nic); unsigned int aq_ring_fill_stats_data(struct aq_ring_s *self, u64 *data); #endif /* AQ_RING_H */