/linux-6.8/drivers/media/common/videobuf2/ |
D | videobuf2-core.c | 36 #define PLANE_INDEX_MASK (BIT_MASK(PLANE_INDEX_BITS) - 1) 38 #define BUFFER_INDEX_MASK (MAX_BUFFER_INDEX - 1) 47 #define dprintk(q, level, fmt, arg...) \ argument 50 pr_info("[%s] %s: " fmt, (q)->name, __func__, \ 103 #define log_qop(q, op) \ argument 104 dprintk(q, 2, "call_qop(%s)%s\n", #op, \ 105 (q)->ops->op ? "" : " (nop)") 107 #define call_qop(q, op, args...) \ argument 111 log_qop(q, op); \ 112 err = (q)->ops->op ? (q)->ops->op(args) : 0; \ [all …]
|
/linux-6.8/net/sched/ |
D | sch_choke.c | 46 #define CHOKE_MAX_QUEUE (128*1024 - 1) 69 unsigned int tab_mask; /* size - 1 */ 75 static unsigned int choke_len(const struct choke_sched_data *q) in choke_len() argument 77 return (q->tail - q->head) & q->tab_mask; in choke_len() 81 static int use_ecn(const struct choke_sched_data *q) in use_ecn() argument 83 return q->flags & TC_RED_ECN; in use_ecn() 87 static int use_harddrop(const struct choke_sched_data *q) in use_harddrop() argument 89 return q->flags & TC_RED_HARDDROP; in use_harddrop() 93 static void choke_zap_head_holes(struct choke_sched_data *q) in choke_zap_head_holes() argument 96 q->head = (q->head + 1) & q->tab_mask; in choke_zap_head_holes() [all …]
|
D | sch_skbprio.c | 40 static u16 calc_new_high_prio(const struct skbprio_sched_data *q) in calc_new_high_prio() argument 44 for (prio = q->highest_prio - 1; prio >= q->lowest_prio; prio--) { in calc_new_high_prio() 45 if (!skb_queue_empty(&q->qdiscs[prio])) in calc_new_high_prio() 53 static u16 calc_new_low_prio(const struct skbprio_sched_data *q) in calc_new_low_prio() argument 57 for (prio = q->lowest_prio + 1; prio <= q->highest_prio; prio++) { in calc_new_low_prio() 58 if (!skb_queue_empty(&q->qdiscs[prio])) in calc_new_low_prio() 62 /* SKB queue is empty, return SKBPRIO_MAX_PRIORITY - 1 in calc_new_low_prio() 65 return SKBPRIO_MAX_PRIORITY - 1; in calc_new_low_prio() 71 const unsigned int max_priority = SKBPRIO_MAX_PRIORITY - 1; in skbprio_enqueue() 72 struct skbprio_sched_data *q = qdisc_priv(sch); in skbprio_enqueue() local [all …]
|
D | sch_sfb.c | 34 #define SFB_NUMBUCKETS (1 << SFB_BUCKET_SHIFT) /* N bins per Level */ 35 #define SFB_BUCKET_MASK (SFB_NUMBUCKETS - 1) 69 u8 slot; /* current active bins (0 or 1) */ 123 static void increment_one_qlen(u32 sfbhash, u32 slot, struct sfb_sched_data *q) in increment_one_qlen() argument 126 struct sfb_bucket *b = &q->bins[slot].bins[0][0]; in increment_one_qlen() 138 static void increment_qlen(const struct sfb_skb_cb *cb, struct sfb_sched_data *q) in increment_qlen() argument 144 increment_one_qlen(sfbhash, 0, q); in increment_qlen() 146 sfbhash = cb->hashes[1]; in increment_qlen() 148 increment_one_qlen(sfbhash, 1, q); in increment_qlen() 152 struct sfb_sched_data *q) in decrement_one_qlen() argument [all …]
|
D | sch_sfq.c | 68 - max mtu to 2^18-1; 76 #define SFQ_MAX_FLOWS (0x10000 - SFQ_MAX_DEPTH - 1) /* max number of flows */ 81 * Scale allot by 8 (1<<3) so that no overflow occurs. 84 #define SFQ_ALLOT_SIZE(X) DIV_ROUND_UP(X, 1 << SFQ_ALLOT_SHIFT) 86 /* This type should contain at least SFQ_MAX_DEPTH + 1 + SFQ_MAX_FLOWS values */ 91 * Small indexes [0 ... SFQ_MAX_FLOWS - 1] are 'pointers' to slots[] array 133 struct sfq_head dep[SFQ_MAX_DEPTH + 1]; 136 * dep[1] : list of flows with 1 packet 150 static inline struct sfq_head *sfq_dep_head(struct sfq_sched_data *q, sfq_index val) in sfq_dep_head() argument 153 return &q->slots[val].dep; in sfq_dep_head() [all …]
|
D | sch_netem.c | 34 Sources: [1] Mark Carson, Darrin Santay, "NIST Net - A Linux-based 59 [1] NetemCLG Home http://netgroup.uniroma2.it/NetemCLG 62 in the Netem module in the Linux kernel", available in [1] 122 TX_IN_GAP_PERIOD = 1, 129 GOOD_STATE = 1, 142 u32 a4; /* p14 for 4-states or 1-k for GE */ 197 rho = (u64)state->rho + 1; in get_crandom() 198 answer = (value * ((1ull<<32) - rho) + state->last * rho) >> 32; in get_crandom() 207 static bool loss_4state(struct netem_sched_data *q) in loss_4state() argument 209 struct clgstate *clg = &q->clg; in loss_4state() [all …]
|
D | sch_cake.c | 101 * @rec_inv_sqrt: reciprocal value of sqrt(count) >> 1 261 CAKE_FLAG_AUTORATE_INGRESS = BIT(1), 300 static u16 quantum_div[CAKE_QUEUES + 1] = {0}; 306 1, 1, 1, 1, 1, 1, 1, 1, 316 2, 0, 1, 2, 4, 2, 2, 2, 317 1, 2, 1, 2, 1, 2, 1, 2, 327 0, 1, 0, 0, 2, 0, 0, 0, 328 1, 0, 0, 0, 0, 0, 0, 0, 338 0, 1, 0, 0, 2, 0, 0, 0, 339 1, 0, 0, 0, 0, 0, 0, 0, [all …]
|
D | sch_fq.c | 66 * If packets have monotically increasing time_to_send, they are placed in O(1) 75 unsigned long age; /* (jiffies | 1UL) when flow was emptied, for gc */ 79 /* Following field is only used for q->internal, 80 * because q->internal is not hashed in fq_root[] 93 struct rb_node rate_node; /* anchor in q->delayed tree */ 125 u8 prio2band[(TC_PRIO_MAX + 1) >> 2]; 173 f->age = jiffies | 1UL; in fq_flow_set_detached() 178 return !!(f->age & 1UL); in fq_flow_is_detached() 194 static void fq_flow_add_tail(struct fq_sched_data *q, struct fq_flow *flow, in fq_flow_add_tail() argument 197 struct fq_perband_flows *pband = &q->band_flows[flow->band]; in fq_flow_add_tail() [all …]
|
D | sch_multiq.c | 32 struct multiq_sched_data *q = qdisc_priv(sch); in multiq_classify() local 35 struct tcf_proto *fl = rcu_dereference_bh(q->filter_list); in multiq_classify() 53 if (band >= q->bands) in multiq_classify() 54 return q->queues[0]; in multiq_classify() 56 return q->queues[band]; in multiq_classify() 79 sch->q.qlen++; in multiq_enqueue() 89 struct multiq_sched_data *q = qdisc_priv(sch); in multiq_dequeue() local 94 for (band = 0; band < q->bands; band++) { in multiq_dequeue() 96 q->curband++; in multiq_dequeue() 97 if (q->curband >= q->bands) in multiq_dequeue() [all …]
|
/linux-6.8/lib/crypto/ |
D | gf128mul.c | 25 1. distributions of this source code include the above copyright 56 #define gf128mul_dat(q) { \ argument 57 q(0x00), q(0x01), q(0x02), q(0x03), q(0x04), q(0x05), q(0x06), q(0x07),\ 58 q(0x08), q(0x09), q(0x0a), q(0x0b), q(0x0c), q(0x0d), q(0x0e), q(0x0f),\ 59 q(0x10), q(0x11), q(0x12), q(0x13), q(0x14), q(0x15), q(0x16), q(0x17),\ 60 q(0x18), q(0x19), q(0x1a), q(0x1b), q(0x1c), q(0x1d), q(0x1e), q(0x1f),\ 61 q(0x20), q(0x21), q(0x22), q(0x23), q(0x24), q(0x25), q(0x26), q(0x27),\ 62 q(0x28), q(0x29), q(0x2a), q(0x2b), q(0x2c), q(0x2d), q(0x2e), q(0x2f),\ 63 q(0x30), q(0x31), q(0x32), q(0x33), q(0x34), q(0x35), q(0x36), q(0x37),\ 64 q(0x38), q(0x39), q(0x3a), q(0x3b), q(0x3c), q(0x3d), q(0x3e), q(0x3f),\ [all …]
|
/linux-6.8/drivers/net/wireless/mediatek/mt76/ |
D | dma.c | 189 mt76_dma_sync_idx(struct mt76_dev *dev, struct mt76_queue *q) in mt76_dma_sync_idx() argument 191 Q_WRITE(q, desc_base, q->desc_dma); in mt76_dma_sync_idx() 192 if (q->flags & MT_QFLAG_WED_RRO_EN) in mt76_dma_sync_idx() 193 Q_WRITE(q, ring_size, MT_DMA_RRO_EN | q->ndesc); in mt76_dma_sync_idx() 195 Q_WRITE(q, ring_size, q->ndesc); in mt76_dma_sync_idx() 196 q->head = Q_READ(q, dma_idx); in mt76_dma_sync_idx() 197 q->tail = q->head; in mt76_dma_sync_idx() 201 __mt76_dma_queue_reset(struct mt76_dev *dev, struct mt76_queue *q, in __mt76_dma_queue_reset() argument 204 if (!q || !q->ndesc) in __mt76_dma_queue_reset() 207 if (!mt76_queue_is_wed_rro_ind(q)) { in __mt76_dma_queue_reset() [all …]
|
/linux-6.8/drivers/net/ethernet/fungible/funeth/ |
D | funeth_rx.c | 22 /* Per packet headroom in non-XDP mode. Present only for 1-frag packets. */ 33 FUN_XDP_FLUSH_REDIR = 1, 50 static void cache_offer(struct funeth_rxq *q, const struct funeth_rxbuf *buf) in cache_offer() argument 52 struct funeth_rx_cache *c = &q->cache; in cache_offer() 58 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE, in cache_offer() 67 static bool cache_get(struct funeth_rxq *q, struct funeth_rxbuf *rb) in cache_get() argument 69 struct funeth_rx_cache *c = &q->cache; in cache_get() 77 dma_sync_single_for_device(q->dma_dev, buf->dma_addr, in cache_get() 88 dma_unmap_page_attrs(q->dma_dev, buf->dma_addr, PAGE_SIZE, in cache_get() 98 static int funeth_alloc_page(struct funeth_rxq *q, struct funeth_rxbuf *rb, in funeth_alloc_page() argument [all …]
|
D | funeth_tx.c | 48 dma_unmap_single(dev, addr[-1], data_len, DMA_TO_DEVICE); in fun_map_pkt() 56 static void *txq_end(const struct funeth_txq *q) in txq_end() argument 58 return (void *)q->hw_wb; in txq_end() 64 static unsigned int txq_to_end(const struct funeth_txq *q, void *p) in txq_to_end() argument 66 return txq_end(q) - p; in txq_to_end() 78 static struct fun_dataop_gl *fun_write_gl(const struct funeth_txq *q, in fun_write_gl() argument 90 i < ngle && txq_to_end(q, gle); i++, gle++) in fun_write_gl() 93 if (txq_to_end(q, gle) == 0) { in fun_write_gl() 94 gle = (struct fun_dataop_gl *)q->desc; in fun_write_gl() 107 static struct sk_buff *fun_tls_tx(struct sk_buff *skb, struct funeth_txq *q, in fun_tls_tx() argument [all …]
|
/linux-6.8/drivers/net/wireless/broadcom/b43/ |
D | pio.c | 24 static u16 generate_cookie(struct b43_pio_txqueue *q, in generate_cookie() argument 37 cookie = (((u16)q->index + 1) << 12); in generate_cookie() 49 struct b43_pio_txqueue *q = NULL; in parse_cookie() local 54 q = pio->tx_queue_AC_BK; in parse_cookie() 57 q = pio->tx_queue_AC_BE; in parse_cookie() 60 q = pio->tx_queue_AC_VI; in parse_cookie() 63 q = pio->tx_queue_AC_VO; in parse_cookie() 66 q = pio->tx_queue_mcast; in parse_cookie() 69 if (B43_WARN_ON(!q)) in parse_cookie() 72 if (B43_WARN_ON(pack_index >= ARRAY_SIZE(q->packets))) in parse_cookie() [all …]
|
/linux-6.8/sound/core/seq/oss/ |
D | seq_oss_readq.c | 20 //#define SNDRV_SEQ_OSS_MAX_TIMEOUT (unsigned long)(-1) 35 struct seq_oss_readq *q; in snd_seq_oss_readq_new() local 37 q = kzalloc(sizeof(*q), GFP_KERNEL); in snd_seq_oss_readq_new() 38 if (!q) in snd_seq_oss_readq_new() 41 q->q = kcalloc(maxlen, sizeof(union evrec), GFP_KERNEL); in snd_seq_oss_readq_new() 42 if (!q->q) { in snd_seq_oss_readq_new() 43 kfree(q); in snd_seq_oss_readq_new() 47 q->maxlen = maxlen; in snd_seq_oss_readq_new() 48 q->qlen = 0; in snd_seq_oss_readq_new() 49 q->head = q->tail = 0; in snd_seq_oss_readq_new() [all …]
|
/linux-6.8/drivers/gpu/drm/xe/ |
D | xe_guc_submit.c | 42 exec_queue_to_guc(struct xe_exec_queue *q) in exec_queue_to_guc() argument 44 return &q->gt->uc.guc; in exec_queue_to_guc() 52 #define EXEC_QUEUE_STATE_REGISTERED (1 << 0) 53 #define ENGINE_STATE_ENABLED (1 << 1) 54 #define EXEC_QUEUE_STATE_PENDING_ENABLE (1 << 2) 55 #define EXEC_QUEUE_STATE_PENDING_DISABLE (1 << 3) 56 #define EXEC_QUEUE_STATE_DESTROYED (1 << 4) 57 #define ENGINE_STATE_SUSPENDED (1 << 5) 58 #define EXEC_QUEUE_STATE_RESET (1 << 6) 59 #define ENGINE_STATE_KILLED (1 << 7) [all …]
|
D | xe_exec_queue.c | 28 XE_EXEC_QUEUE_TIMESLICE = 1, 39 struct xe_exec_queue *q; in __xe_exec_queue_create() local 47 q = kzalloc(sizeof(*q) + sizeof(struct xe_lrc) * width, GFP_KERNEL); in __xe_exec_queue_create() 48 if (!q) in __xe_exec_queue_create() 51 kref_init(&q->refcount); in __xe_exec_queue_create() 52 q->flags = flags; in __xe_exec_queue_create() 53 q->hwe = hwe; in __xe_exec_queue_create() 54 q->gt = gt; in __xe_exec_queue_create() 56 q->vm = xe_vm_get(vm); in __xe_exec_queue_create() 57 q->class = hwe->class; in __xe_exec_queue_create() [all …]
|
/linux-6.8/drivers/net/ethernet/chelsio/cxgb3/ |
D | sge.c | 66 #define FL1_PG_ORDER (PAGE_SIZE > 8192 ? 0 : 1) 94 TXQ_RUNNING = 1 << 0, /* fetch engine is running */ 95 TXQ_LAST_PKT_DB = 1 << 1, /* last packet rang the doorbell */ 141 dma_addr_t addr[MAX_SKB_FRAGS + 1]; 148 * desc = 1 + (flits - 2) / (WR_FLITS - 1). 154 #if SGE_NUM_GENBITS == 1 155 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 160 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 165 # error "SGE_NUM_GENBITS must be 1 or 2" 169 static inline struct sge_qset *rspq_to_qset(const struct sge_rspq *q) in rspq_to_qset() argument [all …]
|
/linux-6.8/drivers/s390/cio/ |
D | qdio_main.c | 41 " lgr 1,%[schid]\n" in do_siga_sync() 50 : "cc", "0", "1", "2", "3"); in do_siga_sync() 61 " lgr 1,%[schid]\n" in do_siga_input() 68 : "cc", "0", "1", "2"); in do_siga_input() 91 " lgr 1,%[schid]\n" in do_siga_output() 100 : "cc", "0", "1", "2", "3"); in do_siga_output() 107 * @q: queue to manipulate 116 static int qdio_do_eqbs(struct qdio_q *q, unsigned char *state, in qdio_do_eqbs() argument 119 int tmp_count = count, tmp_start = start, nr = q->nr; in qdio_do_eqbs() 122 qperf_inc(q, eqbs); in qdio_do_eqbs() [all …]
|
/linux-6.8/arch/s390/kernel/ |
D | fpu.c | 25 asm volatile("stfpc %0" : "=Q" (state->fpc)); in __kernel_fpu_begin() 30 asm volatile("std 0,%0" : "=Q" (state->fprs[0])); in __kernel_fpu_begin() 31 asm volatile("std 1,%0" : "=Q" (state->fprs[1])); in __kernel_fpu_begin() 32 asm volatile("std 2,%0" : "=Q" (state->fprs[2])); in __kernel_fpu_begin() 33 asm volatile("std 3,%0" : "=Q" (state->fprs[3])); in __kernel_fpu_begin() 34 asm volatile("std 4,%0" : "=Q" (state->fprs[4])); in __kernel_fpu_begin() 35 asm volatile("std 5,%0" : "=Q" (state->fprs[5])); in __kernel_fpu_begin() 36 asm volatile("std 6,%0" : "=Q" (state->fprs[6])); in __kernel_fpu_begin() 37 asm volatile("std 7,%0" : "=Q" (state->fprs[7])); in __kernel_fpu_begin() 38 asm volatile("std 8,%0" : "=Q" (state->fprs[8])); in __kernel_fpu_begin() [all …]
|
/linux-6.8/block/ |
D | blk-sysfs.c | 50 static ssize_t queue_requests_show(struct request_queue *q, char *page) in queue_requests_show() argument 52 return queue_var_show(q->nr_requests, page); in queue_requests_show() 56 queue_requests_store(struct request_queue *q, const char *page, size_t count) in queue_requests_store() argument 61 if (!queue_is_mq(q)) in queue_requests_store() 71 err = blk_mq_update_nr_requests(q, nr); in queue_requests_store() 78 static ssize_t queue_ra_show(struct request_queue *q, char *page) in queue_ra_show() argument 82 if (!q->disk) in queue_ra_show() 84 ra_kb = q->disk->bdi->ra_pages << (PAGE_SHIFT - 10); in queue_ra_show() 89 queue_ra_store(struct request_queue *q, const char *page, size_t count) in queue_ra_store() argument 94 if (!q->disk) in queue_ra_store() [all …]
|
D | blk-settings.c | 22 void blk_queue_rq_timeout(struct request_queue *q, unsigned int timeout) in blk_queue_rq_timeout() argument 24 q->rq_timeout = timeout; in blk_queue_rq_timeout() 38 lim->max_discard_segments = 1; in blk_set_default_limits() 90 * @q: the request queue for the device 98 void blk_queue_bounce_limit(struct request_queue *q, enum blk_bounce bounce) in blk_queue_bounce_limit() argument 100 q->limits.bounce = bounce; in blk_queue_bounce_limit() 106 * @q: the request queue for the device 123 void blk_queue_max_hw_sectors(struct request_queue *q, unsigned int max_hw_sectors) in blk_queue_max_hw_sectors() argument 125 struct queue_limits *limits = &q->limits; in blk_queue_max_hw_sectors() 129 max_hw_sectors = 1 << (PAGE_SHIFT - 9); in blk_queue_max_hw_sectors() [all …]
|
/linux-6.8/drivers/infiniband/sw/rxe/ |
D | rxe_queue.h | 18 * - The driver indices are always masked off to q->index_mask 26 * - By passing the type in the parameter list separate from q 83 void rxe_queue_reset(struct rxe_queue *q); 88 int rxe_queue_resize(struct rxe_queue *q, unsigned int *num_elem_p, 95 static inline u32 queue_next_index(struct rxe_queue *q, int index) in queue_next_index() argument 97 return (index + 1) & q->index_mask; in queue_next_index() 100 static inline u32 queue_get_producer(const struct rxe_queue *q, in queue_get_producer() argument 108 prod = smp_load_acquire(&q->buf->producer_index); in queue_get_producer() 112 prod = q->index; in queue_get_producer() 116 prod = q->buf->producer_index; in queue_get_producer() [all …]
|
/linux-6.8/sound/core/seq/ |
D | seq_queue.c | 50 static int queue_list_add(struct snd_seq_queue *q) in queue_list_add() argument 58 queue_list[i] = q; in queue_list_add() 59 q->queue = i; in queue_list_add() 66 return -1; in queue_list_add() 71 struct snd_seq_queue *q; in queue_list_remove() local 75 q = queue_list[id]; in queue_list_remove() 76 if (q) { in queue_list_remove() 77 spin_lock(&q->owner_lock); in queue_list_remove() 78 if (q->owner == client) { in queue_list_remove() 80 q->klocked = 1; in queue_list_remove() [all …]
|
/linux-6.8/drivers/spi/ |
D | spi-fsl-qspi.c | 59 #define QUADSPI_MCR_SWRSTHD_MASK BIT(1) 101 #define QUADSPI_SR_IP_ACC_MASK BIT(1) 125 #define QUADSPI_LCKER_UNLOCK BIT(1) 134 #define LUT_CMD 1 158 #define LUT_PAD(x) (fls(x) - 1) 175 #define QUADSPI_QUIRK_4X_INT_CLK BIT(1) 277 static inline int needs_swap_endian(struct fsl_qspi *q) in needs_swap_endian() argument 279 return q->devtype_data->quirks & QUADSPI_QUIRK_SWAP_ENDIAN; in needs_swap_endian() 282 static inline int needs_4x_clock(struct fsl_qspi *q) in needs_4x_clock() argument 284 return q->devtype_data->quirks & QUADSPI_QUIRK_4X_INT_CLK; in needs_4x_clock() [all …]
|