/linux/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
H A D | crypto.c | 13 * (for example, TLS) after last revalidation in a pool or a bulk. 20 #define MLX5_CRYPTO_DEK_BULK_CALC_FREED(bulk) MLX5_CRYPTO_DEK_CALC_FREED(bulk) argument 22 #define MLX5_CRYPTO_DEK_BULK_IDLE(bulk) \ argument 23 ({ typeof(bulk) _bulk = (bulk); \ 59 int num_deks; /* the total number of keys in a bulk */ 80 struct mlx5_crypto_dek_bulk *bulk; member 182 mlx5_core_dbg(mdev, "DEK objects created, bulk=%d, obj_id=%d\n", in mlx5_crypto_create_dek_bulk() 292 struct mlx5_crypto_dek_bulk *bulk; in mlx5_crypto_dek_bulk_create() local 336 struct mlx5_crypto_dek_bulk *bulk; mlx5_crypto_dek_pool_add_bulk() local 349 mlx5_crypto_dek_bulk_free(struct mlx5_crypto_dek_bulk * bulk) mlx5_crypto_dek_bulk_free() argument 358 mlx5_crypto_dek_pool_remove_bulk(struct mlx5_crypto_dek_pool * pool,struct mlx5_crypto_dek_bulk * bulk,bool delay) mlx5_crypto_dek_pool_remove_bulk() argument 372 struct mlx5_crypto_dek_bulk *bulk; mlx5_crypto_dek_pool_pop() local 429 struct mlx5_crypto_dek_bulk *bulk = dek->bulk; mlx5_crypto_dek_free_locked() local 480 mlx5_crypto_dek_bulk_reset_synced(struct mlx5_crypto_dek_pool * pool,struct mlx5_crypto_dek_bulk * bulk) mlx5_crypto_dek_bulk_reset_synced() argument 511 mlx5_crypto_dek_bulk_handle_avail(struct mlx5_crypto_dek_pool * pool,struct mlx5_crypto_dek_bulk * bulk,struct list_head * destroy_list) mlx5_crypto_dek_bulk_handle_avail() argument 550 struct mlx5_crypto_dek_bulk *bulk, *tmp; mlx5_crypto_dek_pool_reset_synced() local 609 struct mlx5_crypto_dek_bulk *bulk; mlx5_crypto_dek_create() local 664 struct mlx5_crypto_dek_bulk *bulk, *tmp; mlx5_crypto_dek_free_destroy_list() local 709 struct mlx5_crypto_dek_bulk *bulk, *tmp; mlx5_crypto_dek_pool_destroy() local [all...] |
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | fs_pool.c | 29 int mlx5_fs_bulk_get_free_amount(struct mlx5_fs_bulk *bulk) in mlx5_fs_bulk_get_free_amount() argument 31 return bitmap_weight(bulk->bitmask, bulk->bulk_len); in mlx5_fs_bulk_get_free_amount() 78 struct mlx5_fs_bulk *bulk; in mlx5_fs_pool_cleanup() local 81 list_for_each_entry_safe(bulk, tmp, &pool->fully_used, pool_list) in mlx5_fs_pool_cleanup() 82 pool->ops->bulk_destroy(dev, bulk); in mlx5_fs_pool_cleanup() 83 list_for_each_entry_safe(bulk, tmp, &pool->partially_used, pool_list) in mlx5_fs_pool_cleanup() 84 pool->ops->bulk_destroy(dev, bulk); in mlx5_fs_pool_cleanup() 85 list_for_each_entry_safe(bulk, tmp, &pool->unused, pool_list) in mlx5_fs_pool_cleanup() 86 pool->ops->bulk_destroy(dev, bulk); in mlx5_fs_pool_cleanup() 103 mlx5_fs_pool_free_bulk(struct mlx5_fs_pool * fs_pool,struct mlx5_fs_bulk * bulk) mlx5_fs_pool_free_bulk() argument 167 struct mlx5_fs_bulk *bulk = pool_index->fs_bulk; mlx5_fs_pool_release_index() local [all...] |
H A D | fs_counters.c | 41 /* Max number of counters to query in bulk read is 32K */ 105 * - bulk query (single thread workqueue context) 107 * around the same time as the current bulk cmd. 109 * destroyed during a bulk query command. 128 /* Start new bulk query. */ in mlx5_fc_stats_query_all_counters() 129 /* First id must be aligned to 4 when using bulk query. */ in mlx5_fc_stats_query_all_counters() 139 mlx5_core_err(dev, "Error doing bulk query: %d\n", err); in mlx5_fc_stats_query_all_counters() 145 /* Do not update counters added after bulk query was started. */ in mlx5_fc_stats_query_all_counters() 166 if (counter->bulk) in mlx5_fc_release() 183 "Can't increase flow counters bulk quer in mlx5_fc_stats_bulk_query_buf_realloc() 417 mlx5_fc_init(struct mlx5_fc * counter,struct mlx5_fc_bulk * bulk,u32 id) mlx5_fc_init() argument [all...] |
/linux/drivers/staging/vc04_services/interface/ |
H A D | TESTING | 49 Testing bulk transfer for alignment. 50 Testing bulk transfer at PAGE_SIZE. 61 vchi bulk (size 0, 0 async, 0 oneway) -> 546.000000us 62 vchi bulk (size 0, 0 oneway) -> 230.000000us 65 vchi bulk (size 0, 0 async, 0 oneway) -> 296.000000us 66 vchi bulk (size 0, 0 oneway) -> 266.000000us 68 vchi bulk (size 0, 0 oneway) -> 456.000000us 70 vchi bulk (size 0, 0 oneway) -> 640.000000us 72 vchi bulk (size 0, 0 oneway) -> 2309.000000us 78 vchi bulk (siz [all...] |
/linux/include/net/libeth/ |
H A D | xsk.h | 19 * @bq: XDP Tx bulk to queue the head frag to 27 bq->bulk[bq->count++] = (typeof(*bq->bulk)){ in libeth_xsk_tx_queue_head() 36 bq->bulk[bq->count - 1].flags |= LIBETH_XDP_TX_MULTI; in libeth_xsk_tx_queue_head() 43 * @bq: XDP Tx bulk to queue the frag to 49 bq->bulk[bq->count++] = (typeof(*bq->bulk)){ in libeth_xsk_tx_queue_frag() 57 * @bq: XDP Tx bulk to queue the frame to 59 * @flush_bulk: driver callback to flush the bulk to the HW queue 96 bq->bulk[b in libeth_xsk_tx_queue_bulk() 292 const struct libeth_xdp_tx_frame *bulk; libeth_xsk_xmit_do_bulk() local [all...] |
H A D | xdp.h | 278 * @LIBETH_XDP_TX_BULK: one bulk size at which it will be flushed to the queue 368 * struct libeth_xdp_tx_bulk - XDP Tx frame bulk for bulk sending 373 * @count: current number of frames in @bulk 374 * @bulk: array of queued frames for bulk Tx 376 * All XDP Tx operations except XSk xmit queue each frame to the bulk first 377 * and flush it when @count reaches the array end. Bulk is always placed on 378 * the stack for performance. One bulk element contains all the data necessary 381 * to &libeth_xdp_tx_frame as they are compatible and the bulk structur 391 struct libeth_xdp_tx_frame bulk[LIBETH_XDP_TX_BULK]; global() member 517 libeth_xdp_tx_xmit_bulk(const struct libeth_xdp_tx_frame * bulk,void * xdpsq,u32 n,bool unroll,u64 priv,u32 (* prep)(void * xdpsq,struct libeth_xdpsq * sq),struct libeth_xdp_tx_desc (* fill)(struct libeth_xdp_tx_frame frm,u32 i,const struct libeth_xdpsq * sq,u64 priv),void (* xmit)(struct libeth_xdp_tx_desc desc,u32 i,const struct libeth_xdpsq * sq,u64 priv)) libeth_xdp_tx_xmit_bulk() argument 1771 __libeth_xdp_complete_tx(struct libeth_sqe * sqe,struct libeth_cq_pp * cp,typeof(libeth_xdp_return_buff_bulk) bulk,typeof(libeth_xsk_buff_free_slow) xsk) __libeth_xdp_complete_tx() argument [all...] |
/linux/drivers/staging/vc04_services/vchiq-mmal/ |
H A D | mmal-vchiq.c | 138 /* actual buffer used to store bulk reply */ 150 } bulk; /* bulk data */ member 178 /* ordered workqueue to process all bulk operations */ 254 container_of(work, struct mmal_msg_context, u.bulk.work); in buffer_work_cb() 255 struct mmal_buffer *buffer = msg_context->u.bulk.buffer; in buffer_work_cb() 263 buffer->length = msg_context->u.bulk.buffer_used; in buffer_work_cb() 264 buffer->mmal_flags = msg_context->u.bulk.mmal_flags; in buffer_work_cb() 265 buffer->dts = msg_context->u.bulk.dts; in buffer_work_cb() 266 buffer->pts = msg_context->u.bulk in buffer_work_cb() [all...] |
/linux/drivers/media/usb/dvb-usb/ |
H A D | dibusb.h | 25 * bulk msg to/from endpoint 0x01 36 * bulk write: 0x02 ((7bit i2c_addr << 1) | 0x01) register_bytes length_word 37 * bulk read: byte_buffer (length_word bytes) 43 * bulk write: 0x03 (7bit i2c_addr << 1) register_bytes value_bytes 49 * bulk write: 0x04 50 * bulk read: byte_buffer (5 bytes) 59 * bulk write: 0x05 mode_byte
|
/linux/include/linux/usb/ |
H A D | serial.h | 40 * @bulk_in_buffer: pointer to the bulk in buffer for this port. 42 * @read_urb: pointer to the bulk in struct urb for this port. 43 * @bulk_in_endpointAddress: endpoint address for the bulk in pipe for this 45 * @bulk_in_buffers: pointers to the bulk in buffers for this port 46 * @read_urbs: pointers to the bulk in urbs for this port 47 * @read_urbs_free: status bitmap the for bulk in urbs 48 * @bulk_out_buffer: pointer to the bulk out buffer for this port. 50 * @write_urb: pointer to the bulk out struct urb for this port. 52 * @bulk_out_buffers: pointers to the bulk out buffers for this port 53 * @write_urbs: pointers to the bulk ou [all...] |
H A D | storage.h | 30 #define USB_PR_CBI 0x00 /* Control/Bulk/Interrupt */ 31 #define USB_PR_CB 0x01 /* Control/Bulk w/o interrupt */ 32 #define USB_PR_BULK 0x50 /* bulk only */ 48 * Bulk only data structures 81 /* bulk-only class specific requests */
|
/linux/Documentation/w1/masters/ |
H A D | ds2490.rst | 49 - The ds2490 specification doesn't cover short bulk in reads in 51 available, the bulk read will return an error and the hardware will 52 clear the entire bulk in buffer. It would be possible to read the 63 most of the time one of the bulk out or in, and usually the bulk in 64 would fail. qemu sets a 50ms timeout and the bulk in would timeout 65 even when the status shows data available. A bulk out write would
|
/linux/include/drm/ttm/ |
H A D | ttm_resource.h | 284 * @first: first res in the bulk move range 285 * @last: last res in the bulk move range 287 * Range of resources for a lru bulk move. 300 * Container for the current bulk move state. Should be used with 303 * ensure that the bulk as a whole is locked for eviction even if only one BO of 304 * the bulk is evicted. 317 * bulk sublist of @bulk. Protected by the ttm device's lru_lock. 318 * @bulk: Pointer to struct ttm_lru_bulk_move whose subrange @hitch is 323 * This field is valid iff @bulk ! 332 struct ttm_lru_bulk_move *bulk; global() member [all...] |
/linux/drivers/staging/vc04_services/interface/vchiq_arm/ |
H A D | vchiq_core.c | 458 struct vchiq_header *header, struct vchiq_bulk *bulk) in make_service_callback() argument 465 * If a bulk transfer is in progress, pass bulk->cb_*data to the in make_service_callback() 468 if (bulk) { in make_service_callback() 469 cb_data = bulk->cb_data; in make_service_callback() 470 cb_userdata = bulk->cb_userdata; in make_service_callback() 1308 get_bulk_reason(struct vchiq_bulk *bulk) in get_bulk_reason() argument 1310 if (bulk->dir == VCHIQ_BULK_TRANSMIT) { in get_bulk_reason() 1311 if (bulk->actual == VCHIQ_BULK_ACTUAL_ABORTED) in get_bulk_reason() 1317 if (bulk in get_bulk_reason() 1324 service_notify_bulk(struct vchiq_service * service,struct vchiq_bulk * bulk) service_notify_bulk() argument 1375 struct vchiq_bulk *bulk = notify_bulks() local 1495 create_pagelist(struct vchiq_instance * instance,struct vchiq_bulk * bulk) create_pagelist() argument 1755 vchiq_prepare_bulk_data(struct vchiq_instance * instance,struct vchiq_bulk * bulk) vchiq_prepare_bulk_data() argument 1776 vchiq_complete_bulk(struct vchiq_instance * instance,struct vchiq_bulk * bulk) vchiq_complete_bulk() argument 1801 struct vchiq_bulk *bulk = &queue->bulks[BULK_INDEX(queue->process)]; abort_outstanding_bulks() local 2092 struct vchiq_bulk *bulk; parse_message() local 3039 struct vchiq_bulk *bulk; vchiq_bulk_xfer_queue_msg_killable() local [all...] |
H A D | vchiq_core.h | 130 int local_insert; /* Where to insert the next local bulk */ 131 int remote_insert; /* Where to insert the next remote bulk (master) */ 132 int process; /* Bulk to transfer next */ 133 int remote_notify; /* Bulk to notify the remote client of next (mstr) */ 134 int remove; /* Bulk to notify the local client of, and remove, next */ 445 struct vchiq_bulk *bulk; member 453 * is better to use a bulk transfer 506 struct vchiq_bulk *bulk); 510 struct vchiq_bulk *bulk);
|
/linux/Documentation/ABI/testing/ |
H A D | sysfs-driver-w1_therm | 75 * If a bulk read has been triggered, it will directly 76 return the temperature computed when the bulk read 80 * If no bulk read has been triggered, it will trigger 115 (RW) trigger a bulk read conversion. read the status 124 no bulk operation. Reading temperature will 128 'trigger': trigger a bulk read on all supporting 131 Note that if a bulk read is sent but one sensor is not read 134 of the bulk read command (not the current temperature).
|
/linux/drivers/net/ethernet/intel/libeth/ |
H A D | tx.c | 12 DEFINE_STATIC_CALL_NULL(bulk, libeth_xdp_return_buff_bulk); 27 __libeth_xdp_complete_tx(sqe, cp, static_call(bulk), in libeth_tx_complete_any() 38 static_call_update(bulk, ops ? ops->bulk : NULL); in libeth_attach_xdp()
|
H A D | xdp.c | 115 * @bq: XDP Tx frame bulk 116 * @sent: number of frames sent successfully (from this bulk) 121 * the Tx bulk to try again later. 126 const struct libeth_xdp_tx_frame *pos = &bq->bulk[sent]; in libeth_xdp_tx_exception() 133 memmove(bq->bulk, pos, left * sizeof(*bq->bulk)); in libeth_xdp_tx_exception() 284 * @bq: XDP Tx bulk 325 * libeth_xdp_return_buff_bulk - free &xdp_buff as part of a bulk 327 * @bq: XDP frame bulk to store the buffer 431 .bulk [all...] |
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
H A D | fs_hws_pools.c | 128 pr_bulk->prs_data[i].bulk = pr_bulk; in mlx5_fs_hws_pr_bulk_create() 152 mlx5_core_err(dev, "Freeing bulk before all reformats were released\n"); in mlx5_fs_hws_pr_bulk_destroy() 225 struct mlx5_fs_bulk *fs_bulk = &pr_data->bulk->fs_bulk; in mlx5_fs_hws_pr_pool_release_pr() 237 return pr_data->bulk->hws_action; in mlx5_fs_hws_pr_get_action() 282 mh_bulk->mhs_data[i].bulk = mh_bulk; in mlx5_fs_hws_mh_bulk_create() 307 mlx5_core_err(dev, "Freeing bulk before all modify header were released\n"); in mlx5_fs_hws_mh_bulk_destroy() 374 struct mlx5_fs_bulk *fs_bulk = &mh_data->bulk->fs_bulk; in mlx5_fs_hws_mh_pool_release_mh() 409 struct mlx5_fc_bulk *fc_bulk = counter->bulk; in mlx5_fc_get_hws_action() 420 mlx5_fs_put_hws_action(&counter->bulk->hws_data); in mlx5_fc_put_hws_action()
|
/linux/Documentation/driver-api/usb/ |
H A D | bulk-streams.rst | 1 USB bulk streams 7 Bulk endpoint streams were added in the USB 3.0 specification. Streams allow a 8 device driver to overload a bulk endpoint so that multiple transfers can be 41 ID for the bulk IN and OUT endpoints used in a Bi-directional command sequence. 46 declares how many stream IDs it can support, and each bulk endpoint on a
|
/linux/drivers/gpu/drm/msm/ |
H A D | msm_io_utils.c | 17 struct clk *msm_clk_bulk_get_clock(struct clk_bulk_data *bulk, int count, in msm_clk_bulk_get_clock() argument 25 for (i = 0; bulk && i < count; i++) { in msm_clk_bulk_get_clock() 26 if (!strcmp(bulk[i].id, name) || !strcmp(bulk[i].id, n)) in msm_clk_bulk_get_clock() 27 return bulk[i].clk; in msm_clk_bulk_get_clock()
|
/linux/tools/usb/ |
H A D | hcd-tests.sh | 7 # - out, in: out needs 'bulk sink' firmware, in needs 'bulk src' 9 # - halt: needs bulk sink+src, tests halt set/clear from host 10 # - unlink: needs bulk sink and/or src, test HCD unlink processing 17 # this default provides a steady test load for a bulk device 37 # NOTE: the 'in' and 'out' cases are usually bulk, but can be
|
/linux/Documentation/usb/ |
H A D | ehci.rst | 58 At this writing the driver should comfortably handle all control, bulk, 125 and bulk transfers. Shows each active qh and the qtds 160 Bulk transfers are most often used where throughput is an issue. It's 161 good to keep in mind that bulk transfers are always in 512 byte packets, 165 So more than 50 MByte/sec is available for bulk transfers, when both 195 you issue a control or bulk request you can often expect to learn that 203 or using bulk queuing if a series of small requests needs to be issued. 213 I/O be efficient, it's better to just queue up several (bulk) requests
|
/linux/drivers/usb/misc/ |
H A D | usblcd.c | 48 bulk in endpoint */ 50 bulk out endpoint */ 140 /* do a blocking bulk read to get data from the device */ in lcd_read() 208 "nonzero write bulk status received: %d\n", status); in lcd_write_bulk_callback() 270 /* send the data out the bulk port */ in lcd_write() 345 /* use only the first bulk-in and bulk-out endpoints */ in lcd_probe() 350 "Could not find both bulk-in and bulk-out endpoints\n"); in lcd_probe()
|
/linux/drivers/media/usb/dvb-usb-v2/ |
H A D | mxl111sf.c | 30 MODULE_PARM_DESC(isoc, "enable usb isoc xfer (0=bulk, 1=isoc)."); 1040 stream->u.bulk.buffersize = 8192; in mxl111sf_stream_config_bulk() 1059 * bulk EP4/BULK/5/8192 1109 * bulk EP6/BULK/5/8192 1151 * bulk EP5/BULK/5/8192/RAW 1193 * bulk EP6/BULK/ [all...] |
/linux/drivers/media/usb/uvc/ |
H A D | uvc_video.c | 1120 * uvc_video_decode_start is called with URB data at the start of a bulk or 1138 * uvc_video_decode_end is called with header data at the end of a bulk or 1363 * uvc_video_encode_data(). Only bulk transfers are currently supported. 1392 nbytes = min(stream->bulk.max_payload_size - stream->bulk.payload_size, in uvc_video_encode_data() 1581 if (urb->actual_length == 0 && stream->bulk.header_size == 0) in uvc_video_decode_bulk() 1586 stream->bulk.payload_size += len; in uvc_video_decode_bulk() 1592 if (stream->bulk.header_size == 0 && !stream->bulk.skip_payload) { in uvc_video_decode_bulk() 1601 stream->bulk in uvc_video_decode_bulk() [all...] |