/linux-6.14.4/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
D | crypto.c | 13 * (for example, TLS) after last revalidation in a pool or a bulk. 20 #define MLX5_CRYPTO_DEK_BULK_CALC_FREED(bulk) MLX5_CRYPTO_DEK_CALC_FREED(bulk) argument 22 #define MLX5_CRYPTO_DEK_BULK_IDLE(bulk) \ argument 23 ({ typeof(bulk) _bulk = (bulk); \ 59 int num_deks; /* the total number of keys in a bulk */ 80 struct mlx5_crypto_dek_bulk *bulk; member 182 mlx5_core_dbg(mdev, "DEK objects created, bulk=%d, obj_id=%d\n", in mlx5_crypto_create_dek_bulk() 292 struct mlx5_crypto_dek_bulk *bulk; in mlx5_crypto_dek_bulk_create() local 296 bulk = kzalloc(sizeof(*bulk), GFP_KERNEL); in mlx5_crypto_dek_bulk_create() 297 if (!bulk) in mlx5_crypto_dek_bulk_create() [all …]
|
/linux-6.14.4/drivers/net/ethernet/mellanox/mlx5/core/ |
D | fs_pool.c | 29 int mlx5_fs_bulk_get_free_amount(struct mlx5_fs_bulk *bulk) in mlx5_fs_bulk_get_free_amount() argument 31 return bitmap_weight(bulk->bitmask, bulk->bulk_len); in mlx5_fs_bulk_get_free_amount() 78 struct mlx5_fs_bulk *bulk; in mlx5_fs_pool_cleanup() local 81 list_for_each_entry_safe(bulk, tmp, &pool->fully_used, pool_list) in mlx5_fs_pool_cleanup() 82 pool->ops->bulk_destroy(dev, bulk); in mlx5_fs_pool_cleanup() 83 list_for_each_entry_safe(bulk, tmp, &pool->partially_used, pool_list) in mlx5_fs_pool_cleanup() 84 pool->ops->bulk_destroy(dev, bulk); in mlx5_fs_pool_cleanup() 85 list_for_each_entry_safe(bulk, tmp, &pool->unused, pool_list) in mlx5_fs_pool_cleanup() 86 pool->ops->bulk_destroy(dev, bulk); in mlx5_fs_pool_cleanup() 103 mlx5_fs_pool_free_bulk(struct mlx5_fs_pool *fs_pool, struct mlx5_fs_bulk *bulk) in mlx5_fs_pool_free_bulk() argument [all …]
|
D | fs_counters.c | 41 /* Max number of counters to query in bulk read is 32K */ 105 * - bulk query (single thread workqueue context) 107 * around the same time as the current bulk cmd. 109 * destroyed during a bulk query command. 128 /* Start new bulk query. */ in mlx5_fc_stats_query_all_counters() 129 /* First id must be aligned to 4 when using bulk query. */ in mlx5_fc_stats_query_all_counters() 139 mlx5_core_err(dev, "Error doing bulk query: %d\n", err); in mlx5_fc_stats_query_all_counters() 145 /* Do not update counters added after bulk query was started. */ in mlx5_fc_stats_query_all_counters() 166 if (counter->bulk) in mlx5_fc_release() 183 "Can't increase flow counters bulk query buffer size, alloc failed, bulk_query_len(%d)\n", in mlx5_fc_stats_bulk_query_buf_realloc() [all …]
|
/linux-6.14.4/drivers/gpu/drm/ttm/ |
D | ttm_resource.c | 38 /* Detach the cursor from the bulk move list*/ 44 cursor->bulk = NULL; in ttm_resource_cursor_clear_bulk() 48 /* Move the cursor to the end of the bulk move list it's in */ 49 static void ttm_resource_cursor_move_bulk_tail(struct ttm_lru_bulk_move *bulk, in ttm_resource_cursor_move_bulk_tail() argument 56 if (WARN_ON_ONCE(bulk != cursor->bulk)) { in ttm_resource_cursor_move_bulk_tail() 61 pos = &bulk->pos[cursor->mem_type][cursor->priority]; in ttm_resource_cursor_move_bulk_tail() 67 /* Move all cursors attached to a bulk move to its end */ 68 static void ttm_bulk_move_adjust_cursors(struct ttm_lru_bulk_move *bulk) in ttm_bulk_move_adjust_cursors() argument 72 list_for_each_entry_safe(cursor, next, &bulk->cursor_list, bulk_link) in ttm_bulk_move_adjust_cursors() 73 ttm_resource_cursor_move_bulk_tail(bulk, cursor); in ttm_bulk_move_adjust_cursors() [all …]
|
/linux-6.14.4/drivers/staging/vc04_services/interface/ |
D | TESTING | 49 Testing bulk transfer for alignment. 50 Testing bulk transfer at PAGE_SIZE. 61 vchi bulk (size 0, 0 async, 0 oneway) -> 546.000000us 62 vchi bulk (size 0, 0 oneway) -> 230.000000us 65 vchi bulk (size 0, 0 async, 0 oneway) -> 296.000000us 66 vchi bulk (size 0, 0 oneway) -> 266.000000us 68 vchi bulk (size 0, 0 oneway) -> 456.000000us 70 vchi bulk (size 0, 0 oneway) -> 640.000000us 72 vchi bulk (size 0, 0 oneway) -> 2309.000000us 78 vchi bulk (size 0, 0 oneway) -> nanus [all …]
|
/linux-6.14.4/drivers/staging/vc04_services/vchiq-mmal/ |
D | mmal-vchiq.c | 138 /* actual buffer used to store bulk reply */ 150 } bulk; /* bulk data */ member 178 /* ordered workqueue to process all bulk operations */ 254 container_of(work, struct mmal_msg_context, u.bulk.work); in buffer_work_cb() 255 struct mmal_buffer *buffer = msg_context->u.bulk.buffer; in buffer_work_cb() 263 buffer->length = msg_context->u.bulk.buffer_used; in buffer_work_cb() 264 buffer->mmal_flags = msg_context->u.bulk.mmal_flags; in buffer_work_cb() 265 buffer->dts = msg_context->u.bulk.dts; in buffer_work_cb() 266 buffer->pts = msg_context->u.bulk.pts; in buffer_work_cb() 268 atomic_dec(&msg_context->u.bulk.port->buffers_with_vpu); in buffer_work_cb() [all …]
|
/linux-6.14.4/drivers/media/usb/dvb-usb/ |
D | dibusb.h | 25 * bulk msg to/from endpoint 0x01 36 * bulk write: 0x02 ((7bit i2c_addr << 1) | 0x01) register_bytes length_word 37 * bulk read: byte_buffer (length_word bytes) 43 * bulk write: 0x03 (7bit i2c_addr << 1) register_bytes value_bytes 49 * bulk write: 0x04 50 * bulk read: byte_buffer (5 bytes) 59 * bulk write: 0x05 mode_byte
|
/linux-6.14.4/include/linux/usb/ |
D | serial.h | 40 * @bulk_in_buffer: pointer to the bulk in buffer for this port. 42 * @read_urb: pointer to the bulk in struct urb for this port. 43 * @bulk_in_endpointAddress: endpoint address for the bulk in pipe for this 45 * @bulk_in_buffers: pointers to the bulk in buffers for this port 46 * @read_urbs: pointers to the bulk in urbs for this port 47 * @read_urbs_free: status bitmap the for bulk in urbs 48 * @bulk_out_buffer: pointer to the bulk out buffer for this port. 50 * @write_urb: pointer to the bulk out struct urb for this port. 52 * @bulk_out_buffers: pointers to the bulk out buffers for this port 53 * @write_urbs: pointers to the bulk out urbs for this port [all …]
|
D | storage.h | 30 #define USB_PR_CBI 0x00 /* Control/Bulk/Interrupt */ 31 #define USB_PR_CB 0x01 /* Control/Bulk w/o interrupt */ 32 #define USB_PR_BULK 0x50 /* bulk only */ 48 * Bulk only data structures 81 /* bulk-only class specific requests */
|
/linux-6.14.4/Documentation/w1/masters/ |
D | ds2490.rst | 49 - The ds2490 specification doesn't cover short bulk in reads in 51 available, the bulk read will return an error and the hardware will 52 clear the entire bulk in buffer. It would be possible to read the 63 most of the time one of the bulk out or in, and usually the bulk in 64 would fail. qemu sets a 50ms timeout and the bulk in would timeout 65 even when the status shows data available. A bulk out write would
|
/linux-6.14.4/drivers/staging/vc04_services/interface/vchiq_arm/ |
D | vchiq_core.c | 458 struct vchiq_header *header, struct vchiq_bulk *bulk) in make_service_callback() argument 465 * If a bulk transfer is in progress, pass bulk->cb_*data to the in make_service_callback() 468 if (bulk) { in make_service_callback() 469 cb_data = bulk->cb_data; in make_service_callback() 470 cb_userdata = bulk->cb_userdata; in make_service_callback() 1308 get_bulk_reason(struct vchiq_bulk *bulk) in get_bulk_reason() argument 1310 if (bulk->dir == VCHIQ_BULK_TRANSMIT) { in get_bulk_reason() 1311 if (bulk->actual == VCHIQ_BULK_ACTUAL_ABORTED) in get_bulk_reason() 1317 if (bulk->actual == VCHIQ_BULK_ACTUAL_ABORTED) in get_bulk_reason() 1324 struct vchiq_bulk *bulk) in service_notify_bulk() argument [all …]
|
D | vchiq_core.h | 130 int local_insert; /* Where to insert the next local bulk */ 131 int remote_insert; /* Where to insert the next remote bulk (master) */ 132 int process; /* Bulk to transfer next */ 133 int remote_notify; /* Bulk to notify the remote client of next (mstr) */ 134 int remove; /* Bulk to notify the local client of, and remove, next */ 445 struct vchiq_bulk *bulk; member 453 * is better to use a bulk transfer 506 struct vchiq_bulk *bulk); 510 struct vchiq_bulk *bulk);
|
/linux-6.14.4/include/drm/ttm/ |
D | ttm_resource.h | 284 * @first: first res in the bulk move range 285 * @last: last res in the bulk move range 287 * Range of resources for a lru bulk move. 300 * Container for the current bulk move state. Should be used with 303 * ensure that the bulk as a whole is locked for eviction even if only one BO of 304 * the bulk is evicted. 317 * bulk sublist of @bulk. Protected by the ttm device's lru_lock. 318 * @bulk: Pointer to struct ttm_lru_bulk_move whose subrange @hitch is 323 * This field is valid iff @bulk != NULL. 332 struct ttm_lru_bulk_move *bulk; member [all …]
|
/linux-6.14.4/Documentation/ABI/testing/ |
D | sysfs-driver-w1_therm | 75 * If a bulk read has been triggered, it will directly 76 return the temperature computed when the bulk read 80 * If no bulk read has been triggered, it will trigger 115 (RW) trigger a bulk read conversion. read the status 124 no bulk operation. Reading temperature will 128 'trigger': trigger a bulk read on all supporting 131 Note that if a bulk read is sent but one sensor is not read 134 of the bulk read command (not the current temperature).
|
/linux-6.14.4/Documentation/driver-api/usb/ |
D | bulk-streams.rst | 1 USB bulk streams 7 Bulk endpoint streams were added in the USB 3.0 specification. Streams allow a 8 device driver to overload a bulk endpoint so that multiple transfers can be 41 ID for the bulk IN and OUT endpoints used in a Bi-directional command sequence. 46 declares how many stream IDs it can support, and each bulk endpoint on a
|
/linux-6.14.4/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
D | fs_hws_pools.c | 128 pr_bulk->prs_data[i].bulk = pr_bulk; in mlx5_fs_hws_pr_bulk_create() 152 mlx5_core_err(dev, "Freeing bulk before all reformats were released\n"); in mlx5_fs_hws_pr_bulk_destroy() 225 struct mlx5_fs_bulk *fs_bulk = &pr_data->bulk->fs_bulk; in mlx5_fs_hws_pr_pool_release_pr() 237 return pr_data->bulk->hws_action; in mlx5_fs_hws_pr_get_action() 282 mh_bulk->mhs_data[i].bulk = mh_bulk; in mlx5_fs_hws_mh_bulk_create() 307 mlx5_core_err(dev, "Freeing bulk before all modify header were released\n"); in mlx5_fs_hws_mh_bulk_destroy() 374 struct mlx5_fs_bulk *fs_bulk = &mh_data->bulk->fs_bulk; in mlx5_fs_hws_mh_pool_release_mh() 409 struct mlx5_fc_bulk *fc_bulk = counter->bulk; in mlx5_fc_get_hws_action() 436 struct mlx5_fc_bulk_hws_data *fc_bulk_hws = &counter->bulk->hws_data; in mlx5_fc_put_hws_action()
|
/linux-6.14.4/drivers/gpu/drm/msm/ |
D | msm_io_utils.c | 17 struct clk *msm_clk_bulk_get_clock(struct clk_bulk_data *bulk, int count, in msm_clk_bulk_get_clock() argument 25 for (i = 0; bulk && i < count; i++) { in msm_clk_bulk_get_clock() 26 if (!strcmp(bulk[i].id, name) || !strcmp(bulk[i].id, n)) in msm_clk_bulk_get_clock() 27 return bulk[i].clk; in msm_clk_bulk_get_clock()
|
/linux-6.14.4/tools/usb/ |
D | hcd-tests.sh | 7 # - out, in: out needs 'bulk sink' firmware, in needs 'bulk src' 9 # - halt: needs bulk sink+src, tests halt set/clear from host 10 # - unlink: needs bulk sink and/or src, test HCD unlink processing 17 # this default provides a steady test load for a bulk device 37 # NOTE: the 'in' and 'out' cases are usually bulk, but can be
|
/linux-6.14.4/Documentation/usb/ |
D | ehci.rst | 58 At this writing the driver should comfortably handle all control, bulk, 125 and bulk transfers. Shows each active qh and the qtds 160 Bulk transfers are most often used where throughput is an issue. It's 161 good to keep in mind that bulk transfers are always in 512 byte packets, 165 So more than 50 MByte/sec is available for bulk transfers, when both 195 you issue a control or bulk request you can often expect to learn that 203 or using bulk queuing if a series of small requests needs to be issued. 213 I/O be efficient, it's better to just queue up several (bulk) requests
|
D | gadget_serial.rst | 89 with bulk in and bulk out endpoints, and it is treated similarly 93 or any driver that can talk to a device with a simple bulk in/out 122 To load it as a vendor specific bulk in/out device, do this:: 150 serial is loaded as a bulk in/out device, you will want to use the 209 E: Ad=81(I) Atr=02(Bulk) MxPS= 512 Ivl=0ms 210 E: Ad=02(O) Atr=02(Bulk) MxPS= 512 Ivl=0ms 239 E: Ad=81(I) Atr=02(Bulk) MxPS= 512 Ivl=0ms 240 E: Ad=02(O) Atr=02(Bulk) MxPS= 512 Ivl=0ms
|
/linux-6.14.4/drivers/media/usb/dvb-usb-v2/ |
D | mxl111sf.c | 30 MODULE_PARM_DESC(isoc, "enable usb isoc xfer (0=bulk, 1=isoc)."); 1040 stream->u.bulk.buffersize = 8192; in mxl111sf_stream_config_bulk() 1059 * bulk EP4/BULK/5/8192 1109 * bulk EP6/BULK/5/8192 1151 * bulk EP5/BULK/5/8192/RAW 1193 * bulk EP6/BULK/5/8192 EP4/BULK/5/8192 EP5/BULK/5/8192/RAW 1262 * tp bulk EP6/BULK/5/8192 EP4/BULK/5/8192 EP6/BULK/5/8192/RAW 1264 * spi bulk EP6/BULK/5/8192 EP4/BULK/5/8192 EP5/BULK/5/8192/RAW 1341 * tp bulk EP4/BULK/5/8192 EP6/BULK/5/8192/RAW 1343 * spi bulk EP4/BULK/5/8192 EP5/BULK/5/8192/RAW
|
/linux-6.14.4/drivers/usb/misc/ |
D | usblcd.c | 48 bulk in endpoint */ 50 bulk out endpoint */ 140 /* do a blocking bulk read to get data from the device */ in lcd_read() 208 "nonzero write bulk status received: %d\n", status); in lcd_write_bulk_callback() 270 /* send the data out the bulk port */ in lcd_write() 345 /* use only the first bulk-in and bulk-out endpoints */ in lcd_probe() 350 "Could not find both bulk-in and bulk-out endpoints\n"); in lcd_probe()
|
/linux-6.14.4/drivers/media/usb/uvc/ |
D | uvc_video.c | 1111 * uvc_video_decode_start is called with URB data at the start of a bulk or 1129 * uvc_video_decode_end is called with header data at the end of a bulk or 1368 * uvc_video_encode_data(). Only bulk transfers are currently supported. 1397 nbytes = min(stream->bulk.max_payload_size - stream->bulk.payload_size, in uvc_video_encode_data() 1586 if (urb->actual_length == 0 && stream->bulk.header_size == 0) in uvc_video_decode_bulk() 1591 stream->bulk.payload_size += len; in uvc_video_decode_bulk() 1597 if (stream->bulk.header_size == 0 && !stream->bulk.skip_payload) { in uvc_video_decode_bulk() 1606 stream->bulk.skip_payload = 1; in uvc_video_decode_bulk() 1608 memcpy(stream->bulk.header, mem, ret); in uvc_video_decode_bulk() 1609 stream->bulk.header_size = ret; in uvc_video_decode_bulk() [all …]
|
/linux-6.14.4/drivers/usb/serial/ |
D | ipw.c | 90 #define IPW_SIO_RXCTL 0x00 /* control bulk rx channel transmissions, value=1/0 (on/off) */ 159 /* reset the bulk pipes */ in ipw_open() 164 dev_dbg(dev, "%s: setting up bulk read callback\n", __func__); in ipw_open() 167 /*--3: Tell the modem to open the floodgates on the rx bulk channel */ in ipw_open() 178 dev_err(dev, "Enabling bulk RxRead failed (error = %d)\n", result); in ipw_open() 269 /* send RXBULK_off (tell modem to stop transmitting bulk data on in ipw_close() 281 dev_err(dev, "Disabling bulk RxRead failed (error = %d)\n", result); in ipw_close()
|
/linux-6.14.4/fs/xfs/libxfs/ |
D | xfs_btree_staging.c | 25 * use to construct a new btree index using the btree bulk loader code. The 26 * bulk loading code uses the staging btree cursor to abstract the details of 32 * construction by the bulk loading code. Staging btree cursors point to fake 36 * the _stage_cursor function for a specific btree type. When bulk loading is 42 * Bulk Loading for AG Btrees 96 * Bulk Loading for Inode-Rooted Btrees 165 * Bulk Loading of Staged Btrees 177 * "Bulk Loading for AG Btrees" and "Bulk Loading for Inode-Rooted Btrees" for 189 * save them for later use by ->claim_block(). Bulk loading requires all 234 * The btree blocks will be written to disk when bulk loading is finished. [all …]
|