| /linux/fs/ceph/ |
| H A D | caps.c | 579 unsigned issued) in __check_cap_issue() argument 593 (issued & (CEPH_CAP_FILE_CACHE|CEPH_CAP_FILE_LAZYIO)) && in __check_cap_issue() 604 if ((issued & CEPH_CAP_FILE_SHARED) != (had & CEPH_CAP_FILE_SHARED)) { in __check_cap_issue() 605 if (issued & CEPH_CAP_FILE_SHARED) in __check_cap_issue() 615 !(issued & CEPH_CAP_DIR_CREATE)) { in __check_cap_issue() 653 unsigned issued, unsigned wanted, in ceph_add_cap() argument 669 ceph_cap_string(issued), seq); in ceph_add_cap() 678 cap->issued = 0; in ceph_add_cap() 700 cap->issued = cap->implemented = CEPH_CAP_PIN; in ceph_add_cap() 716 issued |= cap->issued; in ceph_add_cap() [all …]
|
| H A D | inode.c | 804 int ceph_fill_file_size(struct inode *inode, int issued, in ceph_fill_file_size() argument 834 WARN_ON_ONCE(issued & (CEPH_CAP_FILE_RD | in ceph_fill_file_size() 842 if ((issued & (CEPH_CAP_FILE_CACHE| in ceph_fill_file_size() 876 void ceph_fill_file_time(struct inode *inode, int issued, in ceph_fill_file_time() argument 887 if (issued & (CEPH_CAP_FILE_EXCL| in ceph_fill_file_time() 916 } else if (issued & CEPH_CAP_FILE_EXCL) { in ceph_fill_file_time() 985 int issued, new_issued, info_caps; in ceph_fill_inode() local 1074 __ceph_caps_issued(ci, &issued); in ceph_fill_inode() 1075 issued |= __ceph_caps_dirty(ci); in ceph_fill_inode() 1076 new_issued = ~issued & info_caps; in ceph_fill_inode() [all …]
|
| H A D | super.h | 211 int issued; /* latest, from the mds */ member 255 int issued, dirty; member 799 int issued; in ceph_caps_issued() local 801 issued = __ceph_caps_issued(ci, NULL); in ceph_caps_issued() 803 return issued; in ceph_caps_issued() 1070 extern int ceph_fill_file_size(struct inode *inode, int issued, 1072 extern void ceph_fill_file_time(struct inode *inode, int issued, 1250 unsigned issued, unsigned wanted,
|
| H A D | xattr.c | 332 int issued; in ceph_vxattrcb_caps() local 335 issued = __ceph_caps_issued(ci, NULL); in ceph_vxattrcb_caps() 339 ceph_cap_string(issued), issued); in ceph_vxattrcb_caps() 1184 int issued; in __ceph_setxattr() local 1233 issued = __ceph_caps_issued(ci, NULL); in __ceph_setxattr() 1235 if ((ci->i_xattrs.version == 0) || !(issued & CEPH_CAP_XATTR_EXCL) || in __ceph_setxattr() 1254 ceph_vinop(inode), name, ceph_cap_string(issued)); in __ceph_setxattr()
|
| /linux/Documentation/driver-api/dmaengine/ |
| H A D | pxa_dma.rst | 14 A transfer which is submitted and issued on a phy doesn't wait for a phy to 20 Any issued transfer with DMA_PREP_INTERRUPT should trigger a callback call. 29 not be issued until the next "start of frame interrupt", hence the need to 44 assigned on the fly when the transfer is issued. 137 Residue granularity will be descriptor based. The issued but not completed 150 - a driver issued tx1+tx2 => channel is running in aligned mode 157 - a driver issued tx4 => tx4 is put in issued queue, not chained 165 This translates into (after tx4 is issued) : 167 - issued queue
|
| /linux/drivers/md/ |
| H A D | dm-cache-background-tracker.c | 20 struct list_head issued; member 41 INIT_LIST_HEAD(&b->issued); in btracker_create() 54 BUG_ON(!list_empty(&b->issued)); in btracker_destroy() 193 list_add(&w->list, &b->issued); in btracker_queue() 213 list_move(&w->list, &b->issued); in btracker_issue()
|
| /linux/drivers/vdpa/mlx5/core/ |
| H A D | resources.c | 333 int issued, in issue_async_cmd() argument 337 struct mlx5_vdpa_async_cmd *cmd = &cmds[issued]; in issue_async_cmd() 347 if (*completed < issued) { in issue_async_cmd() 371 int issued = 0; in mlx5_vdpa_exec_async_cmds() local 377 while (issued < num_cmds) { in mlx5_vdpa_exec_async_cmds() 379 err = issue_async_cmd(mvdev, cmds, issued, &completed); in mlx5_vdpa_exec_async_cmds() 382 issued, num_cmds, err); in mlx5_vdpa_exec_async_cmds() 386 issued++; in mlx5_vdpa_exec_async_cmds() 389 while (completed < issued) in mlx5_vdpa_exec_async_cmds()
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-class-scsi_tape | 18 reads, writes, and other SCSI commands issued to the tape 33 The number of I/O requests issued to the tape drive other 54 Shows the total number of read requests issued to the tape 85 Shows the total number of write requests issued to the tape
|
| H A D | sysfs-platform-hidma-mgmt | 26 reset command is issued. If the value is chosen too short, 63 read transactions that can be issued back to back. 84 write transactions that can be issued back to back.
|
| H A D | sysfs-platform-dfl-port | 61 Description: Read-only. Read this file to get the status of issued command 75 Description: Read-only. Read this file to get the status of issued command
|
| /linux/drivers/dma/ioat/ |
| H A D | dma.c | 174 ioat_chan->issued = ioat_chan->head; in __ioat_issue_pending() 180 ioat_chan->issued, ioat_chan->dmacount); in __ioat_issue_pending() 220 __func__, ioat_chan->head, ioat_chan->tail, ioat_chan->issued); in __ioat_start_null_desc() 252 ioat_chan->issued = ioat_chan->tail; in __ioat_restart_chan() 259 ioat_chan->issued, ioat_chan->dmacount); in __ioat_restart_chan() 474 ioat_chan->tail, ioat_chan->issued); in ioat_check_space_lock() 483 ioat_chan->tail, ioat_chan->issued); in ioat_check_space_lock() 600 __func__, ioat_chan->head, ioat_chan->tail, ioat_chan->issued); in __ioat_cleanup()
|
| H A D | dma.h | 130 u16 issued; member 324 return CIRC_CNT(ioat_chan->head, ioat_chan->issued, in ioat_ring_pending()
|
| /linux/Documentation/admin-guide/ |
| H A D | iostats.rst | 114 read I/Os issued per partition should equal those made to the disks ... 138 Field 1 -- # of reads issued 139 This is the total number of reads issued to this partition. 145 Field 3 -- # of writes issued 146 This is the total number of writes issued to this partition. 162 the number of reads/writes issued can be several times higher than the
|
| H A D | syscall-user-dispatch.rst | 35 architectures, like x86, but at least for Wine, syscalls issued by 39 Since this mechanism is designed to capture syscalls issued by 89 quickly capture system calls issued by a non-native part of the
|
| /linux/Documentation/admin-guide/hw-vuln/ |
| H A D | vmscape.rst | 40 between VM-exit and the next VM-entry, no IBPB is issued. 60 protection. Warning is not issued when: 91 IBPB is issued on every VM-exit. This occurs when other mitigations like
|
| /linux/drivers/mmc/core/ |
| H A D | queue.c | 238 enum mmc_issued issued; in mmc_mq_queue_rq() local 306 issued = mmc_blk_mq_issue_rq(mq, req); in mmc_mq_queue_rq() 308 switch (issued) { in mmc_mq_queue_rq() 320 if (issued != MMC_REQ_STARTED) { in mmc_mq_queue_rq()
|
| /linux/drivers/dma/ |
| H A D | apple-admac.c | 107 struct list_head issued; member 398 list_for_each_entry(adtx, &adchan->issued, node) { in admac_tx_status() 481 list_splice_tail_init(&adchan->submitted, &adchan->issued); in admac_issue_pending() 482 if (!list_empty(&adchan->issued) && !adchan->current_tx) { in admac_issue_pending() 483 tx = list_first_entry(&adchan->issued, struct admac_tx, node); in admac_issue_pending() 529 list_splice_tail_init(&adchan->issued, &adchan->to_free); in admac_terminate_all() 886 INIT_LIST_HEAD(&adchan->issued); in admac_probe()
|
| /linux/Documentation/networking/xfrm/ |
| H A D | xfrm_sync.rst | 136 a) XFRM_MSG_GETAE issued by user-->kernel. 147 b) XFRM_MSG_NEWAE is issued by either user space to configure 155 A response is issued back to the sender in user space to indicate success 159 XFRM_MSG_NEWAE is also issued to any listeners as described in iii).
|
| /linux/fs/f2fs/ |
| H A D | segment.c | 1243 unsigned int *issued) in __submit_zone_reset_cmd() argument 1257 if (issued) in __submit_zone_reset_cmd() 1258 (*issued)++; in __submit_zone_reset_cmd() 1280 struct discard_cmd *dc, int *issued) in __submit_discard_cmd() argument 1306 wait_list, issued); in __submit_discard_cmd() 1330 while (total_len && *issued < dpolicy->max_requests) { in __submit_discard_cmd() 1340 (*issued)++; in __submit_discard_cmd() 1341 if (*issued == dpolicy->max_requests) in __submit_discard_cmd() 1585 struct discard_policy *dpolicy, int *issued) in __issue_discard_cmd_orderly() argument 1615 err = __submit_discard_cmd(sbi, dpolicy, dc, issued); in __issue_discard_cmd_orderly() [all …]
|
| /linux/Documentation/block/ |
| H A D | ublk.rst | 215 Requests which have not been issued to userspace are requeued. Requests 216 which have been issued to userspace are aborted. 220 requests which have been issued to userspace are requeued and will be 221 re-issued to the new process after handling ``UBLK_CMD_END_USER_RECOVERY``. 227 exits, requests which have issued to userspace are failed, as are any 228 subsequently issued requests. Applications continuously issuing I/O against 302 issued to ublk server without data copy. Then, IO backend of ublk server 431 IOs can't be issued concurrently to io_uring if there is any dependency 441 uring_cmd is issued on a different ``io_ring_ctx``, manual buffer
|
| H A D | stat.rst | 87 This value counts the number of I/O requests that have been issued to 89 requests that are in the queue but not yet issued to the device driver.
|
| /linux/Documentation/hid/ |
| H A D | hidraw.rst | 138 state of a report. The format of the buffer issued with this report is identical 149 the regular device read() interface. The format of the buffer issued with this report 159 updates via the regular device write() interface. The format of the buffer issued 169 of the buffer issued with this report is identical to that of HIDIOCGFEATURE.
|
| /linux/Documentation/driver-api/firmware/ |
| H A D | lookup-order.rst | 19 is issued or the custom firmware loading is relied upon for firmware
|
| /linux/Documentation/admin-guide/laptops/ |
| H A D | disk-shock-protection.rst | 46 no further disk head park request has been issued in the meantime, 83 been issued to a device successfully, all I/O operations on the 92 which means that a user-issued head parking request to one device may
|
| /linux/Documentation/arch/powerpc/ |
| H A D | vas-api.rst | 102 several open calls have to be issued. 238 issued. This signal returns with the following siginfo struct:: 251 the thread is exited without closing the window and the request is issued 252 using this window. the signal will be issued to the thread group leader
|