| /linux/drivers/gpu/drm/i915/gvt/ |
| H A D | cmd_parser.c | 522 struct intel_vgpu_workload *workload; member 860 u32 base = s->workload->engine->mmio_base; in is_cmd_update_pdps() 868 struct intel_vgpu_mm *shadow_mm = s->workload->shadow_mm; in cmd_pdp_mmio_update_handler() 884 &s->workload->lri_shadow_mm); in cmd_pdp_mmio_update_handler() 1024 s->workload->ring_context_gpa + 12, &ctx_sr_ctl, 4); in cmd_reg_handler() 1235 s->workload->pending_events); in cmd_handler_pipe_control() 1242 s->workload->pending_events); in cmd_handler_mi_user_interrupt() 1758 s->workload->pending_events); in cmd_handler_mi_flush_dw() 1829 s->vgpu->gtt.ggtt_mm : s->workload->shadow_mm; in find_bb_size() 1845 s->engine->name, s->workload); in find_bb_size() [all …]
|
| H A D | cmd_parser.h | 50 int intel_gvt_scan_and_shadow_ringbuffer(struct intel_vgpu_workload *workload); 56 int intel_gvt_scan_engine_context(struct intel_vgpu_workload *workload);
|
| H A D | scheduler.h | 137 void intel_vgpu_queue_workload(struct intel_vgpu_workload *workload); 164 void intel_vgpu_destroy_workload(struct intel_vgpu_workload *workload);
|
| /linux/Documentation/accel/amdxdna/ |
| H A D | amdnpu.rst | 34 partition which can be bound to a workload context. 55 XDNA Array partition setup, XDNA Array configuration, workload context 56 management and workload orchestration. 59 called ERT to service each workload context. ERT is also used to execute user 60 provided ``ctrlcode`` associated with the workload context. 74 workload context. The user channel is primarily used for submitting work to 125 may be *exclusively* bound to one workload context while another partition may 126 be *temporarily* bound to more than one workload contexts. The microcontroller 134 of 2D array among various workloads. Every workload describes the number 136 component uses hints passed by the workload and its own heuristics to [all …]
|
| /linux/Documentation/ABI/stable/ |
| H A D | sysfs-driver-qaic | 10 ASSIGNED (1) DBC is activated and a workload is running on device 11 BEFORE_SHUTDOWN (2) Sub-system associated with this workload has crashed and 13 AFTER_SHUTDOWN (3) Sub-system associated with this workload has crashed and 15 BEFORE_POWER_UP (4) Sub-system associated with this workload is shutdown and 17 AFTER_POWER_UP (5) Sub-system associated with this workload is now powered up
|
| /linux/Documentation/admin-guide/ |
| H A D | workload-tracing.rst | 4 Discovering Linux kernel subsystems used by a workload 14 * Understanding system resources necessary to build and run a workload 17 in use by a workload. The completeness of the system usage information 18 depends on the completeness of coverage of a workload. 24 * Once we discover and understand the workload needs, we can focus on them 32 the system resources in use by a workload. Once we discover and understand 33 the workload needs, we can focus on them to avoid regressions and use it 37 the workload and doesn't include all the system calls that can be invoked 39 these system calls that are invoked. As an example, if a workload opens a 42 is a workload that provides full coverage of a workload then the method [all …]
|
| /linux/tools/perf/tests/shell/ |
| H A D | stat_bpf_counters.sh | 7 workload="perf test -w sqrtloop" 44 …base_instructions=$(perf stat --no-big-num -e instructions -- $workload 2>&1 | awk '/instructions/… 45 …bpf_instructions=$(perf stat --no-big-num --bpf-counters -e instructions -- $workload 2>&1 | awk … 54 …num -e instructions/name=base_instructions/,instructions/name=bpf_instructions/b -- $workload 2>&1)
|
| H A D | trace_summary.sh | 17 workload="true" 20 echo "testing: perf trace ${args} -- ${workload}" 21 perf trace ${args} -- ${workload} >${OUTPUT} 2>&1
|
| H A D | stat_metrics_values.sh | 14 workload="perf bench futex hash -r 2 -s" 22 $PYTHON $pythonvalidator -rule $rulefile -output_dir $tmpdir -wl "${workload}" \
|
| /linux/tools/perf/tests/ |
| H A D | perf-record.c | 129 err = sched__get_first_possible_cpu(evlist->workload.pid, cpu_mask); in test__PERF_RECORD() 142 if (sched_setaffinity(evlist->workload.pid, cpu_mask_size, cpu_mask) < 0) { in test__PERF_RECORD() 231 if ((pid_t)sample.pid != evlist->workload.pid) { in test__PERF_RECORD() 233 name, evlist->workload.pid, sample.pid); in test__PERF_RECORD() 237 if ((pid_t)sample.tid != evlist->workload.pid) { in test__PERF_RECORD() 239 name, evlist->workload.pid, sample.tid); in test__PERF_RECORD() 248 (pid_t)event->comm.pid != evlist->workload.pid) { in test__PERF_RECORD()
|
| /linux/Documentation/accel/qaic/ |
| H A D | aic100.rst | 67 with the QSM. Except for workload data via the DMA Bridge, all interaction with 83 (Q6) DSP with HVX and HMX. Each NSP can only run one workload at a time, but 84 multiple NSPs may be assigned to a single workload. Since each NSP can only run 85 one workload, AIC100 is limited to 16 concurrent workloads. Workload 95 workload is assigned a single DMA Bridge channel. The DMA Bridge exposes 118 1. Compile the workload into an ELF targeting the NSP(s) 119 2. Make requests to the QSM to load the workload and related artifacts into the 121 3. Make a request to the QSM to activate the workload onto a set of idle NSPs 122 4. Make requests to the DMA Bridge to send input data to the workload to be 124 workload. [all …]
|
| H A D | qaic.rst | 18 if the workload is particularly quick, and the host is responsive. If the host 22 workload's ability to process inputs. The lprnet (license plate reader network) 23 workload is known to trigger this condition, and can generate in excess of 100k 32 sleep for a time to see if the workload will generate more activity. The IRQ 39 workload throughput performance (within run-to-run noise variation). 112 or receive data from a workload. The call will return a GEM handle that 123 get sent where to a workload. This requires a set of DMA transfers for the 166 workload should be allowed to interface with the DBC.
|
| /linux/tools/perf/tests/shell/lib/ |
| H A D | perf_metric_validation.py | 22 \tis {1} in workload(s): {2} \n\ 31 \tis {1} in workload(s): {2}\n\ 39 workload='true', metrics='', cputype='cpu'): argument 49 self.workloads = [x for x in workload.split(",") if x] 380 def _run_perf(self, metric, workload: str): 383 wl = workload.split() 392 def collect_perf(self, workload: str): 416 wl = workload 425 workload = self.workloads[self.wlidx] 427 data = self._run_perf(metric, workload) [all …]
|
| /linux/Documentation/admin-guide/mm/ |
| H A D | idle_page_tracking.rst | 9 accessed by a workload and which are idle. This information can be useful for 10 estimating the workload's working set size, which, in turn, can be taken into 11 account when configuring the workload parameters, setting memory cgroup limits, 12 or deciding where to place the workload within a compute cluster. 51 workload one should: 53 1. Mark all the workload's pages as idle by setting corresponding bits in 55 ``/proc/pid/pagemap`` if the workload is represented by a process, or by 56 filtering out alien pages using ``/proc/kpagecgroup`` in case the workload 59 2. Wait until the workload accesses its working set.
|
| /linux/rust/pin-init/examples/ |
| H A D | static_init.rs | 93 let workload = 1_000; in main() localVariable 100 for _ in 0..workload { in main() 109 for _ in 0..workload { in main() 122 assert_eq!(*mtx.lock(), workload * thread_count * 2); in main()
|
| H A D | pthread_mutex.rs | 158 let workload = 1_000_000; in main() localVariable 165 for _ in 0..workload { in main() 170 for _ in 0..workload { in main() 182 assert_eq!(*mtx.lock(), workload * thread_count * 2); in main()
|
| H A D | mutex.rs | 199 let workload = if cfg!(miri) { 100 } else { 1_000 }; in main() localVariable 206 for _ in 0..workload { in main() 211 for _ in 0..workload { in main() 223 assert_eq!(*mtx.lock(), workload * thread_count * 2); in main()
|
| /linux/Documentation/driver-api/thermal/ |
| H A D | intel_dptf.rst | 186 Available workload types. User space can specify one of the workload type 191 User space can specify any one of the available workload type using 363 workload type and passing hints regarding it to the OS. A special sysfs 364 interface is provided to allow user space to obtain workload type hints from 376 Enable firmware to send workload type hints to user space. 379 Enable firmware to send slow workload type hints to user space. 384 the workload type prediction in the firmware and notifying the OS about 391 Predicted workload type index. User space can get notification of 417 The hardware analyzes workload residencies over an extended period to 418 determine whether the workload classification tends toward idle/battery [all …]
|
| /linux/tools/perf/bench/ |
| H A D | find-bit-bench.c | 34 static noinline void workload(int val) in workload() function 82 workload(bit); in do_for_each_set_bit() 97 workload(bit); in do_for_each_set_bit()
|
| /linux/Documentation/filesystems/nfs/ |
| H A D | knfsd-stats.rst | 54 Depending on the NFS workload patterns and various network stack 58 However this is a more accurate and less workload-dependent measure 74 pool for the NFS workload (the workload is thread-limited), in which 76 performance of the NFS workload. 93 threads configured than can be used by the NFS workload. This is 99 slow; the idle timeout is 60 minutes. Unless the NFS workload
|
| /linux/drivers/gpu/drm/amd/pm/powerplay/hwmgr/ |
| H A D | pp_psm.c | 275 long workload[1]; in psm_adjust_power_state_dynamic() local 300 workload[0] = hwmgr->workload_setting[index]; in psm_adjust_power_state_dynamic() 302 if (hwmgr->power_profile_mode != workload[0] && hwmgr->hwmgr_func->set_power_profile_mode) in psm_adjust_power_state_dynamic() 303 hwmgr->hwmgr_func->set_power_profile_mode(hwmgr, workload, 0); in psm_adjust_power_state_dynamic()
|
| /linux/Documentation/gpu/ |
| H A D | drm-compute.rst | 11 have their memory swapped out from them. Or they simply want their workload 19 waiting for the workload to complete. Effectively this terminates the workload 22 Since this is undesirable, there need to be mitigations to prevent a workload
|
| /linux/tools/perf/Documentation/ |
| H A D | perf-test.txt | 57 --workload=:: 58 Run a built-in workload, to list them use '--list-workloads', current ones include: 71 List the available workloads to use with -w/--workload.
|
| H A D | perf-sched.txt | 18 of an arbitrary workload. 21 and other scheduling properties of the workload. 43 'perf sched script' to see a detailed trace of the workload that 46 'perf sched replay' to simulate the workload that was recorded 48 that mimic the workload based on the events in the trace. These 50 of the workload as it occurred when it was recorded - and can repeat 54 workload captured via perf sched record. Columns stand for 90 scheduler behavior for the workload. 496 repeat the workload n times (0: infinite). Default is 10.
|
| /linux/Documentation/scheduler/ |
| H A D | sched-capacity.rst | 73 With a workload that periodically does a fixed amount of work, you will get an 104 Executing the same workload as described in 1.3.1, which each CPU running at its 112 workload on CPU1 152 One issue that needs to be taken into account is that a workload's duty cycle is 154 periodic workload at a given frequency F:: 163 Now, consider running the *same* workload at frequency F/2:: 185 identical workload on CPUs of different capacity values will yield different 193 Executing a given periodic workload on each CPU at their maximum frequency would 384 workload on CPU0 391 workload on CPU1 [all …]
|