1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3 * uncore-frquency-tpmi: Uncore frequency scaling using TPMI
4 *
5 * Copyright (c) 2023, Intel Corporation.
6 * All Rights Reserved.
7 *
8 * The hardware interface to read/write is basically substitution of
9 * MSR 0x620 and 0x621.
10 * There are specific MMIO offset and bits to get/set minimum and
11 * maximum uncore ratio, similar to MSRs.
12 * The scope of the uncore MSRs was package scope. But TPMI allows
13 * new gen CPUs to have multiple uncore controls at uncore-cluster
14 * level. Each package can have multiple power domains which further
15 * can have multiple clusters.
16 * Here number of power domains = number of resources in this aux
17 * device. There are offsets and bits to discover number of clusters
18 * and offset for each cluster level controls.
19 *
20 */
21
22 #include <linux/auxiliary_bus.h>
23 #include <linux/bitfield.h>
24 #include <linux/bits.h>
25 #include <linux/intel_tpmi.h>
26 #include <linux/intel_vsec.h>
27 #include <linux/io.h>
28 #include <linux/module.h>
29
30 #include "../tpmi_power_domains.h"
31 #include "uncore-frequency-common.h"
32
33 #define UNCORE_MAJOR_VERSION 0
34 #define UNCORE_MINOR_VERSION 2
35 #define UNCORE_ELC_SUPPORTED_VERSION 2
36 #define UNCORE_HEADER_INDEX 0
37 #define UNCORE_FABRIC_CLUSTER_OFFSET 8
38
39 /* status + control + adv_ctl1 + adv_ctl2 */
40 #define UNCORE_FABRIC_CLUSTER_SIZE (4 * 8)
41
42 #define UNCORE_STATUS_INDEX 0
43 #define UNCORE_CONTROL_INDEX 8
44
45 #define UNCORE_FREQ_KHZ_MULTIPLIER 100000
46
47 struct tpmi_uncore_struct;
48
49 /* Information for each cluster */
50 struct tpmi_uncore_cluster_info {
51 bool root_domain;
52 bool elc_supported;
53 u8 __iomem *cluster_base;
54 u16 cdie_id;
55 struct uncore_data uncore_data;
56 struct tpmi_uncore_struct *uncore_root;
57 };
58
59 /* Information for each power domain */
60 struct tpmi_uncore_power_domain_info {
61 u8 __iomem *uncore_base;
62 int ufs_header_ver;
63 int cluster_count;
64 struct tpmi_uncore_cluster_info *cluster_infos;
65 };
66
67 /* Information for all power domains in a package */
68 struct tpmi_uncore_struct {
69 int power_domain_count;
70 int max_ratio;
71 int min_ratio;
72 struct tpmi_uncore_power_domain_info *pd_info;
73 struct tpmi_uncore_cluster_info root_cluster;
74 bool write_blocked;
75 };
76
77 /* Bit definitions for STATUS register */
78 #define UNCORE_CURRENT_RATIO_MASK GENMASK_ULL(6, 0)
79
80 /* Bit definitions for CONTROL register */
81 #define UNCORE_MAX_RATIO_MASK GENMASK_ULL(14, 8)
82 #define UNCORE_MIN_RATIO_MASK GENMASK_ULL(21, 15)
83 #define UNCORE_EFF_LAT_CTRL_RATIO_MASK GENMASK_ULL(28, 22)
84 #define UNCORE_EFF_LAT_CTRL_LOW_THRESHOLD_MASK GENMASK_ULL(38, 32)
85 #define UNCORE_EFF_LAT_CTRL_HIGH_THRESHOLD_ENABLE BIT(39)
86 #define UNCORE_EFF_LAT_CTRL_HIGH_THRESHOLD_MASK GENMASK_ULL(46, 40)
87
88 /* Helper function to read MMIO offset for max/min control frequency */
read_control_freq(struct tpmi_uncore_cluster_info * cluster_info,unsigned int * value,enum uncore_index index)89 static void read_control_freq(struct tpmi_uncore_cluster_info *cluster_info,
90 unsigned int *value, enum uncore_index index)
91 {
92 u64 control;
93
94 control = readq(cluster_info->cluster_base + UNCORE_CONTROL_INDEX);
95 if (index == UNCORE_INDEX_MAX_FREQ)
96 *value = FIELD_GET(UNCORE_MAX_RATIO_MASK, control) * UNCORE_FREQ_KHZ_MULTIPLIER;
97 else
98 *value = FIELD_GET(UNCORE_MIN_RATIO_MASK, control) * UNCORE_FREQ_KHZ_MULTIPLIER;
99 }
100
101 /* Helper function to read efficiency latency control values over MMIO */
read_eff_lat_ctrl(struct uncore_data * data,unsigned int * val,enum uncore_index index)102 static int read_eff_lat_ctrl(struct uncore_data *data, unsigned int *val, enum uncore_index index)
103 {
104 struct tpmi_uncore_cluster_info *cluster_info;
105 u64 ctrl;
106
107 cluster_info = container_of(data, struct tpmi_uncore_cluster_info, uncore_data);
108 if (cluster_info->root_domain)
109 return -ENODATA;
110
111 if (!cluster_info->elc_supported)
112 return -EOPNOTSUPP;
113
114 ctrl = readq(cluster_info->cluster_base + UNCORE_CONTROL_INDEX);
115
116 switch (index) {
117 case UNCORE_INDEX_EFF_LAT_CTRL_LOW_THRESHOLD:
118 *val = FIELD_GET(UNCORE_EFF_LAT_CTRL_LOW_THRESHOLD_MASK, ctrl);
119 *val *= 100;
120 *val = DIV_ROUND_UP(*val, FIELD_MAX(UNCORE_EFF_LAT_CTRL_LOW_THRESHOLD_MASK));
121 break;
122
123 case UNCORE_INDEX_EFF_LAT_CTRL_HIGH_THRESHOLD:
124 *val = FIELD_GET(UNCORE_EFF_LAT_CTRL_HIGH_THRESHOLD_MASK, ctrl);
125 *val *= 100;
126 *val = DIV_ROUND_UP(*val, FIELD_MAX(UNCORE_EFF_LAT_CTRL_HIGH_THRESHOLD_MASK));
127 break;
128
129 case UNCORE_INDEX_EFF_LAT_CTRL_HIGH_THRESHOLD_ENABLE:
130 *val = FIELD_GET(UNCORE_EFF_LAT_CTRL_HIGH_THRESHOLD_ENABLE, ctrl);
131 break;
132 case UNCORE_INDEX_EFF_LAT_CTRL_FREQ:
133 *val = FIELD_GET(UNCORE_EFF_LAT_CTRL_RATIO_MASK, ctrl) * UNCORE_FREQ_KHZ_MULTIPLIER;
134 break;
135
136 default:
137 return -EOPNOTSUPP;
138 }
139
140 return 0;
141 }
142
143 #define UNCORE_MAX_RATIO FIELD_MAX(UNCORE_MAX_RATIO_MASK)
144
145 /* Helper for sysfs read for max/min frequencies. Called under mutex locks */
uncore_read_control_freq(struct uncore_data * data,unsigned int * value,enum uncore_index index)146 static int uncore_read_control_freq(struct uncore_data *data, unsigned int *value,
147 enum uncore_index index)
148 {
149 struct tpmi_uncore_cluster_info *cluster_info;
150
151 cluster_info = container_of(data, struct tpmi_uncore_cluster_info, uncore_data);
152
153 if (cluster_info->root_domain) {
154 struct tpmi_uncore_struct *uncore_root = cluster_info->uncore_root;
155 unsigned int min, max, v;
156 int i;
157
158 min = UNCORE_MAX_RATIO * UNCORE_FREQ_KHZ_MULTIPLIER;
159 max = 0;
160
161 /*
162 * Get the max/min by looking at each cluster. Get the lowest
163 * min and highest max.
164 */
165 for (i = 0; i < uncore_root->power_domain_count; ++i) {
166 int j;
167
168 for (j = 0; j < uncore_root->pd_info[i].cluster_count; ++j) {
169 read_control_freq(&uncore_root->pd_info[i].cluster_infos[j],
170 &v, index);
171 if (v < min)
172 min = v;
173 if (v > max)
174 max = v;
175 }
176 }
177
178 if (index == UNCORE_INDEX_MIN_FREQ)
179 *value = min;
180 else
181 *value = max;
182
183 return 0;
184 }
185
186 read_control_freq(cluster_info, value, index);
187
188 return 0;
189 }
190
191 /* Helper function for writing efficiency latency control values over MMIO */
write_eff_lat_ctrl(struct uncore_data * data,unsigned int val,enum uncore_index index)192 static int write_eff_lat_ctrl(struct uncore_data *data, unsigned int val, enum uncore_index index)
193 {
194 struct tpmi_uncore_cluster_info *cluster_info;
195 u64 control;
196
197 cluster_info = container_of(data, struct tpmi_uncore_cluster_info, uncore_data);
198
199 if (cluster_info->root_domain)
200 return -ENODATA;
201
202 if (!cluster_info->elc_supported)
203 return -EOPNOTSUPP;
204
205 switch (index) {
206 case UNCORE_INDEX_EFF_LAT_CTRL_LOW_THRESHOLD:
207 if (val > 100)
208 return -EINVAL;
209 break;
210
211 case UNCORE_INDEX_EFF_LAT_CTRL_HIGH_THRESHOLD:
212 if (val > 100)
213 return -EINVAL;
214 break;
215
216 case UNCORE_INDEX_EFF_LAT_CTRL_HIGH_THRESHOLD_ENABLE:
217 if (val > 1)
218 return -EINVAL;
219 break;
220
221 case UNCORE_INDEX_EFF_LAT_CTRL_FREQ:
222 val /= UNCORE_FREQ_KHZ_MULTIPLIER;
223 if (val > FIELD_MAX(UNCORE_EFF_LAT_CTRL_RATIO_MASK))
224 return -EINVAL;
225 break;
226
227 default:
228 return -EOPNOTSUPP;
229 }
230
231 control = readq(cluster_info->cluster_base + UNCORE_CONTROL_INDEX);
232
233 switch (index) {
234 case UNCORE_INDEX_EFF_LAT_CTRL_LOW_THRESHOLD:
235 val *= FIELD_MAX(UNCORE_EFF_LAT_CTRL_LOW_THRESHOLD_MASK);
236 val /= 100;
237 control &= ~UNCORE_EFF_LAT_CTRL_LOW_THRESHOLD_MASK;
238 control |= FIELD_PREP(UNCORE_EFF_LAT_CTRL_LOW_THRESHOLD_MASK, val);
239 break;
240
241 case UNCORE_INDEX_EFF_LAT_CTRL_HIGH_THRESHOLD:
242 val *= FIELD_MAX(UNCORE_EFF_LAT_CTRL_HIGH_THRESHOLD_MASK);
243 val /= 100;
244 control &= ~UNCORE_EFF_LAT_CTRL_HIGH_THRESHOLD_MASK;
245 control |= FIELD_PREP(UNCORE_EFF_LAT_CTRL_HIGH_THRESHOLD_MASK, val);
246 break;
247
248 case UNCORE_INDEX_EFF_LAT_CTRL_HIGH_THRESHOLD_ENABLE:
249 control &= ~UNCORE_EFF_LAT_CTRL_HIGH_THRESHOLD_ENABLE;
250 control |= FIELD_PREP(UNCORE_EFF_LAT_CTRL_HIGH_THRESHOLD_ENABLE, val);
251 break;
252
253 case UNCORE_INDEX_EFF_LAT_CTRL_FREQ:
254 control &= ~UNCORE_EFF_LAT_CTRL_RATIO_MASK;
255 control |= FIELD_PREP(UNCORE_EFF_LAT_CTRL_RATIO_MASK, val);
256 break;
257
258 default:
259 break;
260 }
261
262 writeq(control, cluster_info->cluster_base + UNCORE_CONTROL_INDEX);
263
264 return 0;
265 }
266
267 /* Helper function to write MMIO offset for max/min control frequency */
write_control_freq(struct tpmi_uncore_cluster_info * cluster_info,unsigned int input,unsigned int index)268 static void write_control_freq(struct tpmi_uncore_cluster_info *cluster_info, unsigned int input,
269 unsigned int index)
270 {
271 u64 control;
272
273 control = readq(cluster_info->cluster_base + UNCORE_CONTROL_INDEX);
274
275 if (index == UNCORE_INDEX_MAX_FREQ) {
276 control &= ~UNCORE_MAX_RATIO_MASK;
277 control |= FIELD_PREP(UNCORE_MAX_RATIO_MASK, input);
278 } else {
279 control &= ~UNCORE_MIN_RATIO_MASK;
280 control |= FIELD_PREP(UNCORE_MIN_RATIO_MASK, input);
281 }
282
283 writeq(control, (cluster_info->cluster_base + UNCORE_CONTROL_INDEX));
284 }
285
286 /* Helper for sysfs write for max/min frequencies. Called under mutex locks */
uncore_write_control_freq(struct uncore_data * data,unsigned int input,enum uncore_index index)287 static int uncore_write_control_freq(struct uncore_data *data, unsigned int input,
288 enum uncore_index index)
289 {
290 struct tpmi_uncore_cluster_info *cluster_info;
291 struct tpmi_uncore_struct *uncore_root;
292
293 input /= UNCORE_FREQ_KHZ_MULTIPLIER;
294 if (!input || input > UNCORE_MAX_RATIO)
295 return -EINVAL;
296
297 cluster_info = container_of(data, struct tpmi_uncore_cluster_info, uncore_data);
298 uncore_root = cluster_info->uncore_root;
299
300 if (uncore_root->write_blocked)
301 return -EPERM;
302
303 /* Update each cluster in a package */
304 if (cluster_info->root_domain) {
305 struct tpmi_uncore_struct *uncore_root = cluster_info->uncore_root;
306 int i;
307
308 for (i = 0; i < uncore_root->power_domain_count; ++i) {
309 int j;
310
311 for (j = 0; j < uncore_root->pd_info[i].cluster_count; ++j)
312 write_control_freq(&uncore_root->pd_info[i].cluster_infos[j],
313 input, index);
314 }
315
316 if (index == UNCORE_INDEX_MAX_FREQ)
317 uncore_root->max_ratio = input;
318 else
319 uncore_root->min_ratio = input;
320
321 return 0;
322 }
323
324 if (index == UNCORE_INDEX_MAX_FREQ && uncore_root->max_ratio &&
325 uncore_root->max_ratio < input)
326 return -EINVAL;
327
328 if (index == UNCORE_INDEX_MIN_FREQ && uncore_root->min_ratio &&
329 uncore_root->min_ratio > input)
330 return -EINVAL;
331
332 write_control_freq(cluster_info, input, index);
333
334 return 0;
335 }
336
337 /* Helper for sysfs read for the current uncore frequency. Called under mutex locks */
uncore_read_freq(struct uncore_data * data,unsigned int * freq)338 static int uncore_read_freq(struct uncore_data *data, unsigned int *freq)
339 {
340 struct tpmi_uncore_cluster_info *cluster_info;
341 u64 status;
342
343 cluster_info = container_of(data, struct tpmi_uncore_cluster_info, uncore_data);
344 if (cluster_info->root_domain)
345 return -ENODATA;
346
347 status = readq((u8 __iomem *)cluster_info->cluster_base + UNCORE_STATUS_INDEX);
348 *freq = FIELD_GET(UNCORE_CURRENT_RATIO_MASK, status) * UNCORE_FREQ_KHZ_MULTIPLIER;
349
350 return 0;
351 }
352
353 /*
354 * Agent types as per the TPMI UFS Specification for UFS_STATUS
355 * Agent Type - Core Bit: 23
356 * Agent Type - Cache Bit: 24
357 * Agent Type - Memory Bit: 25
358 * Agent Type - IO Bit: 26
359 */
360
361 #define UNCORE_AGENT_TYPES GENMASK_ULL(26, 23)
362
363 /* Helper function to read agent type over MMIO and set the agent type mask */
uncore_set_agent_type(struct tpmi_uncore_cluster_info * cluster_info)364 static void uncore_set_agent_type(struct tpmi_uncore_cluster_info *cluster_info)
365 {
366 u64 status;
367
368 status = readq((u8 __iomem *)cluster_info->cluster_base + UNCORE_STATUS_INDEX);
369 cluster_info->uncore_data.agent_type_mask = FIELD_GET(UNCORE_AGENT_TYPES, status);
370 }
371
372 /* Callback for sysfs read for TPMI uncore values. Called under mutex locks. */
uncore_read(struct uncore_data * data,unsigned int * value,enum uncore_index index)373 static int uncore_read(struct uncore_data *data, unsigned int *value, enum uncore_index index)
374 {
375 struct tpmi_uncore_cluster_info *cluster_info;
376 int ret;
377
378 switch (index) {
379 case UNCORE_INDEX_MIN_FREQ:
380 case UNCORE_INDEX_MAX_FREQ:
381 return uncore_read_control_freq(data, value, index);
382
383 case UNCORE_INDEX_CURRENT_FREQ:
384 return uncore_read_freq(data, value);
385
386 case UNCORE_INDEX_EFF_LAT_CTRL_LOW_THRESHOLD:
387 case UNCORE_INDEX_EFF_LAT_CTRL_HIGH_THRESHOLD:
388 case UNCORE_INDEX_EFF_LAT_CTRL_HIGH_THRESHOLD_ENABLE:
389 case UNCORE_INDEX_EFF_LAT_CTRL_FREQ:
390 return read_eff_lat_ctrl(data, value, index);
391
392 case UNCORE_INDEX_DIE_ID:
393 cluster_info = container_of(data, struct tpmi_uncore_cluster_info, uncore_data);
394 ret = tpmi_get_linux_die_id(cluster_info->uncore_data.package_id,
395 cluster_info->cdie_id);
396 if (ret < 0)
397 return ret;
398
399 *value = ret;
400 return 0;
401
402 default:
403 break;
404 }
405
406 return -EOPNOTSUPP;
407 }
408
409 /* Callback for sysfs write for TPMI uncore data. Called under mutex locks. */
uncore_write(struct uncore_data * data,unsigned int value,enum uncore_index index)410 static int uncore_write(struct uncore_data *data, unsigned int value, enum uncore_index index)
411 {
412 switch (index) {
413 case UNCORE_INDEX_EFF_LAT_CTRL_LOW_THRESHOLD:
414 case UNCORE_INDEX_EFF_LAT_CTRL_HIGH_THRESHOLD:
415 case UNCORE_INDEX_EFF_LAT_CTRL_HIGH_THRESHOLD_ENABLE:
416 case UNCORE_INDEX_EFF_LAT_CTRL_FREQ:
417 return write_eff_lat_ctrl(data, value, index);
418
419 case UNCORE_INDEX_MIN_FREQ:
420 case UNCORE_INDEX_MAX_FREQ:
421 return uncore_write_control_freq(data, value, index);
422
423 default:
424 break;
425 }
426
427 return -EOPNOTSUPP;
428 }
429
remove_cluster_entries(struct tpmi_uncore_struct * tpmi_uncore)430 static void remove_cluster_entries(struct tpmi_uncore_struct *tpmi_uncore)
431 {
432 int i;
433
434 for (i = 0; i < tpmi_uncore->power_domain_count; ++i) {
435 struct tpmi_uncore_power_domain_info *pd_info;
436 int j;
437
438 pd_info = &tpmi_uncore->pd_info[i];
439 if (!pd_info->uncore_base)
440 continue;
441
442 for (j = 0; j < pd_info->cluster_count; ++j) {
443 struct tpmi_uncore_cluster_info *cluster_info;
444
445 cluster_info = &pd_info->cluster_infos[j];
446 uncore_freq_remove_die_entry(&cluster_info->uncore_data);
447 }
448 }
449 }
450
set_cdie_id(int domain_id,struct tpmi_uncore_cluster_info * cluster_info,struct oobmsm_plat_info * plat_info)451 static void set_cdie_id(int domain_id, struct tpmi_uncore_cluster_info *cluster_info,
452 struct oobmsm_plat_info *plat_info)
453 {
454
455 cluster_info->cdie_id = domain_id;
456
457 if (plat_info->cdie_mask && cluster_info->uncore_data.agent_type_mask & AGENT_TYPE_CORE)
458 cluster_info->cdie_id = domain_id + ffs(plat_info->cdie_mask) - 1;
459 }
460
461 #define UNCORE_VERSION_MASK GENMASK_ULL(7, 0)
462 #define UNCORE_LOCAL_FABRIC_CLUSTER_ID_MASK GENMASK_ULL(15, 8)
463 #define UNCORE_CLUSTER_OFF_MASK GENMASK_ULL(7, 0)
464 #define UNCORE_MAX_CLUSTER_PER_DOMAIN 8
465
uncore_probe(struct auxiliary_device * auxdev,const struct auxiliary_device_id * id)466 static int uncore_probe(struct auxiliary_device *auxdev, const struct auxiliary_device_id *id)
467 {
468 bool read_blocked = 0, write_blocked = 0;
469 struct oobmsm_plat_info *plat_info;
470 struct tpmi_uncore_struct *tpmi_uncore;
471 bool uncore_sysfs_added = false;
472 int ret, i, pkg = 0;
473 int num_resources;
474
475 ret = tpmi_get_feature_status(auxdev, TPMI_ID_UNCORE, &read_blocked, &write_blocked);
476 if (ret)
477 dev_info(&auxdev->dev, "Can't read feature status: ignoring blocked status\n");
478
479 if (read_blocked) {
480 dev_info(&auxdev->dev, "Firmware has blocked reads, exiting\n");
481 return -ENODEV;
482 }
483
484 /* Get number of power domains, which is equal to number of resources */
485 num_resources = tpmi_get_resource_count(auxdev);
486 if (!num_resources)
487 return -EINVAL;
488
489 /* Register callbacks to uncore core */
490 ret = uncore_freq_common_init(uncore_read, uncore_write);
491 if (ret)
492 return ret;
493
494 /* Allocate uncore instance per package */
495 tpmi_uncore = devm_kzalloc(&auxdev->dev, sizeof(*tpmi_uncore), GFP_KERNEL);
496 if (!tpmi_uncore) {
497 ret = -ENOMEM;
498 goto err_rem_common;
499 }
500
501 /* Allocate memory for all power domains in a package */
502 tpmi_uncore->pd_info = devm_kcalloc(&auxdev->dev, num_resources,
503 sizeof(*tpmi_uncore->pd_info),
504 GFP_KERNEL);
505 if (!tpmi_uncore->pd_info) {
506 ret = -ENOMEM;
507 goto err_rem_common;
508 }
509
510 tpmi_uncore->power_domain_count = num_resources;
511 tpmi_uncore->write_blocked = write_blocked;
512
513 /* Get the package ID from the TPMI core */
514 plat_info = tpmi_get_platform_data(auxdev);
515 if (unlikely(!plat_info)) {
516 dev_info(&auxdev->dev, "Platform information is NULL\n");
517 ret = -ENODEV;
518 goto err_rem_common;
519 }
520
521 pkg = plat_info->package_id;
522
523 for (i = 0; i < num_resources; ++i) {
524 struct tpmi_uncore_power_domain_info *pd_info;
525 struct resource *res;
526 u64 cluster_offset;
527 u8 cluster_mask;
528 int mask, j;
529 u64 header;
530
531 res = tpmi_get_resource_at_index(auxdev, i);
532 if (!res)
533 continue;
534
535 pd_info = &tpmi_uncore->pd_info[i];
536
537 pd_info->uncore_base = devm_ioremap_resource(&auxdev->dev, res);
538 if (IS_ERR(pd_info->uncore_base)) {
539 ret = PTR_ERR(pd_info->uncore_base);
540 /*
541 * Set to NULL so that clean up can still remove other
542 * entries already created if any by
543 * remove_cluster_entries()
544 */
545 pd_info->uncore_base = NULL;
546 goto remove_clusters;
547 }
548
549 /* Check for version and skip this resource if there is mismatch */
550 header = readq(pd_info->uncore_base);
551 pd_info->ufs_header_ver = header & UNCORE_VERSION_MASK;
552
553 if (pd_info->ufs_header_ver == TPMI_VERSION_INVALID)
554 continue;
555
556 if (TPMI_MAJOR_VERSION(pd_info->ufs_header_ver) != UNCORE_MAJOR_VERSION) {
557 dev_err(&auxdev->dev, "Uncore: Unsupported major version:%lx\n",
558 TPMI_MAJOR_VERSION(pd_info->ufs_header_ver));
559 ret = -ENODEV;
560 goto remove_clusters;
561 }
562
563 if (TPMI_MINOR_VERSION(pd_info->ufs_header_ver) > UNCORE_MINOR_VERSION)
564 dev_info(&auxdev->dev, "Uncore: Ignore: Unsupported minor version:%lx\n",
565 TPMI_MINOR_VERSION(pd_info->ufs_header_ver));
566
567 /* Get Cluster ID Mask */
568 cluster_mask = FIELD_GET(UNCORE_LOCAL_FABRIC_CLUSTER_ID_MASK, header);
569 if (!cluster_mask) {
570 dev_info(&auxdev->dev, "Uncore: Invalid cluster mask:%x\n", cluster_mask);
571 continue;
572 }
573
574 /* Find out number of clusters in this resource */
575 pd_info->cluster_count = hweight8(cluster_mask);
576
577 pd_info->cluster_infos = devm_kcalloc(&auxdev->dev, pd_info->cluster_count,
578 sizeof(struct tpmi_uncore_cluster_info),
579 GFP_KERNEL);
580 if (!pd_info->cluster_infos) {
581 ret = -ENOMEM;
582 goto remove_clusters;
583 }
584 /*
585 * Each byte in the register point to status and control
586 * registers belonging to cluster id 0-8.
587 */
588 cluster_offset = readq(pd_info->uncore_base +
589 UNCORE_FABRIC_CLUSTER_OFFSET);
590
591 for (j = 0; j < pd_info->cluster_count; ++j) {
592 struct tpmi_uncore_cluster_info *cluster_info;
593
594 /* Get the offset for this cluster */
595 mask = (cluster_offset & UNCORE_CLUSTER_OFF_MASK);
596 /* Offset in QWORD, so change to bytes */
597 mask <<= 3;
598
599 cluster_info = &pd_info->cluster_infos[j];
600
601 cluster_info->cluster_base = pd_info->uncore_base + mask;
602
603 uncore_set_agent_type(cluster_info);
604
605 cluster_info->uncore_data.package_id = pkg;
606 /* There are no dies like Cascade Lake */
607 cluster_info->uncore_data.die_id = 0;
608 cluster_info->uncore_data.domain_id = i;
609 cluster_info->uncore_data.cluster_id = j;
610
611 set_cdie_id(i, cluster_info, plat_info);
612
613 cluster_info->uncore_root = tpmi_uncore;
614
615 if (TPMI_MINOR_VERSION(pd_info->ufs_header_ver) >= UNCORE_ELC_SUPPORTED_VERSION)
616 cluster_info->elc_supported = true;
617
618 ret = uncore_freq_add_entry(&cluster_info->uncore_data, 0);
619 if (ret) {
620 cluster_info->cluster_base = NULL;
621 goto remove_clusters;
622 }
623 /* Point to next cluster offset */
624 cluster_offset >>= UNCORE_MAX_CLUSTER_PER_DOMAIN;
625 uncore_sysfs_added = true;
626 }
627 }
628
629 if (!uncore_sysfs_added) {
630 ret = -ENODEV;
631 goto remove_clusters;
632 }
633
634 auxiliary_set_drvdata(auxdev, tpmi_uncore);
635
636 if (topology_max_dies_per_package() > 1)
637 return 0;
638
639 tpmi_uncore->root_cluster.root_domain = true;
640 tpmi_uncore->root_cluster.uncore_root = tpmi_uncore;
641
642 tpmi_uncore->root_cluster.uncore_data.package_id = pkg;
643 tpmi_uncore->root_cluster.uncore_data.domain_id = UNCORE_DOMAIN_ID_INVALID;
644 ret = uncore_freq_add_entry(&tpmi_uncore->root_cluster.uncore_data, 0);
645 if (ret)
646 goto remove_clusters;
647
648 return 0;
649
650 remove_clusters:
651 remove_cluster_entries(tpmi_uncore);
652 err_rem_common:
653 uncore_freq_common_exit();
654
655 return ret;
656 }
657
uncore_remove(struct auxiliary_device * auxdev)658 static void uncore_remove(struct auxiliary_device *auxdev)
659 {
660 struct tpmi_uncore_struct *tpmi_uncore = auxiliary_get_drvdata(auxdev);
661
662 if (tpmi_uncore->root_cluster.root_domain)
663 uncore_freq_remove_die_entry(&tpmi_uncore->root_cluster.uncore_data);
664
665 remove_cluster_entries(tpmi_uncore);
666
667 uncore_freq_common_exit();
668 }
669
670 static const struct auxiliary_device_id intel_uncore_id_table[] = {
671 { .name = "intel_vsec.tpmi-uncore" },
672 {}
673 };
674 MODULE_DEVICE_TABLE(auxiliary, intel_uncore_id_table);
675
676 static struct auxiliary_driver intel_uncore_aux_driver = {
677 .id_table = intel_uncore_id_table,
678 .remove = uncore_remove,
679 .probe = uncore_probe,
680 };
681
682 module_auxiliary_driver(intel_uncore_aux_driver);
683
684 MODULE_IMPORT_NS("INTEL_TPMI");
685 MODULE_IMPORT_NS("INTEL_UNCORE_FREQUENCY");
686 MODULE_IMPORT_NS("INTEL_TPMI_POWER_DOMAIN");
687 MODULE_DESCRIPTION("Intel TPMI UFS Driver");
688 MODULE_LICENSE("GPL");
689