Lines Matching +full:data +full:- +full:mirror
1 // SPDX-License-Identifier: GPL-2.0-only
21 * Interval to flush dirty data for next CTX entry. The interval is measured
37 if (eng_grp->g->engs_num < 0 || in get_cores_bmap()
38 eng_grp->g->engs_num > OTX2_CPT_MAX_ENGINES) { in get_cores_bmap()
40 eng_grp->g->engs_num); in get_cores_bmap()
45 if (eng_grp->engs[i].type) { in get_cores_bmap()
47 eng_grp->engs[i].bmap, in get_cores_bmap()
48 eng_grp->g->engs_num); in get_cores_bmap()
49 bmap.size = eng_grp->g->engs_num; in get_cores_bmap()
56 eng_grp->idx); in get_cores_bmap()
67 if (eng_grp->ucode[1].type) in is_2nd_ucode_used()
76 strscpy(ucode->filename, filename, OTX2_CPT_NAME_LENGTH); in set_ucode_filename()
132 strscpy(tmp_ver_str, ucode_hdr->ver_str, OTX2_CPT_UCODE_VER_STR_SZ); in get_ucode_type()
136 sprintf(ver_str_prefix, "ocpt-%02d", rid); in get_ucode_type()
138 return -EINVAL; in get_ucode_type()
140 nn = ucode_hdr->ver_num.nn; in get_ucode_type()
141 if (strnstr(tmp_ver_str, "se-", OTX2_CPT_UCODE_VER_STR_SZ) && in get_ucode_type()
145 if (strnstr(tmp_ver_str, "ie-", OTX2_CPT_UCODE_VER_STR_SZ) && in get_ucode_type()
156 return -EINVAL; in get_ucode_type()
164 return otx2_cpt_write_af_reg(&cptpf->afpf_mbox, cptpf->pdev, in __write_ucode_base()
177 ret = otx2_cpt_write_af_reg(&cptpf->afpf_mbox, cptpf->pdev, in cptx_set_ucode_base()
179 cptpf->pf_id << RVU_PFVF_PF_SHIFT, blkaddr); in cptx_set_ucode_base()
184 engs = &eng_grp->engs[i]; in cptx_set_ucode_base()
185 if (!engs->type) in cptx_set_ucode_base()
188 dma_addr = engs->ucode->dma; in cptx_set_ucode_base()
194 for_each_set_bit(bit, engs->bmap, eng_grp->g->engs_num) in cptx_set_ucode_base()
195 if (!eng_grp->g->eng_ref_cnt[bit]) { in cptx_set_ucode_base()
210 if (cptpf->has_cpt1) { in cpt_set_ucode_base()
229 ret = otx2_cpt_read_af_reg(&cptpf->afpf_mbox, cptpf->pdev, in cptx_detach_and_disable_cores()
234 if (reg & (1ull << eng_grp->idx)) { in cptx_detach_and_disable_cores()
235 eng_grp->g->eng_ref_cnt[i]--; in cptx_detach_and_disable_cores()
236 reg &= ~(1ull << eng_grp->idx); in cptx_detach_and_disable_cores()
238 ret = otx2_cpt_write_af_reg(&cptpf->afpf_mbox, in cptx_detach_and_disable_cores()
239 cptpf->pdev, in cptx_detach_and_disable_cores()
251 if (timeout-- < 0) in cptx_detach_and_disable_cores()
252 return -EBUSY; in cptx_detach_and_disable_cores()
255 ret = otx2_cpt_read_af_reg(&cptpf->afpf_mbox, in cptx_detach_and_disable_cores()
256 cptpf->pdev, in cptx_detach_and_disable_cores()
271 if (!eng_grp->g->eng_ref_cnt[i]) { in cptx_detach_and_disable_cores()
272 ret = otx2_cpt_write_af_reg(&cptpf->afpf_mbox, in cptx_detach_and_disable_cores()
273 cptpf->pdev, in cptx_detach_and_disable_cores()
291 bmap = get_cores_bmap(&cptpf->pdev->dev, eng_grp); in cpt_detach_and_disable_cores()
293 return -EINVAL; in cpt_detach_and_disable_cores()
295 if (cptpf->has_cpt1) { in cpt_detach_and_disable_cores()
315 ret = otx2_cpt_read_af_reg(&cptpf->afpf_mbox, cptpf->pdev, in cptx_attach_and_enable_cores()
320 if (!(reg & (1ull << eng_grp->idx))) { in cptx_attach_and_enable_cores()
321 eng_grp->g->eng_ref_cnt[i]++; in cptx_attach_and_enable_cores()
322 reg |= 1ull << eng_grp->idx; in cptx_attach_and_enable_cores()
324 ret = otx2_cpt_write_af_reg(&cptpf->afpf_mbox, in cptx_attach_and_enable_cores()
325 cptpf->pdev, in cptx_attach_and_enable_cores()
335 ret = otx2_cpt_add_write_af_reg(&cptpf->afpf_mbox, cptpf->pdev, in cptx_attach_and_enable_cores()
341 return otx2_cpt_send_af_reg_requests(&cptpf->afpf_mbox, cptpf->pdev); in cptx_attach_and_enable_cores()
351 bmap = get_cores_bmap(&cptpf->pdev->dev, eng_grp); in cpt_attach_and_enable_cores()
353 return -EINVAL; in cpt_attach_and_enable_cores()
355 if (cptpf->has_cpt1) { in cpt_attach_and_enable_cores()
374 return -ENOMEM; in load_fw()
376 ret = request_firmware(&uc_info->fw, filename, dev); in load_fw()
380 ucode_hdr = (struct otx2_cpt_ucode_hdr *)uc_info->fw->data; in load_fw()
385 ucode_size = ntohl(ucode_hdr->code_length) * 2; in load_fw()
388 ret = -EINVAL; in load_fw()
392 set_ucode_filename(&uc_info->ucode, filename); in load_fw()
393 memcpy(uc_info->ucode.ver_str, ucode_hdr->ver_str, in load_fw()
395 uc_info->ucode.ver_str[OTX2_CPT_UCODE_VER_STR_SZ] = 0; in load_fw()
396 uc_info->ucode.ver_num = ucode_hdr->ver_num; in load_fw()
397 uc_info->ucode.type = ucode_type; in load_fw()
398 uc_info->ucode.size = ucode_size; in load_fw()
399 list_add_tail(&uc_info->list, &fw_info->ucodes); in load_fw()
404 release_firmware(uc_info->fw); in load_fw()
417 list_for_each_entry_safe(curr, temp, &fw_info->ucodes, list) { in cpt_ucode_release_fw()
418 list_del(&curr->list); in cpt_ucode_release_fw()
419 release_firmware(curr->fw); in cpt_ucode_release_fw()
429 list_for_each_entry(curr, &fw_info->ucodes, list) { in get_ucode()
430 if (!is_eng_type(curr->ucode.type, ucode_type)) in get_ucode()
442 list_for_each_entry(curr, &fw_info->ucodes, list) { in print_uc_info()
443 pr_debug("Ucode filename %s\n", curr->ucode.filename); in print_uc_info()
444 pr_debug("Ucode version string %s\n", curr->ucode.ver_str); in print_uc_info()
446 curr->ucode.ver_num.nn, curr->ucode.ver_num.xx, in print_uc_info()
447 curr->ucode.ver_num.yy, curr->ucode.ver_num.zz); in print_uc_info()
448 pr_debug("Ucode type (%d) %s\n", curr->ucode.type, in print_uc_info()
449 get_ucode_type_str(curr->ucode.type)); in print_uc_info()
450 pr_debug("Ucode size %d\n", curr->ucode.size); in print_uc_info()
451 pr_debug("Ucode ptr %p\n", curr->fw->data); in print_uc_info()
462 INIT_LIST_HEAD(&fw_info->ucodes); in cpt_ucode_load_fw()
472 ret = load_fw(&pdev->dev, fw_info, filename, rid); in cpt_ucode_load_fw()
491 if (!eng_grp->engs[i].type) in find_engines_by_type()
494 if (eng_grp->engs[i].type == eng_type) in find_engines_by_type()
495 return &eng_grp->engs[i]; in find_engines_by_type()
514 switch (engs->type) { in update_engines_avail_count()
516 avail->se_cnt += val; in update_engines_avail_count()
520 avail->ie_cnt += val; in update_engines_avail_count()
524 avail->ae_cnt += val; in update_engines_avail_count()
528 dev_err(dev, "Invalid engine type %d\n", engs->type); in update_engines_avail_count()
529 return -EINVAL; in update_engines_avail_count()
538 switch (engs->type) { in update_engines_offset()
540 engs->offset = 0; in update_engines_offset()
544 engs->offset = avail->max_se_cnt; in update_engines_offset()
548 engs->offset = avail->max_se_cnt + avail->max_ie_cnt; in update_engines_offset()
552 dev_err(dev, "Invalid engine type %d\n", engs->type); in update_engines_offset()
553 return -EINVAL; in update_engines_offset()
564 if (!grp->engs[i].type) in release_engines()
567 if (grp->engs[i].count > 0) { in release_engines()
568 ret = update_engines_avail_count(dev, &grp->g->avail, in release_engines()
569 &grp->engs[i], in release_engines()
570 grp->engs[i].count); in release_engines()
575 grp->engs[i].type = 0; in release_engines()
576 grp->engs[i].count = 0; in release_engines()
577 grp->engs[i].offset = 0; in release_engines()
578 grp->engs[i].ucode = NULL; in release_engines()
579 bitmap_zero(grp->engs[i].bmap, grp->g->engs_num); in release_engines()
592 if (!grp->engs[i].type) { in do_reserve_engines()
593 engs = &grp->engs[i]; in do_reserve_engines()
599 return -ENOMEM; in do_reserve_engines()
601 engs->type = req_engs->type; in do_reserve_engines()
602 engs->count = req_engs->count; in do_reserve_engines()
604 ret = update_engines_offset(dev, &grp->g->avail, engs); in do_reserve_engines()
608 if (engs->count > 0) { in do_reserve_engines()
609 ret = update_engines_avail_count(dev, &grp->g->avail, engs, in do_reserve_engines()
610 -engs->count); in do_reserve_engines()
624 switch (req_eng->type) { in check_engines_availability()
626 avail_cnt = grp->g->avail.se_cnt; in check_engines_availability()
630 avail_cnt = grp->g->avail.ie_cnt; in check_engines_availability()
634 avail_cnt = grp->g->avail.ae_cnt; in check_engines_availability()
638 dev_err(dev, "Invalid engine type %d\n", req_eng->type); in check_engines_availability()
639 return -EINVAL; in check_engines_availability()
642 if (avail_cnt < req_eng->count) { in check_engines_availability()
645 get_eng_type_str(req_eng->type), in check_engines_availability()
646 avail_cnt, req_eng->count); in check_engines_availability()
647 return -EBUSY; in check_engines_availability()
676 if (ucode->va) { in ucode_unload()
677 dma_free_coherent(dev, OTX2_CPT_UCODE_SZ, ucode->va, in ucode_unload()
678 ucode->dma); in ucode_unload()
679 ucode->va = NULL; in ucode_unload()
680 ucode->dma = 0; in ucode_unload()
681 ucode->size = 0; in ucode_unload()
684 memset(&ucode->ver_str, 0, OTX2_CPT_UCODE_VER_STR_SZ); in ucode_unload()
685 memset(&ucode->ver_num, 0, sizeof(struct otx2_cpt_ucode_ver_num)); in ucode_unload()
687 ucode->type = 0; in ucode_unload()
697 ucode->va = dma_alloc_coherent(dev, OTX2_CPT_UCODE_SZ, &ucode->dma, in copy_ucode_to_dma_mem()
699 if (!ucode->va) in copy_ucode_to_dma_mem()
700 return -ENOMEM; in copy_ucode_to_dma_mem()
702 memcpy(ucode->va, ucode_data + sizeof(struct otx2_cpt_ucode_hdr), in copy_ucode_to_dma_mem()
703 ucode->size); in copy_ucode_to_dma_mem()
705 /* Byte swap 64-bit */ in copy_ucode_to_dma_mem()
706 for (i = 0; i < (ucode->size / 8); i++) in copy_ucode_to_dma_mem()
707 cpu_to_be64s(&((u64 *)ucode->va)[i]); in copy_ucode_to_dma_mem()
708 /* Ucode needs 16-bit swap */ in copy_ucode_to_dma_mem()
709 for (i = 0; i < (ucode->size / 2); i++) in copy_ucode_to_dma_mem()
710 cpu_to_be16s(&((u16 *)ucode->va)[i]); in copy_ucode_to_dma_mem()
742 ucode_unload(dev, &eng_grp->ucode[0]); in disable_eng_grp()
743 ucode_unload(dev, &eng_grp->ucode[1]); in disable_eng_grp()
746 if (!eng_grp->engs[i].type) in disable_eng_grp()
749 eng_grp->engs[i].ucode = &eng_grp->ucode[0]; in disable_eng_grp()
762 src_grp->mirror.is_ena = false; in setup_eng_grp_mirroring()
763 src_grp->mirror.idx = 0; in setup_eng_grp_mirroring()
764 src_grp->mirror.ref_count++; in setup_eng_grp_mirroring()
767 dst_grp->mirror.is_ena = true; in setup_eng_grp_mirroring()
768 dst_grp->mirror.idx = src_grp->idx; in setup_eng_grp_mirroring()
769 dst_grp->mirror.ref_count = 0; in setup_eng_grp_mirroring()
776 if (!dst_grp->mirror.is_ena) in remove_eng_grp_mirroring()
779 src_grp = &dst_grp->g->grp[dst_grp->mirror.idx]; in remove_eng_grp_mirroring()
781 src_grp->mirror.ref_count--; in remove_eng_grp_mirroring()
782 dst_grp->mirror.is_ena = false; in remove_eng_grp_mirroring()
783 dst_grp->mirror.idx = 0; in remove_eng_grp_mirroring()
784 dst_grp->mirror.ref_count = 0; in remove_eng_grp_mirroring()
813 engs[i].count -= mirrored_engs->count; in update_requested_engs()
820 struct otx2_cpt_eng_grps *eng_grps = grp->g; in find_mirrored_eng_grp()
824 if (!eng_grps->grp[i].is_enabled) in find_mirrored_eng_grp()
826 if (eng_grps->grp[i].ucode[0].type && in find_mirrored_eng_grp()
827 eng_grps->grp[i].ucode[1].type) in find_mirrored_eng_grp()
829 if (grp->idx == i) in find_mirrored_eng_grp()
831 if (!strncasecmp(eng_grps->grp[i].ucode[0].ver_str, in find_mirrored_eng_grp()
832 grp->ucode[0].ver_str, in find_mirrored_eng_grp()
834 return &eng_grps->grp[i]; in find_mirrored_eng_grp()
846 if (!eng_grps->grp[i].is_enabled) in find_unused_eng_grp()
847 return &eng_grps->grp[i]; in find_unused_eng_grp()
861 engs = &eng_grp->engs[i]; in eng_grp_update_masks()
862 if (!engs->type) in eng_grp_update_masks()
864 if (engs->count <= 0) in eng_grp_update_masks()
867 switch (engs->type) { in eng_grp_update_masks()
869 max_cnt = eng_grp->g->avail.max_se_cnt; in eng_grp_update_masks()
873 max_cnt = eng_grp->g->avail.max_ie_cnt; in eng_grp_update_masks()
877 max_cnt = eng_grp->g->avail.max_ae_cnt; in eng_grp_update_masks()
881 dev_err(dev, "Invalid engine type %d\n", engs->type); in eng_grp_update_masks()
882 return -EINVAL; in eng_grp_update_masks()
885 cnt = engs->count; in eng_grp_update_masks()
886 WARN_ON(engs->offset + max_cnt > OTX2_CPT_MAX_ENGINES); in eng_grp_update_masks()
887 bitmap_zero(tmp_bmap.bits, eng_grp->g->engs_num); in eng_grp_update_masks()
888 for (j = engs->offset; j < engs->offset + max_cnt; j++) { in eng_grp_update_masks()
889 if (!eng_grp->g->eng_ref_cnt[j]) { in eng_grp_update_masks()
891 cnt--; in eng_grp_update_masks()
898 return -ENOSPC; in eng_grp_update_masks()
900 bitmap_copy(engs->bmap, tmp_bmap.bits, eng_grp->g->engs_num); in eng_grp_update_masks()
903 if (!eng_grp->mirror.is_ena) in eng_grp_update_masks()
907 engs = &eng_grp->engs[i]; in eng_grp_update_masks()
908 if (!engs->type) in eng_grp_update_masks()
912 &eng_grp->g->grp[eng_grp->mirror.idx], in eng_grp_update_masks()
913 engs->type); in eng_grp_update_masks()
914 WARN_ON(!mirrored_engs && engs->count <= 0); in eng_grp_update_masks()
918 bitmap_copy(tmp_bmap.bits, mirrored_engs->bmap, in eng_grp_update_masks()
919 eng_grp->g->engs_num); in eng_grp_update_masks()
920 if (engs->count < 0) { in eng_grp_update_masks()
921 bit = find_first_bit(mirrored_engs->bmap, in eng_grp_update_masks()
922 eng_grp->g->engs_num); in eng_grp_update_masks()
923 bitmap_clear(tmp_bmap.bits, bit, -engs->count); in eng_grp_update_masks()
925 bitmap_or(engs->bmap, engs->bmap, tmp_bmap.bits, in eng_grp_update_masks()
926 eng_grp->g->engs_num); in eng_grp_update_masks()
936 if (!eng_grp->is_enabled) in delete_engine_group()
939 if (eng_grp->mirror.ref_count) in delete_engine_group()
940 return -EINVAL; in delete_engine_group()
946 ret = disable_eng_grp(dev, eng_grp, eng_grp->g->obj); in delete_engine_group()
955 eng_grp->is_enabled = false; in delete_engine_group()
964 if (eng_grp->mirror.is_ena) in update_ucode_ptrs()
965 ucode = &eng_grp->g->grp[eng_grp->mirror.idx].ucode[0]; in update_ucode_ptrs()
967 ucode = &eng_grp->ucode[0]; in update_ucode_ptrs()
968 WARN_ON(!eng_grp->engs[0].type); in update_ucode_ptrs()
969 eng_grp->engs[0].ucode = ucode; in update_ucode_ptrs()
971 if (eng_grp->engs[1].type) { in update_ucode_ptrs()
973 eng_grp->engs[1].ucode = &eng_grp->ucode[1]; in update_ucode_ptrs()
975 eng_grp->engs[1].ucode = ucode; in update_ucode_ptrs()
993 return -ENOSPC; in create_engine_group()
998 eng_grp->ucode[i] = uc_info->ucode; in create_engine_group()
999 ret = copy_ucode_to_dma_mem(dev, &eng_grp->ucode[i], in create_engine_group()
1000 uc_info->fw->data); in create_engine_group()
1030 ret = enable_eng_grp(eng_grp, eng_grps->obj); in create_engine_group()
1039 if (eng_grp->mirror.is_ena) in create_engine_group()
1040 ucode_unload(dev, &eng_grp->ucode[0]); in create_engine_group()
1042 eng_grp->is_enabled = true; in create_engine_group()
1050 eng_grp->idx, mirrored_eng_grp->ucode[0].ver_str, in create_engine_group()
1051 mirrored_eng_grp->idx); in create_engine_group()
1054 eng_grp->idx, eng_grp->ucode[0].ver_str); in create_engine_group()
1057 eng_grp->idx, eng_grp->ucode[1].ver_str); in create_engine_group()
1064 ucode_unload(dev, &eng_grp->ucode[0]); in create_engine_group()
1065 ucode_unload(dev, &eng_grp->ucode[1]); in create_engine_group()
1076 if (eng_grps->grp[i].mirror.is_ena) in delete_engine_grps()
1077 delete_engine_group(&pdev->dev, &eng_grps->grp[i]); in delete_engine_grps()
1081 delete_engine_group(&pdev->dev, &eng_grps->grp[i]); in delete_engine_grps()
1104 timeout--; in rnm_to_cpt_errata_fixup()
1125 grp = &eng_grps->grp[i]; in otx2_cpt_get_eng_grp()
1126 if (!grp->is_enabled) in otx2_cpt_get_eng_grp()
1150 struct pci_dev *pdev = cptpf->pdev; in otx2_cpt_create_eng_grps()
1155 mutex_lock(&eng_grps->lock); in otx2_cpt_create_eng_grps()
1160 if (eng_grps->is_grps_created) in otx2_cpt_create_eng_grps()
1163 ret = cpt_ucode_load_fw(pdev, &fw_info, eng_grps->rid); in otx2_cpt_create_eng_grps()
1173 dev_err(&pdev->dev, "Unable to find firmware for SE\n"); in otx2_cpt_create_eng_grps()
1174 ret = -EINVAL; in otx2_cpt_create_eng_grps()
1178 engs[0].count = eng_grps->avail.max_se_cnt; in otx2_cpt_create_eng_grps()
1180 ret = create_engine_group(&pdev->dev, eng_grps, engs, 1, in otx2_cpt_create_eng_grps()
1193 dev_err(&pdev->dev, "Unable to find firmware for IE"); in otx2_cpt_create_eng_grps()
1194 ret = -EINVAL; in otx2_cpt_create_eng_grps()
1198 engs[0].count = eng_grps->avail.max_se_cnt; in otx2_cpt_create_eng_grps()
1200 engs[1].count = eng_grps->avail.max_ie_cnt; in otx2_cpt_create_eng_grps()
1202 ret = create_engine_group(&pdev->dev, eng_grps, engs, 2, in otx2_cpt_create_eng_grps()
1213 dev_err(&pdev->dev, "Unable to find firmware for AE"); in otx2_cpt_create_eng_grps()
1214 ret = -EINVAL; in otx2_cpt_create_eng_grps()
1218 engs[0].count = eng_grps->avail.max_ae_cnt; in otx2_cpt_create_eng_grps()
1220 ret = create_engine_group(&pdev->dev, eng_grps, engs, 1, in otx2_cpt_create_eng_grps()
1225 eng_grps->is_grps_created = true; in otx2_cpt_create_eng_grps()
1236 rnm_to_cpt_errata_fixup(&pdev->dev); in otx2_cpt_create_eng_grps()
1238 otx2_cpt_read_af_reg(&cptpf->afpf_mbox, pdev, CPT_AF_CTL, ®_val, in otx2_cpt_create_eng_grps()
1246 otx2_cpt_write_af_reg(&cptpf->afpf_mbox, pdev, CPT_AF_CTL, in otx2_cpt_create_eng_grps()
1249 * Set interval to periodically flush dirty data for the next in otx2_cpt_create_eng_grps()
1253 otx2_cpt_write_af_reg(&cptpf->afpf_mbox, pdev, CPT_AF_CTX_FLUSH_TIMER, in otx2_cpt_create_eng_grps()
1260 * unpredictable data being delivered to a CPT engine. in otx2_cpt_create_eng_grps()
1263 otx2_cpt_read_af_reg(&cptpf->afpf_mbox, pdev, CPT_AF_DIAG, in otx2_cpt_create_eng_grps()
1265 otx2_cpt_write_af_reg(&cptpf->afpf_mbox, pdev, CPT_AF_DIAG, in otx2_cpt_create_eng_grps()
1269 mutex_unlock(&eng_grps->lock); in otx2_cpt_create_eng_grps()
1277 mutex_unlock(&eng_grps->lock); in otx2_cpt_create_eng_grps()
1290 ret = otx2_cpt_add_write_af_reg(&cptpf->afpf_mbox, cptpf->pdev, in cptx_disable_all_cores()
1296 cptpf->eng_grps.eng_ref_cnt[i] = 0; in cptx_disable_all_cores()
1298 ret = otx2_cpt_send_af_reg_requests(&cptpf->afpf_mbox, cptpf->pdev); in cptx_disable_all_cores()
1306 if (timeout-- < 0) in cptx_disable_all_cores()
1307 return -EBUSY; in cptx_disable_all_cores()
1310 ret = otx2_cpt_read_af_reg(&cptpf->afpf_mbox, in cptx_disable_all_cores()
1311 cptpf->pdev, in cptx_disable_all_cores()
1326 ret = otx2_cpt_add_write_af_reg(&cptpf->afpf_mbox, cptpf->pdev, in cptx_disable_all_cores()
1332 return otx2_cpt_send_af_reg_requests(&cptpf->afpf_mbox, cptpf->pdev); in cptx_disable_all_cores()
1339 total_cores = cptpf->eng_grps.avail.max_se_cnt + in otx2_cpt_disable_all_cores()
1340 cptpf->eng_grps.avail.max_ie_cnt + in otx2_cpt_disable_all_cores()
1341 cptpf->eng_grps.avail.max_ae_cnt; in otx2_cpt_disable_all_cores()
1343 if (cptpf->has_cpt1) { in otx2_cpt_disable_all_cores()
1357 mutex_lock(&eng_grps->lock); in otx2_cpt_cleanup_eng_grps()
1361 grp = &eng_grps->grp[i]; in otx2_cpt_cleanup_eng_grps()
1363 kfree(grp->engs[j].bmap); in otx2_cpt_cleanup_eng_grps()
1364 grp->engs[j].bmap = NULL; in otx2_cpt_cleanup_eng_grps()
1367 mutex_unlock(&eng_grps->lock); in otx2_cpt_cleanup_eng_grps()
1376 mutex_init(&eng_grps->lock); in otx2_cpt_init_eng_grps()
1377 eng_grps->obj = pci_get_drvdata(pdev); in otx2_cpt_init_eng_grps()
1378 eng_grps->avail.se_cnt = eng_grps->avail.max_se_cnt; in otx2_cpt_init_eng_grps()
1379 eng_grps->avail.ie_cnt = eng_grps->avail.max_ie_cnt; in otx2_cpt_init_eng_grps()
1380 eng_grps->avail.ae_cnt = eng_grps->avail.max_ae_cnt; in otx2_cpt_init_eng_grps()
1382 eng_grps->engs_num = eng_grps->avail.max_se_cnt + in otx2_cpt_init_eng_grps()
1383 eng_grps->avail.max_ie_cnt + in otx2_cpt_init_eng_grps()
1384 eng_grps->avail.max_ae_cnt; in otx2_cpt_init_eng_grps()
1385 if (eng_grps->engs_num > OTX2_CPT_MAX_ENGINES) { in otx2_cpt_init_eng_grps()
1386 dev_err(&pdev->dev, in otx2_cpt_init_eng_grps()
1388 eng_grps->engs_num, OTX2_CPT_MAX_ENGINES); in otx2_cpt_init_eng_grps()
1389 ret = -EINVAL; in otx2_cpt_init_eng_grps()
1394 grp = &eng_grps->grp[i]; in otx2_cpt_init_eng_grps()
1395 grp->g = eng_grps; in otx2_cpt_init_eng_grps()
1396 grp->idx = i; in otx2_cpt_init_eng_grps()
1399 grp->engs[j].bmap = in otx2_cpt_init_eng_grps()
1400 kcalloc(BITS_TO_LONGS(eng_grps->engs_num), in otx2_cpt_init_eng_grps()
1402 if (!grp->engs[j].bmap) { in otx2_cpt_init_eng_grps()
1403 ret = -ENOMEM; in otx2_cpt_init_eng_grps()
1423 mutex_lock(&eng_grps->lock); in create_eng_caps_discovery_grps()
1424 ret = cpt_ucode_load_fw(pdev, &fw_info, eng_grps->rid); in create_eng_caps_discovery_grps()
1426 mutex_unlock(&eng_grps->lock); in create_eng_caps_discovery_grps()
1432 dev_err(&pdev->dev, "Unable to find firmware for AE\n"); in create_eng_caps_discovery_grps()
1433 ret = -EINVAL; in create_eng_caps_discovery_grps()
1439 ret = create_engine_group(&pdev->dev, eng_grps, engs, 1, in create_eng_caps_discovery_grps()
1446 dev_err(&pdev->dev, "Unable to find firmware for SE\n"); in create_eng_caps_discovery_grps()
1447 ret = -EINVAL; in create_eng_caps_discovery_grps()
1453 ret = create_engine_group(&pdev->dev, eng_grps, engs, 1, in create_eng_caps_discovery_grps()
1460 dev_err(&pdev->dev, "Unable to find firmware for IE\n"); in create_eng_caps_discovery_grps()
1461 ret = -EINVAL; in create_eng_caps_discovery_grps()
1467 ret = create_engine_group(&pdev->dev, eng_grps, engs, 1, in create_eng_caps_discovery_grps()
1473 mutex_unlock(&eng_grps->lock); in create_eng_caps_discovery_grps()
1480 mutex_unlock(&eng_grps->lock); in create_eng_caps_discovery_grps()
1489 struct otx2_cptlfs_info *lfs = &cptpf->lfs; in otx2_cpt_discover_eng_capabilities()
1504 if (cptpf->is_eng_caps_discovered) in otx2_cpt_discover_eng_capabilities()
1507 pdev = cptpf->pdev; in otx2_cpt_discover_eng_capabilities()
1512 ret = create_eng_caps_discovery_grps(pdev, &cptpf->eng_grps); in otx2_cpt_discover_eng_capabilities()
1516 otx2_cptlf_set_dev_info(lfs, cptpf->pdev, cptpf->reg_base, in otx2_cpt_discover_eng_capabilities()
1517 &cptpf->afpf_mbox, BLKADDR_CPT0); in otx2_cpt_discover_eng_capabilities()
1528 ret = -ENOMEM; in otx2_cpt_discover_eng_capabilities()
1531 rptr_baddr = dma_map_single(&pdev->dev, (void *)result, len, in otx2_cpt_discover_eng_capabilities()
1533 if (dma_mapping_error(&pdev->dev, rptr_baddr)) { in otx2_cpt_discover_eng_capabilities()
1534 dev_err(&pdev->dev, "DMA mapping failed\n"); in otx2_cpt_discover_eng_capabilities()
1535 ret = -EFAULT; in otx2_cpt_discover_eng_capabilities()
1547 /* 64-bit swap for microcode data reads, not needed for addresses */ in otx2_cpt_discover_eng_capabilities()
1554 result->s.compcode = OTX2_CPT_COMPLETION_CODE_INIT; in otx2_cpt_discover_eng_capabilities()
1555 iq_cmd.cptr.s.grp = otx2_cpt_get_eng_grp(&cptpf->eng_grps, in otx2_cpt_discover_eng_capabilities()
1558 lfs->ops->send_cmd(&inst, 1, &cptpf->lfs.lf[0]); in otx2_cpt_discover_eng_capabilities()
1560 while (lfs->ops->cpt_get_compcode(result) == in otx2_cpt_discover_eng_capabilities()
1564 cptpf->eng_caps[etype].u = be64_to_cpup(rptr); in otx2_cpt_discover_eng_capabilities()
1566 dma_unmap_single(&pdev->dev, rptr_baddr, len, DMA_BIDIRECTIONAL); in otx2_cpt_discover_eng_capabilities()
1567 cptpf->is_eng_caps_discovered = true; in otx2_cpt_discover_eng_capabilities()
1574 delete_engine_grps(pdev, &cptpf->eng_grps); in otx2_cpt_discover_eng_capabilities()
1584 struct otx2_cpt_eng_grps *eng_grps = &cptpf->eng_grps; in otx2_cpt_dl_custom_egrp_create()
1587 struct device *dev = &cptpf->pdev->dev; in otx2_cpt_dl_custom_egrp_create()
1589 int grp_idx = 0, ret = -EINVAL; in otx2_cpt_dl_custom_egrp_create()
1594 if (!eng_grps->is_grps_created) { in otx2_cpt_dl_custom_egrp_create()
1596 return -EINVAL; in otx2_cpt_dl_custom_egrp_create()
1599 strscpy(tmp_buf, ctx->val.vstr, strlen(ctx->val.vstr) + 1); in otx2_cpt_dl_custom_egrp_create()
1690 mutex_lock(&eng_grps->lock); in otx2_cpt_dl_custom_egrp_create()
1692 if (cptpf->enabled_vfs) { in otx2_cpt_dl_custom_egrp_create()
1694 ret = -EACCES; in otx2_cpt_dl_custom_egrp_create()
1699 ret = load_fw(dev, &fw_info, ucode_filename[0], eng_grps->rid); in otx2_cpt_dl_custom_egrp_create()
1705 ret = load_fw(dev, &fw_info, ucode_filename[1], eng_grps->rid); in otx2_cpt_dl_custom_egrp_create()
1716 ret = -EINVAL; in otx2_cpt_dl_custom_egrp_create()
1724 ret = -EINVAL; in otx2_cpt_dl_custom_egrp_create()
1734 mutex_unlock(&eng_grps->lock); in otx2_cpt_dl_custom_egrp_create()
1744 struct otx2_cpt_eng_grps *eng_grps = &cptpf->eng_grps; in otx2_cpt_dl_custom_egrp_delete()
1745 struct device *dev = &cptpf->pdev->dev; in otx2_cpt_dl_custom_egrp_delete()
1751 if (strncasecmp(ctx->val.vstr, "egrp", 4)) in otx2_cpt_dl_custom_egrp_delete()
1753 tmp = ctx->val.vstr; in otx2_cpt_dl_custom_egrp_delete()
1762 return -EINVAL; in otx2_cpt_dl_custom_egrp_delete()
1764 if (!eng_grps->grp[egrp].is_enabled) { in otx2_cpt_dl_custom_egrp_delete()
1766 return -EINVAL; in otx2_cpt_dl_custom_egrp_delete()
1768 mutex_lock(&eng_grps->lock); in otx2_cpt_dl_custom_egrp_delete()
1769 ret = delete_engine_group(dev, &eng_grps->grp[egrp]); in otx2_cpt_dl_custom_egrp_delete()
1770 mutex_unlock(&eng_grps->lock); in otx2_cpt_dl_custom_egrp_delete()
1776 return -EINVAL; in otx2_cpt_dl_custom_egrp_delete()