Lines Matching refs:alias_guid

85 	guid_indexes = be64_to_cpu((__force __be64) dev->sriov.alias_guid.
135 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags);
136 if (dev->sriov.alias_guid.ports_guid[port_index].state_flags &
141 alias_guid.ports_guid[port_index].
153 *(__be64 *)&dev->sriov.alias_guid.ports_guid[port_index].
156 dev->sriov.alias_guid.ports_guid[port_index].
159 dev->sriov.alias_guid.ports_guid[port_index].
163 dev->sriov.alias_guid.ports_guid[port_index].
165 dev->sriov.alias_guid.ports_guid[port_index].
170 spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags);
203 rec = &dev->sriov.alias_guid.ports_guid[port_num - 1].
205 guid_indexes = be64_to_cpu((__force __be64) dev->sriov.alias_guid.
235 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags);
248 alias_guid.ag_work_lock, flags);
252 spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock,
306 rec = &dev->sriov.alias_guid.ports_guid[port_index].
326 rec = &dev->sriov.alias_guid.ports_guid[port_index].
329 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags);
423 spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags);
434 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
437 queue_delayed_work(dev->sriov.alias_guid.ports_guid[port_index].wq,
438 &dev->sriov.alias_guid.ports_guid[port_index].
447 spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
457 dev->sriov.alias_guid.ports_guid[port - 1].all_rec_per_port[index].status
463 *(u64 *)&dev->sriov.alias_guid.ports_guid[port - 1].
476 dev->sriov.alias_guid.ports_guid[port - 1].
478 if (dev->sriov.alias_guid.ports_guid[port - 1].
480 dev->sriov.alias_guid.ports_guid[port - 1].
499 &dev->sriov.alias_guid.ports_guid[port - 1].cb_list;
539 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
541 spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
544 ib_sa_guid_info_rec_query(dev->sriov.alias_guid.sa_client,
554 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
557 spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
567 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
570 queue_delayed_work(dev->sriov.alias_guid.ports_guid[port - 1].wq,
571 &dev->sriov.alias_guid.ports_guid[port - 1].alias_guid_work,
574 spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
595 *(__be64 *)&dev->sriov.alias_guid.ports_guid[port - 1].
613 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
615 if (dev->sriov.alias_guid.ports_guid[port - 1].state_flags &
618 dev->sriov.alias_guid.ports_guid[port - 1].state_flags &=
630 cancel_delayed_work(&dev->sriov.alias_guid.
632 queue_delayed_work(dev->sriov.alias_guid.ports_guid[port - 1].wq,
633 &dev->sriov.alias_guid.ports_guid[port - 1].alias_guid_work,
636 spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
650 &dev->sriov.alias_guid.ports_guid[port].
699 rec = dev->sriov.alias_guid.ports_guid[port].
729 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags);
739 spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags);
754 alias_guid);
781 spin_lock_irqsave(&dev->sriov.alias_guid.ag_work_lock, flags1);
787 cancel_delayed_work(&dev->sriov.alias_guid.ports_guid[port].
789 queue_delayed_work(dev->sriov.alias_guid.ports_guid[port].wq,
790 &dev->sriov.alias_guid.ports_guid[port].alias_guid_work, 0);
792 spin_unlock_irqrestore(&dev->sriov.alias_guid.ag_work_lock, flags1);
806 det = &sriov->alias_guid.ports_guid[i];
808 spin_lock_irqsave(&sriov->alias_guid.ag_work_lock, flags);
816 spin_unlock_irqrestore(&sriov->alias_guid.ag_work_lock, flags);
820 spin_lock_irqsave(&sriov->alias_guid.ag_work_lock, flags);
822 spin_unlock_irqrestore(&sriov->alias_guid.ag_work_lock, flags);
825 destroy_workqueue(dev->sriov.alias_guid.ports_guid[i].wq);
826 ib_sa_unregister_client(dev->sriov.alias_guid.sa_client);
827 kfree(dev->sriov.alias_guid.sa_client);
838 dev->sriov.alias_guid.sa_client =
839 kzalloc(sizeof *dev->sriov.alias_guid.sa_client, GFP_KERNEL);
840 if (!dev->sriov.alias_guid.sa_client)
843 ib_sa_register_client(dev->sriov.alias_guid.sa_client);
845 spin_lock_init(&dev->sriov.alias_guid.ag_work_lock);
855 memset(&dev->sriov.alias_guid.ports_guid[i], 0,
857 dev->sriov.alias_guid.ports_guid[i].state_flags |=
861 memset(dev->sriov.alias_guid.ports_guid[i].
863 sizeof(dev->sriov.alias_guid.ports_guid[i].
866 INIT_LIST_HEAD(&dev->sriov.alias_guid.ports_guid[i].cb_list);
874 dev->sriov.alias_guid.ports_guid[i].parent = &dev->sriov.alias_guid;
875 dev->sriov.alias_guid.ports_guid[i].port = i;
877 dev->sriov.alias_guid.ports_guid[i].wq =
878 alloc_ordered_workqueue("alias_guid%d", WQ_MEM_RECLAIM, i);
879 if (!dev->sriov.alias_guid.ports_guid[i].wq) {
883 INIT_DELAYED_WORK(&dev->sriov.alias_guid.ports_guid[i].alias_guid_work,
890 destroy_workqueue(dev->sriov.alias_guid.ports_guid[i].wq);
891 dev->sriov.alias_guid.ports_guid[i].wq = NULL;
895 ib_sa_unregister_client(dev->sriov.alias_guid.sa_client);
896 kfree(dev->sriov.alias_guid.sa_client);
897 dev->sriov.alias_guid.sa_client = NULL;