Lines Matching refs:lynx

85 lynx_get(struct pcilynx *lynx)
87 kref_get(&lynx->kref);
89 return lynx;
99 lynx_put(struct pcilynx *lynx)
101 kref_put(&lynx->kref, lynx_release);
105 struct pcilynx *lynx;
145 list_empty(&client->lynx->link))
216 reg_write(struct pcilynx *lynx, int offset, u32 data)
218 writel(data, lynx->registers + offset);
222 reg_read(struct pcilynx *lynx, int offset)
224 return readl(lynx->registers + offset);
228 reg_set_bits(struct pcilynx *lynx, int offset, u32 mask)
230 reg_write(lynx, offset, (reg_read(lynx, offset) | mask));
238 run_pcl(struct pcilynx *lynx, dma_addr_t pcl_bus,
241 reg_write(lynx, DMA0_CURRENT_PCL + dmachan * 0x20, pcl_bus);
242 reg_write(lynx, DMA0_CHAN_CTRL + dmachan * 0x20,
247 set_phy_reg(struct pcilynx *lynx, int addr, int val)
250 dev_err(&lynx->pci_device->dev,
255 dev_err(&lynx->pci_device->dev,
259 reg_write(lynx, LINK_PHY, LINK_PHY_WRITE |
270 struct pcilynx *tmp, *lynx = NULL;
275 lynx = lynx_get(tmp);
279 if (lynx == NULL)
287 client->lynx = lynx;
298 lynx_put(lynx);
307 struct pcilynx *lynx = client->lynx;
309 spin_lock_irq(&lynx->client_list_lock);
311 spin_unlock_irq(&lynx->client_list_lock);
315 lynx_put(lynx);
331 if (list_empty(&client->lynx->link))
349 spinlock_t *client_list_lock = &client->lynx->client_list_lock;
369 list_add_tail(&client->link, &client->lynx->client_list);
408 packet_irq_handler(struct pcilynx *lynx)
417 length = __le32_to_cpu(lynx->rcv_pcl->pcl_status) & 0x00001fff;
418 tcode = __le32_to_cpu(lynx->rcv_buffer[1]) >> 4 & 0xf;
422 lynx->rcv_buffer[0] = (__force __le32)timestamp;
429 spin_lock(&lynx->client_list_lock);
431 list_for_each_entry(client, &lynx->client_list, link)
434 lynx->rcv_buffer, length + 4);
436 spin_unlock(&lynx->client_list_lock);
440 bus_reset_irq_handler(struct pcilynx *lynx)
449 spin_lock(&lynx->client_list_lock);
451 list_for_each_entry(client, &lynx->client_list, link)
454 spin_unlock(&lynx->client_list_lock);
460 struct pcilynx *lynx = device;
463 pci_int_status = reg_read(lynx, PCI_INT_STATUS);
476 link_int_status = reg_read(lynx, LINK_INT_STATUS);
477 reg_write(lynx, LINK_INT_STATUS, link_int_status);
480 bus_reset_irq_handler(lynx);
487 reg_write(lynx, PCI_INT_STATUS, pci_int_status);
490 packet_irq_handler(lynx);
491 run_pcl(lynx, lynx->rcv_start_pcl_bus, 0);
500 struct pcilynx *lynx = pci_get_drvdata(dev);
504 list_del_init(&lynx->link);
505 misc_deregister(&lynx->misc);
508 reg_write(lynx, PCI_INT_ENABLE, 0);
509 free_irq(lynx->pci_device->irq, lynx);
511 spin_lock_irq(&lynx->client_list_lock);
512 list_for_each_entry(client, &lynx->client_list, link)
514 spin_unlock_irq(&lynx->client_list_lock);
516 dma_free_coherent(&lynx->pci_device->dev, sizeof(struct pcl),
517 lynx->rcv_start_pcl, lynx->rcv_start_pcl_bus);
518 dma_free_coherent(&lynx->pci_device->dev, sizeof(struct pcl),
519 lynx->rcv_pcl, lynx->rcv_pcl_bus);
520 dma_free_coherent(&lynx->pci_device->dev, PAGE_SIZE, lynx->rcv_buffer,
521 lynx->rcv_buffer_bus);
523 iounmap(lynx->registers);
525 lynx_put(lynx);
533 struct pcilynx *lynx;
548 lynx = kzalloc(sizeof *lynx, GFP_KERNEL);
549 if (lynx == NULL) {
554 lynx->pci_device = dev;
555 pci_set_drvdata(dev, lynx);
557 spin_lock_init(&lynx->client_list_lock);
558 INIT_LIST_HEAD(&lynx->client_list);
559 kref_init(&lynx->kref);
561 lynx->registers = ioremap(pci_resource_start(dev, 0),
563 if (lynx->registers == NULL) {
569 lynx->rcv_start_pcl = dma_alloc_coherent(&lynx->pci_device->dev,
571 &lynx->rcv_start_pcl_bus,
573 lynx->rcv_pcl = dma_alloc_coherent(&lynx->pci_device->dev,
575 &lynx->rcv_pcl_bus, GFP_KERNEL);
576 lynx->rcv_buffer = dma_alloc_coherent(&lynx->pci_device->dev,
578 &lynx->rcv_buffer_bus, GFP_KERNEL);
579 if (lynx->rcv_start_pcl == NULL ||
580 lynx->rcv_pcl == NULL ||
581 lynx->rcv_buffer == NULL) {
586 lynx->rcv_start_pcl->next = cpu_to_le32(lynx->rcv_pcl_bus);
587 lynx->rcv_pcl->next = cpu_to_le32(PCL_NEXT_INVALID);
588 lynx->rcv_pcl->async_error_next = cpu_to_le32(PCL_NEXT_INVALID);
590 lynx->rcv_pcl->buffer[0].control =
592 lynx->rcv_pcl->buffer[0].pointer =
593 cpu_to_le32(lynx->rcv_buffer_bus + 4);
594 p = lynx->rcv_buffer_bus + 2048;
595 end = lynx->rcv_buffer_bus + RCV_BUFFER_SIZE;
597 lynx->rcv_pcl->buffer[i].control =
599 lynx->rcv_pcl->buffer[i].pointer = cpu_to_le32(p);
601 lynx->rcv_pcl->buffer[i - 1].control |= cpu_to_le32(PCL_LAST_BUFF);
603 reg_set_bits(lynx, MISC_CONTROL, MISC_CONTROL_SWRESET);
605 reg_write(lynx, DMA0_CHAN_CTRL, 0);
606 reg_write(lynx, DMA_GLOBAL_REGISTER, 0x00 << 24);
610 if ((get_phy_reg(lynx, 2) & 0xe0) == 0xe0) {
611 lynx->phyic.reg_1394a = 1;
612 PRINT(KERN_INFO, lynx->id,
614 lynx->phyic.vendor = get_phy_vendorid(lynx);
615 lynx->phyic.product = get_phy_productid(lynx);
617 lynx->phyic.reg_1394a = 0;
618 PRINT(KERN_INFO, lynx->id, "found old 1394 PHY");
623 reg_write(lynx, FIFO_SIZES, 255);
625 reg_set_bits(lynx, PCI_INT_ENABLE, PCI_INT_DMA_ALL);
627 reg_write(lynx, LINK_INT_ENABLE,
635 set_phy_reg(lynx, 4, 0);
638 reg_set_bits(lynx, LINK_CONTROL, LINK_CONTROL_SNOOP_ENABLE);
640 run_pcl(lynx, lynx->rcv_start_pcl_bus, 0);
643 driver_name, lynx)) {
650 lynx->misc.parent = &dev->dev;
651 lynx->misc.minor = MISC_DYNAMIC_MINOR;
652 lynx->misc.name = "nosy";
653 lynx->misc.fops = &nosy_ops;
656 ret = misc_register(&lynx->misc);
662 list_add_tail(&lynx->link, &card_list);
671 reg_write(lynx, PCI_INT_ENABLE, 0);
672 free_irq(lynx->pci_device->irq, lynx);
675 if (lynx->rcv_start_pcl)
676 dma_free_coherent(&lynx->pci_device->dev, sizeof(struct pcl),
677 lynx->rcv_start_pcl,
678 lynx->rcv_start_pcl_bus);
679 if (lynx->rcv_pcl)
680 dma_free_coherent(&lynx->pci_device->dev, sizeof(struct pcl),
681 lynx->rcv_pcl, lynx->rcv_pcl_bus);
682 if (lynx->rcv_buffer)
683 dma_free_coherent(&lynx->pci_device->dev, PAGE_SIZE,
684 lynx->rcv_buffer, lynx->rcv_buffer_bus);
685 iounmap(lynx->registers);
688 kfree(lynx);