1c110425dSMark Cave-Ayland /* 2c110425dSMark Cave-Ayland * QEMU Sun Happy Meal Ethernet emulation 3c110425dSMark Cave-Ayland * 4c110425dSMark Cave-Ayland * Copyright (c) 2017 Mark Cave-Ayland 5c110425dSMark Cave-Ayland * 6c110425dSMark Cave-Ayland * Permission is hereby granted, free of charge, to any person obtaining a copy 7c110425dSMark Cave-Ayland * of this software and associated documentation files (the "Software"), to deal 8c110425dSMark Cave-Ayland * in the Software without restriction, including without limitation the rights 9c110425dSMark Cave-Ayland * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 10c110425dSMark Cave-Ayland * copies of the Software, and to permit persons to whom the Software is 11c110425dSMark Cave-Ayland * furnished to do so, subject to the following conditions: 12c110425dSMark Cave-Ayland * 13c110425dSMark Cave-Ayland * The above copyright notice and this permission notice shall be included in 14c110425dSMark Cave-Ayland * all copies or substantial portions of the Software. 15c110425dSMark Cave-Ayland * 16c110425dSMark Cave-Ayland * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 17c110425dSMark Cave-Ayland * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 18c110425dSMark Cave-Ayland * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 19c110425dSMark Cave-Ayland * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 20c110425dSMark Cave-Ayland * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 21c110425dSMark Cave-Ayland * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN 22c110425dSMark Cave-Ayland * THE SOFTWARE. 23c110425dSMark Cave-Ayland */ 24c110425dSMark Cave-Ayland 25c110425dSMark Cave-Ayland #include "qemu/osdep.h" 26c110425dSMark Cave-Ayland #include "hw/hw.h" 27c110425dSMark Cave-Ayland #include "hw/pci/pci.h" 28c110425dSMark Cave-Ayland #include "hw/net/mii.h" 29c110425dSMark Cave-Ayland #include "net/net.h" 300b8fa32fSMarkus Armbruster #include "qemu/module.h" 31c110425dSMark Cave-Ayland #include "net/checksum.h" 32c110425dSMark Cave-Ayland #include "net/eth.h" 33c110425dSMark Cave-Ayland #include "sysemu/sysemu.h" 34c110425dSMark Cave-Ayland #include "trace.h" 35c110425dSMark Cave-Ayland 36c110425dSMark Cave-Ayland #define HME_REG_SIZE 0x8000 37c110425dSMark Cave-Ayland 38c110425dSMark Cave-Ayland #define HME_SEB_REG_SIZE 0x2000 39c110425dSMark Cave-Ayland 40c110425dSMark Cave-Ayland #define HME_SEBI_RESET 0x0 41c110425dSMark Cave-Ayland #define HME_SEB_RESET_ETX 0x1 42c110425dSMark Cave-Ayland #define HME_SEB_RESET_ERX 0x2 43c110425dSMark Cave-Ayland 44c110425dSMark Cave-Ayland #define HME_SEBI_STAT 0x100 45c110425dSMark Cave-Ayland #define HME_SEBI_STAT_LINUXBUG 0x108 46c110425dSMark Cave-Ayland #define HME_SEB_STAT_RXTOHOST 0x10000 47c110425dSMark Cave-Ayland #define HME_SEB_STAT_MIFIRQ 0x800000 48c110425dSMark Cave-Ayland #define HME_SEB_STAT_HOSTTOTX 0x1000000 49c110425dSMark Cave-Ayland #define HME_SEB_STAT_TXALL 0x2000000 50c110425dSMark Cave-Ayland 51c110425dSMark Cave-Ayland #define HME_SEBI_IMASK 0x104 52c110425dSMark Cave-Ayland #define HME_SEBI_IMASK_LINUXBUG 0x10c 53c110425dSMark Cave-Ayland 54c110425dSMark Cave-Ayland #define HME_ETX_REG_SIZE 0x2000 55c110425dSMark Cave-Ayland 56c110425dSMark Cave-Ayland #define HME_ETXI_PENDING 0x0 57c110425dSMark Cave-Ayland 58c110425dSMark Cave-Ayland #define HME_ETXI_RING 0x8 59c110425dSMark Cave-Ayland #define HME_ETXI_RING_ADDR 0xffffff00 60c110425dSMark Cave-Ayland #define HME_ETXI_RING_OFFSET 0xff 61c110425dSMark Cave-Ayland 62c110425dSMark Cave-Ayland #define HME_ETXI_RSIZE 0x2c 63c110425dSMark Cave-Ayland 64c110425dSMark Cave-Ayland #define HME_ERX_REG_SIZE 0x2000 65c110425dSMark Cave-Ayland 66c110425dSMark Cave-Ayland #define HME_ERXI_CFG 0x0 67c110425dSMark Cave-Ayland #define HME_ERX_CFG_RINGSIZE 0x600 68c110425dSMark Cave-Ayland #define HME_ERX_CFG_RINGSIZE_SHIFT 9 69c110425dSMark Cave-Ayland #define HME_ERX_CFG_BYTEOFFSET 0x38 70c110425dSMark Cave-Ayland #define HME_ERX_CFG_BYTEOFFSET_SHIFT 3 71c110425dSMark Cave-Ayland #define HME_ERX_CFG_CSUMSTART 0x7f0000 72c110425dSMark Cave-Ayland #define HME_ERX_CFG_CSUMSHIFT 16 73c110425dSMark Cave-Ayland 74c110425dSMark Cave-Ayland #define HME_ERXI_RING 0x4 75c110425dSMark Cave-Ayland #define HME_ERXI_RING_ADDR 0xffffff00 76c110425dSMark Cave-Ayland #define HME_ERXI_RING_OFFSET 0xff 77c110425dSMark Cave-Ayland 78c110425dSMark Cave-Ayland #define HME_MAC_REG_SIZE 0x1000 79c110425dSMark Cave-Ayland 80c110425dSMark Cave-Ayland #define HME_MACI_TXCFG 0x20c 81c110425dSMark Cave-Ayland #define HME_MAC_TXCFG_ENABLE 0x1 82c110425dSMark Cave-Ayland 83c110425dSMark Cave-Ayland #define HME_MACI_RXCFG 0x30c 84c110425dSMark Cave-Ayland #define HME_MAC_RXCFG_ENABLE 0x1 85c110425dSMark Cave-Ayland #define HME_MAC_RXCFG_PMISC 0x40 86c110425dSMark Cave-Ayland #define HME_MAC_RXCFG_HENABLE 0x800 87c110425dSMark Cave-Ayland 88c110425dSMark Cave-Ayland #define HME_MACI_MACADDR2 0x318 89c110425dSMark Cave-Ayland #define HME_MACI_MACADDR1 0x31c 90c110425dSMark Cave-Ayland #define HME_MACI_MACADDR0 0x320 91c110425dSMark Cave-Ayland 92c110425dSMark Cave-Ayland #define HME_MACI_HASHTAB3 0x340 93c110425dSMark Cave-Ayland #define HME_MACI_HASHTAB2 0x344 94c110425dSMark Cave-Ayland #define HME_MACI_HASHTAB1 0x348 95c110425dSMark Cave-Ayland #define HME_MACI_HASHTAB0 0x34c 96c110425dSMark Cave-Ayland 97c110425dSMark Cave-Ayland #define HME_MIF_REG_SIZE 0x20 98c110425dSMark Cave-Ayland 99c110425dSMark Cave-Ayland #define HME_MIFI_FO 0xc 100c110425dSMark Cave-Ayland #define HME_MIF_FO_ST 0xc0000000 101c110425dSMark Cave-Ayland #define HME_MIF_FO_ST_SHIFT 30 102c110425dSMark Cave-Ayland #define HME_MIF_FO_OPC 0x30000000 103c110425dSMark Cave-Ayland #define HME_MIF_FO_OPC_SHIFT 28 104c110425dSMark Cave-Ayland #define HME_MIF_FO_PHYAD 0x0f800000 105c110425dSMark Cave-Ayland #define HME_MIF_FO_PHYAD_SHIFT 23 106c110425dSMark Cave-Ayland #define HME_MIF_FO_REGAD 0x007c0000 107c110425dSMark Cave-Ayland #define HME_MIF_FO_REGAD_SHIFT 18 108c110425dSMark Cave-Ayland #define HME_MIF_FO_TAMSB 0x20000 109c110425dSMark Cave-Ayland #define HME_MIF_FO_TALSB 0x10000 110c110425dSMark Cave-Ayland #define HME_MIF_FO_DATA 0xffff 111c110425dSMark Cave-Ayland 112c110425dSMark Cave-Ayland #define HME_MIFI_CFG 0x10 113c110425dSMark Cave-Ayland #define HME_MIF_CFG_MDI0 0x100 114c110425dSMark Cave-Ayland #define HME_MIF_CFG_MDI1 0x200 115c110425dSMark Cave-Ayland 116c110425dSMark Cave-Ayland #define HME_MIFI_IMASK 0x14 117c110425dSMark Cave-Ayland 118c110425dSMark Cave-Ayland #define HME_MIFI_STAT 0x18 119c110425dSMark Cave-Ayland 120c110425dSMark Cave-Ayland 121c110425dSMark Cave-Ayland /* Wired HME PHY addresses */ 122c110425dSMark Cave-Ayland #define HME_PHYAD_INTERNAL 1 123c110425dSMark Cave-Ayland #define HME_PHYAD_EXTERNAL 0 124c110425dSMark Cave-Ayland 125c110425dSMark Cave-Ayland #define MII_COMMAND_START 0x1 126c110425dSMark Cave-Ayland #define MII_COMMAND_READ 0x2 127c110425dSMark Cave-Ayland #define MII_COMMAND_WRITE 0x1 128c110425dSMark Cave-Ayland 129c110425dSMark Cave-Ayland #define TYPE_SUNHME "sunhme" 130c110425dSMark Cave-Ayland #define SUNHME(obj) OBJECT_CHECK(SunHMEState, (obj), TYPE_SUNHME) 131c110425dSMark Cave-Ayland 132c110425dSMark Cave-Ayland /* Maximum size of buffer */ 133c110425dSMark Cave-Ayland #define HME_FIFO_SIZE 0x800 134c110425dSMark Cave-Ayland 135c110425dSMark Cave-Ayland /* Size of TX/RX descriptor */ 136c110425dSMark Cave-Ayland #define HME_DESC_SIZE 0x8 137c110425dSMark Cave-Ayland 138c110425dSMark Cave-Ayland #define HME_XD_OWN 0x80000000 139c110425dSMark Cave-Ayland #define HME_XD_OFL 0x40000000 140c110425dSMark Cave-Ayland #define HME_XD_SOP 0x40000000 141c110425dSMark Cave-Ayland #define HME_XD_EOP 0x20000000 142c110425dSMark Cave-Ayland #define HME_XD_RXLENMSK 0x3fff0000 143c110425dSMark Cave-Ayland #define HME_XD_RXLENSHIFT 16 144c110425dSMark Cave-Ayland #define HME_XD_RXCKSUM 0xffff 145c110425dSMark Cave-Ayland #define HME_XD_TXLENMSK 0x00001fff 146c110425dSMark Cave-Ayland #define HME_XD_TXCKSUM 0x10000000 147c110425dSMark Cave-Ayland #define HME_XD_TXCSSTUFF 0xff00000 148c110425dSMark Cave-Ayland #define HME_XD_TXCSSTUFFSHIFT 20 149c110425dSMark Cave-Ayland #define HME_XD_TXCSSTART 0xfc000 150c110425dSMark Cave-Ayland #define HME_XD_TXCSSTARTSHIFT 14 151c110425dSMark Cave-Ayland 152c110425dSMark Cave-Ayland #define HME_MII_REGS_SIZE 0x20 153c110425dSMark Cave-Ayland 154c110425dSMark Cave-Ayland typedef struct SunHMEState { 155c110425dSMark Cave-Ayland /*< private >*/ 156c110425dSMark Cave-Ayland PCIDevice parent_obj; 157c110425dSMark Cave-Ayland 158c110425dSMark Cave-Ayland NICState *nic; 159c110425dSMark Cave-Ayland NICConf conf; 160c110425dSMark Cave-Ayland 161c110425dSMark Cave-Ayland MemoryRegion hme; 162c110425dSMark Cave-Ayland MemoryRegion sebreg; 163c110425dSMark Cave-Ayland MemoryRegion etxreg; 164c110425dSMark Cave-Ayland MemoryRegion erxreg; 165c110425dSMark Cave-Ayland MemoryRegion macreg; 166c110425dSMark Cave-Ayland MemoryRegion mifreg; 167c110425dSMark Cave-Ayland 168c110425dSMark Cave-Ayland uint32_t sebregs[HME_SEB_REG_SIZE >> 2]; 169c110425dSMark Cave-Ayland uint32_t etxregs[HME_ETX_REG_SIZE >> 2]; 170c110425dSMark Cave-Ayland uint32_t erxregs[HME_ERX_REG_SIZE >> 2]; 171c110425dSMark Cave-Ayland uint32_t macregs[HME_MAC_REG_SIZE >> 2]; 172c110425dSMark Cave-Ayland uint32_t mifregs[HME_MIF_REG_SIZE >> 2]; 173c110425dSMark Cave-Ayland 174c110425dSMark Cave-Ayland uint16_t miiregs[HME_MII_REGS_SIZE]; 175c110425dSMark Cave-Ayland } SunHMEState; 176c110425dSMark Cave-Ayland 177c110425dSMark Cave-Ayland static Property sunhme_properties[] = { 178c110425dSMark Cave-Ayland DEFINE_NIC_PROPERTIES(SunHMEState, conf), 179c110425dSMark Cave-Ayland DEFINE_PROP_END_OF_LIST(), 180c110425dSMark Cave-Ayland }; 181c110425dSMark Cave-Ayland 182c110425dSMark Cave-Ayland static void sunhme_reset_tx(SunHMEState *s) 183c110425dSMark Cave-Ayland { 184c110425dSMark Cave-Ayland /* Indicate TX reset complete */ 185c110425dSMark Cave-Ayland s->sebregs[HME_SEBI_RESET] &= ~HME_SEB_RESET_ETX; 186c110425dSMark Cave-Ayland } 187c110425dSMark Cave-Ayland 188c110425dSMark Cave-Ayland static void sunhme_reset_rx(SunHMEState *s) 189c110425dSMark Cave-Ayland { 190c110425dSMark Cave-Ayland /* Indicate RX reset complete */ 191c110425dSMark Cave-Ayland s->sebregs[HME_SEBI_RESET] &= ~HME_SEB_RESET_ERX; 192c110425dSMark Cave-Ayland } 193c110425dSMark Cave-Ayland 194c110425dSMark Cave-Ayland static void sunhme_update_irq(SunHMEState *s) 195c110425dSMark Cave-Ayland { 196c110425dSMark Cave-Ayland PCIDevice *d = PCI_DEVICE(s); 197c110425dSMark Cave-Ayland int level; 198c110425dSMark Cave-Ayland 199c110425dSMark Cave-Ayland /* MIF interrupt mask (16-bit) */ 200c110425dSMark Cave-Ayland uint32_t mifmask = ~(s->mifregs[HME_MIFI_IMASK >> 2]) & 0xffff; 201c110425dSMark Cave-Ayland uint32_t mif = s->mifregs[HME_MIFI_STAT >> 2] & mifmask; 202c110425dSMark Cave-Ayland 203c110425dSMark Cave-Ayland /* Main SEB interrupt mask (include MIF status from above) */ 204c110425dSMark Cave-Ayland uint32_t sebmask = ~(s->sebregs[HME_SEBI_IMASK >> 2]) & 205c110425dSMark Cave-Ayland ~HME_SEB_STAT_MIFIRQ; 206c110425dSMark Cave-Ayland uint32_t seb = s->sebregs[HME_SEBI_STAT >> 2] & sebmask; 207c110425dSMark Cave-Ayland if (mif) { 208c110425dSMark Cave-Ayland seb |= HME_SEB_STAT_MIFIRQ; 209c110425dSMark Cave-Ayland } 210c110425dSMark Cave-Ayland 211c110425dSMark Cave-Ayland level = (seb ? 1 : 0); 2126bdc3707SMark Cave-Ayland trace_sunhme_update_irq(mifmask, mif, sebmask, seb, level); 2136bdc3707SMark Cave-Ayland 214c110425dSMark Cave-Ayland pci_set_irq(d, level); 215c110425dSMark Cave-Ayland } 216c110425dSMark Cave-Ayland 217c110425dSMark Cave-Ayland static void sunhme_seb_write(void *opaque, hwaddr addr, 218c110425dSMark Cave-Ayland uint64_t val, unsigned size) 219c110425dSMark Cave-Ayland { 220c110425dSMark Cave-Ayland SunHMEState *s = SUNHME(opaque); 221c110425dSMark Cave-Ayland 222c110425dSMark Cave-Ayland trace_sunhme_seb_write(addr, val); 223c110425dSMark Cave-Ayland 224c110425dSMark Cave-Ayland /* Handly buggy Linux drivers before 4.13 which have 225c110425dSMark Cave-Ayland the wrong offsets for HME_SEBI_STAT and HME_SEBI_IMASK */ 226c110425dSMark Cave-Ayland switch (addr) { 227c110425dSMark Cave-Ayland case HME_SEBI_STAT_LINUXBUG: 228c110425dSMark Cave-Ayland addr = HME_SEBI_STAT; 229c110425dSMark Cave-Ayland break; 230c110425dSMark Cave-Ayland case HME_SEBI_IMASK_LINUXBUG: 231c110425dSMark Cave-Ayland addr = HME_SEBI_IMASK; 232c110425dSMark Cave-Ayland break; 233c110425dSMark Cave-Ayland default: 234c110425dSMark Cave-Ayland break; 235c110425dSMark Cave-Ayland } 236c110425dSMark Cave-Ayland 237c110425dSMark Cave-Ayland switch (addr) { 238c110425dSMark Cave-Ayland case HME_SEBI_RESET: 239c110425dSMark Cave-Ayland if (val & HME_SEB_RESET_ETX) { 240c110425dSMark Cave-Ayland sunhme_reset_tx(s); 241c110425dSMark Cave-Ayland } 242c110425dSMark Cave-Ayland if (val & HME_SEB_RESET_ERX) { 243c110425dSMark Cave-Ayland sunhme_reset_rx(s); 244c110425dSMark Cave-Ayland } 245c110425dSMark Cave-Ayland val = s->sebregs[HME_SEBI_RESET >> 2]; 246c110425dSMark Cave-Ayland break; 247c110425dSMark Cave-Ayland } 248c110425dSMark Cave-Ayland 249c110425dSMark Cave-Ayland s->sebregs[addr >> 2] = val; 250c110425dSMark Cave-Ayland } 251c110425dSMark Cave-Ayland 252c110425dSMark Cave-Ayland static uint64_t sunhme_seb_read(void *opaque, hwaddr addr, 253c110425dSMark Cave-Ayland unsigned size) 254c110425dSMark Cave-Ayland { 255c110425dSMark Cave-Ayland SunHMEState *s = SUNHME(opaque); 256c110425dSMark Cave-Ayland uint64_t val; 257c110425dSMark Cave-Ayland 258c110425dSMark Cave-Ayland /* Handly buggy Linux drivers before 4.13 which have 259c110425dSMark Cave-Ayland the wrong offsets for HME_SEBI_STAT and HME_SEBI_IMASK */ 260c110425dSMark Cave-Ayland switch (addr) { 261c110425dSMark Cave-Ayland case HME_SEBI_STAT_LINUXBUG: 262c110425dSMark Cave-Ayland addr = HME_SEBI_STAT; 263c110425dSMark Cave-Ayland break; 264c110425dSMark Cave-Ayland case HME_SEBI_IMASK_LINUXBUG: 265c110425dSMark Cave-Ayland addr = HME_SEBI_IMASK; 266c110425dSMark Cave-Ayland break; 267c110425dSMark Cave-Ayland default: 268c110425dSMark Cave-Ayland break; 269c110425dSMark Cave-Ayland } 270c110425dSMark Cave-Ayland 271c110425dSMark Cave-Ayland val = s->sebregs[addr >> 2]; 272c110425dSMark Cave-Ayland 273c110425dSMark Cave-Ayland switch (addr) { 274c110425dSMark Cave-Ayland case HME_SEBI_STAT: 275c110425dSMark Cave-Ayland /* Autoclear status (except MIF) */ 276c110425dSMark Cave-Ayland s->sebregs[HME_SEBI_STAT >> 2] &= HME_SEB_STAT_MIFIRQ; 277c110425dSMark Cave-Ayland sunhme_update_irq(s); 278c110425dSMark Cave-Ayland break; 279c110425dSMark Cave-Ayland } 280c110425dSMark Cave-Ayland 281c110425dSMark Cave-Ayland trace_sunhme_seb_read(addr, val); 282c110425dSMark Cave-Ayland 283c110425dSMark Cave-Ayland return val; 284c110425dSMark Cave-Ayland } 285c110425dSMark Cave-Ayland 286c110425dSMark Cave-Ayland static const MemoryRegionOps sunhme_seb_ops = { 287c110425dSMark Cave-Ayland .read = sunhme_seb_read, 288c110425dSMark Cave-Ayland .write = sunhme_seb_write, 289c110425dSMark Cave-Ayland .endianness = DEVICE_LITTLE_ENDIAN, 290c110425dSMark Cave-Ayland .valid = { 291c110425dSMark Cave-Ayland .min_access_size = 4, 292c110425dSMark Cave-Ayland .max_access_size = 4, 293c110425dSMark Cave-Ayland }, 294c110425dSMark Cave-Ayland }; 295c110425dSMark Cave-Ayland 296c110425dSMark Cave-Ayland static void sunhme_transmit(SunHMEState *s); 297c110425dSMark Cave-Ayland 298c110425dSMark Cave-Ayland static void sunhme_etx_write(void *opaque, hwaddr addr, 299c110425dSMark Cave-Ayland uint64_t val, unsigned size) 300c110425dSMark Cave-Ayland { 301c110425dSMark Cave-Ayland SunHMEState *s = SUNHME(opaque); 302c110425dSMark Cave-Ayland 303c110425dSMark Cave-Ayland trace_sunhme_etx_write(addr, val); 304c110425dSMark Cave-Ayland 305c110425dSMark Cave-Ayland switch (addr) { 306c110425dSMark Cave-Ayland case HME_ETXI_PENDING: 307c110425dSMark Cave-Ayland if (val) { 308c110425dSMark Cave-Ayland sunhme_transmit(s); 309c110425dSMark Cave-Ayland } 310c110425dSMark Cave-Ayland break; 311c110425dSMark Cave-Ayland } 312c110425dSMark Cave-Ayland 313c110425dSMark Cave-Ayland s->etxregs[addr >> 2] = val; 314c110425dSMark Cave-Ayland } 315c110425dSMark Cave-Ayland 316c110425dSMark Cave-Ayland static uint64_t sunhme_etx_read(void *opaque, hwaddr addr, 317c110425dSMark Cave-Ayland unsigned size) 318c110425dSMark Cave-Ayland { 319c110425dSMark Cave-Ayland SunHMEState *s = SUNHME(opaque); 320c110425dSMark Cave-Ayland uint64_t val; 321c110425dSMark Cave-Ayland 322c110425dSMark Cave-Ayland val = s->etxregs[addr >> 2]; 323c110425dSMark Cave-Ayland 324c110425dSMark Cave-Ayland trace_sunhme_etx_read(addr, val); 325c110425dSMark Cave-Ayland 326c110425dSMark Cave-Ayland return val; 327c110425dSMark Cave-Ayland } 328c110425dSMark Cave-Ayland 329c110425dSMark Cave-Ayland static const MemoryRegionOps sunhme_etx_ops = { 330c110425dSMark Cave-Ayland .read = sunhme_etx_read, 331c110425dSMark Cave-Ayland .write = sunhme_etx_write, 332c110425dSMark Cave-Ayland .endianness = DEVICE_LITTLE_ENDIAN, 333c110425dSMark Cave-Ayland .valid = { 334c110425dSMark Cave-Ayland .min_access_size = 4, 335c110425dSMark Cave-Ayland .max_access_size = 4, 336c110425dSMark Cave-Ayland }, 337c110425dSMark Cave-Ayland }; 338c110425dSMark Cave-Ayland 339c110425dSMark Cave-Ayland static void sunhme_erx_write(void *opaque, hwaddr addr, 340c110425dSMark Cave-Ayland uint64_t val, unsigned size) 341c110425dSMark Cave-Ayland { 342c110425dSMark Cave-Ayland SunHMEState *s = SUNHME(opaque); 343c110425dSMark Cave-Ayland 344c110425dSMark Cave-Ayland trace_sunhme_erx_write(addr, val); 345c110425dSMark Cave-Ayland 346c110425dSMark Cave-Ayland s->erxregs[addr >> 2] = val; 347c110425dSMark Cave-Ayland } 348c110425dSMark Cave-Ayland 349c110425dSMark Cave-Ayland static uint64_t sunhme_erx_read(void *opaque, hwaddr addr, 350c110425dSMark Cave-Ayland unsigned size) 351c110425dSMark Cave-Ayland { 352c110425dSMark Cave-Ayland SunHMEState *s = SUNHME(opaque); 353c110425dSMark Cave-Ayland uint64_t val; 354c110425dSMark Cave-Ayland 355c110425dSMark Cave-Ayland val = s->erxregs[addr >> 2]; 356c110425dSMark Cave-Ayland 357c110425dSMark Cave-Ayland trace_sunhme_erx_read(addr, val); 358c110425dSMark Cave-Ayland 359c110425dSMark Cave-Ayland return val; 360c110425dSMark Cave-Ayland } 361c110425dSMark Cave-Ayland 362c110425dSMark Cave-Ayland static const MemoryRegionOps sunhme_erx_ops = { 363c110425dSMark Cave-Ayland .read = sunhme_erx_read, 364c110425dSMark Cave-Ayland .write = sunhme_erx_write, 365c110425dSMark Cave-Ayland .endianness = DEVICE_LITTLE_ENDIAN, 366c110425dSMark Cave-Ayland .valid = { 367c110425dSMark Cave-Ayland .min_access_size = 4, 368c110425dSMark Cave-Ayland .max_access_size = 4, 369c110425dSMark Cave-Ayland }, 370c110425dSMark Cave-Ayland }; 371c110425dSMark Cave-Ayland 372c110425dSMark Cave-Ayland static void sunhme_mac_write(void *opaque, hwaddr addr, 373c110425dSMark Cave-Ayland uint64_t val, unsigned size) 374c110425dSMark Cave-Ayland { 375c110425dSMark Cave-Ayland SunHMEState *s = SUNHME(opaque); 376c110425dSMark Cave-Ayland 377c110425dSMark Cave-Ayland trace_sunhme_mac_write(addr, val); 378c110425dSMark Cave-Ayland 379c110425dSMark Cave-Ayland s->macregs[addr >> 2] = val; 380c110425dSMark Cave-Ayland } 381c110425dSMark Cave-Ayland 382c110425dSMark Cave-Ayland static uint64_t sunhme_mac_read(void *opaque, hwaddr addr, 383c110425dSMark Cave-Ayland unsigned size) 384c110425dSMark Cave-Ayland { 385c110425dSMark Cave-Ayland SunHMEState *s = SUNHME(opaque); 386c110425dSMark Cave-Ayland uint64_t val; 387c110425dSMark Cave-Ayland 388c110425dSMark Cave-Ayland val = s->macregs[addr >> 2]; 389c110425dSMark Cave-Ayland 390c110425dSMark Cave-Ayland trace_sunhme_mac_read(addr, val); 391c110425dSMark Cave-Ayland 392c110425dSMark Cave-Ayland return val; 393c110425dSMark Cave-Ayland } 394c110425dSMark Cave-Ayland 395c110425dSMark Cave-Ayland static const MemoryRegionOps sunhme_mac_ops = { 396c110425dSMark Cave-Ayland .read = sunhme_mac_read, 397c110425dSMark Cave-Ayland .write = sunhme_mac_write, 398c110425dSMark Cave-Ayland .endianness = DEVICE_LITTLE_ENDIAN, 399c110425dSMark Cave-Ayland .valid = { 400c110425dSMark Cave-Ayland .min_access_size = 4, 401c110425dSMark Cave-Ayland .max_access_size = 4, 402c110425dSMark Cave-Ayland }, 403c110425dSMark Cave-Ayland }; 404c110425dSMark Cave-Ayland 405c110425dSMark Cave-Ayland static void sunhme_mii_write(SunHMEState *s, uint8_t reg, uint16_t data) 406c110425dSMark Cave-Ayland { 407c110425dSMark Cave-Ayland trace_sunhme_mii_write(reg, data); 408c110425dSMark Cave-Ayland 409c110425dSMark Cave-Ayland switch (reg) { 410c110425dSMark Cave-Ayland case MII_BMCR: 411c110425dSMark Cave-Ayland if (data & MII_BMCR_RESET) { 412c110425dSMark Cave-Ayland /* Autoclear reset bit, enable auto negotiation */ 413c110425dSMark Cave-Ayland data &= ~MII_BMCR_RESET; 414c110425dSMark Cave-Ayland data |= MII_BMCR_AUTOEN; 415c110425dSMark Cave-Ayland } 416c110425dSMark Cave-Ayland if (data & MII_BMCR_ANRESTART) { 417c110425dSMark Cave-Ayland /* Autoclear auto negotiation restart */ 418c110425dSMark Cave-Ayland data &= ~MII_BMCR_ANRESTART; 419c110425dSMark Cave-Ayland 420c110425dSMark Cave-Ayland /* Indicate negotiation complete */ 421c110425dSMark Cave-Ayland s->miiregs[MII_BMSR] |= MII_BMSR_AN_COMP; 422c110425dSMark Cave-Ayland 423c110425dSMark Cave-Ayland if (!qemu_get_queue(s->nic)->link_down) { 424c110425dSMark Cave-Ayland s->miiregs[MII_ANLPAR] |= MII_ANLPAR_TXFD; 425c110425dSMark Cave-Ayland s->miiregs[MII_BMSR] |= MII_BMSR_LINK_ST; 426c110425dSMark Cave-Ayland } 427c110425dSMark Cave-Ayland } 428c110425dSMark Cave-Ayland break; 429c110425dSMark Cave-Ayland } 430c110425dSMark Cave-Ayland 431c110425dSMark Cave-Ayland s->miiregs[reg] = data; 432c110425dSMark Cave-Ayland } 433c110425dSMark Cave-Ayland 434c110425dSMark Cave-Ayland static uint16_t sunhme_mii_read(SunHMEState *s, uint8_t reg) 435c110425dSMark Cave-Ayland { 436c110425dSMark Cave-Ayland uint16_t data = s->miiregs[reg]; 437c110425dSMark Cave-Ayland 438c110425dSMark Cave-Ayland trace_sunhme_mii_read(reg, data); 439c110425dSMark Cave-Ayland 440c110425dSMark Cave-Ayland return data; 441c110425dSMark Cave-Ayland } 442c110425dSMark Cave-Ayland 443c110425dSMark Cave-Ayland static void sunhme_mif_write(void *opaque, hwaddr addr, 444c110425dSMark Cave-Ayland uint64_t val, unsigned size) 445c110425dSMark Cave-Ayland { 446c110425dSMark Cave-Ayland SunHMEState *s = SUNHME(opaque); 447c110425dSMark Cave-Ayland uint8_t cmd, reg; 448c110425dSMark Cave-Ayland uint16_t data; 449c110425dSMark Cave-Ayland 450c110425dSMark Cave-Ayland trace_sunhme_mif_write(addr, val); 451c110425dSMark Cave-Ayland 452c110425dSMark Cave-Ayland switch (addr) { 453c110425dSMark Cave-Ayland case HME_MIFI_CFG: 454c110425dSMark Cave-Ayland /* Mask the read-only bits */ 455c110425dSMark Cave-Ayland val &= ~(HME_MIF_CFG_MDI0 | HME_MIF_CFG_MDI1); 456c110425dSMark Cave-Ayland val |= s->mifregs[HME_MIFI_CFG >> 2] & 457c110425dSMark Cave-Ayland (HME_MIF_CFG_MDI0 | HME_MIF_CFG_MDI1); 458c110425dSMark Cave-Ayland break; 459c110425dSMark Cave-Ayland case HME_MIFI_FO: 460c110425dSMark Cave-Ayland /* Detect start of MII command */ 461c110425dSMark Cave-Ayland if ((val & HME_MIF_FO_ST) >> HME_MIF_FO_ST_SHIFT 462c110425dSMark Cave-Ayland != MII_COMMAND_START) { 463c110425dSMark Cave-Ayland val |= HME_MIF_FO_TALSB; 464c110425dSMark Cave-Ayland break; 465c110425dSMark Cave-Ayland } 466c110425dSMark Cave-Ayland 467c110425dSMark Cave-Ayland /* Internal phy only */ 468c110425dSMark Cave-Ayland if ((val & HME_MIF_FO_PHYAD) >> HME_MIF_FO_PHYAD_SHIFT 469c110425dSMark Cave-Ayland != HME_PHYAD_INTERNAL) { 470c110425dSMark Cave-Ayland val |= HME_MIF_FO_TALSB; 471c110425dSMark Cave-Ayland break; 472c110425dSMark Cave-Ayland } 473c110425dSMark Cave-Ayland 474c110425dSMark Cave-Ayland cmd = (val & HME_MIF_FO_OPC) >> HME_MIF_FO_OPC_SHIFT; 475c110425dSMark Cave-Ayland reg = (val & HME_MIF_FO_REGAD) >> HME_MIF_FO_REGAD_SHIFT; 476c110425dSMark Cave-Ayland data = (val & HME_MIF_FO_DATA); 477c110425dSMark Cave-Ayland 478c110425dSMark Cave-Ayland switch (cmd) { 479c110425dSMark Cave-Ayland case MII_COMMAND_WRITE: 480c110425dSMark Cave-Ayland sunhme_mii_write(s, reg, data); 481c110425dSMark Cave-Ayland break; 482c110425dSMark Cave-Ayland 483c110425dSMark Cave-Ayland case MII_COMMAND_READ: 484c110425dSMark Cave-Ayland val &= ~HME_MIF_FO_DATA; 485c110425dSMark Cave-Ayland val |= sunhme_mii_read(s, reg); 486c110425dSMark Cave-Ayland break; 487c110425dSMark Cave-Ayland } 488c110425dSMark Cave-Ayland 489c110425dSMark Cave-Ayland val |= HME_MIF_FO_TALSB; 490c110425dSMark Cave-Ayland break; 491c110425dSMark Cave-Ayland } 492c110425dSMark Cave-Ayland 493c110425dSMark Cave-Ayland s->mifregs[addr >> 2] = val; 494c110425dSMark Cave-Ayland } 495c110425dSMark Cave-Ayland 496c110425dSMark Cave-Ayland static uint64_t sunhme_mif_read(void *opaque, hwaddr addr, 497c110425dSMark Cave-Ayland unsigned size) 498c110425dSMark Cave-Ayland { 499c110425dSMark Cave-Ayland SunHMEState *s = SUNHME(opaque); 500c110425dSMark Cave-Ayland uint64_t val; 501c110425dSMark Cave-Ayland 502c110425dSMark Cave-Ayland val = s->mifregs[addr >> 2]; 503c110425dSMark Cave-Ayland 504c110425dSMark Cave-Ayland switch (addr) { 505c110425dSMark Cave-Ayland case HME_MIFI_STAT: 506c110425dSMark Cave-Ayland /* Autoclear MIF interrupt status */ 507c110425dSMark Cave-Ayland s->mifregs[HME_MIFI_STAT >> 2] = 0; 508c110425dSMark Cave-Ayland sunhme_update_irq(s); 509c110425dSMark Cave-Ayland break; 510c110425dSMark Cave-Ayland } 511c110425dSMark Cave-Ayland 512c110425dSMark Cave-Ayland trace_sunhme_mif_read(addr, val); 513c110425dSMark Cave-Ayland 514c110425dSMark Cave-Ayland return val; 515c110425dSMark Cave-Ayland } 516c110425dSMark Cave-Ayland 517c110425dSMark Cave-Ayland static const MemoryRegionOps sunhme_mif_ops = { 518c110425dSMark Cave-Ayland .read = sunhme_mif_read, 519c110425dSMark Cave-Ayland .write = sunhme_mif_write, 520c110425dSMark Cave-Ayland .endianness = DEVICE_LITTLE_ENDIAN, 521c110425dSMark Cave-Ayland .valid = { 522c110425dSMark Cave-Ayland .min_access_size = 4, 523c110425dSMark Cave-Ayland .max_access_size = 4, 524c110425dSMark Cave-Ayland }, 525c110425dSMark Cave-Ayland }; 526c110425dSMark Cave-Ayland 527c110425dSMark Cave-Ayland static void sunhme_transmit_frame(SunHMEState *s, uint8_t *buf, int size) 528c110425dSMark Cave-Ayland { 529c110425dSMark Cave-Ayland qemu_send_packet(qemu_get_queue(s->nic), buf, size); 530c110425dSMark Cave-Ayland } 531c110425dSMark Cave-Ayland 532c110425dSMark Cave-Ayland static inline int sunhme_get_tx_ring_count(SunHMEState *s) 533c110425dSMark Cave-Ayland { 534c110425dSMark Cave-Ayland return (s->etxregs[HME_ETXI_RSIZE >> 2] + 1) << 4; 535c110425dSMark Cave-Ayland } 536c110425dSMark Cave-Ayland 537c110425dSMark Cave-Ayland static inline int sunhme_get_tx_ring_nr(SunHMEState *s) 538c110425dSMark Cave-Ayland { 539c110425dSMark Cave-Ayland return s->etxregs[HME_ETXI_RING >> 2] & HME_ETXI_RING_OFFSET; 540c110425dSMark Cave-Ayland } 541c110425dSMark Cave-Ayland 542c110425dSMark Cave-Ayland static inline void sunhme_set_tx_ring_nr(SunHMEState *s, int i) 543c110425dSMark Cave-Ayland { 544c110425dSMark Cave-Ayland uint32_t ring = s->etxregs[HME_ETXI_RING >> 2] & ~HME_ETXI_RING_OFFSET; 545c110425dSMark Cave-Ayland ring |= i & HME_ETXI_RING_OFFSET; 546c110425dSMark Cave-Ayland 547c110425dSMark Cave-Ayland s->etxregs[HME_ETXI_RING >> 2] = ring; 548c110425dSMark Cave-Ayland } 549c110425dSMark Cave-Ayland 550c110425dSMark Cave-Ayland static void sunhme_transmit(SunHMEState *s) 551c110425dSMark Cave-Ayland { 552c110425dSMark Cave-Ayland PCIDevice *d = PCI_DEVICE(s); 553c110425dSMark Cave-Ayland dma_addr_t tb, addr; 554c110425dSMark Cave-Ayland uint32_t intstatus, status, buffer, sum = 0; 555c110425dSMark Cave-Ayland int cr, nr, len, xmit_pos, csum_offset = 0, csum_stuff_offset = 0; 556c110425dSMark Cave-Ayland uint16_t csum = 0; 557c110425dSMark Cave-Ayland uint8_t xmit_buffer[HME_FIFO_SIZE]; 558c110425dSMark Cave-Ayland 559c110425dSMark Cave-Ayland tb = s->etxregs[HME_ETXI_RING >> 2] & HME_ETXI_RING_ADDR; 560c110425dSMark Cave-Ayland nr = sunhme_get_tx_ring_count(s); 561c110425dSMark Cave-Ayland cr = sunhme_get_tx_ring_nr(s); 562c110425dSMark Cave-Ayland 563c110425dSMark Cave-Ayland pci_dma_read(d, tb + cr * HME_DESC_SIZE, &status, 4); 564c110425dSMark Cave-Ayland pci_dma_read(d, tb + cr * HME_DESC_SIZE + 4, &buffer, 4); 565c110425dSMark Cave-Ayland 566c110425dSMark Cave-Ayland xmit_pos = 0; 567c110425dSMark Cave-Ayland while (status & HME_XD_OWN) { 568c110425dSMark Cave-Ayland trace_sunhme_tx_desc(buffer, status, cr, nr); 569c110425dSMark Cave-Ayland 570c110425dSMark Cave-Ayland /* Copy data into transmit buffer */ 571c110425dSMark Cave-Ayland addr = buffer; 572c110425dSMark Cave-Ayland len = status & HME_XD_TXLENMSK; 573c110425dSMark Cave-Ayland 574c110425dSMark Cave-Ayland if (xmit_pos + len > HME_FIFO_SIZE) { 575c110425dSMark Cave-Ayland len = HME_FIFO_SIZE - xmit_pos; 576c110425dSMark Cave-Ayland } 577c110425dSMark Cave-Ayland 578c110425dSMark Cave-Ayland pci_dma_read(d, addr, &xmit_buffer[xmit_pos], len); 579c110425dSMark Cave-Ayland xmit_pos += len; 580c110425dSMark Cave-Ayland 581c110425dSMark Cave-Ayland /* Detect start of packet for TX checksum */ 582c110425dSMark Cave-Ayland if (status & HME_XD_SOP) { 583c110425dSMark Cave-Ayland sum = 0; 584c110425dSMark Cave-Ayland csum_offset = (status & HME_XD_TXCSSTART) >> HME_XD_TXCSSTARTSHIFT; 585c110425dSMark Cave-Ayland csum_stuff_offset = (status & HME_XD_TXCSSTUFF) >> 586c110425dSMark Cave-Ayland HME_XD_TXCSSTUFFSHIFT; 587c110425dSMark Cave-Ayland } 588c110425dSMark Cave-Ayland 589c110425dSMark Cave-Ayland if (status & HME_XD_TXCKSUM) { 590c110425dSMark Cave-Ayland /* Only start calculation from csum_offset */ 591c110425dSMark Cave-Ayland if (xmit_pos - len <= csum_offset && xmit_pos > csum_offset) { 592c110425dSMark Cave-Ayland sum += net_checksum_add(xmit_pos - csum_offset, 593c110425dSMark Cave-Ayland xmit_buffer + csum_offset); 594c110425dSMark Cave-Ayland trace_sunhme_tx_xsum_add(csum_offset, xmit_pos - csum_offset); 595c110425dSMark Cave-Ayland } else { 596c110425dSMark Cave-Ayland sum += net_checksum_add(len, xmit_buffer + xmit_pos - len); 597c110425dSMark Cave-Ayland trace_sunhme_tx_xsum_add(xmit_pos - len, len); 598c110425dSMark Cave-Ayland } 599c110425dSMark Cave-Ayland } 600c110425dSMark Cave-Ayland 601c110425dSMark Cave-Ayland /* Detect end of packet for TX checksum */ 602c110425dSMark Cave-Ayland if (status & HME_XD_EOP) { 603c110425dSMark Cave-Ayland /* Stuff the checksum if required */ 604c110425dSMark Cave-Ayland if (status & HME_XD_TXCKSUM) { 605c110425dSMark Cave-Ayland csum = net_checksum_finish(sum); 606c110425dSMark Cave-Ayland stw_be_p(xmit_buffer + csum_stuff_offset, csum); 607c110425dSMark Cave-Ayland trace_sunhme_tx_xsum_stuff(csum, csum_stuff_offset); 608c110425dSMark Cave-Ayland } 609c110425dSMark Cave-Ayland 610c110425dSMark Cave-Ayland if (s->macregs[HME_MACI_TXCFG >> 2] & HME_MAC_TXCFG_ENABLE) { 611c110425dSMark Cave-Ayland sunhme_transmit_frame(s, xmit_buffer, xmit_pos); 612c110425dSMark Cave-Ayland trace_sunhme_tx_done(xmit_pos); 613c110425dSMark Cave-Ayland } 614c110425dSMark Cave-Ayland } 615c110425dSMark Cave-Ayland 616c110425dSMark Cave-Ayland /* Update status */ 617c110425dSMark Cave-Ayland status &= ~HME_XD_OWN; 618c110425dSMark Cave-Ayland pci_dma_write(d, tb + cr * HME_DESC_SIZE, &status, 4); 619c110425dSMark Cave-Ayland 620c110425dSMark Cave-Ayland /* Move onto next descriptor */ 621c110425dSMark Cave-Ayland cr++; 622c110425dSMark Cave-Ayland if (cr >= nr) { 623c110425dSMark Cave-Ayland cr = 0; 624c110425dSMark Cave-Ayland } 625c110425dSMark Cave-Ayland sunhme_set_tx_ring_nr(s, cr); 626c110425dSMark Cave-Ayland 627c110425dSMark Cave-Ayland pci_dma_read(d, tb + cr * HME_DESC_SIZE, &status, 4); 628c110425dSMark Cave-Ayland pci_dma_read(d, tb + cr * HME_DESC_SIZE + 4, &buffer, 4); 629c110425dSMark Cave-Ayland 630c110425dSMark Cave-Ayland /* Indicate TX complete */ 631c110425dSMark Cave-Ayland intstatus = s->sebregs[HME_SEBI_STAT >> 2]; 632c110425dSMark Cave-Ayland intstatus |= HME_SEB_STAT_HOSTTOTX; 633c110425dSMark Cave-Ayland s->sebregs[HME_SEBI_STAT >> 2] = intstatus; 634c110425dSMark Cave-Ayland 635c110425dSMark Cave-Ayland /* Autoclear TX pending */ 636c110425dSMark Cave-Ayland s->etxregs[HME_ETXI_PENDING >> 2] = 0; 637c110425dSMark Cave-Ayland 638c110425dSMark Cave-Ayland sunhme_update_irq(s); 639c110425dSMark Cave-Ayland } 640c110425dSMark Cave-Ayland 641c110425dSMark Cave-Ayland /* TX FIFO now clear */ 642c110425dSMark Cave-Ayland intstatus = s->sebregs[HME_SEBI_STAT >> 2]; 643c110425dSMark Cave-Ayland intstatus |= HME_SEB_STAT_TXALL; 644c110425dSMark Cave-Ayland s->sebregs[HME_SEBI_STAT >> 2] = intstatus; 645c110425dSMark Cave-Ayland sunhme_update_irq(s); 646c110425dSMark Cave-Ayland } 647c110425dSMark Cave-Ayland 648c110425dSMark Cave-Ayland static int sunhme_can_receive(NetClientState *nc) 649c110425dSMark Cave-Ayland { 650c110425dSMark Cave-Ayland SunHMEState *s = qemu_get_nic_opaque(nc); 651c110425dSMark Cave-Ayland 652*076489c0SMark Cave-Ayland return s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_ENABLE; 653c110425dSMark Cave-Ayland } 654c110425dSMark Cave-Ayland 655c110425dSMark Cave-Ayland static void sunhme_link_status_changed(NetClientState *nc) 656c110425dSMark Cave-Ayland { 657c110425dSMark Cave-Ayland SunHMEState *s = qemu_get_nic_opaque(nc); 658c110425dSMark Cave-Ayland 659c110425dSMark Cave-Ayland if (nc->link_down) { 660c110425dSMark Cave-Ayland s->miiregs[MII_ANLPAR] &= ~MII_ANLPAR_TXFD; 661c110425dSMark Cave-Ayland s->miiregs[MII_BMSR] &= ~MII_BMSR_LINK_ST; 662c110425dSMark Cave-Ayland } else { 663c110425dSMark Cave-Ayland s->miiregs[MII_ANLPAR] |= MII_ANLPAR_TXFD; 664c110425dSMark Cave-Ayland s->miiregs[MII_BMSR] |= MII_BMSR_LINK_ST; 665c110425dSMark Cave-Ayland } 666c110425dSMark Cave-Ayland 667c110425dSMark Cave-Ayland /* Exact bits unknown */ 668c110425dSMark Cave-Ayland s->mifregs[HME_MIFI_STAT >> 2] = 0xffff; 669c110425dSMark Cave-Ayland sunhme_update_irq(s); 670c110425dSMark Cave-Ayland } 671c110425dSMark Cave-Ayland 672c110425dSMark Cave-Ayland static inline int sunhme_get_rx_ring_count(SunHMEState *s) 673c110425dSMark Cave-Ayland { 674c110425dSMark Cave-Ayland uint32_t rings = (s->erxregs[HME_ERXI_CFG >> 2] & HME_ERX_CFG_RINGSIZE) 675c110425dSMark Cave-Ayland >> HME_ERX_CFG_RINGSIZE_SHIFT; 676c110425dSMark Cave-Ayland 677c110425dSMark Cave-Ayland switch (rings) { 678c110425dSMark Cave-Ayland case 0: 679c110425dSMark Cave-Ayland return 32; 680c110425dSMark Cave-Ayland case 1: 681c110425dSMark Cave-Ayland return 64; 682c110425dSMark Cave-Ayland case 2: 683c110425dSMark Cave-Ayland return 128; 684c110425dSMark Cave-Ayland case 3: 685c110425dSMark Cave-Ayland return 256; 686c110425dSMark Cave-Ayland } 687c110425dSMark Cave-Ayland 688c110425dSMark Cave-Ayland return 0; 689c110425dSMark Cave-Ayland } 690c110425dSMark Cave-Ayland 691c110425dSMark Cave-Ayland static inline int sunhme_get_rx_ring_nr(SunHMEState *s) 692c110425dSMark Cave-Ayland { 693c110425dSMark Cave-Ayland return s->erxregs[HME_ERXI_RING >> 2] & HME_ERXI_RING_OFFSET; 694c110425dSMark Cave-Ayland } 695c110425dSMark Cave-Ayland 696c110425dSMark Cave-Ayland static inline void sunhme_set_rx_ring_nr(SunHMEState *s, int i) 697c110425dSMark Cave-Ayland { 698c110425dSMark Cave-Ayland uint32_t ring = s->erxregs[HME_ERXI_RING >> 2] & ~HME_ERXI_RING_OFFSET; 699c110425dSMark Cave-Ayland ring |= i & HME_ERXI_RING_OFFSET; 700c110425dSMark Cave-Ayland 701c110425dSMark Cave-Ayland s->erxregs[HME_ERXI_RING >> 2] = ring; 702c110425dSMark Cave-Ayland } 703c110425dSMark Cave-Ayland 704c110425dSMark Cave-Ayland #define MIN_BUF_SIZE 60 705c110425dSMark Cave-Ayland 706c110425dSMark Cave-Ayland static ssize_t sunhme_receive(NetClientState *nc, const uint8_t *buf, 707c110425dSMark Cave-Ayland size_t size) 708c110425dSMark Cave-Ayland { 709c110425dSMark Cave-Ayland SunHMEState *s = qemu_get_nic_opaque(nc); 710c110425dSMark Cave-Ayland PCIDevice *d = PCI_DEVICE(s); 711c110425dSMark Cave-Ayland dma_addr_t rb, addr; 712c110425dSMark Cave-Ayland uint32_t intstatus, status, buffer, buffersize, sum; 713c110425dSMark Cave-Ayland uint16_t csum; 714c110425dSMark Cave-Ayland uint8_t buf1[60]; 715c110425dSMark Cave-Ayland int nr, cr, len, rxoffset, csum_offset; 716c110425dSMark Cave-Ayland 717c110425dSMark Cave-Ayland trace_sunhme_rx_incoming(size); 718c110425dSMark Cave-Ayland 719c110425dSMark Cave-Ayland /* Do nothing if MAC RX disabled */ 720c110425dSMark Cave-Ayland if (!(s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_ENABLE)) { 721c110425dSMark Cave-Ayland return -1; 722c110425dSMark Cave-Ayland } 723c110425dSMark Cave-Ayland 724c110425dSMark Cave-Ayland trace_sunhme_rx_filter_destmac(buf[0], buf[1], buf[2], 725c110425dSMark Cave-Ayland buf[3], buf[4], buf[5]); 726c110425dSMark Cave-Ayland 727c110425dSMark Cave-Ayland /* Check destination MAC address */ 728c110425dSMark Cave-Ayland if (!(s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_PMISC)) { 729c110425dSMark Cave-Ayland /* Try and match local MAC address */ 730c110425dSMark Cave-Ayland if (((s->macregs[HME_MACI_MACADDR0 >> 2] & 0xff00) >> 8) == buf[0] && 731c110425dSMark Cave-Ayland (s->macregs[HME_MACI_MACADDR0 >> 2] & 0xff) == buf[1] && 732c110425dSMark Cave-Ayland ((s->macregs[HME_MACI_MACADDR1 >> 2] & 0xff00) >> 8) == buf[2] && 733c110425dSMark Cave-Ayland (s->macregs[HME_MACI_MACADDR1 >> 2] & 0xff) == buf[3] && 734c110425dSMark Cave-Ayland ((s->macregs[HME_MACI_MACADDR2 >> 2] & 0xff00) >> 8) == buf[4] && 735c110425dSMark Cave-Ayland (s->macregs[HME_MACI_MACADDR2 >> 2] & 0xff) == buf[5]) { 736c110425dSMark Cave-Ayland /* Matched local MAC address */ 737c110425dSMark Cave-Ayland trace_sunhme_rx_filter_local_match(); 738c110425dSMark Cave-Ayland } else if (buf[0] == 0xff && buf[1] == 0xff && buf[2] == 0xff && 739c110425dSMark Cave-Ayland buf[3] == 0xff && buf[4] == 0xff && buf[5] == 0xff) { 740c110425dSMark Cave-Ayland /* Matched broadcast address */ 741c110425dSMark Cave-Ayland trace_sunhme_rx_filter_bcast_match(); 742c110425dSMark Cave-Ayland } else if (s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_HENABLE) { 743c110425dSMark Cave-Ayland /* Didn't match local address, check hash filter */ 744a89a6b05SMark Cave-Ayland int mcast_idx = net_crc32_le(buf, ETH_ALEN) >> 26; 745c110425dSMark Cave-Ayland if (!(s->macregs[(HME_MACI_HASHTAB0 >> 2) - (mcast_idx >> 4)] & 746c110425dSMark Cave-Ayland (1 << (mcast_idx & 0xf)))) { 747c110425dSMark Cave-Ayland /* Didn't match hash filter */ 748c110425dSMark Cave-Ayland trace_sunhme_rx_filter_hash_nomatch(); 749c110425dSMark Cave-Ayland trace_sunhme_rx_filter_reject(); 750c110425dSMark Cave-Ayland return 0; 751c110425dSMark Cave-Ayland } else { 752c110425dSMark Cave-Ayland trace_sunhme_rx_filter_hash_match(); 753c110425dSMark Cave-Ayland } 754c110425dSMark Cave-Ayland } else { 755c110425dSMark Cave-Ayland /* Not for us */ 756c110425dSMark Cave-Ayland trace_sunhme_rx_filter_reject(); 757c110425dSMark Cave-Ayland return 0; 758c110425dSMark Cave-Ayland } 759c110425dSMark Cave-Ayland } else { 760c110425dSMark Cave-Ayland trace_sunhme_rx_filter_promisc_match(); 761c110425dSMark Cave-Ayland } 762c110425dSMark Cave-Ayland 763c110425dSMark Cave-Ayland trace_sunhme_rx_filter_accept(); 764c110425dSMark Cave-Ayland 765c110425dSMark Cave-Ayland /* If too small buffer, then expand it */ 766c110425dSMark Cave-Ayland if (size < MIN_BUF_SIZE) { 767c110425dSMark Cave-Ayland memcpy(buf1, buf, size); 768c110425dSMark Cave-Ayland memset(buf1 + size, 0, MIN_BUF_SIZE - size); 769c110425dSMark Cave-Ayland buf = buf1; 770c110425dSMark Cave-Ayland size = MIN_BUF_SIZE; 771c110425dSMark Cave-Ayland } 772c110425dSMark Cave-Ayland 773c110425dSMark Cave-Ayland rb = s->erxregs[HME_ERXI_RING >> 2] & HME_ERXI_RING_ADDR; 774c110425dSMark Cave-Ayland nr = sunhme_get_rx_ring_count(s); 775c110425dSMark Cave-Ayland cr = sunhme_get_rx_ring_nr(s); 776c110425dSMark Cave-Ayland 777c110425dSMark Cave-Ayland pci_dma_read(d, rb + cr * HME_DESC_SIZE, &status, 4); 778c110425dSMark Cave-Ayland pci_dma_read(d, rb + cr * HME_DESC_SIZE + 4, &buffer, 4); 779c110425dSMark Cave-Ayland 780c110425dSMark Cave-Ayland rxoffset = (s->erxregs[HME_ERXI_CFG >> 2] & HME_ERX_CFG_BYTEOFFSET) >> 781c110425dSMark Cave-Ayland HME_ERX_CFG_BYTEOFFSET_SHIFT; 782c110425dSMark Cave-Ayland 783c110425dSMark Cave-Ayland addr = buffer + rxoffset; 784c110425dSMark Cave-Ayland buffersize = (status & HME_XD_RXLENMSK) >> HME_XD_RXLENSHIFT; 785c110425dSMark Cave-Ayland 786c110425dSMark Cave-Ayland /* Detect receive overflow */ 787c110425dSMark Cave-Ayland len = size; 788c110425dSMark Cave-Ayland if (size > buffersize) { 789c110425dSMark Cave-Ayland status |= HME_XD_OFL; 790c110425dSMark Cave-Ayland len = buffersize; 791c110425dSMark Cave-Ayland } 792c110425dSMark Cave-Ayland 793c110425dSMark Cave-Ayland pci_dma_write(d, addr, buf, len); 794c110425dSMark Cave-Ayland 795c110425dSMark Cave-Ayland trace_sunhme_rx_desc(buffer, rxoffset, status, len, cr, nr); 796c110425dSMark Cave-Ayland 797c110425dSMark Cave-Ayland /* Calculate the receive checksum */ 798c110425dSMark Cave-Ayland csum_offset = (s->erxregs[HME_ERXI_CFG >> 2] & HME_ERX_CFG_CSUMSTART) >> 799c110425dSMark Cave-Ayland HME_ERX_CFG_CSUMSHIFT << 1; 800c110425dSMark Cave-Ayland sum = 0; 801c110425dSMark Cave-Ayland sum += net_checksum_add(len - csum_offset, (uint8_t *)buf + csum_offset); 802c110425dSMark Cave-Ayland csum = net_checksum_finish(sum); 803c110425dSMark Cave-Ayland 804c110425dSMark Cave-Ayland trace_sunhme_rx_xsum_calc(csum); 805c110425dSMark Cave-Ayland 806c110425dSMark Cave-Ayland /* Update status */ 807c110425dSMark Cave-Ayland status &= ~HME_XD_OWN; 808c110425dSMark Cave-Ayland status &= ~HME_XD_RXLENMSK; 809c110425dSMark Cave-Ayland status |= len << HME_XD_RXLENSHIFT; 810c110425dSMark Cave-Ayland status &= ~HME_XD_RXCKSUM; 811c110425dSMark Cave-Ayland status |= csum; 812c110425dSMark Cave-Ayland 813c110425dSMark Cave-Ayland pci_dma_write(d, rb + cr * HME_DESC_SIZE, &status, 4); 814c110425dSMark Cave-Ayland 815c110425dSMark Cave-Ayland cr++; 816c110425dSMark Cave-Ayland if (cr >= nr) { 817c110425dSMark Cave-Ayland cr = 0; 818c110425dSMark Cave-Ayland } 819c110425dSMark Cave-Ayland 820c110425dSMark Cave-Ayland sunhme_set_rx_ring_nr(s, cr); 821c110425dSMark Cave-Ayland 822c110425dSMark Cave-Ayland /* Indicate RX complete */ 823c110425dSMark Cave-Ayland intstatus = s->sebregs[HME_SEBI_STAT >> 2]; 824c110425dSMark Cave-Ayland intstatus |= HME_SEB_STAT_RXTOHOST; 825c110425dSMark Cave-Ayland s->sebregs[HME_SEBI_STAT >> 2] = intstatus; 826c110425dSMark Cave-Ayland 827c110425dSMark Cave-Ayland sunhme_update_irq(s); 828c110425dSMark Cave-Ayland 829c110425dSMark Cave-Ayland return len; 830c110425dSMark Cave-Ayland } 831c110425dSMark Cave-Ayland 832c110425dSMark Cave-Ayland static NetClientInfo net_sunhme_info = { 833c110425dSMark Cave-Ayland .type = NET_CLIENT_DRIVER_NIC, 834c110425dSMark Cave-Ayland .size = sizeof(NICState), 835c110425dSMark Cave-Ayland .can_receive = sunhme_can_receive, 836c110425dSMark Cave-Ayland .receive = sunhme_receive, 837c110425dSMark Cave-Ayland .link_status_changed = sunhme_link_status_changed, 838c110425dSMark Cave-Ayland }; 839c110425dSMark Cave-Ayland 840c110425dSMark Cave-Ayland static void sunhme_realize(PCIDevice *pci_dev, Error **errp) 841c110425dSMark Cave-Ayland { 842c110425dSMark Cave-Ayland SunHMEState *s = SUNHME(pci_dev); 843c110425dSMark Cave-Ayland DeviceState *d = DEVICE(pci_dev); 844c110425dSMark Cave-Ayland uint8_t *pci_conf; 845c110425dSMark Cave-Ayland 846c110425dSMark Cave-Ayland pci_conf = pci_dev->config; 847c110425dSMark Cave-Ayland pci_conf[PCI_INTERRUPT_PIN] = 1; /* interrupt pin A */ 848c110425dSMark Cave-Ayland 849c110425dSMark Cave-Ayland memory_region_init(&s->hme, OBJECT(pci_dev), "sunhme", HME_REG_SIZE); 850c110425dSMark Cave-Ayland pci_register_bar(pci_dev, 0, PCI_BASE_ADDRESS_SPACE_MEMORY, &s->hme); 851c110425dSMark Cave-Ayland 852c110425dSMark Cave-Ayland memory_region_init_io(&s->sebreg, OBJECT(pci_dev), &sunhme_seb_ops, s, 853c110425dSMark Cave-Ayland "sunhme.seb", HME_SEB_REG_SIZE); 854c110425dSMark Cave-Ayland memory_region_add_subregion(&s->hme, 0, &s->sebreg); 855c110425dSMark Cave-Ayland 856c110425dSMark Cave-Ayland memory_region_init_io(&s->etxreg, OBJECT(pci_dev), &sunhme_etx_ops, s, 857c110425dSMark Cave-Ayland "sunhme.etx", HME_ETX_REG_SIZE); 858c110425dSMark Cave-Ayland memory_region_add_subregion(&s->hme, 0x2000, &s->etxreg); 859c110425dSMark Cave-Ayland 860c110425dSMark Cave-Ayland memory_region_init_io(&s->erxreg, OBJECT(pci_dev), &sunhme_erx_ops, s, 861c110425dSMark Cave-Ayland "sunhme.erx", HME_ERX_REG_SIZE); 862c110425dSMark Cave-Ayland memory_region_add_subregion(&s->hme, 0x4000, &s->erxreg); 863c110425dSMark Cave-Ayland 864c110425dSMark Cave-Ayland memory_region_init_io(&s->macreg, OBJECT(pci_dev), &sunhme_mac_ops, s, 865c110425dSMark Cave-Ayland "sunhme.mac", HME_MAC_REG_SIZE); 866c110425dSMark Cave-Ayland memory_region_add_subregion(&s->hme, 0x6000, &s->macreg); 867c110425dSMark Cave-Ayland 868c110425dSMark Cave-Ayland memory_region_init_io(&s->mifreg, OBJECT(pci_dev), &sunhme_mif_ops, s, 869c110425dSMark Cave-Ayland "sunhme.mif", HME_MIF_REG_SIZE); 870c110425dSMark Cave-Ayland memory_region_add_subregion(&s->hme, 0x7000, &s->mifreg); 871c110425dSMark Cave-Ayland 872c110425dSMark Cave-Ayland qemu_macaddr_default_if_unset(&s->conf.macaddr); 873c110425dSMark Cave-Ayland s->nic = qemu_new_nic(&net_sunhme_info, &s->conf, 874c110425dSMark Cave-Ayland object_get_typename(OBJECT(d)), d->id, s); 875c110425dSMark Cave-Ayland qemu_format_nic_info_str(qemu_get_queue(s->nic), s->conf.macaddr.a); 876c110425dSMark Cave-Ayland } 877c110425dSMark Cave-Ayland 878c110425dSMark Cave-Ayland static void sunhme_instance_init(Object *obj) 879c110425dSMark Cave-Ayland { 880c110425dSMark Cave-Ayland SunHMEState *s = SUNHME(obj); 881c110425dSMark Cave-Ayland 882c110425dSMark Cave-Ayland device_add_bootindex_property(obj, &s->conf.bootindex, 883c110425dSMark Cave-Ayland "bootindex", "/ethernet-phy@0", 884c110425dSMark Cave-Ayland DEVICE(obj), NULL); 885c110425dSMark Cave-Ayland } 886c110425dSMark Cave-Ayland 887c110425dSMark Cave-Ayland static void sunhme_reset(DeviceState *ds) 888c110425dSMark Cave-Ayland { 889c110425dSMark Cave-Ayland SunHMEState *s = SUNHME(ds); 890c110425dSMark Cave-Ayland 891c110425dSMark Cave-Ayland /* Configure internal transceiver */ 892c110425dSMark Cave-Ayland s->mifregs[HME_MIFI_CFG >> 2] |= HME_MIF_CFG_MDI0; 893c110425dSMark Cave-Ayland 894c110425dSMark Cave-Ayland /* Advetise auto, 100Mbps FD */ 895c110425dSMark Cave-Ayland s->miiregs[MII_ANAR] = MII_ANAR_TXFD; 896c110425dSMark Cave-Ayland s->miiregs[MII_BMSR] = MII_BMSR_AUTONEG | MII_BMSR_100TX_FD | 897c110425dSMark Cave-Ayland MII_BMSR_AN_COMP; 898c110425dSMark Cave-Ayland 899c110425dSMark Cave-Ayland if (!qemu_get_queue(s->nic)->link_down) { 900c110425dSMark Cave-Ayland s->miiregs[MII_ANLPAR] |= MII_ANLPAR_TXFD; 901c110425dSMark Cave-Ayland s->miiregs[MII_BMSR] |= MII_BMSR_LINK_ST; 902c110425dSMark Cave-Ayland } 903c110425dSMark Cave-Ayland 904c110425dSMark Cave-Ayland /* Set manufacturer */ 905c110425dSMark Cave-Ayland s->miiregs[MII_PHYID1] = DP83840_PHYID1; 906c110425dSMark Cave-Ayland s->miiregs[MII_PHYID2] = DP83840_PHYID2; 907c110425dSMark Cave-Ayland 908c110425dSMark Cave-Ayland /* Configure default interrupt mask */ 909c110425dSMark Cave-Ayland s->mifregs[HME_MIFI_IMASK >> 2] = 0xffff; 910c110425dSMark Cave-Ayland s->sebregs[HME_SEBI_IMASK >> 2] = 0xff7fffff; 911c110425dSMark Cave-Ayland } 912c110425dSMark Cave-Ayland 913c110425dSMark Cave-Ayland static const VMStateDescription vmstate_hme = { 914c110425dSMark Cave-Ayland .name = "sunhme", 915c110425dSMark Cave-Ayland .version_id = 0, 916c110425dSMark Cave-Ayland .minimum_version_id = 0, 917c110425dSMark Cave-Ayland .fields = (VMStateField[]) { 918c110425dSMark Cave-Ayland VMSTATE_PCI_DEVICE(parent_obj, SunHMEState), 919c110425dSMark Cave-Ayland VMSTATE_MACADDR(conf.macaddr, SunHMEState), 920c110425dSMark Cave-Ayland VMSTATE_UINT32_ARRAY(sebregs, SunHMEState, (HME_SEB_REG_SIZE >> 2)), 921c110425dSMark Cave-Ayland VMSTATE_UINT32_ARRAY(etxregs, SunHMEState, (HME_ETX_REG_SIZE >> 2)), 922c110425dSMark Cave-Ayland VMSTATE_UINT32_ARRAY(erxregs, SunHMEState, (HME_ERX_REG_SIZE >> 2)), 923c110425dSMark Cave-Ayland VMSTATE_UINT32_ARRAY(macregs, SunHMEState, (HME_MAC_REG_SIZE >> 2)), 924c110425dSMark Cave-Ayland VMSTATE_UINT32_ARRAY(mifregs, SunHMEState, (HME_MIF_REG_SIZE >> 2)), 925c110425dSMark Cave-Ayland VMSTATE_UINT16_ARRAY(miiregs, SunHMEState, HME_MII_REGS_SIZE), 926c110425dSMark Cave-Ayland VMSTATE_END_OF_LIST() 927c110425dSMark Cave-Ayland } 928c110425dSMark Cave-Ayland }; 929c110425dSMark Cave-Ayland 930c110425dSMark Cave-Ayland static void sunhme_class_init(ObjectClass *klass, void *data) 931c110425dSMark Cave-Ayland { 932c110425dSMark Cave-Ayland DeviceClass *dc = DEVICE_CLASS(klass); 933c110425dSMark Cave-Ayland PCIDeviceClass *k = PCI_DEVICE_CLASS(klass); 934c110425dSMark Cave-Ayland 935c110425dSMark Cave-Ayland k->realize = sunhme_realize; 936c110425dSMark Cave-Ayland k->vendor_id = PCI_VENDOR_ID_SUN; 937c110425dSMark Cave-Ayland k->device_id = PCI_DEVICE_ID_SUN_HME; 938c110425dSMark Cave-Ayland k->class_id = PCI_CLASS_NETWORK_ETHERNET; 939c110425dSMark Cave-Ayland dc->vmsd = &vmstate_hme; 940c110425dSMark Cave-Ayland dc->reset = sunhme_reset; 941c110425dSMark Cave-Ayland dc->props = sunhme_properties; 942c110425dSMark Cave-Ayland set_bit(DEVICE_CATEGORY_NETWORK, dc->categories); 943c110425dSMark Cave-Ayland } 944c110425dSMark Cave-Ayland 945c110425dSMark Cave-Ayland static const TypeInfo sunhme_info = { 946c110425dSMark Cave-Ayland .name = TYPE_SUNHME, 947c110425dSMark Cave-Ayland .parent = TYPE_PCI_DEVICE, 948c110425dSMark Cave-Ayland .class_init = sunhme_class_init, 949c110425dSMark Cave-Ayland .instance_size = sizeof(SunHMEState), 950c110425dSMark Cave-Ayland .instance_init = sunhme_instance_init, 951fd3b02c8SEduardo Habkost .interfaces = (InterfaceInfo[]) { 952fd3b02c8SEduardo Habkost { INTERFACE_CONVENTIONAL_PCI_DEVICE }, 953fd3b02c8SEduardo Habkost { } 954fd3b02c8SEduardo Habkost } 955c110425dSMark Cave-Ayland }; 956c110425dSMark Cave-Ayland 957c110425dSMark Cave-Ayland static void sunhme_register_types(void) 958c110425dSMark Cave-Ayland { 959c110425dSMark Cave-Ayland type_register_static(&sunhme_info); 960c110425dSMark Cave-Ayland } 961c110425dSMark Cave-Ayland 962c110425dSMark Cave-Ayland type_init(sunhme_register_types) 963