Lines Matching +full:rx +full:- +full:shared

1 // SPDX-License-Identifier: GPL-2.0
3 * For transport using shared mem structure.
5 * Copyright (C) 2019-2024 ARM Ltd.
67 * pre-processor.
98 * until it releases the shared memory, otherwise we may endup in shmem_tx_prepare()
99 * overwriting its response with new message payload or vice-versa. in shmem_tx_prepare()
101 * not to bail-out on intermittent issues where the platform is in shmem_tx_prepare()
104 * Note that after a timeout is detected we bail-out and carry on but in shmem_tx_prepare()
109 stop = ktime_add_ms(ktime_get(), 2 * cinfo->rx_timeout_ms); in shmem_tx_prepare()
110 spin_until_cond((ioread32(&shmem->channel_status) & in shmem_tx_prepare()
113 if (!(ioread32(&shmem->channel_status) & in shmem_tx_prepare()
116 dev_err(cinfo->dev, in shmem_tx_prepare()
122 iowrite32(0x0, &shmem->channel_status); in shmem_tx_prepare()
123 iowrite32(xfer->hdr.poll_completion ? 0 : SCMI_SHMEM_FLAG_INTR_ENABLED, in shmem_tx_prepare()
124 &shmem->flags); in shmem_tx_prepare()
125 iowrite32(sizeof(shmem->msg_header) + xfer->tx.len, &shmem->length); in shmem_tx_prepare()
126 iowrite32(pack_scmi_header(&xfer->hdr), &shmem->msg_header); in shmem_tx_prepare()
127 if (xfer->tx.buf) in shmem_tx_prepare()
128 copy_toio(shmem->msg_payload, xfer->tx.buf, xfer->tx.len); in shmem_tx_prepare()
133 return ioread32(&shmem->msg_header); in shmem_read_header()
140 size_t len = ioread32(&shmem->length); in shmem_fetch_response()
142 xfer->hdr.status = ioread32(shmem->msg_payload); in shmem_fetch_response()
144 xfer->rx.len = min_t(size_t, xfer->rx.len, len > 8 ? len - 8 : 0); in shmem_fetch_response()
146 /* Take a copy to the rx buffer.. */ in shmem_fetch_response()
147 copy_fromio(xfer->rx.buf, shmem->msg_payload + 4, xfer->rx.len); in shmem_fetch_response()
154 size_t len = ioread32(&shmem->length); in shmem_fetch_notification()
157 xfer->rx.len = min_t(size_t, max_len, len > 4 ? len - 4 : 0); in shmem_fetch_notification()
159 /* Take a copy to the rx buffer.. */ in shmem_fetch_notification()
160 copy_fromio(xfer->rx.buf, shmem->msg_payload, xfer->rx.len); in shmem_fetch_notification()
165 iowrite32(SCMI_SHMEM_CHAN_STAT_CHANNEL_FREE, &shmem->channel_status); in shmem_clear_channel()
173 xfer_id = MSG_XTRACT_TOKEN(ioread32(&shmem->msg_header)); in shmem_poll_done()
175 if (xfer->hdr.seq != xfer_id) in shmem_poll_done()
178 return ioread32(&shmem->channel_status) & in shmem_poll_done()
185 return (ioread32(&shmem->channel_status) & in shmem_channel_free()
191 return ioread32(&shmem->flags) & SCMI_SHMEM_FLAG_INTR_ENABLED; in shmem_channel_intr_enabled()
200 const char *desc = tx ? "Tx" : "Rx"; in shmem_setup_iomap()
202 struct device *cdev = cinfo->dev; in shmem_setup_iomap()
208 shmem = of_parse_phandle(cdev->of_node, "shmem", idx); in shmem_setup_iomap()
210 return IOMEM_ERR_PTR(-ENODEV); in shmem_setup_iomap()
212 if (!of_device_is_compatible(shmem, "arm,scmi-shmem")) in shmem_setup_iomap()
213 return IOMEM_ERR_PTR(-ENXIO); in shmem_setup_iomap()
215 /* Use a local on-stack as a working area when not provided */ in shmem_setup_iomap()
221 dev_err(cdev, "failed to get SCMI %s shared memory\n", desc); in shmem_setup_iomap()
226 if (cinfo->max_msg_size + SCMI_SHMEM_LAYOUT_OVERHEAD > size) { in shmem_setup_iomap()
227 dev_err(dev, "misconfigured SCMI shared memory\n"); in shmem_setup_iomap()
228 return IOMEM_ERR_PTR(-ENOSPC); in shmem_setup_iomap()
231 addr = devm_ioremap(dev, res->start, size); in shmem_setup_iomap()
233 dev_err(dev, "failed to ioremap SCMI %s shared memory\n", desc); in shmem_setup_iomap()
234 return IOMEM_ERR_PTR(-EADDRNOTAVAIL); in shmem_setup_iomap()
237 of_property_read_u32(shmem, "reg-io-width", &reg_io_width); in shmem_setup_iomap()