[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[PATCH v7 14/21] multi-process: PCI BAR read/write handling for proxy &
From: |
elena . ufimtseva |
Subject: |
[PATCH v7 14/21] multi-process: PCI BAR read/write handling for proxy & remote endpoints |
Date: |
Sat, 27 Jun 2020 10:09:36 -0700 |
From: Jagannathan Raman <jag.raman@oracle.com>
Proxy device object implements handler for PCI BAR writes and reads.
The handler uses BAR_WRITE/BAR_READ message to communicate to the
remote process with the BAR address and value to be written/read.
The remote process implements handler for BAR_WRITE/BAR_READ
message.
Signed-off-by: Jagannathan Raman <jag.raman@oracle.com>
Signed-off-by: Elena Ufimtseva <elena.ufimtseva@oracle.com>
Signed-off-by: John G Johnson <john.g.johnson@oracle.com>
---
hw/i386/remote-msg.c | 95 ++++++++++++++++++++++++++++++++++++++++
hw/pci/proxy.c | 61 ++++++++++++++++++++++++++
include/hw/pci/proxy.h | 16 ++++++-
include/io/mpqemu-link.h | 10 +++++
io/mpqemu-link.c | 6 +++
5 files changed, 186 insertions(+), 2 deletions(-)
diff --git a/hw/i386/remote-msg.c b/hw/i386/remote-msg.c
index aa5780d521..ffb4143736 100644
--- a/hw/i386/remote-msg.c
+++ b/hw/i386/remote-msg.c
@@ -8,6 +8,7 @@
#include "sysemu/runstate.h"
#include "io/channel-util.h"
#include "hw/pci/pci.h"
+#include "exec/memattrs.h"
static void process_connect_dev_msg(MPQemuMsg *msg, QIOChannel *com,
Error **errp);
@@ -15,6 +16,8 @@ static void process_config_write(QIOChannel *ioc, PCIDevice
*dev,
MPQemuMsg *msg);
static void process_config_read(QIOChannel *ioc, PCIDevice *dev,
MPQemuMsg *msg);
+static void process_bar_write(QIOChannel *ioc, MPQemuMsg *msg, Error **errp);
+static void process_bar_read(QIOChannel *ioc, MPQemuMsg *msg, Error **errp);
gboolean mpqemu_process_msg(QIOChannel *ioc, GIOCondition cond,
gpointer opaque)
@@ -54,6 +57,12 @@ gboolean mpqemu_process_msg(QIOChannel *ioc, GIOCondition
cond,
case PCI_CONFIG_READ:
process_config_read(ioc, pci_dev, &msg);
break;
+ case BAR_WRITE:
+ process_bar_write(ioc, &msg, &local_err);
+ break;
+ case BAR_READ:
+ process_bar_read(ioc, &msg, &local_err);
+ break;
default:
error_setg(&local_err, "Unknown command (%d) received from proxy \
in remote process pid=%d", msg.cmd, getpid());
@@ -143,3 +152,89 @@ static void process_config_read(QIOChannel *ioc, PCIDevice
*dev,
mpqemu_msg_send(&ret, ioc);
}
+
+static void process_bar_write(QIOChannel *ioc, MPQemuMsg *msg, Error **errp)
+{
+ BarAccessMsg *bar_access = &msg->data1.bar_access;
+ AddressSpace *as =
+ bar_access->memory ? &address_space_memory : &address_space_io;
+ MPQemuMsg ret = { 0 };
+ MemTxResult res;
+
+ if (!is_power_of_2(bar_access->size) ||
+ (bar_access->size > sizeof(uint64_t))) {
+ ret.data1.u64 = UINT64_MAX;
+ goto fail;
+ }
+
+ res = address_space_rw(as, bar_access->addr, MEMTXATTRS_UNSPECIFIED,
+ (void *)&bar_access->val, bar_access->size,
+ true);
+
+ if (res != MEMTX_OK) {
+ error_setg(errp, "Could not perform address space write operation,"
+ " inaccessible address: %lx in pid %d.",
+ bar_access->addr, getpid());
+ ret.data1.u64 = -1;
+ }
+
+fail:
+ ret.cmd = RET_MSG;
+ ret.size = sizeof(ret.data1);
+
+ mpqemu_msg_send(&ret, ioc);
+}
+
+static void process_bar_read(QIOChannel *ioc, MPQemuMsg *msg, Error **errp)
+{
+ BarAccessMsg *bar_access = &msg->data1.bar_access;
+ MPQemuMsg ret = { 0 };
+ AddressSpace *as;
+ MemTxResult res;
+ uint64_t val = 0;
+
+ as = bar_access->memory ? &address_space_memory : &address_space_io;
+
+ if (!is_power_of_2(bar_access->size) ||
+ (bar_access->size > sizeof(uint64_t))) {
+ val = UINT64_MAX;
+ goto fail;
+ }
+
+ res = address_space_rw(as, bar_access->addr, MEMTXATTRS_UNSPECIFIED,
+ (void *)&val, bar_access->size, false);
+
+ if (res != MEMTX_OK) {
+ error_setg(errp, "Could not perform address space read operation,"
+ " inaccessible address: %lx in pid %d.",
+ bar_access->addr, getpid());
+ val = UINT64_MAX;
+ goto fail;
+ }
+
+ switch (bar_access->size) {
+ case 8:
+ /* Nothing to do as val is already 8 bytes long */
+ break;
+ case 4:
+ val = *((uint32_t *)&val);
+ break;
+ case 2:
+ val = *((uint16_t *)&val);
+ break;
+ case 1:
+ val = *((uint8_t *)&val);
+ break;
+ default:
+ error_setg(errp, "Invalid PCI BAR read size in pid %d",
+ getpid());
+ val = (uint64_t)-1;
+ }
+
+fail:
+ ret.cmd = RET_MSG;
+ ret.data1.u64 = val;
+ ret.size = sizeof(ret.data1);
+
+ mpqemu_msg_send(&ret, ioc);
+}
diff --git a/hw/pci/proxy.c b/hw/pci/proxy.c
index 8934070a20..fff021a06a 100644
--- a/hw/pci/proxy.c
+++ b/hw/pci/proxy.c
@@ -150,3 +150,64 @@ static void pci_proxy_dev_register_types(void)
}
type_init(pci_proxy_dev_register_types)
+
+static void send_bar_access_msg(PCIProxyDev *pdev, MemoryRegion *mr,
+ bool write, hwaddr addr, uint64_t *val,
+ unsigned size, bool memory)
+{
+ MPQemuMsg msg = { 0 };
+ long ret = -EINVAL;
+ Error *local_err = NULL;
+
+ msg.bytestream = 0;
+ msg.size = sizeof(msg.data1);
+ msg.data1.bar_access.addr = mr->addr + addr;
+ msg.data1.bar_access.size = size;
+ msg.data1.bar_access.memory = memory;
+
+ if (write) {
+ msg.cmd = BAR_WRITE;
+ msg.data1.bar_access.val = *val;
+ } else {
+ msg.cmd = BAR_READ;
+ }
+
+ ret = mpqemu_msg_send_reply_co(&msg, pdev->com, &local_err);
+ if (local_err) {
+ error_report("Failed to send BAR command to the remote process.");
+ }
+
+ if (!write) {
+ *val = ret;
+ }
+}
+
+static void proxy_bar_write(void *opaque, hwaddr addr, uint64_t val,
+ unsigned size)
+{
+ ProxyMemoryRegion *pmr = opaque;
+
+ send_bar_access_msg(pmr->dev, &pmr->mr, true, addr, &val, size,
+ pmr->memory);
+}
+
+static uint64_t proxy_bar_read(void *opaque, hwaddr addr, unsigned size)
+{
+ ProxyMemoryRegion *pmr = opaque;
+ uint64_t val;
+
+ send_bar_access_msg(pmr->dev, &pmr->mr, false, addr, &val, size,
+ pmr->memory);
+
+ return val;
+}
+
+const MemoryRegionOps proxy_mr_ops = {
+ .read = proxy_bar_read,
+ .write = proxy_bar_write,
+ .endianness = DEVICE_NATIVE_ENDIAN,
+ .impl = {
+ .min_access_size = 1,
+ .max_access_size = 1,
+ },
+};
diff --git a/include/hw/pci/proxy.h b/include/hw/pci/proxy.h
index 72dd7e0944..4f9f9c4e15 100644
--- a/include/hw/pci/proxy.h
+++ b/include/hw/pci/proxy.h
@@ -26,12 +26,24 @@
#define PCI_PROXY_DEV_GET_CLASS(obj) \
OBJECT_GET_CLASS(PCIProxyDevClass, (obj), TYPE_PCI_PROXY_DEV)
-typedef struct PCIProxyDev {
+typedef struct PCIProxyDev PCIProxyDev;
+
+typedef struct ProxyMemoryRegion {
+ PCIProxyDev *dev;
+ MemoryRegion mr;
+ bool memory;
+ bool present;
+ uint8_t type;
+} ProxyMemoryRegion;
+
+struct PCIProxyDev {
PCIDevice parent_dev;
char *fd;
QIOChannel *com;
QIOChannel *dev;
-} PCIProxyDev;
+
+ ProxyMemoryRegion region[PCI_NUM_REGIONS];
+};
typedef struct PCIProxyDevClass {
PCIDeviceClass parent_class;
diff --git a/include/io/mpqemu-link.h b/include/io/mpqemu-link.h
index bbd6f3db44..0422213863 100644
--- a/include/io/mpqemu-link.h
+++ b/include/io/mpqemu-link.h
@@ -40,6 +40,8 @@ typedef enum {
RET_MSG,
PCI_CONFIG_WRITE,
PCI_CONFIG_READ,
+ BAR_WRITE,
+ BAR_READ,
MAX = INT_MAX,
} MPQemuCmd;
@@ -55,6 +57,13 @@ struct conf_data_msg {
int l;
};
+typedef struct {
+ hwaddr addr;
+ uint64_t val;
+ unsigned size;
+ bool memory;
+} BarAccessMsg;
+
/**
* Maximum size of data2 field in the message to be transmitted.
*/
@@ -82,6 +91,7 @@ typedef struct {
union {
uint64_t u64;
SyncSysmemMsg sync_sysmem;
+ BarAccessMsg bar_access;
} data1;
int fds[REMOTE_MAX_FDS];
diff --git a/io/mpqemu-link.c b/io/mpqemu-link.c
index 19887611d0..026e25dca4 100644
--- a/io/mpqemu-link.c
+++ b/io/mpqemu-link.c
@@ -256,6 +256,12 @@ bool mpqemu_msg_valid(MPQemuMsg *msg)
return false;
}
break;
+ case BAR_WRITE:
+ case BAR_READ:
+ if (msg->size != sizeof(msg->data1)) {
+ return false;
+ }
+ break;
default:
break;
}
--
2.25.GIT
- [PATCH v7 10/21] multi-process: setup memory manager for remote device, (continued)
- [PATCH v7 10/21] multi-process: setup memory manager for remote device, elena . ufimtseva, 2020/06/27
- [PATCH v7 07/21] multi-process: add co-routines to communicate with remote, elena . ufimtseva, 2020/06/27
- [PATCH v7 11/21] multi-process: introduce proxy object, elena . ufimtseva, 2020/06/27
- [PATCH v7 12/21] multi-process: Connect Proxy Object with device in the remote process, elena . ufimtseva, 2020/06/27
- [PATCH v7 13/21] multi-process: Forward PCI config space acceses to the remote process, elena . ufimtseva, 2020/06/27
- [PATCH v7 16/21] multi-process: create IOHUB object to handle irq, elena . ufimtseva, 2020/06/27
- [PATCH v7 18/21] multi-process: heartbeat messages to remote, elena . ufimtseva, 2020/06/27
- [PATCH v7 19/21] multi-process: perform device reset in the remote process, elena . ufimtseva, 2020/06/27
- [PATCH v7 21/21] multi-process: add configure and usage information, elena . ufimtseva, 2020/06/27
- [PATCH v7 20/21] multi-process: add the concept description to docs/devel/qemu-multiprocess, elena . ufimtseva, 2020/06/27
- [PATCH v7 14/21] multi-process: PCI BAR read/write handling for proxy & remote endpoints,
elena . ufimtseva <=
- [PATCH v7 02/21] multi-process: Add config option for multi-process QEMU, elena . ufimtseva, 2020/06/27
- [PATCH v7 03/21] multi-process: setup PCI host bridge for remote device, elena . ufimtseva, 2020/06/27
- [PATCH v7 04/21] multi-process: setup a machine object for remote device process, elena . ufimtseva, 2020/06/27
- [PATCH v7 05/21] multi-process: add qio channel function to transmit, elena . ufimtseva, 2020/06/27
- [PATCH v7 08/21] multi-process: Initialize communication channel at the remote end, elena . ufimtseva, 2020/06/27
- [PATCH v7 09/21] multi-process: Initialize message handler in remote device, elena . ufimtseva, 2020/06/27