[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[PATCH 33/52] migration/rdma: Drop "@errp is clear" guards around error_
From: |
Markus Armbruster |
Subject: |
[PATCH 33/52] migration/rdma: Drop "@errp is clear" guards around error_setg() |
Date: |
Mon, 18 Sep 2023 16:41:47 +0200 |
These guards are all redundant now.
Signed-off-by: Markus Armbruster <armbru@redhat.com>
---
migration/rdma.c | 164 +++++++++++++++--------------------------------
1 file changed, 51 insertions(+), 113 deletions(-)
diff --git a/migration/rdma.c b/migration/rdma.c
index d29affe410..c88cd1f468 100644
--- a/migration/rdma.c
+++ b/migration/rdma.c
@@ -852,10 +852,8 @@ static int qemu_rdma_broken_ipv6_kernel(struct ibv_context
*verbs, Error **errp)
if (ibv_query_port(verbs, 1, &port_attr)) {
ibv_close_device(verbs);
- if (errp && !*errp) {
- error_setg(errp,
- "RDMA ERROR: Could not query initial IB port");
- }
+ error_setg(errp,
+ "RDMA ERROR: Could not query initial IB port");
return -1;
}
@@ -878,12 +876,10 @@ static int qemu_rdma_broken_ipv6_kernel(struct
ibv_context *verbs, Error **errp)
" migrate over the IB fabric until the kernel "
" fixes the bug.\n");
} else {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: "
- "You only have RoCE / iWARP devices in your
systems"
- " and your management software has specified
'[::]'"
- ", but IPv6 over RoCE / iWARP is not supported
in Linux.");
- }
+ error_setg(errp, "RDMA ERROR: "
+ "You only have RoCE / iWARP devices in your systems"
+ " and your management software has specified '[::]'"
+ ", but IPv6 over RoCE / iWARP is not supported in
Linux.");
return -1;
}
}
@@ -899,18 +895,14 @@ static int qemu_rdma_broken_ipv6_kernel(struct
ibv_context *verbs, Error **errp)
/* IB ports start with 1, not 0 */
if (ibv_query_port(verbs, 1, &port_attr)) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: Could not query initial IB port");
- }
+ error_setg(errp, "RDMA ERROR: Could not query initial IB port");
return -1;
}
if (port_attr.link_layer == IBV_LINK_LAYER_ETHERNET) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: "
- "Linux kernel's RoCE / iWARP does not support IPv6 "
- "(but patches on linux-rdma in progress)");
- }
+ error_setg(errp, "RDMA ERROR: "
+ "Linux kernel's RoCE / iWARP does not support IPv6 "
+ "(but patches on linux-rdma in progress)");
return -1;
}
@@ -935,27 +927,21 @@ static int qemu_rdma_resolve_host(RDMAContext *rdma,
Error **errp)
struct rdma_addrinfo *e;
if (rdma->host == NULL || !strcmp(rdma->host, "")) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: RDMA hostname has not been set");
- }
+ error_setg(errp, "RDMA ERROR: RDMA hostname has not been set");
return -1;
}
/* create CM channel */
rdma->channel = rdma_create_event_channel();
if (!rdma->channel) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: could not create CM channel");
- }
+ error_setg(errp, "RDMA ERROR: could not create CM channel");
return -1;
}
/* create CM id */
ret = rdma_create_id(rdma->channel, &rdma->cm_id, NULL, RDMA_PS_TCP);
if (ret < 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: could not create channel id");
- }
+ error_setg(errp, "RDMA ERROR: could not create channel id");
goto err_resolve_create_id;
}
@@ -964,10 +950,8 @@ static int qemu_rdma_resolve_host(RDMAContext *rdma, Error
**errp)
ret = rdma_getaddrinfo(rdma->host, port_str, NULL, &res);
if (ret) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: could not rdma_getaddrinfo address
%s",
- rdma->host);
- }
+ error_setg(errp, "RDMA ERROR: could not rdma_getaddrinfo address %s",
+ rdma->host);
goto err_resolve_get_addr;
}
@@ -1009,18 +993,14 @@ route:
ret = rdma_get_cm_event(rdma->channel, &cm_event);
if (ret < 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: could not perform
event_addr_resolved");
- }
+ error_setg(errp, "RDMA ERROR: could not perform event_addr_resolved");
goto err_resolve_get_addr;
}
if (cm_event->event != RDMA_CM_EVENT_ADDR_RESOLVED) {
- if (errp && !*errp) {
- error_setg(errp,
- "RDMA ERROR: result not equal to event_addr_resolved
%s",
- rdma_event_str(cm_event->event));
- }
+ error_setg(errp,
+ "RDMA ERROR: result not equal to event_addr_resolved %s",
+ rdma_event_str(cm_event->event));
error_report("rdma_resolve_addr");
rdma_ack_cm_event(cm_event);
goto err_resolve_get_addr;
@@ -1030,25 +1010,19 @@ route:
/* resolve route */
ret = rdma_resolve_route(rdma->cm_id, RDMA_RESOLVE_TIMEOUT_MS);
if (ret < 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: could not resolve rdma route");
- }
+ error_setg(errp, "RDMA ERROR: could not resolve rdma route");
goto err_resolve_get_addr;
}
ret = rdma_get_cm_event(rdma->channel, &cm_event);
if (ret < 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: could not perform
event_route_resolved");
- }
+ error_setg(errp, "RDMA ERROR: could not perform event_route_resolved");
goto err_resolve_get_addr;
}
if (cm_event->event != RDMA_CM_EVENT_ROUTE_RESOLVED) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: "
- "result not equal to event_route_resolved: %s",
- rdma_event_str(cm_event->event));
- }
+ error_setg(errp, "RDMA ERROR: "
+ "result not equal to event_route_resolved: %s",
+ rdma_event_str(cm_event->event));
rdma_ack_cm_event(cm_event);
goto err_resolve_get_addr;
}
@@ -2479,20 +2453,16 @@ static int qemu_rdma_source_init(RDMAContext *rdma,
bool pin_all, Error **errp)
ret = qemu_rdma_alloc_pd_cq(rdma);
if (ret < 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: "
- "rdma migration: error allocating pd and cq! Your
mlock()"
- " limits may be too low. Please check $ ulimit -a # and
"
- "search for 'ulimit -l' in the output");
- }
+ error_setg(errp, "RDMA ERROR: "
+ "rdma migration: error allocating pd and cq! Your mlock()"
+ " limits may be too low. Please check $ ulimit -a # and "
+ "search for 'ulimit -l' in the output");
goto err_rdma_source_init;
}
ret = qemu_rdma_alloc_qp(rdma);
if (ret < 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: rdma migration: error allocating
qp!");
- }
+ error_setg(errp, "RDMA ERROR: rdma migration: error allocating qp!");
goto err_rdma_source_init;
}
@@ -2509,11 +2479,9 @@ static int qemu_rdma_source_init(RDMAContext *rdma, bool
pin_all, Error **errp)
for (idx = 0; idx < RDMA_WRID_MAX; idx++) {
ret = qemu_rdma_reg_control(rdma, idx);
if (ret < 0) {
- if (errp && !*errp) {
- error_setg(errp,
- "RDMA ERROR: rdma migration: error registering %d
control!",
- idx);
- }
+ error_setg(errp,
+ "RDMA ERROR: rdma migration: error registering %d
control!",
+ idx);
goto err_rdma_source_init;
}
}
@@ -2541,29 +2509,21 @@ static int qemu_get_cm_event_timeout(RDMAContext *rdma,
} while (ret < 0 && errno == EINTR);
if (ret == 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: poll cm event timeout");
- }
+ error_setg(errp, "RDMA ERROR: poll cm event timeout");
return -1;
} else if (ret < 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: failed to poll cm event, errno=%i",
- errno);
- }
+ error_setg(errp, "RDMA ERROR: failed to poll cm event, errno=%i",
+ errno);
return -1;
} else if (poll_fd.revents & POLLIN) {
if (rdma_get_cm_event(rdma->channel, cm_event) < 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: failed to get cm event");
- }
+ error_setg(errp, "RDMA ERROR: failed to get cm event");
return -1;
}
return 0;
} else {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: no POLLIN event, revent=%x",
- poll_fd.revents);
- }
+ error_setg(errp, "RDMA ERROR: no POLLIN event, revent=%x",
+ poll_fd.revents);
return -1;
}
}
@@ -2596,18 +2556,14 @@ static int qemu_rdma_connect(RDMAContext *rdma, bool
return_path,
ret = qemu_rdma_post_recv_control(rdma, RDMA_WRID_READY);
if (ret < 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: posting second control recv");
- }
+ error_setg(errp, "RDMA ERROR: posting second control recv");
goto err_rdma_source_connect;
}
ret = rdma_connect(rdma->cm_id, &conn_param);
if (ret < 0) {
perror("rdma_connect");
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: connecting to destination!");
- }
+ error_setg(errp, "RDMA ERROR: connecting to destination!");
goto err_rdma_source_connect;
}
@@ -2616,9 +2572,7 @@ static int qemu_rdma_connect(RDMAContext *rdma, bool
return_path,
} else {
ret = rdma_get_cm_event(rdma->channel, &cm_event);
if (ret < 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: failed to get cm event");
- }
+ error_setg(errp, "RDMA ERROR: failed to get cm event");
}
}
if (ret < 0) {
@@ -2628,9 +2582,7 @@ static int qemu_rdma_connect(RDMAContext *rdma, bool
return_path,
if (cm_event->event != RDMA_CM_EVENT_ESTABLISHED) {
error_report("rdma_get_cm_event != EVENT_ESTABLISHED after
rdma_connect");
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: connecting to destination!");
- }
+ error_setg(errp, "RDMA ERROR: connecting to destination!");
rdma_ack_cm_event(cm_event);
goto err_rdma_source_connect;
}
@@ -2678,18 +2630,14 @@ static int qemu_rdma_dest_init(RDMAContext *rdma, Error
**errp)
}
if (!rdma->host || !rdma->host[0]) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: RDMA host is not set!");
- }
+ error_setg(errp, "RDMA ERROR: RDMA host is not set!");
rdma->errored = true;
return -1;
}
/* create CM channel */
rdma->channel = rdma_create_event_channel();
if (!rdma->channel) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: could not create rdma event
channel");
- }
+ error_setg(errp, "RDMA ERROR: could not create rdma event channel");
rdma->errored = true;
return -1;
}
@@ -2697,9 +2645,7 @@ static int qemu_rdma_dest_init(RDMAContext *rdma, Error
**errp)
/* create CM id */
ret = rdma_create_id(rdma->channel, &listen_id, NULL, RDMA_PS_TCP);
if (ret < 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: could not create cm_id!");
- }
+ error_setg(errp, "RDMA ERROR: could not create cm_id!");
goto err_dest_init_create_listen_id;
}
@@ -2708,19 +2654,15 @@ static int qemu_rdma_dest_init(RDMAContext *rdma, Error
**errp)
ret = rdma_getaddrinfo(rdma->host, port_str, NULL, &res);
if (ret) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: could not rdma_getaddrinfo address
%s",
- rdma->host);
- }
+ error_setg(errp, "RDMA ERROR: could not rdma_getaddrinfo address %s",
+ rdma->host);
goto err_dest_init_bind_addr;
}
ret = rdma_set_option(listen_id, RDMA_OPTION_ID, RDMA_OPTION_ID_REUSEADDR,
&reuse, sizeof reuse);
if (ret < 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: Error: could not set REUSEADDR
option");
- }
+ error_setg(errp, "RDMA ERROR: Error: could not set REUSEADDR option");
goto err_dest_init_bind_addr;
}
@@ -2804,10 +2746,8 @@ static RDMAContext *qemu_rdma_data_init(const char
*host_port, Error **errp)
rdma->host = g_strdup(addr->host);
rdma->host_port = g_strdup(host_port);
} else {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: bad RDMA migration address '%s'",
- host_port);
- }
+ error_setg(errp, "RDMA ERROR: bad RDMA migration address '%s'",
+ host_port);
g_free(rdma);
rdma = NULL;
}
@@ -4193,9 +4133,7 @@ void rdma_start_incoming_migration(const char *host_port,
Error **errp)
ret = rdma_listen(rdma->listen_id, 5);
if (ret < 0) {
- if (errp && !*errp) {
- error_setg(errp, "RDMA ERROR: listening on socket!");
- }
+ error_setg(errp, "RDMA ERROR: listening on socket!");
goto cleanup_rdma;
}
--
2.41.0
- [PATCH 37/52] migration/rdma: Convert qemu_rdma_reg_whole_ram_blocks() to Error, (continued)
- [PATCH 37/52] migration/rdma: Convert qemu_rdma_reg_whole_ram_blocks() to Error, Markus Armbruster, 2023/09/18
- [PATCH 07/52] migration/rdma: Give qio_channel_rdma_source_funcs internal linkage, Markus Armbruster, 2023/09/18
- [PATCH 47/52] migration/rdma: Don't report received completion events as error, Markus Armbruster, 2023/09/18
- [PATCH 19/52] migration/rdma: Fix qemu_get_cm_event_timeout() to always set error, Markus Armbruster, 2023/09/18
- [PATCH 33/52] migration/rdma: Drop "@errp is clear" guards around error_setg(),
Markus Armbruster <=
- [PATCH 10/52] migration/rdma: Eliminate error_propagate(), Markus Armbruster, 2023/09/18
- [PATCH 13/52] migration/rdma: Make qemu_rdma_buffer_mergable() return bool, Markus Armbruster, 2023/09/18