0.185-1 - Upgrade to upstream 0.185

This commit is contained in:
Mark Wielaard 2021-05-26 18:27:57 +02:00
parent 8beea897ae
commit f61a67cdd5
4 changed files with 12 additions and 302 deletions

1
.gitignore vendored
View File

@ -25,3 +25,4 @@
/elfutils-0.182.tar.bz2
/elfutils-0.183.tar.bz2
/elfutils-0.184.tar.bz2
/elfutils-0.185.tar.bz2

View File

@ -1,297 +0,0 @@
commit 1cfd294392762d2707f65ed3c5339cc364f47f1e (HEAD -> rawhide)
Author: Frank Ch. Eigler <fche@redhat.com>
Date: Fri May 14 18:37:30 2021 -0400
PR27859: correct 404-latch bug in debuginfod client reuse
PR27701 implemented curl handle reuse in debuginfod_client objects,
but with an unexpected bug. Server responses returning an error
"latched" because the curl_easy handles for error cases weren't all
systematically removed from the curl multi handle. This prevented
their proper re-addition the next time.
This version of the code simplfies matters by making only the curl
curl_multi handle long-lived. This turns out to be enough, because it
can maintain a pool of long-lived http/https connections and related
data, and lend them out to short-lived curl_easy handles. This mode
handles errors or hung downloads even better, because the easy handles
don't undergo complex state transitions between reuse.
A new test case confirms this correction via the federating debuginfod
instance (cleaning caches between subtests to make sure http* is being
used and reused).
Signed-off-by: Frank Ch. Eigler <fche@redhat.com>
diff --git a/debuginfod/ChangeLog b/debuginfod/ChangeLog
index 97f598f6287f..18e2361ffc5e 100644
--- a/debuginfod/ChangeLog
+++ b/debuginfod/ChangeLog
@@ -1,3 +1,11 @@
+2021-05-14 Frank Ch. Eigler <fche@redhat.com>
+
+ PR27859
+ * debuginfod-client.c (debuginfod_client): Retain only
+ long-lived multi handle from PR27701 work.
+ (debuginfo_begin,debuginfod_end): ctor/dtor for surviving field only.
+ (debuginfod_query_server): Rework to reuse multi handle only.
+
2021-05-04 Alice Zhang <alizhang@redhat.com>
* debuginfod-client.c (cache_miss_default_s): New static time_t,
diff --git a/debuginfod/debuginfod-client.c b/debuginfod/debuginfod-client.c
index 4fa047f5efdb..432d0cfef0ca 100644
--- a/debuginfod/debuginfod-client.c
+++ b/debuginfod/debuginfod-client.c
@@ -119,9 +119,8 @@ struct debuginfod_client
/* File descriptor to output any verbose messages if > 0. */
int verbose_fd;
- /* Count DEBUGINFOD_URLS elements and corresponding curl handles. */
- int num_urls;
- CURL **server_handles;
+ /* Maintain a long-lived curl multi-handle, which keeps a
+ connection/tls/dns cache to recently seen servers. */
CURLM *server_mhandle;
/* Can contain all other context, like cache_path, server_urls,
@@ -541,12 +540,6 @@ debuginfod_query_server (debuginfod_client *c,
/* Is there any server we can query? If not, don't do any work,
just return with ENOSYS. Don't even access the cache. */
- if (c->num_urls == 0)
- {
- rc = -ENOSYS;
- goto out;
- }
-
urls_envvar = getenv(DEBUGINFOD_URLS_ENV_VAR);
if (vfd >= 0)
dprintf (vfd, "server urls \"%s\"\n",
@@ -770,13 +763,20 @@ debuginfod_query_server (debuginfod_client *c,
goto out0;
}
+ /* Count number of URLs. */
+ int num_urls = 0;
+ for (int i = 0; server_urls[i] != '\0'; i++)
+ if (server_urls[i] != url_delim_char
+ && (i == 0 || server_urls[i - 1] == url_delim_char))
+ num_urls++;
+
CURLM *curlm = c->server_mhandle;
assert (curlm != NULL);
/* Tracks which handle should write to fd. Set to the first
handle that is ready to write the target file to the cache. */
CURL *target_handle = NULL;
- struct handle_data *data = malloc(sizeof(struct handle_data) * c->num_urls);
+ struct handle_data *data = malloc(sizeof(struct handle_data) * num_urls);
if (data == NULL)
{
rc = -ENOMEM;
@@ -786,7 +786,7 @@ debuginfod_query_server (debuginfod_client *c,
/* thereafter, goto out1 on error. */
/* Initialize handle_data with default values. */
- for (int i = 0; i < c->num_urls; i++)
+ for (int i = 0; i < num_urls; i++)
{
data[i].handle = NULL;
data[i].fd = -1;
@@ -797,23 +797,20 @@ debuginfod_query_server (debuginfod_client *c,
char *server_url = strtok_r(server_urls, url_delim, &strtok_saveptr);
/* Initialize each handle. */
- for (int i = 0; i < c->num_urls && server_url != NULL; i++)
+ for (int i = 0; i < num_urls && server_url != NULL; i++)
{
if (vfd >= 0)
dprintf (vfd, "init server %d %s\n", i, server_url);
data[i].fd = fd;
data[i].target_handle = &target_handle;
- data[i].handle = c->server_handles[i];
- assert (data[i].handle != NULL);
- curl_easy_reset(data[i].handle); // esp. previously sent http headers
- data[i].client = c;
-
+ data[i].handle = curl_easy_init();
if (data[i].handle == NULL)
{
rc = -ENETUNREACH;
goto out1;
}
+ data[i].client = c;
/* Build handle url. Tolerate both http://foo:999 and
http://foo:999/ forms */
@@ -869,7 +866,7 @@ debuginfod_query_server (debuginfod_client *c,
/* Query servers in parallel. */
if (vfd >= 0)
- dprintf (vfd, "query %d urls in parallel\n", c->num_urls);
+ dprintf (vfd, "query %d urls in parallel\n", num_urls);
int still_running;
long loops = 0;
int committed_to = -1;
@@ -882,7 +879,7 @@ debuginfod_query_server (debuginfod_client *c,
/* If the target file has been found, abort the other queries. */
if (target_handle != NULL)
{
- for (int i = 0; i < c->num_urls; i++)
+ for (int i = 0; i < num_urls; i++)
if (data[i].handle != target_handle)
curl_multi_remove_handle(curlm, data[i].handle);
else
@@ -979,7 +976,7 @@ debuginfod_query_server (debuginfod_client *c,
curl_easy_strerror (msg->data.result));
if (pnl)
c->default_progressfn_printed_p = 0;
- for (int i = 0; i < c->num_urls; i++)
+ for (int i = 0; i < num_urls; i++)
if (msg->easy_handle == data[i].handle)
{
if (strlen (data[i].errbuf) > 0)
@@ -1111,8 +1108,13 @@ debuginfod_query_server (debuginfod_client *c,
/* Perhaps we need not give up right away; could retry or something ... */
}
- curl_multi_remove_handle(curlm, verified_handle);
- assert (verified_handle == target_handle);
+ /* remove all handles from multi */
+ for (int i = 0; i < num_urls; i++)
+ {
+ curl_multi_remove_handle(curlm, data[i].handle); /* ok to repeat */
+ curl_easy_cleanup (data[i].handle);
+ }
+
free (data);
free (server_urls);
@@ -1126,6 +1128,13 @@ debuginfod_query_server (debuginfod_client *c,
/* error exits */
out1:
+ /* remove all handles from multi */
+ for (int i = 0; i < num_urls; i++)
+ {
+ curl_multi_remove_handle(curlm, data[i].handle); /* ok to repeat */
+ curl_easy_cleanup (data[i].handle);
+ }
+
unlink (target_cache_tmppath);
close (fd); /* before the rmdir, otherwise it'll fail */
(void) rmdir (target_cache_dir); /* nop if not empty */
@@ -1174,7 +1183,6 @@ debuginfod_begin (void)
{
debuginfod_client *client;
size_t size = sizeof (struct debuginfod_client);
- const char* server_urls = NULL;
client = (debuginfod_client *) calloc (1, size);
if (client != NULL)
@@ -1187,45 +1195,15 @@ debuginfod_begin (void)
client->verbose_fd = -1;
}
- /* Count the DEBUGINFOD_URLS and create the long-lived curl handles. */
- client->num_urls = 0;
- server_urls = getenv (DEBUGINFOD_URLS_ENV_VAR);
- if (server_urls != NULL)
- for (int i = 0; server_urls[i] != '\0'; i++)
- if (server_urls[i] != url_delim_char
- && (i == 0 || server_urls[i - 1] == url_delim_char))
- client->num_urls++;
-
- client->server_handles = calloc (client->num_urls, sizeof(CURL *));
- if (client->server_handles == NULL)
- goto out1;
-
- // allocate N curl easy handles
- for (int i=0; i<client->num_urls; i++)
- {
- client->server_handles[i] = curl_easy_init ();
- if (client->server_handles[i] == NULL)
- {
- for (i--; i >= 0; i--)
- curl_easy_cleanup (client->server_handles[i]);
- goto out2;
- }
- }
-
// allocate 1 curl multi handle
client->server_mhandle = curl_multi_init ();
if (client->server_mhandle == NULL)
- goto out3;
+ goto out1;
+
+ // extra future initialization
goto out;
- out3:
- for (int i=0; i<client->num_urls; i++)
- curl_easy_cleanup (client->server_handles[i]);
-
- out2:
- free (client->server_handles);
-
out1:
free (client);
client = NULL;
@@ -1259,10 +1237,6 @@ debuginfod_end (debuginfod_client *client)
if (client == NULL)
return;
- // assume that all the easy handles have already been removed from the multi handle
- for (int i=0; i<client->num_urls; i++)
- curl_easy_cleanup (client->server_handles[i]);
- free (client->server_handles);
curl_multi_cleanup (client->server_mhandle);
curl_slist_free_all (client->headers);
free (client->url);
diff --git a/tests/ChangeLog b/tests/ChangeLog
index 35fb2b2cc2f8..d6691b7702bc 100644
--- a/tests/ChangeLog
+++ b/tests/ChangeLog
@@ -1,3 +1,9 @@
+2021-05-14 Frank Ch. Eigler <fche@redhat.com>
+
+ PR27859
+ * run-debuginfod-find.sh: Test absence of 404-latch bug in client
+ curl handle reuse.
+
2021-05-04 Alice Zhang <alizhang@redhat.com>
* run-debuginfod-find.sh: Added tests for negative cache files.
diff --git a/tests/run-debuginfod-find.sh b/tests/run-debuginfod-find.sh
index 64b8290a119e..9183cccb7201 100755
--- a/tests/run-debuginfod-find.sh
+++ b/tests/run-debuginfod-find.sh
@@ -559,12 +559,24 @@ curl -s http://127.0.0.1:$PORT1/metrics | grep 'scanned_bytes_total'
# And generate a few errors into the second debuginfod's logs, for analysis just below
curl -s http://127.0.0.1:$PORT2/badapi > /dev/null || true
-curl -s http://127.0.0.1:$PORT2/buildid/deadbeef/debuginfo > /dev/null || true
+curl -s http://127.0.0.1:$PORT2/buildid/deadbeef/debuginfo > /dev/null || true
+# NB: this error is used to seed the 404 failure for the survive-404 tests
# Confirm bad artifact types are rejected without leaving trace
curl -s http://127.0.0.1:$PORT2/buildid/deadbeef/badtype > /dev/null || true
(curl -s http://127.0.0.1:$PORT2/metrics | grep 'badtype') && false
+# Confirm that reused curl connections survive 404 errors.
+# The rm's force an uncached fetch
+rm -f $DEBUGINFOD_CACHE_PATH/$BUILDID/debuginfo .client_cache*/$BUILDID/debuginfo
+testrun ${abs_top_builddir}/debuginfod/debuginfod-find debuginfo $BUILDID
+rm -f $DEBUGINFOD_CACHE_PATH/$BUILDID/debuginfo .client_cache*/$BUILDID/debuginfo
+testrun ${abs_top_builddir}/debuginfod/debuginfod-find debuginfo $BUILDID
+testrun ${abs_top_builddir}/debuginfod/debuginfod-find debuginfo $BUILDID
+testrun ${abs_top_builddir}/debuginfod/debuginfod-find debuginfo $BUILDID
+rm -f $DEBUGINFOD_CACHE_PATH/$BUILDID/debuginfo .client_cache*/$BUILDID/debuginfo
+testrun ${abs_top_builddir}/debuginfod/debuginfod-find debuginfo $BUILDID
+
# Confirm that some debuginfod client pools are being used
curl -s http://127.0.0.1:$PORT2/metrics | grep 'dc_pool_op.*reuse'

View File

@ -1,6 +1,6 @@
Name: elfutils
Version: 0.184
%global baserelease 5
Version: 0.185
%global baserelease 1
Release: %{baserelease}%{?dist}
URL: http://elfutils.org/
%global source_url ftp://sourceware.org/pub/elfutils/%{version}/
@ -62,7 +62,6 @@ BuildRequires: gettext-devel
%endif
# Patches
Patch1: elfutils-0.184-pr27859.patch
%description
Elfutils is a collection of utilities, including stack (to show
@ -232,7 +231,6 @@ such servers to download those files on demand.
%setup -q
# Apply patches
%patch1 -p1
autoreconf -f -v -i
@ -409,6 +407,14 @@ exit 0
%systemd_postun_with_restart debuginfod.service
%changelog
* Wed May 26 2021 Mark Wielaard <mjw@fedoraproject.org> - 0.185-1
- Upgrade to upstream 0.185
- debuginfod-client: Simplify curl handle reuse so downloads which
return an error are retried.
- elfcompress: Always exit with code 0 when the operation succeeds
(even when nothing was done). On error the exit code
is now always 1.
* Sun May 16 2021 Frank Ch. Eigler <fche@redhat.com> - 0.184-5
- Fix 404-latch problem on reused debuginfod_client. (PR27859)

View File

@ -1 +1 @@
SHA512 (elfutils-0.184.tar.bz2) = 5920c354edd6b28bb7e713d674bf6f2221e8ee91b90d582de39b1f45446aea1e7e3d4b6671024c391675b5e7530fe952c1320d730bb625c02d4c647ceb938bc8
SHA512 (elfutils-0.185.tar.bz2) = 34de0de1355b11740e036e0fc64f2fc063587c8eb121b19216ee5548d3f0f268d8fc3995176c47190466b9d881007cfa11a9d01e9a50e38af6119492bf8bb47f