On Wed, Mar 12, 2025 at 03:17:36PM +1000, Nicholas Piggin wrote: > Currently the fetch code does not fail gracefully when retry limit is > exceeded, it just falls through the loop with no file, which ends up > hitting other errors. > > In preparation for adding more cases where a download gets retried, > add an explicit check for retry limit exceeded. > > Signed-off-by: Nicholas Piggin <npig...@gmail.com> > --- > tests/functional/qemu_test/asset.py | 5 ++++- > 1 file changed, 4 insertions(+), 1 deletion(-) > > diff --git a/tests/functional/qemu_test/asset.py > b/tests/functional/qemu_test/asset.py > index f0730695f09..6a1c92ffbef 100644 > --- a/tests/functional/qemu_test/asset.py > +++ b/tests/functional/qemu_test/asset.py > @@ -116,7 +116,10 @@ def fetch(self): > self.log.info("Downloading %s to %s...", self.url, self.cache_file) > tmp_cache_file = self.cache_file.with_suffix(".download") > > - for retries in range(3): > + for retries in range(4): > + if retries == 3: > + raise Exception("Retries exceeded downloading %s", self.url)
While it works, it feels a bit wierd to me. Given the error retry scenario will unlink the file, I think it would be better todo if not os.path.exists(tmp_cache_file) raise Exception(...) immediately after the for() loop > + > try: > with tmp_cache_file.open("xb") as dst: > with urllib.request.urlopen(self.url) as resp: > -- > 2.47.1 > With regards, Daniel -- |: https://berrange.com -o- https://www.flickr.com/photos/dberrange :| |: https://libvirt.org -o- https://fstop138.berrange.com :| |: https://entangle-photo.org -o- https://www.instagram.com/dberrange :|