From fd2dc15a15951eba45fca5390b0983550b876b0c Mon Sep 17 00:00:00 2001 From: zeskeertwee <55278264+zeskeertwee@users.noreply.github.com> Date: Wed, 10 Jul 2024 18:49:13 +0200 Subject: [PATCH] fix: Follow redirects during scraping (#3875) --- mealie/services/scraper/scraper_strategies.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/mealie/services/scraper/scraper_strategies.py b/mealie/services/scraper/scraper_strategies.py index 95ea2513be6c..752a88e2419b 100644 --- a/mealie/services/scraper/scraper_strategies.py +++ b/mealie/services/scraper/scraper_strategies.py @@ -44,7 +44,9 @@ async def safe_scrape_html(url: str) -> str: """ async with AsyncClient(transport=safehttp.AsyncSafeTransport()) as client: html_bytes = b"" - async with client.stream("GET", url, timeout=SCRAPER_TIMEOUT, headers={"User-Agent": _FIREFOX_UA}) as resp: + async with client.stream( + "GET", url, timeout=SCRAPER_TIMEOUT, headers={"User-Agent": _FIREFOX_UA}, follow_redirects=True + ) as resp: start_time = time.time() async for chunk in resp.aiter_bytes(chunk_size=1024):