From 97959c7eef26832e22228e74b6eaa82609e8158a Mon Sep 17 00:00:00 2001 From: unkn0w7n <51942695+unkn0w7n@users.noreply.github.com> Date: Mon, 6 Nov 2023 13:16:35 +0530 Subject: [PATCH 1/2] wsj - get cover --- recipes/wsj.recipe | 4 ++++ recipes/wsj_free.recipe | 4 ++++ 2 files changed, 8 insertions(+) diff --git a/recipes/wsj.recipe b/recipes/wsj.recipe index 8373ac6b0d..ceeb4fd08f 100644 --- a/recipes/wsj.recipe +++ b/recipes/wsj.recipe @@ -55,6 +55,10 @@ class WSJ(BasicNewsRecipe): .sub, em, i { color: #202020; } ''' + def get_cover_url(self): + soup = self.index_to_soup('https://www.frontpages.com/the-wall-street-journal/') + return 'https://www.frontpages.com' + soup.find('img', attrs={'id':'giornale-img'})['src'] + keep_only_tags = [ dict(name=['h1', 'h2']), dict(attrs={'aria-describedby':'big-top-caption'}), diff --git a/recipes/wsj_free.recipe b/recipes/wsj_free.recipe index 3e5770ba83..6b5a264c50 100644 --- a/recipes/wsj_free.recipe +++ b/recipes/wsj_free.recipe @@ -55,6 +55,10 @@ class WSJ(BasicNewsRecipe): .sub, em, i { color: #202020; } ''' + def get_cover_url(self): + soup = self.index_to_soup('https://www.frontpages.com/the-wall-street-journal/') + return 'https://www.frontpages.com' + soup.find('img', attrs={'id':'giornale-img'})['src'] + keep_only_tags = [ dict(name=['h1', 'h2']), dict(attrs={'aria-describedby':'big-top-caption'}), From 584f05e11c31976e15a75a8d154953ae8e30edc7 Mon Sep 17 00:00:00 2001 From: unkn0w7n <51942695+unkn0w7n@users.noreply.github.com> Date: Mon, 6 Nov 2023 13:18:52 +0530 Subject: [PATCH 2/2] FT cover --- recipes/financial_times.recipe | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/recipes/financial_times.recipe b/recipes/financial_times.recipe index 28534afcfb..aa9163e13d 100644 --- a/recipes/financial_times.recipe +++ b/recipes/financial_times.recipe @@ -44,6 +44,10 @@ class ft(BasicNewsRecipe): classes('in-article-advert') ] + def get_cover_url(self): + soup = self.index_to_soup('https://www.frontpages.com/financial-times/') + return 'https://www.frontpages.com' + soup.find('img', attrs={'id':'giornale-img'})['src'] + # needs_subscription = 'optional' # # def get_browser(self, *args, **kw):