calibre/recipes/new_yorker.recipe
un-pogaz dba73b96b7 comment formating (manual)
ruff 'E114,E115,E116,E261,E262,E265'
2025-01-24 11:14:21 +01:00

146 lines
5.7 KiB
Python

#!/usr/bin/env python
# vim:fileencoding=utf-8
# License: GPLv3 Copyright: 2016, Kovid Goyal <kovid at kovidgoyal.net>
from __future__ import absolute_import, division, print_function, unicode_literals
from collections import defaultdict
from calibre import browser
from calibre.web.feeds.news import BasicNewsRecipe, classes, prefixed_classes
def absurl(x):
if x.startswith('/') and not x.startswith('//'):
x = 'https://www.newyorker.com' + x
return x
class NewYorker(BasicNewsRecipe):
title = 'The New Yorker Magazine'
description = "Articles of the week's New Yorker magazine"
language = 'en_US'
__author__ = 'Kovid Goyal'
no_stylesheets = True
timefmt = ' [%b %d]'
encoding = 'utf-8'
extra_css = '''
img { display:block; margin:0 auto; }
.rubric__name, .byline, time, .caption { font-size:small; }
.byline, .rubric__name { font-size:smaller; font-weight: bold;}
h3 { margin-bottom: 6px; }
.caption, .contributors { font-size: smaller; font-style: italic; font-weight: normal; }
.content-header__accreditation { font-style:italic; }
'''
keep_only_tags = [
classes(
'article__content-header article__body contributors'
),
]
remove_tags = [
classes(
'social-icons'
),
prefixed_classes('ConsentBannerWrapper- ResponsiveCartoonCTA-'),
dict(childtypes='iframe'),
dict(name='svg'),
]
remove_attributes = ['style']
def preprocess_html(self, soup):
w = '/w_320' # use '/w_640' for highres
for img in soup.findAll('img'):
if img.has_attr('srcset'):
for x in img['srcset'].split():
if w in x:
img['src'] = x
elif img.find_previous_sibling('source', attrs={'srcset':True}):
srcset = img.find_previous_sibling('source', attrs={'srcset':True})
for x in srcset['srcset'].split():
if w in x:
img['src'] = x
elif '/w_560' in x:
img['src'] = x
for src in soup.findAll('source'):
src.decompose()
for noscript in soup.findAll('noscript'):
noscript.name = 'div'
return soup
# def preprocess_image(self, img_data, image_url):
# from PIL import Image
# from calibre import fit_image
# from io import BytesIO
# img = Image.open(BytesIO(img_data)).convert('RGB')
# scaled, nwidth, nheight = fit_image(img.width, img.height, 1024, 1024)
# if scaled:
# img = img.resize((nwidth, nheight))
# buf = BytesIO()
# img.save(buf, format='JPEG')
# return buf.getvalue()
recipe_specific_options = {
'date': {
'short': 'The date of the edition to download (YYYY/MM/DD format)',
'long': 'For example, 2024/07/08'
}
}
def parse_index(self):
issue_url = 'https://www.newyorker.com/magazine?intcid=magazine'
d = self.recipe_specific_options.get('date')
if d and isinstance(d, str):
issue_url = 'https://www.newyorker.com/magazine/' + d
soup = self.index_to_soup(issue_url)
cover_img = soup.find('picture',
attrs={'class': lambda x: x and 'asset-embed__responsive-asset' in x})
if cover_img is not None:
self.cover_url = cover_img.img['src']
self.log('Found cover:', self.cover_url)
try:
# the src original resolution w_280 was too low, replace w_280 with w_560
cover_url_width_index = self.cover_url.find('w_')
old_width = self.cover_url[cover_url_width_index:cover_url_width_index+5]
self.cover_url = self.cover_url.replace(old_width, 'w_640')
except Exception:
self.log('Failed enlarging cover img, using the original one')
feeds_dict = defaultdict(list)
for section in soup.findAll('section',
attrs={'class': lambda x: x and 'SummaryRiverSection-' in x}):
for h2 in section.findAll(attrs={'class':lambda x: x and 'SectionTitleHed-' in x}):
secname = self.tag_to_string(h2)
self.log(secname)
for a in section.findAll('a', href=True, attrs={'class':lambda x: x and 'summary-item__hed-link' in x}):
section = secname
url = absurl(a['href'])
title = self.tag_to_string(a)
desc = ''
summ = a.find_next_sibling(attrs={'class':lambda x: x and 'summary-item__dek' in x})
if summ:
desc = self.tag_to_string(summ)
byl = a.find_next_sibling(attrs={'class':lambda x: x and 'summary-item__byline-' in x})
if byl:
desc = self.tag_to_string(byl) + ' | ' + desc
rub = a.find_previous_sibling(attrs={'class':lambda x: x and 'summary-item__rubric' in x})
if rub:
desc = self.tag_to_string(rub) + ' | ' + desc
self.log('\t', title, '\n\t', desc, '\n\t\t', url)
feeds_dict[section].append({'title': title, 'url': url, 'description': desc})
return feeds_dict.items()
# The New Yorker changes the content it delivers based on cookies, so the
# following ensures that we send no cookies
def get_browser(self, *args, **kwargs):
return self
def clone_browser(self, *args, **kwargs):
return self.get_browser()
def open_novisit(self, *args, **kwargs):
br = browser()
return br.open_novisit(*args, **kwargs)
open = open_novisit