This commit is contained in:
Kovid Goyal 2023-09-05 18:29:22 +05:30
commit c7c7b08633
No known key found for this signature in database
GPG Key ID: 06BC317B515ACE7C
5 changed files with 362 additions and 16 deletions

BIN
recipes/icons/natgeohis.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 111 B

BIN
recipes/icons/natgeomag.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 111 B

View File

@ -25,45 +25,53 @@ def extract_json(raw):
def parse_contributors(grp): def parse_contributors(grp):
for item in grp: for item in grp:
line = '<p>' + escape(item['title']) + ' ' line = '<div class="auth">' + escape(item['title']) + ' '
for c in item['contributors']: for c in item['contributors']:
line += escape(c['displayName']) line += escape(c['displayName'])
yield line + '</p>' yield line + '</div>'
def parse_lead_image(media): def parse_lead_image(media):
yield '<div><img src="{}" alt="{}"></div>'.format( if 'dsc' in media['image']:
escape(media['image']['src'], True), escape(media['image']['dsc'], True)) yield '<div><img src="{}" alt="{}"></div>'.format(
yield '<p>' + escape(media['caption']) + '</p>' escape(media['image']['src'], True), escape(media['image']['dsc'], True))
else:
yield '<div><img src="{}"></div>'.format(escape(media['image']['src'], True))
if 'caption' in media:
yield '<div class="cap">' + media['caption'] + '</div>'
if 'credit' in media: if 'credit' in media:
yield '<p>' + escape(media['credit']) + '</p>' yield '<div class="cred">' + media['credit'] + '</div>'
def parse_body(item): def parse_body(item):
c = item['cntnt'] c = item['cntnt']
if item.get('type') == 'inline': if item.get('type') == 'inline':
if c.get('cmsType') == 'listicle': if c.get('cmsType') == 'listicle':
yield '<h3>' + escape(c['title']) + "</h3>" if 'title' in c:
yield '<h3>' + escape(c['title']) + "</h3>"
yield c['text'] yield c['text']
elif c.get('cmsType') == 'image': elif c.get('cmsType') == 'image':
for line in parse_lead_image(c): for line in parse_lead_image(c):
yield line yield line
else: else:
yield '<{tag}>{markup}</{tag}>'.format( if c['mrkup'].strip().startswith('<'):
tag=item['type'], markup=c['mrkup']) yield c['mrkup']
else:
yield '<{tag}>{markup}</{tag}>'.format(
tag=item['type'], markup=c['mrkup'])
def parse_article(edg): def parse_article(edg):
sc = edg['schma'] sc = edg['schma']
yield '<h3>' + escape(edg['sctn']) + '</h3>' yield '<h3 class="sub">' + escape(edg['sctn']) + '</h3>'
yield '<h1>' + escape(sc['sclTtl']) + '</h1>' yield '<h1>' + escape(sc['sclTtl']) + '</h1>'
yield '<div>' + escape(sc['sclDsc']) + '</div>' yield '<div class="byline">' + escape(sc['sclDsc']) + '</div><br>'
for line in parse_contributors(edg['cntrbGrp']): for line in parse_contributors(edg['cntrbGrp']):
yield line yield line
ts = parse_iso8601(edg['mdDt'], as_utc=False).strftime('%B %d, %Y') ts = parse_iso8601(edg['mdDt'], as_utc=False).strftime('%B %d, %Y')
yield '<p>Published: ' + escape(ts) + '</p>' yield '<div class="time">Published: ' + escape(ts) + '</div>'
if 'readTime' in edg: if 'readTime' in edg:
yield '<p>' + escape(edg['readTime']) + '</p>' yield '<div class="time">' + escape(edg['readTime']) + '</div><br>'
if edg.get('ldMda', {}).get('cmsType') == 'image': if edg.get('ldMda', {}).get('cmsType') == 'image':
for line in parse_lead_image(edg['ldMda']): for line in parse_lead_image(edg['ldMda']):
yield line yield line
@ -87,26 +95,54 @@ def article_parse(data):
class NatGeo(BasicNewsRecipe): class NatGeo(BasicNewsRecipe):
title = u'National Geographic' title = u'National Geographic'
description = 'Daily news articles from The National Geographic' description = 'News articles from The National Geographic, Download Monthly.'
language = 'en' language = 'en'
encoding = 'utf8' encoding = 'utf8'
publisher = 'nationalgeographic.com' publisher = 'nationalgeographic.com'
category = 'science, nat geo' category = 'science, nat geo'
__author__ = 'Kovid Goyal' __author__ = 'Kovid Goyal, unkn0wn'
description = 'Inspiring people to care about the planet since 1888' description = 'Inspiring people to care about the planet since 1888'
timefmt = ' [%a, %d %b, %Y]' timefmt = ' [%a, %d %b, %Y]'
no_stylesheets = True no_stylesheets = True
use_embedded_content = False use_embedded_content = False
remove_attributes = ['style'] remove_attributes = ['style']
remove_javascript = False remove_javascript = False
masthead_url = 'https://i.natgeofe.com/n/e76f5368-6797-4794-b7f6-8d757c79ea5c/ng-logo-2fl.png?w=600&h=600'
extra_css = '''
.sub { color:#404040; }
.byline, i { font-style:italic; color:#202020; }
.cap {text-align:center; font-size:small; }
.cred {text-align:center; font-size:small; color:#404040; }
.auth, .time { font-size:small; color:#5c5c5c; }
'''
def parse_index(self): def parse_index(self):
soup = self.index_to_soup('https://www.nationalgeographic.com/latest-stories/') pages = [
'https://www.nationalgeographic.com/animals',
'https://www.nationalgeographic.com/environment',
'https://www.nationalgeographic.com/history',
'https://www.nationalgeographic.com/science',
'https://www.nationalgeographic.com/travel'
]
feeds = []
for sec in pages:
soup = self.index_to_soup(sec)
parsed = self.articles_from_soup(soup)
if parsed:
feeds += parsed
return feeds
def articles_from_soup(self, soup):
ans = {} ans = {}
for article in soup.findAll('article'): for article in soup.findAll('article'):
a = article.find('a') a = article.find('a')
url = a['href'] url = a['href']
section = self.tag_to_string(article.find(**classes('SectionLabel'))) section = self.tag_to_string(article.find(**classes('SectionLabel')))
if section.startswith('Paid Content'):
continue
title = self.tag_to_string(article.find(**classes('PromoTile__Title--truncated'))) title = self.tag_to_string(article.find(**classes('PromoTile__Title--truncated')))
articles = ans.setdefault(section, []) articles = ans.setdefault(section, [])
articles.append({'title': title, 'url': url}) articles.append({'title': title, 'url': url})
@ -116,3 +152,8 @@ class NatGeo(BasicNewsRecipe):
def preprocess_raw_html(self, raw_html, url): def preprocess_raw_html(self, raw_html, url):
data = extract_json(raw_html) data = extract_json(raw_html)
return '\n'.join(article_parse(data)) return '\n'.join(article_parse(data))
def preprocess_html(self, soup):
for img in soup.findAll('img', src=True):
img['src'] = img['src'] + '?w=700&h=700'
return soup

146
recipes/natgeohis.recipe Normal file
View File

@ -0,0 +1,146 @@
#!/usr/bin/env python
# vim:fileencoding=utf-8
from __future__ import absolute_import, division, print_function, unicode_literals
import json
from pprint import pformat
from calibre.web.feeds.news import BasicNewsRecipe
from calibre import prepare_string_for_xml as escape
from calibre.utils.iso8601 import parse_iso8601
def classes(classes):
q = frozenset(classes.split(' '))
return dict(attrs={
'class': lambda x: x and frozenset(x.split()).intersection(q)})
def extract_json(raw):
s = raw.find("window['__natgeo__']")
script = raw[s:raw.find('</script>', s)]
return json.loads(
script[script.find('{'):].rstrip(';'))['page']['content']['article']
def parse_contributors(grp):
for item in grp:
line = '<div class="auth">' + escape(item['title']) + ' '
for c in item['contributors']:
line += escape(c['displayName'])
yield line + '</div>'
def parse_lead_image(media):
if 'dsc' in media['image']:
yield '<div><img src="{}" alt="{}"></div>'.format(
escape(media['image']['src'], True), escape(media['image']['dsc'], True))
else:
yield '<div><img src="{}"></div>'.format(escape(media['image']['src'], True))
if 'caption' in media:
yield '<div class="cap">' + media['caption'] + '</div>'
if 'credit' in media:
yield '<div class="cred">' + media['credit'] + '</div>'
def parse_body(item):
c = item['cntnt']
if item.get('type') == 'inline':
if c.get('cmsType') == 'listicle':
if 'title' in c:
yield '<h3>' + escape(c['title']) + "</h3>"
yield c['text']
elif c.get('cmsType') == 'image':
for line in parse_lead_image(c):
yield line
else:
if c['mrkup'].strip().startswith('<'):
yield c['mrkup']
else:
yield '<{tag}>{markup}</{tag}>'.format(
tag=item['type'], markup=c['mrkup'])
def parse_article(edg):
sc = edg['schma']
yield '<h3 class="sub">' + escape(edg['sctn']) + '</h3>'
yield '<h1>' + escape(sc['sclTtl']) + '</h1>'
yield '<div class="byline">' + escape(sc['sclDsc']) + '</div><br>'
for line in parse_contributors(edg['cntrbGrp']):
yield line
ts = parse_iso8601(edg['mdDt'], as_utc=False).strftime('%B %d, %Y')
yield '<div class="time">Published: ' + escape(ts) + '</div>'
if 'readTime' in edg:
yield '<div class="time">' + escape(edg['readTime']) + '</div><br>'
if edg.get('ldMda', {}).get('cmsType') == 'image':
for line in parse_lead_image(edg['ldMda']):
yield line
for item in edg['bdy']:
for line in parse_body(item):
yield line
def article_parse(data):
yield "<html><body>"
for frm in data['frms']:
if not frm:
continue
for mod in frm.get('mods', ()):
for edg in mod.get('edgs', ()):
if edg.get('cmsType') == 'ArticleBodyTile':
for line in parse_article(edg):
yield line
yield "</body></html>"
class NatGeo(BasicNewsRecipe):
title = u'National Geographic History'
description = (
'From Caesar to Napoleon, the Pyramids to the Parthenon, the Trojan War to the Civil War—National Geographic '
'HISTORY draws readers in with more than 5,000 years of people, places, and things to explore.'
)
language = 'en'
encoding = 'utf8'
publisher = 'nationalgeographic.com'
category = 'science, nat geo'
__author__ = 'Kovid Goyal'
description = 'Inspiring people to care about the planet since 1888'
timefmt = ' [%a, %d %b, %Y]'
no_stylesheets = True
use_embedded_content = False
remove_attributes = ['style']
remove_javascript = False
masthead_url = 'https://i.natgeofe.com/n/e76f5368-6797-4794-b7f6-8d757c79ea5c/ng-logo-2fl.png?w=600&h=600'
extra_css = '''
.sub { color:#404040; }
.byline, i { font-style:italic; color:#202020; }
.cap {text-align:center; font-size:small; }
.cred {text-align:center; font-size:small; color:#404040; }
.auth, .time { font-size:small; color:#5c5c5c; }
'''
def get_cover_url(self):
soup = self.index_to_soup('https://ngsingleissues.nationalgeographic.com/history')
wrap = soup.find(attrs={'class':'product-image-wrapper'})
return wrap.img['src']
def parse_index(self):
soup = self.index_to_soup('https://www.nationalgeographic.com/history/history-magazine')
ans = []
for article in soup.findAll('article'):
a = article.find('a')
url = a['href']
title = self.tag_to_string(article.find(**classes('PromoTile__Title--truncated')))
ans.append({'title': title, 'url': url})
self.log(title, ' ', url)
return [('Articles', ans)]
def preprocess_raw_html(self, raw_html, url):
data = extract_json(raw_html)
return '\n'.join(article_parse(data))
def preprocess_html(self, soup):
for img in soup.findAll('img', src=True):
img['src'] = img['src'] + '?w=1000&h=1000'
return soup

159
recipes/natgeomag.recipe Normal file
View File

@ -0,0 +1,159 @@
#!/usr/bin/env python
# vim:fileencoding=utf-8
from __future__ import absolute_import, division, print_function, unicode_literals
import json, re
from pprint import pformat
from calibre.web.feeds.news import BasicNewsRecipe
from calibre import prepare_string_for_xml as escape
from calibre.utils.iso8601 import parse_iso8601
def classes(classes):
q = frozenset(classes.split(' '))
return dict(attrs={
'class': lambda x: x and frozenset(x.split()).intersection(q)})
def extract_json(raw):
s = raw.find("window['__natgeo__']")
script = raw[s:raw.find('</script>', s)]
return json.loads(
script[script.find('{'):].rstrip(';'))['page']['content']['article']
def parse_contributors(grp):
for item in grp:
line = '<div class="auth">' + escape(item['title']) + ' '
for c in item['contributors']:
line += escape(c['displayName'])
yield line + '</div>'
def parse_lead_image(media):
if 'dsc' in media['image']:
yield '<div><img src="{}" alt="{}"></div>'.format(
escape(media['image']['src'], True), escape(media['image']['dsc'], True))
else:
yield '<div><img src="{}"></div>'.format(escape(media['image']['src'], True))
if 'caption' in media:
yield '<div class="cap">' + media['caption'] + '</div>'
if 'credit' in media:
yield '<div class="cred">' + media['credit'] + '</div>'
def parse_body(item):
c = item['cntnt']
if item.get('type') == 'inline':
if c.get('cmsType') == 'listicle':
if 'title' in c:
yield '<h3>' + escape(c['title']) + "</h3>"
yield c['text']
elif c.get('cmsType') == 'image':
for line in parse_lead_image(c):
yield line
else:
if c['mrkup'].strip().startswith('<'):
yield c['mrkup']
else:
yield '<{tag}>{markup}</{tag}>'.format(
tag=item['type'], markup=c['mrkup'])
def parse_article(edg):
sc = edg['schma']
yield '<h3 class="sub">' + escape(edg['sctn']) + '</h3>'
yield '<h1>' + escape(sc['sclTtl']) + '</h1>'
yield '<div class="byline">' + escape(sc['sclDsc']) + '</div><br>'
for line in parse_contributors(edg['cntrbGrp']):
yield line
ts = parse_iso8601(edg['mdDt'], as_utc=False).strftime('%B %d, %Y')
yield '<div class="time">Published: ' + escape(ts) + '</div>'
if 'readTime' in edg:
yield '<div class="time">' + escape(edg['readTime']) + '</div><br>'
if edg.get('ldMda', {}).get('cmsType') == 'image':
for line in parse_lead_image(edg['ldMda']):
yield line
for item in edg['bdy']:
for line in parse_body(item):
yield line
def article_parse(data):
yield "<html><body>"
for frm in data['frms']:
if not frm:
continue
for mod in frm.get('mods', ()):
for edg in mod.get('edgs', ()):
if edg.get('cmsType') == 'ArticleBodyTile':
for line in parse_article(edg):
yield line
yield "</body></html>"
class NatGeo(BasicNewsRecipe):
title = u'National Geographic Magazine'
description = 'The National Geographic, an American monthly magazine'
language = 'en'
encoding = 'utf8'
publisher = 'nationalgeographic.com'
category = 'science, nat geo'
__author__ = 'Kovid Goyal, unkn0wn'
description = 'Inspiring people to care about the planet since 1888'
timefmt = ' [%a, %d %b, %Y]'
no_stylesheets = True
use_embedded_content = False
remove_attributes = ['style']
remove_javascript = False
masthead_url = 'https://i.natgeofe.com/n/e76f5368-6797-4794-b7f6-8d757c79ea5c/ng-logo-2fl.png?w=600&h=600'
extra_css = '''
.sub { color:#404040; }
.byline, i { font-style:italic; color:#202020; }
.cap {text-align:center; font-size:small; }
.cred {text-align:center; font-size:small; color:#404040; }
.auth, .time { font-size:small; color:#5c5c5c; }
'''
def parse_index(self):
issues = self.index_to_soup('https://www.nationalgeographic.com/magazine')
mag = issues.find('a', attrs={'href':lambda x: x and x.startswith(
'https://www.nationalgeographic.com/magazine/issue/'
)})
self.timefmt = ' [' + self.tag_to_string(mag).replace(' Issue', '') + ']'
soup = self.index_to_soup(mag['href'])
png = re.findall('https://i\.natgeofe\.com\S+?national-geographic-magazine-\S+?\.jpg', soup.decode('utf-8'))
self.cover_url = png[0] + '?w=1000&h=1000'
name = soup.find(attrs={'class':lambda x: x and 'Header__Description' in x.split()})
self.title = 'National Geographic ' + self.tag_to_string(name)
ans = {}
ans2 = None
if photoart := soup.find(attrs={'class':lambda x: x and 'BgImagePromo__Container__Text__Link' in x.split()}):
ans2 = []
title = self.tag_to_string(photoart)
url = 'https://www.nationalgeographic.com' + photoart['href']
ans2.append(('Photo Essay', [{'title': title, 'url': url}]))
for gird in soup.findAll(attrs={'class':'GridPromoTile'}):
for article in soup.findAll('article'):
a = article.find('a')
url = a['href']
section = self.tag_to_string(article.find(**classes('SectionLabel')))
title = self.tag_to_string(article.find(**classes('PromoTile__Title--truncated')))
articles = ans.setdefault(section, [])
articles.append({'title': title, 'url': url})
self.log(pformat(ans))
if ans2:
return list(ans.items()) + ans2
return list(ans.items())
def preprocess_raw_html(self, raw_html, url):
data = extract_json(raw_html)
return '\n'.join(article_parse(data))
def preprocess_html(self, soup):
for img in soup.findAll('img', src=True):
img['src'] = img['src'] + '?w=1000&h=1000'
return soup