Fix parsing of NYTimes Todays paper

JSON format of legacy collection changed.
This commit is contained in:
Kovid Goyal 2025-03-01 12:15:45 +05:30
parent 40d9236470
commit 357555689f
No known key found for this signature in database
GPG Key ID: 06BC317B515ACE7C
2 changed files with 62 additions and 57 deletions

View File

@ -76,7 +76,7 @@ def new_tag(soup, name, attrs=()):
class NewYorkTimes(BasicNewsRecipe):
title = 'The New York Times (Web)'
description = (
'New York Times (Web). You can edit the recipe to remove sections you are not interested in. '
'New York Times (Web). You can edit the recipe to remove sections you are not interested in. '
'Use advanced menu to make changes to fetch Todays Paper'
)
encoding = 'utf-8'
@ -192,8 +192,7 @@ class NewYorkTimes(BasicNewsRecipe):
data = json.loads(json_data.replace(':undefined', ':null'))['initialState']
containers, sections = {}, {}
article_map = {}
gc_pat = re.compile(r'groupings.(\d+).containers.(\d+)')
pat = re.compile(r'groupings.(\d+).containers.(\d+).relations.(\d+)')
sections = []
for key in data:
if 'Article' in key:
adata = data[key]
@ -201,36 +200,39 @@ class NewYorkTimes(BasicNewsRecipe):
url = adata.get('url')
summary = adata.get('summary')
headline = adata.get('headline')
if url and headline and 'id' in headline:
title = data[headline['id']]['default']
if url and headline:
title = headline['default']
article_map[adata['id']] = {
'title': title, 'url': url, 'description': summary or ''}
elif 'Legacy' in key:
sdata = data[key]
tname = sdata.get('__typename')
if tname == 'LegacyCollectionContainer':
m = gc_pat.search(key)
containers[int(m.group(2))] = sdata['label'] or sdata['name']
elif tname == 'LegacyCollectionRelation':
m = pat.search(key)
grouping, container, relation = map(int, m.groups())
asset = sdata['asset']
if asset and asset['typename'] == 'Article' and grouping == 0:
if container not in sections:
sections[container] = []
sections[container].append(asset['id'].split(':', 1)[1])
elif 'LegacyCollection:' in key:
lc = data[key]
if not lc.get('active'):
continue
for sdata in lc['groupings']:
tname = sdata.get('__typename')
if tname != 'LegacyCollectionGrouping':
continue
for cont in sdata['containers']:
if cont.get('__typename') == 'LegacyCollectionContainer':
section_name = cont['label@stripHtml']
articles = []
for rel in cont['relations']:
if rel.get('__typename') == 'LegacyCollectionRelation':
asset = rel['asset']['__ref']
if asset.startswith('Article:'):
articles.append(asset.partition(':')[2])
if articles:
sections.append((section_name, articles))
feeds = []
for container_num in sorted(containers):
section_title = containers[container_num]
if container_num in sections:
articles = sections[container_num]
if articles:
feeds.append((section_title, []))
for artid in articles:
if artid in article_map:
art = article_map[artid]
feeds[-1][1].append(art)
for section_title, article_ids in sections:
articles = []
for aid in article_ids:
if aid in article_map:
art = article_map[aid]
articles.append(art)
if articles:
feeds.append((section_title, articles))
def skey(x):
name = x[0].strip()

View File

@ -192,8 +192,7 @@ class NewYorkTimes(BasicNewsRecipe):
data = json.loads(json_data.replace(':undefined', ':null'))['initialState']
containers, sections = {}, {}
article_map = {}
gc_pat = re.compile(r'groupings.(\d+).containers.(\d+)')
pat = re.compile(r'groupings.(\d+).containers.(\d+).relations.(\d+)')
sections = []
for key in data:
if 'Article' in key:
adata = data[key]
@ -202,35 +201,38 @@ class NewYorkTimes(BasicNewsRecipe):
summary = adata.get('summary')
headline = adata.get('headline')
if url and headline:
title = data[headline['id']]['default']
title = headline['default']
article_map[adata['id']] = {
'title': title, 'url': url, 'description': summary or ''}
elif 'Legacy' in key:
sdata = data[key]
tname = sdata.get('__typename')
if tname == 'LegacyCollectionContainer':
m = gc_pat.search(key)
containers[int(m.group(2))] = sdata['label'] or sdata['name']
elif tname == 'LegacyCollectionRelation':
m = pat.search(key)
grouping, container, relation = map(int, m.groups())
asset = sdata['asset']
if asset and asset['typename'] == 'Article' and grouping == 0:
if container not in sections:
sections[container] = []
sections[container].append(asset['id'].split(':', 1)[1])
elif 'LegacyCollection:' in key:
lc = data[key]
if not lc.get('active'):
continue
for sdata in lc['groupings']:
tname = sdata.get('__typename')
if tname != 'LegacyCollectionGrouping':
continue
for cont in sdata['containers']:
if cont.get('__typename') == 'LegacyCollectionContainer':
section_name = cont['label@stripHtml']
articles = []
for rel in cont['relations']:
if rel.get('__typename') == 'LegacyCollectionRelation':
asset = rel['asset']['__ref']
if asset.startswith('Article:'):
articles.append(asset.partition(':')[2])
if articles:
sections.append((section_name, articles))
feeds = []
for container_num in sorted(containers):
section_title = containers[container_num]
if container_num in sections:
articles = sections[container_num]
if articles:
feeds.append((section_title, []))
for artid in articles:
if artid in article_map:
art = article_map[artid]
feeds[-1][1].append(art)
for section_title, article_ids in sections:
articles = []
for aid in article_ids:
if aid in article_map:
art = article_map[aid]
articles.append(art)
if articles:
feeds.append((section_title, articles))
def skey(x):
name = x[0].strip()
@ -368,5 +370,6 @@ class NewYorkTimes(BasicNewsRecipe):
def get_article_url(self, article):
url = BasicNewsRecipe.get_article_url(self, article)
if not re.search(r'/video/|/athletic/', url):
if not re.search(r'/video/|/athletic/|/card/', url):
return url
self.log('\tSkipping ', url)