mirror of
https://github.com/kovidgoyal/calibre.git
synced 2025-09-29 15:31:08 -04:00
100 lines
3.5 KiB
Python
100 lines
3.5 KiB
Python
from datetime import datetime, timedelta, timezone
|
|
|
|
from calibre.utils.date import parse_date
|
|
from calibre.web.feeds.news import BasicNewsRecipe, classes
|
|
|
|
index = 'https://www.globaltimes.cn/'
|
|
|
|
class GlobalTimes(BasicNewsRecipe):
|
|
title = 'Global Times'
|
|
__author__ = 'unkn0wn'
|
|
description = 'DISCOVER CHINA, DISCOVER THE WORLD'
|
|
language = 'en_CN'
|
|
no_stylesheets = True
|
|
remove_attributes = ['height', 'width', 'style']
|
|
ignore_duplicate_articles = {'url'}
|
|
masthead_url = 'https://www.globaltimes.cn/img/logo1@3x.png'
|
|
encoding = 'utf-8'
|
|
remove_empty_feeds = True
|
|
resolve_internal_links = True
|
|
oldest_article = 1 # days
|
|
|
|
def get_cover_url(self):
|
|
soup = self.index_to_soup('https://en.kiosko.net/cn/np/cn_global_times.html')
|
|
return 'https:' + soup.find('img', attrs={'id':'portada'})['src']
|
|
|
|
extra_css = '''
|
|
.article_column {font-size:small; color:#404040;}
|
|
.author_share_left, .picture, .with_name_card, .pub_time {font-size:small; color:#202020;}
|
|
blockquote, em {color:#202020;}
|
|
'''
|
|
|
|
keep_only_tags = [
|
|
classes(
|
|
'article_column article_title author_share_left article_content'
|
|
)
|
|
]
|
|
remove_tags = [classes('author_card')]
|
|
|
|
def preprocess_raw_html(self, raw, *a):
|
|
return raw.replace('<br /><br />', '</p><p>').replace('<br><br>', '</p><p>')
|
|
|
|
def preprocess_html(self, soup):
|
|
h1 = soup.find(attrs={'class':'article_title'})
|
|
if h1:
|
|
h1.name = 'h1'
|
|
for div in soup.findAll(attrs={'class':'picture'}):
|
|
div.name = 'div'
|
|
p = soup.find(attrs={'class':'author_share_left'})
|
|
if p:
|
|
p.name = 'p'
|
|
return soup
|
|
|
|
def parse_index(self):
|
|
self.log(
|
|
'\n***\nif this recipe fails, report it on: '
|
|
'https://www.mobileread.com/forums/forumdisplay.php?f=228\n***\n'
|
|
)
|
|
sec_url = index + '{}/index.html'
|
|
|
|
section_list = [
|
|
'china', 'source', 'opinion', 'In-depth', 'world', 'life', 'sport', 'cartoon'
|
|
]
|
|
|
|
feeds = []
|
|
|
|
for section in section_list:
|
|
section_title = section.capitalize()
|
|
section_url = sec_url.format(section)
|
|
self.log(section_title, section_url)
|
|
soup = self.index_to_soup(section_url)
|
|
articles = self.articles_from_soup(soup)
|
|
if articles:
|
|
feeds.append((section_title, articles))
|
|
return feeds
|
|
|
|
def articles_from_soup(self, soup):
|
|
ans = []
|
|
dt = datetime.today().strftime('%Y%m')
|
|
for a in soup.findAll('a', attrs={'href':lambda x: x and x.startswith(index + 'page/' + dt + '/')}):
|
|
if a.find('img'):
|
|
continue
|
|
url = a['href']
|
|
title = self.tag_to_string(a).strip()
|
|
desc = ''
|
|
p = a.find_next_sibling('p')
|
|
if p:
|
|
desc = self.tag_to_string(p).strip()
|
|
src_time = a.find_next_sibling(attrs={'class':'source_time'})
|
|
if src_time:
|
|
time = self.tag_to_string(src_time).strip()
|
|
if '|' in time:
|
|
time = time.split('|')[1].strip()
|
|
date = parse_date(time)
|
|
today = (datetime.now(timezone.utc)).replace(microsecond=0)
|
|
if (today - date) > timedelta(self.oldest_article):
|
|
continue
|
|
self.log('\t', title, '\n\t', desc, '\n\t\t', url)
|
|
ans.append({'title': title, 'url': url, 'description': desc})
|
|
return ans
|