mirror of
https://github.com/kovidgoyal/calibre.git
synced 2025-07-09 03:04:10 -04:00
Merge branch 'master' of https://github.com/unkn0w7n/calibre
This commit is contained in:
commit
3d5f70de36
Binary file not shown.
Before Width: | Height: | Size: 157 B After Width: | Height: | Size: 286 B |
BIN
recipes/icons/the_week_uk.png
Normal file
BIN
recipes/icons/the_week_uk.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 286 B |
@ -16,6 +16,7 @@ class MoneyControlRecipe(BasicNewsRecipe):
|
|||||||
ignore_duplicate_articles = {'title', 'url'}
|
ignore_duplicate_articles = {'title', 'url'}
|
||||||
remove_empty_feeds = True
|
remove_empty_feeds = True
|
||||||
resolve_internal_links = True
|
resolve_internal_links = True
|
||||||
|
oldest_article = 1 # days
|
||||||
|
|
||||||
extra_css = '''
|
extra_css = '''
|
||||||
img {display:block; margin:0 auto;}
|
img {display:block; margin:0 auto;}
|
||||||
@ -65,7 +66,7 @@ class MoneyControlRecipe(BasicNewsRecipe):
|
|||||||
|
|
||||||
feeds = []
|
feeds = []
|
||||||
|
|
||||||
when = 27 # hours
|
when = oldest_article*24
|
||||||
index = 'https://www.moneycontrol.com/'
|
index = 'https://www.moneycontrol.com/'
|
||||||
|
|
||||||
business_sections = [
|
business_sections = [
|
||||||
@ -73,12 +74,12 @@ class MoneyControlRecipe(BasicNewsRecipe):
|
|||||||
'personal-finance', 'commodities', 'trade', 'companies'
|
'personal-finance', 'commodities', 'trade', 'companies'
|
||||||
]
|
]
|
||||||
|
|
||||||
a = 'https://news.google.com/rss/search?q=when:{}h+allinurl:{}{}&hl=en-IN&gl=IN&ceid=IN:en'
|
a = 'https://news.google.com/rss/search?q=when:{}h+allinurl:{}&hl=en-IN&gl=IN&ceid=IN:en'
|
||||||
|
|
||||||
for sec in business_sections:
|
for sec in business_sections:
|
||||||
allinurl_a = index + 'news/business'
|
allinurl_a = index + 'news/business'
|
||||||
feeds.append((sec.capitalize(), a.format(when, quote(allinurl_a, safe=''), '%2F' + sec + '%2F')))
|
feeds.append((sec.capitalize(), a.format(when, quote(allinurl_a + sec, safe=''))))
|
||||||
feeds.append(('Business' , a.format(when, quote(allinurl_a, safe=''), '')))
|
feeds.append(('Business' , a.format(when, quote(allinurl_a + sec, safe=''))))
|
||||||
|
|
||||||
news_sections = [
|
news_sections = [
|
||||||
'india', 'world', 'opinion', 'politics', 'technology', 'trends', 'lifestyle'
|
'india', 'world', 'opinion', 'politics', 'technology', 'trends', 'lifestyle'
|
||||||
@ -86,8 +87,8 @@ class MoneyControlRecipe(BasicNewsRecipe):
|
|||||||
|
|
||||||
for sec in news_sections:
|
for sec in news_sections:
|
||||||
allinurl_b = index + 'news'
|
allinurl_b = index + 'news'
|
||||||
feeds.append((sec.capitalize(), a.format(when, quote(allinurl_b, safe=''), '%2F' + sec + '%2F')))
|
feeds.append((sec.capitalize(), a.format(when, quote(allinurl_b + sec, safe=''))))
|
||||||
feeds.append(('News', a.format(when, quote(allinurl_b, safe=''), '')))
|
feeds.append(('News', a.format(when, quote(allinurl_b + sec, safe=''), '')))
|
||||||
feeds.append(
|
feeds.append(
|
||||||
('Others', 'https://news.google.com/rss/search?q=when:{}h+allinurl:{}&hl=en-IN&gl=IN&ceid=IN:en'.format(when, quote(index, safe='')))
|
('Others', 'https://news.google.com/rss/search?q=when:{}h+allinurl:{}&hl=en-IN&gl=IN&ceid=IN:en'.format(when, quote(index, safe='')))
|
||||||
)
|
)
|
||||||
|
@ -1,27 +1,95 @@
|
|||||||
__license__ = 'GPL v3'
|
|
||||||
__copyright__ = '2010, JOlo'
|
|
||||||
'''
|
'''
|
||||||
www.theweek.com
|
www.theweek.com
|
||||||
'''
|
'''
|
||||||
|
from calibre.web.feeds.news import BasicNewsRecipe, classes
|
||||||
from calibre.web.feeds.news import BasicNewsRecipe
|
from urllib.parse import quote
|
||||||
|
|
||||||
|
|
||||||
class TheWeek(BasicNewsRecipe):
|
class TheWeek(BasicNewsRecipe):
|
||||||
title = 'TheWeek.com'
|
title = 'The Week'
|
||||||
__author__ = 'Jim Olo'
|
__author__ = 'unkn0wn'
|
||||||
description = "The best of the US and international media. Daily coverage of commentary and analysis of the day's events, as well as arts, entertainment, people and gossip, and political cartoons." # noqa
|
description = (
|
||||||
publisher = 'The Week Publications, Inc.'
|
'The Week is for readers who want to know what\'s going on in the world, without having to read '
|
||||||
masthead_url = 'http://test.theweek.com/images/logo_theweek.gif'
|
'several daily newspapers or get wrapped up in the endless news cycle. For every important story, '
|
||||||
cover_url = masthead_url
|
'our editors carefully select commentary from all sides of the debate and artfully stitch them together '
|
||||||
category = 'news, politics, USA'
|
'into one concise read. By showing you every perspective, we enable you to form your own opinion.'
|
||||||
oldest_article = 7
|
)
|
||||||
max_articles_per_feed = 100
|
language = 'en_US'
|
||||||
no_stylesheets = True
|
|
||||||
encoding = 'utf-8'
|
encoding = 'utf-8'
|
||||||
use_embedded_content = False
|
no_stylesheets = True
|
||||||
language = 'en'
|
remove_javascript = True
|
||||||
auto_cleanup = True
|
remove_attributes = ['width', 'height', 'style']
|
||||||
feeds = [
|
|
||||||
(u'Latest articles', u'http://theweek.com/rss.xml'),
|
ignore_duplicate_articles = {'title', 'url'}
|
||||||
|
remove_empty_feeds = True
|
||||||
|
resolve_internal_links = True
|
||||||
|
simultaneous_downloads = 1
|
||||||
|
oldest_article = 7 # days
|
||||||
|
web_url = ''
|
||||||
|
|
||||||
|
extra_css = '''
|
||||||
|
img {display:block; margin:0 auto;}
|
||||||
|
.caption__text--hero, .credit { font-size:small; text-align:center; }
|
||||||
|
.header__strapline, em, i { color:#202020; }
|
||||||
|
.article-type__breadcrumb { color:grey; }
|
||||||
|
.author-byline__author-text {font-size:small; }
|
||||||
|
'''
|
||||||
|
|
||||||
|
def get_cover_url(self):
|
||||||
|
import json
|
||||||
|
url = 'https://usmagazine.theweek.com/timelines.json'
|
||||||
|
data = json.loads(self.index_to_soup(url, raw=True))
|
||||||
|
for x in data['timelines'][:5]:
|
||||||
|
if '-cover-' in x['image']:
|
||||||
|
return 'https://usmagazine.theweek.com' + x['image'][1:]
|
||||||
|
|
||||||
|
articles_are_obfuscated = True
|
||||||
|
|
||||||
|
def get_obfuscated_article(self, url):
|
||||||
|
br = self.get_browser()
|
||||||
|
soup = self.index_to_soup(url)
|
||||||
|
link = soup.a['href']
|
||||||
|
skip_sections =[ # add sections you want to skip
|
||||||
|
'/video/', '/videos/', '/multimedia/',
|
||||||
|
]
|
||||||
|
if any(x in link for x in skip_sections):
|
||||||
|
self.abort_article('skipping video links ', link)
|
||||||
|
self.web_url = link
|
||||||
|
html = br.open(link).read()
|
||||||
|
return ({ 'data': html, 'url': link })
|
||||||
|
|
||||||
|
keep_only_tags = [
|
||||||
|
classes('article-type__breadcrumb header__title header__strapline image image--hero author-byline__author-text article__body')
|
||||||
]
|
]
|
||||||
|
|
||||||
|
remove_tags = [
|
||||||
|
dict(name='aside'),
|
||||||
|
classes(
|
||||||
|
'blueconic-article__wrapper ad-unit van_vid_carousel tag-links'
|
||||||
|
)
|
||||||
|
]
|
||||||
|
|
||||||
|
def preprocess_html(self, soup):
|
||||||
|
for img in soup.findAll('img', attrs={'data-pin-media':True}):
|
||||||
|
img['src'] = img['data-pin-media'].replace('.jpg', '-768-80.jpg')
|
||||||
|
return soup
|
||||||
|
|
||||||
|
feeds = []
|
||||||
|
when = oldest_article*24
|
||||||
|
index = 'https://theweek.com/'
|
||||||
|
sections = [
|
||||||
|
'politics', 'news', 'cartoons', 'tech', 'science', 'health',
|
||||||
|
'culture-life', 'business', 'travel', 'arts-life', 'history'
|
||||||
|
]
|
||||||
|
for sec in sections:
|
||||||
|
a = 'https://news.google.com/rss/search?q=when:{}h+allinurl:{}&hl=en-IN&gl=US&ceid=US:en'
|
||||||
|
feeds.append((sec.capitalize(), a.format(when, quote(index + sec, safe=''))))
|
||||||
|
feeds.append(('Others', a.format(when, quote(index, safe=''), '')))
|
||||||
|
|
||||||
|
def populate_article_metadata(self, article, soup, first):
|
||||||
|
article.title = article.title.replace(' - The Week', '')
|
||||||
|
desc = soup.find(**classes('header__strapline'))
|
||||||
|
if desc:
|
||||||
|
article.summary = self.tag_to_string(desc)
|
||||||
|
article.text_summary = article.summary
|
||||||
|
article.url = self.web_url
|
||||||
|
95
recipes/the_week_uk.recipe
Normal file
95
recipes/the_week_uk.recipe
Normal file
@ -0,0 +1,95 @@
|
|||||||
|
'''
|
||||||
|
www.theweek.com
|
||||||
|
'''
|
||||||
|
from calibre.web.feeds.news import BasicNewsRecipe, classes
|
||||||
|
from urllib.parse import quote
|
||||||
|
|
||||||
|
|
||||||
|
class TheWeek(BasicNewsRecipe):
|
||||||
|
title = 'The Week'
|
||||||
|
__author__ = 'unkn0wn'
|
||||||
|
description = (
|
||||||
|
'The Week is for readers who want to know what\'s going on in the world, without having to read '
|
||||||
|
'several daily newspapers or get wrapped up in the endless news cycle. For every important story, '
|
||||||
|
'our editors carefully select commentary from all sides of the debate and artfully stitch them together '
|
||||||
|
'into one concise read. By showing you every perspective, we enable you to form your own opinion.'
|
||||||
|
)
|
||||||
|
language = 'en_UK'
|
||||||
|
encoding = 'utf-8'
|
||||||
|
no_stylesheets = True
|
||||||
|
remove_javascript = True
|
||||||
|
remove_attributes = ['width', 'height', 'style']
|
||||||
|
|
||||||
|
ignore_duplicate_articles = {'title', 'url'}
|
||||||
|
remove_empty_feeds = True
|
||||||
|
resolve_internal_links = True
|
||||||
|
simultaneous_downloads = 1
|
||||||
|
oldest_article = 7 # days
|
||||||
|
web_url = ''
|
||||||
|
|
||||||
|
extra_css = '''
|
||||||
|
img {display:block; margin:0 auto;}
|
||||||
|
.caption__text--hero, .credit { font-size:small; text-align:center; }
|
||||||
|
.header__strapline, em, i { color:#202020; }
|
||||||
|
.article-type__breadcrumb { color:grey; }
|
||||||
|
.author-byline__author-text {font-size:small; }
|
||||||
|
'''
|
||||||
|
|
||||||
|
def get_cover_url(self):
|
||||||
|
import json
|
||||||
|
url = 'https://ukmagazine.theweek.com/timelines.json'
|
||||||
|
data = json.loads(self.index_to_soup(url, raw=True))
|
||||||
|
for x in data['timelines'][:5]:
|
||||||
|
if '-cover-' in x['image']:
|
||||||
|
return 'https://ukmagazine.theweek.com' + x['image'][1:]
|
||||||
|
|
||||||
|
articles_are_obfuscated = True
|
||||||
|
|
||||||
|
def get_obfuscated_article(self, url):
|
||||||
|
br = self.get_browser()
|
||||||
|
soup = self.index_to_soup(url)
|
||||||
|
link = soup.a['href']
|
||||||
|
skip_sections =[ # add sections you want to skip
|
||||||
|
'/video/', '/videos/', '/multimedia/',
|
||||||
|
]
|
||||||
|
if any(x in link for x in skip_sections):
|
||||||
|
self.abort_article('skipping video links ', link)
|
||||||
|
self.web_url = link
|
||||||
|
html = br.open(link).read()
|
||||||
|
return ({ 'data': html, 'url': link })
|
||||||
|
|
||||||
|
keep_only_tags = [
|
||||||
|
classes('article-type__breadcrumb header__title header__strapline image image--hero author-byline__author-text article__body')
|
||||||
|
]
|
||||||
|
|
||||||
|
remove_tags = [
|
||||||
|
dict(name='aside'),
|
||||||
|
classes(
|
||||||
|
'blueconic-article__wrapper ad-unit van_vid_carousel tag-links'
|
||||||
|
)
|
||||||
|
]
|
||||||
|
|
||||||
|
def preprocess_html(self, soup):
|
||||||
|
for img in soup.findAll('img', attrs={'data-pin-media':True}):
|
||||||
|
img['src'] = img['data-pin-media'].replace('.jpg', '-768-80.jpg')
|
||||||
|
return soup
|
||||||
|
|
||||||
|
feeds = []
|
||||||
|
when = oldest_article*24
|
||||||
|
index = 'https://theweek.com/'
|
||||||
|
sections = [
|
||||||
|
'politics', 'news', 'cartoons', 'tech', 'science', 'health',
|
||||||
|
'culture-life', 'business', 'travel', 'arts-life', 'history'
|
||||||
|
]
|
||||||
|
for sec in sections:
|
||||||
|
a = 'https://news.google.com/rss/search?q=when:{}h+allinurl:{}&hl=en-IN&gl=US&ceid=US:en'
|
||||||
|
feeds.append((sec.capitalize(), a.format(when, quote(index + sec, safe=''))))
|
||||||
|
feeds.append(('Others', a.format(when, quote(index, safe=''), '')))
|
||||||
|
|
||||||
|
def populate_article_metadata(self, article, soup, first):
|
||||||
|
article.title = article.title.replace(' - The Week', '')
|
||||||
|
desc = soup.find(**classes('header__strapline'))
|
||||||
|
if desc:
|
||||||
|
article.summary = self.tag_to_string(desc)
|
||||||
|
article.text_summary = article.summary
|
||||||
|
article.url = self.web_url
|
Loading…
x
Reference in New Issue
Block a user