mirror of
https://github.com/kovidgoyal/calibre.git
synced 2025-06-23 15:30:45 -04:00
59 lines
2.8 KiB
Plaintext
59 lines
2.8 KiB
Plaintext
from calibre.web.feeds.news import BasicNewsRecipe
|
|
from calibre.ebooks.BeautifulSoup import Comment
|
|
|
|
|
|
class LinuxJournal(BasicNewsRecipe):
|
|
title = u'Linux Journal'
|
|
__author__ = 'fenuks'
|
|
description = u'The monthly magazine of the Linux community, promoting the use of Linux worldwide.'
|
|
cover_url = 'http://www.linuxjournal.com/files/linuxjournal.com/ufiles/logo-lj.jpg'
|
|
category = 'IT, Linux'
|
|
language = 'en'
|
|
oldest_article = 7
|
|
max_articles_per_feed = 100
|
|
no_stylesheets = True
|
|
use_embedded_content = False
|
|
remove_empty_feeds = True
|
|
keep_only_tags = [dict(id='content-inner')]
|
|
remove_tags_after = dict(attrs={'class': 'user-signature clear-block'})
|
|
remove_tags = [dict(attrs={
|
|
'class': ['user-signature clear-block', 'breadcrumb', 'terms terms-inline']})]
|
|
feeds = [
|
|
(u'Front Page', u'http://feeds.feedburner.com/linuxjournalcom'),
|
|
(u'News', u'http://feeds.feedburner.com/LinuxJournal-BreakingNews'),
|
|
(u'Blogs', u'http://www.linuxjournal.com/blog/feed'),
|
|
(u'Audio/Video', u'http://www.linuxjournal.com/taxonomy/term/28/0/feed'),
|
|
(u'Community', u'http://www.linuxjournal.com/taxonomy/term/18/0/feed'),
|
|
(u'Education', u'http://www.linuxjournal.com/taxonomy/term/25/0/feed'),
|
|
(u'Embedded', u'http://www.linuxjournal.com/taxonomy/term/27/0/feed'),
|
|
(u'Hardware', u'http://www.linuxjournal.com/taxonomy/term/23/0/feed'),
|
|
(u'HOWTOs', u'http://www.linuxjournal.com/taxonomy/term/19/0/feed'),
|
|
(u'International', u'http://www.linuxjournal.com/taxonomy/term/30/0/feed'),
|
|
(u'Security', u'http://www.linuxjournal.com/taxonomy/term/31/0/feed'),
|
|
(u'Software', u'http://www.linuxjournal.com/taxonomy/term/17/0/feed'),
|
|
(u'Sysadmin', u'http://www.linuxjournal.com/taxonomy/term/21/0/feed'),
|
|
(u'Webmaster', u'http://www.linuxjournal.com/taxonomy/term/24/0/feed')]
|
|
|
|
def append_page(self, soup, appendtag):
|
|
next = appendtag.find('li', attrs={'class': 'pager-next'})
|
|
while next:
|
|
nexturl = next.a['href']
|
|
appendtag.find('div', attrs={'class': 'links'}).extract()
|
|
soup2 = self.index_to_soup('http://www.linuxjournal.com' + nexturl)
|
|
pagetext = soup2.find(
|
|
attrs={'class': 'node-inner'}).find(attrs={'class': 'content'})
|
|
next = appendtag.find('li', attrs={'class': 'pager-next'})
|
|
comments = pagetext.findAll(
|
|
text=lambda text: isinstance(text, Comment))
|
|
for comment in comments:
|
|
comment.extract()
|
|
pos = len(appendtag.contents)
|
|
appendtag.insert(pos, pagetext)
|
|
tag = appendtag.find('div', attrs={'class': 'links'})
|
|
if tag:
|
|
tag.extract()
|
|
|
|
def preprocess_html(self, soup):
|
|
self.append_page(soup, soup.body)
|
|
return soup
|