mirror of
https://github.com/kovidgoyal/calibre.git
synced 2025-07-07 10:14:46 -04:00
Formula AS by Silviu Cotoara
This commit is contained in:
parent
64966c6a99
commit
dbc74fa064
50
recipes/formulaas.recipe
Normal file
50
recipes/formulaas.recipe
Normal file
@ -0,0 +1,50 @@
|
|||||||
|
# -*- coding: utf-8 -*-
|
||||||
|
#!/usr/bin/env python
|
||||||
|
|
||||||
|
__license__ = 'GPL v3'
|
||||||
|
__copyright__ = u'2011, Silviu Cotoar\u0103'
|
||||||
|
'''
|
||||||
|
formula-as.ro
|
||||||
|
'''
|
||||||
|
|
||||||
|
from calibre.web.feeds.news import BasicNewsRecipe
|
||||||
|
|
||||||
|
class FormulaAS(BasicNewsRecipe):
|
||||||
|
title = u'Formula AS'
|
||||||
|
__author__ = u'Silviu Cotoar\u0103'
|
||||||
|
publisher = u'Formula AS'
|
||||||
|
description = u'Formula AS'
|
||||||
|
oldest_article = 5
|
||||||
|
language = 'ro'
|
||||||
|
max_articles_per_feed = 100
|
||||||
|
no_stylesheets = True
|
||||||
|
use_embedded_content = False
|
||||||
|
category = 'Ziare,Romania'
|
||||||
|
encoding = 'utf-8'
|
||||||
|
cover_url = 'http://www.formula-as.ro/_client/img/header_logo.png'
|
||||||
|
|
||||||
|
conversion_options = {
|
||||||
|
'comments' : description
|
||||||
|
,'tags' : category
|
||||||
|
,'language' : language
|
||||||
|
,'publisher' : publisher
|
||||||
|
}
|
||||||
|
|
||||||
|
keep_only_tags = [
|
||||||
|
dict(name='div', attrs={'class':'item padded'})
|
||||||
|
]
|
||||||
|
|
||||||
|
remove_tags = [
|
||||||
|
dict(name='ul', attrs={'class':'subtitle lower'})
|
||||||
|
]
|
||||||
|
|
||||||
|
remove_tags_after = [
|
||||||
|
dict(name='ul', attrs={'class':'subtitle lower'}),
|
||||||
|
dict(name='div', attrs={'class':'item-brief-options'})
|
||||||
|
]
|
||||||
|
feeds = [
|
||||||
|
(u'\u0218tiri', u'http://www.formula-as.ro/rss/articole.xml')
|
||||||
|
]
|
||||||
|
|
||||||
|
def preprocess_html(self, soup):
|
||||||
|
return self.adeify_images(soup)
|
BIN
recipes/icons/formulaas.png
Normal file
BIN
recipes/icons/formulaas.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 687 B |
@ -1,27 +1,26 @@
|
|||||||
# adapted from old recipe by Darko Miletic <darko.miletic at gmail.com>
|
# adapted from old recipe by Darko Miletic <darko.miletic at gmail.com>
|
||||||
|
|
||||||
import string, re
|
import re
|
||||||
from calibre import strftime
|
|
||||||
from calibre.web.feeds.recipes import BasicNewsRecipe
|
from calibre.web.feeds.recipes import BasicNewsRecipe
|
||||||
from calibre.ebooks.BeautifulSoup import BeautifulSoup, Tag, NavigableString
|
from calibre.ebooks.BeautifulSoup import Tag, NavigableString
|
||||||
|
|
||||||
|
|
||||||
class TheIndependentNew(BasicNewsRecipe):
|
class TheIndependentNew(BasicNewsRecipe):
|
||||||
|
|
||||||
# flag to enable/disable article graphics on business pages/some others
|
# flag to enable/disable article graphics on business pages/some others
|
||||||
# eg http://www.independent.co.uk/news/world/europe/berlusconi-departure-fails-to-calm-the-markets-6259682.html
|
# eg http://www.independent.co.uk/news/world/europe/berlusconi-departure-fails-to-calm-the-markets-6259682.html
|
||||||
# -max dimensions can be altered using the .pictureContainer img selector in the css
|
# -max dimensions can be altered using the .pictureContainer img selector in the css
|
||||||
_FETCH_ARTICLE_GRAPHICS = True
|
_FETCH_ARTICLE_GRAPHICS = True
|
||||||
|
|
||||||
#Flag to enable/disable image fetching (not business)
|
#Flag to enable/disable image fetching (not business)
|
||||||
_FETCH_IMAGES = True
|
_FETCH_IMAGES = True
|
||||||
|
|
||||||
|
|
||||||
#used for converting rating to stars
|
#used for converting rating to stars
|
||||||
_STAR_URL = 'http://www.independent.co.uk/skins/ind/images/rating_star.png'
|
_STAR_URL = 'http://www.independent.co.uk/skins/ind/images/rating_star.png'
|
||||||
_NO_STAR_URL = 'http://www.independent.co.uk/skins/ind/images/rating_star_grey.png'
|
_NO_STAR_URL = 'http://www.independent.co.uk/skins/ind/images/rating_star_grey.png'
|
||||||
|
|
||||||
|
|
||||||
title = u'The Independent'
|
title = u'The Independent'
|
||||||
__author__ = 'Will'
|
__author__ = 'Will'
|
||||||
description = 'The latest in UK News and World News from The \
|
description = 'The latest in UK News and World News from The \
|
||||||
@ -42,26 +41,26 @@ class TheIndependentNew(BasicNewsRecipe):
|
|||||||
dict(attrs={'id' : ['RelatedArtTag','renderBiography']}),
|
dict(attrs={'id' : ['RelatedArtTag','renderBiography']}),
|
||||||
dict(attrs={'class' : ['autoplay','openBiogPopup']})
|
dict(attrs={'class' : ['autoplay','openBiogPopup']})
|
||||||
]
|
]
|
||||||
|
|
||||||
keep_only_tags =[dict(attrs={'id':'main'})]
|
keep_only_tags =[dict(attrs={'id':'main'})]
|
||||||
recursions = 0
|
recursions = 0
|
||||||
|
|
||||||
# fixes non compliant html nesting and 'marks' article graphics links
|
# fixes non compliant html nesting and 'marks' article graphics links
|
||||||
preprocess_regexps = [
|
preprocess_regexps = [
|
||||||
(re.compile('<span class="storyTop ">(?P<nested>.*?)</span>', re.DOTALL),
|
(re.compile('<span class="storyTop ">(?P<nested>.*?)</span>', re.DOTALL),
|
||||||
lambda match: '<div class="storyTop">' + match.group('nested') + '</div>'),
|
lambda match: '<div class="storyTop">' + match.group('nested') + '</div>'),
|
||||||
(re.compile('(<strong>.*?[Cc]lick.*?<a.*?((HERE)|([Hh]ere)).*?</strong>)', re.DOTALL),
|
(re.compile('(<strong>.*?[Cc]lick.*?<a.*?((HERE)|([Hh]ere)).*?</strong>)', re.DOTALL),
|
||||||
lambda match: '<div class="article-graphic">' + match.group(0) + '</div>'),
|
lambda match: '<div class="article-graphic">' + match.group(0) + '</div>'),
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
conversion_options = {
|
conversion_options = {
|
||||||
'comment' : description
|
'comment' : description
|
||||||
, 'tags' : category
|
, 'tags' : category
|
||||||
, 'publisher' : publisher
|
, 'publisher' : publisher
|
||||||
, 'language' : language
|
, 'language' : language
|
||||||
}
|
}
|
||||||
|
|
||||||
extra_css = """
|
extra_css = """
|
||||||
h1{font-family: Georgia,serif }
|
h1{font-family: Georgia,serif }
|
||||||
body{font-family: Verdana,Arial,Helvetica,sans-serif}
|
body{font-family: Verdana,Arial,Helvetica,sans-serif}
|
||||||
@ -81,22 +80,22 @@ class TheIndependentNew(BasicNewsRecipe):
|
|||||||
.articleContent {display: block; clear:left;}
|
.articleContent {display: block; clear:left;}
|
||||||
.storyTop{}
|
.storyTop{}
|
||||||
.pictureContainer img { max-width: 400px; max-height: 400px;}
|
.pictureContainer img { max-width: 400px; max-height: 400px;}
|
||||||
"""
|
"""
|
||||||
|
|
||||||
oldest_article = 1
|
oldest_article = 1
|
||||||
max_articles_per_feed = 100
|
max_articles_per_feed = 100
|
||||||
|
|
||||||
_processed_urls = []
|
_processed_urls = []
|
||||||
|
|
||||||
|
|
||||||
def get_article_url(self, article):
|
def get_article_url(self, article):
|
||||||
url = super(self.__class__,self).get_article_url(article)
|
url = super(self.__class__,self).get_article_url(article)
|
||||||
|
|
||||||
title = article.get('title', None)
|
title = article.get('title', None)
|
||||||
if title and re.search("^Video:",title):
|
if title and re.search("^Video:",title):
|
||||||
return None
|
return None
|
||||||
|
|
||||||
#remove duplicates
|
#remove duplicates
|
||||||
if not (url in self._processed_urls):
|
if not (url in self._processed_urls):
|
||||||
self._processed_urls.append(url)
|
self._processed_urls.append(url)
|
||||||
else:
|
else:
|
||||||
@ -104,101 +103,101 @@ class TheIndependentNew(BasicNewsRecipe):
|
|||||||
return url
|
return url
|
||||||
|
|
||||||
def preprocess_html(self, soup):
|
def preprocess_html(self, soup):
|
||||||
|
|
||||||
#remove 'advertorial articles'
|
#remove 'advertorial articles'
|
||||||
strapline = soup.find('div',attrs={'class' : re.compile('.*strapLine.*')})
|
strapline = soup.find('div',attrs={'class' : re.compile('.*strapLine.*')})
|
||||||
if strapline:
|
if strapline:
|
||||||
for para in strapline.findAll('p'):
|
for para in strapline.findAll('p'):
|
||||||
if len(para.contents) and isinstance(para.contents[0],NavigableString) \
|
if len(para.contents) and isinstance(para.contents[0],NavigableString) \
|
||||||
and para.contents[0] == 'ADVERTORIAL FEATURE':
|
and para.contents[0] == 'ADVERTORIAL FEATURE':
|
||||||
return None
|
return None
|
||||||
|
|
||||||
items_to_extract = []
|
items_to_extract = []
|
||||||
|
|
||||||
for item in soup.findAll(attrs={'class' : re.compile("widget.*")}):
|
for item in soup.findAll(attrs={'class' : re.compile("widget.*")}):
|
||||||
remove = True
|
remove = True
|
||||||
pattern = re.compile('((articleContent)|(title))$')
|
pattern = re.compile('((articleContent)|(title))$')
|
||||||
if (pattern.search(item['class'])) is not None:
|
if (pattern.search(item['class'])) is not None:
|
||||||
remove = False
|
remove = False
|
||||||
|
|
||||||
# corrections
|
# corrections
|
||||||
# story content always good
|
# story content always good
|
||||||
pattern = re.compile('storyContent')
|
pattern = re.compile('storyContent')
|
||||||
if (pattern.search(item['class'])) is not None:
|
if (pattern.search(item['class'])) is not None:
|
||||||
remove = False
|
remove = False
|
||||||
|
|
||||||
#images
|
#images
|
||||||
pattern = re.compile('slideshow')
|
pattern = re.compile('slideshow')
|
||||||
if (pattern.search(item['class'])) is not None:
|
if (pattern.search(item['class'])) is not None:
|
||||||
if self._FETCH_IMAGES:
|
if self._FETCH_IMAGES:
|
||||||
remove = False
|
remove = False
|
||||||
else:
|
else:
|
||||||
remove = True
|
remove = True
|
||||||
|
|
||||||
#social widgets always bad
|
#social widgets always bad
|
||||||
pattern = re.compile('socialwidget')
|
pattern = re.compile('socialwidget')
|
||||||
if (pattern.search(item['class'])) is not None:
|
if (pattern.search(item['class'])) is not None:
|
||||||
remove = True
|
remove = True
|
||||||
|
|
||||||
if remove:
|
if remove:
|
||||||
items_to_extract.append(item)
|
items_to_extract.append(item)
|
||||||
|
|
||||||
for item in items_to_extract:
|
for item in items_to_extract:
|
||||||
item.extract()
|
item.extract()
|
||||||
|
|
||||||
items_to_extract = []
|
items_to_extract = []
|
||||||
|
|
||||||
if self._FETCH_IMAGES:
|
if self._FETCH_IMAGES:
|
||||||
for item in soup.findAll('a',attrs={'href' : re.compile('.*')}):
|
for item in soup.findAll('a',attrs={'href' : re.compile('.*')}):
|
||||||
if item.img is not None:
|
if item.img is not None:
|
||||||
#use full size image
|
#use full size image
|
||||||
img = item.findNext('img')
|
img = item.findNext('img')
|
||||||
|
|
||||||
img['src'] = item['href']
|
img['src'] = item['href']
|
||||||
|
|
||||||
#insert caption if available
|
#insert caption if available
|
||||||
if img['title'] is not None and (len(img['title']) > 1):
|
if img['title'] is not None and (len(img['title']) > 1):
|
||||||
tag = Tag(soup,'h3')
|
tag = Tag(soup,'h3')
|
||||||
text = NavigableString(img['title'])
|
text = NavigableString(img['title'])
|
||||||
tag.insert(0,text)
|
tag.insert(0,text)
|
||||||
|
|
||||||
#picture before text
|
#picture before text
|
||||||
img.extract()
|
img.extract()
|
||||||
item.insert(0,img)
|
item.insert(0,img)
|
||||||
item.insert(1,tag)
|
item.insert(1,tag)
|
||||||
|
|
||||||
# remove link
|
# remove link
|
||||||
item.name = "div"
|
item.name = "div"
|
||||||
item["class"]='image'
|
item["class"]='image'
|
||||||
del item["href"]
|
del item["href"]
|
||||||
|
|
||||||
|
|
||||||
#remove empty subtitles
|
#remove empty subtitles
|
||||||
"""
|
"""
|
||||||
currently the subtitle is located in first paragraph after
|
currently the subtitle is located in first paragraph after
|
||||||
sibling <h3 class="subtitle"> tag. This may be 'fixed' at
|
sibling <h3 class="subtitle"> tag. This may be 'fixed' at
|
||||||
some point.
|
some point.
|
||||||
"""
|
"""
|
||||||
subtitle = soup.find('h3',attrs={'class' : 'subtitle'})
|
subtitle = soup.find('h3',attrs={'class' : 'subtitle'})
|
||||||
if subtitle is not None:
|
if subtitle is not None:
|
||||||
subtitleText = subtitle.findNext('p')
|
subtitleText = subtitle.findNext('p')
|
||||||
if subtitleText is not None:
|
if subtitleText is not None:
|
||||||
if len(subtitleText.contents[0]) <= 1 :
|
if len(subtitleText.contents[0]) <= 1 :
|
||||||
subtitleText.extract()
|
subtitleText.extract()
|
||||||
subtitle.extract()
|
subtitle.extract()
|
||||||
|
|
||||||
|
|
||||||
#replace rating numbers with stars
|
#replace rating numbers with stars
|
||||||
for item in soup.findAll('div',attrs={ 'class' : 'starRating'}):
|
for item in soup.findAll('div',attrs={ 'class' : 'starRating'}):
|
||||||
if item is not None:
|
if item is not None:
|
||||||
soup2 = self._insertRatingStars(soup,item)
|
soup2 = self._insertRatingStars(soup,item)
|
||||||
if soup2 is not None:
|
if soup2 is not None:
|
||||||
soup = soup2
|
soup = soup2
|
||||||
|
|
||||||
|
|
||||||
#remove empty paragraph tags in storyTop which can leave a space
|
#remove empty paragraph tags in storyTop which can leave a space
|
||||||
#between first paragraph and rest of story
|
#between first paragraph and rest of story
|
||||||
nested_content = False
|
nested_content = False
|
||||||
storyTop = soup.find('div',attrs={ 'class' : ['storyTop']})
|
storyTop = soup.find('div',attrs={ 'class' : ['storyTop']})
|
||||||
for item in storyTop.findAll('p'):
|
for item in storyTop.findAll('p'):
|
||||||
for nested in item:
|
for nested in item:
|
||||||
@ -207,19 +206,19 @@ class TheIndependentNew(BasicNewsRecipe):
|
|||||||
break
|
break
|
||||||
if not nested_content and item.contents is not None and len(item.contents[0]) <= 1 :
|
if not nested_content and item.contents is not None and len(item.contents[0]) <= 1 :
|
||||||
items_to_extract.append(item)
|
items_to_extract.append(item)
|
||||||
|
|
||||||
for item in items_to_extract:
|
for item in items_to_extract:
|
||||||
item.extract()
|
item.extract()
|
||||||
|
|
||||||
items_to_extract = []
|
items_to_extract = []
|
||||||
|
|
||||||
|
|
||||||
#remove line breaks immediately next to tags with default margins
|
#remove line breaks immediately next to tags with default margins
|
||||||
#to prevent double line spacing and narrow columns of text
|
#to prevent double line spacing and narrow columns of text
|
||||||
storyTop = soup.find('div',attrs={ 'class' : ['storyTop']})
|
storyTop = soup.find('div',attrs={ 'class' : ['storyTop']})
|
||||||
self._remove_undesired_line_breaks_from_tag(storyTop,soup)
|
self._remove_undesired_line_breaks_from_tag(storyTop,soup)
|
||||||
|
|
||||||
|
|
||||||
#replace article graphics link with the graphics themselves
|
#replace article graphics link with the graphics themselves
|
||||||
if self._FETCH_ARTICLE_GRAPHICS:
|
if self._FETCH_ARTICLE_GRAPHICS:
|
||||||
items_to_insert = []
|
items_to_insert = []
|
||||||
@ -231,20 +230,20 @@ class TheIndependentNew(BasicNewsRecipe):
|
|||||||
if isinstance(child,Tag):
|
if isinstance(child,Tag):
|
||||||
if str(child.name) == 'a':
|
if str(child.name) == 'a':
|
||||||
items_to_insert.extend(self._get_article_graphic(strong,child['href'],soup))
|
items_to_insert.extend(self._get_article_graphic(strong,child['href'],soup))
|
||||||
|
|
||||||
for item in items_to_insert:
|
for item in items_to_insert:
|
||||||
item[0].replaceWith(item[1])
|
item[0].replaceWith(item[1])
|
||||||
|
|
||||||
for item in items_to_extract:
|
for item in items_to_extract:
|
||||||
item.extract()
|
item.extract()
|
||||||
|
|
||||||
return soup
|
return soup
|
||||||
|
|
||||||
|
|
||||||
def _get_article_graphic(self,old_item,url,soup):
|
def _get_article_graphic(self,old_item,url,soup):
|
||||||
|
|
||||||
items_to_insert = []
|
items_to_insert = []
|
||||||
|
|
||||||
if re.search('\.jpg$',str(url)):
|
if re.search('\.jpg$',str(url)):
|
||||||
div = Tag(soup,'div')
|
div = Tag(soup,'div')
|
||||||
div['class'] = 'pictureContainer'
|
div['class'] = 'pictureContainer'
|
||||||
@ -254,20 +253,20 @@ class TheIndependentNew(BasicNewsRecipe):
|
|||||||
div.insert(0,img)
|
div.insert(0,img)
|
||||||
items_to_insert.append((old_item,div,))
|
items_to_insert.append((old_item,div,))
|
||||||
return items_to_insert
|
return items_to_insert
|
||||||
|
|
||||||
soup2 = self.index_to_soup(url)
|
soup2 = self.index_to_soup(url)
|
||||||
for item in soup2.findAll('div',attrs={'class' : re.compile("widget picture article.*")}):
|
for item in soup2.findAll('div',attrs={'class' : re.compile("widget picture article.*")}):
|
||||||
items_to_insert.append((old_item,item),)
|
items_to_insert.append((old_item,item),)
|
||||||
return items_to_insert
|
return items_to_insert
|
||||||
|
|
||||||
|
|
||||||
def _insertRatingStars(self,soup,item):
|
def _insertRatingStars(self,soup,item):
|
||||||
if item.contents is None:
|
if item.contents is None:
|
||||||
return
|
return
|
||||||
rating = item.contents[0]
|
rating = item.contents[0]
|
||||||
if not rating.isdigit():
|
if not rating.isdigit():
|
||||||
return None
|
return None
|
||||||
rating = int(item.contents[0])
|
rating = int(item.contents[0])
|
||||||
for i in range(1,6):
|
for i in range(1,6):
|
||||||
star = Tag(soup,'img')
|
star = Tag(soup,'img')
|
||||||
if i <= rating:
|
if i <= rating:
|
||||||
@ -277,26 +276,26 @@ class TheIndependentNew(BasicNewsRecipe):
|
|||||||
star['alt'] = 'star number ' + str(i)
|
star['alt'] = 'star number ' + str(i)
|
||||||
item.insert(i,star)
|
item.insert(i,star)
|
||||||
#item.contents[0] = NavigableString('(' + str(rating) + ')')
|
#item.contents[0] = NavigableString('(' + str(rating) + ')')
|
||||||
item.contents[0] = ''
|
item.contents[0] = ''
|
||||||
|
|
||||||
def postprocess_html(self,soup, first_fetch):
|
def postprocess_html(self,soup, first_fetch):
|
||||||
#find broken images and remove captions
|
#find broken images and remove captions
|
||||||
items_to_extract = []
|
items_to_extract = []
|
||||||
for item in soup.findAll('div', attrs={'class' : 'image'}):
|
for item in soup.findAll('div', attrs={'class' : 'image'}):
|
||||||
img = item.findNext('img')
|
img = item.findNext('img')
|
||||||
if img is not None and img['src'] is not None:
|
if img is not None and img['src'] is not None:
|
||||||
# broken images still point to remote url
|
# broken images still point to remote url
|
||||||
pattern = re.compile('http://www.independent.co.uk.*')
|
pattern = re.compile('http://www.independent.co.uk.*')
|
||||||
if pattern.match(img["src"]) is not None:
|
if pattern.match(img["src"]) is not None:
|
||||||
caption = img.findNextSibling('h3')
|
caption = img.findNextSibling('h3')
|
||||||
if caption is not None:
|
if caption is not None:
|
||||||
items_to_extract.append(caption)
|
items_to_extract.append(caption)
|
||||||
items_to_extract.append(img)
|
items_to_extract.append(img)
|
||||||
|
|
||||||
for item in items_to_extract:
|
for item in items_to_extract:
|
||||||
item.extract()
|
item.extract()
|
||||||
return soup
|
return soup
|
||||||
|
|
||||||
def _recurisvely_linearise_tag_tree(
|
def _recurisvely_linearise_tag_tree(
|
||||||
self,
|
self,
|
||||||
item,
|
item,
|
||||||
@ -311,25 +310,25 @@ class TheIndependentNew(BasicNewsRecipe):
|
|||||||
if not (isinstance(item,Tag)):
|
if not (isinstance(item,Tag)):
|
||||||
return linearised
|
return linearised
|
||||||
for nested in item:
|
for nested in item:
|
||||||
linearised.append(nested)
|
linearised.append(nested)
|
||||||
linearised = self._recurisvely_linearise_tag_tree(nested,linearised, count)
|
linearised = self._recurisvely_linearise_tag_tree(nested,linearised, count)
|
||||||
return linearised
|
return linearised
|
||||||
|
|
||||||
|
|
||||||
def _get_previous_tag(self,current_index, tag_tree):
|
def _get_previous_tag(self,current_index, tag_tree):
|
||||||
if current_index == 0:
|
if current_index == 0:
|
||||||
return None
|
return None
|
||||||
else:
|
else:
|
||||||
return tag_tree[current_index - 1]
|
return tag_tree[current_index - 1]
|
||||||
|
|
||||||
|
|
||||||
def _get_next_tag(self,current_index, tag_tree):
|
def _get_next_tag(self,current_index, tag_tree):
|
||||||
if current_index < len(tag_tree) - 1:
|
if current_index < len(tag_tree) - 1:
|
||||||
return tag_tree[current_index + 1]
|
return tag_tree[current_index + 1]
|
||||||
else:
|
else:
|
||||||
return None
|
return None
|
||||||
|
|
||||||
|
|
||||||
def _list_match(self,test_str, list_regex):
|
def _list_match(self,test_str, list_regex):
|
||||||
for regex in list_regex:
|
for regex in list_regex:
|
||||||
match = re.match(regex, test_str)
|
match = re.match(regex, test_str)
|
||||||
@ -338,24 +337,24 @@ class TheIndependentNew(BasicNewsRecipe):
|
|||||||
return False
|
return False
|
||||||
|
|
||||||
def _remove_undesired_line_breaks_from_tag(self,parent,soup):
|
def _remove_undesired_line_breaks_from_tag(self,parent,soup):
|
||||||
|
|
||||||
if parent is None:
|
if parent is None:
|
||||||
return
|
return
|
||||||
|
|
||||||
|
|
||||||
tag_tree = self._recurisvely_linearise_tag_tree(parent)
|
tag_tree = self._recurisvely_linearise_tag_tree(parent)
|
||||||
items_to_remove = []
|
items_to_remove = []
|
||||||
|
|
||||||
|
|
||||||
for item in tag_tree:
|
for item in tag_tree:
|
||||||
if item == u'\n':
|
if item == u'\n':
|
||||||
items_to_remove.append(item)
|
items_to_remove.append(item)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
for item in items_to_remove:
|
for item in items_to_remove:
|
||||||
tag_tree.remove(item)
|
tag_tree.remove(item)
|
||||||
|
|
||||||
|
|
||||||
spaced_tags = [r'p', r'h\d', r'blockquote']
|
spaced_tags = [r'p', r'h\d', r'blockquote']
|
||||||
tags_to_extract = []
|
tags_to_extract = []
|
||||||
tags_to_replace = []
|
tags_to_replace = []
|
||||||
@ -363,41 +362,41 @@ class TheIndependentNew(BasicNewsRecipe):
|
|||||||
if isinstance(tag, Tag):
|
if isinstance(tag, Tag):
|
||||||
if str(tag) == '<br />':
|
if str(tag) == '<br />':
|
||||||
previous_tag = self._get_previous_tag(i, tag_tree)
|
previous_tag = self._get_previous_tag(i, tag_tree)
|
||||||
|
|
||||||
if isinstance(previous_tag, Tag):
|
if isinstance(previous_tag, Tag):
|
||||||
previous_tag_is_spaced = previous_tag is not None\
|
previous_tag_is_spaced = previous_tag is not None\
|
||||||
and self._list_match(str(previous_tag.name),
|
and self._list_match(str(previous_tag.name),
|
||||||
spaced_tags)
|
spaced_tags)
|
||||||
else:
|
else:
|
||||||
previous_tag_is_spaced = False
|
previous_tag_is_spaced = False
|
||||||
|
|
||||||
next_tag = self._get_next_tag(i, tag_tree)
|
next_tag = self._get_next_tag(i, tag_tree)
|
||||||
|
|
||||||
if isinstance(next_tag, Tag):
|
if isinstance(next_tag, Tag):
|
||||||
next_tag_is_spaced = next_tag is not None\
|
next_tag_is_spaced = next_tag is not None\
|
||||||
and self._list_match(str(next_tag.name), spaced_tags)
|
and self._list_match(str(next_tag.name), spaced_tags)
|
||||||
else:
|
else:
|
||||||
next_tag_is_spaced = False
|
next_tag_is_spaced = False
|
||||||
|
|
||||||
if previous_tag_is_spaced or next_tag_is_spaced or i == 0\
|
if previous_tag_is_spaced or next_tag_is_spaced or i == 0\
|
||||||
or i == len(tag_tree) - 1:
|
or i == len(tag_tree) - 1:
|
||||||
tags_to_extract.append(tag)
|
tags_to_extract.append(tag)
|
||||||
else:
|
else:
|
||||||
tags_to_replace.append((tag,NavigableString(' '),))
|
tags_to_replace.append((tag,NavigableString(' '),))
|
||||||
|
|
||||||
|
|
||||||
for pair in tags_to_replace:
|
for pair in tags_to_replace:
|
||||||
pair[0].replaceWith(pair[1])
|
pair[0].replaceWith(pair[1])
|
||||||
for tag in tags_to_extract:
|
for tag in tags_to_extract:
|
||||||
tag.extract()
|
tag.extract()
|
||||||
|
|
||||||
feeds = [
|
feeds = [
|
||||||
(u'News - UK',
|
(u'News - UK',
|
||||||
u'http://www.independent.co.uk/news/uk/?service=rss'),
|
u'http://www.independent.co.uk/news/uk/?service=rss'),
|
||||||
(u'News - World',
|
(u'News - World',
|
||||||
u'http://www.independent.co.uk/news/world/?service=rss'),
|
u'http://www.independent.co.uk/news/world/?service=rss'),
|
||||||
(u'News - Business',
|
(u'News - Business',
|
||||||
u'http://www.independent.co.uk/news/business/?service=rss'),
|
u'http://www.independent.co.uk/news/business/?service=rss'),
|
||||||
(u'News - People',
|
(u'News - People',
|
||||||
u'http://www.independent.co.uk/news/people/?service=rss'),
|
u'http://www.independent.co.uk/news/people/?service=rss'),
|
||||||
(u'News - Science',
|
(u'News - Science',
|
||||||
@ -497,4 +496,4 @@ class TheIndependentNew(BasicNewsRecipe):
|
|||||||
(u'IndyBest',
|
(u'IndyBest',
|
||||||
u'http://www.independent.co.uk/extras/indybest/?service=rss'),
|
u'http://www.independent.co.uk/extras/indybest/?service=rss'),
|
||||||
]
|
]
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user