#!/usr/bin/env python
# vim:fileencoding=utf-8
from __future__ import unicode_literals
__license__ = 'GPL v3'
__copyright__ = '2013'
'''
monde-diplomatique.fr
'''
import re
from calibre.web.feeds import feeds_from_index
from calibre.web.feeds.news import BasicNewsRecipe
def absurl(url):
if url.startswith('/'):
url = 'http://www.monde-diplomatique.fr' + url
return url
class LeMondeDiplomatiqueSiteWeb(BasicNewsRecipe):
title = u'Le Monde diplomatique.fr'
__author__ = 'Gaëtan Lehmann'
description = "Le Monde diplomatique est un mensuel français d’information et d’opinion à la ligne éditoriale nettement engagée en faveur d'une gauche de rupture avec le capitalisme. Il aborde de nombreux sujets — géopolitique, relations internationales, économie, questions sociales, écologie, culture, médias, …" # noqa
oldest_article = 30
max_articles_per_feed = 100
auto_cleanup = True
publisher = 'monde-diplomatique.fr'
category = 'news, France, world'
language = 'fr'
masthead_url = 'http://www.monde-diplomatique.fr/squelettes/images/logotyfa.png'
timefmt = ' [%d %b %Y]'
no_stylesheets = True
recipe_specific_options = {
'days': {
'short': 'Oldest article to download from this news source. In days ',
'long': 'For example, 0.5, gives you articles from the past 12 hours',
'default': str(oldest_article)
}
}
def __init__(self, *args, **kwargs):
BasicNewsRecipe.__init__(self, *args, **kwargs)
d = self.recipe_specific_options.get('days')
if d and isinstance(d, str):
self.oldest_article = float(d)
feeds = [(u'Blogs', u'http://blog.mondediplo.net/spip.php?page=backend'),
(u'Archives', u'http://www.monde-diplomatique.fr/rss/')]
preprocess_regexps = [
(re.compile(r'
(.*) - Les blogs du Diplo'),
lambda m: '' + m.group(1) + ''),
(re.compile(r'(.*) - Les blogs du Diplo
'),
lambda m: '' + m.group(1) + '
'),
(re.compile(r'(.*) \(Le Monde diplomatique\)'),
lambda m: '' + m.group(1) + ''),
(re.compile(r'(.*) \(Le Monde diplomatique\)
'),
lambda m: '' + m.group(1) + '
'),
(re.compile(r'Grand format
'), lambda m: '')]
remove_tags = [dict(name='div', attrs={'class': 'voiraussi liste'}),
dict(name='ul', attrs={
'class': 'hermetique carto hombre_demi_inverse'}),
dict(name='a', attrs={'class': 'tousles'}),
dict(name='h3', attrs={'class': 'cat'}),
dict(name='div', attrs={'class': 'logodiplo'}),
dict(name='img', attrs={'class': 'spip_logos'}),
dict(name='p', attrs={'id': 'hierarchie'}),
dict(name='div', attrs={'class': 'espace'})]
conversion_options = {
'comments': description,
'tags': category,
'publisher': publisher,
'linearize_tables': True
}
remove_empty_feeds = True
filterDuplicates = True
# don't use parse_index - we need it to send an exception so we can mix
# feed and parse_index results in parse_feeds
def parse_index_valise(self):
articles = []
soup = self.index_to_soup('http://www.monde-diplomatique.fr/carnet/')
cnt = soup.find('ul', attrs={'class': 'liste double'})
for item in cnt.findAll('li'):
description = ''
feed_link = item.find('a', href=True)
title = self.tag_to_string(item.find('h3'))
desc = item.find('div', attrs={'class': 'intro'})
date = item.find('div', attrs={'class': 'dates_auteurs'})
if desc:
description = desc.string
if feed_link:
articles.append({
'title': title,
'date': self.tag_to_string(date),
'url': absurl(feed_link['href']),
'description': description
})
return [("La valise diplomatique", articles)]
def parse_index_cartes(self):
articles = []
soup = self.index_to_soup('http://www.monde-diplomatique.fr/cartes/')
cnt = soup.find('ul', attrs={'class': 'liste_vignettes hautcartes'})
for li in cnt.findAll('li'):
feed_link = li.find('a', href=True)
h3 = li.find('h3')
authorAndDate = li.find('div', attrs={'class': 'dates_auteurs'})
author_date = self.tag_to_string(authorAndDate).split(', ')
author = author_date[0]
date = author_date[-1]
if feed_link:
title = self.tag_to_string(h3)
articles.append({
'title': title,
'date': date,
'url': absurl(feed_link['href']),
'description': author
})
return [("Cartes", articles)]
def parse_feeds(self):
feeds = BasicNewsRecipe.parse_feeds(self)
valise = feeds_from_index(self.parse_index_valise(), oldest_article=self.oldest_article,
max_articles_per_feed=self.max_articles_per_feed,
log=self.log)
cartes = feeds_from_index(self.parse_index_cartes(), oldest_article=self.oldest_article,
max_articles_per_feed=self.max_articles_per_feed,
log=self.log)
feeds = valise + feeds + cartes
return feeds