#!/usr/bin/env python __license__ = 'GPL v3' __copyright__ = '2009, Darko Miletic ' ''' elargentino.com ''' from calibre.web.feeds.news import BasicNewsRecipe from calibre.ebooks.BeautifulSoup import Tag def new_tag(soup, name, attrs=()): impl = getattr(soup, 'new_tag', None) if impl is not None: return impl(name, attrs=dict(attrs)) return Tag(soup, name, attrs=attrs or None) class Veintitres(BasicNewsRecipe): title = 'Veintitres' __author__ = 'Darko Miletic' description = 'Revista Argentino dedicada a politica' publisher = 'Veintitres' category = 'news, politics, Argentina' oldest_article = 7 max_articles_per_feed = 100 no_stylesheets = True use_embedded_content = False encoding = 'utf-8' language = 'es_AR' lang = 'es-AR' direction = 'ltr' INDEX = 'http://www.elargentino.com/medios/120/veintitres.html' extra_css = ' .titulo{font-size: x-large; font-weight: bold} .volantaImp{font-size: small; font-weight: bold} ' html2lrf_options = [ '--comment', description, '--category', category, '--publisher', publisher ] html2epub_options = 'publisher="' + publisher + '"\ncomments="' + description + '"\ntags="' + \ category + \ '"\noverride_css=" p {text-indent: 0cm; margin-top: 0em; margin-bottom: 0.5em} "' keep_only_tags = [dict(name='div', attrs={'class': 'ContainerPop'})] remove_tags = [dict(name='link')] feeds = [(u'Articulos', u'http://www.elargentino.com/Highlights.aspx?ParentType=Section&ParentId=120&Content-Type=text/xml&ChannelDesc=Veintitres')] def print_version(self, url): main, sep, article_part = url.partition('/nota-') article_id, rsep, rrest = article_part.partition('-') return u'http://www.elargentino.com/Impresion.aspx?Id=' + article_id def preprocess_html(self, soup): for item in soup.findAll(style=True): del item['style'] soup.html['lang'] = self.lang soup.html['dir'] = self.direction mlang = new_tag(soup, 'meta', [ ("http-equiv", "Content-Language"), ("content", self.lang)]) mcharset = new_tag(soup, 'meta', [ ("http-equiv", "Content-Type"), ("content", "text/html; charset=utf-8")]) soup.head.insert(0, mlang) soup.head.insert(1, mcharset) return soup def get_cover_url(self): cover_url = None soup = self.index_to_soup(self.INDEX) cover_item = soup.find('div', attrs={'class': 'colder'}) if cover_item: clean_url = self.image_url_processor( None, cover_item.div.img['src']) cover_url = 'http://www.elargentino.com' + clean_url + '&height=600' return cover_url def image_url_processor(self, baseurl, url): base, sep, rest = url.rpartition('?Id=') img, sep2, rrest = rest.partition('&') return base + sep + img