From e9478054a8ef767c3a5753852a4d2ac769a27a0f Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Tomasz=20D=C5=82ugosz?= Date: Sun, 14 Oct 2018 23:05:51 +0200 Subject: [PATCH] site moved and feed is gone --- recipes/icons/poradnia_pwn.png | Bin 145 -> 0 bytes recipes/poradnia_pwn.recipe | 63 --------------------------------- 2 files changed, 63 deletions(-) delete mode 100644 recipes/icons/poradnia_pwn.png delete mode 100644 recipes/poradnia_pwn.recipe diff --git a/recipes/icons/poradnia_pwn.png b/recipes/icons/poradnia_pwn.png deleted file mode 100644 index 22ed7364b8af39785aab00071f10d129ada94f44..0000000000000000000000000000000000000000 GIT binary patch literal 0 HcmV?d00001 literal 145 zcmeAS@N?(olHy`uVBq!ia0vp^0wBx*Bp9q_EZ7UAB0OCjLn;_q4>B?`FmN1RaJ1gS zTCpXXS7F^M#-g`j%uD*XE?nKn=C?YTH)pMzY|QJd?R;;pg}!AdUn|uRyM5N}zO~)A sLN}XjT$Lr86dKyRzopr00uQP-T(jq diff --git a/recipes/poradnia_pwn.recipe b/recipes/poradnia_pwn.recipe deleted file mode 100644 index ce89cb73f0..0000000000 --- a/recipes/poradnia_pwn.recipe +++ /dev/null @@ -1,63 +0,0 @@ -# vim:fileencoding=UTF-8:ts=4:sw=4:sta:et:sts=4:fdm=marker:ai -from calibre.web.feeds.news import BasicNewsRecipe - - -class PoradniaPWN(BasicNewsRecipe): - title = u'Poradnia Językowa PWN' - __author__ = 'fenuks' - description = u'Internetowa poradnia językowa Wydawnictwa Naukowego PWN. Poradnię prowadzi Redaktor Naczelny Słowników Języka Polskiego, prof. Mirosław Bańko. Pomagają mu eksperci - znani polscy językoznawcy. Współpracuje z nami m.in. prof. Jerzy Bralczyk oraz dr Jan Grzenia.' # noqa - category = 'language' - language = 'pl' - oldest_article = 14 - max_articles_per_feed = 100000 - INDEX = "http://poradnia.pwn.pl/" - no_stylesheets = True - remove_attributes = ['style'] - remove_javascript = True - use_embedded_content = False - keep_only_tags = [dict(name="div", attrs={"class": "searchhi"})] - feeds = [(u'Poradnia', u'http://rss.pwn.pl/poradnia.rss')] - - '''def find_articles(self, url): - articles = [] - soup=self.index_to_soup(url) - counter = int(soup.find(name='p', attrs={'class':'count'}).findAll('b')[-1].string) - counter = 500 - pos = 0 - next = url - while next: - soup=self.index_to_soup(next) - tag=soup.find(id="listapytan") - art=tag.findAll(name='li') - for i in art: - if i.h4: - title=i.h4.a.string - url=self.INDEX+i.h4.a['href'] - #date=soup.find(id='footer').ul.li.string[41:-1] - articles.append({'title' : title, - 'url' : url, - 'date' : '', - 'description' : '' - }) - pos += 10 - if not pos >=counter: - next = 'http://poradnia.pwn.pl/lista.php?kat=18&od=' + str(pos) - print u'Tworzenie listy artykułów dla', next - else: - next = None - print articles - return articles - - def parse_index(self): - feeds = [] - feeds.append((u"Poradnia", self.find_articles('http://poradnia.pwn.pl/lista.php'))) - - return feeds''' - - def preprocess_html(self, soup): - for i in soup.findAll(name=['ul', 'li']): - i.name = "div" - for z in soup.findAll(name='a'): - if not z['href'].startswith('http'): - z['href'] = 'http://poradnia.pwn.pl/' + z['href'] - return soup