diff --git a/recipes/icons/nowa_fantastyka.png b/recipes/icons/nowa_fantastyka.png
deleted file mode 100644
index 66170e3398..0000000000
Binary files a/recipes/icons/nowa_fantastyka.png and /dev/null differ
diff --git a/recipes/nowa_fantastyka.recipe b/recipes/nowa_fantastyka.recipe
deleted file mode 100644
index 28c5408890..0000000000
--- a/recipes/nowa_fantastyka.recipe
+++ /dev/null
@@ -1,87 +0,0 @@
-# -*- coding: utf-8 -*-
-from calibre.web.feeds.news import BasicNewsRecipe
-import re
-
-
-class Nowa_Fantastyka(BasicNewsRecipe):
- title = u'Nowa Fantastyka'
- oldest_article = 7
- __author__ = 'fenuks'
- __modified_by__ = 'zaslav'
- language = 'pl'
- encoding = 'latin2'
- description = u'Strona dla miłośników fantastyki'
- category = 'fantasy'
- masthead_url = 'http://farm5.static.flickr.com/4133/4956658792_7ba7fbf562.jpg'
- # extra_css='.tytul {font-size: 20px;}' #not working
- max_articles_per_feed = 100
- INDEX = 'http://www.fantastyka.pl/'
- no_stylesheets = True
- needs_subscription = 'optional'
- remove_tags_before = dict(attrs={'class': 'naglowek2'})
- remove_tags_after = dict(name='form', attrs={'name': 'form1'})
- remove_tags = [dict(attrs={'class': ['avatar2', 'belka-margin', 'naglowek2']}), dict(name='span', attrs={'class': 'alert-oceny'}), dict(name='img', attrs={'src': ['obrazki/sledz1.png', 'obrazki/print.gif', 'obrazki/mlnf.gif']}), dict(name='b', text='Dodaj komentarz'), dict(name='a', attrs={'href': 'http://www.fantastyka.pl/10,1727.html'}), dict(name='form')] # noqa
- preprocess_regexps = [
- (re.compile(r'\
'), lambda match: ''),
- (re.compile(r'\'), lambda match: ''),
- (re.compile(r'\'), lambda match: '')]
-
- def find_articles(self, url):
- articles = []
- soup = self.index_to_soup(url)
- tag = soup.find(attrs={'class': 'belka1-tlo-m'})
- art = tag.findAll(name='a', attrs={'class': 'a-box'})
- for i in art:
- title = i.string
- url = self.INDEX + i['href']
- # date=soup.find(id='footer').ul.li.string[41:-1]
- articles.append({'title': title,
- 'url': url,
- 'date': '',
- 'description': ''
- })
- return articles
-
- def parse_index(self):
- feeds = []
- feeds.append((u"Opowiadania", self.find_articles(
- 'http://www.fantastyka.pl/3.html')))
- feeds.append((u"Publicystyka", self.find_articles(
- 'http://www.fantastyka.pl/6.html')))
- feeds.append((u"Hype Park", self.find_articles(
- 'http://www.fantastyka.pl/9.html')))
-
- return feeds
-
- def get_cover_url(self):
- soup = self.index_to_soup('http://www.e-kiosk.pl/nowa_fantastyka')
- self.cover_url = 'http://www.e-kiosk.pl' + \
- soup.find(name='a', attrs={'class': 'img'})['href']
- return getattr(self, 'cover_url', self.cover_url)
-
- def get_browser(self):
- br = BasicNewsRecipe.get_browser(self)
- if self.username is not None and self.password is not None:
- br.open('http://www.fantastyka.pl/')
- br.select_form(nr=0)
- br['login'] = self.username
- br['pass'] = self.password
- br.submit()
- return br
-
- def preprocess_html(self, soup):
- for item in soup.findAll(style=True):
- del item['style']
- for item in soup.findAll(font=True):
- del item['font']
- for item in soup.findAll(align=True):
- del item['align']
- for item in soup.findAll(name='tr'):
- item.name = 'div'
- title = soup.find(attrs={'class': 'tytul'})
- if title:
- title['style'] = 'font-size: 20px; font-weight: bold;'
- for a in soup('a'):
- if a.has_key('href') and 'http://' not in a['href'] and 'https://' not in a['href']: # noqa
- a['href'] = self.INDEX + a['href']
- return soup