from calibre.web.feeds.news import BasicNewsRecipe from html5_parser import parse from datetime import datetime import json today = datetime.today().strftime('%d-%m-%Y') # today = '20-09-2023' day, month, year = (int(x) for x in today.split('-')) dt = datetime(year, month, day) class BusinessStandardPrint(BasicNewsRecipe): title = 'Business Standard Print Edition' __author__ = 'unkn0wn' description = "India's most respected business daily, Articles from Today's Paper" language = 'en_IN' masthead_url = 'https://bsmedia.business-standard.com/include/_mod/site/html5/images/business-standard-logo.png' encoding = 'utf-8' timefmt = ' [' + dt.strftime('%b %d, %Y') + ']' resolve_internal_links = True remove_empty_feeds = True no_stylesheets = True remove_javascript = True remove_attributes = ['width', 'height', 'float', 'style'] def __init__(self, *args, **kwargs): BasicNewsRecipe.__init__(self, *args, **kwargs) if self.output_profile.short_name.startswith('kindle'): self.title = 'Business Standard ' + dt.strftime('%b %d, %Y') def get_browser(self): return BasicNewsRecipe.get_browser(self, user_agent='common_words/based') ignore_duplicate_articles = {'title', 'url'} remove_empty_feeds = True resolve_internal_links = True extra_css = ''' img {display:block; margin:0 auto;} .auth, .cat { font-size:small; color:#202020; } .cap { font-size:small; text-align:center; } ''' def get_cover_url(self): soup = self.index_to_soup('https://www.magzter.com/IN/Business-Standard-Private-Ltd/Business-Standard/Newspaper/') for citem in soup.findAll('meta', content=lambda s: s and s.endswith('view/3.jpg')): return citem['content'] def parse_index(self): if dt.weekday() == 6: self.log.warn( 'Business Standard Does Not Have A Print Publication On Sunday. The Reports' ' And Columns On This Page Today Appeared In The Newspaper\'s Saturday Edition.' ) url = 'https://apibs.business-standard.com/category/today-paper?sortBy=' + today raw = self.index_to_soup(url, raw=True) data = json.loads(raw) data = data['data'] feeds = [] for section in data: if section == 'EpaperImage': continue self.log(section) articles = [] for article in data[section]: title = article['heading1'] desc = article['sub_heading'] url = 'https://www.business-standard.com' + article['article_url'] self.log('\t', title, '\n\t', desc, '\n\t\t', url) articles.append({'title': title, 'description':desc, 'url': url}) if articles: feeds.append((section, articles)) return feeds def preprocess_raw_html(self, raw, *a): root = parse(raw) m = root.xpath('//script[@id="__NEXT_DATA__"]') data = json.loads(m[0].text) img_url = None if 'articleImageUrl' in data['props']['pageProps']['articleSchema']: img_url = data['props']['pageProps']['articleSchema']['articleImageUrl'] art_url = 'https://www.business-standard.com' + data['props']['pageProps']['url'] data = data['props']['pageProps']['data'] title = '
' + data['defaultArticleCat']['h1_tag'] + '
' + ', '.join(authors) + ' | ' + data['placeName'] + ' | ' + date + '
'.format(img_url)
else:
lede = '
'.format(data['featuredImageObj']['url'])
if 'alt_text' in data['featuredImageObj']:
caption = '' + data['featuredImageObj']['alt_text'] + '