#!/usr/bin/env python
import re

from calibre.web.feeds.news import BasicNewsRecipe

use_wayback_machine = False


class NytFeeds(BasicNewsRecipe):
    title = 'NYT News'
    __author__ = 'unkn0wn'
    description = (
        'The New York Times is dedicated to helping people understand the world through '
        'on-the-ground, expert and deeply reported independent journalism. Feeds based recipe.'
    )
    oldest_article = 1
    encoding = 'utf-8'
    use_embedded_content = False
    language = 'en_US'
    remove_empty_feeds = True
    resolve_internal_links = True
    ignore_duplicate_articles = {'title', 'url'}
    masthead_url = 'https://static01.nytimes.com/newsgraphics/2015-12-23-masthead-2016/b15c3d81d3d7b59065fff9a3f3afe85aa2e2dff5/_assets/nyt-logo.png'

    # https://www.nytimes.com/rss
    # https://developer.nytimes.com/docs/rss-api/1/overview
    feeds = [
        # to filter out all opinions from other sections first
        'https://rss.nytimes.com/services/xml/rss/nyt/Opinion.xml',
        'https://rss.nytimes.com/services/xml/rss/nyt/HomePage.xml',
        'https://rss.nytimes.com/services/xml/rss/nyt/World.xml',
        'https://rss.nytimes.com/services/xml/rss/nyt/US.xml',
        'https://rss.nytimes.com/services/xml/rss/nyt/Business.xml',
        'https://rss.nytimes.com/services/xml/rss/nyt/YourMoney.xml',
        # 'https://rss.nytimes.com/services/xml/rss/nyt/Technology.xml',
        'https://rss.nytimes.com/services/xml/rss/nyt/Science.xml',
        'https://rss.nytimes.com/services/xml/rss/nyt/Climate.xml',
        'https://rss.nytimes.com/services/xml/rss/nyt/Health.xml',
        'https://rss.nytimes.com/services/xml/rss/nyt/Arts.xml',
        # 'https://rss.nytimes.com/services/xml/rss/nyt/FashionandStyle.xml',
        # 'https://rss.nytimes.com/services/xml/rss/nyt/tmagazine.xml',
        # 'https://rss.nytimes.com/services/xml/rss/nyt/books.xml',
        'https://www.nytimes.com/services/xml/rss/nyt/Travel.xml',
        # 'https://rss.nytimes.com/services/xml/rss/nyt/well.xml',
        'https://rss.nytimes.com/services/xml/rss/nyt/Sports.xml',
        # 'https://rss.nytimes.com/services/xml/rss/nyt/Magazine.xml',
        # 'http://nytimes.com/timeswire/feeds/',
    ]

    def get_cover_url(self):
        soup = self.index_to_soup('https://www.frontpages.com/the-new-york-times/')
        return (
            'https://www.frontpages.com'
            + soup.find('img', attrs={'id': 'giornale-img'})['src']
        )

    recipe_specific_options = {
        'days': {
            'short': 'Oldest article to download from this news source. In days ',
            'long': 'For example, 0.5, gives you articles from the past 12 hours',
            'default': str(oldest_article),
        },
        'comp': {'short': 'Compress News Images?', 'long': 'enter yes', 'default': 'no'},
        'rev': {
            'short': 'Reverse the order of articles in each feed?',
            'long': 'enter yes',
            'default': 'no',
        },
        'res': {
            'short': (
                'For hi-res images, select a resolution from the following\noptions: '
                'popup, jumbo, mobileMasterAt3x, superJumbo'
            ),
            'long': (
                'This is useful for non e-ink devices, and for a lower file size\nthan '
                'the default, use mediumThreeByTwo440, mediumThreeByTwo225, articleInline.'
            ),
        },
    }

    def __init__(self, *args, **kwargs):
        BasicNewsRecipe.__init__(self, *args, **kwargs)
        d = self.recipe_specific_options.get('days')
        if d and isinstance(d, str):
            self.oldest_article = float(d)
        r = self.recipe_specific_options.get('rev')
        if r and isinstance(r, str):
            if r.lower() == 'yes':
                self.reverse_article_order = True
        c = self.recipe_specific_options.get('comp')
        if c and isinstance(c, str):
            if c.lower() == 'yes':
                self.compress_news_images = True

    extra_css = '''
        .byl, .time { font-size:small; color:#202020; }
        .cap { font-size:small; text-align:center; }
        .cred { font-style:italic; font-size:small; }
        em, blockquote { color: #202020; }
        .sc { font-variant: small-caps; }
        .lbl { font-size:small; color:#404040; }
        img { display:block; margin:0 auto; }
    '''

    @property
    def nyt_parser(self):
        ans = getattr(self, '_nyt_parser', None)
        if ans is None:
            from calibre.live import load_module

            self._nyt_parser = ans = load_module('calibre.web.site_parsers.nytimes')
        return ans

    def get_nyt_page(self, url, skip_wayback=False):
        if use_wayback_machine and not skip_wayback:
            from calibre import browser

            return self.nyt_parser.download_url(url, browser())
        return self.index_to_soup(url, raw=True)

    articles_are_obfuscated = use_wayback_machine

    if use_wayback_machine:

        def get_obfuscated_article(self, url):
            from calibre.ptempfile import PersistentTemporaryFile

            with PersistentTemporaryFile() as tf:
                tf.write(self.get_nyt_page(url))
            return tf.name

    def preprocess_raw_html(self, raw_html, url):
        return self.nyt_parser.extract_html(self.index_to_soup(raw_html), url)

    def get_browser(self, *args, **kwargs):
        kwargs['user_agent'] = (
            'Mozilla/5.0 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)'
        )
        br = BasicNewsRecipe.get_browser(self, *args, **kwargs)
        br.addheaders += [
            ('Referer', 'https://www.google.com/'),
            ('X-Forwarded-For', '66.249.66.1'),
        ]
        return br

    def preprocess_html(self, soup):
        w = self.recipe_specific_options.get('res')
        if w and isinstance(w, str):
            res = '-' + w
            for img in soup.findAll('img', attrs={'src': True}):
                if '-article' in img['src']:
                    ext = img['src'].split('?')[0].split('.')[-1]
                    img['src'] = img['src'].rsplit('-article', 1)[0] + res + '.' + ext
        for c in soup.findAll('div', attrs={'class': 'cap'}):
            for p in c.findAll(['p', 'div']):
                p.name = 'span'
        return soup

    def get_article_url(self, article):
        url = BasicNewsRecipe.get_article_url(self, article)
        # you can remove '|/espanol/' from code below to include spanish articles.
        if not re.search(r'/video/|/live/|/athletic/|/espanol/|/card/', url):
            return url
        self.log('\tSkipped URL: ', url)
        return None
