View Single Post
Old 03-28-2017, 03:28 AM   #1
lennyw
Member
lennyw began at the beginning.
 
Posts: 13
Karma: 10
Join Date: Oct 2011
Location: Berlin, Germany
Device: Kindle 3G
The Spectator - how do I download back issues?

Hi all,

I've been looking at the recipe for The Spectator, hoping there'd be a line which I could edit to download a complete back issue, but cannot find it (I have read the threads on downloading back issues of The Economist that one just needs to edit the line INDEX = 'https://www.economist.com/printedition').

On The Spectator's website each issue's index has an issue-specific url, eg:
https://www.spectator.co.uk/issues/11-march-2017/

Is there any way I can incorporate this into the feed to download specific issues?

Thanks in advance for your help
Lenny

PS Here's the recipe code from Calibre:
Spoiler:
Code:
from calibre.web.feeds.recipes import BasicNewsRecipe


def class_sel(cls):
    def f(x):
        return x and cls in x.split()
    return f


class Spectator(BasicNewsRecipe):

    title = 'Spectator Magazine'
    __author__ = 'Kovid Goyal'
    description = 'Magazine'
    language = 'en'

    no_stylesheets = True

    keep_only_tags = dict(name='div', attrs={
                          'class': ['article-header__text', 'featured-image', 'article-content']})
    remove_tags = [
        dict(name='div', attrs={'id': ['disqus_thread']}),
        dict(attrs={'class': ['middle-promo',
                              'sharing', 'mejs-player-holder']}),
        dict(name='a', onclick=lambda x: x and '__gaTracker' in x and 'outbound-article' in x),
    ]
    remove_tags_after = [
        dict(name='hr', attrs={'class': 'sticky-clear'}),
    ]

    def parse_spec_section(self, div):
        h2 = div.find('h2')
        sectitle = self.tag_to_string(h2)
        self.log('Section:', sectitle)
        articles = []
        for div in div.findAll('div', id=lambda x: x and x.startswith('post-')):
            h2 = div.find('h2', attrs={'class': class_sel('term-item__title')})
            if h2 is None:
                h2 = div.find(attrs={'class': class_sel('news-listing__title')})
            title = self.tag_to_string(h2)
            a = h2.find('a')
            url = a['href']
            desc = ''
            self.log('\tArticle:', title)
            p = div.find(attrs={'class': class_sel('term-item__excerpt')})
            if p is not None:
                desc = self.tag_to_string(p)
            articles.append({'title': title, 'url': url, 'description': desc})
        return sectitle, articles

    def parse_index(self):
        soup = self.index_to_soup('https://www.spectator.co.uk/magazine/')
        a = soup.find('a', attrs={'class': 'issue-details__cover-link'})
        self.timefmt = ' [%s]' % a['title']
        self.cover_url = a['href']
        if self.cover_url.startswith('//'):
            self.cover_url = 'http:' + self.cover_url

        feeds = []

        div = soup.find(attrs={'class': class_sel('content-area')})
        for x in div.findAll(attrs={'class': class_sel('magazine-section-holder')}):
            title, articles = self.parse_spec_section(x)
            if articles:
                feeds.append((title, articles))
        return feeds
lennyw is offline   Reply With Quote