Ok, now I'm attempting to remove duplicate urls that might appear in multiple feeds that I'm aggregating.
I created a list called 'added_links' then overloaded the is_link_wanted method like so:
Code:
def is_link_wanted(self, url, tag):
wanted = False
if url not in self.added_links:
self.added_links.append(url)
wanted = True
return wanted
This seems to accurately catch duplicate urls. If I print out the "added_links" array, each url is only listed once. However, the duplicate articles/urls still appear in final ebook.