Benutzer:Florian K/category.py

aus Memory Alpha, der freien deutschen Star-Trek-Datenbank
Spring zu: Navigation, suche
# -*- coding: utf-8 -*-

"""
Scripts to manage categories.

Syntax: python category.py action [-option]

where action can be one of these:
 * add    - mass-add a category to a list of pages
 * remove - remove category tag from all pages in a category
 * move - move all pages in a category to another category
 * tidy   - tidy up a category by moving its articles into subcategories
 * tree   - show a tree of subcategories of a given category

and option can be one of these:
 * person  - sort persons by their last name (for action 'add')
 * episode - sort for episode-title (e.g. 'Koerperlosee Wesen, Das')
 * rebuild - reset the database

For the actions tidy and tree, the bot will store the category structure locally
in category.dump. This saves time and server load, but if it uses these data
later, they may be outdated; use the -rebuild parameter in this case.

For example, to create a new category from a list of persons, type:
    
  python category.py add -person
    
and follow the on-screen instructions.
"""

#
# (C) Rob W.W. Hooft, 2004
# (C) Daniel Herding, 2004
#
__version__ = '$Id: category.py,v 1.71 2005/03/09 09:46:44 a_engels Exp $'
#
# Distributed under the terms of the PSF license.
# 
import re, sys, pickle
import wikipedia, catlib, config, interwiki

# Summary messages
msg_add={
    'da':u'Robot: Tilføjer [[Kategori:%s]]',
    'de':u'Bot: Ergänze [[Kategorie:%s]]',
    'en':u'Robot: Adding [[Category:%s]]',
    'es':u'Bot: Añadida [[Categoría:%s]]',
    'is':u'Vélmenni: Bæti við [[Flokkur:%s]]',
    'pt':u'Bot: Adicionando [[Categoria:%s]]',
    }

msg_change={
    'da':u'Robot: Ændrer kategori %s',
    'de':u'Bot: Ändere Kategorie %s',
    'en':u'Robot: Changing category %s',
    'es':u'Bot: Cambiada categoría %s',
    'is':u'Vélmenni: Breyti flokknum [[Flokkur:%s]]',
    'nl':u'Bot: Wijziging Categorie %s',
    'pt':u'Bot: Modificando [[Categoria:%s]]',
    }

deletion_reason_move = {
    'de':u'Bot: Kategorie wurde nach %s verschoben',
    'en':u'Robot: Category was moved to %s',
    }

class CategoryDatabase:
    '''
    This is a temporary knowledge base saving for each category the contained
    subcategories and articles, so that category pages don't need to
    be loaded over and over again
    '''
    def __init__(self, rebuild = False, filename = 'category.dump'):
        if rebuild:
            self.rebuild()
        else:
            try:
                f = open(filename, 'r')
                databases = pickle.load(f)
                f.close()
                # keys are categories, values are 2-tuples with lists as entries.
                self.catContentDB = databases['catContentDB'] 
                # like the above, but for supercategories
                self.superclassDB = databases['superclassDB']
                del databases
            except:
                # If something goes wrong, just rebuild the database
                self.rebuild()
        
    def rebuild(self):
        self.catContentDB={}
        self.superclassDB={}
    
    def getSubcats(self, supercat):
        '''
        For a given supercategory, return a list of CatLinks for all its
        subcategories.
        Saves this list in a temporary database so that it won't be loaded from the
        server next time it's required.
        '''
        # if we already know which subcategories exist here
        if supercat in self.catContentDB:
            return self.catContentDB[supercat][0]
        else:
            subcatlist = supercat.subcategories()
            articlelist = supercat.articles()
            # add to dictionary
            self.catContentDB[supercat] = (subcatlist, articlelist)
            return subcatlist
    
    def getArticles(self, cat):
        '''
        For a given category, return a list of PageLinks for all its articles.
        Saves this list in a temporary database so that it won't be loaded from the
        server next time it's required.
        '''
        # if we already know which articles exist here
        if cat in self.catContentDB:
            return self.catContentDB[cat][1]
        else:
            subcatlist = cat.subcategories()
            articlelist = cat.articles()
            # add to dictionary
            self.catContentDB[cat] = (subcatlist, articlelist)
            return articlelist
    
    def getSupercats(self, subcat):
        # if we already know which subcategories exist here
        if subcat in self.superclassDB:
            return self.superclassDB[subcat]
        else:
            supercatlist = subcat.supercategories()
            # add to dictionary
            self.superclassDB[subcat] = supercatlist
            return supercatlist

    def dump(self, filename = 'category.dump'):
        '''
        Saves the contents of the dictionaries superclassDB and catContentDB to disk.
        '''
        # this is currently only used by print_treeview(). We might want to add it
        # for others like the tidy bot.
        f = open(filename, 'w')
        databases = {
            'catContentDB': self.catContentDB,
            'superclassDB': self.superclassDB
        }
        # store dump to disk in binary format
        pickle.dump(databases, f, bin=1)
        f.close()
        
def sorted_by_last_name(catlink, pagelink):
        '''
        given a CatLink, returns a CatLink which has an explicit sort key which
        sorts persons by their last names.
        Trailing words in brackets will be removed.
        Example: If category_name is 'Author' and pl is a PageLink to
        [[Alexandre Dumas (senior)]], this function will return this CatLink:
        [[Category:Author|Dumas, Alexandre]]
        '''
        page_name = pagelink.linkname()
        site = pagelink.site()
        # regular expression that matches a name followed by a space and
        # disambiguation brackets. Group 1 is the name without the rest.
        bracketsR = re.compile('(.*) \(.+?\)')
        match_object = bracketsR.match(page_name)
        if match_object:
            page_name = match_object.group(1)
        split_string = page_name.split(' ')
        if len(split_string) > 1:
            # pull last part of the name to the beginning, and append the rest after a comma
            # e.g. "John von Neumann" becomes "Neumann, John von"
            sorted_key = split_string[-1] + ', ' + ' '.join(split_string[:-1])
            # give explicit sort key
            return wikipedia.PageLink(site, catlink.linkname() + '|' + sorted_key)
        else:
            return wikipedia.PageLink(site, catlink.linkname())

def sorted_by_episode(catlink, pagelink):
        '''
        given a CatLink, returns a CatLink which has an explicit sort key which
        sorts persons by their last names.
        Trailing words in brackets will be removed.
        Example: If category_name is 'Author' and pl is a PageLink to
        [[Alexandre Dumas (senior)]], this function will return this CatLink:
        [[Category:Author|Dumas, Alexandre]]
        '''
        page_name = pagelink.linkname()
        site = pagelink.site()
        # getting rid of accents for natural sort-order instead of ASCII sort-order
        page_name = re.compile(r'[áàâ]').sub(r'a',page_name)
        page_name = re.compile(r'[éèê]').sub(r'e',page_name)
        page_name = re.compile(r'[íìî]').sub(r'i',page_name)
        page_name = re.compile(r'[óòô]').sub(r'o',page_name)
        page_name = re.compile(r'[úùû]').sub(r'u',page_name)
        page_name = re.compile(r'[ÁÀÂ]').sub(r'A',page_name)
        page_name = re.compile(r'[ÉÈÊ]').sub(r'E',page_name)
        page_name = re.compile(r'[ÍÌÎ]').sub(r'I',page_name)
        page_name = re.compile(r'[ÓÒÔ]').sub(r'O',page_name)
        page_name = re.compile(r'[ÚÙÛ]').sub(r'U',page_name)
        # regular expression that matches a name followed by a space and
        # disambiguation brackets. Group 1 is the name without the rest.
        bracketsR = re.compile('(.*) \(.+?\)')
        match_object = bracketsR.match(page_name)
        if match_object:
            page_name = match_object.group(1)
        split_string = page_name.split(' ')
        if len(split_string) > 1 and split_string[0] in ('Der', 'Die', 'Das', 'Ein', 'Eine'):
            # pull last part of the name to the beginning, and append the rest after a comma
            # e.g. "John von Neumann" becomes "Neumann, John von"
            sorted_key = ' '.join(split_string[1:]) + ', ' + split_string[0]
            sorted_key = sorted_key[0].upper() + sorted_key[1:]
            # give explicit sort key
            page_name = sorted_key
        else:
            print""
        page_name = re.compile(r'[Ä]').sub(r'Ae',page_name)
        page_name = re.compile(r'[Ö]').sub(r'Oe',page_name)
        page_name = re.compile(r'[Ü]').sub(r'Ue',page_name)
        page_name = re.compile(r'[ä]').sub(r'ae',page_name)
        page_name = re.compile(r'[ö]').sub(r'oe',page_name)
        page_name = re.compile(r'[ü]').sub(r'ue',page_name)
        page_name = re.compile(r'[ß]').sub(r'ss',page_name)
        if page_name != pagelink.linkname():
            return wikipedia.PageLink(site, catlink.linkname() + '|' + page_name)
            print page_name
        else:
            return wikipedia.PageLink(site, catlink.linkname())

def add_category(sort_by_last_name = False):
    '''
    A robot to mass-add a category to a list of pages.
    '''
    print "This bot has two modes: you can add a category link to all"
    print "pages mentioned in a List that is now in another wikipedia page"
    print "or you can add a category link to all pages that link to a"
    print "specific page. If you want the second, please give an empty"
    print "answer to the first question."
    listpage = wikipedia.input(u'Wikipedia page with list of pages to change:')
    if listpage:
        try:
            pl = wikipedia.PageLink(wikipedia.getSite(), listpage)
        except NoPage:
            wikipedia.output(u'The page ' + listpage + ' could not be loaded from the server.')
            sys.exit()
        pagenames = pl.links()
    else:
        refpage = wikipedia.input(u'Wikipedia page that is now linked to:')
        pl = wikipedia.PageLink(wikipedia.getSite(), refpage)
        pagenames = wikipedia.getReferences(pl)
    print "  ==> %d pages to process"%len(pagenames)
    print
    newcat = wikipedia.input(u'Category to add (do not give namespace):')
    newcat = newcat[:1].upper() + newcat[1:]

    # get edit summary message
    wikipedia.setAction(wikipedia.translate(wikipedia.getSite(), msg_add) % newcat)
    
    cat_namespace = wikipedia.getSite().category_namespaces()[0]

    answer = ''
    for nm in pagenames:
        pl2 = wikipedia.PageLink(wikipedia.getSite(), nm)
        if answer != 'a':
            answer = ''
            
        while answer not in ('y','n','a'):
            answer = wikipedia.input(u'%s [y/n/a(ll)]:' % (pl2.aslink()))
            if answer == 'a':
                confirm = ''
		while confirm not in ('y','n'):
	            confirm = wikipedia.input(u'This should be used if and only if you are sure that your links are correct! Are you sure? [y/n]:')
	
	if answer == 'y' or answer == 'a':
            try:
                cats = pl2.categories()
            except wikipedia.NoPage:
                wikipedia.output(u"%s doesn't exist yet. Ignoring."%(pl2.aslocallink()))
                pass
            except wikipedia.IsRedirectPage,arg:
                pl3 = wikipedia.PageLink(wikipedia.getSite(),arg.args[0])
                wikipedia.output(u"WARNING: %s is redirect to [[%s]]. Ignoring."%(pl2.aslocallink(),pl3.aslocallink()))
            else:
                wikipedia.output(u"Current categories:")
                for curpl in cats:
                    wikipedia.output(u"* %s" % pl.aslink())
                catpl = wikipedia.PageLink(wikipedia.getSite(), cat_namespace + ':' + newcat)
                if sort_by_last_name:
                    catpl = sorted_by_last_name(catpl, pl2) 
                if sort_by_episode:
                    catpl = sorted_by_episode(catpl, pl2) 
                if catpl in cats:
                    wikipedia.output(u"%s already has %s"%(pl2.aslocallink(), catpl.aslocallink()))
                else:
                    wikipedia.output(u'Adding %s' % catpl.aslocallink())
                    cats.append(catpl)
                    text = pl2.get()
                    text = wikipedia.replaceCategoryLinks(text, cats)
                    pl2.put(text)

class CategoryMoveRobot:
    def __init__(self, oldCatTitle, newCatTitle):
        self.oldCatTitle = oldCatTitle
        self.newCatTitle = newCatTitle
        # get edit summary message
        wikipedia.setAction(wikipedia.translate(wikipedia.getSite(),msg_change) % oldCatTitle)

    def run(self):
        old_cat = catlib.CatLink(wikipedia.getSite(), self.oldCatTitle)
        articles = old_cat.articles(recurse = 0)
        if len(articles) == 0:
            wikipedia.output(u'There are no articles in category ' + self.oldCatTitle)
        else:
            for article in articles:
                catlib.change_category(article, self.oldCatTitle, self.newCatTitle)
        
        subcategories = old_cat.subcategories(recurse = 0)
        if len(subcategories) == 0:
            wikipedia.output(u'There are no subcategories in category ' + self.oldCatTitle)
        else:
            for subcategory in subcategories:
                catlib.change_category(subcategory, self.oldCatTitle, self.newCatTitle)
        # try to copy page contents to new cat page
        if old_cat.copyTo(newCatTitle):
            if old_cat.isEmpty():
                reason = wikipedia.translate(wikipedia.getSite(), deletion_reason_move) % newCatTitle
                # TODO: only try to delete if bot has admin status
                old_cat.delete(reason)
            else:
                wikipedia.output('Couldn\'t copy contents of %s because %s already exists.' % (self.oldCatTitle, self.newCatTitle))

class CategoryRemoveRobot:
    '''
    Removes the category tag from all pages in a given category and from the
    category pages of all subcategories, without prompting.
    Doesn't remove category tags pointing at subcategories.
    '''
    deletion_reason_remove = {
        'de':u'Bot: Kategorie wurde aufgelöst',
        'en':u'Robot: Category was disbanded',
    }
    
    msg_remove={
        'da':u'Robot: Fjerner fra kategori %s',
        'de':u'Bot: Entferne aus Kategorie %s',
        'en':u'Robot: Removing from category %s',
        'es':u'Bot: Eliminada de la categoría %s',
        'is':u'Vélmenni: Fjarlægi [[Flokkur:%s]]',
        'nl':u'Bot: Verwijderd uit Categorie %s',
        'pt':u'Bot: Removendo [[Categoria:%s]]',
    }
    
    def __init__(self, catTitle):
        self.catTitle = catTitle
        # get edit summary message
        wikipedia.setAction(wikipedia.translate(wikipedia.getSite(), self.msg_remove) % self.catTitle)
        
    def run(self):
        cat = catlib.CatLink(wikipedia.getSite(), self.catTitle)
      
        articles = cat.articles(recurse = 0)
        if len(articles) == 0:
            wikipedia.output(u'There are no articles in category %s' % self.catTitle)
        else:
            for article in articles:
                catlib.change_category(article, self.catTitle, None)
        # Also removes the category tag from subcategories' pages 
        subcategories = cat.subcategories(recurse = 0)
        if len(subcategories) == 0:
            wikipedia.output(u'There are no subcategories in category %s' % self.catTitle)
        else:
            for subcategory in subcategories:
                catlib.change_category(subcategory, self.catTitle, None)
        # TODO: only try to delete if bot has admin status
        if cat.isEmpty():
            reason = wikipedia.translate(wikipedia.getSite(), self.deletion_reason_remove)
            cat.delete(reason)

class CategoryTidyRobot:
    """
    Script to help a human to tidy up a category by moving its articles into
    subcategories
    
    Specify the category name on the command line. The program will pick up the
    page, and look for all subcategories and supercategories, and show them with
    a number adjacent to them. It will then automatically loop over all pages
    in the category. It will ask you to type the number of the appropriate
    replacement, and perform the change robotically.
    
    If you don't want to move the article to a subcategory or supercategory, but to
    another category, you can use the 'j' (jump) command.
    
    Typing 's' will leave the complete page unchanged.
    
    Typing '?' will show you the first few bytes of the current page, helping
    you to find out what the article is about and in which other categories it
    currently is.
    
    Important:
     * this bot is written to work with the MonoBook skin, so make sure your bot
       account uses this skin
    """
    def __init__(self, catTitle, catDB):
        self.catTitle = catTitle
        self.catDB = catDB        

        # This is a purely interactive robot. We set the delays lower.
        wikipedia.get_throttle.setDelay(1)
        wikipedia.put_throttle.setDelay(10)
    
    def move_to_category(self, article, original_cat, current_cat):
        '''
        Given an article which is in category original_cat, ask the user if
        it should be moved to one of original_cat's subcategories.
        Recursively run through subcategories' subcategories.
        NOTE: current_cat is only used for internal recursion. You should
        always use current_cat = original_cat.
        '''
        print
        wikipedia.output(u'Treating page %s, currently in category %s' % (article.linkname(), current_cat.linkname()))
        subcatlist = self.catDB.getSubcats(current_cat)
        supercatlist = self.catDB.getSupercats(current_cat)
        print
        if len(subcatlist) == 0:
            print 'This category has no subcategories.'
            print
        if len(supercatlist) == 0:
            print 'This category has no supercategories.'
            print
        # show subcategories as possible choices (with numbers)
        for i,x in enumerate(supercatlist):
            # layout: we don't expect a cat to have more than 10 supercats
            wikipedia.output(u'u%d - Move up to %s' % (i, x.linkname()))
        for i,x in enumerate(subcatlist):
            # layout: we don't expect a cat to have more than 100 subcats
            wikipedia.output(u'%2d - Move down to %s' % (i, x.linkname()))
        print ' j - Jump to another category'
        print ' n - Skip this article'
        print ' r - Remove this category tag'
        print ' ? - Read the page'
        wikipedia.output(u'Enter - Save category as %s' % current_cat.linkname())

        length = 1000
        while True:
            print ''
            choice=wikipedia.input(u'Choice:')
            if choice == 'n':
                break
            elif choice == '':
                wikipedia.output(u'Saving category as %s' % current_cat.linkname())
                if current_cat == original_cat:
                    print 'No changes necessary.'
                else:
                    catlib.change_category(article, original_cat.catname(), current_cat.catname())
                break
            elif choice == 'j':
                newCatTitle = wikipedia.input(u'Please enter the category the article should be moved to:')
                newCat = catlib.CatLink(wikipedia.getSite(), newCatTitle)
                # recurse into chosen category
                self.move_to_category(article, original_cat, newCat)
                break
            elif choice == 'r':
                # remove the category tag
                catlib.change_category(article, original_cat.catname(), None)
                break
            elif choice == '?':
                print ''
                full_text = article.get()
                print ''
                wikipedia.output(full_text[0:length])
                
                # if categories possibly weren't visible, show them additionally
                # (maybe this should always be shown?)
                if len(full_text) > length:
                    print ''
                    print 'Original categories: '
                    for cat in article.categories(): 
                        wikipedia.output(u'* %s' % cat.linkname()) 
                    # show more text if the user uses this function again
                    length += 500
            elif choice[0] == 'u':
                try:
                    choice=int(choice[1:])
                except ValueError:
                    # user pressed an unknown command. Prompt him again.
                    continue
                self.move_to_category(article, original_cat, supercatlist[choice])
                break
            else:
                try:
                    choice=int(choice)
                except ValueError:
                    # user pressed an unknown command. Prompt him again.
                    continue
                # recurse into subcategory
                self.move_to_category(article, original_cat, subcatlist[choice])
                break
    
    def run(self):
        catlink = catlib.CatLink(wikipedia.getSite(), self.catTitle)
        
        # get edit summary message
        wikipedia.setAction(wikipedia.translate(wikipedia.getSite(), msg_change) % self.catTitle)
        
        articles = catlink.articles(recurse = 0)
        if len(articles) == 0:
            wikipedia.output(u'There are no articles in category ' + catTitle)
        else:
            for article in articles:
                print
                print '==================================================================='
                self.move_to_category(article, catlink, catlink)

class CategoryTreeRobot:
    '''
    Robot to create tree overviews of the category structure.
    
    Parameters:
        * catTitle - The category which will be the tree's root.
        * catDB    - A CategoryDatabase object
        * maxDepth - The limit beyond which no subcategories will be listed.
                     This also guarantees that loops in the category structure
                     won't be a problem.
        * filename - The textfile where the tree should be saved; None to print
                     the tree to stdout.
    '''
    
    def __init__(self, catTitle, catDB, filename = None, maxDepth = 10):
        self.catTitle = catTitle
        self.catDB = catDB
        self.filename = filename
        # TODO: make maxDepth changeable with a parameter or config file entry
        self.maxDepth = maxDepth
        
    def treeview(self, cat, currentDepth = 0, parent = None):
        '''
        Returns a multi-line string which contains a tree view of all subcategories
        of cat, up to level maxDepth. Recursively calls itself.
        
        Parameters:
            * cat - the CatLink of the node we're currently opening
            * currentDepth - the current level in the tree (for recursion)
            * parent - the CatLink of the category we're coming from
        '''
        
        # Translations to say that the current category is in more categories than
        # the one we're coming from
        also_in_cats = {
            'da': u'(ogsÃ¥ i %s)',
            'de': u'(auch in %s)',
            'en': u'(also in %s)',
            'is': u'(einnig í %s)',
            'pt': u'(também em %s)',
            }
            
        result = u'#' * currentDepth
        result += '[[:%s|%s]]' % (cat.linkname(), cat.linkname().split(':', 1)[1])
        result += ' (%d)' % len(self.catDB.getArticles(cat))
        # We will remove an element of this array, but will need the original array
        # later, so we create a shallow copy with [:]
        supercats = self.catDB.getSupercats(cat)[:]
        # Find out which other cats are supercats of the current cat
        try:
            supercats.remove(parent)
        except:
            pass
        if supercats != []:
            supercat_names = []
            for x in supercats:
                # create a list of wiki links to the supercategories
                supercat_names.append('[[:%s|%s]]' % (x.linkname(), x.linkname().split(':', 1)[1]))
                # print this list, separated with commas, using translations given in also_in_cats
            result += ' ' + wikipedia.translate(wikipedia.getSite(), also_in_cats) % ', '.join(supercat_names)
        result += '\n'
        if currentDepth < self.maxDepth:
            for subcat in self.catDB.getSubcats(cat):
                # recurse into subdirectories
                result += self.treeview(subcat, currentDepth + 1, parent = cat)
        else:
            if self.catDB.getSubcats(cat) != []:
                # show that there are more categories beyond the depth limit
                result += '#' * (currentDepth + 1) + '[...]\n'
        return result

    def run(self):
        """
        Prints the multi-line string generated by treeview or saves it to a file.
    
        Parameters:
            * catTitle - the title of the category which will be the tree's root
            * maxDepth - the limit beyond which no subcategories will be listed
        """
        cat = catlib.CatLink(wikipedia.getSite(), self.catTitle)
        tree = self.treeview(cat)
        if filename:
            wikipedia.output(u'Saving results in %s' % filename)
            import codecs
            f = codecs.open(filename, 'a', 'utf-8')
            f.write(tree)
            f.close()
        else:
            wikipedia.output(tree)

if __name__ == "__main__":
    try:
        catDB = CategoryDatabase()
        action = None
        sort_by_last_name = False
        restore = False
        for arg in sys.argv[1:]:
            arg = wikipedia.argHandler(arg)
            if arg:
                if arg == 'add':
                    action = 'add'
                elif arg == 'remove':
                    action = 'remove'
                elif arg == 'move':
                    action = 'move'
                elif arg == 'tidy':
                    action = 'tidy'
                elif arg == 'tree':
                    action = 'tree'
                elif arg == '-person':
                    sort_by_last_name = True
                elif arg == '-episode':
                    sort_by_episode = True
                elif arg == '-rebuild':
                    catDB.rebuild()
                
        if action == 'add':
            add_category(sort_by_last_name)
        elif action == 'remove':
            catTitle = wikipedia.input(u'Please enter the name of the category that should be removed:')
            bot = CategoryRemoveRobot(catTitle)
            bot.run()
        elif action == 'move':
            oldCatTitle = wikipedia.input(u'Please enter the old name of the category:')
            newCatTitle = wikipedia.input(u'Please enter the new name of the category:')
            bot = CategoryMoveRobot(oldCatTitle, newCatTitle)
            bot.run()
        elif action == 'tidy':
            catTitle = wikipedia.input(u'Which category do you want to tidy up?')
            bot = CategoryTidyRobot(catTitle, catDB)
            bot.run()
        elif action == 'tree':
            catTitle = wikipedia.input(u'For which category do you want to create a tree view?')
            filename = wikipedia.input(u'Please enter the name of the file where the tree should be saved, or press enter to simply show the tree:')
            bot = CategoryTreeRobot(catTitle, catDB, filename)
            bot.run()
        else:
            # show help
            wikipedia.output(__doc__, 'utf-8')
    finally:
        catDB.dump()
        wikipedia.stopme()