copyright_put.py :  » Network » Python-Wikipedia-Robot-Framework » pywikipedia » Python Open Source

Home
Python Open Source
1.3.1.2 Python
2.Ajax
3.Aspect Oriented
4.Blog
5.Build
6.Business Application
7.Chart Report
8.Content Management Systems
9.Cryptographic
10.Database
11.Development
12.Editor
13.Email
14.ERP
15.Game 2D 3D
16.GIS
17.GUI
18.IDE
19.Installer
20.IRC
21.Issue Tracker
22.Language Interface
23.Log
24.Math
25.Media Sound Audio
26.Mobile
27.Network
28.Parser
29.PDF
30.Project Management
31.RSS
32.Search
33.Security
34.Template Engines
35.Test
36.UML
37.USB Serial
38.Web Frameworks
39.Web Server
40.Web Services
41.Web Unit
42.Wiki
43.Windows
44.XML
Python Open Source » Network » Python Wikipedia Robot Framework 
Python Wikipedia Robot Framework » pywikipedia » copyright_put.py
# -*- coding: utf-8  -*-
"""
"""
__version__ = '$Id: copyright_put.py 7895 2010-01-27 07:36:13Z xqt $'

#
# (C) Francesco Cosoleto, 2006
#
# Distributed under the terms of the MIT license.
#

import sys, re, codecs, os, time, shutil
import wikipedia, config, date

from copyright import put,join_family_data,appdir,reports_cat

#
# Month + Year save method (e.g. User:BotName/Report_December_2007)
append_date_to_wiki_save_path = True

#
# Append day of month to wiki save path (e.g. User:BotName/Report_25_December_2007)
append_day_to_wiki_save_path = False

#
# Add pubblication date to entries (template:botdate)
append_date_to_entries = False

msg_table = {
    'en': {'_default': [u'New entries', u'New entries']},
    'es': {'_default': [u'Entradas nuevas', u'Entradas nuevas']},
    'it': {'_default': [u'Pagine nuove', u'Nuove voci'],
           'feed': [u'Aggiunte a voci esistenti', u'Testo aggiunto in']},
}

template_cat = {
  '_default': [u'This template is used by copyright.py, a script part of [[:m:Using the python wikipediabot|PyWikipediaBot]].', u''],
  'it': [u'Questo template  usato dallo script copyright.py del [[:m:Using the python wikipediabot|PyWikipediaBot]].', u'Template usati da bot'],
}

stat_msg = {
    'ar': [u'', u'', u'', u'', u'', ''],
    'en': [u'Statistics', u'Page', u'Entries', u'Size', u'Total', 'Update'],
    'es': [u'Estadsticas', u'Pgina', u'Entradas', u'Tamao', u'Total', u'Actualizacion'],
    'it': [u'Statistiche', u'Pagina', u'Segnalazioni', u'Lunghezza', u'Totale', u'Ultimo aggiornamento'],
}

separatorC = re.compile('(?m)^== +')

def get_wiki_save_page(stat_page = False):

    site = wikipedia.getSite()

    wiki_save_path = {
        '_default': u'User:%s/Report' % config.usernames[site.family.name][site.lang],
        'es': u'Usuario:%s/Reporte' % config.usernames[site.family.name][site.lang],
        'it': u'Utente:RevertBot/Report'
    }

    save_path = wikipedia.translate(site, wiki_save_path)

    if stat_page:
        return wikipedia.Page(site, '%s/%s' % (save_path, wikipedia.translate(site, stat_msg)[0]))

    if append_date_to_wiki_save_path:
        t = time.localtime()
        day = ''
        if append_day_to_wiki_save_path:
            day = '_' + str(t[2])

        save_path += day + '_' + date.monthName(site.language(), t[1]) + '_' + str(t[0])

    return wikipedia.Page(site, save_path)

def set_template(name = None):

    site = wikipedia.getSite()

    tcat = wikipedia.translate(site, template_cat)

    url = "%s://%s%s" % (site.protocol(), site.hostname(), site.path())

    botdate = u"""
<div style="text-align:right">{{{1}}}</div><noinclude>%s\n[[%s:%s]]</noinclude>
""" % (tcat[0], site.namespace(14), tcat[1])

    botbox = """
<div class=plainlinks style="text-align:right">[%s?title={{{1}}}&diff={{{2}}}&oldid={{{3}}} diff] - [%s?title={{{1}}}&action=history cron] - [%s?title=Special:Log&page={{{1}}} log]</div><noinclude>%s\n[[%s:%s]]</noinclude>
""" % (url, url, url, tcat[0], site.namespace(14), tcat[1])

    if name == 'botdate':
        p = wikipedia.Page(site, 'Template:botdate')
        if not p.exists():
            p.put(botdate, comment = 'Init.')

    if name == 'botbox':
        p = wikipedia.Page(site, 'Template:botbox')
        if not p.exists():
            p.put(botbox, comment = 'Init.')

def stat_sum(engine, text):
    return len(re.findall('(?im)^\*.*?' + engine + '.*?- ', text))

def get_stats():

    import catlib, pagegenerators

    msg = wikipedia.translate(wikipedia.getSite(), stat_msg)

    cat = catlib.Category(wikipedia.getSite(), 'Category:%s' % wikipedia.translate(wikipedia.getSite(), reports_cat))
    gen = pagegenerators.CategorizedPageGenerator(cat, recurse = True)

    output = u"""{| {{prettytable|width=|align=|text-align=left}}
! %s
! %s
! %s
! %s
! %s
! %s
|-
""" % ( msg[1], msg[2], msg[3], 'Google', 'Yahoo', 'Live Search' )

    gnt = 0 ; ynt = 0 ; mnt = 0 ; ent = 0 ; sn = 0 ; snt = 0

    for page in gen:
        data = page.get()

        gn = stat_sum('google', data)
        yn = stat_sum('yahoo', data)
        mn = stat_sum('(msn|live)', data)

        en = len(re.findall('=== \[\[', data))
        sn = len(data)

        gnt += gn ; ynt += yn ; mnt += mn ; ent += en ; snt += sn

        if en > 0:
            output += u"|%s||%s||%s KB||%s||%s||%s\n|-\n" % (page.aslink(), en, sn / 1024, gn, yn, mn)

    output += u"""|&nbsp;||||||||
|-
|'''%s'''||%s||%s KB||%s||%s||%s
|-
|colspan="6" align=right style="background-color:#eeeeee;"|<small>''%s: %s''</small>
|}
""" % (msg[4], ent, snt / 1024, gnt, ynt, mnt, msg[5], time.strftime("%d " + "%s" % (date.monthName(wikipedia.getSite().language(), time.localtime()[1])) + " %Y"))

    return output

def put_stats():
    page = get_wiki_save_page(stat_page = True)
    page.put(get_stats(), comment = wikipedia.translate(wikipedia.getSite(), stat_msg)[0])

def output_files_gen():
    for f in os.listdir(appdir):
        if 'output' in f and not '_pending' in f:
            m = re.search('output_(.*?)\.txt', f)
            if m:
                tag = m.group(1)
            else:
                tag = '_default'

            section_name_and_summary = wikipedia.translate(wikipedia.getSite(), msg_table)[tag]

            section = section_name_and_summary[0]
            summary = section_name_and_summary[1]

            yield os.path.join(appdir, f), section, summary

def read_output_file(filename):
    if os.path.isfile(filename + '_pending'):
        shutil.move(filename, filename + '_temp')
        ap = codecs.open(filename + '_pending', 'a', 'utf-8')
        ot = codecs.open(filename + '_temp', 'r', 'utf-8')
        ap.write(ot.read())
        ap.close()
        ot.close()
        os.remove(filename + '_temp')
    else:
        shutil.move(filename, filename + '_pending')

    f = codecs.open(filename + '_pending', 'r', 'utf-8')
    data = f.read()
    f.close()

    return data

def run(send_stats = False):
    page = get_wiki_save_page()

    try:
        wikitext = page.get()
    except wikipedia.NoPage:
        wikipedia.output("%s not found." % page.aslink())
        wikitext = '[[%s:%s]]\n' % (wikipedia.getSite().namespace(14), wikipedia.translate(wikipedia.getSite(), reports_cat))

    final_summary = u''
    output_files = list()

    for f, section, summary in output_files_gen():
        wikipedia.output('File: \'%s\'\nSection: %s\n' % (f, section))

        output_data = read_output_file(f)
        output_files.append(f)

        entries = re.findall('=== (.*?) ===', output_data)

        if not entries:
            continue

        if append_date_to_entries:
            dt = time.strftime('%d-%m-%Y %H:%M', time.localtime())
            output_data = re.sub("(?m)^(=== \[\[.*?\]\] ===\n)", r"\1{{botdate|%s}}\n" % dt, output_data)

        m = re.search('(?m)^==\s*%s\s*==' % section, wikitext)
        if m:
            m_end = re.search(separatorC, wikitext[m.end():])
            if m_end:
                wikitext = wikitext[:m_end.start() + m.end()] + output_data + wikitext[m_end.start() + m.end():]
            else:
                wikitext += '\n' + output_data
        else:
            wikitext += '\n' + output_data

        if final_summary:
            final_summary += ' '
        final_summary += u'%s: %s' % (summary, ', '.join(entries))

    if final_summary:
        wikipedia.output(final_summary + '\n')

        # if a page in 'Image' or 'Category' namespace is checked then fix
        # title section by adding ':' in order to avoid wiki code effects.

        wikitext = re.sub(u'(?i)=== \[\[%s:' % join_family_data('Image', 6), ur'=== [[:\1:', wikitext)
        wikitext = re.sub(u'(?i)=== \[\[%s:' % join_family_data('Category', 14), ur'=== [[:\1:', wikitext)

        # TODO:
        # List of frequent rejected address to improve upload process.

        wikitext = re.sub('http://(.*?)((forumcommunity|forumfree).net)',r'<blacklist>\1\2', wikitext)

        if len(final_summary)>=200:
            final_summary = final_summary[:200]
            final_summary = final_summary[:final_summary.rindex("[")-3] + "..."

        try:
            put(page, wikitext, comment = final_summary)
            for f in output_files:
                os.remove(f + '_pending')
                wikipedia.output("\'%s\' deleted." % f)
        except wikipedia.PageNotSaved:
            raise

        if append_date_to_entries:
            set_template(name =  'botdate')
        if '{{botbox' in wikitext:
            set_template(name =  'botbox')

    if send_stats:
        put_stats()

def main():
    #
    # Send statistics
    send_stats = False

    for arg in wikipedia.handleArgs():
        if arg == "-stats":
            send_stats = True
    run(send_stats = send_stats)

if __name__ == "__main__":
    try:
        main()
    finally:
        wikipedia.stopme()
www.java2java.com | Contact Us
Copyright 2009 - 12 Demo Source and Support. All rights reserved.
All other trademarks are property of their respective owners.