#!/usr/bin/env python2 # -*- coding: utf-8 -*- # Copyright (C) 2011-2016 WikiTeam developers # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see . import datetime import json try: from hashlib import md5 except ImportError: # Python 2.4 compatibility from md5 import new as md5 import os import requests import shutil import time import unittest import urllib import urllib2 import tempfile from dumpgenerator import delay, domain2prefix, getImageNames, getPageTitles, getUserAgent, getWikiEngine, mwGetAPIAndIndex class TestDumpgenerator(unittest.TestCase): # Documentation # http://revista.python.org.ar/1/html/unittest.html # https://docs.python.org/2/library/unittest.html # Ideas: # - Check one wiki per wikifarm at least (page titles & images, with/out API) def test_delay(self): # This test checks several delays print '\n', '#'*73, '\n', 'test_delay', '\n', '#'*73 for i in [0, 1, 2, 3]: print 'Testing delay:', i config = {'delay': i} t1 = time.time() delay(config=config) t2 = time.time() - t1 print 'Elapsed time in seconds (approx.):', t2 self.assertTrue(t2 + 0.01 > i and t2 < i + 1) def test_getImages(self): # This test download the image list using API and index.php # Compare both lists in length and file by file # Check the presence of some special files, like odd chars filenames # The tested wikis are from different wikifarms and some alone print '\n', '#'*73, '\n', 'test_getImages', '\n', '#'*73 tests = [ # Alone wikis #['http://wiki.annotation.jp/index.php', 'http://wiki.annotation.jp/api.php', u'かずさアノテーション - ソーシャル・ゲノム・アノテーション.jpg'], ['https://archiveteam.org/index.php', 'https://archiveteam.org/api.php', u'Archive-is 2013-07-02 17-05-40.png'], #['http://skilledtests.com/wiki/index.php', 'http://skilledtests.com/wiki/api.php', u'Benham\'s disc (animated).gif'], # Editthis wikifarm # It has a page view limit # Gamepedia wikifarm #['http://dawngate.gamepedia.com/index.php', 'http://dawngate.gamepedia.com/api.php', u'Spell Vanquish.png'], # Neoseeker wikifarm #['http://digimon.neoseeker.com/w/index.php', 'http://digimon.neoseeker.com/w/api.php', u'Ogremon card.png'], # Orain wikifarm #['http://mc.orain.org/w/index.php', 'http://mc.orain.org/w/api.php', u'Mojang logo.svg'], # Referata wikifarm #['http://wikipapers.referata.com/w/index.php', 'http://wikipapers.referata.com/w/api.php', u'Avbot logo.png'], # ShoutWiki wikifarm #['http://commandos.shoutwiki.com/w/index.php', 'http://commandos.shoutwiki.com/w/api.php', u'Night of the Wolves loading.png'], # Wiki-site wikifarm #['http://minlingo.wiki-site.com/index.php', 'http://minlingo.wiki-site.com/api.php', u'一 (書方灋ᅗᅩ).png'], # Wikkii wikifarm # It seems offline ] session = requests.Session() session.headers = {'User-Agent': getUserAgent()} for index, api, filetocheck in tests: # Testing with API print '\nTesting', api config_api = {'api': api, 'delay': 0, 'retries': 5, 'date': "20150807", } config_api['path'] = tempfile.mkdtemp() req = urllib2.Request(url=api, data=urllib.urlencode({'action': 'query', 'meta': 'siteinfo', 'siprop': 'statistics', 'format': 'json'}), headers={'User-Agent': getUserAgent()}) f = urllib2.urlopen(req) imagecount = int(json.loads(f.read())['query']['statistics']['images']) f.close() print 'Trying to parse', filetocheck, 'with API' result_api = getImageNames(config=config_api, session=session) self.assertEqual(len(result_api), imagecount) self.assertTrue(filetocheck in [filename for filename, url, uploader in result_api]) # Testing with index print '\nTesting', index config_index = {'index': index, 'delay': 0, 'retries': 5, 'date': "20150807", } config_api['path'] = tempfile.mkdtemp() req = urllib2.Request(url=api, data=urllib.urlencode({'action': 'query', 'meta': 'siteinfo', 'siprop': 'statistics', 'format': 'json'}), headers={'User-Agent': getUserAgent()}) f = urllib2.urlopen(req) imagecount = int(json.loads(f.read())['query']['statistics']['images']) f.close() print 'Trying to parse', filetocheck, 'with index' result_index = getImageNames(config=config_index, session=session) #print 111, set([filename for filename, url, uploader in result_api]) - set([filename for filename, url, uploader in result_index]) self.assertEqual(len(result_index), imagecount) self.assertTrue(filetocheck in [filename for filename, url, uploader in result_index]) # Compare every image in both lists, with/without API c = 0 for filename_api, url_api, uploader_api in result_api: self.assertEqual(filename_api, result_index[c][0], u'{0} and {1} are different'.format(filename_api, result_index[c][0])) self.assertEqual(url_api, result_index[c][1], u'{0} and {1} are different'.format(url_api, result_index[c][1])) self.assertEqual(uploader_api, result_index[c][2], u'{0} and {1} are different'.format(uploader_api, result_index[c][2])) c += 1 def test_getPageTitles(self): # This test download the title list using API and index.php # Compare both lists in length and title by title # Check the presence of some special titles, like odd chars # The tested wikis are from different wikifarms and some alone print '\n', '#'*73, '\n', 'test_getPageTitles', '\n', '#'*73 tests = [ # Alone wikis ['https://archiveteam.org/index.php', 'https://archiveteam.org/api.php', u'April Fools\' Day'], #['http://skilledtests.com/wiki/index.php', 'http://skilledtests.com/wiki/api.php', u'Conway\'s Game of Life'], # Test old allpages API behaviour #['http://wiki.damirsystems.com/index.php', 'http://wiki.damirsystems.com/api.php', 'SQL Server Tips'], # Test BOM encoding #['http://www.libreidea.org/w/index.php', 'http://www.libreidea.org/w/api.php', 'Main Page'], ] session = requests.Session() session.headers = {'User-Agent': getUserAgent()} for index, api, pagetocheck in tests: # Testing with API print '\nTesting', api print 'Trying to parse', pagetocheck, 'with API' config_api = {'api': api, 'index': '', 'delay': 0, 'namespaces': ['all'], 'exnamespaces': [], 'date': datetime.datetime.now().strftime('%Y%m%d'), 'path': '.', 'retries': 5, } titles_api = getPageTitles(config=config_api, session=session) result_api = open(titles_api, 'r').read().decode('utf8').splitlines() os.remove(titles_api) self.assertTrue(pagetocheck in result_api) # Testing with index print 'Testing', index print 'Trying to parse', pagetocheck, 'with index' config_index = {'index': index, 'api': '', 'delay': 0, 'namespaces': ['all'], 'exnamespaces': [], 'date': datetime.datetime.now().strftime('%Y%m%d'), 'path': '.', 'retries': 5} titles_index = getPageTitles(config=config_index, session=session) result_index = open(titles_index, 'r').read().decode('utf8').splitlines() os.remove(titles_index) self.assertTrue(pagetocheck in result_index) self.assertEqual(len(result_api), len(result_index)) # Compare every page in both lists, with/without API c = 0 for pagename_api in result_api: chk = pagename_api in result_index self.assertEqual(chk, True, u'%s not in result_index'%(pagename_api)) c += 1 def test_getWikiEngine(self): print '\n', '#'*73, '\n', 'test_getWikiEngine', '\n', '#'*73 tests = [ ['https://www.dokuwiki.org', 'DokuWiki'], #['http://wiki.openwrt.org', 'DokuWiki'], #['http://skilledtests.com/wiki/', 'MediaWiki'], #['http://moinmo.in', 'MoinMoin'], ['https://wiki.debian.org', 'MoinMoin'], ['http://twiki.org/cgi-bin/view/', 'TWiki'], #['http://nuclearinfo.net/Nuclearpower/CurrentReactors', 'TWiki'], ['http://www.pmwiki.org/', 'PmWiki'], ['http://www.apfelwiki.de/', 'PmWiki'], ['http://wiki.wlug.org.nz/', 'PhpWiki'], #['http://wiki.greenmuseum.org/', 'PhpWiki'], #['http://www.cmswiki.com/tiki-index.php', 'TikiWiki'], ['http://www.wasteflake.com/', 'TikiWiki'], ['http://foswiki.org/', 'FosWiki'], ['http://www.w3c.br/Home/WebHome', 'FosWiki'], #['http://mojomojo.org/', 'MojoMojo'], #['http://wiki.catalystframework.org/wiki/', 'MojoMojo'], #['https://www.ictu.nl/archief/wiki.noiv.nl/xwiki/bin/view/Main', 'XWiki'], #['https://web.archive.org/web/20080517021020id_/http://berlin.xwiki.com/xwiki/bin/view/Main/WebHome', 'XWiki'], ['http://www.xwiki.org/xwiki/bin/view/Main/WebHome', 'XWiki'], ['https://confluence.atlassian.com/', 'Confluence'], #['https://wiki.hybris.com/dashboard.action', 'Confluence'], ['https://confluence.sakaiproject.org/', 'Confluence'], #['http://demo.bananadance.org/', 'Banana Dance'], ['http://wagn.org/', 'Wagn'], #['http://wiki.ace-mod.net/', 'Wagn'], #['https://success.mindtouch.com/', 'MindTouch'], #['https://jspwiki.apache.org/', 'JSPWiki'], ['http://www.ihear.com/FreeCLAS/', 'JSPWiki'], ['http://www.wikkawiki.org/HomePage', 'WikkaWiki'], #['http://puppylinux.org/wikka/', 'WikkaWiki'], ['https://www.cybersphere.net/', 'MediaWiki'], #['http://web.archive.org/web/20060717202033id_/http://www.comawiki.org/CoMa.php?CoMa=startseite', 'CoMaWiki'], ['http://bootbook.de/CoMa.php', 'CoMaWiki'], #['http://wikini.net/wakka.php', 'WikiNi'], ['http://wiki.raydium.org/wiki/', 'WikiNi'], #['http://wiki.cs.cityu.edu.hk/CitiWiki/SourceCode', 'CitiWiki'], #['http://wackowiki.sourceforge.net/test/', 'WackoWiki'], ['http://www.sw4me.com/wiki/', 'WackoWiki'], #['http://lslwiki.net/lslwiki/wakka.php', 'WakkaWiki'], ['http://kw.pm.org/wiki/index.cgi', 'Kwiki'], ['http://wiki.wubi.org/index.cgi', 'Kwiki'], #['http://perl.bristolbath.org/index.cgi', 'Kwiki'], #['http://www.anwiki.com/', 'Anwiki'], #['http://www.anw.fr/', 'Anwiki'], ['http://www.aneuch.org/', 'Aneuch'], ['http://doc.myunixhost.com/', 'Aneuch'], ['http://www.bitweaver.org/wiki/index.php', 'bitweaver'], ['http://wiki.e-shell.org/Home', 'Zwiki'], ['http://leo.zwiki.org/', 'Zwiki'], #['http://accessibility4all.wikispaces.com/', 'Wikispaces'], ['http://darksouls.wikidot.com/', 'Wikidot'], ['http://www.wikifoundrycentral.com/', 'Wetpaint'], ['http://wiki.openid.net/', 'PBworks'], ] for wiki, engine in tests: print 'Testing', wiki try: guess_engine = getWikiEngine(wiki) except ConnectionError: print "%s failed to load, skipping..." % (wiki) continue print 'Got: %s, expected: %s' % (guess_engine, engine) self.assertEqual(guess_engine, engine) def test_mwGetAPIAndIndex(self): print '\n', '#'*73, '\n', 'test_mwGetAPIAndIndex', '\n', '#'*73 tests = [ # Alone wikis ['https://archiveteam.org', 'https://archiveteam.org/api.php', 'https://archiveteam.org/index.php'], #['http://skilledtests.com/wiki/', 'http://skilledtests.com/wiki/api.php', 'http://skilledtests.com/wiki/index.php'], # Editthis wikifarm # It has a page view limit # Gamepedia wikifarm #['http://dawngate.gamepedia.com', 'http://dawngate.gamepedia.com/api.php', 'http://dawngate.gamepedia.com/index.php'], # Neoseeker wikifarm #['http://digimon.neoseeker.com', 'http://digimon.neoseeker.com/w/api.php', 'http://digimon.neoseeker.com/w/index.php'], # Orain wikifarm #['http://mc.orain.org', 'http://mc.orain.org/w/api.php', 'http://mc.orain.org/w/index.php'], # Referata wikifarm # ['http://wikipapers.referata.com', 'http://wikipapers.referata.com/w/api.php', 'http://wikipapers.referata.com/w/index.php'], # ShoutWiki wikifarm #['http://commandos.shoutwiki.com', 'http://commandos.shoutwiki.com/w/api.php', 'http://commandos.shoutwiki.com/w/index.php'], # Wiki-site wikifarm #['http://minlingo.wiki-site.com', 'http://minlingo.wiki-site.com/api.php', 'http://minlingo.wiki-site.com/index.php'], # Wikkii wikifarm # It seems offline ] for wiki, api, index in tests: print 'Testing', wiki api2, index2 = mwGetAPIAndIndex(wiki) self.assertEqual(api, api2) self.assertEqual(index, index2) if __name__ == '__main__': #copying dumpgenerator.py to this directory #shutil.copy2('../dumpgenerator.py', './dumpgenerator.py') unittest.main()