When downloading Wikipedia/Wikimedia Commons dumps, pages-meta-history.xml.7z and 2 are the same, but 7z is usually smaller (better compress ratio), so use 7z. Don't issue commands you don't understand, especially batch commands which use loops or find and xargs, unless you're ready to lose all the data you got. There is another site of MediaWiki dumps located here on Scott's website. When you've uploaded enough wikis, you'll probably be made a collection admin to save others the effort to move your stuff.įor a manually curated list, visit the download section on GitHub. If you want an item to land there, just upload it in "opensource" collection and remember the "WikiTeam" keyword, it will be moved at some point. Most of our dumps are in the wikiteam collection at the Internet Archive. UseModWiki: use wget/curl and raw mode (might have a different URL scheme, like this).wikipediadownloader.py to download Wikipedia dumps from : python wikipediadownloader.py.dumpgenerator.py to download MediaWiki wikis: python dumpgenerator.py -api= -xml -images.Tools and source code Official WikiTeam tools We're trying to decide which other wiki engines to work on: suggestions needed! Wikia - a website that allows the creation and hosting of wikis.But there is no image dump available, only the image descriptions.Wikimedia Commons - a wiki of media files available for free usage.The transfer of the dumps to the Internet Archive is automated and is currently managed by Hydriz.It offers public backups (also for sister projects): Wikipedia - arguably the largest and one of the oldest wikis on the planet.Also, you can edit those pages to link existing dumps! You'll help others focus their work. If you don't know where to start, pick a wiki which was not archived yet from the lists on WikiApiary. Remember that there are thousands of wikis we don't even know about yet. Please add a wiki to WikiApiary if you want someone to archive it sooner or later or tell us on IRC ( #wikiteam (on hackint)) if it's particularly urgent. Data from February 2013 and earlier saved. Allegedly pending move to Orain (which became offline too). ~100 wikis cannot be archived because they are private. Dumps were made in August 2013, January 2014 and August 2015.Ĭheck why there are dozens of wikis without dump. Last dumped in June 2022, is a broken wiki Help:Database download, Their dumping code Perhaps some pages are in the Wayback Machine. ![]() Uses MediaWiki 1.15.1 and therefore is not supported by dumpgenerator.py WARC dumps were done in July 2022. This wikifarm is not well covered in WikiApiary. Last dumped Mar/Apr 2022, fr. shows wrong XML when dumping There are Python scripts to generate those lists for many wikifarms. For a full list, please use WikiApiary wikifarms main page.īefore backing up a wikifarm, try to update the list of wikis for it. Here we only create pages for those we have some special information about that we don't want to lose (like archiving history and tips). There are also wikifarms with hundreds of wikis. Īs of 2019, our collection at Internet Archive holds dumps for 250,000 wikis (including independent, wikifarm wikis, some packages of wikis and Wikiedia). The oldest dumps are probably some 2001 dumps of Wikipedia when it used UseModWiki. ![]() By number of files: Wikimedia Commons (57 million), English Wikipedia (800,000).By number of pages: Wikimedia Commons (77 million), Wikidata (72 million), English Wikipedia (49 million), DailyWeeKee (35 million), WikiBusiness (22 million). ![]() Most wikis are small, containing about 100 pages or less, but there are some very large wikis: The URLs in this project were added to WikiApiary in the past too.Ī number of wikifarms have vanished and about 180 are still online. This list was imported into WikiApiary.Īccording to WikiIndex, there are 20,698 wikis. But it doesn't include 400,000+ Wikia wikis, and the independent list coverage can be improved for sure.Īccording to Pavlo's list generated in December 2008, there are 20,000 wikis. The total number of MediaWiki wikis is unknown, but some estimates exist.Īccording to WikiApiary, which is the most updated database, there are 21,139 independent wikis (1,718 are semantic) and 4,819 in wikifarms as of.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |