Backing up the wiki: Difference between revisions
No edit summary |
No edit summary |
||
(6 intermediate revisions by the same user not shown) | |||
Line 1: | Line 1: | ||
=== |
=== WikiTeam3 === |
||
You can easily generate a database dump and file dump using the |
You can easily generate a database dump and file dump using Save the Web Project's [https://github.com/saveweb/wikiteam3/ WikiTeam3] Python script, (full instructions are at that link). |
||
{{note|<nowiki>Windows: When using --images, because NTFS does not allow characters such as :*?"<>| in filenames, some files may not be downloaded, please check the errors.log file.</nowiki>}} |
|||
The result will include an XML dump with full page history, a dump of all images and files along with associated descriptions and a siteinfo.json file containing information about features, such as the installed extensions and skins. |
|||
e.g. <code><nowiki>wikiteam3dumpgenerator https://WIKINAME.miraheze.org --xml --xmlrevisions --images --bypass-cdn-image-compression --force</nowiki></code> |
|||
To dump a private wiki you will have to use a login that has at least read permission on the wiki. |
To dump a private wiki you will have to use a login that has at least read permission on the wiki. |
||
e.g. <code><nowiki>wikiteam3dumpgenerator https://WIKINAME.miraheze.org --xml --xmlrevisions --images --bypass-cdn-image-compression --force --user USER --pass PASSWORD</nowiki></code> |
|||
⚫ | |||
⚫ | |||
After installing MediaWiki and extensions, in the shell use importDump.php to import the XML, this can take a long time. e.g. from the mediawiki folder<br /><code>php maintenance/importDump.php --conf LocalSettings.php --dry-run < ''your_dumpfile.xml''</code> |
|||
If that works repeat without --dry-run. It won't matter if the XML dump file has the file extension .gz or .bz2 (is compressed). |
|||
* --xml exports an XML dump, uses Special:Export by default when no other xmldump method is specified. |
|||
Due to the bug [https://phabricator.wikimedia.org/T206683 T206683] it may be necessary to also include <code>--user-prefix=""</code> in the command. |
|||
* --xmlrevisions uses API:Allrevisions xmldump method. Recommended as it's quicker and puts almost no pressure on the MediaWiki backend compared to Special:Export. |
|||
* --images generates an image dump |
|||
* --bypass-cdn-image-compression appends random parameters to URL when downloading image |
|||
* --force generates a dump even if there is one already at Internet Archive |
|||
If you encounter any problem with running the script, please [https://github.com/saveweb/wikiteam3/issues raise a new issue] at the Save the Web Project's [https://github.com/saveweb/wikiteam3/ saveweb/WikiTeam3] GitHub repository. |
|||
Afterwards use ImportImages.php to import the images<br /><code>php maintenance/importImages.php ''your_files''/</code> |
|||
⚫ | |||
Afterwards run <code>php maintenance/rebuildrecentchanges.php</code> in order to update the content of Special:Recentchanges. |
|||
⚫ | |||
== External links == |
== External links == |
Latest revision as of 10:18, 1 July 2024
WikiTeam3
You can easily generate a database dump and file dump using Save the Web Project's WikiTeam3 Python script, (full instructions are at that link).
Windows: When using --images, because NTFS does not allow characters such as :*?"<>| in filenames, some files may not be downloaded, please check the errors.log file.
e.g. wikiteam3dumpgenerator https://WIKINAME.miraheze.org --xml --xmlrevisions --images --bypass-cdn-image-compression --force
To dump a private wiki you will have to use a login that has at least read permission on the wiki.
e.g. wikiteam3dumpgenerator https://WIKINAME.miraheze.org --xml --xmlrevisions --images --bypass-cdn-image-compression --force --user USER --pass PASSWORD
- --xml exports an XML dump, uses Special:Export by default when no other xmldump method is specified.
- --xmlrevisions uses API:Allrevisions xmldump method. Recommended as it's quicker and puts almost no pressure on the MediaWiki backend compared to Special:Export.
- --images generates an image dump
- --bypass-cdn-image-compression appends random parameters to URL when downloading image
- --force generates a dump even if there is one already at Internet Archive
If you encounter any problem with running the script, please raise a new issue at the Save the Web Project's saveweb/WikiTeam3 GitHub repository.
Restoring from backup
See MediaWiki.org, specifically Manual:Importing XML dumps and Manual:importImages.php.
External links
- Backups, Miraheze Meta
- Manual:Backing up a wiki, MediaWiki