Backups
Miraheze heeft interne en externe procedures omtrent back-ups. Kritische toepassingen worden automatisch geback-upt naar een externe server (deze server is onder het beheer van Miraheze, maar wordt gehost door een andere provider en ander land ten opzichte van onze normale servers). Deze 'interne' back-ups bestaan uit volledige 'database dumps', welke weer bestaan uit onder andere gebruikersinformatie, artikelen en CheckUser-informatie. Deze back-ups kunnen alleen worden ingezien door de systeembeheerders zodat deze ze kunnen gebruiken om de website zo snel mogelijk online te krijgen in het geval van problemen.
As such, Miraheze has a total of three types of backups which are taken. On top of this, users may generate their own backups, quickly and conveniently on demand using our DataDump tool.
Backup types
Miraheze takes three types of backups to ensure as much resiliency as possible.
- Internal backups are backups kept on hand which the Site Reliability Engineering team can use to bring the entire site quickly up in the event of a catastrophic failure. These backups include full database dumps, which include user account information and CheckUser information. See the schedule below for more information.
- External backups are automatic backups kept on servers controlled by us but on a different host and in a different country. This is done to ensure that a failure on one host or in the power grid of one country, etc., doesn't cause extended downtime or data loss to our users. These types of backups include critical parts of our infrastructure such as the databases of all wikis, private Git repository data, Phorge configurations, and much more. See the schedule below for more information.
- Public backups are XML backups which we upload every month to archive.org of all public wikis. We do this to make sure we have a reliable backup of all wikis on an external site, along with to ensure users have peace of mind by seeing a readily available backup for usage by us/them.
General backup schedules
Up to date as of 12 January 2023.
Miraheze automatically runs the following backups for disaster recovery purposes:
- Internal/External
- De volgende gegevens worden elke zondag volledig geback-upt:
- Onze interne git repository (puppet1)
- Hierin worden wachtwoorden en private keys opgeslagen
- Ganglia (misc2)
- ACME (mw1)
- Hierin staan onze private keys en certificaten van wiki's die gebruik maken van SSL-certificaten uitgegeven door Let's Encrypt
- Onze interne git repository (puppet1)
- De volgende gegevens worden op de eerste zondag van de maand volledig ge-backupt. Op de derde zondag van de maand worden veranderingen (dus tussen de eerste en derde zondag van de maand) ge-backupt:
- Databases van alle wiki's (bevat o.a. gebruikersinformatie, artikelen, logboeken, etc.)
- Zgn. 'static' content (afbeeldingen, XML dumps)
- Phabricator static (afbeeldingen en documenten opgeslagen door onze bugtracker)
- Not currently ran: Static images for all wikis.
- Public
- Monthly: All public wikis; XML dumps uploaded to archive.org.
Manual backups
On top of our internal, external, and public backups, users may generate their own using different ways.
Users may upload their wiki dumps to Archive.org for a further level of backup. Please include "miraheze" in the list of subject tabs.
Back-ups van wiki's
Naast deze back-ups, welke privacygevoelige informatie bevatten, kunnen systeembeheerders op aanvraag XML-dumps beschikbaar stellen (via onze sites of e-mail, zowel voor openbare als privéwiki's). Deze back-ups bestaan uit artikelen en logboeken, en kunnen probleemloos worden geïmporteerd op elke wiki die de MediaWiki draait.
To use DataDump, go to Special:DataDump on your wiki and select what backup you want. Once you submit your request, your backups will be generated. Depending on the size of the wiki, it may take from a few seconds up to a few hours to generate a database dump.
DataDump API
DataDump offers an API module which lets users use DataDump via the command line. As yet, there are no scripts that make use of this.
Zie ook
Substitute WIKINAME with the wiki's subdomain.
The result will include an XML dump with full page history, a dump of all images and files along with associated descriptions and a siteinfo.json
file containing information about features, such as the installed extensions and skins.
Private wikis
To dump a private wiki you will have to use a login that has at least read permission on the wiki.
dumpgenerator --xml --xmlrevisions --images -g=--user -g=USER -g=--pass -g=PASSWORD https://WIKINAME.miraheze.org
Substitute WIKINAME with the wiki's subdomain, USER with the login name and PASSWORD with an actual password.
Issues
If you encounter any problem with running the script, please raise a new issue at the MediaWiki Scraper GitHub repository.
Restoring from backup
Templates, modules, CSS, Javascripts and Gadgets get imported with the XML. Boilerplates, abuse filters don't.
- Import XML via Special:RequestImportDump
- Import images and descriptions via create a task at Phorge
- Import ManageWiki settings (JSON) also via the Phorge task, (extensions, preferences, and so on).
- Configure any bots if required.
- Import any custom abuse filters.
- Check everything still functions as expected.
See also
- Bacula (the former backup system)