There are a number of these tools available.
All the links will be converted, so you can open the main page, say, and then click on links and get to all the site. You could use a webcrawler tool which will save the site as HTML files. If the wiki is not part of Wikimedia or doesn't have a dump there is a project on github that downloads that wiki using its API: WikiTeam - We archive wikis, from Wikipedia to tiniest wikis (I tested the bellow article completely offline) However, when you are done, you will have a complete, recentĬopy of English Wikipedia with images that can fit on a 128GB SD card.īut the offline version is very much like the online version, includes photos etc: Numbers increase to 100GB of disk space and 30 hours of processing The import itself will require 80GB of disk space and five hours Setting all this up on your computer will not be a quick process. There are 13.9+ million pages with 20.0+ GB of text, as well as 3.7+ million thumbnails. If you want a complete mirror of Wikipedia (including images) full HTML formatting intact that will download in aprox 30 hours, you should use:Įnglish Wikipedia has a lot of data. Later edit for the case you want also the images offline: If too many people would do that it can flood the sites with requests. Taking Wikimedia stuff with wget is not good practice. As the comment specify in order to take other MediaWiki sites for kiwix mwoffliner can be used, it may not work with all since they may have custom differences but it is the only variant I came across. Or you can use Kiwix, faster to set up because it also provides the already processed dumps (. The experience is very similar to the browser experience. That file will be used by WikiTaxi program to search through articles. taxi(around 15Gb for Wikipedia) file out of the. You can take the 2 from the Wikimedia dumps site and process them with WikiTaxi (download in upper left corner).