bassac.blogg.se

Kiwix zim file dump
Kiwix zim file dump










  1. #Kiwix zim file dump how to
  2. #Kiwix zim file dump archive
  3. #Kiwix zim file dump code
  4. #Kiwix zim file dump download

#Kiwix zim file dump download

Bulk download is (as of September 2013) available from mirrors but not offered directly from Wikimedia servers. Images and other uploaded media are available from mirrors in addition to being served directly from Wikimedia servers. Where are the uploaded files (image, audio, video, etc.)? These dumps are also available from the Internet Archive. simple, nostalgia) exist, with the same structure.

#Kiwix zim file dump code

The sub-directories are named for the language code and the appropriate project. org directory you will find the latest SQL and XML dumps for the projects, not just English.

#Kiwix zim file dump how to

See for info about such multistream files and about how to decompress them with python see also and related files for an old working toy. You could then either bzip2 decompress it or use bzip2recover, and search the first file for the article ID.

#Kiwix zim file dump archive

The first field of this index is the number of bytes to seek into the compressed archive 2, the second is the article ID, the third the article title.Ĭut a small part out of the archive with dd using the byte offset as found in the index. Each separate 'stream' (or really, file) in the multistream dump contains 100 pages, except possibly the last one.įor multistream, you can get an index file, 2. NOTE THAT the multistream dump file contains multiple bz2 'streams' (bz2 header, body, footer) concatenated together into one file, in contrast to the vanilla file which contains one stream. And it will unpack to ~5-10 times its original size. You might be tempted to get the smaller non-multistream archive, but this will be useless if you don't unpack it.

kiwix zim file dump

The only downside to multistream is that it is marginally larger. Your reader should handle this for you, if your reader doesn't support it it will work anyway since multistream and non-multistream contain the same xml. But with multistream, it is possible to get an article from the archive without unpacking the whole thing. So if you unpack either, you get the same data. 2 and 2 both contain the same xml contents. GET THE MULTISTREAM VERSION! (and the corresponding index file, 2) To download a subset of the database in XML format, such as a specific category or a list of articles see: Special:Export, usage of which is described at Help:Export.

kiwix zim file dump

Go to Latest Dumps and look out for all the files that have 'pages-meta-history' in their name. Please only download these if you know you can cope with this quantity of data.

  • All revisions, all pages: These files expand to multiple terabytes of text.
  • SQL files for the pages and links are also available.
  • all-titles-in-ns0.gz – Article titles only (with redirects).
  • 2 – Current revisions only, all pages (including talk).
  • 2 – Current revisions only, no talk or user pages this is probably what you want, and is approximately 18 GB compressed (expands to over 78 GB when decompressed).
  • Download the data dump using a BitTorrent client (torrenting has many benefits and reduces server load, saving bandwidth costs).
  • English Wikipedia dumps in SQL and XML: dumps.
  • Dumps from any Wikimedia Foundation project: dumps.
  • Where do I get it? English-language Wikipedia Some of them are mobile applications – see " list of Wikipedia mobile applications".
  • Wikipedia on rockbox: § Wikiviewer for Rockbox.
  • Selected Wikipedia articles as a printed book: Help:Books/Printed books.
  • Selected Wikipedia articles as a PDF, OpenDocument, etc.: Wikipedia:Books.
  • BzReader: § BzReader and MzReader (for Windows).
  • kiwix zim file dump

    Some of the many ways to read Wikipedia while offline:

  • 12.4 BzReader and MzReader (for Windows).
  • 12 Dynamic HTML generation from a local XML database dump.
  • 11 Static HTML tree dumps for mirroring or CD distribution.
  • 9.1 Doing Hadoop MapReduce on the Wikipedia current database dump.
  • 9 Help to parse dumps for use in scripts.
  • 7.2 Doing SQL queries on the current database dump.
  • 7 Why not just retrieve data from at runtime?.
  • 4 Where are the uploaded files (image, audio, video, etc.)?.











  • Kiwix zim file dump