Wikislices
In this context, a Wikislice is collection of articles pulled from Wikipedia for the WikiReader activity on the XO. The goal is to select from wikipedia well written, structured, and cited articles while excluding the rest. The entire english Wikipedia is very large and wouldn't fit on the XO. Nor are 1000+ articles on Pokemon characters important education materials for the developing world.
Questions
Universalism
The question of universal use of this content needs to be considered. Do we run this project under OLPC entirely? Or do we try to create logical bundles for anyone with a wikireader? What are our ideas that may differ from other Wikipedians?
Bundles
We are planning on shipping a general collection of material with the XO. Additional packages will be shipped with the XS School Server or available online. Collections could be added to a student's XO based on classroom assignments or simply a child's interest in a subject.
Health
- In conjunction with other Health Content, a wikislice of relevant health materials would be very useful.
- A bundle of Where There is No Doctor is underway, with a pdf - to - html conversion care of Pascals
Science
- An update to the Biology bundle with fungi and protists is slowly underway. Ditto an update clarifying licenses of the images (all cc-by)
- Something for a bug blitz would be most helpful, drawing on the above and related zipcodezoo and misha h's content.
Bundling scripts
Scripts that are currently used to make bundles:
HTML dumps
- wikiwix export (being built) : takes in a list of wikiwix entries, outputs ?
- wikiwix interface (being improved) : allow selection via firefox plugin of a set of articles for a collection
Summaries
- Summarize list : takes in a list of article titles, outputs a directory of one-paragraph html files with css. (by Zdenek, not published yet)
- Compress images : take a set of pages and images, reduce images according to a slider
- no images (remove altogether) v. hotlink images (include original thumbnail, alt text when offline)
- include first {0-10} images on a page, with metadata
- thumbnail only v. include full image (but not extra large) v. include all image sizes (full screen and more-than-fullscreen where available)
- bonus: assume local resize tool v. store 3 images for large instances
Older notes
see also the talk page
A Wikislice is a collection of materials gathered from a public wiki and packaged into a reusable form. Common examples are topical wikislices from Wikipedia, resulting in books such as the "Solar system" wikijunior text and various wikireaders. Tools used to make wikislices are regular expression toolkits.
See User:ZdenekBroz and the library grid for some examples.
WikiProject Wikislice
Please visit the project about wikislices on the english wikipedia.
Code libraries
- KsanaForge and their KsanaWiki project have a set of scripts that process raw xml dumps from MediaWiki. They are working on producing read-only flash drives and SD cards for distribution.
- Linterweb, developer of one of the freely-available static selections of Wikipedia, has an open source toolchain for building it; they are also working on wiki search engines (see Kiwix) and have offered to help build the local-filesystem search for the journal.
- The Moulinwiki project and Renaud Gaudin have a toolchain from processing html output from the MediaWiki parser. They are now combining forces with Linterweb.
- PediaPress has an "mwlib" library for parsing mediawiki text which is freely available
- the "Wikipedia 1.0" team and Andrew Cates (user:BozMo on en:wp) is using their own scripts to generate and review static collections from a list of constantly changing wiki articles.