【求助】维基百科怎么把他全部下载下来?
zhwiki dump progress on 20080625This is the Wikimedia dump service. Please read the copyrights information. See Meta:Data dumps for documentation on the provided data formats.
See all databases list.
Last dumped on 2008-06-20
Dump complete
Verify downloaded files against the MD5 checksums to check for corrupted files.
2008-06-26 04:43:26 done All pages with complete edit history (.7z)
Everything is Ok
These dumps can be *very* large, uncompressing up to 100 times the archive download size. Suitable for archival and statistical use, most mirror sites won't want or need this.
pages-meta-history.xml.7z 705.2 MB
2008-06-25 20:08:52 done All pages with complete page edit history (.bz2)
2008-06-25 20:08:52: zhwiki 612928 pages (12.644/sec), 6923068 revs (142.817/sec), 99.3% prefetched, ETA 2008-06-25 21:00:01
These dumps can be *very* large, uncompressing up to 20 times the archive download size. Suitable for archival and statistical use, most mirror sites won't want or need this.
pages-meta-history.xml.bz2 2.8 GB
2008-06-25 06:40:54 done All pages, current versions only.
2008-06-25 06:40:53: zhwiki 612928 pages (477.623/sec), 612928 revs (477.623/sec), 96.6% prefetched, ETA 2008-06-25 06:47:07
Discussion and user pages are included in this complete archive. Most mirrors won't want this extra material.
pages-meta-current.xml.bz2 297.5 MB
2008-06-25 06:19:27 done Articles, templates, image descriptions, and primary meta-pages.
2008-06-25 06:19:26: zhwiki 445322 pages (503.257/sec), 445322 revs (503.257/sec), 95.9% prefetched, ETA 2008-06-25 06:30:54
This contains current versions of article content, and is the archive most mirror sites will probably want.
pages-articles.xml.bz2 238.7 MB
2008-06-25 06:04:38 done Creating split stub dumps...
2008-06-25 06:04:35: zhwiki 612928 pages (385.398/sec), 6923068 revs (4353.096/sec), ETA 2008-06-25 06:06:15
These files contain no page text, only revision metadata.
stub-meta-history.xml.gz 222.5 MB
stub-meta-current.xml.gz 30.1 MB
stub-articles.xml.gz 23.3 MB
2008-06-25 05:37:42 done Extracted page abstracts for Yahoo
2008-06-25 05:37:42: zhwiki 612882 pages (35.730/sec), 612886 revs (35.730/sec), ETA 2008-06-25 07:01:01
abstract.xml 185.0 MB
abstract-zh-cn.xml 185.0 MB
abstract-zh-tw.xml 185.0 MB
2008-06-25 00:51:48 done List of page titles
all-titles-in-ns0.gz 1.8 MB
2008-06-25 00:51:47 done Redirect list
redirect.sql.gz 1.3 MB
2008-06-25 00:51:46 done Nonexistent pages that have been protected.
protected_titles.sql.gz 8 KB
2008-06-25 00:51:46 done Name/value pairs for pages.
page_props.sql.gz 3 KB
2008-06-25 00:51:46 done Newer per-page restrictions table.
page_restrictions.sql.gz 17 KB
2008-06-25 00:51:46 done Base per-page data (id, title, old restrictions, etc).
page.sql.gz 18.5 MB
2008-06-25 00:51:41 done Category information.
category.sql.gz 548 KB
2008-06-25 00:51:40 done User group assignments.
user_groups.sql.gz 1021 bytes
2008-06-25 00:51:40 done Data for various events (deletions, uploads, etc).
logging.sql.gz 22.9 MB
2008-06-25 00:51:31 done Set of defined interwiki prefixes and links for this wiki.
interwiki.sql.gz 7 KB
2008-06-25 00:51:31 done Wiki interlanguage link records.
langlinks.sql.gz 19.8 MB
2008-06-25 00:50:56 done Wiki external URL link records.
externallinks.sql.gz 15.3 MB
2008-06-25 00:50:39 done Wiki template inclusion link records.
templatelinks.sql.gz 7.6 MB
2008-06-25 00:50:26 done Wiki image usage records.
imagelinks.sql.gz 5.7 MB
2008-06-25 00:50:18 done Wiki category membership link records.
categorylinks.sql.gz 10.8 MB
2008-06-25 00:50:14 done Wiki page-to-page link records.
pagelinks.sql.gz 80.9 MB
2008-06-25 00:47:35 done Metadata on prior versions of uploaded images.
oldimage.sql.gz 236 KB
2008-06-25 00:47:35 done Metadata on current versions of uploaded images.
image.sql.gz 2.7 MB
2008-06-25 00:47:33 done A few statistics such as the page count.
site_stats.sql.gz 451 bytes
2008-06-25 00:47:33 done Update dataset for OAI updater system. (private)
2008-06-25 00:47:29 done Deleted page and revision data. (private)
2008-06-25 00:47:20 done Data for blocks of IP addresses, ranges, and users. (private)
2008-06-25 00:47:20 done Users' watchlist settings. (private)
2008-06-25 00:47:10 done User account data. (private)
----------------------------------------------------------------------------------------------------------------------------------------
这么多东西,不知道下载哪一个好 你可能需要1PB的硬盘吧 老大 没必要吧 找网站要数据拷一份 只要中文的就行了,维基提供的下载服务搞不来哦 用webaroo 去人家公司把硬盘拿走就是了。 其实babylon英文字典中的离线wikipedia(Englsih)字典所占几乎1G,可惜是比较老的版本喔。如果稀饭,不妨一试。 webaroo 好像不好使啊 维基百科的内容是不断更新的,下载一个镜像来看的话就失去了wiki的意义。而且现在维基百科不是可以正常访问了么?而且英文版的wiki仿佛从来就没有被封过吧。 中文版可使用for mdict的词典
在手机以及电脑上皆可 Posted by zhangyy1979 on 2008-8-3 21:44 http://www.ibmnb.com/images/common/back.gif
去人家公司把硬盘拿走就是了。
兄弟,还是你厉害!!! 没听说过这玩意有用吗. Posted by kopkop on 2008-8-4 15:57 http://www.ibmnb.com/images/common/back.gif
维基百科的内容是不断更新的,下载一个镜像来看的话就失去了wiki的意义。而且现在维基百科不是可以正常访问了么?而且英文版的wiki仿佛从来就没有被封过吧。
最多两个月又要封了,英文版也封的吧,包括HTTPS的,都不能访问,只能用工具访问啊 为什么要下载下来?应该简明版靠平时的一些使用了。 wiki的优势在于随时更新,都下载下来就没有意义了 召唤VPN不就解决访问问题了 wiki的优势在于随时更新,都下载下来就没有意义了
页:
[1]