Hi,
As we all know, download.wikimedia.org is temporarily offline. Does
somebody have a recent stub-meta-history.xml available (any language
is okay)?
Best regards,
Diederik
Hello all:
I'm going to construct an ontology database with wikipedia. What I
want to do is importing datadumps into a database and then extract knowledge
from the databse. But I find a problem .As you know that Chinese contains
Simplified Chinese and Traditional Chinese. When I check the data in the
dumps, I find both Simplified Chinese and Traditional Chinese mixes
together. I don't know how to convert Traditional Chinese to Simplified
Chinese. Is that possible I use the datadumps to construct my ontology
database?
The datadumps I download is "zhwiki-20101014".
Thanks!
David
We noticed a kernel panic message and stack trace in the logs on the
server that servers XML dumps. The web server that provides access to
these files is temporarily out of commission; we hope to have it back on
line in 12 hours or less. Dumps themselves have been suspended while we
investigate. I hope to have an update on this tomorrow as well.
Ariel
I have download the "Database backup dumps" of chinese Edition. There are
files with XML and sql format. I want to have data all in database like
MySQL. Can I get this data (especially the XML format) to MySQL database
without using MediaWiki? How to do this if possible?
Where can I get the format details of each dump? Because I have read
contents in "zhwiki-20101014-pages-articles.xml" , but chiness have two
eddition: "Simplified Chinese" and “Traditional Chinese”. Both format exits
raffertily In file "zhwiki-20101014-pages-articles.xml" . I don't known how
to get rid it.
Thanks!
Y Mucho amor
--
Este mensaje le ha llegado mediante el servicio de correo electronico
que ofrece Infomed para respaldar el cumplimiento de las misiones del Sistem
a Nacional de Salud. La persona que envia este correo asume el compromiso de
usar el servicio a tales fines y cumplir con las regulaciones establecidas
Infomed: http://www.sld.cu/
Hi all;
Are there statistics about how many people download the dumps? Not only the
hits, also the completed downloads (is it possible?), if not, the wasted
bandwidth would be a good measure.
Regards,
emijrp
Saludos a Todos.Que es?
ondbzip2???? Es posible mejor ,para traduccion por favor.Gracias,Mr.Serguey
puede ayudar usted en las traducciones,por favor,he leido todo,y me parece
interesante.Dr.Juan Cesar Martinez
----- Original Message -----
From: <xmldatadumps-l-request(a)lists.wikimedia.org>
To: <xmldatadumps-l(a)lists.wikimedia.org>
Sent: Monday, October 04, 2010 2:21 PM
Subject: Xmldatadumps-l Digest, Vol 9, Issue 1
> Send Xmldatadumps-l mailing list submissions to
> xmldatadumps-l(a)lists.wikimedia.org
>
> To subscribe or unsubscribe via the World Wide Web, visit
> https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l
> or, via email, send a message with subject or body 'help' to
> xmldatadumps-l-request(a)lists.wikimedia.org
>
> You can reach the person managing the list at
> xmldatadumps-l-owner(a)lists.wikimedia.org
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of Xmldatadumps-l digest..."
>
>
> Today's Topics:
>
> 1. Re: Dumps, dumps, dumps (Jamie Morken)
> 2. testing one phase at a time parallelized en wiki dumps
> (Ariel T. Glenn)
> 3. Enwiki stopped (Andreas Meier)
> 4. Re: Enwiki stopped (Ariel T. Glenn)
> 5. Domas visits logs (emijrp)
> 6. Re: Domas visits logs (Ariel T. Glenn)
> 7. dataset1 maintenance Sat Oct 1 (dumps unavailable)
> (Ariel T. Glenn)
> 8. Re: dataset1 maintenance Sat Oct 1 (dumps unavailable)
> (Ariel T. Glenn)
> 9. posting (yllaermdm27(a)gmail.com)
> 10. Re: dataset1 maintenance Sat Oct 1 (dumps unavailable) (emijrp)
>
>
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Sun, 15 Aug 2010 09:13:23 -0700
> From: Jamie Morken <jmorken(a)shaw.ca>
> Subject: Re: [Xmldatadumps-l] Dumps, dumps, dumps
> To: "Ariel T. Glenn" <ariel(a)wikimedia.org>
> Cc: xmldatadumps-l(a)lists.wikimedia.org
> Message-ID: <cdc2e0a421fc9.4c67afb3(a)shaw.ca>
> Content-Type: text/plain; charset="iso-8859-1"
>
>
> Hi,
>
> ----- Original Message -----
> From: "Ariel T. Glenn" <ariel(a)wikimedia.org>
> Date: Sunday, August 15, 2010 12:15 am
> Subject: Re: [Xmldatadumps-l] Dumps, dumps, dumps
> To: Jamie Morken <jmorken(a)shaw.ca>
> Cc: emijrp <emijrp(a)gmail.com>, xmldatadumps-l(a)lists.wikimedia.org
>
> > Images take up 8T or more these days (of course that includes deletes
> > and earlier versions but those aren't the bulk of it).?
> > Hosting 8T
> > tarballs seems out of the question... who would download them anyways?
> >
> > Having said that, hosting small subsets of images is qute
> > possible and
> > is something that has been discussed in the past.? I would
> > love to hear
> > which subsets of images people want and would actually use.
>
> There is the script wikix that people have used to manually download
images from wikis:
>
> http://meta.wikimedia.org/wiki/Wikix
>
> It generates a list of all the images in an XML dump and then downloads
them.? The only thing missing is the image scaling, without that the enwiki
image dump will be too large for most people to use right now.? ImageMagick,
http://en.wikipedia.org/wiki/ImageMagick could work to scale the various
formats of images to smaller sizes.
>
> Here's a script snippet I found using it in the bash shell:
>
> #!/bin/sh
> find /media/SHAWN\ IPOD/Songs/ -iname "*.png"| while read file;
> do
> convert -size 75x75 "$file" -resize 100x100 "cover.bmp"
> cp cover.bmp "${file%/*}"/.
> done
>
> If wikimedia foundation provides a dump of images I think people will find
good ways to use them in interesting ways.? Dumps of enwiki images with a
max size of 640x480 or 800x600 and also enwiki thumbnails are the two
subsets I think would be most valuable.
>
> cheers,
> Jamie
>
>
> >
> > Ariel
> >
> >
>
He leido todo en absoluto.Hace falta involucrar mas La Wiki,y enlazarla mas.
Repito necesario elazar mas la wikipedia,atraer gente joven,voluntaria, no dejar que la Wikipedia ,toque fondo,Yo propongo muy respetuosamente enlazar con Facebook,Twitter,es el lugar mas vistado por los Jovenes,tambien revisar prensas,Blogs, y Los Bloguer por obligacion deben condicionar el anuncio de la Wikipedia.
Recordar este es y fue un trabajo excepcional.
Fuente de Info.
Respetuosamente.JCMS.
--
Este mensaje le ha llegado mediante el servicio de correo electronico
que ofrece Infomed para respaldar el cumplimiento de las misiones del Sistem
a Nacional de Salud. La persona que envia este correo asume el compromiso de
usar el servicio a tales fines y cumplir con las regulaciones establecidas
Infomed: http://www.sld.cu/
Hi;
Do you know more projects which publish public dumps? I know Wikimedia,
Wikia and Citizendium. Any more? I'm working in a tool for analysing dumps,
and I want to add support to all of them.
Thanks,
emijrp