Saludos a Todos.Que es?
ondbzip2???? Es posible mejor ,para traduccion por favor.Gracias,Mr.Serguey
puede ayudar usted en las traducciones,por favor,he leido todo,y me parece
interesante.Dr.Juan Cesar Martinez
----- Original Message -----
From: <xmldatadumps-l-request(a)lists.wikimedia.org>
To: <xmldatadumps-l(a)lists.wikimedia.org>
Sent: Monday, October 04, 2010 2:21 PM
Subject: Xmldatadumps-l Digest, Vol 9, Issue 1
> Send Xmldatadumps-l mailing list submissions to
> xmldatadumps-l(a)lists.wikimedia.org
>
> To subscribe or unsubscribe via the World Wide Web, visit
> https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l
> or, via email, send a message with subject or body 'help' to
> xmldatadumps-l-request(a)lists.wikimedia.org
>
> You can reach the person managing the list at
> xmldatadumps-l-owner(a)lists.wikimedia.org
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of Xmldatadumps-l digest..."
>
>
> Today's Topics:
>
> 1. Re: Dumps, dumps, dumps (Jamie Morken)
> 2. testing one phase at a time parallelized en wiki dumps
> (Ariel T. Glenn)
> 3. Enwiki stopped (Andreas Meier)
> 4. Re: Enwiki stopped (Ariel T. Glenn)
> 5. Domas visits logs (emijrp)
> 6. Re: Domas visits logs (Ariel T. Glenn)
> 7. dataset1 maintenance Sat Oct 1 (dumps unavailable)
> (Ariel T. Glenn)
> 8. Re: dataset1 maintenance Sat Oct 1 (dumps unavailable)
> (Ariel T. Glenn)
> 9. posting (yllaermdm27(a)gmail.com)
> 10. Re: dataset1 maintenance Sat Oct 1 (dumps unavailable) (emijrp)
>
>
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Sun, 15 Aug 2010 09:13:23 -0700
> From: Jamie Morken <jmorken(a)shaw.ca>
> Subject: Re: [Xmldatadumps-l] Dumps, dumps, dumps
> To: "Ariel T. Glenn" <ariel(a)wikimedia.org>
> Cc: xmldatadumps-l(a)lists.wikimedia.org
> Message-ID: <cdc2e0a421fc9.4c67afb3(a)shaw.ca>
> Content-Type: text/plain; charset="iso-8859-1"
>
>
> Hi,
>
> ----- Original Message -----
> From: "Ariel T. Glenn" <ariel(a)wikimedia.org>
> Date: Sunday, August 15, 2010 12:15 am
> Subject: Re: [Xmldatadumps-l] Dumps, dumps, dumps
> To: Jamie Morken <jmorken(a)shaw.ca>
> Cc: emijrp <emijrp(a)gmail.com>, xmldatadumps-l(a)lists.wikimedia.org
>
> > Images take up 8T or more these days (of course that includes deletes
> > and earlier versions but those aren't the bulk of it).?
> > Hosting 8T
> > tarballs seems out of the question... who would download them anyways?
> >
> > Having said that, hosting small subsets of images is qute
> > possible and
> > is something that has been discussed in the past.? I would
> > love to hear
> > which subsets of images people want and would actually use.
>
> There is the script wikix that people have used to manually download
images from wikis:
>
> http://meta.wikimedia.org/wiki/Wikix
>
> It generates a list of all the images in an XML dump and then downloads
them.? The only thing missing is the image scaling, without that the enwiki
image dump will be too large for most people to use right now.? ImageMagick,
http://en.wikipedia.org/wiki/ImageMagick could work to scale the various
formats of images to smaller sizes.
>
> Here's a script snippet I found using it in the bash shell:
>
> #!/bin/sh
> find /media/SHAWN\ IPOD/Songs/ -iname "*.png"| while read file;
> do
> convert -size 75x75 "$file" -resize 100x100 "cover.bmp"
> cp cover.bmp "${file%/*}"/.
> done
>
> If wikimedia foundation provides a dump of images I think people will find
good ways to use them in interesting ways.? Dumps of enwiki images with a
max size of 640x480 or 800x600 and also enwiki thumbnails are the two
subsets I think would be most valuable.
>
> cheers,
> Jamie
>
>
> >
> > Ariel
> >
> >
>
He leido todo en absoluto.Hace falta involucrar mas La Wiki,y enlazarla mas.
Repito necesario elazar mas la wikipedia,atraer gente joven,voluntaria, no dejar que la Wikipedia ,toque fondo,Yo propongo muy respetuosamente enlazar con Facebook,Twitter,es el lugar mas vistado por los Jovenes,tambien revisar prensas,Blogs, y Los Bloguer por obligacion deben condicionar el anuncio de la Wikipedia.
Recordar este es y fue un trabajo excepcional.
Fuente de Info.
Respetuosamente.JCMS.
--
Este mensaje le ha llegado mediante el servicio de correo electronico
que ofrece Infomed para respaldar el cumplimiento de las misiones del Sistem
a Nacional de Salud. La persona que envia este correo asume el compromiso de
usar el servicio a tales fines y cumplir con las regulaciones establecidas
Infomed: http://www.sld.cu/
Hi;
Do you know more projects which publish public dumps? I know Wikimedia,
Wikia and Citizendium. Any more? I'm working in a tool for analysing dumps,
and I want to add support to all of them.
Thanks,
emijrp
Hi;
Some weeks ago, I read about WMF had downloaded every hour log from the
Domas website. In Internet Archive are only in the date range from December
2007 to September 2009. Now, at Domas website, are available the last few
months (from April 2010 to now). So, the dta from October 2009 to March 2010
is missing.
Can be enabled a new section in download.wikimedia.org with a link to the
directory where WMF saves a copy of these logs?
Regards,
emijrp
Folks,
The server that hosts the XML dumps will be undergoing maintenance (it's
going to be moved to another rack), on Saturday Oct 1 starting at about
15:00 GMT. We expect the server to be back up by 17:00 GMT. During
that time XML dumps will be unavailable.
In other news the first run of the full en.wikipedia history in chunks
has completed. The recompression to 7z has not been done, nor the
recompression into a single large bz2 file for people who prefer it.
However, for those interested, please have a look at the files:
http://dumps.wikimedia.org/enwiki/20100904/
Each file has its own mediawiki header and footer, each covering a range
of 2 million (sequential) page IDs, except for the last "chunk" which
covers rather more than it should.
As you can see, the chunk sizes are rather disparate. The next such run
should split up more evenly with roughly the same number of revisions in
each chunk, and as such, they should all take nearly the same time to
complete.
Ariel Glenn