Hello, I'm looking for the the best way to dump the information in a locally
installed Media Wiki so it can be subsequently parsed and then displayed in
a network visualization program. For an example of this kind of
visualization program please see the attached file, click here:
http://www.centennial-software.com/_gfx/screenshots/discovery_visual_map.jpg
or perform a google search for "network visualization"
This will show how each page is related and allow curators to keep an eye on
how the wiki is growing. I can write a parsing program myself and
visualization programs are freely available on the web. I'm simply
wondering what is the best way to dump everything into a static file.
Thanks for your time.
>> but i 'm not so good in pointer variable and how to follow to the
>> right function to change?
>
> Like this:
>
> class TiffHandler extends BitmapHandler {
> function getThumbType( $ext, $mime ) {
> return array( 'png', 'image/png' );
> }
> }
>
> $wgMediaHandlers['image/tiff'] = 'TiffHandler';
>
Perhaps this could be called BitmapAsPngHanlder and be included in the core? It
could be used for any image format ImageMagic can deal with. Hm, since it can
(maybe) even work for stuff like PS and WMF, it could even be called
ImageAsPngHandler.
-- Daniel
Hallo wikitech-l group,
can someone here look what is going on with the tif supported code, it doesn't convert anymore in 1.11.0 "MediaHandler::getHandler: no handler found for image/tiff." to a png thumbnail to support the browser
or is this corrected in the 1.12.0 code sources then i go to update it already , i tried to find but i am not a php programmer sorry.
or what are i doing wrong ? what can i check ?
thanks
kris
Op Thu, 13 Dec 2007 18:07:50 +0100
Kris Coryn (Mailinglists) <k.coryn.ml(a)telenet.be> Schreef:
> Thanks boris for doing the test.
>
> this is what i excepted from trying to debug and reading the sources,
> i have almost the same versions like boris, it is not a
> misconfiguration but the support for tif is changed from version
> 1.10.0 i checked the svn code for the 1.12.0 and that part isn't
> changed for the image/tiff code support.
>
> someone of the developers team ? that can help me, why probably there
> is no more support for tiff images for the moment?
>
>
> Kris
>
> Op Wed, 12 Dec 2007 21:01:59 -0500
> Boris Steipe <boris.steipe(a)utoronto.ca> Schreef:
>
> > Kris -
> >
> > No joy with that. Running MediaWiki: 1.11.0 PHP: 5.2.4
> > (apache2handler) and MySQL: 5.0.45; ImageMagick 6.3.6 10/20/07 Q16 :
> >
> > No thumbnail, no image, but file uploads oK since the clickable
> > link displays the image in my browser without problem.
> >
> > Tried the following:
> > Converted to uncompressed tiff - No
> > Converted to gif - Yes
> > Converted to png - Yes
> > Converted to jpeg - yes
> >
> > On my installation, the two tif files were not supported, the same
> > file as gif, png or jpeg was.
> >
> > HTH,
> > Boris
> >
> >
> > On 12-Dec-07, at 5:12 PM, Kris Coryn (Mailinglists) wrote:
> >
> > > Hallo,
> > >
> > > 't is possible that several person's who have access to a
> > > mediawiki 1.11.0 server do the simple test with a tiff file
> > > uploaded
> > >
> > > check the $wgFileExtensions = array( 'png', 'gif', 'jpg',
> > > 'jpeg','tif' ); for allow tif files to upload
> > >
> > > download this little tif file http://users.pandora.be/kris.coryn/
> > > YinYang_symbool_bw.tif
> > >
> > > and do the upload test , if you see the little thumbnail on a
> > > page than 't is ok for the mediawiki 1.11.0 software do the
> > > convertion from a tif image
> > > if you have a blank box (like me) than is it possible that
> > > 1.11.0 with the standard configuration doesn't convert a tiff
> > > image good anymore.
> > >
> > > I know tif is not a standard web image file , but i have all my
> > > scanned BW photo's in tiff format on my wiki ;-((
> > >
> > > thanks for your time to test it.
> > >
> > > kris
> > >
> > >
> > > Op Tue, 11 Dec 2007 22:07:23 +0100
> > > Kris Coryn (Mailinglists) <k.coryn.ml(a)telenet.be> Schreef:
> > >
> > >> Op Mon, 10 Dec 2007 23:05:26 -0800
> > >> Hiram Clawson <hiram(a)soe.ucsc.edu> Schreef:
> > >>
> > >>> You might try adding something like this to your local settings:
> > >>>
> > >>> $wgMimeDetectorCommand = "/usr/bin/file -bi";
> > >>>
> > >>> I had trouble similar to this after an upgrade.
> > >>
> > >> that's one of the thinks a have tested also installed
> > >> php-pecl-Fileinfo
> > >> http://www.mediawiki.org/wiki/Manual:FAQ#.
> > >> 22File_is_corrupt_or_has_an_invalid_extension.22
> > >>
> > >> but for me there is no mime type problem because in the
> > >> $wgDebugLogFile :
> > >>
> > >> MediaHandler::getHandler: no handler found for image/tiff.
> > >>
> > >> he knows it is a image/tiff but doesn't know how to handler this
> > >> format
> > >>
> > >> for example a png file that have already thumbnails of the good
> > >> size output this in the $wgDebugLogFile :
> > >>
> > >>
> > >> File::transform: Doing stat
> > >> for /var_data/www/localhost/htdocs/taijiquan.homelinux.net/wiki/
> > >> images/thumb/5/56/Yin_yang.svg.png/120px-Yin_yang.svg.png
> > >> BitmapHandler::doTransform: creating 120x120 thumbnail
> > >> at /var_data/www/localhost/htdocs/taijiquan.homelinux.net/wiki/
> > >> images/thumb/5/56/Yin_yang.svg.png/120px-
> > >> Yin_yang.svg.png
> > >>
> > >> if i remove the 120px image then the ImageMagick convertor do his
> > >> job like this
> > >>
> > >> File::transform: Doing stat
> > >> for /var_data/www/localhost/htdocs/taijiquan.homelinux.net/wiki/
> > >> images/thumb/5/56/Yin_yang.svg.png/120px-Yin_yang.svg.png
> > >> BitmapHandler::doTransform: creating 120x120 thumbnail
> > >> at /var_data/www/localhost/htdocs/taijiquan.homelinux.net/wiki/
> > >> images/thumb/5/56/Yin_yang.svg.png/120px-
> > >> Yin_yang.svg.png BitmapHandler::doTransform: running ImageMagick:
> > >> '/usr/bin/convert' -quality 95 -background white -size 120
> > >> '/var_data/www/localhost/htdocs/taijiquan.homelinux .net/wiki/
> > >> images/5/56/Yin_yang.svg.png'
> > >> -coalesce -thumbnail '120x120!' -depth 8
> > >> '/var_data/www/localhost/htdocs/taijiquan.homelinux.net/wiki/
> > >> images/thumb/5/
> > >> 56/Yin_yang.svg.png/120px-Yin_yang.svg.png' 2>&1 wfShellExec:
> > >> '/var_data/www/localhost/htdocs/taijiquan.homelinux.net/wiki/bin/
> > >> ulimit4.sh'
> > >> 30 102400 102400 ''\''/usr/bin/convert'\'' -quality 95
> > >> -background wh ite -size 120
> > >> '\''/var_data/www/localhost/htdocs/taijiquan.homelinux.net/wiki/
> > >> images/5/56/Yin_yang.svg.png'\''
> > >> -coalesce -thumbnail '\''120x120!'\'' -depth 8
> > >> '\''/var_data/www/localhost/htdocs/taijiquan.homelinux.net/wiki/
> > >> images/thumb/5/56/Yin_yang.svg.png/120px-Yin_yang.svg.png'\''
> > >> 2>&1'
> > >>
> > >> but no actions are taken to convert a image/tiff just one line :
> > >>
> > >> MediaHandler::getHandler: no handler found for image/tiff.
> > >>
> > >> this is defined in the DefaultSettings.php
> > >>
> > >> $wgMediaHandlers = array(
> > >> 'image/jpeg' => 'BitmapHandler',
> > >> 'image/png' => 'BitmapHandler',
> > >> 'image/gif' => 'BitmapHandler',
> > >> 'image/x-ms-bmp' => 'BmpHandler',
> > >> 'image/svg+xml' => 'SvgHandler', // official
> > >> 'image/svg' => 'SvgHandler', // compat
> > >> 'image/vnd.djvu' => 'DjVuHandler', // official
> > >> 'image/x.djvu' => 'DjVuHandler', // compat
> > >> 'image/x-djvu' => 'DjVuHandler', // compat
> > >> );
> > >>
> > >> but i don't know and find no doc's or faq what the Handler is
> > >> for a image/tiff ? so i can add this to the Localsettings.php
> > >> this is changed in version 1.10.0 maybe better in the upcoming
> > >> next version 1.12.0 ?
> > >>
> > >>
> > >>
> > >> Kris
> > >>
> > >>>
> > >>> --Hiram
> > >>>
> > >>> Kris Coryn (Mailinglists) wrote:
> > >>>> Hallo,
> > >>>>
> > >>>> i have upgraded my webserver to Mediawiki 1.11.0 , almost
> > >>>> perfect but i can't see my tiff documents. jpeg or png is no
> > >>>> problem so LocalSettings.php is good configured for using
> > >>>> ImageMagick to created thumbnails pictures
> > >>>>
> > >>>> after checking FAQ and debug with $wgDebugLogFile i found this
> > >>>> error for all the tiff images
> > >>>>
> > >>>> MediaHandler::getHandler: no handler found for image/tiff.
> > >>>>
> > >>>> i traced this back to wgMediaHandlers where no tiff images are
> > >>>> present
> > >>>>
> > >>>>
> > >>>> http://www.mediawiki.org/wiki/Manual:%24wgMediaHandlers say's
> > >>>>
> > >>>> $wgMediaHandlers = array(
> > >>>> 'image/jpeg' => 'BitmapHandler',
> > >>>> 'image/png' => 'BitmapHandler',
> > >>>> 'image/gif' => 'BitmapHandler',
> > >>>> 'image/x-ms-bmp' => 'BmpHandler',
> > >>>> 'image/svg+xml' => 'SvgHandler', // official
> > >>>> 'image/svg' => 'SvgHandler', // compat
> > >>>> 'image/vnd.djvu' => 'DjVuHandler', // official
> > >>>> 'image/x.djvu' => 'DjVuHandler', // compat
> > >>>> 'image/x-djvu' => 'DjVuHandler', // compat
> > >>>> );
> > >>>>
> > >>>> this wil results in the script wiki/includes/media/Generic.php
> > >>>> for the clas MediaHandler to say "no handler found for $type."
> > >>>>
> > >>>> how to configure the wiki for using tiff files again?
> > >>>> the previous version was mediawiki 1.8.2 that has no problems
> > >>>> with tiff images. wgMediaHandlers is introduced in version
> > >>>> 1.10.0
> > >>>>
> > >>>> thanks
> > >>>>
> > >>>> Kris
> > >>>>
> > >>>> _______________________________________________
>
> _______________________________________________
> MediaWiki-l mailing list
> MediaWiki-l(a)lists.wikimedia.org
> http://lists.wikimedia.org/mailman/listinfo/mediawiki-l
>
One week ago, Domas Mituzas started to publish hourly logfiles
that summarize the page views for each page of Wikipedia in all
languages. I've downloaded these files and tried to dig out some
interesting information. I also tried (user:LA2) to spark a
discussion on the "village pump" of some languages about how these
statistics could be made useful. See for example
[[cs:Wikipedie:Pod lípou]], [[de:Wikipedia:Café]],
[[nl:Wikipedia:De_kroeg]], [[ru:Википедия:Форум/Технический]],
[[sv:Wikipedia:Bybrunnen]].
The only clear answer for now is that a simple weekly summary of
the very most popular (a few hundred or so) pages could serve as a
to-do list for focusing improvement work. This is similar to the
existing "WikiCharts top 100" on the Toolserver,
http://tools.wikimedia.de/~leon/stats/wikicharts/
The new logfiles allow a much more accurate, up-to-date and deeper
analysis. Is anybody developing a Toolserver service for analyzing
and presenting such statistics?
>From "long tail" theory I borrowed the idea to use the logarithm
of the rank as the metric, and this seems to be very useful. The
actual page view count can vary, but many articles keep their
stable ranking over many days. When a page moves from rank 4711
to rank 471, that is a factor of 10 (logarithm +1.0), just as much
of an achievement as moving from rank 471 to rank 47.
The day-to-day trend analysis (see village pumps) is fascinating
to look at, for example how [[Ike Turner]] quickly became a
prominent article at the announcement of his death, but I don't
know if this analysis has any real practical use.
It would be nice to correlate page popularity to page quality.
However, I'm not sure how to determine page quality.
One simple metric could be the number of interlanguage links. If
an article is among the top 50 in one language, it probably
deserves being translated into other languages.
Perhaps the page count analysis should be combined with category
tree tools already available on the Toolserver? One could present
a category as a "tag cloud" with large fonts for the more popular
pages.
--
Lars Aronsson (lars(a)aronsson.se)
Aronsson Datateknik - http://aronsson.se
On 12/15/07, werdna(a)svn.wikimedia.org <werdna(a)svn.wikimedia.org> wrote:
> Log Message:
> -----------
> * New special page: Special:Protectedtitles, which lists titles protected from creation.
Wouldn't it make more sense to work this into Special:Protectedpages?
They're logically almost the same, and we have *way* too many special
pages. (I hope to merge all the "most linked" ones eventually, for
instance.)
I've been editing [[Commons:File types]] (the xiph.org press release
on Ogg and HTML5 directs to it) and see it lists allowable file
formats as of late 2006.
* What's the current list?
* Is there a handy place to look it up?
- d.
Hi,
I am quite new to wiki. I tried to write a dynamic extension of the
sidebar, which is context-sensitive, and when the user select an
article - it can react and adds some good links to the sidebar.
First I tried to insert a spec tag into the article, like
<sidebarExt context="projects"/>. My TAG parser reads the arcticle,
returns an empty string, and writes the context into the session
array. Then I attached a hook function to the SkinTemplateOutputPageBeforeExec,
to read the session and generate the Sidebar links.
First it worked well. But after a few going back and forth,
it seemed that the session value does not change. I think it is
because the cache of the sidebar and/or the cache of the article.
As far as I know, the sidebar cache can be disabled - the article
cache? I don't know.
I think it is not the best way, because the article cache is a good
thing, but in this case the article is not read always by the myTag
extension function, so it won't change the value. Other words - it
wont detect, that a change is needed in the sidebar.
Second thought is to add a new meta tag to the article, but I don't
know how to do that.
Anybody has a new idea how to solve this? Either adding a new meta
tag, either change my mind, and do it another way?
Thanks in advance...
Zoltan Hernyak
A question: what do our browser usage stats presently look like? Just
raw page hits per claimed user-agent. Has anyone collected the data on
this?
- d.
Hello,
we have now some kind of 'what pages are visited' statistics. It
isn't very trivial to separate exact pageviews, but this regular
expression should does the rough job :)
urlre = re.compile('^http://([^\.]+)\.wikipedia.org/wiki/([^?]+)')
It is applied to our squid access-log stream and redirected to
profiling agent (webstatscollector - then the hourly snapshots are
written in very trivial format.
This can be used to both noticing strange activities, as well as
spotting trends (specific events show up really nicely - let it be a
movie premiere, a national holiday or any scandal :). Last March,
when I was experimenting with it, it was impossible not to notice
that "300" did hit theatres, St.Patrick's day revealed Ireland, and
there was some crazy DDoS against us.
Anyway, log files for now are at:
http://dammit.lt/wikistats/
- didn't figure out yet retention policy, but as there're few gigs
available, at least few weeks should be up.
A normal snapshot contains ~3.5M page titles and extracted is over
100MB. Entries inside are grouped by project, and in semi-alphabetic
order.
I'm experimenting with visualization software too, so if you have any
ideas and are too lazy to implement - share them anyway :)
Cheers,
--
Domas Mituzas -- http://dammit.lt/ -- [[user:midom]]
I have had a lot of fun already, playing around with Domas' log
files posted in the last four days. However, the log files contain
parts of URLs that need to be decoded. Removing the underscore in
United_Kingdom is not a problem. Neither is decoding the correct
UTF-8 as in Sm%C3%B6rg%C3%A5sbord (Smörgåsbord). But for the
Russian Wikipedia, many URLs found in these log files are not
proper UTF-8. What method or algorithm should I use to decode
these URLs, and how can I tell them apart from the majority?
Does the MediaWiki software make assumptions about ISO 8859-1 for
Swedish or KOI-8 for Russian URLs?
Currently I use the following simple Perl code for decoding and
unifying URLs, running in an 8-bit binary environment:
$text =~ s/\+/_/g;
$text =~ s/%([A-Fa-f0-9][A-Fa-f0-9])/sprintf("%c",hex($1))/eg;
$text =~ s/ /_/g;
--
Lars Aronsson (lars(a)aronsson.se)
Aronsson Datateknik - http://aronsson.se