Actually it would, each new file introduces more disk read time, file open, file close time. Then depending on the language it may have to create a new json parser, parse the file, and then destroy the json parser. With 1-2 files its not that big of a deal, but as it scales up the issue becomes more and more of a bottle neck
https://bugzilla.wikimedia.org/show_bug.cgi?id=63327
--- Comment #2 from Niklas Laxström <niklas.laxstrom@gmail.com> ---
It shouldn't matter too much whether N messages are in 50 or 1000 files (made
up numbers) on the time how much it takes to parse them.
--
You are receiving this mail because:
You are the assignee for the bug.
_______________________________________________
Pywikipedia-bugs mailing list
Pywikipedia-bugs@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/pywikipedia-bugs