I have only 2 comments:
1. Please nice https://en.wikipedia.org/wiki/Nice_(Unix) any heavy long running local processes, so that others can continue to use the machine.
2. For large data, consider using the Hadoop cluster! I think you are getting your data from the webrequest logs in Hadoop anyway, so you might as well continue to do processing there, no? If you do, you shouldn’t have to worry (too much) about resource contention: https://yarn.wikimedia.org/cluster/scheduler
:)
- Andrew Otto Systems Engineer, WMF
On Sat, Aug 12, 2017 at 2:20 PM, Erik Zachte ezachte@wikimedia.org wrote:
I will soon start the two Wikistats jobs which run for about several weeks each month,
They might use two cores each, one for unzip, one for perl.
How many cores are there anyway?
Cheers,
Erik
*From:* Analytics [mailto:analytics-bounces@lists.wikimedia.org] *On Behalf Of *Adrian Bielefeldt *Sent:* Saturday, August 12, 2017 19:44 *To:* analytics@lists.wikimedia.org *Subject:* [Analytics] Resources stat1005
Hello everyone,
I wanted to ask about resource allocation on stat1005. We https://meta.wikimedia.org/wiki/Research:Understanding_Wikidata_Queries need quite a bit since we process every entry in wdqs_extract and I was wondering how many cores and how much memory we can use without conflicting with anyone else.
Greetings,
Adrian
Analytics mailing list Analytics@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/analytics