I want to mention that data in Hadoop is only available from Aug 27th 2015. Older data is only available in mysql.

On Tue, Dec 15, 2015 at 11:27 AM, Roan Kattouw <rkattouw@wikimedia.org> wrote:
If the data is going to be retained but would just become harder to query (i.e. still in Hadoop but not in mysql), maybe we could nuke data that's more than a year old (or 6 months old or something) from mysql?

On Tue, Dec 15, 2015 at 9:35 AM, Andrew Otto <aotto@wikimedia.org> wrote:
We could blacklist this schema from the mysql database, and still keep producing it.  It would be available in Hadoop either way.


On Dec 15, 2015, at 12:22, Jonathan Morgan <jmorgan@wikimedia.org> wrote:

Hi Nuria,

FWIW: Although I'm not using this right now, but I could see it being useful for understanding the impact of new notification updates that are coming down the pike.[1][2]

What are the costs involved in keeping this schema up?

Best,

On Tue, Dec 15, 2015 at 8:22 AM, Nuria Ruiz <nuria@wikimedia.org> wrote:
Roan:

The data for Echo schema(https://meta.wikimedia.org/wiki/Schema:Echo) is quite large and we are not sure is even used. 

Can you confirm either way? If it is no longer used we will stop collecting it.


Thanks, 

Nuria

_______________________________________________
Analytics mailing list
Analytics@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/analytics




--
Jonathan T. Morgan
Senior Design Researcher
Wikimedia Foundation

_______________________________________________
Analytics mailing list
Analytics@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/analytics



_______________________________________________
Analytics mailing list
Analytics@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/analytics




--
--Madhu :)