Friday's events on RIPE Atlas
Dear all, Here is a description of the events that took place last Friday and some 'lessons learned' that we took away from it. The high level summary of events is that an Atlas user was authorised to create an extreme number of measurements involving a large number of probes. This effectively overloaded the back-end machinery in various different ways. Even though this event could only happen because of the exception in resource use limits, we have implemented workarounds and countermeasures to avoid a repetition in future, and we will be investigating some of the more fundamental issues in the coming period. More information is available below for those interested. Observations: - Problems started shortly before 11am when one of the Atlas users created a large number of new measurements each involving all available probes. The user had been given an exceptional amount of credits as part of a special experiment. Therefore the normal limitations on the impact any individual user can have on the system were not active when the measurements were created and activated. - The results of the newly created measurements put a lot of strain on the measurement scheduler, which triggered our interest. After some investigation the cause of the overload was identified and the related measurements were ended. - However, by this time the majority of results, up to that moment, had already reached our queuing servers and the consumers were already ingesting the results into our Hadoop storage platform. - At this phase we discovered a capacity problem with the process that consumes the Atlas results, so we doubled the capacity of that component on the fly. - This exposed the next bottleneck in our platform in the form of an accumulation of the created results on a very small number of processing nodes. Normally the incoming measurement results are distributed over several storage nodes, so this strongly reduced the consumption rate of new data. - A third factor that contributed was the fact that, in attempt to curb growth of the Atlas data, we have migrated the Atlas data sets to a more efficient compression algorithm earlier in the year. This saved us some 40-50% of storage space for the Atlas data, at the expense of some compute power. Under normal circumstances, even at high loads, this compute power is abundantly available on the storage cluster. Under the specific circumstances of last Friday's events, it turned out that the change of the compression algorithm had increased processing time for some Hadoop system tasks by up to a factor of 8, which had a direct impact on the data consumption speed. Immediate actions taken: - Removed special privileges of the end-user in question - Added capacity to the Atlas consumer processes - Returned (temporarily) to less efficient compression on the Atlas data sets. Lessons learned and further planned action: - Granting special privileges for some of the Atlas users needs (even) more attention than it already receives. - We need to better communicate "best practices" to these power users so they can use their extra allowances responsibly. - Improved compression of Atlas data has decreased our storage demands but also decreased our processing capacity. This needs further investigation to find the optimum configuration. - Investigate possibilities to better spread incoming results over more worker nodes (reduce hotspots). - Investigate and quantify reasonable boundaries of scalability of the whole system, to guide the limits for granting credits to end users. Kind regards, Romeo Zwart
On 9/22/15 5:34 PM, Romeo Zwart wrote:
- We need to better communicate "best practices" to these power users so they can use their extra allowances responsibly.
hi, who is a power user? maybe an anchor host or an atlas sponsor? thank you -- antonio
On 2015-09-22 17:56, Antonio Prado wrote:
On 9/22/15 5:34 PM, Romeo Zwart wrote:
- We need to better communicate "best practices" to these power users so they can use their extra allowances responsibly.
hi,
who is a power user? maybe an anchor host or an atlas sponsor?
thank you -- antonio
Hi, This was a Dutch researcher whose proposal (DNS anycast checks) was ultimately seen as useful, so we granted the higher than usual resource limits. Cheers, Robert
On Wed, 23 Sep 2015 09:05:24 +0200 Antonio Prado <thinkofit@gmail.com> wrote:
On 9/23/15 9:01 AM, Randy Bush wrote:
who is a power user?
i don't think it is productive to go down this path
randy, you read my mind
You mean not productive trying to place the blame, or not productive to have shady anonymous "power users" being arbitrarily and behind-the-scenes granted the whole atlas infrastructure as their personal toy to play with? ^^ -- With respect, Roman
who is a power user? i don't think it is productive to go down this path randy, you read my mind
You mean not productive trying to place the blame, or not productive to have shady anonymous "power users" being arbitrarily and behind-the-scenes granted the whole atlas infrastructure as their personal toy to play with? ^^
QED
Hi, On Wed, Sep 23, 2015 at 12:09:03PM +0500, Roman Mamedov wrote:
You mean not productive trying to place the blame, or not productive to have shady anonymous "power users" being arbitrarily and behind-the-scenes granted the whole atlas infrastructure as their personal toy to play with? ^^
I have always understood that if you need "more resources than your credits allowed", you could talk to the atlas team, explain your needs, and find a solution - so it seems this is what was done, as documented...? That it blew up the thing did obviously not work as planned, but the general option to have exceptions for tests that "really need all of the probes!" sounds useful to me. Document the process and criteria better? Maybe. Make a strict check-list of things, with 3 copies on paper required to circulate to all Atlas probe hosts? Certainly not. Balance between "things can be done" and "bureaucracy" must be found. gert -- have you enabled IPv6 on something today...? SpaceNet AG Vorstand: Sebastian v. Bomhard Joseph-Dollinger-Bogen 14 Aufsichtsratsvors.: A. Grundner-Culemann D-80807 Muenchen HRB: 136055 (AG Muenchen) Tel: +49 (0)89/32356-444 USt-IdNr.: DE813185279
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA256 On 2015/09/23 10:22 , Gert Doering wrote:
I have always understood that if you need "more resources than your credits allowed", you could talk to the atlas team, explain your needs, and find a solution - so it seems this is what was done, as documented...?
That it blew up the thing did obviously not work as planned, but the general option to have exceptions for tests that "really need all of the probes!" sounds useful to me.
Document the process and criteria better? Maybe.
By nature, researchers want to do things that nobody else is doing. And in many cases they want it all. More data is better. Unfortunately, such an experiment may overload parts of Atlas in unexpected ways. I take it that documenting the process and criteria better is for the most part something internal to the Atlas team. Philip -----BEGIN PGP SIGNATURE----- Version: GnuPG v2 iQIcBAEBCAAGBQJWAmmCAAoJEPr6076EDopyYdAP+wZs9jonz8qsX6sT1kWbu2aa ZHQZNjcLODxcf+KNkV4uRuhJzjsu0+2+CqDxI/ZV5q7caOhxtQCYmY5zKN3k9FVg tzLZxle8Qy8MEJeHg68KX/Sv0r3Wu/dvnqVnfwy2mOknuV/Dy4kSNCahMkkzJbxN W7cw3ynXzllwBGK078BzGodruRivk8vSPa8OJQDuocD5iR/1JRih6Zfp9aMX0HcU jnDpBy44k6f8NN2Yr9kbQ+Kn5j3MACL3vSwMYrm3wToY/GNnwE1appJ+eh9Q3W0B keyDK2S6VLwA4ned3VEP6kxgVmNnJ9uLmKnd0n2cIcq/yXzPU+rHOmPtJs0B1DhS SEzp+NrpYjoDl9fxPuKrOgyc5uC2r9VnpvyPnBuUQmV2QdViG42C2+dqr3xEMLqM Se/mixm+0gzFlq9nbKw4gaW4cTrfgRpNAdOtUbcTjbZr+EuQUmzzDzbnOutyaDlF 35UmAG+WhjovdxAWQ/RE3CDxMYETHzTX6iZuu+2gls8M6LXJedbO/JYOSgYtibsC XBBy8+Kd6dNLeB4xnPiNb3jpz1L58eRuHSKye5oVIglfEKJ9h+4CTMNx2iOQPZCB TMApEwnHCX4W3Jmzy9MvoYMlafLt3alCh6yzTT0ISNdoKOmE5KTLGWLrRRLhWETc 549XiqyOkklaQszsj30S =jRS5 -----END PGP SIGNATURE-----
On 23/09/2015 09:22, Gert Doering wrote:
Document the process and criteria better? Maybe.
+ learn lessons + move on + definitely don't stop doing interesting and exciting stuff just because there was a blip. Nick
a great post mortem, thanks! my uneducated thoughts o disk is cheap; don't compress o occasionally, large scale measurements in narrow time windows will be useful randy
On Wed, Sep 23, 2015 at 9:00 AM, Randy Bush <randy@psg.com> wrote:
o disk is cheap; don't compress
You need to update your cliches :-) Disk might be cheap, but I/O certainly isn't.
maybe compress the data info on the probes before it is sent and save to backend storage with the data already compress’d ? just a thought Colin
On 2015-09-23 12:29, Colin Johnston wrote:
maybe compress the data info on the probes before it is sent and save to backend storage with the data already compress’d ?
just a thought
Colin
It may be interesting to know, although not at all surprising: compressing individual results achieves almost nothing (and has the drawback of having to unpack/repack while the result moves in the pipeline). Compressing small chunks, say dozens of related results (e.g. from the same measurement) achieves 40-50% saving (a lot depends on what kind of result it is), while compressing large blocks (multi-megabyte chunks) can achieve 90-95% in most cases. It makes sense to compress data once it's at rest, especially if it's not often accessed. Therefore our processing pipeline is constructed such that it applies compression once a large enough batch has been consumed. As Romeo explained, this part was stretched a lot on Friday. Cheers, Robert
Hi all, quick techie question how much power is pulled via usb for the probev2 devices ? Colin
On 2015/09/23 14:10 , Colin Johnston wrote:
Hi all, quick techie question how much power is pulled via usb for the probev2 devices ?
Our highly specialized measuring equipment says... 0.22A :-) See attached picture.
participants (10)
-
Antonio Prado
-
Colin Johnston
-
Gert Doering
-
Mark Santcroos
-
Nick Hilliard
-
Philip Homburg
-
Randy Bush
-
Robert Kisteleki
-
Roman Mamedov
-
Romeo Zwart