Uptime Sanity Checking

coderman coderman at gmail.com
Fri Mar 9 06:01:16 UTC 2007


On 3/8/07, Nick Mathewson <nickm at freehaven.net> wrote:
> ...
> I think a fix_able_ cap probably gets us most of the benefit: if we
> change the cap, only the directory servers need to change their code
> or configuration.

seems reasonable; the nature of the network is going to vary (perhaps
significantly) with size and age...

as for a particular tunable cap:
2 nodes with uptime over 20 million
14 over 10 million
47 over 5 million
131 over 2 million
215 over 1 million
239 over 500k
456 over 200k
545 over 100k
647 over 50k
702 over 20k
753 over 10k
...

> Really, though, this is a band-aid, and we don't want to make it too
> sophisticated.  Remember that 'uptime' is a bad proxy for the property
> we want the 'Stable' flag to measure.  Mean time between failures
> would approximate stability better, I think.

agreed.  previously long lived instances are overly penalized for a restart.

are there scenarios where a restart indicates possible vulnerability
that make aversion useful?  for instance, a server seized/cracked,
keys copied, and a rogue node comes up in it's place?

(that is, could MTBF open up other attacks that are avoided by uptime
measurement - an email in the morning: "remove my node from the
directory, it's been compromised")



More information about the tor-dev mailing list