Uptime objectives » History » Version 8
Nico Schottelius, 07/01/2019 07:41 PM
1 | 1 | Nico Schottelius | h1. Uptime objectives |
---|---|---|---|
2 | |||
3 | 2 | Nico Schottelius | {{toc}} |
4 | 1 | Nico Schottelius | |
5 | 4 | Nico Schottelius | h2. Uptime definitons |
6 | |||
7 | |||
8 | | % | Downtime / year | |
||
9 | | 99 | 87h or 3.65 days | |
||
10 | | 99.9 | 8.76h | |
||
11 | | 99.99 | 0.876h or 52.55 minutes | |
||
12 | | 99.999 | 5.25 minutes | |
||
13 | |||
14 | 1 | Nico Schottelius | h2. Power Supply |
15 | |||
16 | * What: Power supply to all systems |
||
17 | * Setup: |
||
18 | ** Core systems are connected to UPS that last between 7-30 minutes |
||
19 | 4 | Nico Schottelius | ** Virtualisation systems are not (yet) fully connected to UPS (to be finished 2019-07) |
20 | 8 | Nico Schottelius | * Expected outages |
21 | 4 | Nico Schottelius | ** Prior to full UPS installation: <= 24h/year (99%) |
22 | ** After UPS installation: 99.9% |
||
23 | 1 | Nico Schottelius | *** Probably less, as most power outages are <1m |
24 | 8 | Nico Schottelius | ** Values for 2020/2021 are estimated, need to be confirmed with actual power outages |
25 | * Uptime objective |
||
26 | ** 2019: 99% |
||
27 | ** 2020: 99.9% |
||
28 | ** 2021: 99.99% |
||
29 | 1 | Nico Schottelius | |
30 | 5 | Nico Schottelius | h2. L2 Internal Network |
31 | 1 | Nico Schottelius | |
32 | * What: The connection between servers, routers and switches. |
||
33 | * Setup: All systems are connected twice internally, usually via fiber |
||
34 | * Expected outages |
||
35 | ** Single switch outage: no outage, maybe short packet loss (LACP link detection might take some seconds) |
||
36 | ** Double switch outage: full outage, manual replacement |
||
37 | * Uptime objectives |
||
38 | ** From 2019: >= 99.999% |
||
39 | |||
40 | h2. L2 external Network |
||
41 | |||
42 | * What: the network between the different locations |
||
43 | * Setup: |
||
44 | ** Provided by local (electricity) companies. |
||
45 | ** No additional active equipment / same as internal network |
||
46 | * Expected outages |
||
47 | ** 1 in 2018 that could be bridged by Wifi |
||
48 | ** If an outage happens, it's long (digging through the cable) |
||
49 | ** But it happens very rarely |
||
50 | ** Mid term geo redundant lines planned |
||
51 | 4 | Nico Schottelius | ** Geo redundancy might be achieved starting 2020 |
52 | 1 | Nico Schottelius | * Uptime objectives |
53 | ** 2019: >= 99.99% |
||
54 | 4 | Nico Schottelius | ** 2020: >= 99.999% |
55 | ** 2021: >= 99.999% |
||
56 | 1 | Nico Schottelius | |
57 | |||
58 | h2. L3 external Network |
||
59 | |||
60 | * What: the external (uplink) networks |
||
61 | * Setup |
||
62 | 4 | Nico Schottelius | ** Currently 2 uplinks |
63 | ** Soon 2 individual plus a third central uplink |
||
64 | 1 | Nico Schottelius | * Expected outages |
65 | 4 | Nico Schottelius | ** 2019 added bgp support |
66 | ** Outage simulations still due |
||
67 | 1 | Nico Schottelius | * Uptime objectives |
68 | 4 | Nico Schottelius | ** 2019: >= 99.99% |
69 | ** 2020: >= 99.999% |
||
70 | ** 2021: >= 99.999% |
||
71 | 1 | Nico Schottelius | |
72 | |||
73 | h2. Routers |
||
74 | |||
75 | 4 | Nico Schottelius | * What: the central routers |
76 | * Setup |
||
77 | ** Two routers running Linux with keepalived |
||
78 | ** Both routers are rebooted periodically -> downtime during that time is critical, but unlikely |
||
79 | ** Routers are connected to UPS |
||
80 | ** Routers are running raid1 |
||
81 | * Expected outages |
||
82 | ** Machines are rather reliable |
||
83 | ** If one machines has to be replaced, replacement can be prepared while other routers are active |
||
84 | ** Rare events, nice 2017 no router related downtime |
||
85 | * Uptime objectives |
||
86 | ** 2019: >= 99.99% |
||
87 | ** 2020: >= 99.999% |
||
88 | ** 2021: >= 99.999% |
||
89 | 1 | Nico Schottelius | |
90 | 6 | Nico Schottelius | h2. VMs on servers |
91 | 1 | Nico Schottelius | |
92 | * What: Servers host VMs and in case of a defect VMs need to be restarted on a different server |
||
93 | * Setup: |
||
94 | ** Servers are dual power connected |
||
95 | ** Servers are used hardware |
||
96 | ** Servers are being monitored (prometheus+consul) |
||
97 | ** Not yet sure how to detect soon failng servers |
||
98 | ** So far 3 servers affected (out of about 30) |
||
99 | 6 | Nico Schottelius | ** Restart of a VM takes a couple of seconds, as data is distributed in ceph |
100 | ** Detection is not yet reliably automated -> needs to be finished in 2019 |
||
101 | 4 | Nico Schottelius | * Expected outages |
102 | ** At the moment servers "run until they die" |
||
103 | ** In the future servers should be periodically rebooted to detect broken hardware (live migrations enable this) |
||
104 | ** While a server downtime effects all VMs (up to 100 per server), it's a rare event |
||
105 | * Uptime objectives (per VM) |
||
106 | ** 2019: >= 99.99% |
||
107 | ** 2020: >= 99.999% |
||
108 | ** 2021: >= 99.999% |
||
109 | 7 | Nico Schottelius | |
110 | h2. Storage backends |
||
111 | |||
112 | * What: the ceph storage that contains the data of VMs and services |
||
113 | * Setup |
||
114 | ** A disk is striped into 4MB blocks |
||
115 | ** Each block is saved 3x |
||
116 | * Expected outages |
||
117 | ** Downtime happens at 3 failures at the same time in a near time window |
||
118 | ** 1 disk failure triggers instant replication |
||
119 | ** Disks (HDD, SSD) are ranging from 600GB to 10TB |
||
120 | ** Slow rebuild speed is around 200MB/s |
||
121 | ** Thus slowest rebuild window is 14.56h |
||
122 | * Uptime objectives (per image) |
||
123 | ** 2019: >= 99.999% |
||
124 | ** 2020: >= 99.999% |
||
125 | ** 2021: >= 99.999% |