IP Scheme
The goal is to create a scheme which eases daily life. IP mapping should be more or less straightforward and extensible. For servers look at:
Management Network
For this we will utilize parts of the RFC1918 172.x.y.z/12 network. Thus, please use a netmask of 255.240.0.0 and NOT 255.255.0.0!
Note that 172.25.100.0/24 is special (legacy, historic reasons :()...
type/name |
IP address |
range/comment |
name |
rack CMC |
172.25.0.x |
x: 1..24 (one IP per two racks) |
rack01..rack24 |
open rack CMC |
172.25.0.x |
x: 101.. (one IP per open rack) |
open rack 1 ... |
LCP |
172.25.1.x |
x: 1..24 |
lcp01..lcp24 |
Allied Telesys Switches |
172.25.10.x |
x: 1..48 |
at01...at48 (physical rack number!) |
ProCurveManagement |
172.25.10.x |
x: 51,53,55 |
pr51...pr55 (Rack 51,53,55!) |
FS1000 A |
172.25.20.2 |
|
fs1000_a |
FS1000 B |
172.25.20.3 |
|
fs1000_b |
Management Core Switch |
172.25.20.40 |
|
|
FS500 |
172.25.21.x |
x: 1-24 |
fs500_01a, ....,10z |
Compute nodes |
172.26.x.y |
x: 1..13, y:0..99 |
mgn0001...mgn1344 |
IPMI Compute nodes |
172.27.x.y |
x: 1..13, y:0..99 |
ipn0001..ipn1344 |
IPMI head nodes |
172.28.0.x |
x:1..4 |
iph1..iph4 |
IPMI data nodes |
172.28.1.x |
x:1..31 |
ipd01..ipd31 |
Areca head nodes |
172.28.10.x |
x:1..4 |
arecah1..arecah4 |
ddc4000 and monitoring |
172.28.100.x |
1,2 |
special dedicated network for point to point connection between ddc4000 and monitoring server |
Areca data nodes |
172.28.11.x |
x:1..31 |
arecad01..arecad31 |
RAID misc. servers |
172.28.14.x |
x:1... |
raid-HOSTNAME should be the same as for 172.28.4.xx |
IPMI/ILOM SUN |
172.28.2.x |
x:1..13 |
ilom01..ilom13 |
Head nodes Mgmt |
172.28.20.x |
x:1..4 |
mgh1..mgh4 |
Data nodes Mgmt |
172.28.21.x |
x:1..32 |
mgd01..mgd31 |
server Mgmt |
172.28.22.x |
... |
mgXXXXX |
gpuXXX Mgmt |
172.28.23.x |
x:1..66 |
mggpuXXX |
HSM_TL |
172.28.3.0 |
mgtl |
mgtl |
HSM Management Switch |
172.28.3.250 |
hsmmgn |
hsmmgn |
ILOM/HSMNFS |
172.28.3.x |
iphsmx |
iphsm01..iphsm06 |
Server IPMI |
172.28.4.x |
x:1.. |
ipHOSTNAME (corresponds to 10.20.50.x) |
Special Server IPMI |
172.28.5.x |
x:1.. |
ipHOSTNAME (corresponds to 10.20.50.x) |
10k controller (1,2) |
172.28.6.(1,2) |
|
fc10k1 and fc10k2 |
Data Network
For this we will utilize a 10.X.Y.Z/8 network (also RFC1918). The corresponding scheme for the host names can be found in
hostnames .
The first free octet will tell what kind of thing we are talking about:
type |
Octet X |
Octet Y |
Octet Z |
result |
example |
Compute node |
10 |
hundreds |
singles |
10.10.H.S |
n1123 => 10.10.11.23 |
GPGPU |
12 |
0 |
singles |
10.12.0.S |
gpu001, gpu054 |
Test Nodes |
11 |
0 |
singles |
10.15.0.S |
gpu01 => 10.11.0.1 |
file server (Linux) |
20 |
10 |
singles |
10.20.10.S |
d31 => 10.20.10.31 |
file server (Sun ATLAS.Internal.X4500) |
20 |
20 |
singles |
10.20.20.S |
s09 => 10.20.20.9 |
head nodes |
20 |
30 |
singles |
10.20.30.S |
h03 => 10.20.30.2 |
internal nodes |
20 |
40 |
singles |
10.20.40.S |
|
special purpose nodes |
20 |
50 |
singles |
10.20.50.S |
e.g. einstein-dl, einstein-db |
TRX switch |
25 |
0 |
singles |
10.25.0.S |
rack05 => 10.25.0.5 |
By Computer Class
This section is only for "classes" of computers, i.e. not individual servers
type |
mgmt. network |
data network |
IPMI |
misc |
compute node |
172.26.x.y |
10.10.x.y |
172.27.x.y |
|
Sun thumper |
|
10.20.20.x |
172.28.2.x |
|
Head nodes |
172.28.20.x |
10.20.30.x |
172.28.0.x |
raid: 172.28.10.x |
Areca nodes |
172.28.21.x |
10.20.10.x |
172.28.1.x |
raid: 172.28.11.x |
test nodes and hardware
type |
IP range |
data network |
10.255.255.0/24 |
ipmi |
172.31.1.0/24 |
management |
172.31.0.0/24 |