Project

General

Profile

ОБЗОР » History » Version 8

Dmitry Chernyak, 11/29/2012 02:16 AM

1 1 Dmitry Chernyak
h1. OVERVIEW
2 1 Dmitry Chernyak
3 1 Dmitry Chernyak
{{toc}}
4 1 Dmitry Chernyak
5 3 Владимир Ипатов
[[INSTALL]] | [[BUILD-ISO]] | [[SETUP]] | [[OPERATIONS]] | [[GITMAGIC]] | [[LICENSE]] | [[STATUS]]
6 8 Dmitry Chernyak
на Русском: [[ОБЗОР]] | [[СБОРКА-ISO]] | [[УСТАНОВКА]] | [[НАСТРОЙКА]] | [[ОПЕРАЦИИ]] | [[МАГИЯ GIT]] | [[ЛИЦЕНЗИЯ]] | [[СОСТОЯНИЕ]]
7 1 Dmitry Chernyak
8 1 Dmitry Chernyak
h2. THE PROJECTS
9 1 Dmitry Chernyak
10 1 Dmitry Chernyak
"SkyCover Infrastructure" (*SCI*) is the multipurpose, mostly automated high
11 1 Dmitry Chernyak
availability virtual server infrastructure, equipped with automatic monitoring,
12 1 Dmitry Chernyak
backup and audit processes.
13 1 Dmitry Chernyak
14 1 Dmitry Chernyak
"SCI-CD" is the deployment engine for cheap and flexible virtual high
15 1 Dmitry Chernyak
availability cluster, the ground level for *SCI*.
16 1 Dmitry Chernyak
17 1 Dmitry Chernyak
h2. INTRODUCTION
18 1 Dmitry Chernyak
19 1 Dmitry Chernyak
SCI-CD is created to make easy the deployment of the virtual infrastructure,
20 1 Dmitry Chernyak
based on the Ganeti project (http://code.google.com/p/ganeti/).
21 1 Dmitry Chernyak
22 1 Dmitry Chernyak
With Ganeti you will be able to install the "farm" of several computers -
23 1 Dmitry Chernyak
"nodes", each with it's own storage and to create the cluster environment over
24 1 Dmitry Chernyak
them when each virtual instance will be started on the one node and backed up
25 1 Dmitry Chernyak
on the another node with on-line disk synchronization (via DRBD) and fast
26 1 Dmitry Chernyak
failover.
27 1 Dmitry Chernyak
28 1 Dmitry Chernyak
!cluster-sync.jpg!
29 1 Dmitry Chernyak
30 1 Dmitry Chernyak
This provides the way to made cheap redundant high available systems.
31 1 Dmitry Chernyak
32 1 Dmitry Chernyak
Ganeti is also able to manage "regular", stand-alone virtual instances.
33 1 Dmitry Chernyak
34 1 Dmitry Chernyak
h2. SOFTWARE
35 1 Dmitry Chernyak
36 1 Dmitry Chernyak
SCI-CD is based on the Debian/GNU project and uses it's installer (simple-cdd)
37 1 Dmitry Chernyak
and also it's package repositories.
38 1 Dmitry Chernyak
Installing SCI-CD you'll not got a new thing - you'll got a regular Debian/GNU
39 1 Dmitry Chernyak
Linux system, just by the easy way and tuned as complex cluster platform.
40 1 Dmitry Chernyak
41 1 Dmitry Chernyak
SCI-CD contains the minimum of original software, instead it is focused on the
42 1 Dmitry Chernyak
proper usage of existing open source components.
43 1 Dmitry Chernyak
44 1 Dmitry Chernyak
h2. CONTENTS
45 1 Dmitry Chernyak
46 1 Dmitry Chernyak
SCI core platform consists of two Ganeti nodes and one *service instance* (virtual
47 5 Владимир Ипатов
machine, named *sci*).
48 1 Dmitry Chernyak
49 1 Dmitry Chernyak
The service instance provides:
50 1 Dmitry Chernyak
* DNS for local zones, forwarders, hint (BIND9 chrooted),
51 4 Владимир Ипатов
* DHCP server for lan(disabled by default),
52 1 Dmitry Chernyak
* Configuation management, able to tune up new nodes and instances (Puppet),
53 1 Dmitry Chernyak
* Apt-proxy, with uplink to Debian mirrors and also with local repository,
54 1 Dmitry Chernyak
copied from SCI-CD CD-ROM (Approx).
55 1 Dmitry Chernyak
56 1 Dmitry Chernyak
More virtual instances may be created and more functions may be assigned to them
57 1 Dmitry Chernyak
using the regular Ganeti commands.
58 1 Dmitry Chernyak
59 1 Dmitry Chernyak
h2. HARDWARE
60 1 Dmitry Chernyak
61 1 Dmitry Chernyak
Minimal SCI core hardware is based on the two nodes which can be any two computers -
62 1 Dmitry Chernyak
the major brands, such as HP, IBM, the "China brands", such as
63 1 Dmitry Chernyak
SuperMicro or self-maiden servers or even workstations.
64 1 Dmitry Chernyak
65 1 Dmitry Chernyak
Of course, the performance of your system will be very dependent on the
66 1 Dmitry Chernyak
hardware you have chosed - the CPU, memory, RAID subsystem - all will matters
67 1 Dmitry Chernyak
as usual for the performance, but mostly - not for high availability
68 1 Dmitry Chernyak
capabilities.
69 1 Dmitry Chernyak
70 1 Dmitry Chernyak
The HA features are supported by providing two separate nodes with their own
71 1 Dmitry Chernyak
storage subsystems and on-line storage synchronization between them. In the
72 1 Dmitry Chernyak
most cases this level of redundancy is sufficient to cover the data losses and
73 1 Dmitry Chernyak
service interrupts even with the cheap hardware.
74 1 Dmitry Chernyak
75 1 Dmitry Chernyak
Of course, the systems which can't be interrupted even for a few minutes must
76 1 Dmitry Chernyak
be sized and analyzed excluively.
77 1 Dmitry Chernyak
78 1 Dmitry Chernyak
h2. LOCAL NETWORKING
79 1 Dmitry Chernyak
80 1 Dmitry Chernyak
All nodes should be attached to one common TCP/IP network segment via the main
81 1 Dmitry Chernyak
interface (exceptions are possible - see Ganeti manuals).
82 1 Dmitry Chernyak
83 1 Dmitry Chernyak
The SCI setup supports 802.1q VLAN tagging, so you can easy give your nodes and
84 1 Dmitry Chernyak
instances the interfaces in different network segments (the network switch with
85 1 Dmitry Chernyak
802.1q support is required).
86 1 Dmitry Chernyak
87 1 Dmitry Chernyak
h2. STORAGE NETWORKING
88 1 Dmitry Chernyak
89 1 Dmitry Chernyak
The nodes need to be interconnected with the fast TCP/IP interlink to transfer
90 1 Dmitry Chernyak
storage data (DRBD synchronization). 1GigE is recommended.
91 1 Dmitry Chernyak
92 1 Dmitry Chernyak
It is possible (and recommended) that the storage interlink should be
93 1 Dmitry Chernyak
separated from the main TCP/IP access interfaces. In the other words, each
94 1 Dmitry Chernyak
node should have a two network adapters and at least one of them should
95 1 Dmitry Chernyak
support 1GigE.
96 1 Dmitry Chernyak
97 1 Dmitry Chernyak
For only two nodes the storage network may be constructed without an Ethernet
98 6 Владимир Ипатов
switch, using simple cat5e cable.
99 1 Dmitry Chernyak
100 1 Dmitry Chernyak
For the simple and demo tasks there is possible to use only one network adapter
101 1 Dmitry Chernyak
on the node, but it *MUST* support 1GigE speed.
102 1 Dmitry Chernyak
103 1 Dmitry Chernyak
h2. INTERNET NETWORKING
104 1 Dmitry Chernyak
105 1 Dmitry Chernyak
In order to access to the Internet you should connect your local network segment
106 1 Dmitry Chernyak
to the Internet router.
107 1 Dmitry Chernyak
108 1 Dmitry Chernyak
The simplest way is to use the separate router with NAT.
109 1 Dmitry Chernyak
110 1 Dmitry Chernyak
The more advanced way is to construct the router/firewall on the cluster's virtual
111 1 Dmitry Chernyak
instance. This setup is not the part SCI-CD project but can be implemented on the
112 1 Dmitry Chernyak
top of it.
113 1 Dmitry Chernyak
114 1 Dmitry Chernyak
h2. DEBIAN REPOSITORIES
115 1 Dmitry Chernyak
116 1 Dmitry Chernyak
The service instance provides "approx" apt proxy, which can be uplinked to
117 1 Dmitry Chernyak
regular Debian mirrors or to the intermediate mirrors or apt proxies in your LAN. 
118 1 Dmitry Chernyak
119 1 Dmitry Chernyak
In any case, the service instance's "approx" is loaded by the copy of the
120 1 Dmitry Chernyak
repository from the SCI-CD CD-ROM. It can be used even in absence of the uplink
121 1 Dmitry Chernyak
to the external sources.
122 1 Dmitry Chernyak
123 1 Dmitry Chernyak
The standard reposiporise list is pushed automatically into the sources.lists
124 1 Dmitry Chernyak
of the puppet-client instances (which are by default all the nodes and
125 1 Dmitry Chernyak
instances).
126 1 Dmitry Chernyak
127 1 Dmitry Chernyak
h2. INSTALLATION
128 1 Dmitry Chernyak
129 1 Dmitry Chernyak
Please read [[INSTALL]] for instructions.