User Tools

Site Tools


pharos:pharos_specification

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Last revision Both sides next revision
pharos:pharos_specification [2015/04/08 23:00]
Trevor Cooper [Hardware]
pharos:pharos_specification [2015/05/11 13:36]
Victor Laza Moved all pharos_specifications under docs/spec.rst in git
Line 1: Line 1:
-===== Pharos Specification ===== +{{scrape>​http://artifacts.opnfv.org/pharos/docs/spec.html}}
- +
-==== Objectives / Scope ==== +
- +
-Objectives of environment are for OPNFV platform __deployment__ and __testing__ with requirements to ... +
-    * Provide a secure, scalable, standard and HA environment +
-    * Support full deployment lifecycle (this requires a bare metal environment) +
-    * Support functional and performance testing +
-    * Provide common tooling and test scenarios (including test cases and workloads) +
-    * Provide mechanisms and procedures for secure remote access to the environment  +
- +
-Virtualized environments will be useful but do not provide a fully featured deployment/​test capability +
- +
-The high level architecture may be summarized as follow: +
-{{ :​pharos:​pharos-archi1.jpg?​500 |}} +
- +
-==== Constraints of a Pharos compliant OPNFV test-bed environment ==== +
- +
-   * One jump (provisioning server) in which the installer runs in a VM +
-   * 2 - 5 compute / controller nodes +
-   * Jump server provisioned with CentOS7 +
-   * Installer (e.g. Foreman) runs its own DHCP server therefore management network should not have DHCP server +
-   * Remote access +
-   * A lights-out network is required for remote management and bare metal provisioning capability +
- +
-==== Target Systems State (for Rls 1 follows BGS) ==== +
- +
-   * Target system state includes default software components, network configuration,​ storage requirements https://wiki.opnfv.org/get_started/get_started_system_state +
- +
-  Rls 1 specification is modeled from Bootstrap/Get-started project requirements (BGS) ... +
-  * First draft of environment for BGS https://​wiki.opnfv.org/​get_started/​get_started_work_environment +
-  * Fuel environment https://​wiki.opnfv.org/​get_started/​networkingblueprint +
-  * Foreman environment https://​wiki.opnfv.org/​get_started_experiment1#​topology +
- +
- +
- +
- +
- +
-==== Hardware ==== +
- +
- +
-=== Servers === +
- +
-  CPU: +
-  * Intel Xeon E5-2600 (IvyBridge at least, or similar) +
- +
-  Local Storage: +
-  * Disks: 4 x 500G-2T + 1 x 300GB SSD (leave some room for experiments)  +
-  * First 2 disks should be combined to form a 1 TB virtual store for the OS/Software etc +
-  * Remaining should be combined to form a virtual disk for CEPH storage +
-  * The 5'th disk (SSD) for distributed storage (CEPH) journal towards SSD technology. +
-  * Performance testing requires a mix of compute nodes that have CEPH(swift+Cinder) and without CEPH storage +
-  * Virtual ISO boot capabilities or a separate PXE boot server (DHCP/tftp or Cobbler) +
- +
-  Memory: +
-  * 32G RAM Minimum +
- +
-  Power Supply Single +
-  * Single power supply acceptable (redundant power not required/​nice to have) +
- +
- +
- +
-  Pre-provisioning Jump Server +
-    * OS - CentOS7 +
-    * KVM / Qemu +
-    * Installer (Foreman, Fuel, ...) in a VM +
-    * Collabration Tools +
-        * VNC +
-    * Test Tools +
-        * Specified per Functest +
-        * Foreman +
-        * Fuel +
-        *  +
-   +
-  Provisioning Nodes +
-  * Controller nodes - bare metal +
-  * Compute nodes - bare metal +
- +
-===Security=== +
- +
-  * Default permissions +
-  * Server Logins +
- +
- +
-=== Remote management=== +
- +
-  Lights-out Management:​ +
-  * Out-of-band management for power on/​off/​reset and bare-metal provisioning +
-  * Access to server is through lights-out-management tool and/or a serial console +
-  * Intel lights-out =>  +
-    RMM http://​www.intel.com/​content/​www/​us/​en/​server-management/​intel-remote-management-module.html +
-  * HP lights-out => ILO +
-    http://​www8.hp.com/​us/​en/​products/​servers/​ilo/​index.html +
-  * CISCO UCS => +
-    https://​developer.cisco.com/​site/​ucs-dev-center/​index.gsp +
- +
- +
-==== Networking ==== +
- +
-  Test-bed network +
-  * 24 or 48 Port TOR Switch +
-  * NICS - 1GE, 10GE - per server can be on-board or PCI-e +
-  * Connectivity for each data/​control network is through a separate NIC. This simplifies Switch Management however requires more NICs on the server and also more switch ports +
-  * Lights-out network can share with Admin/​Management +
- +
-  Network Interfaces +
-  * Option 1: 4x1G Control, 2x40G Data, 48 Port Switch +
-    * 1 x 1G for ILMI (Lights out Management ) +
-    * 1 x 1G for Admin/PXE boot +
-    * 1 x 1G for control Plane connectivity +
-    * 1 x 1G for storage +
-    * 2 x 40G (or 10G) for data network (redundancy,​ NIC bonding, High bandwidth testing) +
-  * Option II: 1x1G Control, 2x 40G (or 10G) Data, 24 Port Switch  +
-    * Connectivity to networks is through VLANs on the Control NIC. Data NIC used for VNF traffic and storage traffic segmented through VLANs +
-  * Option III: 2x1G Control, 2x10G Data, 2x40G Storage, 24 Port Switch  +
-    * Data NIC used for VNF traffic, storage NIC used for control plane and Storage segmented through VLANs (separate host traffic from VNF) +
-    * 1 x 1G for IPMI +
-    * 1 x 1G for Admin/PXE boot +
-    * 2 x 10G for control plane connectivity/​Storage +
-    * 2 x 40G (or 10G) for data network +
-  +
-=== Storage Network === +
-    * Needs specification +
- +
- +
- +
- +
- +
-== Topology == +
- +
-    * Subnet, VLANs (want to standardize but may be constrained by existing lab setups or rules) +
-    * IPs +
-    * Types of NW - lights-out, public, private, admin, storage +
-    * May be special NW requirements for performance related projects +
-    * Default gateways +
- +
-{{ :​pharos:​bridge1.png?​500 | }} +
-controller node bridge topology overview +
- +
-{{ :​pharos:​bridge2.png?​500 | }} +
-compute node bridge topology overview +
- +
-=== Remote access (Jump server) ===   +
-    * Access rules / restrictions +
-    * SSH to Jump server +
-    * VPN is optional and dependent on company security rules (out of Pharos scope) +
-    * Access by LF environment for automated deployment (type of test-bed to be declared to LF CI system)  +
- +
- +
-==== Architecture ==== +
- +
-=== Network Diagram === +
-The Pharos architecture may be described as follow: +
-{{ :​pharos:​opnfv-pharos-diagram-v01.jpg |}} +
-//Figure 1: Standard Deployment Environment//​ +
- +
- +
-==== Tools ==== +
-    * Jenkins agent (server hosted by CI infrastructure) +
-    * Tempest / Rally +
-    * Robot +
-    * Repo for test scenarios +
-    * Jira - hosted by LF +
-    * FAQ channel - hosted by LF +
-    * Juju and MAAS on JUMP SERVER +
-  +
- +
- +
- +
- +
----- +
- +
- +
- +
- +
- +
- +
-==== Sample Network Drawings ==== +
- +
-Files for documenting lab network layout. These were contributed as Visio VSDX format compressed as a ZIP file. Here is a sample of what the visio looks like. +
- +
-Download the visio zip file here: {{:​opnfv-example-lab-diagram.vsdx.zip|}} +
- +
-{{:​opnfv-example-lab-diagram.png?​direct&​200|}} +
- +
-FYI: [[http://​www.opendaylight.org/​community/​community-labs|Here]] is what the OpenDaylight lab wiki pages look like.+
pharos/pharos_specification.txt · Last modified: 2015/06/03 13:44 by Trevor Cooper