This is an old revision of the document!
This wiki defines the target system state that is created by a successful execution of the BGS. This target system state should be independent from the installer approach taken.
The Pharos specification describes a "standard" OPNFV deployment environment (compute, network, storage)
Component Type | Flavor | Version | Notes |
---|---|---|---|
Base OS | CentOS | 7 | Including current updates for all installed packages |
SDN Controller | OpenDaylight | Helium SR2 | With Open vSwitch |
Infrastructure controller | OpenStack | Juno |
Install only core components on all servers. Additional dependencies will be included when specific packages are added. This list is derived from with fuel specific components removed OpenStack Fuel Template. Puppet is included since OpenSteak, Foreman and Fuel installers are all using Puppet.
Package | Version | Note |
---|---|---|
authconfig | ||
bind-utils | ||
cronie | ||
crontabs | ||
curl | ||
daemonize | ||
dhcp | ||
gdisk | ||
lrzip | ||
lsof | ||
man | ||
mlocate | ||
nmap-ncat | ||
ntp | ||
openssh-clients | ||
policycoreutils | ||
rsync | ||
ruby21-puppet | ||
ruby21-rubygem-netaddr | ||
ruby21-rubygem-openstack | ||
selinux-policy-targeted | ||
strace | ||
subscription-manager | ||
sysstat | ||
system-config-firewall-base | ||
tcpdump | ||
telnet | ||
vim-enhanced | ||
virt-what | ||
wget | ||
yum |
OpenStack Juno Components OpenStack Juno
Component | Package | Version | Notes |
---|---|---|---|
Nova | Juno | ||
Glance | Juno | ||
Neutron | Juno | ||
Keystone | Juno | ||
MySQL | Juno | ||
RabbitMQ | Juno | ||
Pacemaker cluster stack | Juno | ||
Corosync | Juno | ||
Ceilometer | Juno | Possibility to drop from first release | |
Horizon | Juno | Possibility to drop from first release | |
Heat | Juno | Possibility to drop from first release | |
Tempest | Juno | What about documenting on the CI page and not the getting started page? | |
Robot | Juno | What about documenting on the CI page and not the getting started page? |
Component | Sub-Component | Version | Notes |
---|---|---|---|
odl-dlux-all | 0.1.2-Helium-SR2 | ||
odl-config-persister-all | 0.2.6-Helium-SR2 | OpenDaylight :: Config Persister:: All | |
odl-aaa-all | 0.1.2-Helium-SR2 | OpenDaylight :: AAA :: Authentication :: All Featu | |
odl-ovsdb-all | 1.0.2-Helium-SR2 | OpenDaylight :: OVSDB :: all | |
odl-ttp-all | 0.0.3-Helium-SR2 | OpenDaylight :: ttp :: All | |
odl-openflowplugin-all | 0.0.5-Helium-SR2 | OpenDaylight :: Openflow Plugin :: All | |
odl-adsal-compatibility-all | 1.4.4-Helium-SR2 | OpenDaylight :: controller :: All | |
odl-tcpmd5-all | 1.0.2-Helium-SR2 | ||
odl-adsal-all | 0.8.3-Helium-SR2 | OpenDaylight AD-SAL All Features | |
odl-config-all | 0.2.7-Helium-SR2 | OpenDaylight :: Config :: All | |
odl-netconf-all | 0.2.7-Helium-SR2 | OpenDaylight :: Netconf :: All | |
odl-base-all | 1.4.4-Helium-SR2 | OpenDaylight Controller | |
odl-mdsal-all | 1.1.2-Helium-SR2 | OpenDaylight :: MDSAL :: All | |
odl-yangtools-all | 0.6.4-Helium-SR2 | OpenDaylight Yangtools All | |
odl-restconf-all | 1.1.2-Helium-SR2 | OpenDaylight :: Restconf :: All | |
odl-integration-compatible-with-all | 0.2.2-Helium-SR2 | ||
odl-netconf-connector-all | 1.1.2-Helium-SR2 | OpenDaylight :: Netconf Connector :: All | |
odl-akka-all | 1.4.4-Helium-SR2 | OpenDaylight :: Akka :: All |
Component | Package | Version | Notes |
---|---|---|---|
Hypervisor: KVM | |||
Forwarder: OVS | 2.3.0 | ||
Node config: Puppet | |||
Example VNF1: Linux | Centos | 7 | |
Example VNF2: OpenWRT | version 14.07 – barrier braker) |
Describe which L2 segments are configured (i.e. for management, control, use by client VNFs, etc.), how these segments are realized (e.g. VXLAN between OVSs) and which segment numbering (e.g. VLAN IDs, VXLAN IDs) are used. Describe which IP addresses are used, which DNS entries (if any are configured), default gateways, etc. Describe if/how segments are interconnected etc.
List and purpose of used subnets as defined here: Network addressing and topology blueprint
Following picture shows ODL connects to Neutron through ML2 pugin and to nova-compute through OVS bridge. (not yet finished, ceph storage will be added, approach with ODL in docker container will be added.)
Local storage will be used.
Multiple labs will eventually be working together across geographic boundaries.