MidoNet Reference and Testing Deployment Module for OpenStack.
Version 8.0 / Mitaka
####Table of Contents
- Overview - What is the puppetlabs-openstack module?
- A Note on Versioning
- Module Description - What does the module do?
- Setup - The basics of getting started with OpenStack
- Usage - Configuration and customization options
- Reference - An under-the-hood peek at what the module is doing
- Limitations - OS compatibility, etc.
- License
##Overview
The puppetlabs-openstack module is used to deploy a multi-node, all-in-one, or swift-only installation of OpenStack Mitaka. This module superseeds puppetlabs-openstack by defining roles and profiles that configure OpenStack with MidoNet as Neutron driver
##Versioning
This module has been given version 4 to track the puppet-openstack modules. The versioning for the puppet-openstack modules are as follows:
Puppet Module :: OpenStack Version :: OpenStack Codename
2.0.0 -> 2013.1.0 -> Grizzly
3.0.0 -> 2013.2.0 -> Havana
4.0.0 -> 2014.1.0 -> Icehouse
5.0.0 -> 2014.2.0 -> Juno
8.0.0 -> 2016.04.07 -> Mitaka
##Module Description
Using the stable/mitaka branch of the puppet-openstack modules, midonet-midonet_openstack allows for the rapid deployment of an installation of OpenStack Mitaka. For the multi-node, up to four types of nodes are created for the deployment:
- A controller node that hosts databases, message queues and caches, and most api services.
- A compute node to run guest operating systems with the MidoNet agent
- A nsdb node for handling topology, data flow information
- An analytics node for insights
You can still use the storage and tempest nodes defined in puppetlabs-openstack
##Setup
###Setup Requirements
This module assumes nodes running on a RedHat 7 variant (RHEL, CentOS, or Scientific Linux) or Ubuntu 14.04 (Trusty) or Ubuntu 16.04 (Xenial Xerus) with either Puppet Enterprise or Puppet.
Each node needs a minimum of two network interfaces, and up to four. The network interfaces are divided into two groups.
- Public interfaces:
- API network.
- External network.
- Internal interfaces:
- Management network.
- Data network.
This module have been tested with Puppet 4.X and Puppet Enterprise. This module depends upon Hiera. Object store support (Swift) depends upon exported resources and PuppetDB.
###Beginning with OpenStack
To begin, you will need to do some basic setup on the compute node. SElinux needs to be disabled on the compute nodes to give OpenStack full control over the KVM hypervisor and other necessary services. This is the only node that SELinux needs to be disabled on.
Additionally, you need to know the network address ranges for all four of the public/private networks, and the specific ip addresses of the controller node and the storage node. Keep in mind that your public networks can overlap with one another, as can the private networks.
The examples directory contains Vagrantfiles with CentOS 7 boxes to test out all-in-one, multi-node, or swift-only deployments.
##Usage
###Params Configuration The first step to using the puppetlabs-openstack module is to configure params.pp with settings specific to your installation. In this module, the example directory contains sample params.pp (for multi-node) and for all-in-one files with all of the settings required by this module
###Controller Node For your controller node, you need to assign your node the controller role. For example:
node 'control.localdomain' {
include ::openstack::role::controller
}
It's important to apply this configuration to the controller node before any of the other nodes are applied. The other nodes depend upon the service and database setup in the controller node.
###Other Nodes
For the remainder nodes, there are roles to assign for each. For example:
node /compute[0-9]+.localdomain/ {
include ::openstack::role::compute
}
node /nsdb[0-9]+.localdomain/ {
class {'::openstack::role::nsdb' :
zk_servers => [{
'ip' => $::ipaddress}
]
}
node /analytics.localdomain/ {
include ::openstack::role::allinone_analytics
}
For this deployment, it's assumed that there is only one analytics node ( there can't be more than one). There may be multiple compute nodes.
After applying the configuration to the controller node, apply the remaining configurations to the worker nodes.
You will need to reboot all of the nodes after installation to ensure that the kernel module that provides network namespaces, required by Open VSwitch, is loaded.
##Reference
The midonet-midonet_openstack module is built on the 'Roles and Profiles' pattern. Every node in a deployment is assigned a single role. Every role is composed of some number of profiles, which ideally should be independent of one another, allowing for composition of new roles. The puppetlabs-openstack module does not strictly adhere to this pattern, but should serve as a useful example of how to build profiles from modules for customized and maintainable OpenStack deployments.
##Limitations
- High availability and SSL-enabled endpoints are not provided by this module.
- Only one Analytics node is supported
##License
Copyright (c) 2016 Midokura SARL, All Rights Reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
strongly based on
Puppet Labs OpenStack - A Puppet Module for a Multi-Node OpenStack Mitaka Installation.
Copyright (C) 2013, 2014 Puppet Labs, Inc. and Authors
Original Author - Christian Hoge
Puppet Labs can be contacted at: [email protected]
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.