Add models with mitaka/newton os releases for maintenance purposes
Change-Id: I555bf37aba6c48be4d3513bb8f413fb809b6b25c
diff --git a/classes/cluster/virtual-mcp-newton-dvr/openstack/control.yml b/classes/cluster/virtual-mcp-newton-dvr/openstack/control.yml
new file mode 100644
index 0000000..6202b33
--- /dev/null
+++ b/classes/cluster/virtual-mcp-newton-dvr/openstack/control.yml
@@ -0,0 +1,111 @@
+classes:
+- system.linux.system.lowmem
+- system.linux.storage.loopback
+- system.linux.system.repo.glusterfs
+- system.linux.system.repo.mcp.openstack
+- system.linux.system.repo.mcp.extra
+- system.linux.system.repo.saltstack.xenial
+- system.memcached.server.single
+- system.rabbitmq.server.cluster
+- system.rabbitmq.server.vhost.openstack
+- system.keystone.server.wsgi
+- system.keystone.server.cluster
+- system.glusterfs.client.cluster
+- system.glusterfs.client.volume.glance
+- system.glusterfs.client.volume.keystone
+- system.glusterfs.server.volume.glance
+- system.glusterfs.server.volume.keystone
+- system.glusterfs.server.cluster
+- system.glance.control.cluster
+- system.nova.control.cluster
+- system.neutron.control.openvswitch.cluster
+- system.cinder.control.cluster
+- system.cinder.control.backend.lvm
+- system.cinder.volume.single
+- system.cinder.volume.backend.lvm
+- system.heat.server.cluster
+- system.ceilometer.server.cluster
+- system.ceilometer.server.backend.influxdb
+- system.aodh.server.cluster
+- system.heka.ceilometer_collector.single
+- system.galera.server.cluster
+- system.galera.server.database.aodh
+- system.galera.server.database.ceilometer
+- system.galera.server.database.cinder
+- system.galera.server.database.glance
+- system.galera.server.database.grafana
+- system.galera.server.database.heat
+- system.galera.server.database.keystone
+- system.galera.server.database.nova
+- cluster.virtual-mcp-newton-dvr
+
+parameters:
+ _param:
+ keepalived_vip_interface: ens4
+ linux:
+ system:
+ package:
+ python-msgpack:
+ version: latest
+ network:
+ interface:
+ ens4:
+ enabled: true
+ type: eth
+ proto: static
+ address: ${_param:single_address}
+ netmask: 255.255.255.0
+ keepalived:
+ cluster:
+ instance:
+ VIP:
+ virtual_router_id: 150
+ keystone:
+ server:
+ admin_email: ${_param:admin_email}
+ notification:
+ driver: messagingv2
+ topics: "notifications,${_param:stacklight_notification_topic}"
+ glance:
+ server:
+ storage:
+ engine: file
+ images: []
+ workers: 1
+ notification:
+ driver: messagingv2
+ topics: "notifications,${_param:stacklight_notification_topic}"
+ heat:
+ server:
+ notification:
+ driver: messagingv2
+ topics: "notifications,${_param:stacklight_notification_topic}"
+ neutron:
+ server:
+ notification:
+ driver: messagingv2
+ topics: "notifications,${_param:stacklight_notification_topic}"
+ nova:
+ controller:
+ networking: dvr
+ cpu_allocation: 54
+ metadata:
+ password: ${_param:metadata_password}
+ bind:
+ private_address: ${_param:cluster_local_address}
+ public_address: ${_param:cluster_vip_address}
+ novncproxy_port: 6080
+ vncproxy_url: http://${_param:cluster_vip_address}:6080
+ workers: 1
+ notification:
+ driver: messagingv2
+ topics: "notifications,${_param:stacklight_notification_topic}"
+ cinder:
+ volume:
+ notification:
+ driver: messagingv2
+ topics: "notifications,${_param:stacklight_notification_topic}"
+ controller:
+ notification:
+ driver: messagingv2
+ topics: "notifications,${_param:stacklight_notification_topic}"