====== Installation VM compute ======
Les noeuds de calculs correspondent aux hyperviseurs. Ils sont pilotés par la brique Nova, et possèdent en général un agent neutron pour la configuration des réseaux dédiés aux VMs.
===== Installation de la partie Nova =====
Pour installer les paquets :
yum install openstack-nova-compute
===== Configuration de Nova =====
Fichier de configuration /etc/nova/nova.conf :
[DEFAULT]
auth_strategy=keystone
my_ip=192.168.2.69
use_neutron=true
enabled_apis=osapi_compute,metadata
firewall_driver = nova.virt.firewall.NoopFirewallDriver
rpc_backend=rabbit
[api_database]
[barbican]
[cache]
[cells]
[cinder]
[cloudpipe]
[conductor]
[cors]
[cors.subdomain]
[crypto]
[database]
[ephemeral_storage_encryption]
[glance]
api_servers = http://glance.internal:9292
[guestfs]
[hyperv]
[image_file_url]
[ironic]
[key_manager]
[keystone_authtoken]
auth_uri = http://controller.internal:5000
auth_url = http://controller.internal:35357
memcached_servers = controller.internal:11211
auth_type = password
project_domain_name = default
user_domain_name = default
project_name = service
username = nova
password = nova
[libvirt]
[matchmaker_redis]
[metrics]
[mks]
[neutron]
[osapi_v21]
[oslo_concurrency]
lock_path = /var/lib/nova/tmp
[oslo_messaging_amqp]
[oslo_messaging_notifications]
[oslo_messaging_rabbit]
rabbit_host = bdd
rabbit_userid = openstack
rabbit_password = RABBIT_PASS
[oslo_messaging_zmq]
[oslo_middleware]
[oslo_policy]
[placement]
[placement_database]
[rdp]
[remote_debug]
[serial_console]
[spice]
[ssl]
[trusted_computing]
[upgrade_levels]
[vmware]
[vnc]
enabled = True
vncserver_listen = 0.0.0.0
vncserver_proxyclient_address = $my_ip
novncproxy_base_url = http://controller.public:6080/vnc_auto.html
[workarounds]
[wsgi]
[xenserver]
[xvp]
Activer et demarrer les services Nova de l'hyperviseur :
systemctl enable libvirtd openstack-nova-compute
systemctl start libvirtd openstack-nova-compute
===== Configuration de la partie Neutron =====
Installer les paquets des agents nécessaires :
yum install openstack-neutron-openvswitch ipset
Configuration de l'agent openvswitch dans /etc/neutron/plugins/ml2/openvswitch_agent.ini :
[DEFAULT]
[agent]
tunnel_types = gre,vxlan
l2_population = True
[ovs]
local_ip = 192.168.200.4
bridge_mappings = vlan:br-vlan
[securitygroup]
firewall_driver = iptables_hybrid
Configuration de la partie Neutron dans /etc/neutron/neutron.conf :
[DEFAULT]
auth_strategy = keystone
rpc_backend = rabbit
[agent]
[cors]
[cors.subdomain]
[database]
[keystone_authtoken]
auth_uri = http://controller.internal:5000
auth_url = http://controller.internal:35357
memcached_servers = controller.internal:11211
auth_type = password
project_domain_name = default
user_domain_name = default
project_name = service
username = neutron
password = neutron
[matchmaker_redis]
[nova]
[oslo_concurrency]
lock_path = /var/lib/neutron/tmp
[oslo_messaging_amqp]
[oslo_messaging_notifications]
[oslo_messaging_rabbit]
rabbit_host = bdd
rabbit_userid = openstack
rabbit_password = RABBIT_PASS
[oslo_messaging_zmq]
[oslo_policy]
[qos]
[quotas]
[ssl]
Activer et démarrer les services suivants :
systemctl enable neutron-openvswitch-agent openvswitch
systemctl start neutron-openvswitch-agent openvswitch
===== Configuration de la libvirt =====
Pour permettre les migrations à chaud, il faut ajouter l'option --listen via le fichier /etc/sysconfig/libvirtd. Il faut de plus autoriser le tcp, sans le TLS (à voir pour plus tard), ainsi que d'accepter les connexions sans authentification en TCP. Cela se fait dans /etc/libvirt/libvirtd.conf (auth_tcp = none).