学习openstack(四)

一。KVM虚拟化

1.KVM的安装:

     yum install qemu-kvm qemu-kvm-tools virt-manager libvirt virt-install

     /etc/init.d/libvirtd start      

2.创建一个5G的大小的虚拟机:

    qemu-img create -f raw /opt/centos-6.5-x86_64.raw 5G

3.查看虚拟机空间使用大小:

    qemu-img info /opt/centos-6.5-x86_64.raw

4.启动虚拟机,并指定内存512,磁盘位置,和CDROM的位置,并启动VNC。

virt-install --virt-type kvm --name centos-6.6-64 --ram 512 --cdrom=/opt/centos-6.5.iso --disk path=/opt/centos-6.5-x86_64.raw --network network=default --graphics vnc,listen=0.0.0.0 --noautoconsole --os-type=linux --os-variant=rhel6

5.使用VNC访问虚拟机:

  172.16.2.210 5900端口        (第一台虚拟机默认是5900,第二台就是5901)

6.查看所有创建的虚拟机:

    virsh list --all

7.启动虚拟机virsh start centos-6.6-64:

  virsh start centos-6.6-64 

7.1.直接进入虚拟机命令行:

   virsh console centos-6.6-64

8.查看xml文件:

  vim /etc/libvirt/qemu/centos-6.6-64.xml

   virsh edit centos-6.6-64        (修改xml文件,不能直接用vim修改)

9.根据xml文件新建虚拟机:

   virsh define /opt/centos-6.6-64.xml

10.虚拟机监控命令:virt-top

11.查看虚拟网桥状态:brctl show

12.创建桥接网卡并且关联到eth0上面:

   brctl addbr br0

   brctl addif br0 eth0 && ip del dev eth0 172.16.1.210/24 && ifconfig br0 172.16.1.210/24 up

   (创建了桥接网卡以后需要把原来的eth0的IP去掉,把原来的IP配置在br0上面)

13。修改虚拟机的网卡为刚才创建的桥接网卡:

     virsh edit centos-6.6-64   (修改虚拟机xml文件)

  1.  
    <interface type='bridge'>
  2.  
    <mac address='52:54:00:e4:46:c7'/>
  3.  
    <source bridge='br0'/>
  4.  
    <model type='virtio'/>
    virsh destroy entos-6.6-64 (关闭在启动虚拟机生效)    virsh start entos-6.6-64
 
二。OPENSTACK 云计算与虚拟化 (I 版)
 

 1.安装Openstack基础环境:
  1.  
    wget https://dl.fedoraproject.org/pub/epel/epel-release-latest-6.noarch.rpm (下载epel源)
  2.  
    rpm -ivh epel-release-latest-6.noarch.rpm
  3.  
    yum install -y python-pip gcc gcc-c++ make libtool patch automake python-devel libxslt-devel MySQL-python openssl-devel libudev-devel git wget libvirt-python libvirt qemu-kvm gedit python-numdisplay python-eventlet device-mapper bridge-utils libffi-devel libffi
 2.安装mysql,并且在my.cnf里面加上配置。
  1.  
    default-storage-engine = innodb
  2.  
    innodb_file_per_table
  3.  
    collation-server = utf8_general_ci
  4.  
    init-connect = 'SET NAMES utf8'
  5.  
    character-set-server = utf8
3.给openstack创建Mysql相关库。
  1.  
    create database keystone;
  2.  
    grant all on keystone.* to keystone@'172.16.0.0/255.255.0.0' identified by 'keystone';(改权限和限制访问的Ip段)
  3.  
     
  4.  
    create database glance;
  5.  
    grant all on glance.* to glance@'172.16.0.0/255.255.0.0' identified by 'glance';
  6.  
     
  7.  
    create database nova;
  8.  
    grant all on nova.* to nova@'172.16.0.0/255.255.0.0' identified by 'nova';
  9.  
     
  10.  
    create database neutron;
  11.  
    grant all on neutron.* to neutron@'172.16.0.0/255.255.0.0' identified by 'neutron';
  12.  
     
  13.  
    create database cinder;
  14.  
    grant all on cinder.* to cinder@'172.16.0.0/255.255.0.0' identified by 'cinder';
  15.  
     
 4.安装rabbitmq消息队列:
   yum install rabbitmq-server
5.安装rabbitmq的web管理插件:
   cd /usr/lib/rabbitmq/bin
   ./rabbitmq-plugins enable rabbitmq_management
   /etc/init.d/rabbitmq-server restart
   ./rabbitmq-plugins list  (检查插件是否正常安装)
  http://172.16.1.210:15672  (通过这个网址访问,默认账号密码都是guest)
二。认证服务keystone相关(5000,35357)
6.配置openstack的官方安装源:
  1.  
    [openstack]
  2.  
    name=openstack
  3.  
    baseurl=http://repos.fedorapeople.org/repos/openstack/EOL/openstack-icehouse/epel-6
  4.  
    enabled=1
  5.  
    gpgcheck=0
  6.  
    gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-EPEL-6
 7.安装openstack-keystone:
yum install openstack-keystone python-keystoneclient
 8.创建PKI的目录给keystone使用:
 keystone-manage pki_setup --keystone-user keystone --keystone-group keystone
 chown -R keystone:keystone /etc/keystone/ssl
 chmod -R o-rwx /etc/keystone/ssl   (删掉其他用户的权限)
 9.修改keystone的配置文件:
vim /etc/keystone/keystone.conf 
 
  1.  
    admin_token=ADMIN
  2.  
    log_file=/var/log/keystone/keystone.log
  3.  
    connection=mysql://keystone:keystone@172.16.1.210/keystone
 10.初始化keystone的数据库表结构:
   keystone-manage db_sync  (没有报错就是成功,最好在进mysql里面show tables看下)
  rm /var/log/keystone/keystone.log  (不操作会权限报错)
11.启动keystone:
  /etc/init.d/openstack-keystone start

 12.配置连接keystone连接需要的环境变量:

    export OS_SERVICE_TOKEN=ADMIN

    export OS_SERVICE_ENDPOINT=http://172.16.1.210:35357/v2.0

 13.初始化话keystone的用户数据(admin,demo):

   创建admin,demo用户,创建admin角色,创建service,admin租户

  1.  
    keystone user-create --name=admin --pass=admin --email=admin@example.com
  2.  
    keystone role-create --name=admin
  3.  
    keystone tenant-create --name=admin --description="Admin Tenant"
  4.  
    keystone user-role-add --user=admin --tenant=admin --role=admin
  5.  
    keystone user-role-add --user=admin --role=_member_ --tenant=admin
  6.  
    keystone user-create --name=demo --pass=demo
  7.  
    keystone tenant-create --name=demo --description="demo Tenant"
  8.  
    keystone user-role-add --user=demo --role=_member_ --tenant=demo
  9.  
    keystone tenant-create --name=service
  10.  
    keystone service-create --name=keystone --type=identity
  11.  
    keystone endpoint-create --service-id=$(keystone service-list | awk '/ identity / {print $2}') --publicurl=http://172.16.1.210:5000/v2.0 --internalurl=http://172.16.1.210:5000/v2.0 --adminurl=http://172.16.1.210:35357/v2.0

 创建完成后,使用keystone user-list 命令查看是否有admin和demo两个账号

 

14.查看admin的token

unset OS_SERVICE_TOKEN

unset OS_SERVICE_ENDPOINT

keystone --os-username=admin --os-password=admin --os-tenant-name=admin --os-auth-url=http://172.16.1.210:35357/v2.0 token-get

15.创建admin和demo用户的环境变量文件:

vim /root/keystone-admin

 

  1.  
    export OS_TENANT_NAME=admin
  2.  
    export OS_USERNAME=admin
  3.  
    export OS_PASSWORD=admin
  4.  
    export OS_AUTH_URL=http://172.16.1.210:35357/v2.0

vim /root/keystone-demo

 

  1.  
    export OS_TENANT_NAME=demo
  2.  
    export OS_USERNAME=demo
  3.  
    export OS_PASSWORD=demo
  4.  
    export OS_AUTH_URL=http://172.16.1.210:35357/v2.0

 三。镜像服务Glance(9292,9191)


1.安装glance:

yum install openstack-glance python-glance python-glanceclient

 

2.配置glance:

vim /etc/glance/glance-api.conf

 

  1.  
    [DEFAULT]
  2.  
    debug=True
  3.  
    default_store=file
  4.  
    filesystem_store_datadir=/data/glance/images/
  5.  
    log_file=/var/log/glance/api.log
  6.  
    notifier_strategy = rabbit
  7.  
    rabbit_host=172.16.1.210
  8.  
    rabbit_port=5672
  9.  
    rabbit_use_ssl=false
  10.  
    rabbit_userid=guest
  11.  
    rabbit_password=guest
  12.  
    rabbit_virtual_host=/
  13.  
    rabbit_notification_exchange=glance
  14.  
    rabbit_notification_topic=notifications
  15.  
    rabbit_durable_queues=False
  16.  
    [database]
  17.  
    connection=mysql://glance:glance@172.16.1.210/glance
  18.  
    [keystone_authtoken]
  19.  
    auth_host=172.16.1.210
  20.  
    auth_port=35357
  21.  
    auth_protocol=http
  22.  
    admin_tenant_name=service
  23.  
    admin_user=glance
  24.  
    admin_password=glance
  25.  
    [paste_deploy]
  26.  
    flavor=keystone

vim /etc/glance/glance-registry.conf

 

  1.  
    debug=True
  2.  
    log_file=/var/log/glance/registry.log
  3.  
    connection=mysql://glance:glance@172.16.1.210/glance
  4.  
     
  5.  
    [keystone_authtoken]
  6.  
    auth_host=172.16.1.210
  7.  
    auth_port=35357
  8.  
    auth_protocol=http
  9.  
    admin_tenant_name=service
  10.  
    admin_user=glance
  11.  
    admin_password=glance
  12.  
     
  13.  
    [paste_deploy]
  14.  
    flavor=keystone

 

3.同步glance的mysql数据库:

glance-manage db_sync  (警告报错可以忽略)

chown -R glance:glance /var/log/glance

4.在keystone中创建glance的用户:

keystone user-create --name=glance --pass=glance  (创建glance密码也是一样)

keystone user-role-add --user=glance --tenant=service --role=admin (讲glance用户加入到admin角色service租户中)

5.将glance在keystone注册服务和注册url:

  1.  
    keystone service-create --name=glance --type=image
  2.  
    keystone endpoint-create --service-id=$(keystone service-list|awk '/ image / {print $2}') --publicurl=http://172.16.1.210:9292 --internalurl=http://172.16.1.210:9292 --adminurl=http://172.16.1.210:9292

 6.启动glance:

/etc/init.d/openstack-glance-api status

/etc/init.d/openstack-glance-registry start 

7.查看镜像列表:

   glance image-list 

8.下载开发镜像并进行导入:

  1.  
    wget http://download.cirros-cloud.net/0.3.2/cirros-0.3.2-x86_64-disk.img
  2.  
    glance image-create --name "cirros-0.3.2-x86_64" --disk-format qcow2 --container-format bare --is-public True --file cirros-0.3.2-x86_64-disk.img

 

 

四。计算服务nova相关  (5000,35357)

 1.安装控制节点的nova服务:

yum install openstack-nova-api openstack-nova-cert openstack-nova-conductor openstack-nova-console openstack-nova-novncproxy openstack-nova-scheduler python-novaclient

pip install websockify==0.5.1   (解决novnc启动不了的问题)

 2.修改nova的配置文件:

 vim /etc/nova/nova.conf

  1.  
    rabbit_host=172.16.1.210
  2.  
    rabbit_port=5672
  3.  
    rabbit_use_ssl=false
  4.  
    rabbit_userid=guest
  5.  
    rabbit_password=guest
  6.  
    rpc_backend=rabbit
  7.  
    my_ip=172.16.1.210
  8.  
    auth_strategy=keystone
  9.  
     
  10.  
    network_api_class=nova.network.neutronv2.api.API
  11.  
    linuxnet_interface_driver=nova.network.linux_net.LinuxBridgeInterfaceDriver
  12.  
    neutron_url=http://172.16.1.210:9696
  13.  
    neutron_admin_username=neutron
  14.  
    neutron_admin_password=neutron
  15.  
    neutron_admin_tenant_id=96616014997f4f79b7dbd9e319912154 ("keystone tenant-list"命令看到的 service_id)
  16.  
    neutron_admin_tenant_name=service
  17.  
    neutron_admin_auth_url=http://172.16.1.210:5000/v2.0
  18.  
    neutron_auth_strategy=keystone
  19.  
    firewall_driver=nova.virt.firewall.NoopFirewallDriver
  20.  
     
  21.  
    novncproxy_base_url=http://172.16.1.210:6080/vnc_auto.html
  22.  
    vncserver_listen=0.0.0.0
  23.  
    vncserver_proxyclient_address=172.16.1.210
  24.  
    vnc_enabled=true
  25.  
    vnc_keymap=en-us
  26.  
    connection=mysql://nova:nova@172.16.1.210/nova
  27.  
    auth_host=172.16.1.210
  28.  
    auth_port=35357
  29.  
    auth_protocol=http
  30.  
    auth_uri=http://172.16.1.210:5000
  31.  
    auth_version=v2.0
  32.  
    admin_user=nova
  33.  
    admin_password=nova
  34.  
    admin_tenant_name=service
  35.  
    vif_driver=nova.virt.libvirt.vif.NeutronLinuxBridgeVIFDriver

 

3.初始化nova的数据库表结构:

nova-manage db sync

4在keystone中创建nova的用户:

source /root/keystone-admin

keystone user-create --name=nova --pass=nova

keystone user-role-add --user=nova --tenant=service --role=admin

5.将nova在keystone注册服务和注册url:

 

  1.  
    source /root/keystone-admin (根据自己存放变量的文件)
  2.  
    keystone service-create --name=nova --type=compute
  3.  
    keystone endpoint-create --service-id=$(keystone service-list|awk '/ compute / {print $2}') --publicurl=http://172.16.1.210:8774/v2/%\(tenant_id\)s --internalurl=http://172.16.1.210:8774/v2/%\(tenant_id\)s --adminurl=http://172.16.1.210:8774/v2/%\(tenant_id\)s

 6.启动nova所有相关服务:

 

for i in {api,cert,conductor,consoleauth,novncproxy,scheduler};do service openstack-nova-$i start; done
 

 7.在计算节点上面安装相关服务:

    yum install -y qemu-kvm libvirt openstack-nova-compute python-novaclient

    yum upgrade device-mapper-libs

 8.将控制节点nova配置文件复制到计算节点:

  scp /etc/nova/nova.conf 172.16.1.211:/etc/nova/nova.conf

  修改vncserver_proxyclient_address=172.16.1.211 配置的IP为计算节点的。

 9.启动计算节点的相关服务:

/etc/init.d/libvirtd start

/etc/init.d/messagebus start

/etc/init.d/openstack-nova-compute start

10.在控制节点上面查看是否注册上来:

  nova host-list

六。网络服务neutron相关  (9696)

1.在控制节点安装neutron服务:

yum install openstack-neutron openstack-neutron-ml2 python-neutronclient openstack-neutron-linuxbridge

2.修改neutron配置文件:

vim /etc/neutron/neutron.conf

  1.  
    state_path = /var/lib/neutron
  2.  
    lock_path = $state_path/lock
  3.  
    core_plugin = ml2
  4.  
    service_plugins = router,firewall,lbaas
  5.  
    api_paste_config = /usr/share/neutron/api-paste.ini
  6.  
    auth_strategy = keystone
  7.  
    rabbit_host = 172.16.1.210
  8.  
    rabbit_password = guest
  9.  
    rabbit_port = 5672
  10.  
    rabbit_userid = guest
  11.  
    rabbit_virtual_host = /
  12.  
    notify_nova_on_port_status_changes = true
  13.  
    notify_nova_on_port_data_changes = true
  14.  
    nova_url = http://172.16.1.210:8774/v2
  15.  
    nova_admin_username = nova
  16.  
    nova_admin_tenant_id = 96616014997f4f79b7dbd9e319912154 ("keystone tenant-list"命令看到的 service_id)
  17.  
    nova_admin_password = nova
  18.  
    nova_admin_auth_url = http://172.16.1.210:35357/v2.0
  19.  
    root_helper = sudo neutron-rootwrap /etc/neutron/rootwrap.conf
  20.  
    auth_host = 172.16.1.210
  21.  
    auth_port = 35357
  22.  
    auth_protocol = http
  23.  
    admin_tenant_name = service
  24.  
    admin_user = neutron
  25.  
    admin_password = neutron
  26.  
    connection = mysql://neutron:neutron@172.16.1.210:3306/neutron
  27.  
    service_provider=LOADBALANCER:Haproxy:neutron.services.loadbalancer.drivers.haproxy.plugin_driver.HaproxyOnHostPluginDriver:default
  28.  
    service_provider=VPN:openswan:neutron.services.vpn.service_drivers.ipsec.IPsecVPNDriver:default

 vim /etc/neutron/plugins/ml2/ml2_conf.ini  (下面是单一扁平网络的配置)

  1.  
    type_drivers = flat,vlan,gre,vxlan
  2.  
    tenant_network_types = flat,vlan,gre,vxlan
  3.  
    mechanism_drivers = linuxbridge,openvswitch
  4.  
    flat_networks = physnet1
  5.  
    enable_security_group = True

 vim /etc/neutron/plugins/linuxbridge/linuxbridge_conf.ini    (下面是单一扁平网络的配置)

  1.  
    network_vlan_ranges = physnet1
  2.  
    physical_interface_mappings = physnet1:eth0 (根据网卡来配置)
  3.  
    firewall_driver = neutron.agent.linux.iptables_firewall.IptablesFirewallDriver
  4.  
    enable_security_group = True

 

3.在keystone中创建neutron的用户:

  1.  
    keystone user-create --name neutron --pass neutron
  2.  
    keystone user-role-add --user neutron --tenant service --role admin

 4.这里需要改下nova的neutron相关配置(我前面已经整合进去了)

 5.将neutron在keystone注册服务和注册url:

  1.  
    keystone service-create --name neutron --type network
  2.  
    keystone endpoint-create --service-id=$(keystone service-list |awk '/ network / {print $2}') --publicurl=http://172.16.1.210:9696 --internalurl=http://172.16.1.210:9696 --adminurl=http://172.16.1.210:9696

 6.先使用手动启动看是否有报错(启动成功会监听9696端口):

neutron-server --config-file=/etc/neutron/neutron.conf --config-file=/etc/neutron/pluns/ml2/ml2_conf.ini  --config-file=/etc/neutron/plugins/linuxbridge/linuxbridge_conf.ini
 

7.修改neutron服务的init启动脚本:

vim /etc/init.d/neutron-server

vim /etc/init.d/neutron-linuxbridge-agent              (两个文件修改相同的地方)

  1.  
    configs=(
  2.  
    "/etc/neutron/neutron.conf" \
  3.  
    "/etc/neutron/plugins/linuxbridge/linuxbridge_conf.ini" \
  4.  
    "/etc/neutron/plugins/ml2/ml2_conf.ini" \
  5.  
    )

 8.启动neutron服务:

/etc/init.d/neutron-server status

/etc/init.d/neutron-linuxbridge-agent start

9.查看neutron的连接情况:

neutron agent-list

 

10.在计算节点安装neutron服务:

yum install openstack-neutron openstack-neutron-ml2 python-neutronclient openstack-neutron-linuxbridge

11.将控制节点的配置文件复制到计算节点上:

  1.  
    scp /etc/init.d/neutron-* 172.16.1.211:/etc/init.d/
  2.  
    scp /etc/neutron/neutron.conf 172.16.1.211:/etc/neutron/neutron.conf
  3.  
    scp /etc/neutron/plugins/ml2/ml2_conf.ini 172.16.1.211:/etc/neutron/plugins/ml2/ml2_conf.ini
  4.  
    scp /etc/neutron/plugins/linuxbridge/linuxbridge_conf.ini 172.16.1.211:/etc/neutron/plugins/linuxbridge/linuxbridge_conf.ini

 12.在计算节点上启动neutron的客户端程序:

/etc/init.d/neutron-linuxbridge-agent start

 七。web服务dashboard相关  (80)

1.安装dashboard服务:

yum install httpd mod_wsgi memcached python-memcached openstack-dashboard

2.修改dashboard的配置文件:

vim /etc/openstack-dashboard/local_settings    (修改下面两项)

  1.  
    ALLOWED_HOSTS = ['horizon.example.com', 'localhost','172.16.1.210'] (允许的地址)
  2.  
    CACHES = {
  3.  
    'default': {
  4.  
    'BACKEND' : 'django.core.cache.backends.memcached.MemcachedCache',
  5.  
    'LOCATION' : '127.0.0.1:11211',
  6.  
    }
  7.  
    }
  8.  
     
  9.  
    OPENSTACK_HOST = "172.16.1.210"

 3.启动dashboard服务:

/etc/init.d/memcached start

/etc/init.d/httpd start

4.登录dashboard:

http://172.16.1.210/dashboard/

默认账户密码:admin/admin

 八。调试Openstack

1.为demo创建一个单一扁平网络:

neutron net-create --tenant-id 9d18b0a337064af386cc0d599dd172fd flat_net --shared --provider:network_type flat --provider:physical_network physnet1
 

 id为keystone tenant-list 查看的demo_id,  flat_net是创建网络的名称,--shared代表共享网络,后面是网络的类型,最后是对应neutron网络名称.

2.查看创建的网络:

   neutron net-list

3.在网页上面创建子网:

 管理员-网络-点击网络-创建子网-扁平网络网段就用和eth0一样的就可以

4.虚拟机创建流程图:


5.安装备注:

  1.如果安装完成以后虚拟机无法ping同外网,请检查网卡是否打开“混杂模式”

  2.强行修改虚拟机的状态为“运行”

   nova reset-state 6986b3f8-be2c-4931-b3b9-90d8077210b6 --active

  3.启动“主机集合”功能需要修改配置文件:

  /etc/nova/nova.conf

  scheduler_default_filters=AvailabilityZoneFilter,RetryFilter,RamFilter,ComputeFilter,ComputeCapabilitiesFilter,ImagePropertiesFilter,CoreFilter

 4.将db4加入到gigold-2的主机集合里面:

 nova aggregate-add-host gigold-2 db4

 nova aggregate-remove-host gigold-2 db4 (这个是删除)

 5.关闭默认的virbr0虚拟网卡

  1.  
    virsh net-destroy default
  2.  
    virsh net-undefine default
  3.  
    service libvirtd restart

 

  • 大小: 386.1 KB
  • 大小: 214.9 KB

 #####################################################################################################

一。块存储服务Cinder控制节点

1.安装cinder:

yum install openstack-cinder python-cinderclient

2.修改配置文件:

vim /etc/cinder/cinder.conf

  1.  
    rabbit_host=172.16.1.210
  2.  
    rabbit_port=5672
  3.  
    rabbit_use_ssl=false
  4.  
    rabbit_userid=guest
  5.  
    rabbit_password=guest
  6.  
    rpc_backend=rabbit
  7.  
    my_ip=172.16.1.210
  8.  
    glance_host=$my_ip
  9.  
    auth_strategy=keystone
  10.  
    connection=mysql://cinder:cinder@172.16.1.210/cinder
  11.  
    auth_host=172.16.1.210
  12.  
    auth_port=35357
  13.  
    auth_protocol=http
  14.  
    auth_uri=http://172.16.1.210:5000
  15.  
    identity_uri=http://172.16.1.210:35357/
  16.  
    auth_version=v2.0
  17.  
    admin_user=cinder
  18.  
    admin_password=cinder
  19.  
    admin_tenant_name=service

 

3.初始化cinder数据库:

cinder-manage db sync

4.在keystone上面创建cinder用户:

keystone user-create --name=cinder --pass=cinder

keystone user-role-add --user=cinder --tenant=service --role=admin

5.将cinder在keystone注册服务和注册url:

  1.  
    keystone service-create --name=cinder --type=volume
  2.  
    keystone endpoint-create --service-id=$(keystone service-list|awk '/ cinder / {print $2}') --publicurl=http://172.16.1.210:8776/v1/%\(tenant_id\)s --internalurl=http://172.16.1.210:8776/v1/%\(tenant_id\)s --adminurl=http://172.16.1.210:8776/v1/%\(tenant_id\)s
  3.  
     
  4.  
    keystone service-create --name=cinderv2 --type=volumev2
  5.  
    keystone endpoint-create --service-id=$(keystone service-list|awk '/ cinderv2 / {print $2}') --publicurl=http://172.16.1.210:8776/v2/%\(tenant_id\)s --internalurl=http://172.16.1.210:8776/v2/%\(tenant_id\)s --adminurl=http://172.16.1.210:8776/v2/%\(tenant_id\)s

 6.启动cinder服务:

 /etc/init.d/openstack-cinder-api start

/etc/init.d/openstack-cinder-scheduler start

7.查看cinder服务上面注册了哪些云硬盘:

cinder service-list

二。块存储服务Cinder存储节点(LVM)


1.在存储节点上面创建一块硬盘,然后用pv初始化:

pvcreate /dev/sdb

vgcreate cinder-volumes /dev/sdb

2.修改配置文件:

vim /etc/lvm/lvm.conf

 

  1.  
    # Use anchors if you want to be really specific
  2.  
    filter = [ "a/sda1/","a/sdb/","r/.*/" ]

3.安装共享存储target服务:

yum install scsi-target-utils

4.修改配置文件:

vim /etc/tgt/targets.conf

 

include /etc/cinder/volumes/*  (添加)
 

 4.安装cinder服务:

yum install openstack-cinder

5.将控制节点的配置文件拷贝到存储节点:

scp /etc/cinder/cinder.conf 172.16.1.211:/etc/cinder/cinder.conf

6.修改配置文件:

vim /etc/cinder/cinder.conf

  1.  
    my_ip=172.16.1.211
  2.  
    glance_host=172.16.1.210
  3.  
     
  4.  
    iscsi_ip_address=$my_ip
  5.  
    volume_backend_name=iSCSI-Storage
  6.  
    iscsi_helper=tgtadm
  7.  
    volume_driver=cinder.volume.drivers.lvm.LVMISCSIDriver

 7.启动相关服务:

/etc/init.d/tgtd start

/etc/init.d/openstack-cinder-volume start

8.创建cinder的iscsi类型(方便选择后端的存储位置):

  1.  
    cinder type-create iSCSI
  2.  
    cinder type-key iSCSI set volume_backend_name=iSCSI-Storage
  3.  
    (命令在控制节点上操作,“iSCSI-Storage”这个是配置文件里面命名的)

9.下面可以通过网页的云硬盘操作了。

。块存储服务Cinder存储节点(NFS)


1.创建NFS共享(不说明了):

2.修改cinder配置文件:

vim /etc/cinder/cinder.conf

  1.  
    volume_backend_name=Nfs-Storage
  2.  
    nfs_shares_config=/etc/cinder/nfs_shares
  3.  
    nfs_mount_point_base=$state_path/mnt
  4.  
    volume_driver=cinder.volume.drivers.nfs.NfsDriver

 vim /etc/cinder/nfs_shares

  1.  
    172.16.1.210:/data/nfs
  2.  
    NFS地址

3.启动cinder服务:

/etc/init.d/openstack-cinder-volume start

4.创建cinder的nfs类型:

 

  1.  
    cinder type-create NFS
  2.  
    cinder type-key NFS set volume_backend_name=Nfs-Storage
5.下面可以通过网页的云硬盘操作了。

 

 

。块存储服务Cinder存储节点(glusterfs

1.安装glusterfs服务(两个存储节点都装上):

 

  1.  
    wget http://download.gluster.org/pub/gluster/glusterfs/3.7/3.7.14/CentOS/glusterfs-epel.repo
  2.  
    mv glusterfs-epel.repo /etc/yum.repos.d/
  3.  
    yum install glusterfs-server
 2.启动两台机器的glusterfs:

 

/etc/init.d/glusterd start

3.创建glusterfs劵:

 

  1.  
    mkdir /data/glusterfs/exp1 (两台机器都要操作)
  2.  
    gluster peer probe 172.16.1.211 (创建对等节点,在210上面操作)
  3.  
    gluster volume create cinder-volume01 replica 2 172.16.1.210:/data/glusterfs/exp1 172.16.1.211:/data/glusterfs/exp1 force (创建cinder-volume01劵)
  4.  
    gluster vol start cinder-volume01 (启动卷)
  5.  
    gluster vol info (查看劵的状态)
 4.修改cinder配置文件:

 

vim /etc/cinder/glusterfs_shares   (本来没有,需要创建)

172.16.1.210:/cinder-volome01
 

 vim /etc/cinder/cinder.conf

  1.  
    volume_backend_name=GLS-Storage
  2.  
    glusterfs_shares_config=/etc/cinder/glusterfs_shares
  3.  
    volume_driver=cinder.volume.drivers.glusterfs.GlusterfsDriver

 5.创建cinder的glusterfs类型:

  1.  
    cinder type-create GLS
  2.  
    cinder type-key GLS set volume_backend_name=GLS-Storage

 6.现在可以到网页上面创建云硬盘了

####################################################################################

一。负载均衡服务LBaas


1.在dashboard中打开lbaas菜单:

  1.  
    'enable_lb': True,
  2.  
    'enable_firewall': True,
  3.  
    'enable_quotas': True,
  4.  
    'enable_vpn': True,
  5.  
     
  6.  
    将原来的False改为True.(注意大写)

 2.重启dashboard服务:

/etc/init.d/httpd restart

3.安装 haproxy服务:

yum install haproxy

4.修改neutron的配置文件:

vim /etc/neutron/lbaas_agent.ini

  1.  
    interface_driver = neutron.agent.linux.interface.BridgeInterfaceDriver
  2.  
    device_driver = neutron.services.loadbalancer.drivers.haproxy.namespace_driver.HaproxyNSDriver
  3.  
    user_group = haproxy

5.安装namespace支持:

ip netns list   (输入命令没有报错说明支持,不要在安装了)

yum update iproute

6.修改启动lbaas脚本:

vim /etc/init.d/neutron-lbaas-agent

  1.  
    configs=(
  2.  
    "/etc/neutron/neutron.conf" \
  3.  
    "/etc/neutron/lbaas_agent.ini" \
  4.  
    )

 7.启动lbaas服务:

/etc/init.d/neutron-lbaas-agent start

8.下面可以在WEB界面添加负载均衡了

posted on 2019-01-03 11:01  五光十色  阅读(492)  评论(0编辑  收藏  举报

导航