Appendix B. Example gdeploy configuration files
B.1. Example gdeploy configuration file for setting up TLS/SSL
set_up_encryption.conf
# IPs that corresponds to the Gluster Network [hosts] <Gluster_Network_NodeA> <Gluster_Network_NodeB> <Gluster_Network_NodeC> # STEP-1: Generate Keys, Certificates & CA files # The following section generates the keys,certicates, creates # ca file and distributes it to all the hosts [volume1] action=enable-ssl volname=engine ssl_clients=<Gluster_Network_NodeA>,<Gluster_Network_NodeB>,<Gluster_Network_NodeC> ignore_volume_errors=no # As the certificates are already generated, its enough to stop # rest of the volumes,set TLS/SSL related volume options, and # start the volume # STEP-2: Stop all the volumes [volume2] action=stop volname=vmstore [volume3] action=stop volname=data # STEP-3: Set volume options on all the volumes to enable TLS/SSL on the volumes [volume4] action=set volname=vmstore key=client.ssl,server.ssl,auth.ssl-allow value=on,on,"<Gluster_Network_NodeA>;<Gluster_Network_NodeB>;<Gluster_Network_NodeC>" ignore_volume_errors=no [volume5] action=set volname=data key=client.ssl,server.ssl,auth.ssl-allow value=on,on,"<Gluster_Network_NodeA>;<Gluster_Network_NodeB>;<Gluster_Network_NodeC>" ignore_volume_errors=no # STEP-4: Start all the volumes [volume6] action=start volname=vmstore [volume7] action=start volname=data
B.2. Example gdeploy configuration file for preparing to replace a node
If the disks must be replaced as well as the node, ensure that the [pv], [vg], and [lv] sections are not commented out of this file.
For details about how to safely replace a node, see Chapter 9, Replacing a Gluster Storage Node.
replace_node_prep.conf
# EDITME: @1: Change to IP addresses of the network intended for gluster traffic # Values provided here are used to probe the gluster hosts. [hosts] 10.70.X1.Y1 #EDITME : @2: Change to IP addresses of the network intended for gluster traffic #of the node which is going to be replaced. [script1] action=execute ignore_script_errors=no file=/usr/share/ansible/gdeploy/scripts/grafton-sanity-check.sh -d sdc -h 10.70.X1.Y1 # EDITME: @3: Specify the number of data disks in RAID configuration [disktype] raid6 [diskcount] 4 [stripesize] 256 # EDITME : @4: UNCOMMENT SECTION (RHEL ONLY) :Provide the subscription details # Register to RHSM only on the node which needs to be replaced #[RH-subscription1:10.70.X1.Y1] #action=register #username=<username> #password=<passwd> #pool=<pool-id> #[RH-subscription2] #action=disable-repos #repos= #[RH-subscription3] #action=enable-repos #repos=rhel-7-server-rpms,rh-gluster-3-for-rhel-7-server-rpms,rhel-7-server-rhv-4-mgmt-agent-rpms #[yum1] #action=install #packages=vdsm,vdsm-gluster,ovirt-hosted-engine-setup,screen,gluster-nagios-addons #update=yes [service1] action=enable service=chronyd [service2] action=restart service=chronyd [shell1] action=execute command=gluster pool list [shell2] action=execute command=vdsm-tool configure --force # Disable multipath [script3] action=execute file=/usr/share/ansible/gdeploy/scripts/disable-multipath.sh #EDIT ME: @5: UNCOMMENT SECTIONS ONLY: if original brick disks have to be replaced. #[pv1] #action=create #devices=sdc #ignore_pv_errors=no #[vg1] #action=create #vgname=gluster_vg_sdc #pvname=sdc #ignore_vg_errors=no #[lv2:10.70.X1:Y1] #action=create #poolname=gluster_thinpool_sdc #ignore_lv_errors=no #vgname=gluster_vg_sdc #lvtype=thinpool #poolmetadatasize=16GB #size=14TB #[lv3:10.70.X1:Y1] #action=create #lvname=gluster_lv_engine #ignore_lv_errors=no #vgname=gluster_vg_sdc #mount=/gluster_bricks/engine #size=100GB #lvtype=thick #[lv5:10.70.X1:Y1] #action=create #lvname=gluster_lv_data #ignore_lv_errors=no #vgname=gluster_vg_sdc #mount=/gluster_bricks/data #lvtype=thinlv #poolname=gluster_thinpool_sdc #virtualsize=12TB #[lv7:10.70.X1:Y1] #action=create #lvname=gluster_lv_vmstore #ignore_lv_errors=no #vgname=gluster_vg_sdc #mount=/gluster_bricks/vmstore #lvtype=thinlv #poolname=gluster_thinpool_sdc #virtualsize=1TB #[selinux] #yes #[lv9:10.70.X1:Y1] #action=setup-cache #ssd=sdb #vgname=gluster_vg_sdc #poolname=lvthinpool #cache_lv=lvcache #cache_lvsize=180GB [service3] action=start service=glusterd slice_setup=yes [firewalld] action=add ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,54322/tcp services=glusterfs [script2] action=execute file=/usr/share/ansible/gdeploy/scripts/disable-gluster-hooks.sh
B.3. Example gdeploy configuration file for scaling to additional nodes
add_nodes.conf
# Add the hosts to be added [hosts] <Gluster_Network_NodeD> <Gluster_Network_NodeE> <Gluster_Network_NodeF> # If using RHEL 7 as platform, enable required repos # RHVH has all the packages available #[RH-subscription] #ignore_register_errors=no #ignore_attach_pool_errors=no #ignore_enable_errors=no #action=register #username=<username> #password=<mypassword> #pool=<pool-id> #repos=rhel-7-server-rpms,rh-gluster-3-for-rhel-7-server-rpms,rhel-7-server-rhv-4-mgmt-agent-rpms #disable-repos=yes # If using RHEL 7 as platform, have the following section to install packages [yum1] action=install packages=vdsm-gluster,ovirt-hosted-engine-setup,screen update=yes gpgcheck=yes ignore_yum_errors=no # enable chronyd [service1] action=enable service=chronyd # start chronyd service [service2] action=restart service=chronyd # Setup glusterfs slice [service3] action=restart service=glusterd slice_setup=yes # Open the required ports and firewalld services [firewalld] action=add ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,54322/tcp services=glusterfs # Disable gluster hook scripts [script2] action=execute file=/usr/share/ansible/gdeploy/scripts/disable-gluster-hooks.sh

Where did the comment section go?
Red Hat's documentation publication system recently went through an upgrade to enable speedier, more mobile-friendly content. We decided to re-evaluate our commenting platform to ensure that it meets your expectations and serves as an optimal feedback mechanism. During this redesign, we invite your input on providing feedback on Red Hat documentation via the discussion platform.