diff --git a/README.md b/README.md index 53785a8..cc9e519 100644 --- a/README.md +++ b/README.md @@ -55,17 +55,17 @@ Also don't forget to add the new VM's hostname to your ansible inventory file (d ### VM Customization -Noe the fun begins. The cloud-init.data file contains information common to all +Now the fun begins. The ``cloud-init.data`` file contains information common to all VMs you'll create. There should properly be a meta-data file for the VM-specific stuff, but I haven't been able to get that to work and thus I dynamically create a tempory composite cloud-init for the actual VM creation. -Clone the make_ceph_osdx.sh file to make a custom VM. Edit the variables that +Clone the ``make_cephxx.sh`` file to make a custom VM. Edit the variables that define the hostname, MAC address and LVM Logical Volume that will hold the OSD data. Note that the default MAC address for libvirt is randomly generated, so I manually supply my own to make DHCP assign a predictable IP address. -Use LVM's "lvreate" command to create the Logical Volume you'll reference here +Use LVM's "lvcreate" command to create the Logical Volume you'll reference here and edit the script to edit it". As presently configured, the VM will present the LVM logical volume as device "/dev/sda" - the OS lives on /dev/vda. The device ID will vary if you use a different VM bus type than "scsi", but since I don't know @@ -78,13 +78,13 @@ Of course, if you are as error-prone as I am, this may require a few tweaks. Fea If you're extra paranoid, you can delete VM disk and (if it got created), the VM itself. Once everything is happy, the boot process will run and log to your command-line -console. At its end, you'll be presented with a login prompy. +console. At its end, you'll be presented with a login prompt. ***Caution*** It's best to wait a minute or 2, as some setup may still be running even after the login prompt comes up! As supplied, the login is userid "almalinux" and password "redhat". These are defined in the cloud-init.data file and if you like, you can change them. -Now you're ready to run the Ansible stage. Use ctrl-] to return to you VM's original shell (disconnect from the VM console). You don't need it anymore. +Now you're ready to run the Ansible stage. Use ctrl-] to return to your VM host's original shell (disconnect from the VM console). You don't need it anymore. ## Ansible provisioning @@ -98,25 +98,26 @@ prep work: password is "redhat". "mynewosd" is, of course, the hostname you gave to the new OSD VM. -Use the ansible-playbook to run the ceph OSD playbook. This playbook provisions -using the "ceph_osd_hot" role you installed. +Use the ansible-playbook command to run the ceph OSD playbook. This playbook provisions +using the "ceph_osd_host" role you installed. It does the following: 1. Install the ceph repository into ``yum.repos.d``. -1. Install the cephamd utility from the ceph repository. -1. Copy in the ``/etc/ceph`` configuration information files. -1. Do an initial run of cephadm to cause it to pull the container(s) needed to tun cephadm and the ceph daemons. +1. Install the cephadm utility from the ceph repository. +1. Copy in the ``/etc/ceph`` configuration information files from your master copy in the role/files directory.. +1. Do an initial run of cephadm to cause it to pull the container(s) needed to run cephadm and the ceph daemons. Note that if you like, you can also install the "ceph-common" package and be able to run ceph commands without needing "cephadm shell" to run them. ## Rejoice! -Congratulations! You have just created a new ceph host. You can confirm if you like, by using ssh to login to "almalinux@mynewcephosd" and issuing the "sudo cephadm shell" command to enter the cephadm shell and then type "ceph orch ps" to +Congratulations! You have just created a new ceph host. You can confirm if you like, by using ssh to login to "almalinux@mynewcephosd", issuing the "sudo cephadm shell" command to enter the cephadm shell and then type "ceph orch ps" to list the running daemons in your system. Note that if the above fails, the most likely cause will be that your /etc/ceph config files are wrong. You did replace mine with your own in the ansible role file, didn't you? -## Going live, You're now a full-fledged ceph node and you only need to issue the +## Going live +You're now a full-fledged ceph node and you only need to issue the "ceph orch host add" command to add this new VM to the Ceph host list. Ceph will automatically see the unused OSD data device (/dev/sdb) and make an OSD out of it.