Fixed much badspelling

This commit is contained in:
tim holloway 2024-07-20 14:42:41 -04:00
parent 1d81e11179
commit 05cbfd67cb

View File

@ -55,17 +55,17 @@ Also don't forget to add the new VM's hostname to your ansible inventory file (d
### VM Customization
Noe the fun begins. The cloud-init.data file contains information common to all
Now the fun begins. The ``cloud-init.data`` file contains information common to all
VMs you'll create. There should properly be a meta-data file for the VM-specific
stuff, but I haven't been able to get that to work and thus I dynamically create
a tempory composite cloud-init for the actual VM creation.
Clone the make_ceph_osdx.sh file to make a custom VM. Edit the variables that
Clone the ``make_cephxx.sh`` file to make a custom VM. Edit the variables that
define the hostname, MAC address and LVM Logical Volume that will hold the OSD data.
Note that the default MAC address for libvirt is randomly generated, so I manually supply my own to make DHCP assign a predictable IP address.
Use LVM's "lvreate" command to create the Logical Volume you'll reference here
Use LVM's "lvcreate" command to create the Logical Volume you'll reference here
and edit the script to edit it". As presently configured, the VM will present the
LVM logical volume as device "/dev/sda" - the OS lives on /dev/vda. The device ID
will vary if you use a different VM bus type than "scsi", but since I don't know
@ -78,13 +78,13 @@ Of course, if you are as error-prone as I am, this may require a few tweaks. Fea
If you're extra paranoid, you can delete VM disk and (if it got created), the VM itself.
Once everything is happy, the boot process will run and log to your command-line
console. At its end, you'll be presented with a login prompy.
console. At its end, you'll be presented with a login prompt.
***Caution*** It's best to wait a minute or 2, as some setup may still be running even after the login prompt comes up!
As supplied, the login is userid "almalinux" and password "redhat". These are defined in the cloud-init.data file and if you like, you can change them.
Now you're ready to run the Ansible stage. Use ctrl-] to return to you VM's original shell (disconnect from the VM console). You don't need it anymore.
Now you're ready to run the Ansible stage. Use ctrl-] to return to your VM host's original shell (disconnect from the VM console). You don't need it anymore.
## Ansible provisioning
@ -98,25 +98,26 @@ prep work:
password is "redhat". "mynewosd" is, of course, the hostname you gave to the new
OSD VM.
Use the ansible-playbook to run the ceph OSD playbook. This playbook provisions
using the "ceph_osd_hot" role you installed.
Use the ansible-playbook command to run the ceph OSD playbook. This playbook provisions
using the "ceph_osd_host" role you installed.
It does the following:
1. Install the ceph repository into ``yum.repos.d``.
1. Install the cephamd utility from the ceph repository.
1. Copy in the ``/etc/ceph`` configuration information files.
1. Do an initial run of cephadm to cause it to pull the container(s) needed to tun cephadm and the ceph daemons.
1. Install the cephadm utility from the ceph repository.
1. Copy in the ``/etc/ceph`` configuration information files from your master copy in the role/files directory..
1. Do an initial run of cephadm to cause it to pull the container(s) needed to run cephadm and the ceph daemons.
Note that if you like, you can also install the "ceph-common" package and be able to run ceph commands without needing "cephadm shell" to run them.
## Rejoice!
Congratulations! You have just created a new ceph host. You can confirm if you like, by using ssh to login to "almalinux@mynewcephosd" and issuing the "sudo cephadm shell" command to enter the cephadm shell and then type "ceph orch ps" to
Congratulations! You have just created a new ceph host. You can confirm if you like, by using ssh to login to "almalinux@mynewcephosd", issuing the "sudo cephadm shell" command to enter the cephadm shell and then type "ceph orch ps" to
list the running daemons in your system.
Note that if the above fails, the most likely cause will be that your /etc/ceph config files are wrong. You did replace mine with your own in the ansible role file, didn't you?
## Going live, You're now a full-fledged ceph node and you only need to issue the
## Going live
You're now a full-fledged ceph node and you only need to issue the
"ceph orch host add" command to add this new VM to the Ceph host list. Ceph will
automatically see the unused OSD data device (/dev/sdb) and make an OSD out of it.