User Tools

Site Tools


hardware:ivan

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
hardware:ivan [2022/06/08 15:13]
djgalloway [Summary]
hardware:ivan [2022/06/08 15:17] (current)
djgalloway
Line 26: Line 26:
  
 ===== OSD/Block Device Information ===== ===== OSD/Block Device Information =====
-The ivan have 9x 12TB HDD, 2x 1.5TB NVMe, and 1x 350GB NVMe. +I used the Orchestrator ​to deploy OSDs on the ivan hosts (I did this one by one to avoid a mass data rebalance all to one rack).
- +
-The 12TB were added to so we can say we're testing ​on drives larger than 8TB. +
- +
-The smaller NVMe device is split into eleven equal logical volumes. One for each OSD's journal. +
- +
-<​code>​ +
-root@ivan04:​~#​ lsblk +
-NAME                 ​MAJ:​MIN RM   SIZE RO TYPE  MOUNTPOINT +
-sda                    8:0    0 894.3G ​ 0 disk   +
-`-sda1 ​                ​8:​1 ​   0 894.3G ​ 0 part   +
-  `-md0                9:0    0 894.1G ​ 0 raid1 / +
-sdb                    8:16   0 894.3G ​ 0 disk   +
-`-sdb1 ​                ​8:​17 ​  0 894.3G ​ 0 part   +
-  `-md0                9:0    0 894.1G ​ 0 raid1 / +
-sdc                    8:32   ​0 ​ 10.9T  0 disk   +
-sdd                    8:48   ​0 ​ 10.9T  0 disk   +
-sde                    8:64   ​0 ​ 10.9T  0 disk   +
-sdf                    8:80   ​0 ​ 10.9T  0 disk   +
-sdg                    8:96   ​0 ​ 10.9T  0 disk   +
-sdh                    8:112  0  10.9T  0 disk   +
-sdi                    8:128  0  10.9T  0 disk   +
-sdj                    8:144  0  10.9T  0 disk   +
-sdk                    8:160  0  10.9T  0 disk   +
-sr0                   ​11:​0 ​   1   ​841M ​ 0 rom    +
-nvme0n1 ​             259:0    0 349.3G ​ 0 disk   +
-`-nvme0n1p1 ​         259:3    0 349.3G ​ 0 part   +
-  |-journals-sdc ​    ​253:​0 ​   0    31G  0 lvm    +
-  |-journals-sdd ​    ​253:​1 ​   0    31G  0 lvm    +
-  |-journals-sde ​    ​253:​2 ​   0    31G  0 lvm    +
-  |-journals-sdf ​    ​253:​3 ​   0    31G  0 lvm    +
-  |-journals-sdg ​    ​253:​4 ​   0    31G  0 lvm    +
-  |-journals-sdh ​    ​253:​5 ​   0    31G  0 lvm    +
-  |-journals-sdi ​    ​253:​6 ​   0    31G  0 lvm    +
-  |-journals-sdj ​    ​253:​7 ​   0    31G  0 lvm    +
-  |-journals-sdk ​    ​253:​8 ​   0    31G  0 lvm    +
-  |-journals-nvme1n1 253:9    0    31G  0 lvm    +
-  `-journals-nvme2n1 253:​10 ​  ​0 ​   31G  0 lvm    +
-nvme1n1 ​             259:1    0   ​1.5T ​ 0 disk   +
-nvme2n1 ​             259:2    0   ​1.5T ​ 0 disk +
-</​code>​ +
- +
-==== How to partition/​re-partition ​the NVMe device ==== +
-Here's my bash history that can be used to set up an ivan machine'​s 375GB NVMe card.+
  
 <​code>​ <​code>​
-ansible ​-a "sudo parted ​-/dev/nvme0n1 mktable gpt" ivan +root@reesi001:​~#​ cat ivan_osd_spec.yml  
-ansible ​-a "sudo parted ​/dev/nvme0n1 unit '​%'​ mkpart foo 0 100" ivan +service_type:​ osd 
-ansible ​-a "sudo pvcreate ​/dev/nvme0n1p1"​ ivan +service_id: osd_using_paths 
-ansible ​-a "sudo vgcreate journals ​/dev/nvme0n1p1"​ ivan +placement:​ 
-for disk in sd{c..k} nvme1n1 nvme2n1; do ansible ​-a "sudo lvcreate ​-L 31G -n $disk journals"​ ivan; done+  hosts: 
 +    ​ivan01 
 +    - ivan02 
 +    - ivan03 
 +    - ivan04 
 +    - ivan05 
 +    - ivan06 
 +    - ivan07 
 +spec: 
 +  data_devices:​ 
 +    paths: 
 +    ​- /dev/sdc 
 +    - /dev/sdd 
 +    - /dev/sde 
 +    - /dev/sdf 
 +    /dev/sdg 
 +    ​/dev/sdh 
 +    ​/dev/sdi 
 +    - /dev/sdj 
 +    - /dev/sdk 
 +    - /​dev/​nvme1n1 
 +    - /​dev/​nvme2n1 
 +  db_devices:​ 
 +    paths: 
 +    - /​dev/​nvme0n1
 </​code>​ </​code>​
  
Line 93: Line 74:
 I added the hosts to the cluster using the ''​back''​ IPs. The cluster became very unhappy complaining about slow OPs. Come to find out the ivan servers couldn'​t get **out** from their ''​back''​ interfaces so the OSDs defaulted back to the 1Gb link. I added the hosts to the cluster using the ''​back''​ IPs. The cluster became very unhappy complaining about slow OPs. Come to find out the ivan servers couldn'​t get **out** from their ''​back''​ interfaces so the OSDs defaulted back to the 1Gb link.
  
-I reached out to Red Hat IT to have the 25Gb network ports switched over to VLAN100. After that, I struggled to get eno1 (the 1Gb interface) to not come up on boot.+I reached out to Red Hat IT to have the 25Gb network ports switched over to VLAN100. After that, I struggled to get eno1 (the 1Gb interface) to **not** come up on boot since I didn't need it anymore.
  
 Finally I figured out<​code>​ Finally I figured out<​code>​
hardware/ivan.1654701233.txt.gz · Last modified: 2022/06/08 15:13 by djgalloway