0. ntp설정
1. iml 다운로드 및 서버 추가
2. 러스터 설치
3. lnet 설정 및 시작
4. 러스터 모듈 추가
5. zfs 모듈 추가
6. zfspool 생성
7. dataset 생성
8. 마운트
9. iml
-
모든 러스터 볼륨 umount 후 러스터 모듈 제거
-
tunefs.lustre --reformat --nomgs --mgsnode=${mds01 ip} --mgsnode=${mds02 ip} --writeconf /dev/mapper/scratch-mdt
-
home-mdt 는 --writeconf 옵션만 주고 러스터 모듈 올리고 다시마운트
-
ost 의 경유 --reformat --writeconf 옵션 준
-
기존에 scratch에 마운트되어있던 것들은 --erase-param한후 --mgsnode --servicenode
#yum repolist
#yum-config-manager --disable updates
#yum install -y createrepo vsftpd
#mkdir /repo
#mount -t nfs 10.73.10.33:/root/Downloads /repo
#tar -xvzf /repo/local_repos.tar.gz -C /var/ftp
#createrepo /var/ftp
#vi /etc/yum.repos.d/remote.repo
[remote]
name=reomte
baseurl=ftp://10.73.10.10
enabled=1
gpgcheck=0
#yum-config-manager --enable remote
#yum repolist
#yum install -y python2-iml-manager
#chroma-config setup admin lustre localhost --no-dbspace-check -v
#chroma --username admin --password lustre server add mds1.local --server_profile base_managed_rh7
lnetctl set discovery [0 | 1]
createrepo -g /path/to/mygroups.xml /srv/my/repo //그룹list 추가
mkfs.lustre --mgs --failnode=10.73.20.12@tcp0 --backfstype=zfs --mkfsoptions="mountpoint=none" mgtpool/MGS
mkfs.lustre --mgs --servicenode 192.168.227.11@tcp1 --servicenode 192.168.227.12@tcp1 --backfstype=ldiskfs /dev/dm-1
mkfs.lustre --mdt --mgsnode=10.73.20.11@tcp0 --mgsnode=10.73.20.12@tcp0 --failnode=10.73.20.11@tcp0 --index=0 --backfstype=zfs --fsname=hello --mkfsoptions="mountpoint=none" mdtpool/hello-MDT0000
mkfs.lustre --mdt --fsname demo --index 0 --mgsnode 192.168.227.11@tcp1 --mgsnode 192.168.227.12@tcp1 --servicenode 192.168.227.12@tcp1 --servicenode 192.168.227.11@tcp1 /dev/dm-2
mkfs.lustre --ost --mgsnode=10.73.20.11@tcp0 --mgsnode=10.73.20.12@tcp0 --failnode=10.73.20.22@tcp0 --index=0 --backfstype=zfs --fsname=hello --mkfsoptions="mountpoint=none" ostpool1/hello-OST0000
mkfs.lustre --ost --mgsnode=10.73.20.11@tcp0 --mgsnode=10.73.20.12@tcp0 --failnode=10.73.20.21@tcp0 --index=1 --backfstype=zfs --fsname=hello --mkfsoptions="mountpoint=none" ostpool2/hello-OST0001
mkfs.lustre --ost --fsname demo --index 0 --mgsnode 192.168.227.11@tcp1 --mgsnode 192.168.227.12@tcp1 --servicenode 192.168.227.21@tcp1 --servicenode 192.168.227.22@tcp1 /dev/dm-3
#yum install -y --exclude kernel-debug python2-iml-agent-management kernel-devel-lustre pcs fence-agents fence-agents-virsh lustre-resource-agents lustre-ldiskfs-zfs python2-iml-agent-4.2.0-1.el7
#systemctl reboot
#lnetctl net del --net tcp
#lnetctl net add --net tcp --if eth1
config ntp
#### python2-iml-agent-management 설치 해야 fence_chroma 사용 가능 ####
#### managerforlustre-manager-for-lustre-5.1 레포지토리에 패키지 있음 ####
https://wiki.whamcloud.com/display/PUB/Using+Pacemaker+1.1+with+a+Lustre+File+System
#passwd hacluster //모든 서버에서 설정
#pcs cluster auth mds1.local mds2.local
username:
password:
mds1.local: Authorized
mds2.local: Authorized
#pcs cluster setup --start --name mds-cluster mds1.local mds2.local --enable --token 17000
#pcs stonith create st-fencing fence_chroma
#pcs cluster auth oss1.local oss2.local
username:
password:
oss1.local: Authorized
oss2.local: Authorized
#pcs cluster setup --start --name oss-cluster oss1.local oss2.local --enable --token 17000
#pcs stonith create st-fencing fence_chroma
#pcs cluster node add 10.128.0.11
Disabling SBD service...
10.128.0.11: sbd disabled
Sending remote node configuration files to '10.128.0.11'
10.128.0.11: successful distribution of the file 'pacemaker_remote authkey'
10.128.0.12: Corosync updated
Setting up corosync...
10.128.0.11: Succeeded
Synchronizing pcsd certificates on nodes 10.128.0.11...
10.128.0.11: Success
Restarting pcsd on the nodes in order to reload the certificates...
10.128.0.11: Success
#config corosync
#pcs cluster auth 10.128.0.12 -u hacluster -p ********************
10.128.0.12: Authorized
#pcs cluster setup --name lustre-ha-cluster --force 10.128.0.12 --transport udp --rrpmode passive --addr0 10.128.0.0 --mcast0 226.94.0.1 --mcastport0 42227 --addr1 10.73.10.0 --mcast1 226.94.1.1 --mcastport1 42227 --token 17000 --fail_recv_const 10
Destroying cluster on nodes: 10.128.0.12...
10.128.0.12: Stopping Cluster (pacemaker)...
10.128.0.12: Successfully destroyed cluster
Sending 'pacemaker_remote authkey' to '10.128.0.12'
10.128.0.12: successful distribution of the file 'pacemaker_remote authkey'
Sending cluster config files to the nodes...
10.128.0.12: Succeeded
Synchronizing pcsd certificates on nodes 10.128.0.12...
10.128.0.12: Success
Restarting pcsd on the nodes in order to reload the certificates...
10.128.0.12: Success
stonith resource 만 등록되어있음
-
smtp 서버 설정
sendmail 설치 # yum install -y sendmail sendmail-cf //sendmail : 메일전송 메이전트, sendmail설정 # vi /etc/mail/sendmail.mc //52,53번 라인 주석 제거 (dnl제거) # m4 sendmail.mc > sendmail.cf # systemctl restart sendmail # yum install -y mailx // mail 송신 test # mail [email protected] Subject: title context context . EOT 메일 전송이 되지 않고, 로그에 rejecting connections on daemon MTA: load average: 12 가 표시될 경우 - sendmail.cf 395번째 라인의 주석(#)제거 후 값을 12 -> 30으로 변경 # vi sendmail.cf ---- 변경 전 ---- 394 # load average at which we refuse connections 395 #O RefuseLA=12 ---- 변경 후 ---- 394 # load average at which we refuse connections 395 O RefuseLA=30 - QueueLA : 처리중인 메일 프로세스가 해당 수치에 도달하면 queue에 적재 - 메일 프로세스가 해당 수치에 도달하면 접속을 거부
-
iml 설정 # vi /usr/share/chroma-manager/local_settings.py //chroma설치 및 초기 설정 이후 추가 설정시 local_settings.py EMAIL_HOST='hostname.domainname' EMAIL_SENDER='[email protected]' //email의 from에 명시될 이름
# chroma-config restart