Giter VIP home page Giter VIP logo

lustre_prac's Introduction

lustre_prac

  0. ntp설정
  1. iml 다운로드 및 서버 추가
  2. 러스터 설치
  3. lnet 설정 및 시작
  4. 러스터 모듈 추가
  5. zfs 모듈 추가 
  6. zfspool 생성
  7. dataset 생성
  8. 마운트
  9. iml

tunefs nomgs 사용

  1. 모든 러스터 볼륨 umount 후 러스터 모듈 제거

  2. tunefs.lustre --reformat --nomgs --mgsnode=${mds01 ip} --mgsnode=${mds02 ip} --writeconf /dev/mapper/scratch-mdt

  3. home-mdt 는 --writeconf 옵션만 주고 러스터 모듈 올리고 다시마운트

  4. ost 의 경유 --reformat --writeconf 옵션 준

  5. 기존에 scratch에 마운트되어있던 것들은 --erase-param한후 --mgsnode --servicenode


managed lustre

  #yum repolist

  #yum-config-manager --disable updates

  #yum install -y createrepo vsftpd
  
  #mkdir /repo

  #mount -t nfs 10.73.10.33:/root/Downloads /repo

  #tar -xvzf /repo/local_repos.tar.gz -C /var/ftp

  #createrepo /var/ftp

  #vi /etc/yum.repos.d/remote.repo
  [remote]
  name=reomte
  baseurl=ftp://10.73.10.10
  enabled=1
  gpgcheck=0

  #yum-config-manager --enable remote
  
  #yum repolist
  
  #yum install -y python2-iml-manager
  
  #chroma-config setup admin lustre localhost --no-dbspace-check -v

  #chroma --username admin --password lustre server add mds1.local --server_profile base_managed_rh7

multirail

  lnetctl set discovery [0 | 1]

yum grouplist 추가

createrepo -g /path/to/mygroups.xml /srv/my/repo //그룹list 추가

디스크 포맷

mgt

  mkfs.lustre --mgs --failnode=10.73.20.12@tcp0 --backfstype=zfs --mkfsoptions="mountpoint=none" mgtpool/MGS
  mkfs.lustre --mgs --servicenode 192.168.227.11@tcp1 --servicenode 192.168.227.12@tcp1 --backfstype=ldiskfs /dev/dm-1

mdt

  mkfs.lustre --mdt --mgsnode=10.73.20.11@tcp0 --mgsnode=10.73.20.12@tcp0 --failnode=10.73.20.11@tcp0 --index=0 --backfstype=zfs --fsname=hello --mkfsoptions="mountpoint=none" mdtpool/hello-MDT0000
  
  mkfs.lustre --mdt --fsname demo --index 0 --mgsnode 192.168.227.11@tcp1 --mgsnode 192.168.227.12@tcp1 --servicenode 192.168.227.12@tcp1 --servicenode 192.168.227.11@tcp1 /dev/dm-2

ost

  mkfs.lustre --ost --mgsnode=10.73.20.11@tcp0 --mgsnode=10.73.20.12@tcp0 --failnode=10.73.20.22@tcp0 --index=0 --backfstype=zfs --fsname=hello --mkfsoptions="mountpoint=none" ostpool1/hello-OST0000

  mkfs.lustre --ost --mgsnode=10.73.20.11@tcp0 --mgsnode=10.73.20.12@tcp0 --failnode=10.73.20.21@tcp0 --index=1 --backfstype=zfs --fsname=hello --mkfsoptions="mountpoint=none" ostpool2/hello-OST0001
  
  
  
  mkfs.lustre --ost --fsname demo --index 0 --mgsnode 192.168.227.11@tcp1 --mgsnode 192.168.227.12@tcp1 --servicenode 192.168.227.21@tcp1 --servicenode 192.168.227.22@tcp1 /dev/dm-3

monitored zfs

mds[1,2] oss[1,2]

  #yum install -y --exclude kernel-debug python2-iml-agent-management kernel-devel-lustre pcs fence-agents fence-agents-virsh lustre-resource-agents lustre-ldiskfs-zfs python2-iml-agent-4.2.0-1.el7

  #systemctl reboot

  #lnetctl net del --net tcp

  #lnetctl net add --net tcp --if eth1

  config ntp
  
  #### python2-iml-agent-management 설치 해야 fence_chroma 사용 가능 ####
  #### managerforlustre-manager-for-lustre-5.1 레포지토리에 패키지 있음 ####

mds1

https://wiki.whamcloud.com/display/PUB/Using+Pacemaker+1.1+with+a+Lustre+File+System
#passwd hacluster //모든 서버에서 설정

  #pcs cluster auth mds1.local mds2.local 
  username:
  password:
  mds1.local: Authorized
  mds2.local: Authorized
  #pcs cluster setup --start --name mds-cluster mds1.local mds2.local --enable --token 17000
  #pcs stonith create st-fencing fence_chroma

oss1

  #pcs cluster auth oss1.local oss2.local 
  username:
  password:
  oss1.local: Authorized
  oss2.local: Authorized
  #pcs cluster setup --start --name oss-cluster oss1.local oss2.local --enable --token 17000
  #pcs stonith create st-fencing fence_chroma

iml상 pacemaker 설정

              #pcs cluster node add 10.128.0.11
              Disabling SBD service...
              10.128.0.11: sbd disabled
              Sending remote node configuration files to '10.128.0.11'
              10.128.0.11: successful distribution of the file 'pacemaker_remote authkey'
              10.128.0.12: Corosync updated
              Setting up corosync...
              10.128.0.11: Succeeded
              Synchronizing pcsd certificates on nodes 10.128.0.11...
              10.128.0.11: Success
              Restarting pcsd on the nodes in order to reload the certificates...
              10.128.0.11: Success

              #config corosync 
              #pcs cluster auth 10.128.0.12 -u hacluster -p ********************
              10.128.0.12: Authorized

              #pcs cluster setup --name lustre-ha-cluster --force 10.128.0.12 --transport udp --rrpmode passive --addr0 10.128.0.0 --mcast0 226.94.0.1 --mcastport0 42227 --addr1 10.73.10.0 --mcast1 226.94.1.1 --mcastport1 42227 --token 17000 --fail_recv_const 10
              Destroying cluster on nodes: 10.128.0.12...
              10.128.0.12: Stopping Cluster (pacemaker)...
              10.128.0.12: Successfully destroyed cluster

              Sending 'pacemaker_remote authkey' to '10.128.0.12'
              10.128.0.12: successful distribution of the file 'pacemaker_remote authkey'
              Sending cluster config files to the nodes...
              10.128.0.12: Succeeded

              Synchronizing pcsd certificates on nodes 10.128.0.12...
              10.128.0.12: Success
              Restarting pcsd on the nodes in order to reload the certificates...
              10.128.0.12: Success

stonith resource 만 등록되어있음

iml email 설정

  1. smtp 서버 설정

         sendmail 설치
         # yum install -y sendmail sendmail-cf    //sendmail : 메일전송 메이전트, sendmail설정
         # vi /etc/mail/sendmail.mc               //52,53번 라인 주석 제거 (dnl제거)
         # m4 sendmail.mc > sendmail.cf
         # systemctl restart sendmail
         # yum install -y mailx                   // mail 송신 test
         # mail [email protected]
         Subject: title
         context
         context
         .
         EOT
         
         메일 전송이 되지 않고, 로그에 rejecting connections on daemon MTA: load average: 12 가 표시될 경우
         - sendmail.cf 395번째 라인의 주석(#)제거 후 값을 12 -> 30으로 변경
         # vi sendmail.cf
    
         ----  변경 전 ----
         394 # load average at which we refuse connections
         395 #O RefuseLA=12
    
         ----  변경 후 ----
         394 # load average at which we refuse connections
         395 O RefuseLA=30
          - QueueLA : 처리중인 메일 프로세스가 해당 수치에 도달하면 queue에 적재
          - 메일 프로세스가 해당 수치에 도달하면 접속을 거부
    
  2. iml 설정 # vi /usr/share/chroma-manager/local_settings.py //chroma설치 및 초기 설정 이후 추가 설정시 local_settings.py EMAIL_HOST='hostname.domainname' EMAIL_SENDER='[email protected]' //email의 from에 명시될 이름

         # chroma-config restart
    

lustre_prac's People

Contributors

rohguentak76 avatar

Watchers

 avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.