centos7 搭建集群环境openshift3.10

centos7 搭建集群环境openshift3.10

操作系统:
centos7.9
节点规划:
master:
8G/4H(cpu的虚拟化引擎选项全都打勾开启)
硬盘120G
node1:
4G/4H(cpu的虚拟化引擎选项全都打勾开启)
硬盘120G
node2:
4G/4H(cpu的虚拟化引擎选项全都打勾开启)
硬盘120G
下面安装部分的参考文档如下:
https://blog.youkuaiyun.com/huqigang/article/details/82351972
https://blog.youkuaiyun.com/qq_31977125/article/details/82912970
https://www.cnblogs.com/dukuan/p/8711717.html
https://www.cnblogs.com/mascot1/p/9571601.html

第一部分 初始化设置:

  1. 配置主机名
    hostnamectl set-hostname master
    hostnamectl set-hostname node1
    hostnamectl set-hostname node2

  2. 配置hosts文件(各个节点操作):

    cat /etc/hosts
    ip地址 master
    ip地址 node1
    ip地址 node2
    
  3. 开启selinux(各个节点操作)

    cat /etc/sysconfig/selinux 
    SELINUX=enforcing 
    SELINUXTYPE=targeted
    

    重启主机生效

  4. 修改配置文件ifcfg-ens33(各个节点操作)

    cat /etc/sysconfig/network-scripts/ifcfg-ens33 
    
    添加如下参数 
    NM_CONTROLLED=yes //启用网络控制,即修改后可以实时生效,修改后无需要重启网卡立即生效 
    
    修改好之后重启网络 
    service network restart
    
  5. 停掉NetworkManager,iptables,firewalld(各个节点操作)

    systemctl stop NetworkManager
    systemctl disable NetworkManager
    systemctl stop firewalld
    systemctl stop iptables
    systemctl disable firewalld
    systemctl disable iptables
    
  6. 在master节点生成密钥,分发到各node节点(master节点操作)

    ssh-keygen -t rsa
    ssh-copy-id -i .ssh/id_rsa.pub master
    ssh-copy-id -i .ssh/id_rsa.pub node1
    ssh-copy-id -i .ssh/id_rsa.pub node2
    
  7. ntpdate做时间同步(各个节点操作)

    ntpdate time2.aliyun.com
    
  8. 安装基础包(各个节点操作)

    yum update -y
    
    yum install -y wget git net-tools bind-utils yum-utils iptables-services bridge-utils bash-completion kexec-tools sos psacct java-1.8.0-openjdk-headless python-passlib
    
    yum -y install wget net-tools nfs-utils lrzsz gcc gcc-c++ make cmake libxml2-devel openssl-devel curl curl-devel unzip sudo ntp libaio-devel wget vim ncurses-devel autoconf automake zlib-devel python-devel epel-release lrzsz openssh-server socat ipvsadm conntrack
    
  9. 安装ansible2.6.5(master节点操作.openshift3.10需要安装ansible2.6.5,否则会报错)
    在这里插入图片描述
    yum install ansible-2.6.5-1.el7.ans.noarch.rpm

  10. 上传openshift-3.10安装包到master节点
    https://mirror.openshift.com/pub/openshift-v4/clients/ocp/
    在这里插入图片描述
    解压:unzip openshift-ansible-release-3.10.zip /root/

  11. 安装docker(各个几点操作)

yum install -y docker-1.13.1

(1)修改docker配置文件

cat /etc/sysconfig/docker

之前的OPTIONS注释掉,变成下面这行
OPTIONS='--selinux-enabled=false --signature-verification=False'

(2)配置docker镜像加速器

cat /etc/docker/daemon.json
{"registry-mirrors": ["http://6e9e5b27.m.daocloud.io"]}

重启docker
systemctl daemon-reload
systemctl restart docker.service

(3)配置私有仓库(在master节点操作)

docker pull registry:2.5

yum install httpd -y

service httpd start

chkconfig httpd on

mkdir -p /opt/registry-var/auth/

#生成用户名/密码访问 (这里设置用户名admin , 密码 admin)
docker run --entrypoint htpasswd registry:2.5 -Bbn admin admin  >> /opt/registry-var/auth/htpasswd

设置配置文件(注意yml文件格式)
mkdir -p /opt/registry-var/config
vim /opt/registry-var/config/config.yml
在这里插入图片描述
启动服务

docker run -d -p 5000:5000 --restart=always  --name=registry -v /opt/registry-var/config/:/etc/docker/registry/ -v /opt/registry-var/auth/:/auth/ -e "REGISTRY_AUTH=htpasswd"  -e "REGISTRY_AUTH_HTPASSWD_REALM=Registry Realm" -e REGISTRY_AUTH_HTPASSWD_PATH=/auth/htpasswd -v /opt/registry-var/:/var/lib/registry/ registry:2.5

配置https权限支持(在master,node1,node2上均做修改)
由于harbor为了安全性考虑,默认是需要https证书支持的

在这里插入代码片

vim /etc/docker/daemon.json
{
“registry-mirrors”: [“http://6e9e5b27.m.daocloud.io”],
“insecure-registries”:[“103.104.222.33:5000”]
}

重启docker

systemctl daemon-reload
systemctl restart docker.service

测试私有仓库是否可以登陆(各个节点均操作)

docker login 103.104.222.33:5000

输入用户名和密码显示如下说明登陆成功,其他机器如果想要push或者pull镜像到这个私有仓库,需要docker login 103.104.222.33:5000 输入用户名和密码登陆成功之后才可以上传下载
在这里插入图片描述

1、报错现象及原因 
报错:Warning: failed to get default registry endpoint from daemon (Cannot connect to the Docker daemon. Is the docker daemon running on this host?). 
原因:刚装完docker服务没有启动,启动命令如下 

$ sudo systemctl start docker.service 

2、其它报错 
启动后不加sudo 操作诸如docker ps的命令依然会报以下错误 
Got permission denied while trying to connect to the Docker daemon socket at unix:///var/run/docker.sock: Get http://%2Fvar%2Frun%2Fdocker.sock/v1.38/info: dial unix /var/run/docker.sock: connect: permission denied 需要添加当前用户到docker用户组,命令如下 

$ sudo usermod -aG docker $USER && newgrp docker

(4)设置docker开机自启动

systemctl enable docker
systemctl is-active docker

(5)配置docker 后端存储,
所有节点
/etc/sysconfig/docker-storage-setup修改如下内容:

vi /etc/sysconfig/docker-storage-setup
VG=centos #使用一个已建立的卷组,作为docker的后端存储

所有Node节点执行

docker-storage-setup

内部做的操作:

将空余块设备(可以是分区)创建成physical volume(pv) 
在由这些PV组成volume group(vg) 
从vg中建立两个logical volume(lv),data和matedata 
将data和matedata映射成thin-pool

12.下载镜像
master节点:

docker pull quay.io/coreos/etcd:v3.2.22
docker pull openshift/origin-control-plane:v3.10
docker pull docker.io/openshift/origin-service-catalog:v3.10
docker pull openshift/origin-node:v3.10
docker pull openshift/origin-deployer:v3.10
docker pull openshift/origin-deployer:v3.10.0
docker pull openshift/origin-template-service-broker:v3.10
docker pull openshift/origin-pod:v3.10
docker pull openshift/origin-pod:v3.10.0
docker pull openshift/origin-web-console:v3.10
docker pull openshift/origin-docker-registry:v3.10
docker pull openshift/origin-docker-registry:v3.10
docker pull openshift/origin-haproxy-router:v3.10
docker pull cockpit/kubernetes:latest
docker pull docker.io/cockpit/kubernetes:latest
docker pull docker.io/openshift/origin-control-plane:v3.10
docker pull docker.io/openshift/origin-deployer:v3.10
docker pull docker.io/openshift/origin-docker-registry:v3.10
docker pull docker.io/openshift/origin-haproxy-router:v3.10
docker pull docker.io/openshift/origin-pod:v3.10
docker pull quay.io/coreos/etcd:v3.2.22

node1和node2节点:

docker pull quay.io/coreos/etcd:v3.2.22
docker pull openshift/origin-control-plane:v3.10
docker pull openshift/origin-node:v3.10
docker pull docker.io/openshift/origin-node:v3.10
docker pull quay.io/coreos/etcd:v3.2.22
docker pull openshift/origin-haproxy-router:v3.10
docker pull openshift/origin-deployer:v3.10
docker pull openshift/origin-pod:v3.10
docker pull ansibleplaybookbundle/origin-ansible-service-broker:v3.10
docker pull openshift/origin-docker-registry:v3.10
docker pull openshift/origin-docker-registry:v3.10
docker pull cockpit/kubernetes:latest
docker pull openshift/origin-haproxy-router:v3.10
docker pull docker.io/cockpit/kubernetes:latest
docker pull docker.io/openshift/origin-control-plane:v3.10
docker pull docker.io/openshift/origin-deployer:v3.10
docker pull docker.io/openshift/origin-docker-registry:v3.10
docker pull docker.io/openshift/origin-haproxy-router:v3.10
docker pull docker.io/openshift/origin-pod:v3.10
docker pull quay.io/coreos/etcd:v3.2.22

可以将所有命令放入文本文件registry,由脚本执行docker pull

#!/bin/bash
cat ./registry|while read line;do
    ${line}
    sleep 120     #每pull一个镜像,等待120秒
done

13.配置ansible的hosts文件
cat /etc/ansible/hosts

[OSEv3:children]
masters
nodes
etcd
[OSEv3:vars]
openshift_deployment_type=origin
ansible_ssh_user=root
ansible_become=yes
openshift_repos_enable_testing=true
openshift_enable_service_catalog=false
template_service_broker_install=false
debug_level=4
openshift_clock_enabled=true
openshift_version=3.10.0
openshift_image_tag=v3.10
openshift_disable_check=disk_availability,docker_storage,memory_availability,docker_image_availability,os_sdn_network_plugin_name=redhat/openshift-ovs-multitenant i
openshift_master_identity_providers=[{'name': 'htpasswd_auth','login': 'true', 'challenge': 'true','kind': 'HTPasswdPasswordIdentityProvider'}]
[masters]
master
[nodes]
master openshift_node_group_name='node-config-master'
node1 openshift_node_group_name='node-config-master'
node2 openshift_node_group_name='node-config-master'
[etcd]
master

14.安装集群
(1)安装前预配置检查

ansible-playbook -i /etc/ansible/hosts openshift-ansible-release-3.10/playbooks/prerequisites.yml

检查如果显示如下,则没有报错,那么开始安装
在这里插入图片描述
(2)安装

ansible-playbook -i /etc/ansible/hosts openshift-ansible-release-3.10/playbooks/deploy_cluster.yml

执行deploy时主机dns导致连外网失败(在执行上面deploy时,需要在每个节点ping www.baidu.com,如果ping不通,解决方案如下)
临时解决方案更改/etc/resolv.conf
当部署的时候看到retry,就需要在master和node节点执行下面命令,这样就可以继续ping通外网
echo nameserver 8.8.8.8 >>/etc/resolv.conf

需要给节点打标签
TASK [openshift_manage_node : Set node schedulability]
到这个task之后执行下面部分

oc label node node1 node-role.kubernetes.io/infra=true
oc label node node2 node-role.kubernetes.io/infra=true

echo nameserver 8.8.8.8 >>/etc/resolv.conf
显示如下,说明安装成功
在这里插入图片描述
15.创建管理员账号

首次新建用户密码
htpasswd -cb /etc/origin/master/htpasswd admin admin

添加用户密码
htpasswd -b /etc/origin/master/htpasswd dev dev

以集群管理员登录
oc login -u system:admin

给用户分配一个集群管理员角色
oc adm policy add-cluster-role-to-user cluster-admin admin

16.在浏览器登陆console控制台
https://master:8443
注意这里有个坑,openshift的web console要用域名登录,做好域名IP映射(etc/hosts)
账号:admin
密码:admin
在这里插入图片描述

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

clarence.wei

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值