centos7 搭建集群环境openshift3.10
操作系统:
centos7.9
节点规划:
master:
8G/4H(cpu的虚拟化引擎选项全都打勾开启)
硬盘120G
node1:
4G/4H(cpu的虚拟化引擎选项全都打勾开启)
硬盘120G
node2:
4G/4H(cpu的虚拟化引擎选项全都打勾开启)
硬盘120G
下面安装部分的参考文档如下:
https://blog.youkuaiyun.com/huqigang/article/details/82351972
https://blog.youkuaiyun.com/qq_31977125/article/details/82912970
https://www.cnblogs.com/dukuan/p/8711717.html
https://www.cnblogs.com/mascot1/p/9571601.html
第一部分 初始化设置:
-
配置主机名
hostnamectl set-hostname master
hostnamectl set-hostname node1
hostnamectl set-hostname node2 -
配置hosts文件(各个节点操作):
cat /etc/hosts ip地址 master ip地址 node1 ip地址 node2
-
开启selinux(各个节点操作)
cat /etc/sysconfig/selinux SELINUX=enforcing SELINUXTYPE=targeted
重启主机生效
-
修改配置文件ifcfg-ens33(各个节点操作)
cat /etc/sysconfig/network-scripts/ifcfg-ens33 添加如下参数 NM_CONTROLLED=yes //启用网络控制,即修改后可以实时生效,修改后无需要重启网卡立即生效 修改好之后重启网络 service network restart
-
停掉NetworkManager,iptables,firewalld(各个节点操作)
systemctl stop NetworkManager systemctl disable NetworkManager systemctl stop firewalld systemctl stop iptables systemctl disable firewalld systemctl disable iptables
-
在master节点生成密钥,分发到各node节点(master节点操作)
ssh-keygen -t rsa ssh-copy-id -i .ssh/id_rsa.pub master ssh-copy-id -i .ssh/id_rsa.pub node1 ssh-copy-id -i .ssh/id_rsa.pub node2
-
ntpdate做时间同步(各个节点操作)
ntpdate time2.aliyun.com
-
安装基础包(各个节点操作)
yum update -y yum install -y wget git net-tools bind-utils yum-utils iptables-services bridge-utils bash-completion kexec-tools sos psacct java-1.8.0-openjdk-headless python-passlib yum -y install wget net-tools nfs-utils lrzsz gcc gcc-c++ make cmake libxml2-devel openssl-devel curl curl-devel unzip sudo ntp libaio-devel wget vim ncurses-devel autoconf automake zlib-devel python-devel epel-release lrzsz openssh-server socat ipvsadm conntrack
-
安装ansible2.6.5(master节点操作.openshift3.10需要安装ansible2.6.5,否则会报错)
yum install ansible-2.6.5-1.el7.ans.noarch.rpm -
上传openshift-3.10安装包到master节点
https://mirror.openshift.com/pub/openshift-v4/clients/ocp/
解压:unzip openshift-ansible-release-3.10.zip /root/
-
安装docker(各个几点操作)
yum install -y docker-1.13.1
(1)修改docker配置文件
cat /etc/sysconfig/docker
之前的OPTIONS注释掉,变成下面这行
OPTIONS='--selinux-enabled=false --signature-verification=False'
(2)配置docker镜像加速器
cat /etc/docker/daemon.json
{"registry-mirrors": ["http://6e9e5b27.m.daocloud.io"]}
重启docker
systemctl daemon-reload
systemctl restart docker.service
(3)配置私有仓库(在master节点操作)
docker pull registry:2.5
yum install httpd -y
service httpd start
chkconfig httpd on
mkdir -p /opt/registry-var/auth/
#生成用户名/密码访问 (这里设置用户名admin , 密码 admin)
docker run --entrypoint htpasswd registry:2.5 -Bbn admin admin >> /opt/registry-var/auth/htpasswd
设置配置文件(注意yml文件格式)
mkdir -p /opt/registry-var/config
vim /opt/registry-var/config/config.yml
启动服务
docker run -d -p 5000:5000 --restart=always --name=registry -v /opt/registry-var/config/:/etc/docker/registry/ -v /opt/registry-var/auth/:/auth/ -e "REGISTRY_AUTH=htpasswd" -e "REGISTRY_AUTH_HTPASSWD_REALM=Registry Realm" -e REGISTRY_AUTH_HTPASSWD_PATH=/auth/htpasswd -v /opt/registry-var/:/var/lib/registry/ registry:2.5
配置https权限支持(在master,node1,node2上均做修改)
由于harbor为了安全性考虑,默认是需要https证书支持的
在这里插入代码片
vim /etc/docker/daemon.json
{
“registry-mirrors”: [“http://6e9e5b27.m.daocloud.io”],
“insecure-registries”:[“103.104.222.33:5000”]
}
重启docker
systemctl daemon-reload
systemctl restart docker.service
测试私有仓库是否可以登陆(各个节点均操作)
docker login 103.104.222.33:5000
输入用户名和密码显示如下说明登陆成功,其他机器如果想要push或者pull镜像到这个私有仓库,需要docker login 103.104.222.33:5000 输入用户名和密码登陆成功之后才可以上传下载
1、报错现象及原因
报错:Warning: failed to get default registry endpoint from daemon (Cannot connect to the Docker daemon. Is the docker daemon running on this host?).
原因:刚装完docker服务没有启动,启动命令如下
$ sudo systemctl start docker.service
2、其它报错
启动后不加sudo 操作诸如docker ps的命令依然会报以下错误
Got permission denied while trying to connect to the Docker daemon socket at unix:///var/run/docker.sock: Get http://%2Fvar%2Frun%2Fdocker.sock/v1.38/info: dial unix /var/run/docker.sock: connect: permission denied 需要添加当前用户到docker用户组,命令如下
$ sudo usermod -aG docker $USER && newgrp docker
(4)设置docker开机自启动
systemctl enable docker
systemctl is-active docker
(5)配置docker 后端存储,
所有节点
/etc/sysconfig/docker-storage-setup修改如下内容:
vi /etc/sysconfig/docker-storage-setup
VG=centos #使用一个已建立的卷组,作为docker的后端存储
所有Node节点执行
docker-storage-setup
内部做的操作:
将空余块设备(可以是分区)创建成physical volume(pv)
在由这些PV组成volume group(vg)
从vg中建立两个logical volume(lv),data和matedata
将data和matedata映射成thin-pool
12.下载镜像
master节点:
docker pull quay.io/coreos/etcd:v3.2.22
docker pull openshift/origin-control-plane:v3.10
docker pull docker.io/openshift/origin-service-catalog:v3.10
docker pull openshift/origin-node:v3.10
docker pull openshift/origin-deployer:v3.10
docker pull openshift/origin-deployer:v3.10.0
docker pull openshift/origin-template-service-broker:v3.10
docker pull openshift/origin-pod:v3.10
docker pull openshift/origin-pod:v3.10.0
docker pull openshift/origin-web-console:v3.10
docker pull openshift/origin-docker-registry:v3.10
docker pull openshift/origin-docker-registry:v3.10
docker pull openshift/origin-haproxy-router:v3.10
docker pull cockpit/kubernetes:latest
docker pull docker.io/cockpit/kubernetes:latest
docker pull docker.io/openshift/origin-control-plane:v3.10
docker pull docker.io/openshift/origin-deployer:v3.10
docker pull docker.io/openshift/origin-docker-registry:v3.10
docker pull docker.io/openshift/origin-haproxy-router:v3.10
docker pull docker.io/openshift/origin-pod:v3.10
docker pull quay.io/coreos/etcd:v3.2.22
node1和node2节点:
docker pull quay.io/coreos/etcd:v3.2.22
docker pull openshift/origin-control-plane:v3.10
docker pull openshift/origin-node:v3.10
docker pull docker.io/openshift/origin-node:v3.10
docker pull quay.io/coreos/etcd:v3.2.22
docker pull openshift/origin-haproxy-router:v3.10
docker pull openshift/origin-deployer:v3.10
docker pull openshift/origin-pod:v3.10
docker pull ansibleplaybookbundle/origin-ansible-service-broker:v3.10
docker pull openshift/origin-docker-registry:v3.10
docker pull openshift/origin-docker-registry:v3.10
docker pull cockpit/kubernetes:latest
docker pull openshift/origin-haproxy-router:v3.10
docker pull docker.io/cockpit/kubernetes:latest
docker pull docker.io/openshift/origin-control-plane:v3.10
docker pull docker.io/openshift/origin-deployer:v3.10
docker pull docker.io/openshift/origin-docker-registry:v3.10
docker pull docker.io/openshift/origin-haproxy-router:v3.10
docker pull docker.io/openshift/origin-pod:v3.10
docker pull quay.io/coreos/etcd:v3.2.22
可以将所有命令放入文本文件registry,由脚本执行docker pull
#!/bin/bash
cat ./registry|while read line;do
${line}
sleep 120 #每pull一个镜像,等待120秒
done
13.配置ansible的hosts文件
cat /etc/ansible/hosts
[OSEv3:children]
masters
nodes
etcd
[OSEv3:vars]
openshift_deployment_type=origin
ansible_ssh_user=root
ansible_become=yes
openshift_repos_enable_testing=true
openshift_enable_service_catalog=false
template_service_broker_install=false
debug_level=4
openshift_clock_enabled=true
openshift_version=3.10.0
openshift_image_tag=v3.10
openshift_disable_check=disk_availability,docker_storage,memory_availability,docker_image_availability,os_sdn_network_plugin_name=redhat/openshift-ovs-multitenant i
openshift_master_identity_providers=[{'name': 'htpasswd_auth','login': 'true', 'challenge': 'true','kind': 'HTPasswdPasswordIdentityProvider'}]
[masters]
master
[nodes]
master openshift_node_group_name='node-config-master'
node1 openshift_node_group_name='node-config-master'
node2 openshift_node_group_name='node-config-master'
[etcd]
master
14.安装集群
(1)安装前预配置检查
ansible-playbook -i /etc/ansible/hosts openshift-ansible-release-3.10/playbooks/prerequisites.yml
检查如果显示如下,则没有报错,那么开始安装
(2)安装
ansible-playbook -i /etc/ansible/hosts openshift-ansible-release-3.10/playbooks/deploy_cluster.yml
执行deploy时主机dns导致连外网失败(在执行上面deploy时,需要在每个节点ping www.baidu.com,如果ping不通,解决方案如下)
临时解决方案更改/etc/resolv.conf
当部署的时候看到retry,就需要在master和node节点执行下面命令,这样就可以继续ping通外网
echo nameserver 8.8.8.8 >>/etc/resolv.conf
需要给节点打标签
TASK [openshift_manage_node : Set node schedulability]
到这个task之后执行下面部分
oc label node node1 node-role.kubernetes.io/infra=true
oc label node node2 node-role.kubernetes.io/infra=true
echo nameserver 8.8.8.8 >>/etc/resolv.conf
显示如下,说明安装成功
15.创建管理员账号
首次新建用户密码
htpasswd -cb /etc/origin/master/htpasswd admin admin
添加用户密码
htpasswd -b /etc/origin/master/htpasswd dev dev
以集群管理员登录
oc login -u system:admin
给用户分配一个集群管理员角色
oc adm policy add-cluster-role-to-user cluster-admin admin
16.在浏览器登陆console控制台
https://master:8443
注意这里有个坑,openshift的web console要用域名登录,做好域名IP映射(etc/hosts)
账号:admin
密码:admin