This commit is contained in:
2021-11-14 14:32:08 +08:00
parent f75ad8bedd
commit b0f6120151
152 changed files with 22219 additions and 8 deletions

832
content/post/k8s-install.md Normal file
View File

@@ -0,0 +1,832 @@
---
title: "K8s 二进制安装"
date: 2019-10-30T01:09:48+08:00
lastmod: 2020-02-03T11:43:30+08:00
keywords: []
tags: ["kubernetes", "k8s", "二进制"]
categories: ["container"]
---
# 环境
角色 | 主机名 | 内网 IP | 集群 IP | 操作系统 | 服务 | 执行目录
---- | ---- | ---- | ---- | ---- | ---- | ----
部署机 k8s-master | master120 | 10.0.4.120 | - | CentOS | kube-apiserver kube-scheduler kube-controller-manager | /opt/kubernetes/
etcd-node | etcd121 | 10.0.4.121 | 10.10.10.121 | CentOS | etcd | /opt/etcd/
etcd-node | etcd122 | 10.0.4.122 | 10.10.10.122 | CentOS | etcd | /opt/etcd/
etcd-node | etcd123 | 10.0.4.123 | 10.10.10.123 | CentOS | etcd | /opt/etcd/
k8s-node | node124 | 10.0.4.124 | - | CentOS | docker flannel kubelet kube-proxy | /opt/kubernetes/
k8s-node | node125 | 10.0.4.125 | - | CentOS | docker flannel kubelet kube-proxy | /opt/kubernetes/
k8s-node | node126 | 10.0.4.126 | - | CentOS | docker flannel kubelet kube-proxy | /opt/kubernetes/
- 全部服务器关闭 firewalld 和 selinux禁用 swap部署机(master120)可免密 ssh 登陆其他服务器
- 软件版本
- CentOS: 7.7
- etcd: 3.3.18
- docker: ce-19.03.5
- flannel: 0.11.0
- kubernetes: 1.17.2
- k8s牵扯到多个网段这里说明下
- 10.0.4.0/24 该网段是服务器物理网卡 IP 地址段,通过该地址访问互联网
- 10.10.10.0/24 该网段是杜撰的,但也配置在服务器物理网卡上,用于 etcd 集群节点间通信,与 k8s 集群无关
- 10.10.9.0/24 该网段是杜撰的,分配 k8s service 的 clusterIP
- 172.17.0.0/24 该网段是杜撰的,是 docker 网桥自带网段,也是 flannel 提供的网段,实现不同节点间的容器互通
- 172.16.0.0/24 该网段是杜撰的,是 k8s pod 的 IP 地址区间,用于区别流量来源
# 部署 etcd 集群
- 在部署机(master120)上操作下面步骤
- 创建 etcd 部署目录
```bash
mkdir /home/deploy/etcd/{bin,cfg,ssl} -p
mkdir /home/deploy/ssl/etcd -p
```
- 安装 cfssl 工具
```bash
curl -o /usr/local/bin/cfssl https://pkg.cfssl.org/R1.2/cfssl_linux-amd64
curl -o /usr/local/bin/cfssljson https://pkg.cfssl.org/R1.2/cfssljson_linux-amd64
curl -o /usr/local/bin/cfssl-certinfo https://pkg.cfssl.org/R1.2/cfssl-certinfo_linux-amd64
chmod 0755 /usr/local/bin/cfssl*
```
- 创建 ca-config.json 文件
```bash
cat > /home/deploy/ssl/etcd/ca-config.json <<< '
{
"signing": {
"default": {
"expiry": "87600h"
},
"profiles": {
"etcd": {
"expiry": "87600h",
"usages": [
"signing",
"key encipherment",
"server auth",
"client auth"
]
}
}
}
}
'
```
- 创建 ca-csr.json 文件
```bash
cat > /home/deploy/ssl/etcd/ca-csr.json <<< '
{
"CN": "etcd",
"key": {
"algo": "rsa",
"size": 2048
},
"names": [
{
"C": "CN",
"L": "Beijing",
"ST": "Beijing"
}
]
}
'
```
- 创建 server-csr.json 文件
```bash
# 注意修改 "10.0.4.*" 和 "10.10.10.*" 为自己环境地址
cat > /home/deploy/ssl/etcd/server-csr.json <<< '
{
"CN": "etcd",
"hosts": [
"10.0.4.121",
"10.0.4.122",
"10.0.4.123",
"10.10.10.121",
"10.10.10.122",
"10.10.10.123",
"127.0.0.1"
],
"key": {
"algo": "rsa",
"size": 2048
},
"names": [
{
"C": "CN",
"L": "BeiJing",
"ST": "BeiJing"
}
]
}
'
```
- 生成证书
```bash
cd /home/deploy/ssl/etcd/
# 生成 ca.pem ca-key.pem
cfssl gencert -initca ca-csr.json | cfssljson -bare ca -
# 生成 server.pem server-key.pem
cfssl gencert \
-ca=ca.pem \
-ca-key=ca-key.pem \
-config=ca-config.json \
-profile=etcd \
server-csr.json | cfssljson -bare server
# 复制证书到部署目录
scp *.pem /home/deploy/etcd/ssl/
```
- 下载 etcd 二进制包
```bash
cd /home/deploy/
curl -L -O https://github.com/etcd-io/etcd/releases/download/v3.3.18/etcd-v3.3.18-linux-amd64.tar.gz
tar zxf etcd-v3.3.18-linux-amd64.tar.gz
scp etcd-v3.3.18-linux-amd64/{etcd,etcdctl} etcd/bin/
```
- 创建 etcd 配置文件
```bash
# 这里的 etcd 虚拟机都有两个网卡,一个用于提供服务,另一个用于集群通信
# 注意修改 "10.0.4.*" 和 "10.10.10.*" 为自己环境地址
cat > /home/deploy/etcd/cfg/etcd <<<'
# [Member]
ETCD_NAME="etcdXXX"
ETCD_DATA_DIR="/var/lib/etcd/default.etcd"
ETCD_LISTEN_PEER_URLS="https://10.10.10.XXX:2380"
ETCD_LISTEN_CLIENT_URLS="https://10.0.4.XXX:2379"
# [Clustering]
ETCD_INITIAL_ADVERTISE_PEER_URLS="https://10.10.10.XXX:2380"
ETCD_ADVERTISE_CLIENT_URLS="https://10.0.4.XXX:2379"
ETCD_INITIAL_CLUSTER="etcd121=https://10.10.10.121:2380,etcd122=https://10.10.10.122:2380,etcd123=https://10.10.10.123:2380"
ETCD_INITIAL_CLUSTER_TOKEN="fucking-etcd-cluster"
ETCD_INITIAL_CLUSTER_STATE="new"
'
```
- 创建 etcd.service
```bash
cat > /home/deploy/etcd.service <<<'
[Unit]
Description=Etcd Server
After=network.target
After=network-online.target
Wants=network-online.target
[Service]
Type=notify
EnvironmentFile=/opt/etcd/cfg/etcd
ExecStart=/opt/etcd/bin/etcd \
--name=${ETCD_NAME} \
--data-dir=${ETCD_DATA_DIR} \
--listen-peer-urls=${ETCD_LISTEN_PEER_URLS} \
--listen-client-urls=${ETCD_LISTEN_CLIENT_URLS},http://127.0.0.1:2379 \
--advertise-client-urls=${ETCD_ADVERTISE_CLIENT_URLS} \
--initial-advertise-peer-urls=${ETCD_INITIAL_ADVERTISE_PEER_URLS} \
--initial-cluster=${ETCD_INITIAL_CLUSTER} \
--initial-cluster-token=${ETCD_INITIAL_CLUSTER_TOKEN} \
--initial-cluster-state=${ETCD_INITIAL_CLUSTER_STATE} \
--cert-file=/opt/etcd/ssl/server.pem \
--key-file=/opt/etcd/ssl/server-key.pem \
--peer-cert-file=/opt/etcd/ssl/server.pem \
--peer-key-file=/opt/etcd/ssl/server-key.pem \
--trusted-ca-file=/opt/etcd/ssl/ca.pem \
--peer-trusted-ca-file=/opt/etcd/ssl/ca.pem
Restart=on-failure
LimitNOFILE=65536
[Install]
WantedBy=multi-user.target
'
```
- 部署到远程三个 etcd 节点(etcd121、etcd122、etcd123)
```bash
# 注意修改 "10.0.4.*" 为自己环境地址
cd /home/deploy
for id in $(seq 121 123); do
ip="10.0.4.$id"
scp -r etcd $ip:/opt/
ssh $ip "sed -i 's/XXX/$id/g' /opt/etcd/cfg/etcd"
scp etcd.service $ip:/usr/lib/systemd/system/
systemctl -H $ip daemon-reload
systemctl -H $ip enable etcd
done
```
- 启动三个 etcd 节点的 etcd 服务
```bash
# 注意修改 "10.0.4.*" 为自己环境地址
for ip in $(seq -f'10.0.4.%g' 121 123); do
systemctl -H $ip start etcd
done
```
- 查看 etcd 集群状态
```bash
# 注意修改 "10.0.4.*" 为自己环境地址
cd /home/deploy/etcd/ssl
../bin/etcdctl \
--ca-file=ca.pem \
--cert-file=server.pem \
--key-file=server-key.pem \
--endpoints="https://10.0.4.121:2379" \
cluster-health
```
# 安装 Docker
- 在每个 k8s node 服务器(node124、node125、node126)上操作下面步骤
- 安装 docker-ce[参考这里](https://my.oschina.net/colben/blog/1505141#h1_2)
# 部署 Flannel 网络
- 在部署机(master120)上操作下面步骤
- 创建 flannel 部署目录
```bash
cd /home/deploy
mkdir flannel/{bin,cfg,ssl} -p
# 复制 etcd 证书到 flannel 证书目录下
rm -rf flannel/ssl/etcd && scp -r etcd/ssl flannel/ssl/etcd
```
- 连接 etcd写入预定义的子网段
```bash
# 这里的预定义字段是 "172.17.0.0/16",推荐用这个,与 docker 原生网段一致
# 注意修改 "10.0.4.*" 为自己环境地址
cd /home/deploy/etcd/ssl
../bin/etcdctl \
--ca-file=ca.pem \
--cert-file=server.pem \
--key-file=server-key.pem \
--endpoints="https://10.0.4.122:2379" \
set /coreos.com/network/config '{"Network": "172.17.0.0/16", "Backend": {"Type": "vxlan"}}'
```
- 获取 flannel 二进制包
```bash
cd /home/deploy
curl -L -O https://github.com/coreos/flannel/releases/download/v0.11.0/flannel-v0.11.0-linux-amd64.tar.gz
tar zxf flannel-v0.11.0-linux-amd64.tar.gz
scp flanneld mk-docker-opts.sh flannel/bin/
```
- 创建 flannel 配置文件
```bash
# 注意修改 "10.0.4.*" 为自己环境地址
cat > /home/deploy/flannel/cfg/flanneld <<< '
FLANNEL_OPTIONS=" \
--etcd-endpoints=https://10.0.4.121:2379,https://10.0.4.122:2379,https://10.0.4.123:2379 \
-etcd-cafile=/opt/kubernetes/ssl/etcd/ca.pem \
-etcd-certfile=/opt/kubernetes/ssl/etcd/server.pem \
-etcd-keyfile=/opt/kubernetes/ssl/etcd/server-key.pem \
"
'
```
- 创建 flanneld.service
```bash
cat > /home/deploy/flanneld.service <<< '
[Unit]
Description=Flanneld overlay address etcd agent
After=network-online.target network.target
Before=docker.service
[Service]
Type=notify
EnvironmentFile=/opt/kubernetes/cfg/flanneld
ExecStart=/opt/kubernetes/bin/flanneld --ip-masq $FLANNEL_OPTIONS
ExecStartPost=/opt/kubernetes/bin/mk-docker-opts.sh -k DOCKER_NETWORK_OPTIONS -d /run/flannel/subnet.env
Restart=on-failure
[Install]
WantedBy=multi-user.target
'
```
- 修改 docker.service用于从指定的子网段启动 docker
```bash
# 关键在 "EnvironmentFile=/run/flannel/subnet.env"
# 该文件就是 flanneld 服务生成的 docker 参数
# 这里仅作记录,具体实现已移到下一步的部署脚本中
```
- 部署到远程三个 k8s node 节点(node124、node125、node126)
```bash
# 注意修改 "10.0.4.*" 为自己环境地址
cd /home/deploy
for ip in $(seq -f'10.0.4.%g' 124 126); do
systemctl -H $ip stop docker
ssh $ip "mkdir /opt/kubernetes"
scp -r flannel/* $ip:/opt/kubernetes/
scp flanneld.service $ip:/usr/lib/systemd/system/
ssh $ip 'sed -i \
-e "/^Type/aEnvironmentFile=/run/flannel/subnet.env" \
-e "/^ExecStart/s/$/\$DOCKER_NETWORK_OPTIONS/" \
/usr/lib/systemd/system/docker.service \
'
systemctl -H $ip daemon-reload
systemctl -H $ip enable flanneld
done
```
- 启动三个 k8s node 节点的 flanneld 和 docker 服务
```bash
# 注意修改 "10.0.4.*" 为自己环境地址
for ip in $(seq -f'10.0.4.%g' 124 126); do
systemctl -H $ip start flanneld
systemctl -H $ip start docker
done
```
- 启动完成后,不同节点的 docker0 网卡 ip 互通
# 部署 k8s master 节点
- 部署前确保前面的 etcd 集群、flannel 网络和 docker 都是正常的
- 在部署机(master120即当前节点)上操作下面步骤
- 创建 master 部署目录
```bash
cd /home/deploy
mkdir master/{bin,cfg,ssl} -p
mkdir ssl/master -p
# 复制 etcd 证书到 master 证书目录下
rm -rf master/ssl/etcd && scp -r etcd/ssl master/ssl/etcd
```
- 创建 ca-config.json 文件
```bash
cat > /home/deploy/ssl/master/ca-config.json <<< '
{
"signing": {
"default": {
"expiry": "87600h"
},
"profiles": {
"kubernetes": {
"expiry": "87600h",
"usages": [
"signing",
"key encipherment",
"server auth",
"client auth"
]
}
}
}
}
'
```
- 创建 ca-csr.json 文件
```bash
cat > /home/deploy/ssl/master/ca-csr.json <<< '
{
"CN": "kubernetes",
"key": {
"algo": "rsa",
"size": 2048
},
"names": [
{
"C": "CN",
"L": "Beijing",
"ST": "Beijing",
"O": "k8s",
"OU": "System"
}
]
}
'
```
- 创建 kube-apiserver-csr.json 文件
```bash
# 这里的 10.10.9.1 是 kubernetes service 的集群地址
# 该地址默认是下文中的 service-cluster-ip-range 网段的第一个 ip
# dns 组件会用到
# 注意修改 "10.0.4.*" 为自己环境地址
cat > /home/deploy/ssl/master/kube-apiserver-csr.json <<< '
{
"CN": "kubernetes",
"hosts": [
"127.0.0.1",
"10.0.4.120",
"10.10.9.1",
"kubernetes",
"kubernetes.default",
"kubernetes.default.svc",
"kubernetes.default.svc.cluster",
"kubernetes.default.svc.cluster.local"
],
"key": {
"algo": "rsa",
"size": 2048
},
"names": [
{
"C": "CN",
"L": "BeiJing",
"ST": "BeiJing",
"O": "k8s",
"OU": "System"
}
]
}
'
```
- 创建 kube-proxy-csr.json 文件
```bash
cat > /home/deploy/ssl/master/kube-proxy-csr.json <<< '
{
"CN": "system:kube-proxy",
"hosts": [],
"key": {
"algo": "rsa",
"size": 2048
},
"names": [
{
"C": "CN",
"L": "BeiJing",
"ST": "BeiJing",
"O": "k8s",
"OU": "System"
}
]
}
'
```
- 生成证书
```bash
cd /home/deploy/ssl/master
cfssl gencert -initca ca-csr.json | cfssljson -bare ca -
cfssl gencert -ca=ca.pem -ca-key=ca-key.pem -config=ca-config.json -profile=kubernetes kube-apiserver-csr.json | cfssljson -bare kube-apiserver
cfssl gencert -ca=ca.pem -ca-key=ca-key.pem -config=ca-config.json -profile=kubernetes kube-proxy-csr.json | cfssljson -bare kube-proxy
# 复制证书到部署目录
scp *.pem /home/deploy/master/ssl/
```
- 获取 kubernetes 二进制包
```bash
cd /home/deploy
curl -L -O https://dl.k8s.io/v1.17.2/kubernetes-server-linux-amd64.tar.gz
tar zxf kubernetes-server-linux-amd64.tar.gz
cd kubernetes/server/bin
scp kube-apiserver kube-scheduler kube-controller-manager kubectl /home/deploy/master/bin/
```
- 创建 token 文件
```bash
# 第一个字符串随机写的,看心情 ……
echo '1111222233334444aaaabbbbccccdddd,kubelet-bootstrap,10001,"system:kubelet-bootstrap"' > /home/deploy/master/cfg/token.csv
```
- 创建 kube-apiserver 配置文件
```bash
# 这里的 service-cluster-ip-range 是 k8s service 地址区间,提供一个与现有网络不通的网段
# 注意修改 "10.0.4.*" 为自己环境地址
cat > /home/deploy/master/cfg/kube-apiserver <<< '
KUBE_APISERVER_OPTS=" \
--logtostderr=true \
--v=4 \
--etcd-servers=https://10.0.4.121:2379,https://10.0.4.122:2379,https://10.0.4.123:2379 \
--bind-address=10.0.4.120 \
--secure-port=6443 \
--advertise-address=10.0.4.120 \
--allow-privileged=true \
--service-cluster-ip-range=10.10.9.0/24 \
--enable-admission-plugins=NamespaceLifecycle,LimitRanger,SecurityContextDeny,ServiceAccount,ResourceQuota,NodeRestriction \
--authorization-mode=RBAC,Node \
--enable-bootstrap-token-auth \
--token-auth-file=/opt/kubernetes/cfg/token.csv \
--service-node-port-range=30000-50000 \
--tls-cert-file=/opt/kubernetes/ssl/kube-apiserver.pem \
--tls-private-key-file=/opt/kubernetes/ssl/kube-apiserver-key.pem \
--client-ca-file=/opt/kubernetes/ssl/ca.pem \
--service-account-key-file=/opt/kubernetes/ssl/ca-key.pem \
--etcd-cafile=/opt/kubernetes/ssl/etcd/ca.pem \
--etcd-certfile=/opt/kubernetes/ssl/etcd/server.pem \
--etcd-keyfile=/opt/kubernetes/ssl/etcd/server-key.pem \
"
'
```
- 创建 kube-apiserver.service 文件
```bash
cat > /home/deploy/kube-apiserver.service <<< '
[Unit]
Description=Kubernetes API Server
Documentation=https://github.com/kubernetes/kubernetes
[Service]
EnvironmentFile=-/opt/kubernetes/cfg/kube-apiserver
ExecStart=/opt/kubernetes/bin/kube-apiserver $KUBE_APISERVER_OPTS
Restart=on-failure
[Install]
WantedBy=multi-user.target
'
```
- 创建 kube-scheduler 配置文件
```bash
cat > /home/deploy/master/cfg/kube-scheduler <<< '
KUBE_SCHEDULER_OPTS=" \
--logtostderr=true \
--v=4 \
--master=127.0.0.1:8080 \
--leader-elect \
"
'
```
- 创建 kube-scheduler.service
```bash
cat > /home/deploy/kube-scheduler.service <<< '
[Unit]
Description=Kubernetes Scheduler
Documentation=https://github.com/kubernetes/kubernetes
[Service]
EnvironmentFile=-/opt/kubernetes/cfg/kube-scheduler
ExecStart=/opt/kubernetes/bin/kube-scheduler $KUBE_SCHEDULER_OPTS
Restart=on-failure
[Install]
WantedBy=multi-user.target
'
```
- 创建 kube-controller-mananger 配置文件
```bash
# 注意这里设置了 cluster-name 为 "my_k8s_cluster"
# 这个名字在后面会用到
# 这里的 service-cluster-ip-range 是 k8s service 地址区间,与之前配置的网段相同
cat > /home/deploy/master/cfg/kube-controller-manager <<< '
KUBE_CONTROLLER_MANAGER_OPTS=" \
--logtostderr=true \
--v=4 \
--master=127.0.0.1:8080 \
--leader-elect=true \
--address=127.0.0.1 \
--service-cluster-ip-range=10.10.9.0/24 \
--cluster-name=my_k8s_cluster \
--cluster-signing-cert-file=/opt/kubernetes/ssl/ca.pem \
--cluster-signing-key-file=/opt/kubernetes/ssl/ca-key.pem \
--root-ca-file=/opt/kubernetes/ssl/ca.pem \
--service-account-private-key-file=/opt/kubernetes/ssl/ca-key.pem \
"
'
```
- 创建 kube-controller-manager.service
```bash
cat > /home/deploy/kube-controller-manager.service <<< '
[Unit]
Description=Kubernetes Controller Manager
Documentation=https://github.com/kubernetes/kubernetes
[Service]
EnvironmentFile=-/opt/kubernetes/cfg/kube-controller-manager
ExecStart=/opt/kubernetes/bin/kube-controller-manager $KUBE_CONTROLLER_MANAGER_OPTS
Restart=on-failure
[Install]
WantedBy=multi-user.target
'
```
- 部署到执行目录
```bash
cd /home/deploy
mkdir /opt/kubernetes
scp -r master/* /opt/kubernetes/
scp kube-apiserver.service kube-scheduler.service kube-controller-manager.service /usr/lib/systemd/system/
systemctl daemon-reload
systemctl enable kube-apiserver kube-scheduler kube-controller-manager
ln -sf /opt/kubernetes/bin/kubectl /usr/local/bin/
```
- 启动 k8s master 各组件
```bash
systemctl start kube-apiserver
systemctl start kube-scheduler
systemctl start kube-controller-manager
```
- 查看集群各组件状态
```bash
kubectl get cs
```
# 部署 k8s node 节点
- 在部署机(master120)上操作下面步骤
- 创建 node 部署目录
```bash
cd /home/deploy
mkdir node/{bin,cfg,ssl} -p
scp kubernetes/server/bin/{kubelet,kube-proxy} node/bin/
```
- 将kubelet-bootstrap用户绑定到系统集群角色
```bash
kubectl create clusterrolebinding kubelet-bootstrap \
--clusterrole=system:node-bootstrapper \
--user=kubelet-bootstrap
```
- 生成 bootstrap.kubeconfig 文件
```bash
export BOOTSTRAP_TOKEN=1111222233334444aaaabbbbccccdddd
export KUBE_APISERVER="https://10.0.4.120:6443"
cd /home/deploy/master/ssl
# 设置集群参数,这里的指定的集群就是前面设置的 "my_k8s_cluster"
kubectl config set-cluster my_k8s_cluster \
--certificate-authority=ca.pem \
--embed-certs=true \
--server=${KUBE_APISERVER} \
--kubeconfig=/home/deploy/node/cfg/bootstrap.kubeconfig
# 设置客户端认证参数
kubectl config set-credentials kubelet-bootstrap \
--token=${BOOTSTRAP_TOKEN} \
--kubeconfig=/home/deploy/node/cfg/bootstrap.kubeconfig
# 设置上下文参数
kubectl config set-context default \
--cluster=my_k8s_cluster \
--user=kubelet-bootstrap \
--kubeconfig=/home/deploy/node/cfg/bootstrap.kubeconfig
# 设置默认上下文
kubectl config use-context default --kubeconfig=/home/deploy/node/cfg/bootstrap.kubeconfig
```
- 生成 kube-proxy.kubeconfig 文件
```bash
export KUBE_APISERVER="https://10.0.4.120:6443"
cd /home/deploy/master/ssl
# 设置集群参数,这里的指定的集群就是前面设置的 "my_k8s_cluster"
kubectl config set-cluster my_k8s_cluster \
--certificate-authority=ca.pem \
--embed-certs=true \
--server=${KUBE_APISERVER} \
--kubeconfig=/home/deploy/node/cfg/kube-proxy.kubeconfig
kubectl config set-credentials kube-proxy \
--client-certificate=kube-proxy.pem \
--client-key=kube-proxy-key.pem \
--embed-certs=true \
--kubeconfig=/home/deploy/node/cfg/kube-proxy.kubeconfig
kubectl config set-context default \
--cluster=my_k8s_cluster \
--user=kube-proxy \
--kubeconfig=/home/deploy/node/cfg/kube-proxy.kubeconfig
kubectl config use-context default --kubeconfig=/home/deploy/node/cfg/kube-proxy.kubeconfig
```
- 创建 kubelet 配置文件
```bash
# --kubeconfig 指定 kubeconfig 文件位置,会自动生成
# --cert-dir 颁发证书待存放位置
# 注意修改 "10.0.4.*" 为自己环境地址
# 这里保留 "XXX",后面部署命令会统一替换
cat > /home/deploy/node/cfg/kubelet <<< '
KUBELET_OPTS=" \
--logtostderr=true \
--v=4 \
--hostname-override=10.0.4.XXX \
--kubeconfig=/opt/kubernetes/cfg/kubelet.kubeconfig \
--bootstrap-kubeconfig=/opt/kubernetes/cfg/bootstrap.kubeconfig \
--config=/opt/kubernetes/cfg/kubelet.config \
--cert-dir=/opt/kubernetes/ssl \
--pod-infra-container-image=registry.cn-hangzhou.aliyuncs.com/google-containers/pause-amd64:3.0 \
"
'
# registry.cn-hangzhou.aliyuncs.com/google-containers/pause-amd64:3.0 该镜像可以提前导入本地局域网中的私有 docker 仓库中
```
- 创建 kubelet.config 配置文件
```bash
# 这里的 clusterDNS 是 DNS service 的集群地址,这里分配 10.10.9.2
# 注意修改 "10.0.4.*" 为自己环境地址
# 这里保留 "XXX",后面部署命令会统一替换
cat > /home/deploy/node/cfg/kubelet.config <<< '
kind: KubeletConfiguration
apiVersion: kubelet.config.k8s.io/v1beta1
address: 10.0.4.XXX
port: 10250
readOnlyPort: 10255
cgroupDriver: cgroupfs
clusterDNS: ["10.10.9.2"]
clusterDomain: cluster.local.
failSwapOn: false
authentication:
anonymous:
enabled: true
'
```
- 创建 kubelet.service
```bash
cat > /home/deploy/kubelet.service <<< '
[Unit]
Description=Kubernetes Kubelet
After=docker.service
Requires=docker.service
[Service]
EnvironmentFile=/opt/kubernetes/cfg/kubelet
ExecStart=/opt/kubernetes/bin/kubelet $KUBELET_OPTS
Restart=on-failure
KillMode=process
[Install]
WantedBy=multi-user.target
'
```
- 创建 kube-proxy 配置文件
```bash
# 这里的 cluster-cidr 是 pod ip 地址区间
# 注意修改 "10.0.4.*" 为自己环境地址
# 这里保留 "XXX",后面部署命令会统一替换
cat > /home/deploy/node/cfg/kube-proxy <<< '
KUBE_PROXY_OPTS=" \
--logtostderr=true \
--v=4 \
--hostname-override=10.0.4.XXX \
--cluster-cidr=172.16.0.0/16 \
--kubeconfig=/opt/kubernetes/cfg/kube-proxy.kubeconfig \
"
'
```
- 创建 kube-proxy.service
```bash
cat > /home/deploy/kube-proxy.service <<< '
[Unit]
Description=Kubernetes Proxy
After=network.target
[Service]
EnvironmentFile=-/opt/kubernetes/cfg/kube-proxy
ExecStart=/opt/kubernetes/bin/kube-proxy $KUBE_PROXY_OPTS
Restart=on-failure
[Install]
WantedBy=multi-user.target
'
```
- 部署到远程三个 k8s node 节点(node124、node125、node126)
```bash
# 注意修改 "10.0.4.*" 为自己环境地址
cd /home/deploy
for id in $(seq 124 126); do
ip="10.0.4.$id"
scp -r node/* $ip:/opt/kubernetes/
ssh $ip "sed -i 's/XXX/$id/g' /opt/kubernetes/cfg/kubelet"
ssh $ip "sed -i 's/XXX/$id/g' /opt/kubernetes/cfg/kubelet.config"
ssh $ip "sed -i 's/XXX/$id/g' /opt/kubernetes/cfg/kube-proxy"
scp kubelet.service kube-proxy.service $ip:/usr/lib/systemd/system/
systemctl -H $ip daemon-reload
systemctl -H $ip enable kubelet kube-proxy
done
```
- 启动三个 k8s node 节点的 kubelet 和 kube-proxy 服务
```bash
# 注意修改 "10.0.4.*" 为自己环境地址
for ip in $(seq -f'10.0.4.%g' 124 126); do
systemctl -H $ip start kubelet
systemctl -H $ip start kube-proxy
done
```
- 审批 node 加入集群
```bash
kubectl get csr
# "XXXX" 上一命令输出的 NAME 列
kubectl certificate approve XXXX
kubectl get node
```
- 查看集群状态
```bash
kubectl get node
kubectl get cs
```
# 部署其他组件
- [coreDNS](https://colben.cn/post/k8s-coredns/)
# 运行一个测试示例
- 暂无