21 Jan 2018
不参照[CentOS]和[Using kubeadm to Create a Cluster]的原因是,前者已经废弃,后者在beta阶段。另外此文档只是一个大纲,这样能够更深入的了解kubernetes的组件和原理。
文档中有很多细节,实际操作之外的步骤大部分忽略掉了,推荐详读一遍文档。
| items | version | comment |
|---|---|---|
| OS | centos7 | |
| kubernetes | 1.9.1 | 最新稳定版本 |
| docker | 17.09.0-ce | |
| etcd | 3.0.7 | |
| flannel | 使用flannel做overlay网络,支持不同主机间pods间网络互通 |
- docker(或者rkt)是必备的,因为kubernetes本身就是一个容器的编排工具
- etcd给kubernetes和flannel提供数据存储支持,可部署在kubernetes master节点上,也可以单独启用一个集群
- flannel给kubernetes提供了overlay网络支持(可选,也有其他选择,详细见文章开头的文档链接中的描述),实现了不同主机pods之间的直接互通
- kubernetes包含以下组件
- 在master节点上运行的kube-apiserver,kube-controller-manager,kube-scheduler
- 在node节点上运行的kubelet,kube-proxy
| hostname | ip address | service | comment |
|---|---|---|---|
| master | 172.16.1.100 | etcd,kube-apiserver,kube-controller-manager,kube-scheduler,docker | 主节点 |
| node01 | 172.16.1.101 | flannel,docker,kubelet,kube-proxy | node 1 |
| node02 | 172.16.1.102 | flannel,docker,kubelet,kube-proxy | node 2 |
| node03 | 172.16.1.103 | flannel,docker,kubelet,kube-proxy | node 3 |
为了将系统环境和软件环境对安装的影响度降低,需要确保以下几项需求满足
安装必要的工具包
yum install -y wget vim iptables iptables-services
关闭selinux
sed -i "s/SELINUX=enforcing/SELINUX=disabled/g" /etc/selinux/config setenforce 0
systemctl stop firewalld;systemctl stop iptablesbash
echo "172.16.1.100 master
172.16.1.101 node01
172.16.1.102 node02
172.16.1.103 node03" >> /etc/hosts
设定sysctl中的net.ipv4.ip_forward = 1
echo "net.ipv4.ip_forward = 1" >> /etc/sysctl.conf sysctl -p
net.ipv4.ip_forward = 1的配置确保了可以通过映射docker容器端口到外网,否则我们无法通过外网ip访问容器
- 关闭系统swap
bash swapoff -a
注释swap的开机挂载项,修改/etc/fstab
#/dev/mapper/VolGroup00-LogVol01 swap swap defaults 0 0
关闭系统swap,是为了严格的按照cpu和内存的限制,这样scheduler在规划pod的时候就不会把pod放进swap中了,这是为了性能考虑。
echo 'export MASTER_IP=172.16.1.100 export SERVICE_CLUSTER_IP_RANGE=10.254.0.0/16 export CLUSTER_NAME=KubeTest export PATH=$PATH:/usr/local/kubernetes/bin' > /etc/profile.d/kubernetes.sh source /etc/profile.d/kubernetes.sh
规划集群中需要重复使用的内容为变量
MASTER_IP- master的静态ipSERVICE_CLUSTER_IP_RANGE- service对象使用的ip范围CLUSTER_NAME- kubernetes集群的名称
kubernetes的二进制包里面包含了kubernetes的二进制文件和支持的etcd版本
# 下载kubernetes wget https://dl.k8s.io/v1.9.1/kubernetes-server-linux-amd64.tar.gz tar zxvf kubernetes-server-linux-amd64.tar.gz # 拷贝二进制文件到server端 mkdir -p /usr/local/kubernetes/{bin,security,conf} cp kubernetes/server/bin/{kube-apiserver,kube-scheduler,kube-controller-manager,kubectl} /usr/local/kubernetes/bin/ chmod 750 /usr/local/kubernetes/bin/* # 如果使用docker启动kube-apiserver,kube-scheduler,kube-controller-manager这三个服务的话,不需要拷贝它们的二进制文件,只需要拷贝kubectl即可 # 拷贝二进制文件到node端(提前做好ssh信任) scp kubernetes/server/bin/{kubelet,kube-proxy} root@node01:/usr/local/bin scp kubernetes/server/bin/{kubelet,kube-proxy} root@node02:/usr/local/bin scp kubernetes/server/bin/{kubelet,kube-proxy} root@node03:/usr/local/bin
因为kubernetes这个项目是使用go语言编写,而go语言程序的部署方式很简单,就是拷贝二进制文件就可以,所以在这里,我们通过简单的复制各服务的二进制文件,就可以通过启动它们来启动相应的服务。
本文开头的参照文档中说:
node需要运行的kubelet,kube-proxy,docker,推荐直接在系统层面上启动服务;
而对于etcd, kube-apiserver, kube-controller-manager 和 kube-scheduler,推荐我们使用容器来运行它们,文档中给出了几种镜像的获取方式,当然,我们下载的二进制文件中也有这样的镜像文件(bin目录中tar结尾的文件)可以本地加载(使用docker load命令)镜像到本机的docker中。
生成的 CA 证书和秘钥文件如下:
使用证书的组件如下:
安装cfssl
curl -s -L -o /usr/local/bin/cfssl https://pkg.cfssl.org/R1.2/cfssl_linux-amd64 curl -s -L -o /usr/local/bin/cfssljson https://pkg.cfssl.org/R1.2/cfssljson_linux-amd64 curl -s -L -o /usr/local/bin/cfssl-certinfo https://pkg.cfssl.org/R1.2/cfssl-certinfo_linux-amd64 chmod +x /usr/local/bin/* export PATH=$PATH:/usr/local/bin
创建k8s-ssl目录
mkdir ~/k8s-ssl
cd ~/k8s-ssl
此目录只是临时存放ca生成文件,可随意更换位置
创建 CA 证书签名请求
cat > ca-csr.json << EOF { "CN": "kubernetes", "key": { "algo": "rsa", "size": 2048 }, "names": [ { "C": "CN", "ST": "BeiJing", "L": "BeiJing", "O": "k8s", "OU": "System" } ] } EOF
字段说明
“CN”:Common Name,kube-apiserver 从证书中提取该字段作为请求的用户名 (User Name);浏览器使用该字段验证网站是否合法;
“O”:Organization,kube-apiserver 从证书中提取该字段作为请求用户所属的组 (Group);
生成 CA 证书和私钥
cfssl gencert -initca ca-csr.json | cfssljson -bare ca
ls ca*
ca-config.json ca-csr.json ca-key.pem ca.csr ca.pem
创建CA配置文件
cat > ca-config.json <<EOF { "signing": { "default": { "expiry": "87600h" }, "profiles": { "kubernetes": { "usages": [ "signing", "key encipherment", "server auth", "client auth" ], "expiry": "87600h" } } } } EOF
字段说明
ca-config.json:可以定义多个 profiles,分别指定不同的过期时间、使用场景等参数;后续在签名证书时使用某个 profile;
signing:表示该证书可用于签名其它证书;生成的 ca.pem 证书中 CA=TRUE;
server auth:表示client可以用该 CA 对server提供的证书进行验证;
client auth:表示server可以用该CA对client提供的证书进行验证;
创建 server 证书
cat > server-csr.json << EOF { "CN": "kubernetes", "hosts": [ "127.0.0.1", "172.16.1.100", "172.16.1.101", "172.16.1.102", "172.16.1.103", "10.254.0.1", "kubernetes", "kubernetes.default", "kubernetes.default.svc", "kubernetes.default.svc.cluster", "kubernetes.default.svc.cluster.local" ], "key": { "algo": "rsa", "size": 2048 }, "names": [ { "C": "CN", "ST": "BeiJing", "L": "BeiJing", "O": "k8s", "OU": "System" } ] } EOF
如果 hosts 字段不为空则需要指定授权使用该证书的 IP 或域名列表,由于该证书后续被 etcd 集群和 kubernetes master 集群使用,所以上面分别指定了 etcd 集群、kubernetes master 集群的主机 IP 和 kubernetes 服务的服务 IP(一般是 kube-apiserver 指定的 service-cluster-ip-range 网段的第一个IP,如 10.254.0.1),另外还有kubenetes的集群节点ip注。以上物理节点的IP也可以更换为主机名。
生成 kubernetes 证书和私钥
cfssl gencert -ca=ca.pem -ca-key=ca-key.pem -config=ca-config.json -profile=kubernetes server-csr.json | cfssljson -bare server ls server* server-csr.json server-key.pem server.csr server.pem
创建 admin 证书
cat > admin-csr.json << EOF { "CN": "admin", "hosts": [], "key": { "algo": "rsa", "size": 2048 }, "names": [ { "C": "CN", "ST": "BeiJing", "L": "BeiJing", "O": "system:masters", "OU": "System" } ] } EOF
后续 kube-apiserver 使用 RBAC 对客户端(如 kubelet、kube-proxy、Pod)请求进行授权;
kube-apiserver 预定义了一些 RBAC 使用的 RoleBindings,如 cluster-admin 将 Group system:masters 与 Role cluster-admin 绑定,该 Role 授予了调用kube-apiserver 的所有 API的权限;
O 指定该证书的 Group 为 system:masters,kubelet 使用该证书访问 kube-apiserver 时 ,由于证书被 CA 签名,所以认证通过,同时由于证书用户组为经过预授权的 system:masters,所以被授予访问所有 API 的权限;
注意:这个admin 证书,是将来生成管理员用的kube config 配置文件用的,现在我们一般建议使用RBAC 来对kubernetes 进行角色权限控制, kubernetes 将证书中的CN 字段 作为User, O 字段作为 Group(具体参考 Kubernetes中的用户与身份认证授权中 X509 Client Certs 一段)。在搭建完 kubernetes 集群后,我们可以通过命令: kubectl get clusterrolebinding cluster-admin -o yaml ,查看到 clusterrolebinding cluster-admin 的 subjects 的 kind 是 Group,name 是 system:masters。 roleRef 对象是 ClusterRole cluster-admin。 意思是凡是 system:masters Group 的 user 或者 serviceAccount 都拥有 cluster-admin 的角色。 因此我们在使用 kubectl 命令时候,才拥有整个集群的管理权限。可以使用 kubectl get clusterrolebinding cluster-admin -o yaml 来查看。
生成 admin 证书和私钥:
cfssl gencert -ca=ca.pem -ca-key=ca-key.pem -config=ca-config.json -profile=kubernetes admin-csr.json | cfssljson -bare admin ls admin* admin-csr.json admin-key.pem admin.csr admin.pem
创建 kube-proxy 证书
cat > kube-proxy-csr.json << EOF { "CN": "system:kube-proxy", "hosts": [], "key": { "algo": "rsa", "size": 2048 }, "names": [ { "C": "CN", "ST": "BeiJing", "L": "BeiJing", "O": "k8s", "OU": "System" } ] } EOF
CN 指定该证书的 User 为 system:kube-proxy;
kube-apiserver 预定义的 RoleBinding cluster-admin 将User system:kube-proxy 与 Role system:node-proxier 绑定,该 Role 授予了调用 kube-apiserver Proxy 相关 API 的权限;
生成 kube-proxy 客户端证书和私钥
cfssl gencert -ca=ca.pem -ca-key=ca-key.pem -config=ca-config.json -profile=kubernetes kube-proxy-csr.json | cfssljson -bare kube-proxy ls kube-proxy* kube-proxy-csr.json kube-proxy-key.pem kube-proxy.csr kube-proxy.pem
校验证书
cfssl-certinfo -cert server.pem { "subject": { "common_name": "kubernetes", "country": "CN", "organization": "k8s", "organizational_unit": "System", "locality": "BeiJing", "province": "BeiJing", "names": [ "CN", "BeiJing", "BeiJing", "k8s", "System", "kubernetes" ] }, "issuer": { "common_name": "kubernetes", "country": "CN", "organization": "k8s", "organizational_unit": "System", "locality": "BeiJing", "province": "BeiJing", "names": [ "CN", "BeiJing", "BeiJing", "k8s", "System", "kubernetes" ] }, "serial_number": "424094439737352687284679460316316317005227399373", "sans": [ "kubernetes", "kubernetes.default", "kubernetes.default.svc", "kubernetes.default.svc.cluster", "kubernetes.default.svc.cluster.local", "127.0.0.1", "172.16.1.100", "172.16.1.101", "172.16.1.102", "172.16.1.103", "10.254.0.1" ], "not_before": "2018-01-20T14:21:00Z", "not_after": "2028-01-18T14:21:00Z", "sigalg": "SHA256WithRSA", "authority_key_id": "93:B8:B7:BB:D1:1B:8A:C4:BA:6F:C2:E3:5C:EA:1E:1:66:D6:3F:B4", "subject_key_id": "73:B4:8:D8:2A:E7:B2:5B:15:F1:26:DE:6C:26:1:F2:AB:CB:D1:5D", "pem": "-----BEGIN CERTIFICATE-----证书内容-----END CERTIFICATE-----\n" }
拷贝证书到指定位置
cp *.pem /usr/local/kubernetes/security for node in node01 node02 node03 do scp *.pem root@$node:/usr/local/kubernetes/security done
路径可以随意指定,只要上下文中一致即可
kubelet、kube-proxy 等 Node 机器上的进程与 Master 机器的 kube-apiserver 进程通信时需要认证和授权;
kubernetes 1.4 开始支持由 kube-apiserver 为客户端生成 TLS 证书的 TLS Bootstrapping 功能,这样就不需要为每个客户端生成证书了;该功能当前仅支持为 kubelet 生成证书;
以下操作只需要在master节点上执行,生成的*.kubeconfig文件可以直接拷贝到node节点的/etc/kubernetes目录下。
创建 TLS Bootstrapping Token
Token auth file
创建k8s-config目录
mkdir -p ~/k8s-config
cd ~/k8s-config
Token可以是任意的包涵128 bit的字符串,可以使用安全的随机数发生器生成。
export BOOTSTRAP_TOKEN=$(head -c 16 /dev/urandom | od -An -t x | tr -d ' ') cat > token.csv <<EOF ${BOOTSTRAP_TOKEN},kubelet-bootstrap,10001,"system:kubelet-bootstrap" EOF cat token.csv aa0ed3c0a...4910c61a9f1,kubelet-bootstrap,10001,"system:kubelet-bootstrap"
注意:在进行后续操作前请检查 token.csv 文件,确认其中的 ${BOOTSTRAP_TOKEN} 环境变量已经被真实的值替换。
BOOTSTRAP_TOKEN将被写入到 kube-apiserver 使用的 token.csv 文件和 kubelet 使用的 bootstrap.kubeconfig 文件,如果后续重新生成了 BOOTSTRAP_TOKEN,则需要:
- 更新 token.csv 文件,分发到所有机器 (master 和 node),分发到node节点上非必需;
- 重新生成 bootstrap.kubeconfig 文件,分发到所有 node 机器;
- 重启 kube-apiserver 和 kubelet 进程;
- 重新 approve kubelet 的 csr 请求;
cp token.csv /usr/local/kubernetes/
创建 kubelet bootstrapping kubeconfig 文件
cd /usr/local/kubernetes export KUBE_APISERVER="https://172.16.1.100:6443" # 设置集群参数 kubectl config set-cluster KubeTest \ --certificate-authority=/usr/local/kubernetes/security/ca.pem \ --embed-certs=true \ --server=${KUBE_APISERVER} \ --kubeconfig=bootstrap.kubeconfig cat bootstrap.kubeconfig apiVersion: v1 clusters: - cluster: certificate-authority-data: LS0t...LS0K server: https://172.16.1.100:6443 name: KubeTest contexts: [] current-context: "" kind: Config preferences: {} users: [] # 设置客户端认证参数 kubectl config set-credentials kubelet-bootstrap \ --token=${BOOTSTRAP_TOKEN} \ --kubeconfig=bootstrap.kubeconfig cat bootstrap.kubeconfig apiVersion: v1 clusters: - cluster: certificate-authority-data: LS0t...LS0K server: https://172.16.1.100:6443 name: KubeTest contexts: [] current-context: "" kind: Config preferences: {} users: - name: kubelet-bootstrap user: as-user-extra: {} token: aa0ed3c0acc4cfc30a2cd4910c61a9f1 # 设置上下文参数 kubectl config set-context default \ --cluster=KubeTest \ --user=kubelet-bootstrap \ --kubeconfig=bootstrap.kubeconfig cat bootstrap.kubeconfig apiVersion: v1 clusters: - cluster: certificate-authority-data: LS0t...LS0K server: https://172.16.1.100:6443 name: KubeTest contexts: - context: cluster: KubeTest user: kubelet-bootstrap name: default current-context: "" kind: Config preferences: {} users: - name: kubelet-bootstrap user: as-user-extra: {} token: aa0ed3c0acc4cfc30a2cd4910c61a9f1 # 设置默认上下文 kubectl config use-context default --kubeconfig=bootstrap.kubeconfig cat bootstrap.kubeconfig apiVersion: v1 clusters: - cluster: certificate-authority-data: LS0t...LS0K server: https://172.16.1.100:6443 name: KubeTest contexts: - context: cluster: KubeTest user: kubelet-bootstrap name: default current-context: default kind: Config preferences: {} users: - name: kubelet-bootstrap user: as-user-extra: {} token: aa0ed3c0acc4cfc30a2cd4910c61a9f1
--embed-certs为 true 时表示将 certificate-authority 证书写入到生成的 bootstrap.kubeconfig 文件中;
设置客户端认证参数时没有指定秘钥和证书,后续由 kube-apiserver 自动生成;
创建 kube-proxy kubeconfig 文件
export KUBE_APISERVER="https://172.16.1.100:6443" # 设置集群参数 kubectl config set-cluster KubeTest \ --certificate-authority=/usr/local/kubernetes/security/ca.pem \ --embed-certs=true \ --server=${KUBE_APISERVER} \ --kubeconfig=kube-proxy.kubeconfig cat kube-proxy.kubeconfig apiVersion: v1 clusters: - cluster: certificate-authority-data: LS0t...LS0K server: https://172.16.1.100:6443 name: KubeTest contexts: [] current-context: "" kind: Config preferences: {} users: [] # 设置客户端认证参数 kubectl config set-credentials kube-proxy \ --client-certificate=/usr/local/kubernetes/security/kube-proxy.pem \ --client-key=/usr/local/kubernetes/security/kube-proxy-key.pem \ --embed-certs=true \ --kubeconfig=kube-proxy.kubeconfig cat kube-proxy.kubeconfig apiVersion: v1 clusters: - cluster: certificate-authority-data: LS0t...LS0K server: https://172.16.1.100:6443 name: KubeTest contexts: [] current-context: "" kind: Config preferences: {} users: - name: kube-proxy user: as-user-extra: {} client-certificate-data: LS0t...LQo= client-key-data: LS0t...Cg== # 设置上下文参数 kubectl config set-context default \ --cluster=KubeTest \ --user=kube-proxy \ --kubeconfig=kube-proxy.kubeconfig cat kube-proxy.kubeconfig apiVersion: v1 clusters: - cluster: certificate-authority-data: LS0t...LS0K server: https://172.16.1.100:6443 name: KubeTest contexts: - context: cluster: KubeTest user: kube-proxy name: default current-context: "" kind: Config preferences: {} users: - name: kube-proxy user: as-user-extra: {} client-certificate-data: LS0t...LQo= client-key-data: LS0t...Cg== # 设置默认上下文 kubectl config use-context default --kubeconfig=kube-proxy.kubeconfig cat kube-proxy.kubeconfig apiVersion: v1 clusters: - cluster: certificate-authority-data: LS0t...LS0K server: https://172.16.1.100:6443 name: KubeTest contexts: - context: cluster: KubeTest user: kube-proxy name: default current-context: "default" kind: Config preferences: {} users: - name: kube-proxy user: as-user-extra: {} client-certificate-data: LS0t...LQo= client-key-data: LS0t...Cg==
设置集群参数和客户端认证参数时 –embed-certs 都为 true,这会将 certificate-authority、client-certificate 和 client-key 指向的证书文件内容写入到生成的 kube-proxy.kubeconfig 文件中;
kube-proxy.pem 证书中 CN 为 system:kube-proxy,kube-apiserver 预定义的 RoleBinding cluster-admin 将User system:kube-proxy 与 Role system:node-proxier 绑定,该 Role 授予了调用 kube-apiserver Proxy 相关 API 的权限;
分发 kubeconfig 文件 将两个 kubeconfig 文件分发到所有 Node 机器
for node in node01 node02 node03 do scp bootstrap.kubeconfig kube-proxy.kubeconfig root@$node:/usr/local/kubernetes/conf done
在kubernetes/cluster/images/etcd/Makefile中查找到对应的etcd版本
# step 1 下载安装etcd ETCD_VER=v3.0.17 DOWNLOAD_URL=https://github.com/coreos/etcd/releases/download curl -L ${DOWNLOAD_URL}/${ETCD_VER}/etcd-${ETCD_VER}-linux-amd64.tar.gz -o etcd-${ETCD_VER}-linux-amd64.tar.gz tar xzvf etcd-${ETCD_VER}-linux-amd64.tar.gz mv etcd-${ETCD_VER}-linux-amd64/etcd* /usr/local/bin # step 2 准备systemd unit文件 echo '[Unit] Description=etcd key-value store Documentation=https://github.com/coreos/etcd After=network.target [Service] User=etcd Type=notify Environment=ETCD_DATA_DIR=/var/lib/etcd Environment=ETCD_NAME=%m Environment=ETCD_LISTEN_CLIENT_URLS=https://172.16.1.100:2379,https://172.16.1.100:4001 Environment=ETCD_ADVERTISE_CLIENT_URLS=https://172.16.1.100:2379 Environment=ETCD_TRUSTED_CA_FILE=/usr/local/kubernetes/security/ca.pem Environment=ETCD_CERT_FILE=/usr/local/kubernetes/security/server.pem Environment=ETCD_KEY_FILE=/usr/local/kubernetes/security/server-key.pem ExecStart=/usr/local/bin/etcd Restart=always RestartSec=10s LimitNOFILE=40000 [Install] WantedBy=multi-user.target' > /usr/lib/systemd/system/etcd.service # step 3 准备etcd用户 useradd -r -s /sbin/nologin etcd # step 4 准备etcd数据目录 mkdir -p /var/lib/etcd chown -R etcd:etcd /var/lib/etcd # step 5 启动etcd systemctl daemon-reload systemctl enable etcd.service systemctl start etcd.service # 验证etcd etcdctl \ --endpoints=https://172.16.1.100:2379 \ --ca-file=/usr/local/kubernetes/security/ca.pem \ --cert-file=/usr/local/kubernetes/security/server.pem \ --key-file=/usr/local/kubernetes/security/server-key.pem \ cluster-health
使用etcd储存flannel的网络配置
etcdctl \ --endpoints https://172.16.1.100:2379 \ --ca-file=/usr/local/kubernetes/security/ca.pem \ --cert-file=/usr/local/kubernetes/security/server.pem \ --key-file=/usr/local/kubernetes/security/server-key.pem \ set /kube-centos/network/config '{ "Network": "10.5.0.0/16", "Backend": {"Type": "vxlan"}}'
为了测试,在主节点上只启动一个节点的etcd,etcd集群参照etcd 集群文档
后来启动etcd出错,是因为没有权限访问key文件,为了方便,临时给key加上777权限,生产环境时需要想办法解决
export KUBE_APISERVER="https://172.16.1.100:6443" # 设置集群参数 kubectl config set-cluster KubeTest \ --certificate-authority=/usr/local/kubernetes/security/ca.pem \ --embed-certs=true \ --server=${KUBE_APISERVER} # 设置客户端认证参数 kubectl config set-credentials admin \ --client-certificate=/usr/local/kubernetes/security/admin.pem \ --embed-certs=true \ --client-key=/usr/local/kubernetes/security/admin-key.pem # 设置上下文参数 kubectl config set-context kubernetes \ --cluster=kubernetes \ --user=admin # 设置默认上下文 kubectl config use-context kubernetes
注意:~/.kube/config文件拥有对该集群的最高权限,请妥善保管。
准备配置文件:
cat > /usr/local/kubernetes/conf/config << EOF ### # kubernetes system config # # The following values are used to configure various aspects of all # kubernetes services, including # # kube-apiserver.service # kube-controller-manager.service # kube-scheduler.service # kubelet.service # kube-proxy.service # logging to stderr means we get it in the systemd journal KUBE_LOGTOSTDERR="--logtostderr=true" # journal message level, 0 is debug KUBE_LOG_LEVEL="--v=0" # Should this cluster be allowed to run privileged docker containers KUBE_ALLOW_PRIV="--allow-privileged=true" # How the controller-manager, scheduler, and proxy find the apiserver KUBE_MASTER="--master=http://127.0.0.1:8080" EOF cat > /usr/local/kubernetes/conf/apiserver << EOF ### # kubernetes system config # # The following values are used to configure the kube-apiserver # # The address on the local server to listen to. KUBE_API_ADDRESS="--advertise-address=172.16.1.100 --bind-address=172.16.1.100 --insecure-bind-address=127.0.0.1" # The port on the local server to listen on. #KUBE_API_PORT="--insecure-port=8080" # Port minions listen on # KUBELET_PORT="--kubelet-port=10250" # Comma separated list of nodes in the etcd cluster KUBE_ETCD_SERVERS="--etcd-servers=https://172.16.1.100:2379,https://172.16.1.100:4001" # Address range to use for services KUBE_SERVICE_ADDRESSES="--service-cluster-ip-range=10.254.0.0/16" # default admission control policies KUBE_ADMISSION_CONTROL="--admission-control=NamespaceLifecycle,LimitRanger,SecurityContextDeny,ServiceAccount,ResourceQuota" # Add your own! KUBE_API_ARGS="--authorization-mode=RBAC --runtime-config=rbac.authorization.k8s.io/v1beta1 --kubelet-https=true --enable-bootstrap-token-auth --token-auth-file=/usr/local/kubernetes/token.csv --service-node-port-range=30000-32767 --tls-cert-file=/usr/local/kubernetes/security/server.pem --tls-private-key-file=/usr/local/kubernetes/security/server-key.pem --client-ca-file=/usr/local/kubernetes/security/ca.pem --service-account-key-file=/usr/local/kubernetes/security/ca-key.pem --etcd-cafile=/usr/local/kubernetes/security/ca.pem --etcd-certfile=/usr/local/kubernetes/security/server.pem --etcd-keyfile=/usr/local/kubernetes/security/server-key.pem --enable-swagger-ui=true --apiserver-count=1 --audit-log-maxage=30 --audit-log-maxbackup=3 --audit-log-maxsize=100 --audit-log-path=/var/lib/audit.log --event-ttl=1h --storage-backend=etcd2" EOF cat > /usr/local/kubernetes/conf/controller-manager << EOF ### # The following values are used to configure the kubernetes controller-manager # defaults from config and apiserver should be adequate # Add your own! KUBE_CONTROLLER_MANAGER_ARGS="--address=127.0.0.1 --service-cluster-ip-range=10.254.0.0/16 --cluster-name=KubeTest --cluster-signing-cert-file=/usr/local/kubernetes/security/ca.pem --cluster-signing-key-file=/usr/local/kubernetes/security/ca-key.pem --service-account-private-key-file=/usr/local/kubernetes/security/ca-key.pem --root-ca-file=/usr/local/kubernetes/security/ca.pem --leader-elect=true" EOF cat > /usr/local/kubernetes/conf/scheduler << EOF ### # kubernetes scheduler config # default config should be adequate # Add your own! KUBE_SCHEDULER_ARGS="--leader-elect=true --address=127.0.0.1" EOF
--storage-backend=etcd2,增加此option,是因为kube-api-server在增加了key认证之后,和etcd的交互有问题。git issure 讲解
准备systemd unit文件:
echo '[Unit] Description=Kubernetes API Server Documentation=https://github.com/GoogleCloudPlatform/kubernetes After=network.target After=etcd.service [Service] EnvironmentFile=-/usr/local/kubernetes/conf/config EnvironmentFile=-/usr/local/kubernetes/conf/apiserver User=kube ExecStart=/usr/local/kubernetes/bin/kube-apiserver \ $KUBE_LOGTOSTDERR \ $KUBE_LOG_LEVEL \ $KUBE_ETCD_SERVERS \ $KUBE_API_ADDRESS \ $KUBE_API_PORT \ $KUBELET_PORT \ $KUBE_ALLOW_PRIV \ $KUBE_SERVICE_ADDRESSES \ $KUBE_ADMISSION_CONTROL \ $KUBE_API_ARGS Restart=on-failure Type=notify LimitNOFILE=65536 [Install] WantedBy=multi-user.target' > /usr/lib/systemd/system/kube-apiserver.service mkdir /usr/lib/systemd/system/kube-apiserver.service.d echo '[Service] PermissionsStartOnly=yes ExecStartPre=/usr/bin/mkdir -p /var/run/kubernetes ExecStartPre=/usr/bin/chown kube.kube /var/run/kubernetes' > /usr/lib/systemd/system/kube-apiserver.service.d/pre-start.conf echo '[Unit] Description=Kubernetes Controller Manager Documentation=https://github.com/GoogleCloudPlatform/kubernetes [Service] EnvironmentFile=-/usr/local/kubernetes/conf/config EnvironmentFile=-/usr/local/kubernetes/conf/controller-manager User=kube ExecStart=/usr/local/kubernetes/bin/kube-controller-manager \ $KUBE_LOGTOSTDERR \ $KUBE_LOG_LEVEL \ $KUBE_MASTER \ $KUBE_CONTROLLER_MANAGER_ARGS Restart=on-failure LimitNOFILE=65536 [Install] WantedBy=multi-user.target' > /usr/lib/systemd/system/kube-controller-manager.service echo '[Unit] Description=Kubernetes Scheduler Plugin Documentation=https://github.com/GoogleCloudPlatform/kubernetes [Service] EnvironmentFile=-/usr/local/kubernetes/conf/config EnvironmentFile=-/usr/local/kubernetes/conf/scheduler User=kube ExecStart=/usr/local/kubernetes/bin/kube-scheduler \ $KUBE_LOGTOSTDERR \ $KUBE_LOG_LEVEL \ $KUBE_MASTER \ $KUBE_SCHEDULER_ARGS Restart=on-failure LimitNOFILE=65536 [Install] WantedBy=multi-user.target' > /usr/lib/systemd/system/kube-scheduler.service
依次启动kube-apiserver.service, kube-controller-manager.service, kube-scheduler.service
# 重载systemd unit文件 systemctl daemon-reload # 创建spawn服务的用户kube(在配置文件中配置) useradd -r -s /sbin/nologin kube chown :kube /usr/local/kubernetes/bin/* systemctl enable kube-apiserver.service systemctl enable kube-controller-manager.service systemctl enable kube-scheduler.service systemctl start kube-apiserver.service systemctl start kube-controller-manager.service systemctl start kube-scheduler.service # 检查集群状态 kubectl get componentstatuses NAME STATUS MESSAGE ERROR scheduler Healthy ok controller-manager Healthy ok etcd-1 Healthy {"health": "true"} etcd-0 Healthy {"health": "true"}
# 下载flannel FLANNEL_VER=v0.9.1 wget https://github.com/coreos/flannel/releases/download/${FLANNEL_VER}/flannel-${FLANNEL_VER}-linux-amd64.tar.gz mkdir flannel tar zxvf flannel-${FLANNEL_VER}-linux-amd64.tar.gz -C flannel cp flannel/flanneld /usr/local/bin mkdir -p /usr/libexec/flannel cp flannel/mk-docker-opts.sh /usr/libexec/flannel/ # 准备flannel配置文件 ## !!重点!! ## # -iface,根据实际情况设定 # FLANNELD_PUBLIC_IP,每个节点不同 ############# cat > /etc/sysconfig/flanneld << EOF FLANNELD_PUBLIC_IP="172.16.1.101" FLANNELD_ETCD_ENDPOINTS="https://172.16.1.100:2379" FLANNELD_ETCD_PREFIX="/kube-centos/network" # Any additional options that you want to pass FLANNELD_OPTIONS="-iface=eth1 -etcd-cafile=/usr/local/kubernetes/security/ca.pem -etcd-certfile=/usr/local/kubernetes/security/server.pem -etcd-keyfile=/usr/local/kubernetes/security/server-key.pem" EOF # 准备flannel systemd unit文件 echo '[Unit] Description=Flanneld overlay address etcd agent After=network.target After=network-online.target Wants=network-online.target Before=docker.service [Service] Type=notify EnvironmentFile=/etc/sysconfig/flanneld ExecStart=/usr/local/bin/flanneld $FLANNELD_OPTIONS ExecStartPost=/usr/libexec/flannel/mk-docker-opts.sh -c Restart=on-failure [Install] WantedBy=multi-user.target RequiredBy=docker.service' > /usr/lib/systemd/system/flannel.service systemctl daemon-reload systemctl enable flannel systemctl start flannel
每个节点的flannel需要根据自己情况来填写配置文件
flannel启动后生成了以下文件:
- /var/run/flannel/subnet.env, 从etcd中获取信息然后生成的flanneld配置文件
- /run/docker_opts.env, flannel service文件中指定的/usr/libexec/flannel/mk-docker-opts.sh生成的docker环境变量文件
# 安装docker底包 yum install -y git libcgroup libcgroup-tools systemctl enable cgconfig systemctl start cgconfig # 下载安装docker DOCKER_VER=17.09.0 wget https://download.docker.com/linux/static/stable/x86_64/docker-${DOCKER_VER}-ce.tgz tar zxvf docker-${DOCKER_VER}-ce.tgz cp docker/* /usr/local/bin/ wget https://github.com/docker/compose/releases/download/1.17.1/docker-compose-Linux-x86_64 cp docker-compose-Linux-x86_64 /usr/local/bin/docker-compose chmod 755 /usr/local/bin/* # 准备systemd unit文件 echo '[Unit] Description=Docker Application Container Engine Documentation=https://docs.docker.com After=network-online.target docker.socket flannel.service Wants=network-online.target Requires=docker.socket [Service] Type=notify # the default is not to use systemd for cgroups because the delegate issues still # exists and systemd currently does not support the cgroup feature set required # for containers run by docker EnvironmentFile=/run/docker_opts.env ExecStart=/usr/local/bin/dockerd -H fd:// $DOCKER_OPTS ExecReload=/bin/kill -s HUP $MAINPID LimitNOFILE=1048576 # Having non-zero Limit*s causes performance problems due to accounting overhead # in the kernel. We recommend using cgroups to do container-local accounting. LimitNPROC=infinity LimitCORE=infinity # Uncomment TasksMax if your systemd version supports it. # Only systemd 226 and above support this version. #TasksMax=infinity TimeoutStartSec=0 # set delegate yes so that systemd does not reset the cgroups of docker containers Delegate=yes # kill only the docker process, not all processes in the cgroup KillMode=process # restart the docker process if it exits prematurely Restart=on-failure StartLimitBurst=3 StartLimitInterval=60s [Install] WantedBy=multi-user.target' > /usr/lib/systemd/system/docker.service echo '[Unit] Description=Docker Socket for the API PartOf=docker.service [Socket] ListenStream=/var/run/docker.sock SocketMode=0660 SocketUser=root SocketGroup=docker [Install] WantedBy=sockets.target' > /usr/lib/systemd/system/docker.socket groupadd docker systemctl daemon-reload systemctl enable docker systemctl start docker
docker systemd
kubelet 启动时向 kube-apiserver 发送 TLS bootstrapping 请求,需要先将 bootstrap token 文件中的 kubelet-bootstrap 用户赋予 system:node-bootstrapper cluster 角色(role), 然后 kubelet 才能有权限创建认证请求(certificate signing requests):
# master节点执行 kubectl create clusterrolebinding kubelet-bootstrap \ --clusterrole=system:node-bootstrapper \ --user=kubelet-bootstrap
–user=kubelet-bootstrap 是在 /etc/kubernetes/token.csv 文件中指定的用户名,同时也写入了 /etc/kubernetes/bootstrap.kubeconfig 文件;
准备配置文件:
mkdir /usr/local/kubernetes/conf -p cat > /usr/local/kubernetes/conf/config << EOF ### # kubernetes system config # # The following values are used to configure various aspects of all # kubernetes services, including # # kube-apiserver.service # kube-controller-manager.service # kube-scheduler.service # kubelet.service # kube-proxy.service # logging to stderr means we get it in the systemd journal KUBE_LOGTOSTDERR="--logtostderr=true" # journal message level, 0 is debug KUBE_LOG_LEVEL="--v=0" # Should this cluster be allowed to run privileged docker containers KUBE_ALLOW_PRIV="--allow-privileged=false" # How the controller-manager, scheduler, and proxy find the apiserver KUBE_MASTER="--master=https://172.16.1.100:6443" EOF cat > /usr/local/kubernetes/conf/kubelet << EOF ### # kubernetes kubelet (minion) config # --kubeconfig for kubeconfig KUBELET_KUBECONFIG="--kubeconfig=/usr/local/kubernetes/conf/kubelet.kubeconfig" # The address for the info server to serve on (set to 0.0.0.0 or "" for all interfaces) KUBELET_ADDRESS="--address=0.0.0.0" # The port for the info server to serve on # KUBELET_PORT="--port=10250" # You may leave this blank to use the actual hostname KUBELET_HOSTNAME="--hostname-override=" # Add your own! KUBELET_ARGS="--cluster-dns=10.254.0.2 --bootstrap-kubeconfig=/usr/local/kubernetes/conf/bootstrap.kubeconfig --cert-dir=/usr/local/kubernetes/security --cluster-domain=cluster.local --hairpin-mode promiscuous-bridge --serialize-image-pulls=false" EOF cat > /usr/local/kubernetes/conf/proxy << EOF ### # kubernetes proxy config # default config should be adequate # Add your own! KUBE_PROXY_ARGS="--bind-address=172.16.1.101 --hostname-override= --kubeconfig=/usr/local/kubernetes/conf/kube-proxy.kubeconfig --cluster-cidr=10.254.0.0/16" EOF
准备systemd unit文件:
echo '[Unit] Description=Kubernetes Kubelet Server Documentation=https://github.com/GoogleCloudPlatform/kubernetes After=docker.service Requires=docker.service [Service] WorkingDirectory=/var/lib/kubelet EnvironmentFile=-/usr/local/kubernetes/conf/config EnvironmentFile=-/usr/local/kubernetes/conf/kubelet ExecStart=/usr/local/bin/kubelet \ $KUBE_LOGTOSTDERR \ $KUBE_LOG_LEVEL \ $KUBELET_KUBECONFIG \ $KUBELET_ADDRESS \ $KUBELET_PORT \ $KUBELET_HOSTNAME \ $KUBE_ALLOW_PRIV \ $KUBELET_ARGS Restart=on-failure KillMode=process [Install] WantedBy=multi-user.target' > /usr/lib/systemd/system/kubelet.service echo '[Unit] Description=Kubernetes Kube-Proxy Server Documentation=https://github.com/GoogleCloudPlatform/kubernetes After=network.target [Service] EnvironmentFile=-/usr/local/kubernetes/conf/config EnvironmentFile=-/usr/local/kubernetes/conf/proxy ExecStart=/usr/local/bin/kube-proxy \ $KUBE_LOGTOSTDERR \ $KUBE_LOG_LEVEL \ $KUBE_MASTER \ $KUBE_PROXY_ARGS Restart=on-failure LimitNOFILE=65536 [Install] WantedBy=multi-user.target' > /usr/lib/systemd/system/kube-proxy.service
依次启动kubelet,kube-proxy服务
# 重载systemd units文件 systemctl daemon-reload # 创建kubelet工作目录 mkdir /var/lib/kubelet # 启动服务 systemctl enable kubelet systemctl enable kube-proxy systemctl start kubelet systemctl start kube-proxy
kubectl get csr | awk '/Pending/ {print $1}' | xargs -i echo {} node-csr-IVYfMcoBJGmTyLVceabnuoDRpFbCxWiisev6pQ0ynYA node-csr-TGGZ3Gy4dxaApbOMSPGCsaZy4xKeuwu8Iuc8RWxYkro node-csr-TIZ6m0T5hZD6rgZBLCxzgdwp_O-p4DGvSTAZX3dQ3YU kubectl get csr | awk '/Pending/ {print $1}' | xargs -i kubectl certificate approve {} certificatesigningrequest "node-csr-IVYfMcoBJGmTyLVceabnuoDRpFbCxWiisev6pQ0ynYA" approved certificatesigningrequest "node-csr-TGGZ3Gy4dxaApbOMSPGCsaZy4xKeuwu8Iuc8RWxYkro" approved certificatesigningrequest "node-csr-TIZ6m0T5hZD6rgZBLCxzgdwp_O-p4DGvSTAZX3dQ3YU" approved