k8s二 | 使用kubeadm部署K8S

上篇文章主要介紹了k8s(一) 基本概念與組件原理,下面我們來使用Kubeadm快速的部署一個(gè) Kubernetes 集群,來理解 Kubernetes 組件的工作方式和架構(gòu)。

一. kubeadm介紹

kubeadm是Kubernetes官方提供的用于快速安裝 Kubernetes 集群的工具,它提供了 kubeadm init以及 kubeadm join 這兩個(gè)命令作為快速創(chuàng)建 kubernetes 集群的最佳實(shí)踐,只需將kubeadm,kubelet,kubectl安裝到服務(wù)器,其他核心組件以容器化方式快速部署,不過目前kubeadm還處于beta狀態(tài),還不能用于生產(chǎn)環(huán)境。

二. 環(huán)境準(zhǔn)備

系統(tǒng)版本: Centos7.2
內(nèi)核版本: 3.10.0
k8s 版本: 1.15.3
Docker版本:19.03

節(jié)點(diǎn)添加hosts信息

$ cat <<EOF >> /etc/hosts
172.16.1.100  k8s-master 
172.16.1.101  k8s-node01  
EOF

禁用Firewalld,Selinux,Swap

$ systemctl stop firewalld 
$ systemctl disable firewalld 
$ setenforce 0
$ sed -i "s/enforcing/disabled/g"  /etc/selinux/config
$ swapoff -a 

修改內(nèi)核參數(shù)

$ cat << EOF > /etc/sysctl.d/k8s.conf
net.bridge.bridge-nf-call-ip6tables = 1
net.bridge.bridge-nf-call-iptables = 1
net.ipv4.ip_forward = 1
vm.swappiness=0
EOF
$ modprobe br_netfilter #報(bào)錯(cuò)使用yum -y update 更新內(nèi)核模塊
$ sysctl -p /etc/sysctl.d/k8s.conf

安裝ipvs(負(fù)載均衡器,kube-proxy使用ipvs模式)

$ cat > /etc/sysconfig/modules/ipvs.modules <<EOF
#!/bin/bash
modprobe -- ip_vs
modprobe -- ip_vs_rr
modprobe -- ip_vs_wrr
modprobe -- ip_vs_sh
modprobe -- nf_conntrack_ipv4
EOF
$ chmod 755 /etc/sysconfig/modules/ipvs.modules && bash /etc/sysconfig/modules/ipvs.modules && lsmod | grep -e ip_vs -e nf_conntrack_ipv4

安裝ipset (iptables的擴(kuò)展)

$ yum install ipvsadm ipset

同步服務(wù)器時(shí)間

$ yum -y install  ntp
$ ntpdate  ntp1.aliyun.com

三. 安裝docker

添加docker源并安裝

$ yum install -y yum-utils \
  device-mapper-persistent-data \
  lvm2
$ yum-config-manager \
    --add-repo \
    https://download.docker.com/linux/centos/docker-ce.repo
$ yum list docker-ce --showduplicates | sort -r
$ yum install docker-ce-19.03.1-3.el7

配置 Docker 鏡像加速

$ mkdir /etc/docker
$ cat  << EOF  > /etc/docker/daemon.json
{
  "exec-opts": ["native.cgroupdriver=systemd"],
  "registry-mirrors" : [
    "https://ot2k4d59.mirror.aliyuncs.com/"
  ]
}
EOF

啟動(dòng)docker并設(shè)置開機(jī)自啟

$ systemctl start docker
$ systemctl enable docker

四. 安裝Kubeadm

添加鏡像源

$ cat <<EOF > /etc/yum.repos.d/kubernetes.repo
[kubernetes]
name=Kubernetes
baseurl=http://mirrors.aliyun.com/kubernetes/yum/repos/kubernetes-el7-x86_64
enabled=1
gpgcheck=0
repo_gpgcheck=0
gpgkey=http://mirrors.aliyun.com/kubernetes/yum/doc/yum-key.gpg
        http://mirrors.aliyun.com/kubernetes/yum/doc/rpm-package-key.gpg
EOF

安裝 kubeadm、kubelet、kubectl

$ yum -y install kubectl-1.15.3-0 kubeadm-1.15.3-0  kubelet-1.15.3-0
$ kubeadm version  #查看版本
$ kubeadm version: &version.Info{Major:"1", Minor:"15", GitVersion:"v1.15.3", GitCommit:"2d3c76f9091b6bec110a5e63777c332469e0cba2", GitTreeState:"clean", BuildDate:"2019-08-19T11:11:18Z", GoVersion:"go1.12.9", Compiler:"gc", Platform:"linux/amd64"}
$ systemctl enable kubelet.service # 設(shè)置開機(jī)自啟

五. 初始化集群

kubeadm初始化工作流程

首先我們可以使用kubeadm init命令來進(jìn)行初始化工作,其中kubeadm 首先要做的,是一系列的檢查工作,以確定這臺(tái)機(jī)器可以用來部署 Kubernetes,比如檢查內(nèi)核版本是否是3.10以上,Cgroups 模塊是否可用,Docker是否正確安裝等,然后以Pod的形式來部署kube-apiserver、kube-controller-manager、kube-scheduler這些組件,最后則是部署kube-proxyDNS這些插件。

如果我們需要使用一些自定義的配置,在Master節(jié)點(diǎn)可以導(dǎo)出默認(rèn)的初始化文件進(jìn)行修改。

$ kubeadm config print init-defaults > kubeadm.yaml

根據(jù)自己需求修改默認(rèn)配置

apiVersion: kubeadm.k8s.io/v1beta2
bootstrapTokens:
- groups:
  - system:bootstrappers:kubeadm:default-node-token
  token: abcdef.0123456789abcdef
  ttl: 24h0m0s
  usages:
  - signing
  - authentication
kind: InitConfiguration
localAPIEndpoint:
  advertiseAddress: 172.16.1.100   #修改apiserverIP
  bindPort: 6443
nodeRegistration:
  criSocket: /var/run/dockershim.sock
  name: kubesphere
  taints:
  - effect: NoSchedule
    key: node-role.kubernetes.io/master
---
apiServer:
  timeoutForControlPlane: 4m0s
apiVersion: kubeadm.k8s.io/v1beta2
certificatesDir: /etc/kubernetes/pki
clusterName: kubernetes
controllerManager: {}
dns:
  type: CoreDNS
etcd:
  local:
    dataDir: /var/lib/etcd
imageRepository: gcr.azk8s.cn/google_containers #修改鏡像倉庫地址
kind: ClusterConfiguration
kubernetesVersion: v1.15.3
networking:
  dnsDomain: cluster.local
  podSubnet: 192.168.0.0/16  #Pod的IP網(wǎng)段,后面使用calico插件
  serviceSubnet: 10.96.0.0/12
scheduler: {}
---  #添加以下,修改kube-proxy模式
apiVersion: kubeproxy.config.k8s.io/v1alpha1 
kind: KubeProxyConfiguration
mode: ipvs  

初始化集群

$ kubeadm init --config kubeadm.yaml
------
[addons] Applied essential addon: CoreDNS
[addons] Applied essential addon: kube-proxy

Your Kubernetes control-plane has initialized successfully!

To start using your cluster, you need to run the following as a regular user:

  mkdir -p $HOME/.kube
  sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
  sudo chown $(id -u):$(id -g) $HOME/.kube/config

You should now deploy a pod network to the cluster.
Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at:
  https://kubernetes.io/docs/concepts/cluster-administration/addons/

Then you can join any number of worker nodes by running the following on each as root:
kubeadm join 172.16.1.100:6443 --token szu5t8.z6m03rxaamo8jzy1     --discovery-token-ca-cert-hash sha256:0455a39d0ff4cca1a9c947fa902ac635c09da5b4d7a30363e9376a9a2eb97a24

拷貝 kubeconfig 文件

$ mkdir -p $HOME/.kube
$ sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config
$ sudo chown $(id -u):$(id -g) $HOME/.kube/config

六. 節(jié)點(diǎn)加入集群

kubeadm join工作流程

在Master節(jié)點(diǎn)生成token后,然后在任意一臺(tái)安裝了 kubelet 和 kubeadm 的機(jī)器上執(zhí)行 kubeadm join命令 即可加入到kubernetes集群中。

$ kubeadm join 172.16.1.100:6443 --token szu5t8.z6m03rxaamo8jzy1     --discovery-token-ca-cert-hash sha256:0455a39d0ff4cca1a9c947fa902ac635c09da5b4d7a30363e9376a9a2eb97a24

上面的Key如果忘掉可以在master使用命令kubeadm token create --print-join-command重新獲取。

七. 安裝集群插件

1. 安裝calico網(wǎng)絡(luò)插件
$ wget https://docs.projectcalico.org/v3.8/manifests/calico.yaml
$ kubectl apply -f calico.yaml

查看Pod運(yùn)行狀態(tài)

$ kubectl get pods -n kube-system
NAME                                       READY   STATUS    RESTARTS   AGE
calico-kube-controllers-5df986d44c-hpqrr   1/1     Running   0          67m
calico-node-nvhfh                          1/1     Running   0          63m
calico-node-vgft9                          1/1     Running   0          63m
coredns-cf8fb6d7f-q5kw6                    1/1     Running   0          2d19h
coredns-cf8fb6d7f-z92hh                    1/1     Running   0          2d19h
etcd-kubesphere                            1/1     Running   0          2d19h
kube-apiserver-kubesphere                  1/1     Running   0          2d19h
kube-controller-manager-kubesphere         1/1     Running   0          2d19h
kube-proxy-68n9f                           1/1     Running   0          2d19h
kube-proxy-6ht99                           1/1     Running   0          73m
kube-scheduler-kubesphere                  1/1     Running   0          2d19h
tiller-deploy-74cd79795-p26l5              1/1     Running   0          173m

查看節(jié)點(diǎn)運(yùn)行狀態(tài)

$ kubectl  get nodes
NAME         STATUS   ROLES    AGE     VERSION
k8s-node01   Ready    <none>   74m     v1.15.3
k8s-master   Ready    master   2d19h   v1.15.3
2. 安裝Dashboard可視化插件
$ wget https://raw.githubusercontent.com/kubernetes/dashboard/v1.10.1/src/deploy/recommended/kubernetes-dashboard.yaml
$ vim  kubernetes-dashboard.yaml #修改鏡像名稱
......
containers:
- args:
  - --auto-generate-certificates
  image: gcr.azk8s.cn/google_containers/kubernetes-dashboard-amd64:v1.10.1 # 修改鏡像名稱
  imagePullPolicy: IfNotPresent
......
selector:
  k8s-app: kubernetes-dashboard
type: NodePort  # 修改Service為NodePort類型
......

創(chuàng)建服務(wù)

$ kubectl apply -f kubernetes-dashboard.yaml
$ kubectl get pods -n kube-system -l k8s-app=kubernetes-dashboard
NAME                                  READY   STATUS    RESTARTS   AGE
kubernetes-dashboard-fcfb4cbc-wr79d   1/1     Running   0          39s
$ kubectl get svc -n kube-system -l k8s-app=kubernetes-dashboard
NAME                   TYPE       CLUSTER-IP     EXTERNAL-IP   PORT(S)         AGE
kubernetes-dashboard   NodePort   10.96.168.30   <none>        443:31445/TCP   53s

創(chuàng)建一個(gè)具有所有權(quán)限的用戶來登錄Dashboard:

$ vim admin.yaml
kind: ClusterRoleBinding
apiVersion: rbac.authorization.k8s.io/v1beta1
metadata:
  name: admin
  annotations:
    rbac.authorization.kubernetes.io/autoupdate: "true"
roleRef:
  kind: ClusterRole
  name: cluster-admin
  apiGroup: rbac.authorization.k8s.io
subjects:
- kind: ServiceAccount
  name: admin
  namespace: kube-system

---
apiVersion: v1
kind: ServiceAccount
metadata:
  name: admin
  namespace: kube-system
  labels:
    kubernetes.io/cluster-service: "true"
    addonmanager.kubernetes.io/mode: Reconcile

創(chuàng)建用戶獲取token

$ kubectl apply -f admin.yaml
$ kubectl get secret -n kube-system|grep admin-token
admin-token-4fjvq                                kubernetes.io/service-account-token   3      58s
$ kubectl get secret admin-token-4fjvq  -o jsonpath={.data.token} -n kube-system |base64 -d 

使用火狐瀏覽器訪問Dashboard的NodePort端口
https://172.16.1.100:31445/

公眾號(hào):前行技術(shù)圈

總結(jié):在這篇文章中,我們使用kubeadm工具來快速部署了k8s集群,理解了集群工作方式和架構(gòu),但因?yàn)?kubeadm 目前還是不能一鍵部署高可用的 Kubernetes 集群,如:Etcd、Apiserver等組件都應(yīng)該是多節(jié)點(diǎn)集群,所以目前還是不能用于生產(chǎn)環(huán)境,關(guān)注公眾號(hào)回復(fù)k8s部署獲取二進(jìn)制生產(chǎn)高可用部署文檔。


上篇文章:k8s一 | 基本概念與組件原理
系列文章:深入理解Kuerneters
參考資料:從Docker到Kubernetes進(jìn)階-陽明

最后編輯于
?著作權(quán)歸作者所有,轉(zhuǎn)載或內(nèi)容合作請(qǐng)聯(lián)系作者
【社區(qū)內(nèi)容提示】社區(qū)部分內(nèi)容疑似由AI輔助生成,瀏覽時(shí)請(qǐng)結(jié)合常識(shí)與多方信息審慎甄別。
平臺(tái)聲明:文章內(nèi)容(如有圖片或視頻亦包括在內(nèi))由作者上傳并發(fā)布,文章內(nèi)容僅代表作者本人觀點(diǎn),簡(jiǎn)書系信息發(fā)布平臺(tái),僅提供信息存儲(chǔ)服務(wù)。

友情鏈接更多精彩內(nèi)容