怎樣做網(wǎng)站外鏈最近新聞大事
微服務(wù)
問:用控制器來完成集群的工作負(fù)載,那么應(yīng)用如何暴漏出去?
答:需要通過微服務(wù)暴漏出去后才能被訪問
- Service 是一組提供相同服務(wù)的Pod對外開放的接口
- 借助Service,應(yīng)用可以實(shí)現(xiàn)服務(wù)發(fā)現(xiàn)和負(fù)載均衡
- Service 默認(rèn)只支持4層負(fù)載均衡能力,沒有7層功能,需要借助 Ingress 實(shí)現(xiàn)
微服務(wù)類型
微服務(wù)類型 | 作用描述 |
---|---|
ClusterIP | 默認(rèn)值,k8s系統(tǒng)給service自動分配的虛擬IP,只能在集群內(nèi)部訪問 |
NodePort | 將Service通過指定的Node上的端口暴露給外部,訪問任意一個NodeIP:nodePort都將路由到ClusterIP |
LoadBalancer | 在NodePort的基礎(chǔ)上,借助cloud provider創(chuàng)建一個外部的負(fù)載均衡器,并將請求轉(zhuǎn)發(fā)到 NodeIP:NodePort,此模式只能在云服務(wù)器上使用 |
ExternalName | 將服務(wù)通過 DNS CNAME 記錄方式轉(zhuǎn)發(fā)到指定的域名(通過 spec.externlName 設(shè)定 |
用例
[root@k8s-master ~]# kubectl create deployment mini--image myapp:v1 --replicas 2
# 生成控制器文件并建立控制器
[root@k8s-master ~]# kubectl create deployment mini--image myapp:v1 --replicas 2 --dry-run=client -o yaml > mini.yaml
# 生成微服務(wù)Yaml追加到已有Yaml中
[root@k8s-master ~]# kubectl expose deployment mini--port 80 --target-port 80 --dry-run=client -o yaml >> mini.yaml
[root@k8s-master ~]# kubectl delete deployments.apps mini
[root@k8s-master ~]# vim mini.yaml
[root@k8s-master ~]# kubectl apply -f mini.yaml
[root@k8s-master ~]# kubectl get service
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
kubernetes ClusterIP 10.96.0.1 <none> 443/TCP 9d
mini ClusterIP 10.104.255.78 <none> 80/TCP 42s
微服務(wù)默認(rèn)使用 iptables 調(diào)度
# 可以在火墻中查看到策略信息(一般在底下)
[root@k8s-master ~]# iptables -t nat -nL
...
KUBE-MARK-MASQ 6 -- !10.244.0.0/16 10.104.255.78 /* default/mini cluster IP */ tcp dpt:80
...
IPVS 模式
- Service 是由 kube-proxy 組件,加上 iptables 來共同實(shí)現(xiàn)的
- kube-proxy 通過 iptables 處理 Service 的過程,需要在宿主機(jī)上設(shè)置相當(dāng)多的 iptables 規(guī)則,如果宿主機(jī)有大量的Pod,不斷刷新iptables規(guī)則,會消耗大量的CPU資源
- IPVS模式的service,可以使K8s集群支持更多量級的Pod
IPVS 配置
# 所有節(jié)點(diǎn)安裝 ipvsadm
dnf install ipvsadm -y
# 修改Master節(jié)點(diǎn)的代理配置
[root@k8s-master ~]# kubectl -n kube-system edit cm kube-proxymetricsBindAddress: ""mode: "ipvs"# 設(shè)置kube-system使用IPVS模式nftables:
# 當(dāng)改變配置文件后,已運(yùn)行的Pod狀態(tài)不會改變,需要重啟Pod
[root@k8s-master ~]# kubectl -n kube-system get pods | awk '/kube-proxy/{system("kubectl -n kube-system delete pods "$1)}'
[root@k8s-master ~]# kubectl get pods -o wide
NAME READY STATUS RESTARTS AGE IP NODE
mini-d5496d8f4-75khx 1/1 Running 0 15m 10.244.2.47 k8s-node2.org
mini-d5496d8f4-792mb 1/1 Running 0 15m 10.244.1.70 k8s-node1.org
[root@k8s-master ~]# ipvsadm -Ln
...
TCP 10.104.255.78:80 rr-> 10.244.1.70:80 Masq 1 0 0-> 10.244.2.47:80 Masq 1 0 0
...
切換 IPVS 模式后,kube-proxy會在宿主機(jī)上添加一個虛擬網(wǎng)卡:kube-ipvs0,并分配所有service IP
[root@k8s-master ~]# ip a | tail
...inet 10.96.0.10/32 scope global kube-ipvs0valid_lft forever preferred_lft forever
深入微服務(wù)類型
ClusterIP
ClusterIP 模式只能在集群內(nèi)訪問,并對集群內(nèi)的Pod提供健康檢測和自動發(fā)現(xiàn)功能
ClusterIP 用例
[root@k8s-master ~]# vim mini.yaml
apiVersion: apps/v1
kind: Deployment
metadata:labels:app: mininame: mini
spec:replicas: 2selector:matchLabels:app: minitemplate:metadata:creationTimestamp: nulllabels:app: minispec:containers:- image: myapp:v1name: myapp
---
apiVersion: v1
kind: Service
metadata:labels:app: mininame: mini
spec:ports:- port: 80protocol: TCPtargetPort: 80selector:app: minitype: ClusterIP
# service 創(chuàng)建后 集群DNS 提供解析
[root@k8s-master ~]# dnf install bind-utils -y
[root@k8s-master ~]# dig mini.default.svc.cluster.local @10.96.0.10
...
;; ANSWER SECTION:
mini.default.svc.cluster.local. 30 IN A 10.104.255.78;; Query time: 3 msec
;; SERVER: 10.96.0.10#53(10.96.0.10)
;; WHEN: Wed Oct 09 21:07:57 CST 2024
;; MSG SIZE rcvd: 117
ClusterIP的另一種模式:HeadLess
HeadLess(無頭服務(wù))
對于無頭 Services 并不會分配 Cluster IP,kube-proxy 不會處理它們, 而且平臺也不會為它們進(jìn)行負(fù)載均衡和路由,集群訪問通過 DNS 解析直接指向到業(yè)務(wù) Pod 上的 IP,所有的調(diào)度由 DNS 單獨(dú)完成
HeadLess 用例
[root@k8s-master ~]# vim mini.yaml
...selector:app: minitype: ClusterIPclusterIP: None
[root@k8s-master ~]# kubectl delete -f mini.yaml
[root@k8s-master ~]# kubectl apply -f mini.yaml
# 測試
[root@k8s-master ~]# kubectl get service mini
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
mini ClusterIP None <none> 80/TCP 18s
[root@k8s-master ~]# dig mini.default.svc.cluster.local @10.96.0.10
# mini.default.svc.cluster.local. 集群DNS
...
;; ANSWER SECTION:
mini.default.svc.cluster.local. 30 IN A 10.244.2.48
# 解析到Pod上
mini.default.svc.cluster.local. 30 IN A 10.244.1.71
...
kubectl get services mini
[root@k8s-master ~]# kubectl run ovo --image busyboxplus -it
/ # nslookup mini
/ # nslookup mini.default.svc.cluster.local.
Server: 10.96.0.10
Address 1: 10.96.0.10 kube-dns.kube-system.svc.cluster.localName: mini.default.svc.cluster.local.
Address 1: 10.244.1.71 10-244-1-71.mini.default.svc.cluster.local
Address 2: 10.244.2.48 10-244-2-48.mini.default.svc.cluster.local
/ # curl mini
Hello MyApp | Version: v1 | <a href="hostname.html">Pod Name</a>
/ # curl mini/hostname.html
mini-d5496d8f4-228g5
[root@k8s-master ~]# kubectl describe service mini
...
Endpoints: 10.244.1.71:80,10.244.2.48:80
...
NodePort
通過 IPVS 暴漏端口,從而使外部主機(jī)通過 Mater 節(jié)點(diǎn)的對外 IP:Port 來訪問 Pod 業(yè)務(wù)
訪問過程:NodePort ——> ClusterIP ——> Pods
NodePort 用例
[root@k8s-master ~]# vim mini.yaml
...selector:app: minitype: NodePort
[root@k8s-master ~]# kubectl delete -f mini.yaml
[root@k8s-master ~]# kubectl apply -f mini.yaml
[root@k8s-master ~]# kubectl get services mini
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
mini NodePort 10.96.170.18 <none> 80:30835/TCP 3m22s
# nodeport在集群節(jié)點(diǎn)上綁定端口,一個端口對應(yīng)一個服務(wù)
[root@k8s-master ~]# kubectl describe service mini
...
NodePort: <unset> 30835/TCP
...
[root@k8s-master ~]# for i in {1..5}
> do
> curl 172.25.254.200:30835/hostname.html
> done
mini-d5496d8f4-cts8s
mini-d5496d8f4-9v24v
mini-d5496d8f4-cts8s
mini-d5496d8f4-9v24v
mini-d5496d8f4-cts8s
NodePort 默認(rèn)端口是 30000—32767,超出會報(bào)錯
如果需要使用范圍外的端口,就需要特殊設(shè)定
vim /etc/kubernetes/manifests/kube-apiserver.yaml
# 需要增加到- command:
- --service-node-port-range=30000-40000
添加 --service-node-port-range= 參數(shù),端口范圍可以自定義
修改后 api-server 會自動重啟,等 apiserver 正常啟動后才能操作集群
集群重啟自動完成在修改完參數(shù)后,全程不需要人為干預(yù)
LoadBalancer
云平臺會為我們分配vip并實(shí)現(xiàn)訪問,如果是裸金屬主機(jī)那么需要metallb來實(shí)現(xiàn)ip的分配
過程:LoadBalancer ——> NodePort ——> ClusterIP ——> Pods
LoadBalancer 用例
[root@k8s-master ~]# vim mini.yaml
...selector:app: minitype: LoadBalancer
[root@k8s-master ~]# kubectl delete -f mini.yaml
[root@k8s-master ~]# kubectl apply -f mini.yaml
# 默認(rèn)無法分配外部訪問IP
[root@k8s-master ~]# kubectl get svc mini
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
mini LoadBalancer 10.102.105.5 <pending> 80:31759/TCP 18s
LoadBalancer 模式適用云平臺,裸金屬環(huán)境需要安裝 MetalLB提供支持
MetalLB
官網(wǎng):https://metallb.universe.tf/installation/
MetalLB功能:為 LoadBalancer 分配 VIP
MetalLB 配置
# 設(shè)置 IPVS 模式
[root@k8s-master ~]# kubectl edit cm -n kube-system kube-proxy
...metricsBindAddress: ""mode: "ipvs"ipvs:strictARP: true
...
[root@k8s-master ~]# kubectl -n kube-system get pods | awk '/kube-proxy/{system("kubectl -n kube-system delete pods "$1)}'
# 下載部署文件
[root@k8s-master ~]# dnf install wget -y
[root@k8s-master ~]# wget https://raw.githubusercontent.com/metallb/metallb/v0.13.12/config/manifests/metallb-native.yaml
# 修改文件鏡像拉取地址(配置好Docker拉取鏡像默認(rèn)地址)
...
image: metallb/controller:v0.14.8
...
image: metallb/speaker:v0.14.8
...
# 上傳鏡像到harbor倉庫
[root@k8s-master ~]# docker pull quay.io/metallb/controller:v0.14.8
[root@k8s-master ~]# docker pull quay.io/metallb/speaker:v0.14.8
[root@k8s-master ~]# docker tag quay.io/metallb/speaker:v0.14.8 ooovooo.org/metallb/speaker:v0.14.8
[root@k8s-master ~]# docker tag quay.io/metallb/controller:v0.14.8 ooovooo.org/metallb/controller:v0.14.8
[root@k8s-master ~]# docker push ooovooo.org/metallb/speaker:v0.14.8
[root@k8s-master ~]# docker push ooovooo.org/metallb/controller:v0.14.8
# 部署服務(wù)
[root@k8s-master ~]# kubectl apply -f metallb-native.yaml
[root@k8s-master ~]# kubectl -n metallb-system get pods
NAME READY STATUS RESTARTS AGE
controller-65957f77c8-c9lrv 1/1 Running 0 23s
speaker-5g4hz 1/1 Running 0 23s
speaker-bw4qh 1/1 Running 0 23s
speaker-t7d7f 1/1 Running 0 23s
# 配置分配地址段
[root@k8s-master ~]# vim configmap.yml
apiVersion: metallb.io/v1beta1
kind: IPAddressPool
metadata:name: name# 地址池名稱namespace: metallb-system
spec:addresses:- 172.25.254.25-172.25.254.50# 地址池段
---
# 不同的kind之間使用---分割
apiVersion: metallb.io/v1beta1
kind: L2Advertisement
metadata:name: examplenamespace: metallb-system
spec:ipAddressPools:- name# 使用的地址池
[root@k8s-master ~]# kubectl apply -f configmap.yml
[root@k8s-master ~]# kubectl get services mini
# 自動分配IP
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
mini LoadBalancer 10.102.105.5 172.25.254.25 80:31759/TCP 62m
# 通過分配地址從集群外訪問服務(wù)
[root@k8s-master ~]# curl 172.25.254.25
Hello MyApp | Version: v1 | <a href="hostname.html">Pod Name</a>
ExternalName
- 開啟 services 后,不會被分配 IP,而是用 DNS 解析 CNAME 固定域名來解決 IP 變化問題
- 一般應(yīng)用于外部業(yè)務(wù)和 Pod 溝通或外部業(yè)務(wù)遷移到 Pod 內(nèi)時(shí)
- 在應(yīng)用向集群遷移過程中,ExternalName在過度階段就可以起作用了
- 集群外的資源遷移到集群時(shí),在遷移的過程中 IP 可能會變化,但是 域名+DNS解析 能完美解決此問題
ExternalName 用例
[root@k8s-master ~]# vim mini.yaml
...selector:app: minitype: ExternalNameexternalName: www.mini.org
[root@k8s-master ~]# kubectl delete -f mini.yaml
[root@k8s-master ~]# kubectl apply -f mini.yaml
[root@k8s-master ~]# kubectl get services mini
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
mini ExternalName <none> www.mini.org 80/TCP 5s
Ingress-Nginx
官網(wǎng):https://kubernetes.github.io/ingress-nginx/deploy/#bare-metal-clusters
Ingress-Nginx 功能
- 一種全局的、為了代理不同后端 Service 而設(shè)置的負(fù)載均衡服務(wù),支持7層
- Ingress由兩部分組成:Ingress controller和Ingress服務(wù)
- Ingress Controller 會根據(jù)你定義的 Ingress 對象,提供對應(yīng)的代理能力
部署 Ingress
[root@k8s-master ~]# mkdir ingress
[root@k8s-master ~]# cd ingress/
# 下載部署文件
[root@k8s-master ingress]# wget https://raw.githubusercontent.com/kubernetes/ingress-nginx/controller-v1.11.2/deploy/static/provider/baremetal/deploy.yaml
# 還要下載ingress-nginx的鏡像
# 上傳鏡像到harbor
[root@k8s-master ~]# docker tag reg.harbor.org/ingress-nginx/controller:v1.11.2 ooovooo.org/ingress-nginx/controller:v1.11.2
[root@k8s-master ~]# docker tag reg.harbor.org/ingress-nginx/kube-webhook-certgen:v1.4.3 ooovooo.org/ingress-nginx/kube-webhook-certgen:v1.4.3
[root@k8s-master ~]# docker push ooovooo.org/ingress-nginx/controller:v1.11.2
[root@k8s-master ~]# docker push ooovooo.org/ingress-nginx/kube-webhook-certgen:v1.4.3
# 安裝Ingress
[root@k8s-master ~]# vim deploy.yaml
...
image: ingress-nginx/controller:v1.11.2
...
image: ingress-nginx/kube-webhook-certgen:v1.4.3
...
image: ingress-nginx/kube-webhook-certgen:v1.4.3
[root@k8s-master ingress]# kubectl apply -f deploy.yaml
[root@k8s-master ingress]# kubectl -n ingress-nginx get pods
# 一開始可能會有一個error,刪掉再加載就好了
NAME READY STATUS RESTARTS AGE
ingress-nginx-admission-create-w6jz9 0/1 Completed 0 14s
ingress-nginx-admission-patch-bbsn6 0/1 Completed 1 14s
ingress-nginx-controller-bb7d8f97c-nx96n 1/1 Running 0 14s
# ingress-nginx-controller 1/1 Running 即運(yùn)行成功
[root@k8s-master ~]# kubectl -n ingress-nginx get svc
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
ingress-nginx-controller NodePort 10.100.33.214 <none> 80:32416/TCP,443:30320/TCP 30s
ingress-nginx-controller-admission ClusterIP 10.98.75.102 <none> 443/TCP 30s
# 修改微服務(wù)為loadbalancer
[root@k8s-master ~]# kubectl -n ingress-nginx edit svc ingress-nginx-controller
...
49 type: LoadBalancer
# 查看修改結(jié)果
[root@k8s-master ~]# kubectl -n ingress-nginx get services
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S)
# 需要配置有 MetalLB
ingress-nginx-controller LoadBalancer 10.100.33.214 172.25.254.25 80:32416/TCP,443:30320/TCP
ingress-nginx-controller-admission ClusterIP 10.98.75.102 <none> 443/TCP
在 kubectl -n ingress-nginx get services 中 的 EXTERNAL-IP
即 Ingress 最終對外的 IP
測試 Ingress
[root@k8s-master ingress]# kubectl create deployment myappv1 --image myapp:v1 --dry-run=client -o yaml > myappv1.yml
[root@k8s-master ingress]# kubectl apply -f myappv1.yml
[root@k8s-master ingress]# kubectl expose deployment myappv1 --port 80 --target-port 80 --dry-run=client -o yaml >> myappv1.yml
[root@k8s-master ingress]# vim myappv1.yml
apiVersion: apps/v1
kind: Deployment
metadata:labels:app: myappv1name: myappv1
spec:replicas: 1selector:matchLabels:app: myappv1strategy: {}template:metadata:labels:app: myappv1spec:containers:- image: myapp:v1name: myapp
---
apiVersion: v1
kind: Service
metadata:labels:app: myappv1name: myappv1
spec:ports:- port: 80protocol: TCPtargetPort: 80selector:app: myappv1
kubectl apply -f myappv1.yml
kubectl create ingress webcluster --rule '*/=ooovooo-svc:80' --dry-run=client -o yaml > ingress.yml
[root@k8s-master ingress]# vim ingress.yml
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:name: myappv1
spec:ingressClassName: nginxrules:- http:paths:- backend:service:name: myappv1# 與自己的服務(wù)名保持一致port:number: 80path: /pathType: Prefix# Exact(精確匹配)# ImplementationSpecific(特定實(shí)現(xiàn))# Prefix(前綴匹配)# Regular expression(正則表達(dá)式匹配)
# 建立Ingress控制器
kubectl apply -f ingress.yml
# 根據(jù)kubectl -n ingress-nginx get services中的IP進(jìn)行訪問
[root@k8s-master ingress]# curl 172.25.254.25
Hello MyApp | Version: v1 | <a href="hostname.html">Pod Name</a>
Ingress 必須和輸出的 service 資源處于同一 namespace 中
Ingress 高級用法
基于路徑的訪問
[root@k8s-master ingress]# vim ingress1.yml
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:annotations:nginx.ingress.kubernetes.io/rewrite-target: /name: ingress1
spec:ingressClassName: nginxrules:- host: www.ooovooo.orghttp:paths:- backend:service:name: myappv1port:number: 80path: /v1pathType: Prefix- backend:service:name: myappv2port:number: 80path: /v2pathType: Prefix
[root@k8s-master ingress]# kubectl apply -f ingress1.yml
[root@k8s-master ingress]# echo 172.25.254.25 www.ooovooo.org >> /etc/hosts
[root@k8s-master ingress]# curl www.ooovooo.org/v1
Hello MyApp | Version: v1 | <a href="hostname.html">Pod Name</a>
[root@k8s-master ingress]# curl www.ooovooo.org/v2
Hello MyApp | Version: v2 | <a href="hostname.html">Pod Name</a>
[root@k8s-master ingress]# curl www.ooovooo.org/v2/haha
Hello MyApp | Version: v2 | <a href="hostname.html">Pod Name</a>
[root@k8s-master ingress]# curl www.ooovooo.org/v1/gaga
Hello MyApp | Version: v1 | <a href="hostname.html">Pod Name</a>
基于域名的訪問
[root@k8s-master ingress]# vim ingress2.yml
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:annotations:nginx.ingress.kubernetes.io/rewrite-target: /name: ingress2
spec:ingressClassName: nginxrules:- host: myappv1.ooovooo.orghttp:paths:- backend:service:name: myappv1port:number: 80path: /pathType: Prefix- host: myappv2.ooovooo.orghttp:paths:- backend:service:name: myappv2port:number: 80path: /pathType: Prefix
[root@k8s-master ingress]# kubectl apply -f ingress2.yml
[root@k8s-master ingress]# kubectl delete -f ingress1.yml
[root@k8s-master ingress]# kubectl describe ingress ingress2
...Host Path Backends---- ---- --------myappv1.ooovooo.org/ myappv1:80 (10.244.1.89:80)myappv2.ooovooo.org/ myappv2:80 (10.244.1.90:80)
[root@k8s-master ingress]# curl myappv1.ooovooo.org
Hello MyApp | Version: v1 | <a href="hostname.html">Pod Name</a>
[root@k8s-master ingress]# curl myappv2.ooovooo.org
Hello MyApp | Version: v2 | <a href="hostname.html">Pod Name</a>
建立 TLS 加密
[root@k8s-master ingress]# openssl req -newkey rsa:2048 -nodes -keyout tls.key -x509 -days 365 -subj "/CN=nginxsvc/O=nginxsvc" -out tls.crt
[root@k8s-master ingress]# kubectl create secret tls web-tls-secret --key tls.key --cert tls.crt
[root@k8s-master ingress]# vim ingress3.yml
[root@k8s-master ingress]# echo 172.25.254.25 myapp-tls.ooovooo.org >> /etc/hosts
[root@k8s-master ingress]# kubectl apply -f ingress3.yml
[root@k8s-master ingress]# kubectl delete -f ingress2.yml
# 在Windows主機(jī)添加解析,并進(jìn)行訪問
建立 AUTH 認(rèn)證
[root@k8s-master ingress]# vim ingress4.yml
[root@k8s-master ingress]# kubectl delete -f ingress3.yml
[root@k8s-master ingress]# kubectl apply -f ingress4.yml
[root@k8s-master ingress]# kubectl describe ingress ingress4
...
TLS:web-tls-secret terminates myapp-tls.ooovooo.org
...myapp-tls.ooovooo.org/ myappv1:80 (10.244.1.89:80)
...
[root@k8s-master ingress]# curl -k https://myapp-tls.ooovooo.org
<html>
<head><title>401 Authorization Required</title></head>
<body>
<center><h1>401 Authorization Required</h1></center>
<hr><center>nginx</center>
</body>
</html>
[root@k8s-master ingress]# curl -k https://myapp-tls.ooovooo.org -u ovo:aaa
Hello MyApp | Version: v1 | <a href="hostname.html">Pod Name</a>
# Windows主機(jī)訪問同樣需要登錄
Rewrite 重定向
# 將指定訪問文件重定向到hostname.html上
[root@k8s-master ingress]# vim ingress5.yml
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:annotations:nginx.ingress.kubernetes.io/app-root: /hostname.htmlname: ingress5
spec:ingressClassName: nginxrules:- host: myapp-tls.ooovooo.orghttp:paths:- backend:service:name: myappv1port:number: 80path: /# 當(dāng)訪問/時(shí),重定向到hostname.htmlpathType: Prefix
[root@k8s-master ingress]# kubectl delete -f ingress4.yml
[root@k8s-master ingress]# kubectl apply -f ingress5.yml
[root@k8s-master ingress]# curl -Lk https://myapp-tls.ooovooo.org -u ovo:aaa
myappv1-586444467f-w4dxn
[root@k8s-master ingress]# curl -Lk https://myapp-tls.ooovooo.org/haha/hostname.html -u ovo:aaa
<html>
<head><title>404 Not Found</title></head>
<body bgcolor="white">
<center><h1>404 Not Found</h1></center>
<hr><center>nginx/1.12.2</center>
</body>
</html>
# 以上存在一個問題,當(dāng)有多路徑時(shí)需要重定向時(shí),需要配置多個,費(fèi)人力
# 正則解決指定路徑問題
[root@k8s-master ingress]# vim ingress6.yml
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:annotations:nginx.ingress.kubernetes.io/rewrite-target: /$2nginx.ingress.kubernetes.io/use-regex: "true"name: ingress6
spec:ingressClassName: nginxrules:- host: myapp-tls.ooovooo.orghttp:paths:- backend:service:name: myapp-v1port:number: 80path: /pathType: Prefix- backend:service:name: myappv1port:number: 80path: /haha(/|$)(.*)pathType: ImplementationSpecific
[root@k8s-master ingress]# kubectl delete -f ingress5.yml
[root@k8s-master ingress]# kubectl apply -f ingress6.yml
[root@k8s-master ingress]# curl -Lk https://myapp-tls.ooovooo.org/haha/hostname.html -u ovo:aaa
myappv1-586444467f-w4dxn
Canary 金絲雀發(fā)布
金絲雀發(fā)布(Canary Release)也稱為灰度發(fā)布,是一種軟件發(fā)布策略
主要目的是在將新版本的軟件全面推廣到生產(chǎn)環(huán)境之前,先在一小部分用戶或服務(wù)器上進(jìn)行測試和驗(yàn)證,以降低因新版本引入重大問題而對整個系統(tǒng)造成的影響,是一種 Pod 的發(fā)布方式
金絲雀發(fā)布采取先添加、再刪除的方式,保證Pod的總量不低于期望值。并且在更新部分Pod后,暫停更新,當(dāng)確認(rèn)新Pod版本運(yùn)行正常后再進(jìn)行其他版本的Pod的更新
發(fā)布方式
Header > Cookie > Weiht
其中 Header 和 Weiht 中的最多
基于Header (HTTP包頭)灰度
- 通過Annotaion擴(kuò)展
- 創(chuàng)建灰度 Ingress,配置灰度頭部 key 以及 value
- 灰度流量驗(yàn)證完畢后,切換正式 Ingress 到新版本
- 之前我們在做升級時(shí)可以通過控制器做滾動更新,默認(rèn)25%利用Header 可以使升級更為平滑,通過 key 和 value 測試新的業(yè)務(wù)體系是否有問題
# 創(chuàng)建版本v1的ingress
[root@k8s-master ingress]# kubectl delete -f ingress6.yml
[root@k8s-master ingress]# vim ingress7.yml
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:annotations:name: myapp-v1-ingress
spec:ingressClassName: nginxrules:- host: myapp-tls.ooovooo.orghttp:paths:- backend:service:name: myappv1port:number: 80path: /pathType: Prefix
[root@k8s-master ingress]# kubectl apply -f ingress7.yml
# 建立基于header的ingress
[root@k8s-master ingress]# vim ingress8.yml
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:annotations:nginx.ingress.kubernetes.io/canary: "true"nginx.ingress.kubernetes.io/canary-by-header: "version"nginx.ingress.kubernetes.io/canary-by-header-value: "2"name: myapp-v2-ingress
spec:ingressClassName: nginxrules:- host: myapp-tls.ooovooo.orghttp:paths:- backend:service:name: myappv2port:number: 80path: /pathType: Prefix
[root@k8s-master ingress]# kubectl apply -f ingress8.yml
# 進(jìn)行測試
[root@k8s-master ingress]# curl myapp-tls.ooovooo.org
Hello MyApp | Version: v1 | <a href="hostname.html">Pod Name</a>
[root@k8s-master ingress]# curl -H "version: 2" myapp-tls.ooovooo.org
Hello MyApp | Version: v2 | <a href="hostname.html">Pod Name</a>
基于權(quán)重的灰度發(fā)布
- 通過 Annotaion 拓展
- 創(chuàng)建灰度 Ingress,配置灰度權(quán)重以及總權(quán)重
- 灰度流量驗(yàn)證完畢后,切換正式 Ingress 到新版本
[root@k8s-master ingress]# vim ingress9.yml
apiVersion: networking.k8s.io/v1
kind: Ingress
metadata:annotations:nginx.ingress.kubernetes.io/canary: "true"nginx.ingress.kubernetes.io/canary-weight: "10"# 灰度權(quán)重nginx.ingress.kubernetes.io/canary-weight-total: "100"# 總權(quán)重name: myapp-v2-ingress
spec:ingressClassName: nginxrules:- host: myapp-tls.ooovooo.orghttp:paths:- backend:service:name: myappv2port:number: 80path: /pathType: Prefix
[root@k8s-master ingress]# kubectl delete -f ingress8.yml
[root@k8s-master ingress]# kubectl apply -f ingress9.yml
[root@k8s-master ingress]# vim check_ingress.sh
#!/bin/bash
v1=0
v2=0for (( i=0; i<100; i++))
doresponse=`curl -s myapp-tls.ooovooo.org |grep -c v1`v1=`expr $v1 + $response`v2=`expr $v2 + 1 - $response`done
echo "v1:$v1, v2:$v2"
[root@k8s-master ingress]# chmod +x check_ingress.sh
[root@k8s-master ingress]# sh check_ingress.sh
v1:89, v2:11
# 根據(jù)不同灰度權(quán)重而不同