服务发现,说白了就是服务(应用)之间相互定位的过程。node
服务发现须要解决的问题:nginx
一、服务动态性强--容器在k8s中ip变化或迁移git
二、更新发布频繁--版本迭代快github
三、支持自动伸缩--大促或流量高峰docker
咱们为了解决pod地址变化的问题,咱们以前部署了service资源,将pod地址经过service资源暴露的固定地址,来解决以上问题,bootstrap
那么,如何解决service资源名称和service资源暴露出来的集群网络IP作自动的对应呢,从而达到服务的自动发现呢?api
在k8s中,coredns就是为了解决以上问题。网络
从coredns开始,咱们采用向k8s中交付容器的方式,来部署服务,而且使用声明式的方式,来部署服务。app
首先在hdss7-200上建立一个nginx虚拟主机,用来获取资源配置清单:tcp
vi /etc/nginx/conf.d/k8s-yaml.od.com.conf
server { listen 80; server_name k8s-yaml.od.com; location / { autoindex on; default_type text/plain; root /data/k8s-yaml; } }
# mkdir -p /data/k8s-yaml/coredns
# nginx -t # nginx -s reload
添加域名解析:hdss-11上
# vi /var/named/od.com.zone 在最后添加一条解析记录 $ORIGIN od.com. $TTL 600 ; 10 minutes @ IN SOA dns.od.com. dnsadmin.od.com. ( 2019061803 ; serial 10800 ; refresh (3 hours) 900 ; retry (15 minutes) 604800 ; expire (1 week) 86400 ; minimum (1 day) ) NS dns.od.com. $TTL 60 ; 1 minute dns A 10.4.7.11 harbor A 10.4.7.200 k8s-yaml A 10.4.7.200
# systemctl restart named
coredns github地址:
# cd /data/k8s-yaml/coredns
# docker pull docker.io/coredns/coredns:1.6.1
# docker tag c0f6e815079e harbor.od.com/public/coredns:v1.6.1
# docker push harbor.od.com/public/coredns:v1.6.1
而后编辑资源配置清单:能够从官网上参考资源配置清单
1.rbac.yaml--拿到集群相关权限
# vi rbac.yaml
apiVersion: v1 kind: ServiceAccount metadata: name: coredns namespace: kube-system labels: kubernetes.io/cluster-service: "true" addonmanager.kubernetes.io/mode: Reconcile --- apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRole metadata: labels: kubernetes.io/bootstrapping: rbac-defaults addonmanager.kubernetes.io/mode: Reconcile name: system:coredns rules: - apiGroups: - "" resources: - endpoints - services - pods - namespaces verbs: - list - watch --- apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRoleBinding metadata: annotations: rbac.authorization.kubernetes.io/autoupdate: "true" labels: kubernetes.io/bootstrapping: rbac-defaults addonmanager.kubernetes.io/mode: EnsureExists name: system:coredns roleRef: apiGroup: rbac.authorization.k8s.io kind: ClusterRole name: system:coredns subjects: - kind: ServiceAccount name: coredns namespace: kube-system
2.cm.yaml--configmap 对集群的相关配置
# vi cm.yaml
apiVersion: v1 kind: ConfigMap metadata: name: coredns namespace: kube-system data: Corefile: | .:53 { errors log health ready kubernetes cluster.local 192.168.0.0/16 #service资源cluster地址 forward . 10.4.7.11 #上级DNS地址 cache 30 loop reload loadbalance }
3.dp.yaml---pod控制器
# vi dp.yaml
apiVersion: apps/v1 kind: Deployment metadata: name: coredns namespace: kube-system labels: k8s-app: coredns kubernetes.io/name: "CoreDNS" spec: replicas: 1 selector: matchLabels: k8s-app: coredns template: metadata: labels: k8s-app: coredns spec: priorityClassName: system-cluster-critical serviceAccountName: coredns containers: - name: coredns image: harbor.od.com/public/coredns:v1.6.1 args: - -conf - /etc/coredns/Corefile volumeMounts: - name: config-volume mountPath: /etc/coredns ports: - containerPort: 53 name: dns protocol: UDP - containerPort: 53 name: dns-tcp protocol: TCP - containerPort: 9153 name: metrics protocol: TCP livenessProbe: httpGet: path: /health port: 8080 scheme: HTTP initialDelaySeconds: 60 timeoutSeconds: 5 successThreshold: 1 failureThreshold: 5 dnsPolicy: Default volumes: - name: config-volume configMap: name: coredns items: - key: Corefile path: Corefile
4.svc.yaml---service资源
# vi svc.yaml
apiVersion: v1 kind: Service metadata: name: coredns namespace: kube-system labels: k8s-app: coredns kubernetes.io/cluster-service: "true" kubernetes.io/name: "CoreDNS" spec: selector: k8s-app: coredns clusterIP: 192.168.0.2 ports: - name: dns port: 53 protocol: UDP - name: dns-tcp port: 53 - name: metrics port: 9153 protocol: TCP
而后使用http请求资源配置清单yaml的方式来建立资源:在任意node节点上建立
# kubectl create -f http://k8s-yaml.od.com/coredns/rbac.yaml # kubectl create -f http://k8s-yaml.od.com/coredns/cm.yaml # kubectl create -f http://k8s-yaml.od.com/coredns/dp.yaml # kubectl create -f http://k8s-yaml.od.com/coredns/svc.yaml
查看运行状况:
# kubectl get all -n kube-system
查看coredns的cluster ip:
# kubectl get svc -o wide -n kube-system
测试coredns:
# dig -t A www.baidu.com @192.168.0.2 +short
看到已经能够解析到百度。
测试coredns解析service资源名称,首先查看kube-public下是否有service资源,若是没有,建立一个,使用kubectl expose nginx-dp --port=80 -n kube-public
# kubectl expose nginx-dp --port=80 -n kube-public
测试:使用coredns测试解析,须要使用SQDN规则
# dig -t A nginx-dp.kube-public.svc.cluster.local. @192.168.0.2 +short
能够看到咱们没有手动添加任何解析记录,咱们nginx-dp的service资源的IP,已经被解析了:
那么为何呢?
推荐你们了解一下coredns都作了什么:Kubernetes内部域名解析原理、弊端及优化方式
你们能够看到,当我进入到pod内部之后,咱们会发现咱们的dns地址是咱们的coredns地址,以及搜索域: