k8s中负载均衡器【ingress-nginx】部署
在kubernetes中,服务和pod的ip地址仅可以在集群网络内部使用,对于集群外的应用是不可见的。为了使外部的应用能够访问集群内的服务,在kubernetes 目前 提供了以下几种方案:
nodeport
loadbalancer
ingress
本节主要就ingress和ingress控制器ingress-nginx-controller的部署作简单介绍和记录。
以下系统组件版本:
云服务器:centos版本7.6.1810、k8s版本1.15.0、docker版本18.06.1-ce、ingress-nginx-controller版本0.25.0
ingress
ingress 组成?
将nginx的配置抽象成一个ingress对象,每添加一个新的服务只需写一个新的ingress的yaml文件即可
将新加入的ingress转化成nginx的配置文件并使之生效
ingress controller
ingress服务
ingress 工作原理?
- ingress controller通过和kubernetes api交互,动态的去感知集群中ingress规则变化,
- 然后读取它,按照自定义的规则,规则就是写明了哪个域名对应哪个service,生成一段nginx配置,
- 再写到nginx-ingress-controller的pod里,这个ingress
controller的pod里运行着一个nginx服务,控制器会把生成的nginx配置写入/etc/nginx.conf文件中, - 然后reload一下使配置生效。以此达到域名分配置和动态更新的问题。
ingress 可以解决什么问题?
动态配置服务
如果按照传统方式, 当新增加一个服务时, 我们可能需要在流量入口加一个反向代理指向我们新的服务. 而如果用了ingress, 只需要配置好这个服务, 当服务启动时, 会自动注册到ingress的中, 不需要而外的操作.
减少不必要的端口暴露
配置过k8s的都清楚, 第一步是要关闭防火墙的, 主要原因是k8s的很多服务会以nodeport方式映射出去, 这样就相当于给宿主机打了很多孔, 既不安全也不优雅. 而ingress可以避免这个问题, 除了ingress自身服务可能需要映射出去, 其他服务都不要用nodeport方式
ingress当前的实现方式?
ingress-nginx-controller
目前最新版本的ingress-nginx-controller,用lua实现了当upstream变化时不用reload,大大减少了生产环境中由于服务的重启、升级引起的ip变化导致的nginx reload。
以下就ingress-nginx-controller的部署做简单记录:
yaml如下:
kubectl apply -f {如下文件}
apiversion: v1 kind: namespace metadata: name: ingress-nginx labels: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx --- kind: configmap apiversion: v1 metadata: name: nginx-configuration namespace: ingress-nginx labels: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx --- kind: configmap apiversion: v1 metadata: name: tcp-services namespace: ingress-nginx labels: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx --- kind: configmap apiversion: v1 metadata: name: udp-services namespace: ingress-nginx labels: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx --- apiversion: v1 kind: serviceaccount metadata: name: nginx-ingress-serviceaccount namespace: ingress-nginx labels: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx --- apiversion: rbac.authorization.k8s.io/v1beta1 kind: clusterrole metadata: name: nginx-ingress-clusterrole labels: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx rules: - apigroups: - "" resources: - configmaps - endpoints - nodes - pods - secrets verbs: - list - watch - apigroups: - "" resources: - nodes verbs: - get - apigroups: - "" resources: - services verbs: - get - list - watch - apigroups: - "" resources: - events verbs: - create - patch - apigroups: - "extensions" - "networking.k8s.io" resources: - ingresses verbs: - get - list - watch - apigroups: - "extensions" - "networking.k8s.io" resources: - ingresses/status verbs: - update --- apiversion: rbac.authorization.k8s.io/v1beta1 kind: role metadata: name: nginx-ingress-role namespace: ingress-nginx labels: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx rules: - apigroups: - "" resources: - configmaps - pods - secrets - namespaces verbs: - get - apigroups: - "" resources: - configmaps resourcenames: # defaults to "<election-id>-<ingress-class>" # here: "<ingress-controller-leader>-<nginx>" # this has to be adapted if you change either parameter # when launching the nginx-ingress-controller. - "ingress-controller-leader-nginx" verbs: - get - update - apigroups: - "" resources: - configmaps verbs: - create - apigroups: - "" resources: - endpoints verbs: - get --- apiversion: rbac.authorization.k8s.io/v1beta1 kind: rolebinding metadata: name: nginx-ingress-role-nisa-binding namespace: ingress-nginx labels: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx roleref: apigroup: rbac.authorization.k8s.io kind: role name: nginx-ingress-role subjects: - kind: serviceaccount name: nginx-ingress-serviceaccount namespace: ingress-nginx --- apiversion: rbac.authorization.k8s.io/v1beta1 kind: clusterrolebinding metadata: name: nginx-ingress-clusterrole-nisa-binding labels: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx roleref: apigroup: rbac.authorization.k8s.io kind: clusterrole name: nginx-ingress-clusterrole subjects: - kind: serviceaccount name: nginx-ingress-serviceaccount namespace: ingress-nginx --- apiversion: apps/v1 kind: deployment metadata: name: nginx-ingress-controller namespace: ingress-nginx labels: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx spec: replicas: 1 selector: matchlabels: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx template: metadata: labels: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx annotations: prometheus.io/port: "10254" prometheus.io/scrape: "true" spec: serviceaccountname: nginx-ingress-serviceaccount containers: - name: nginx-ingress-controller image: quay.io/kubernetes-ingress-controller/nginx-ingress-controller:0.25.0 args: - /nginx-ingress-controller - --configmap=$(pod_namespace)/nginx-configuration - --tcp-services-configmap=$(pod_namespace)/tcp-services - --udp-services-configmap=$(pod_namespace)/udp-services - --publish-service=$(pod_namespace)/ingress-nginx - --annotations-prefix=nginx.ingress.kubernetes.io securitycontext: allowprivilegeescalation: true capabilities: drop: - all add: - net_bind_service # www-data -> 33 runasuser: 33 env: - name: pod_name valuefrom: fieldref: fieldpath: metadata.name - name: pod_namespace valuefrom: fieldref: fieldpath: metadata.namespace ports: - name: http containerport: 80 - name: https containerport: 443 livenessprobe: failurethreshold: 3 httpget: path: /healthz port: 10254 scheme: http initialdelayseconds: 10 periodseconds: 10 successthreshold: 1 timeoutseconds: 10 readinessprobe: failurethreshold: 3 httpget: path: /healthz port: 10254 scheme: http periodseconds: 10 successthreshold: 1 timeoutseconds: 10 ---
在墙内会拉取不到以下镜像:
quay.io/kubernetes-ingress-controller/nginx-ingress-controller:0.25.0
可以使用国内阿里云镜像:
docker pull registry.cn-hangzhou.aliyuncs.com/google_containers/nginx-ingress-controller:0.25.0
安装后,在ingress-nginx命名空间下可以看到pod一直pending,describe pod报如下警告:
查看master节点默认加了污点,一般不允许pod调度到master节点:
如果k8s集群只有一个节点,可以在pod的spec下设置容忍该污点:
即:
spec: tolerations: - effect: noschedule key: node-role.kubernetes.io/master
可以看到ingress-nginx pod被调度到master节点,且变为running
看日志报以下警告:
w0714 13:31:04.883127 6 queue.go:130] requeuing &objectmeta{name:sync status,generatename:,namespace:,selflink:,uid:,resourceversion:,generation:0,creationtimestamp:0001-01-01 00:00:00 +0000 utc,deletiontimestamp:<nil>,deletiongraceperiodseconds:nil,labels:map[string]string{},annotations:map[string]string{},ownerreferences:[],finalizers:[],clustername:,initializers:nil,managedfields:[],}, err services "ingress-nginx" not found
需要创一个名为ingress-nginx的service:
kubectl apply -f {如下文件} kind: service apiversion: v1 metadata: name: ingress-nginx namespace: ingress-nginx labels: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx spec: externaltrafficpolicy: local type: loadbalancer selector: app.kubernetes.io/name: ingress-nginx app.kubernetes.io/part-of: ingress-nginx ports: - name: http port: 80 targetport: http - name: https port: 443 targetport: https
参考:
ingress deploy
taint和toleration(污点和容忍)
k8s 1.12部署ingress-nginx
本公众号免费提供csdn下载服务,海量it学习资源,如果你准备入it坑,励志成为优秀的程序猿,那么这些资源很适合你,包括但不限于java、go、python、springcloud、elk、嵌入式 、大数据、面试资料、前端 等资源。同时我们组建了一个技术交流群,里面有很多大佬,会不定时分享技术文章,如果你想来一起学习提高,可以公众号后台回复【2】,免费邀请加技术交流群互相学习提高,会不定期分享编程it相关资源。
扫码关注,精彩内容第一时间推给你
下一篇: Flask框架 请求与响应 & 模板语法