簡介:服務網格ASM的Mixerless Telemetry技術,爲業務容器提供了無侵入式的遙測數據。遙測數據一方面做爲監控指標被ARMPS/prometheus採集,用於服務網格可觀測性;另外一方面被HPA和flaggers使用,成爲應用級擴縮容和漸進式灰度發佈的基石。 本系列聚焦於遙測數據在應用級擴縮容和漸進式灰度發佈上的實踐,將分三篇介紹遙測數據(監控指標)、應用級擴縮容,和漸進式灰度發佈。
服務網格ASM的Mixerless Telemetry技術,爲業務容器提供了無侵入式的遙測數據。遙測數據一方面做爲監控指標被ARMPS/prometheus採集,用於服務網格可觀測性;另外一方面被HPA和flaggers使用,成爲應用級擴縮容和漸進式灰度發佈的基石。html
本系列聚焦於遙測數據在應用級擴縮容和漸進式灰度發佈上的實踐,將分三篇介紹遙測數據(監控指標)、應用級擴縮容,和漸進式灰度發佈。node
本系列的整體架構以下圖所示:git
Flagger官網描述了漸進式發佈流程,這裏翻譯以下:github
acceptance-test
驗證load-test
驗證原文以下:json
With the above configuration, Flagger will run a canary release with the following steps:api
- detect new revision (deployment spec, secrets or configmaps changes)
- scale from zero the canary deployment
- wait for the HPA to set the canary minimum replicas
- check canary pods health
- run the acceptance tests
- abort the canary release if tests fail
- start the load tests
- mirror 100% of the traffic from primary to canary
- check request success rate and request duration every minute
- abort the canary release if the metrics check failure threshold is reached
- stop traffic mirroring after the number of iterations is reached
- route live traffic to the canary pods
- promote the canary (update the primary secrets, configmaps and deployment spec)
- wait for the primary deployment rollout to finish
- wait for the HPA to set the primary minimum replicas
- check primary pods health
- switch live traffic back to primary
- scale to zero the canary
- send notification with the canary analysis result
本篇將介紹如何基於ASM配置並採集應用級監控指標(好比請求數量總數istio_requests_total
和請求延遲istio_request_duration
等)。主要步驟包括建立EnvoyFilter、校驗envoy遙測數據和校驗Prometheus採集遙測數據。瀏覽器
登陸ASM控制檯,左側導航欄選擇服務網格 >網格管理,並進入ASM實例的功能配置頁面。架構
`prometheus:9090
(本系列將使用社區版Prometheus,後文將使用這個配置)。若是使用阿里雲產品ARMS,請參考集成ARMS Prometheus實現網格監控。點擊肯定後,咱們將在控制平面看到ASM生成的相關EnvoyFilter列表:app
執行以下命令安裝Prometheus(完整腳本參見:demo\_mixerless.sh)。less
kubectl --kubeconfig "$USER_CONFIG" apply -f $ISTIO_SRC/samples/addons/prometheus.yaml
安裝完Prometheus,咱們須要爲其配置添加istio相關的監控指標。登陸ACK控制檯,左側導航欄選擇配置管理>配置項,在istio-system
下找到prometheus
一行,點擊編輯。
在prometheus.yaml
配置中,將scrape\_configs.yaml中的配置追加到scrape_configs
中。
保存配置後,左側導航欄選擇工做負載>容器組,在istio-system
下找到prometheus
一行,刪除Prometheus POD,以確保配置在新的POD中生效。
能夠執行以下命令查看Prometheus配置中的job_name
:
kubectl --kubeconfig "$USER_CONFIG" get cm prometheus -n istio-system -o jsonpath={.data.prometheus\\.yml} | grep job_name - job_name: 'istio-mesh' - job_name: 'envoy-stats' - job_name: 'istio-policy' - job_name: 'istio-telemetry' - job_name: 'pilot' - job_name: 'sidecar-injector' - job_name: prometheus job_name: kubernetes-apiservers job_name: kubernetes-nodes job_name: kubernetes-nodes-cadvisor - job_name: kubernetes-service-endpoints - job_name: kubernetes-service-endpoints-slow job_name: prometheus-pushgateway - job_name: kubernetes-services - job_name: kubernetes-pods - job_name: kubernetes-pods-slow
使用以下命令部署本系列的示例應用podinfo:
kubectl --kubeconfig "$USER_CONFIG" apply -f $PODINFO_SRC/kustomize/deployment.yaml -n test kubectl --kubeconfig "$USER_CONFIG" apply -f $PODINFO_SRC/kustomize/service.yaml -n test
使用以下命令請求podinfo,以產生監控指標數據
podinfo_pod=$(k get po -n test -l app=podinfo -o jsonpath={.items..metadata.name}) for i in {1..10}; do kubectl --kubeconfig "$USER_CONFIG" exec $podinfo_pod -c podinfod -n test -- curl -s podinfo:9898/version echo done
本系列重點關注的監控指標項是istio_requests_total
和istio_request_duration
。首先,咱們在envoy容器內確認這些指標已經生成。
使用以下命令請求envoy獲取stats相關指標數據,並確認包含istio_requests_total
。
kubectl --kubeconfig "$USER_CONFIG" exec $podinfo_pod -n test -c istio-proxy -- curl -s localhost:15090/stats/prometheus | grep istio_requests_total
返回結果信息以下:
:::: istio_requests_total :::: # TYPE istio_requests_total counter istio_requests_total{response_code="200",reporter="destination",source_workload="podinfo",source_workload_namespace="test",source_principal="spiffe://cluster.local/ns/test/sa/default",source_app="podinfo",source_version="unknown",source_cluster="c199d81d4e3104a5d90254b2a210914c8",destination_workload="podinfo",destination_workload_namespace="test",destination_principal="spiffe://cluster.local/ns/test/sa/default",destination_app="podinfo",destination_version="unknown",destination_service="podinfo.test.svc.cluster.local",destination_service_name="podinfo",destination_service_namespace="test",destination_cluster="c199d81d4e3104a5d90254b2a210914c8",request_protocol="http",response_flags="-",grpc_response_status="",connection_security_policy="mutual_tls",source_canonical_service="podinfo",destination_canonical_service="podinfo",source_canonical_revision="latest",destination_canonical_revision="latest"} 10 istio_requests_total{response_code="200",reporter="source",source_workload="podinfo",source_workload_namespace="test",source_principal="spiffe://cluster.local/ns/test/sa/default",source_app="podinfo",source_version="unknown",source_cluster="c199d81d4e3104a5d90254b2a210914c8",destination_workload="podinfo",destination_workload_namespace="test",destination_principal="spiffe://cluster.local/ns/test/sa/default",destination_app="podinfo",destination_version="unknown",destination_service="podinfo.test.svc.cluster.local",destination_service_name="podinfo",destination_service_namespace="test",destination_cluster="c199d81d4e3104a5d90254b2a210914c8",request_protocol="http",response_flags="-",grpc_response_status="",connection_security_policy="unknown",source_canonical_service="podinfo",destination_canonical_service="podinfo",source_canonical_revision="latest",destination_canonical_revision="latest"} 10
使用以下命令請求envoy獲取stats相關指標數據,並確認包含istio_request_duration
。
kubectl --kubeconfig "$USER_CONFIG" exec $podinfo_pod -n test -c istio-proxy -- curl -s localhost:15090/stats/prometheus | grep istio_request_duration
返回結果信息以下:
:::: istio_request_duration :::: # TYPE istio_request_duration_milliseconds histogram istio_request_duration_milliseconds_bucket{response_code="200",reporter="destination",source_workload="podinfo",source_workload_namespace="test",source_principal="spiffe://cluster.local/ns/test/sa/default",source_app="podinfo",source_version="unknown",source_cluster="c199d81d4e3104a5d90254b2a210914c8",destination_workload="podinfo",destination_workload_namespace="test",destination_principal="spiffe://cluster.local/ns/test/sa/default",destination_app="podinfo",destination_version="unknown",destination_service="podinfo.test.svc.cluster.local",destination_service_name="podinfo",destination_service_namespace="test",destination_cluster="c199d81d4e3104a5d90254b2a210914c8",request_protocol="http",response_flags="-",grpc_response_status="",connection_security_policy="mutual_tls",source_canonical_service="podinfo",destination_canonical_service="podinfo",source_canonical_revision="latest",destination_canonical_revision="latest",le="0.5"} 10 istio_request_duration_milliseconds_bucket{response_code="200",reporter="destination",source_workload="podinfo",source_workload_namespace="test",source_principal="spiffe://cluster.local/ns/test/sa/default",source_app="podinfo",source_version="unknown",source_cluster="c199d81d4e3104a5d90254b2a210914c8",destination_workload="podinfo",destination_workload_namespace="test",destination_principal="spiffe://cluster.local/ns/test/sa/default",destination_app="podinfo",destination_version="unknown",destination_service="podinfo.test.svc.cluster.local",destination_service_name="podinfo",destination_service_namespace="test",destination_cluster="c199d81d4e3104a5d90254b2a210914c8",request_protocol="http",response_flags="-",grpc_response_status="",connection_security_policy="mutual_tls",source_canonical_service="podinfo",destination_canonical_service="podinfo",source_canonical_revision="latest",destination_canonical_revision="latest",le="1"} 10 ...
最後,咱們驗證Envoy生成的監控指標數據,是否被Prometheus實時採集上來。對外暴露Prometheus服務,並使用瀏覽器請求該服務。而後在查詢框輸入istio_requests_total
,獲得結果以下圖所示。
本文內容由阿里雲實名註冊用戶自發貢獻,版權歸原做者全部,阿里雲開發者社區不擁有其著做權,亦不承擔相應法律責任。具體規則請查看《阿里雲開發者社區用戶服務協議》和《阿里雲開發者社區知識產權保護指引》。若是您發現本社區中有涉嫌抄襲的內容,填寫侵權投訴表單進行舉報,一經查實,本社區將馬上刪除涉嫌侵權內容。