Ini ialah panduan langkah demi langkah untuk menyediakan Kubernetes pada Scaleway bare-metal ARM dan x86-64. Sebab utama saya mengusahakan projek ini ialah saya ingin mengautomasikan penciptaan persekitaran ujian untuk OpenFaaS dan Weave Net pada ARM. Saya sedang mencari penyelesaian murah untuk menjalankan ujian integrasi dan selepas mencuba beberapa penyedia awan, saya telah menyelesaikan Scaleway. Scaleway ialah penyedia awan Perancis yang menawarkan pelayan ARM dan x86-64 tanpa logam pada harga yang berpatutan. Menggunakan penyedia Terraform Scaleway bersama-sama dengan kubeadm anda boleh mempunyai kluster Kubernetes yang berfungsi sepenuhnya dalam masa sepuluh minit
Persediaan awal
Klon repositori dan pasang kebergantungan:
$ git klon httpsgithub.com/stefanprodan/k8s-scw-baremetal.git $ cd k8s-scw-baremetal $ terraform init
Ambil perhatian bahawa anda memerlukan Terraform v0.10 atau lebih baharu untuk menjalankan projek ini
Sebelum menjalankan projek, anda perlu membuat token akses untuk Terraform menyambung ke Scaleway API. Menggunakan token dan kunci akses anda, buat dua pembolehubah persekitaran:
$ eksport SCALEWAY_ORGANIZATIONACCESS-KEY>"$ eksport SCALEWAY_TOKENACCESS-TOKEN>"Penggunaan
Buat gugusan Kubernetes bare-metal ARMv7 dengan satu induk dan dua nod:
$ ruang kerja terraform lengan baharu $ terraform apply \ -var region=par1 \ -var arch=arm \ -var server_type=C1 \ -var nodes=2 \ -var weave_passwd=ChangeMe \ -var k8s_version=stable-1.9 \ -var docker_version =17.03.0~ce-0~ubuntu-xenial
Ini akan melakukan perkara berikut:
- menyimpan IP awam untuk setiap pelayan
- menyediakan tiga pelayan bare-metal dengan Ubuntu 16.04.1 LTS
- menyambung ke pelayan induk melalui SSH dan memasang pakej Docker CE dan kubeadm armhf apt
- menjalankan kubeadm init pada pelayan induk dan mengkonfigurasi kubectl
- memuat turun fail konfigurasi pentadbir kubectl pada mesin tempatan anda dan menggantikan IP peribadi dengan yang awam
- mencipta rahsia Kubernetes dengan kata laluan Weave Net
- memasang Weave Net dengan tindanan yang disulitkan
- memasang alat tambah kluster (papan pemuka Kubernetes, pelayan metrik dan Heapster)
- memulakan nod pekerja secara selari dan memasang Docker CE dan kubeadm
- menyertai nod pekerja dalam kelompok menggunakan token kubeadm yang diperoleh daripada induk
Skalakan dengan menambah bilangan nod:
$ terraform digunakan -var nodes=3
Runtuhkan keseluruhan infrastruktur dengan:
terraform-force
Buat gugusan Kubernetes bare-metal AMD64 dengan satu induk dan satu nod:
$ ruang kerja terraform amd64 baharu $ terraform digunakan \ -var region=par1 \ -var arch=x86_64 \ -var server_type=C2S \ -var nodes=1 \ -var weave_passwd=ChangeMe \ -var k8s_version=stable-1.9 \ -var docker_version =17.03.0~ce-0~ubuntu-xenial
Alat kawalan jauh
Selepas menggunakan pelan Terraform, anda akan melihat beberapa pembolehubah output seperti IP awam induk, arahan sertai kubeadmn dan konfigurasi pentadbir ruang kerja semasa
Untuk berlari
perintah kubectl terhadap kluster Scaleway yang anda boleh gunakan
pembolehubah output kubectl_config:
Semak sama ada Heapster berfungsi:
$ kubectl --kubeconfig terraform output kubectl_config) nod atas NAMA CPU(teras) CPU% MEMORY(bait) MEMORY% arm-master-1 655m 16% 873Mi 45% arm-node-1 147m 3% 618Mi 32% arm-node- 2 101m 2% 584Mi 30%
The
format fail konfigurasi kubectl ialah
.conf as in
arm.conf or
amd64.conf
In order to access the dashboard youâÂÂll need to find its cluster IP:
$ kubectl --kubeconfig terraform output kubectl_config) \ -n kube-system get svc --selector=k8s-app=kubernetes-dashboard NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE kubernetes-dashboard ClusterIP 10.107.37.220 80/TCP 6m
Open a SSH tunnel:
ssh -L 8888::80 [email protected]
Now you can access the dashboard on your computer at
httplocalhost:8888
Expose services outside the cluster
Since weâÂÂre running on bare-metal and Scaleway doesnâÂÂt offer a load balancer, the easiest way to expose applications outside of Kubernetes is using a NodePort service
LetâÂÂs deploy the podinfo app in the default namespace. Podinfo has a multi-arch Docker image and it will work on arm, arm64 or amd64
Create the podinfo nodeport service:
$ kubectl --kubeconfig terraform output kubectl_config) \ apply -f httpsraw.githubusercontent.com/stefanprodan/k8s-podinfo/master/deploy/auto-scaling/podinfo-svc-nodeport.yaml service "podinfo-nodeport" created
Create the podinfo deployment:
$ kubectl --kubeconfig terraform output kubectl_config) \ apply -f httpsraw.githubusercontent.com/stefanprodan/k8s-podinfo/master/deploy/auto-scaling/podinfo-dep.yaml deployment "podinfo" created
Inspect the podinfo service to obtain the port number:
$ kubectl --kubeconfig terraform output kubectl_config) \ get svc --selector=app=podinfo NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE podinfo-nodeport NodePort 10.104.132.14 9898:31190/TCP 3m
You can access podinfo at
httpMASTER_PUBLIC_IP>:31190 or using curl:
$ curl httpterraform output k8s_master_public_ip):31190 runtime: arch: arm max_procs: "4" num_cpu: "4" num_goroutine: "12" os: linux version: go1.9.2 labels: app: podinfo pod-template-hash: "1847780700" annotations: kubernetes.io/config.seen: 2018-01-08T00:39:45.580597397Z kubernetes.io/config.source: api environment: HOME: /root HOSTNAME: podinfo-5d8ccd4c44-zrczc KUBERNETES_PORT: tcp10.96.0.1:443 KUBERNETES_PORT_443_TCP: tcp10.96.0.1:443 KUBERNETES_PORT_443_TCP_ADDR: 10.96.0.1 KUBERNETES_PORT_443_TCP_PORT: "443" KUBERNETES_PORT_443_TCP_PROTO: tcp KUBERNETES_SERVICE_HOST: 10.96.0.1 KUBERNETES_SERVICE_PORT: "443" KUBERNETES_SERVICE_PORT_HTTPS: "443" PATH: /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin externalIP: IPv4: 163.172.139.112
OpenFaaS
You can deploy OpenFaaS on Kubernetes with Helm or by using the YAML files form the faas-netes repository
Clone the faas-netes repo:
git clone httpsgithub.com/openfaas/faas-netes cd faas-netes
Deploy OpenFaaS for ARM:
$ kubectl --kubeconfig terraform output kubectl_config) \ apply -f ./namespaces.ymlyaml_armhf
Deploy OpenFaaS for AMD64:
$ kubectl --kubeconfig terraform output kubectl_config) \ apply -f ./namespaces.ymlyaml
You can access the OpenFaaS gateway at
httpMASTER_PUBLIC_IP>:31112
Horizontal Pod Autoscaling
Starting from Kubernetes 1.9
kube-controller-manager is configured by default with
horizontal-pod-autoscaler-use-rest-clients
In order to use HPA we need to install the metrics server to enable the new metrics API used by HPA v2
Both Heapster and the metrics server have been deployed from Terraform
when the master node was provisioned
The metric server collects resource usage data from each node using Kubelet Summary API. Check if the metrics server is running:
$ kubectl --kubeconfig terraform output kubectl_config) \ get --raw "/apis/metrics.k8s.io/v1beta1/nodes" | jq
{ "kind": "NodeMetricsList", "apiVersion": "metrics.k8s.io/v1beta1", "metadata": { "selfLink": "/apis/metrics.k8s.io/v1beta1/nodes" }, "items": [ { "metadata": { "name": "arm-master-1", "selfLink": "/apis/metrics.k8s.io/v1beta1/nodes/arm-master-1", "creationTimestamp": "2018-01-08T15:17:09Z" }, "timestamp": "2018-01-08T15:17:00Z", "window": "1m0s", "usage": { "cpu": "384m", "memory": "935792Ki" } }, { "metadata": { "name": "arm-node-1", "selfLink": "/apis/metrics.k8s.io/v1beta1/nodes/arm-node-1", "creationTimestamp": "2018-01-08T15:17:09Z" }, "timestamp": "2018-01-08T15:17:00Z", "window": "1m0s", "usage": { "cpu": "130m", "memory": "649020Ki" } }, { "metadata": { "name": "arm-node-2", "selfLink": "/apis/metrics.k8s.io/v1beta1/nodes/arm-node-2", "creationTimestamp": "2018-01-08T15:17:09Z" }, "timestamp": "2018-01-08T15:17:00Z", "window": "1m0s", "usage": { "cpu": "120m", "memory": "614180Ki" } } ] }
LetâÂÂs define a HPA that will maintain a minimum of two replicas and will scale up to ten if the CPU average is over 80% or if the memory goes over 200Mi
apiVersion: autoscaling/v2beta1 kind: HorizontalPodAutoscaler metadata: name: podinfo spec: scaleTargetRef: apiVersion: apps/v1beta1 kind: Deployment name: podinfo minReplicas: 2 maxReplicas: 10 metrics: - type: Resource resource: name: cpu targetAverageUtilization: 80 - type: Resource resource: name: memory targetAverageValue: 200Mi
Apply the podinfo HPA:
$ kubectl --kubeconfig terraform output kubectl_config) \ apply -f httpsraw.githubusercontent.com/stefanprodan/k8s-podinfo/master/deploy/auto-scaling/podinfo-hpa.yaml horizontalpodautoscaler "podinfo" created
After a couple of seconds the HPA controller will contact the metrics server and will fetch the CPU and memory usage:
$ kubectl --kubeconfig terraform output kubectl_config) get hpa NAME REFERENCE TARGETS MINPODS MAXPODS REPLICAS AGE podinfo Deployment/podinfo 2826240 / 200Mi, 15% / 80% 2 10 2 5m
In order to increase the CPU usage we could run a load test with hey:
#install hey go get -u github.com/rakyll/hey #do 10K requests rate limited at 20 QPS hey -n 10000 -q 10 -c 5 httpterraform output k8s_master_public_ip):31190
You can monitor the autoscaler events with:
$ kubectl --kubeconfig terraform output kubectl_config) describe hpa Events: Type Reason Age From MessageNormal SuccessfulRescale 7m horizontal-pod-autoscaler New size: 4; reason: cpu resource utilization (percentage of request) above target Normal SuccessfulRescale 3m horizontal-pod-autoscaler New size: 8; reason: cpu resource utilization (percentage of request) above target
After the load tests finishes the autoscaler will remove replicas until the deployment reaches the initial replica count:
Events: Type Reason Age From MessageNormal SuccessfulRescale 20m horizontal-pod-autoscaler New size: 4; reason: cpu resource utilization (percentage of request) above target Normal SuccessfulRescale 16m horizontal-pod-autoscaler New size: 8; reason: cpu resource utilization (percentage of request) above target Normal SuccessfulRescale 12m horizontal-pod-autoscaler New size: 10; reason: cpu resource utilization (percentage of request) above target Normal SuccessfulRescale 6m horizontal-pod-autoscaler New size: 2; reason: All metrics below target
Conclusions
Thanks to kubeadm and Terraform, bootstrapping a Kubernetes cluster on bare-metal can be done with a single command and it takes just ten minutes to have a fully functional setup. If you have any suggestion on improving this guide please submit an issue or PR on GitHub at stefanprodan/k8s-scw-baremetal. Contributions are more than welcome!