You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

128 lines
6.1 KiB

7 years ago
  1. # Getting started
  2. ## Building your own inventory
  3. Ansible inventory can be stored in 3 formats: YAML, JSON, or INI-like. There is
  4. an example inventory located
  5. [here](https://github.com/kubernetes-sigs/kubespray/blob/master/inventory/sample/inventory.ini).
  6. You can use an
  7. [inventory generator](https://github.com/kubernetes-sigs/kubespray/blob/master/contrib/inventory_builder/inventory.py)
  8. to create or modify an Ansible inventory. Currently, it is limited in
  9. functionality and is only used for configuring a basic Kubespray cluster inventory, but it does
  10. support creating inventory file for large clusters as well. It now supports
  11. separated ETCD and Kubernetes master roles from node role if the size exceeds a
  12. certain threshold. Run `python3 contrib/inventory_builder/inventory.py help` help for more information.
  13. Example inventory generator usage:
  14. ```ShellSession
  15. cp -r inventory/sample inventory/mycluster
  16. declare -a IPS=(10.10.1.3 10.10.1.4 10.10.1.5)
  17. CONFIG_FILE=inventory/mycluster/hosts.yml python3 contrib/inventory_builder/inventory.py ${IPS[@]}
  18. ```
  19. Then use `inventory/mycluster/hosts.yml` as inventory file.
  20. ## Starting custom deployment
  21. Once you have an inventory, you may want to customize deployment data vars
  22. and start the deployment:
  23. **IMPORTANT**: Edit my\_inventory/groups\_vars/\*.yaml to override data vars:
  24. ```ShellSession
  25. ansible-playbook -i inventory/mycluster/hosts.yml cluster.yml -b -v \
  26. --private-key=~/.ssh/private_key
  27. ```
  28. See more details in the [ansible guide](ansible.md).
  29. ### Adding nodes
  30. You may want to add worker, master or etcd nodes to your existing cluster. This can be done by re-running the `cluster.yml` playbook, or you can target the bare minimum needed to get kubelet installed on the worker and talking to your masters. This is especially helpful when doing something like autoscaling your clusters.
  31. - Add the new worker node to your inventory in the appropriate group (or utilize a [dynamic inventory](https://docs.ansible.com/ansible/intro_dynamic_inventory.html)).
  32. - Run the ansible-playbook command, substituting `cluster.yml` for `scale.yml`:
  33. ```ShellSession
  34. ansible-playbook -i inventory/mycluster/hosts.yml scale.yml -b -v \
  35. --private-key=~/.ssh/private_key
  36. ```
  37. ### Remove nodes
  38. You may want to remove **master**, **worker**, or **etcd** nodes from your
  39. existing cluster. This can be done by re-running the `remove-node.yml`
  40. playbook. First, all specified nodes will be drained, then stop some
  41. kubernetes services and delete some certificates,
  42. and finally execute the kubectl command to delete these nodes.
  43. This can be combined with the add node function. This is generally helpful
  44. when doing something like autoscaling your clusters. Of course, if a node
  45. is not working, you can remove the node and install it again.
  46. Use `--extra-vars "node=<nodename>,<nodename2>"` to select the node(s) you want to delete.
  47. ```ShellSession
  48. ansible-playbook -i inventory/mycluster/hosts.yml remove-node.yml -b -v \
  49. --private-key=~/.ssh/private_key \
  50. --extra-vars "node=nodename,nodename2"
  51. ```
  52. If a node is completely unreachable by ssh, add `--extra-vars reset_nodes=no`
  53. to skip the node reset step. If one node is unavailable, but others you wish
  54. to remove are able to connect via SSH, you could set reset_nodes=no as a host
  55. var in inventory.
  56. ## Connecting to Kubernetes
  57. By default, Kubespray configures kube-master hosts with insecure access to
  58. kube-apiserver via port 8080. A kubeconfig file is not necessary in this case,
  59. because kubectl will use <http://localhost:8080> to connect. The kubeconfig files
  60. generated will point to localhost (on kube-masters) and kube-node hosts will
  61. connect either to a localhost nginx proxy or to a loadbalancer if configured.
  62. More details on this process are in the [HA guide](ha-mode.md).
  63. Kubespray permits connecting to the cluster remotely on any IP of any
  64. kube-master host on port 6443 by default. However, this requires
  65. authentication. One can get a kubeconfig from kube-master hosts
  66. (see [below](#accessing-kubernetes-api)) or connect with a [username and password](vars.md#user-accounts).
  67. For more information on kubeconfig and accessing a Kubernetes cluster, refer to
  68. the Kubernetes [documentation](https://kubernetes.io/docs/tasks/access-application-cluster/configure-access-multiple-clusters/).
  69. ## Accessing Kubernetes Dashboard
  70. As of kubernetes-dashboard v1.7.x:
  71. - New login options that use apiserver auth proxying of token/basic/kubeconfig by default
  72. - Requires RBAC in authorization\_modes
  73. - Only serves over https
  74. - No longer available at <https://first_master:6443/ui> until apiserver is updated with the https proxy URL
  75. If the variable `dashboard_enabled` is set (default is true), then you can access the Kubernetes Dashboard at the following URL, You will be prompted for credentials:
  76. <https://first_master:6443/api/v1/namespaces/kube-system/services/https:kubernetes-dashboard:/proxy/#!/login>
  77. Or you can run 'kubectl proxy' from your local machine to access dashboard in your browser from:
  78. <http://localhost:8001/api/v1/namespaces/kube-system/services/https:kubernetes-dashboard:/proxy/#!/login>
  79. It is recommended to access dashboard from behind a gateway (like Ingress Controller) that enforces an authentication token. Details and other access options here: <https://github.com/kubernetes/dashboard/wiki/Accessing-Dashboard---1.7.X-and-above>
  80. ## Accessing Kubernetes API
  81. The main client of Kubernetes is `kubectl`. It is installed on each kube-master
  82. host and can optionally be configured on your ansible host by setting
  83. `kubectl_localhost: true` and `kubeconfig_localhost: true` in the configuration:
  84. - If `kubectl_localhost` enabled, `kubectl` will download onto `/usr/local/bin/` and setup with bash completion. A helper script `inventory/mycluster/artifacts/kubectl.sh` also created for setup with below `admin.conf`.
  85. - If `kubeconfig_localhost` enabled `admin.conf` will appear in the `inventory/mycluster/artifacts/` directory after deployment.
  86. - The location where these files are downloaded to can be configured via the `artifacts_dir` variable.
  87. You can see a list of nodes by running the following commands:
  88. ```ShellSession
  89. cd inventory/mycluster/artifacts
  90. ./kubectl.sh get nodes
  91. ```
  92. If desired, copy admin.conf to ~/.kube/config.