You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

117 lines
5.9 KiB

7 years ago
  1. Getting started
  2. ===============
  3. Building your own inventory
  4. ---------------------------
  5. Ansible inventory can be stored in 3 formats: YAML, JSON, or INI-like. There is
  6. an example inventory located
  7. [here](https://github.com/kubernetes-incubator/kubespray/blob/master/inventory/sample/hosts.ini).
  8. You can use an
  9. [inventory generator](https://github.com/kubernetes-incubator/kubespray/blob/master/contrib/inventory_builder/inventory.py)
  10. to create or modify an Ansible inventory. Currently, it is limited in
  11. functionality and is only used for configuring a basic Kubespray cluster inventory, but it does
  12. support creating inventory file for large clusters as well. It now supports
  13. separated ETCD and Kubernetes master roles from node role if the size exceeds a
  14. certain threshold. Run `python3 contrib/inventory_builder/inventory.py help` help for more information.
  15. Example inventory generator usage:
  16. ```
  17. cp -r inventory/sample inventory/mycluster
  18. declare -a IPS=(10.10.1.3 10.10.1.4 10.10.1.5)
  19. CONFIG_FILE=inventory/mycluster/hosts.ini python3 contrib/inventory_builder/inventory.py ${IPS[@]}
  20. ```
  21. Starting custom deployment
  22. --------------------------
  23. Once you have an inventory, you may want to customize deployment data vars
  24. and start the deployment:
  25. **IMPORTANT: Edit my_inventory/groups_vars/*.yaml to override data vars**
  26. ```
  27. ansible-playbook -i inventory/mycluster/hosts.ini cluster.yml -b -v \
  28. --private-key=~/.ssh/private_key
  29. ```
  30. See more details in the [ansible guide](ansible.md).
  31. Adding nodes
  32. ------------
  33. You may want to add **worker** nodes to your existing cluster. This can be done by re-running the `cluster.yml` playbook, or you can target the bare minimum needed to get kubelet installed on the worker and talking to your masters. This is especially helpful when doing something like autoscaling your clusters.
  34. - Add the new worker node to your inventory under kube-node (or utilize a [dynamic inventory](https://docs.ansible.com/ansible/intro_dynamic_inventory.html)).
  35. - Run the ansible-playbook command, substituting `scale.yml` for `cluster.yml`:
  36. ```
  37. ansible-playbook -i inventory/mycluster/hosts.ini scale.yml -b -v \
  38. --private-key=~/.ssh/private_key
  39. ```
  40. Remove nodes
  41. ------------
  42. You may want to remove **worker** nodes to your existing cluster. This can be done by re-running the `remove-node.yml` playbook. First, all nodes will be drained, then stop some kubernetes services and delete some certificates, and finally execute the kubectl command to delete these nodes. This can be combined with the add node function, This is generally helpful when doing something like autoscaling your clusters. Of course if a node is not working, you can remove the node and install it again.
  43. - Add worker nodes to the list under kube-node if you want to delete them (or utilize a [dynamic inventory](https://docs.ansible.com/ansible/intro_dynamic_inventory.html)).
  44. - Run the ansible-playbook command, substituting `remove-node.yml`:
  45. ```
  46. ansible-playbook -i inventory/mycluster/hosts.ini remove-node.yml -b -v \
  47. --private-key=~/.ssh/private_key
  48. ```
  49. Connecting to Kubernetes
  50. ------------------------
  51. By default, Kubespray configures kube-master hosts with insecure access to
  52. kube-apiserver via port 8080. A kubeconfig file is not necessary in this case,
  53. because kubectl will use http://localhost:8080 to connect. The kubeconfig files
  54. generated will point to localhost (on kube-masters) and kube-node hosts will
  55. connect either to a localhost nginx proxy or to a loadbalancer if configured.
  56. More details on this process are in the [HA guide](ha-mode.md).
  57. Kubespray permits connecting to the cluster remotely on any IP of any
  58. kube-master host on port 6443 by default. However, this requires
  59. authentication. One could generate a kubeconfig based on one installed
  60. kube-master hosts (needs improvement) or connect with a username and password.
  61. By default, a user with admin rights is created, named `kube`.
  62. The password can be viewed after deployment by looking at the file
  63. `PATH_TO_KUBESPRAY/credentials/kube_user`. This contains a randomly generated
  64. password. If you wish to set your own password, just precreate/modify this
  65. file yourself.
  66. For more information on kubeconfig and accessing a Kubernetes cluster, refer to
  67. the Kubernetes [documentation](https://kubernetes.io/docs/tasks/access-application-cluster/configure-access-multiple-clusters/).
  68. Accessing Kubernetes Dashboard
  69. ------------------------------
  70. As of kubernetes-dashboard v1.7.x:
  71. * New login options that use apiserver auth proxying of token/basic/kubeconfig by default
  72. * Requires RBAC in authorization_modes
  73. * Only serves over https
  74. * No longer available at https://first_master:6443/ui until apiserver is updated with the https proxy URL
  75. If the variable `dashboard_enabled` is set (default is true), then you can access the Kubernetes Dashboard at the following URL, You will be prompted for credentials:
  76. https://first_master:6443/api/v1/namespaces/kube-system/services/https:kubernetes-dashboard:/proxy/#!/login
  77. Or you can run 'kubectl proxy' from your local machine to access dashboard in your browser from:
  78. http://localhost:8001/api/v1/namespaces/kube-system/services/https:kubernetes-dashboard:/proxy/#!/login
  79. It is recommended to access dashboard from behind a gateway (like Ingress Controller) that enforces an authentication token. Details and other access options here: https://github.com/kubernetes/dashboard/wiki/Accessing-Dashboard---1.7.X-and-above
  80. Accessing Kubernetes API
  81. ------------------------
  82. The main client of Kubernetes is `kubectl`. It is installed on each kube-master
  83. host and can optionally be configured on your ansible host by setting
  84. `kubeconfig_localhost: true` in the configuration. If enabled, kubectl and
  85. admin.conf will appear in the artifacts/ directory after deployment. You can
  86. see a list of nodes by running the following commands:
  87. cd artifacts/
  88. ./kubectl --kubeconfig admin.conf get nodes
  89. If desired, copy kubectl to your bin dir and admin.conf to ~/.kube/config.