Browse Source
Add native Vagrant support
Add native Vagrant support
This allows you to simply run `vagrant up` to get a 3 node HA cluster. * Creates a dynamic inventory and uses the inventory/group_vars/all.yml * commented lines in inventory.example so that ansible doesn't try to use it. * added requirements.txt to give easy way to install ansible/ipaddr * added gitignore files to stop attempts to save unwated files * changed `Check if kube-system exists` to `failed_when: false` instead of `ignore_errors`pull/227/head
Paul Czarkowski
8 years ago
7 changed files with 165 additions and 30 deletions
Split View
Diff Options
-
3.gitignore
-
2README.md
-
124Vagrantfile
-
58inventory/inventory.example
-
2requirements.txt
-
2roles/kubernetes/master/tasks/main.yml
-
4roles/kubernetes/secrets/files/.gitignore
@ -0,0 +1,3 @@ |
|||
.vagrant |
|||
*.retry |
|||
inventory/vagrant_ansible_inventory |
@ -0,0 +1,124 @@ |
|||
# -*- mode: ruby -*- |
|||
# # vi: set ft=ruby : |
|||
|
|||
require 'fileutils' |
|||
|
|||
Vagrant.require_version ">= 1.8.0" |
|||
|
|||
CONFIG = File.join(File.dirname(__FILE__), "vagrant/config.rb") |
|||
|
|||
# Defaults for config options defined in CONFIG |
|||
$num_instances = 3 |
|||
$instance_name_prefix = "k8s" |
|||
$vm_gui = false |
|||
$vm_memory = 1024 |
|||
$vm_cpus = 1 |
|||
$shared_folders = {} |
|||
$forwarded_ports = {} |
|||
$subnet = "172.17.8" |
|||
|
|||
host_vars = {} |
|||
|
|||
if File.exist?(CONFIG) |
|||
require CONFIG |
|||
end |
|||
|
|||
# if $inventory is not set, try to use example |
|||
$inventory = File.join(File.dirname(__FILE__), "inventory") if ! $inventory |
|||
|
|||
# if $inventory has a hosts file use it, otherwise copy over vars etc |
|||
# to where vagrant expects dynamic inventory to be. |
|||
if ! File.exist?(File.join(File.dirname($inventory), "hosts")) |
|||
$vagrant_ansible = File.join(File.dirname(__FILE__), ".vagrant", |
|||
"provisioners", "ansible") |
|||
FileUtils.mkdir_p($vagrant_ansible) if ! File.exist?($vagrant_ansible) |
|||
if ! File.exist?(File.join($vagrant_ansible,"inventory")) |
|||
FileUtils.ln_s($inventory, $vagrant_ansible) |
|||
end |
|||
end |
|||
|
|||
Vagrant.configure("2") do |config| |
|||
# always use Vagrants insecure key |
|||
config.ssh.insert_key = false |
|||
|
|||
config.vm.box = "ubuntu-14.04" |
|||
config.vm.box_url = "https://storage.googleapis.com/%s.release.core-os.net/amd64-usr/%s/coreos_production_vagrant.json" % [$update_channel, $image_version] |
|||
|
|||
["vmware_fusion", "vmware_workstation"].each do |vmware| |
|||
config.vm.provider vmware do |v, override| |
|||
override.vm.box_url = "https://storage.googleapis.com/%s.release.core-os.net/amd64-usr/%s/coreos_production_vagrant_vmware_fusion.json" % [$update_channel, $image_version] |
|||
end |
|||
end |
|||
|
|||
config.vm.provider :virtualbox do |v| |
|||
# On VirtualBox, we don't have guest additions or a functional vboxsf |
|||
# in CoreOS, so tell Vagrant that so it can be smarter. |
|||
v.check_guest_additions = false |
|||
v.functional_vboxsf = false |
|||
end |
|||
|
|||
# plugin conflict |
|||
if Vagrant.has_plugin?("vagrant-vbguest") then |
|||
config.vbguest.auto_update = false |
|||
end |
|||
|
|||
(1..$num_instances).each do |i| |
|||
config.vm.define vm_name = "%s-%02d" % [$instance_name_prefix, i] do |config| |
|||
config.vm.hostname = vm_name |
|||
|
|||
if $expose_docker_tcp |
|||
config.vm.network "forwarded_port", guest: 2375, host: ($expose_docker_tcp + i - 1), auto_correct: true |
|||
end |
|||
|
|||
$forwarded_ports.each do |guest, host| |
|||
config.vm.network "forwarded_port", guest: guest, host: host, auto_correct: true |
|||
end |
|||
|
|||
["vmware_fusion", "vmware_workstation"].each do |vmware| |
|||
config.vm.provider vmware do |v| |
|||
v.vmx['memsize'] = $vm_memory |
|||
v.vmx['numvcpus'] = $vm_cpus |
|||
end |
|||
end |
|||
|
|||
config.vm.provider :virtualbox do |vb| |
|||
vb.gui = $vm_gui |
|||
vb.memory = $vm_memory |
|||
vb.cpus = $vm_cpus |
|||
end |
|||
|
|||
ip = "#{$subnet}.#{i+100}" |
|||
host_vars[vm_name] = { |
|||
"ip" => ip, |
|||
"access_ip" => ip |
|||
} |
|||
config.vm.network :private_network, ip: ip |
|||
|
|||
# Only execute once the Ansible provisioner, |
|||
# when all the machines are up and ready. |
|||
if i == $num_instances |
|||
config.vm.provision "ansible" do |ansible| |
|||
ansible.playbook = "cluster.yml" |
|||
if File.exist?(File.join(File.dirname($inventory), "hosts")) |
|||
ansible.inventory_path = $inventory |
|||
end |
|||
ansible.sudo = true |
|||
ansible.limit = "all" |
|||
ansible.host_key_checking = false |
|||
ansible.raw_arguments = ["--forks=#{$num_instances}"] |
|||
ansible.host_vars = host_vars |
|||
ansible.groups = { |
|||
# The first three nodes should be etcd servers |
|||
"etcd" => ["k8s-0[1:3]"], |
|||
# The first two nodes should be masters |
|||
"kube-master" => ["k8s-0[1:2]"], |
|||
# all nodes should be kube nodes |
|||
"kube-node" => ["k8s-0[1:#{$num_instances}]"], |
|||
"k8s-cluster:children" => ["kube-master", "kube-node"], |
|||
} |
|||
end |
|||
end |
|||
|
|||
end |
|||
end |
|||
end |
@ -1,29 +1,29 @@ |
|||
[kube-master] |
|||
node1 ansible_ssh_host=10.99.0.26 |
|||
node2 ansible_ssh_host=10.99.0.27 |
|||
|
|||
[etcd] |
|||
node1 ansible_ssh_host=10.99.0.26 |
|||
node2 ansible_ssh_host=10.99.0.27 |
|||
node3 ansible_ssh_host=10.99.0.4 |
|||
|
|||
[kube-node] |
|||
node2 ansible_ssh_host=10.99.0.27 |
|||
node3 ansible_ssh_host=10.99.0.4 |
|||
node4 ansible_ssh_host=10.99.0.5 |
|||
node5 ansible_ssh_host=10.99.0.36 |
|||
node6 ansible_ssh_host=10.99.0.37 |
|||
|
|||
[paris] |
|||
node1 ansible_ssh_host=10.99.0.26 |
|||
node3 ansible_ssh_host=10.99.0.4 local_as=xxxxxxxx |
|||
node4 ansible_ssh_host=10.99.0.5 local_as=xxxxxxxx |
|||
|
|||
[new-york] |
|||
node2 ansible_ssh_host=10.99.0.27 |
|||
node5 ansible_ssh_host=10.99.0.36 local_as=xxxxxxxx |
|||
node6 ansible_ssh_host=10.99.0.37 local_as=xxxxxxxx |
|||
|
|||
[k8s-cluster:children] |
|||
kube-node |
|||
kube-master |
|||
#[kube-master] |
|||
#node1 ansible_ssh_host=10.99.0.26 |
|||
#node2 ansible_ssh_host=10.99.0.27 |
|||
# |
|||
#[etcd] |
|||
#node1 ansible_ssh_host=10.99.0.26 |
|||
#node2 ansible_ssh_host=10.99.0.27 |
|||
#node3 ansible_ssh_host=10.99.0.4 |
|||
# |
|||
#[kube-node] |
|||
#node2 ansible_ssh_host=10.99.0.27 |
|||
#node3 ansible_ssh_host=10.99.0.4 |
|||
#node4 ansible_ssh_host=10.99.0.5 |
|||
#node5 ansible_ssh_host=10.99.0.36 |
|||
#node6 ansible_ssh_host=10.99.0.37 |
|||
# |
|||
#[paris] |
|||
#node1 ansible_ssh_host=10.99.0.26 |
|||
#node3 ansible_ssh_host=10.99.0.4 local_as=xxxxxxxx |
|||
#node4 ansible_ssh_host=10.99.0.5 local_as=xxxxxxxx |
|||
# |
|||
#[new-york] |
|||
#node2 ansible_ssh_host=10.99.0.27 |
|||
#node5 ansible_ssh_host=10.99.0.36 local_as=xxxxxxxx |
|||
#node6 ansible_ssh_host=10.99.0.37 local_as=xxxxxxxx |
|||
# |
|||
#[k8s-cluster:children] |
|||
#kube-node |
|||
#kube-master |
@ -0,0 +1,2 @@ |
|||
ansible |
|||
netaddr |
@ -0,0 +1,4 @@ |
|||
tokens/*.token |
|||
tokens/known_tokens.csv |
|||
certs/*.pem |
|||
openssl.conf |
Write
Preview
Loading…
Cancel
Save