You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
 

814 lines
23 KiB

stages:
- deploy-do
- moderator
- unit-tests
- deploy-gce-part1
- deploy-gce-part2
- deploy-gce-special
variables:
FAILFASTCI_NAMESPACE: 'kargo-ci'
# DOCKER_HOST: tcp://localhost:2375
ANSIBLE_FORCE_COLOR: "true"
MAGIC: "ci check this"
TEST_ID: "$CI_PIPELINE_ID-$CI_BUILD_ID"
CI_TEST_VARS: "./tests/files/${CI_JOB_NAME}.yml"
GS_ACCESS_KEY_ID: $GS_KEY
GS_SECRET_ACCESS_KEY: $GS_SECRET
CONTAINER_ENGINE: docker
SSH_USER: $GCE_USER
GCE_PREEMPTIBLE: "false"
ANSIBLE_KEEP_REMOTE_FILES: "1"
ANSIBLE_CONFIG: ./tests/ansible.cfg
IDEMPOT_CHECK: "false"
RESET_CHECK: "false"
UPGRADE_TEST: "false"
KUBEADM_ENABLED: "false"
LOG_LEVEL: "-vvv"
# asia-east1-a
# asia-northeast1-a
# europe-west1-b
# us-central1-a
# us-east1-b
# us-west1-a
before_script:
- pip install -r tests/requirements.txt
- mkdir -p /.ssh
.job: &job
tags:
- kubernetes
- docker
image: quay.io/kubespray/kubespray:master
.docker_service: &docker_service
services:
- docker:dind
.create_cluster: &create_cluster
<<: *job
<<: *docker_service
.gce_variables: &gce_variables
GCE_USER: travis
SSH_USER: $GCE_USER
CLOUD_MACHINE_TYPE: "g1-small"
CI_PLATFORM: "gce"
PRIVATE_KEY: $GCE_PRIVATE_KEY
.do_variableS: &do_variables
PRIVATE_KEY: $DO_PRIVATE_KEY
CI_PLATFORM: "do"
.testcases: &testcases
<<: *job
<<: *docker_service
cache:
key: "$CI_BUILD_REF_NAME"
paths:
- downloads/
- $HOME/.cache
before_script:
- docker info
- pip install -r tests/requirements.txt
- mkdir -p /.ssh
- mkdir -p $HOME/.ssh
- ansible-playbook --version
- export PYPATH=$([[ ! "$CI_JOB_NAME" =~ "coreos" ]] && echo /usr/bin/python || echo /opt/bin/python)
- echo "CI_JOB_NAME is $CI_JOB_NAME"
- echo "PYPATH is $PYPATH"
script:
- pwd
- ls
- echo ${PWD}
- echo "${STARTUP_SCRIPT}"
- cd tests && make create-${CI_PLATFORM} -s ; cd -
# Check out latest tag if testing upgrade
# Uncomment when gitlab kargo repo has tags
#- test "${UPGRADE_TEST}" != "false" && git fetch --all && git checkout $(git describe --tags $(git rev-list --tags --max-count=1))
- test "${UPGRADE_TEST}" != "false" && git checkout ba0a03a8ba2d97a73d06242ec4bb3c7e2012e58c
# Checkout the CI vars file so it is available
- test "${UPGRADE_TEST}" != "false" && git checkout "${CI_BUILD_REF}" tests/files/${CI_JOB_NAME}.yml
# Workaround https://github.com/kubernetes-incubator/kubespray/issues/2021
- 'sh -c "echo ignore_assert_errors: true | tee -a tests/files/${CI_JOB_NAME}.yml"'
# Create cluster
- >
ansible-playbook
-i inventory/sample/hosts.ini
-b --become-user=root
--private-key=${HOME}/.ssh/id_rsa
-u $SSH_USER
${SSH_ARGS}
${LOG_LEVEL}
-e @${CI_TEST_VARS}
-e ansible_python_interpreter=${PYPATH}
-e ansible_ssh_user=${SSH_USER}
-e local_release_dir=${PWD}/downloads
--limit "all:!fake_hosts"
cluster.yml
# Repeat deployment if testing upgrade
- >
if [ "${UPGRADE_TEST}" != "false" ]; then
test "${UPGRADE_TEST}" == "basic" && PLAYBOOK="cluster.yml";
test "${UPGRADE_TEST}" == "graceful" && PLAYBOOK="upgrade-cluster.yml";
git checkout "${CI_BUILD_REF}";
ansible-playbook
-i inventory/sample/hosts.ini
-b --become-user=root
--private-key=${HOME}/.ssh/id_rsa
-u $SSH_USER
${SSH_ARGS}
${LOG_LEVEL}
-e @${CI_TEST_VARS}
-e ansible_python_interpreter=${PYPATH}
-e ansible_ssh_user=${SSH_USER}
-e local_release_dir=${PWD}/downloads
--limit "all:!fake_hosts"
$PLAYBOOK;
fi
# Tests Cases
## Test Master API
- >
ansible-playbook -i inventory/sample/hosts.ini -e ansible_python_interpreter=${PYPATH} -u $SSH_USER -e ansible_ssh_user=$SSH_USER $SSH_ARGS -b --become-user=root --limit "all:!fake_hosts" tests/testcases/010_check-apiserver.yml $LOG_LEVEL
-e "{kubeadm_enabled: ${KUBEADM_ENABLED}}"
## Ping the between 2 pod
- ansible-playbook -i inventory/sample/hosts.ini -e ansible_python_interpreter=${PYPATH} -u $SSH_USER -e ansible_ssh_user=$SSH_USER $SSH_ARGS -b --become-user=root --limit "all:!fake_hosts" tests/testcases/030_check-network.yml $LOG_LEVEL
## Advanced DNS checks
- ansible-playbook -i inventory/sample/hosts.ini -e ansible_python_interpreter=${PYPATH} -u $SSH_USER -e ansible_ssh_user=$SSH_USER $SSH_ARGS -b --become-user=root --limit "all:!fake_hosts" tests/testcases/040_check-network-adv.yml $LOG_LEVEL
## Idempotency checks 1/5 (repeat deployment)
- >
if [ "${IDEMPOT_CHECK}" = "true" ]; then
ansible-playbook
-i inventory/sample/hosts.ini
-b --become-user=root
--private-key=${HOME}/.ssh/id_rsa
-u $SSH_USER
${SSH_ARGS}
${LOG_LEVEL}
-e @${CI_TEST_VARS}
-e ansible_python_interpreter=${PYPATH}
-e local_release_dir=${PWD}/downloads
--limit "all:!fake_hosts"
cluster.yml;
fi
## Idempotency checks 2/5 (Advanced DNS checks)
- >
if [ "${IDEMPOT_CHECK}" = "true" ]; then
ansible-playbook
-i inventory/sample/hosts.ini
-b --become-user=root
--private-key=${HOME}/.ssh/id_rsa
-u $SSH_USER
${SSH_ARGS}
${LOG_LEVEL}
-e @${CI_TEST_VARS}
--limit "all:!fake_hosts"
tests/testcases/040_check-network-adv.yml $LOG_LEVEL;
fi
## Idempotency checks 3/5 (reset deployment)
- >
if [ "${IDEMPOT_CHECK}" = "true" -a "${RESET_CHECK}" = "true" ]; then
ansible-playbook
-i inventory/sample/hosts.ini
-b --become-user=root
--private-key=${HOME}/.ssh/id_rsa
-u $SSH_USER
${SSH_ARGS}
${LOG_LEVEL}
-e @${CI_TEST_VARS}
-e ansible_python_interpreter=${PYPATH}
-e reset_confirmation=yes
--limit "all:!fake_hosts"
reset.yml;
fi
## Idempotency checks 4/5 (redeploy after reset)
- >
if [ "${IDEMPOT_CHECK}" = "true" -a "${RESET_CHECK}" = "true" ]; then
ansible-playbook
-i inventory/sample/hosts.ini
-b --become-user=root
--private-key=${HOME}/.ssh/id_rsa
-u $SSH_USER
${SSH_ARGS}
${LOG_LEVEL}
-e @${CI_TEST_VARS}
-e ansible_python_interpreter=${PYPATH}
-e local_release_dir=${PWD}/downloads
--limit "all:!fake_hosts"
cluster.yml;
fi
## Idempotency checks 5/5 (Advanced DNS checks)
- >
if [ "${IDEMPOT_CHECK}" = "true" -a "${RESET_CHECK}" = "true" ]; then
ansible-playbook -i inventory/sample/hosts.ini -e ansible_python_interpreter=${PYPATH}
-u $SSH_USER -e ansible_ssh_user=$SSH_USER $SSH_ARGS -b --become-user=root
--limit "all:!fake_hosts"
tests/testcases/040_check-network-adv.yml $LOG_LEVEL;
fi
after_script:
- cd tests && make delete-${CI_PLATFORM} -s ; cd -
.gce: &gce
<<: *testcases
variables:
<<: *gce_variables
.do: &do
variables:
<<: *do_variables
<<: *testcases
# .do: &do
# <<: *job
# <<: *gce
# cache:
# key: "$CI_BUILD_REF_NAME"
# paths:
# - downloads/
# - $HOME/.cache
# before_script:
# - docker info
# - pip install ansible==2.2.1.0
# - pip install netaddr
# - pip install apache-libcloud==0.20.1
# - pip install boto==2.9.0
# - pip install dopy==0.3.5
# - mkdir -p /.ssh
# - mkdir -p $HOME/.ssh
# - echo $DO_PRIVATE_KEY | base64 -d > $HOME/.ssh/id_rsa
# - chmod 400 $HOME/.ssh/id_rsa
# - md5sum ~/.ssh/id_rsa
# - ansible-playbook --version
# - export PYPATH=$([ $BOOTSTRAP_OS = coreos ] && echo /opt/bin/python || echo /usr/bin/python )
# - echo $PYPATH
# script:
# - pwd
# - ls
# - echo ${PWD}
# - >
# ansible-playbook tests/cloud_playbooks/create-do.yml -i tests/local_inventory/hosts.cfg -c local
# ${LOG_LEVEL}
# -e cloud_image=${CLOUD_IMAGE}
# -e cloud_region=${CLOUD_REGION}
# -e inventory_path=${PWD}/inventory/inventory.ini
# -e kube_network_plugin=${KUBE_NETWORK_PLUGIN}
# -e mode=${CLUSTER_MODE}
# -e test_id=${TEST_ID}
# # Check out latest tag if testing upgrade
# # Uncomment when gitlab kargo repo has tags
# #- test "${UPGRADE_TEST}" = "true" && git fetch --all && git checkout $(git describe --tags $(git rev-list --tags --max-count=1))
# - test "${UPGRADE_TEST}" = "true" && git checkout 031cf565ec3ccd3ebbe80eeef3454c3780e5c598 && pip install ansible==2.2.0
# # Create cluster
# - >
# ansible-playbook -i inventory/inventory.ini -b --become-user=root --private-key=${HOME}/.ssh/id_rsa -u root
# ${SSH_ARGS}
# ${LOG_LEVEL}
# -e state=present
# -e ansible_python_interpreter=${PYPATH}
# -e ansible_ssh_user=root
# -e bootstrap_os=${BOOTSTRAP_OS}
# -e cert_management=${CERT_MGMT:-script}
# -e cloud_provider=gce
# -e deploy_netchecker=true
# -e download_localhost=true
# -e download_run_once=true
# -e etcd_deployment_type=${ETCD_DEPLOYMENT}
# -e kube_network_plugin=${KUBE_NETWORK_PLUGIN}
# -e kubelet_deployment_type=${KUBELET_DEPLOYMENT}
# -e local_release_dir=${PWD}/downloads
# -e resolvconf_mode=${RESOLVCONF_MODE}
# -e vault_deployment_type=${VAULT_DEPLOYMENT}
# cluster.yml
# # Repeat deployment if testing upgrade
# #FIXME(mattymo): repeat "Create cluster" above without duplicating code
# - >
# if [ "${UPGRADE_TEST}" = "true" ]; then
# pip install ansible==2.2.1.0;
# git checkout "${CI_BUILD_REF}";
# ansible-playbook -i inventory/inventory.ini -b --become-user=root --private-key=${HOME}/.ssh/id_rsa -u root
# ${SSH_ARGS}
# ${LOG_LEVEL}
# -e ansible_python_interpreter=${PYPATH}
# -e ansible_ssh_user=root
# -e bootstrap_os=${BOOTSTRAP_OS}
# -e cloud_provider=gce
# -e deploy_netchecker=true
# -e download_localhost=true
# -e download_run_once=true
# -e etcd_deployment_type=${ETCD_DEPLOYMENT}
# -e kube_network_plugin=${KUBE_NETWORK_PLUGIN}
# -e kubelet_deployment_type=${KUBELET_DEPLOYMENT}
# -e local_release_dir=${PWD}/downloads
# -e resolvconf_mode=${RESOLVCONF_MODE}
# -e weave_cpu_requests=${WEAVE_CPU_LIMIT}
# -e weave_cpu_limit=${WEAVE_CPU_LIMIT}
# cluster.yml;
# fi
# # Tests Cases
# ## Test Master API
# - ansible-playbook -i inventory/inventory.ini -e ansible_python_interpreter=${PYPATH} -u root -e ansible_ssh_user=root $SSH_ARGS -b --become-user=root tests/testcases/010_check-apiserver.yml $LOG_LEVEL
# ## Ping the between 2 pod
# - ansible-playbook -i inventory/inventory.ini -e ansible_python_interpreter=${PYPATH} -u root -e ansible_ssh_user=root $SSH_ARGS -b --become-user=root tests/testcases/030_check-network.yml $LOG_LEVEL
# ## Advanced DNS checks
# - ansible-playbook -i inventory/inventory.ini -e ansible_python_interpreter=${PYPATH} -u root -e ansible_ssh_user=root $SSH_ARGS -b --become-user=root tests/testcases/040_check-network-adv.yml $LOG_LEVEL
# ## Idempotency checks 1/5 (repeat deployment)
# - >
# if [ "${IDEMPOT_CHECK}" = "true" ]; then
# ansible-playbook -i inventory/inventory.ini -u root -e ansible_ssh_user=root $SSH_ARGS
# -b --become-user=root -e cloud_provider=gce $LOG_LEVEL -e kube_network_plugin=${KUBE_NETWORK_PLUGIN}
# --private-key=${HOME}/.ssh/id_rsa
# -e bootstrap_os=${BOOTSTRAP_OS}
# -e ansible_python_interpreter=${PYPATH}
# -e download_run_once=true
# -e download_localhost=true
# -e deploy_netchecker=true
# -e resolvconf_mode=${RESOLVCONF_MODE}
# -e local_release_dir=${PWD}/downloads
# -e etcd_deployment_type=${ETCD_DEPLOYMENT}
# -e kubelet_deployment_type=${KUBELET_DEPLOYMENT}
# cluster.yml;
# fi
# ## Idempotency checks 2/5 (Advanced DNS checks)
# - >
# if [ "${IDEMPOT_CHECK}" = "true" ]; then
# ansible-playbook -i inventory/inventory.ini -e ansible_python_interpreter=${PYPATH}
# -u root -e ansible_ssh_user=root $SSH_ARGS -b --become-user=root
# tests/testcases/040_check-network-adv.yml $LOG_LEVEL;
# fi
# ## Idempotency checks 3/5 (reset deployment)
# - >
# if [ "${IDEMPOT_CHECK}" = "true" ]; then
# ansible-playbook -i inventory/inventory.ini -u root -e ansible_ssh_user=root $SSH_ARGS
# -b --become-user=root -e cloud_provider=gce $LOG_LEVEL -e kube_network_plugin=${KUBE_NETWORK_PLUGIN}
# --private-key=${HOME}/.ssh/id_rsa
# -e bootstrap_os=${BOOTSTRAP_OS}
# -e ansible_python_interpreter=${PYPATH}
# reset.yml;
# fi
# ## Idempotency checks 4/5 (redeploy after reset)
# - >
# if [ "${IDEMPOT_CHECK}" = "true" ]; then
# ansible-playbook -i inventory/inventory.ini -u root -e ansible_ssh_user=root $SSH_ARGS
# -b --become-user=root -e cloud_provider=gce $LOG_LEVEL -e kube_network_plugin=${KUBE_NETWORK_PLUGIN}
# --private-key=${HOME}/.ssh/id_rsa
# -e bootstrap_os=${BOOTSTRAP_OS}
# -e ansible_python_interpreter=${PYPATH}
# -e download_run_once=true
# -e download_localhost=true
# -e deploy_netchecker=true
# -e resolvconf_mode=${RESOLVCONF_MODE}
# -e local_release_dir=${PWD}/downloads
# -e etcd_deployment_type=${ETCD_DEPLOYMENT}
# -e kubelet_deployment_type=${KUBELET_DEPLOYMENT}
# cluster.yml;
# fi
# ## Idempotency checks 5/5 (Advanced DNS checks)
# - >
# if [ "${IDEMPOT_CHECK}" = "true" ]; then
# ansible-playbook -i inventory/inventory.ini -e ansible_python_interpreter=${PYPATH}
# -u root -e ansible_ssh_user=root $SSH_ARGS -b --become-user=root
# tests/testcases/040_check-network-adv.yml $LOG_LEVEL;
# fi
# after_script:
# - >
# ansible-playbook -i inventory/inventory.ini tests/cloud_playbooks/create-do.yml -c local $LOG_LEVEL
# -e state=absent
# -e mode=${CLUSTER_MODE}
# -e test_id=${TEST_ID}
# -e kube_network_plugin=${KUBE_NETWORK_PLUGIN}
# -e gce_project_id=${GCE_PROJECT_ID}
# -e gce_service_account_email=${GCE_ACCOUNT}
# -e gce_credentials_file=${HOME}/.ssh/gce.json
# -e cloud_image=${CLOUD_IMAGE}
# -e inventory_path=${PWD}/inventory/inventory.ini
# -e cloud_region=${CLOUD_REGION}
# Test matrix. Leave the comments for markup scripts.
.coreos_calico_aio_variables: &coreos_calico_aio_variables
# stage: deploy-gce-part1
MOVED_TO_GROUP_VARS: "true"
.ubuntu_canal_ha_variables: &ubuntu_canal_ha_variables
# stage: deploy-gce-part1
UPGRADE_TEST: "graceful"
.centos_weave_kubeadm_variables: &centos_weave_kubeadm_variables
# stage: deploy-gce-part1
UPGRADE_TEST: "graceful"
.ubuntu_canal_kubeadm_variables: &ubuntu_canal_kubeadm_variables
# stage: deploy-gce-part1
MOVED_TO_GROUP_VARS: "true"
.ubuntu_contiv_sep_variables: &ubuntu_contiv_sep_variables
# stage: deploy-gce-special
MOVED_TO_GROUP_VARS: "true"
.rhel7_weave_variables: &rhel7_weave_variables
# stage: deploy-gce-part1
MOVED_TO_GROUP_VARS: "true"
.centos7_flannel_addons_variables: &centos7_flannel_addons_variables
# stage: deploy-gce-part2
MOVED_TO_GROUP_VARS: "true"
.debian8_calico_variables: &debian8_calico_variables
# stage: deploy-gce-part2
MOVED_TO_GROUP_VARS: "true"
.coreos_canal_variables: &coreos_canal_variables
# stage: deploy-gce-part2
MOVED_TO_GROUP_VARS: "true"
.rhel7_canal_sep_variables: &rhel7_canal_sep_variables
# stage: deploy-gce-special
MOVED_TO_GROUP_VARS: "true"
.ubuntu_weave_sep_variables: &ubuntu_weave_sep_variables
# stage: deploy-gce-special
MOVED_TO_GROUP_VARS: "true"
.centos7_calico_ha_variables: &centos7_calico_ha_variables
# stage: deploy-gce-special
MOVED_TO_GROUP_VARS: "true"
.coreos_alpha_weave_ha_variables: &coreos_alpha_weave_ha_variables
# stage: deploy-gce-special
MOVED_TO_GROUP_VARS: "true"
.ubuntu_rkt_sep_variables: &ubuntu_rkt_sep_variables
# stage: deploy-gce-part1
MOVED_TO_GROUP_VARS: "true"
.ubuntu_vault_sep_variables: &ubuntu_vault_sep_variables
# stage: deploy-gce-part1
MOVED_TO_GROUP_VARS: "true"
.ubuntu_flannel_variables: &ubuntu_flannel_variables
# stage: deploy-gce-special
MOVED_TO_GROUP_VARS: "true"
# Builds for PRs only (premoderated by unit-tests step) and triggers (auto)
coreos-calico-aio:
stage: deploy-do
# stage: deploy-gce-part1
<<: *job
<<: *gce
variables:
<<: *coreos_calico_aio_variables
<<: *gce_variables
when: on_success
except: ['triggers']
# only: [/^pr-.*$/]
ubuntu-canal-ha-do:
stage: deploy-do
<<: *job
<<: *do
variables:
<<: *do_variables
when: on_success
except: ['triggers']
# only: [/^pr-.*$/]
# coreos-calico-sep-triggers:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *coreos_calico_aio_variables
# when: on_success
# only: ['triggers']
# centos7-flannel-addons:
# stage: deploy-gce-part2
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *centos7_flannel_addons_variables
# when: on_success
# except: ['triggers']
# only: [/^pr-.*$/]
# centos7-flannel-addons-triggers:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *centos7_flannel_addons_variables
# when: on_success
# only: ['triggers']
# ubuntu-weave-sep:
# stage: deploy-gce-special
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *ubuntu_weave_sep_variables
# when: on_success
# except: ['triggers']
# only: [/^pr-.*$/]
# ubuntu-weave-sep-triggers:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *ubuntu_weave_sep_variables
# when: on_success
# only: ['triggers']
# # More builds for PRs/merges (manual) and triggers (auto)
# ubuntu-canal-ha:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *ubuntu_canal_ha_variables
# when: manual
# except: ['triggers']
# only: ['master', /^pr-.*$/]
# ubuntu-canal-ha-triggers:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *ubuntu_canal_ha_variables
# when: on_success
# only: ['triggers']
# ubuntu-canal-kubeadm:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *ubuntu_canal_kubeadm_variables
# when: manual
# except: ['triggers']
# only: ['master', /^pr-.*$/]
# ubuntu-canal-kubeadm-triggers:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *ubuntu_canal_kubeadm_variables
# when: on_success
# only: ['triggers']
# centos-weave-kubeadm:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *centos_weave_kubeadm_variables
# when: manual
# except: ['triggers']
# only: ['master', /^pr-.*$/]
# centos-weave-kubeadm-triggers:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *centos_weave_kubeadm_variables
# when: on_success
# only: ['triggers']
# ubuntu-contiv-sep:
# stage: deploy-gce-special
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *ubuntu_contiv_sep_variables
# when: manual
# except: ['triggers']
# only: ['master', /^pr-.*$/]
# rhel7-weave:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *rhel7_weave_variables
# when: manual
# except: ['triggers']
# only: ['master', /^pr-.*$/]
# rhel7-weave-triggers:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *rhel7_weave_variables
# when: on_success
# only: ['triggers']
# debian8-calico-upgrade:
# stage: deploy-gce-part2
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *debian8_calico_variables
# when: manual
# except: ['triggers']
# only: ['master', /^pr-.*$/]
# debian8-calico-triggers:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *debian8_calico_variables
# when: on_success
# only: ['triggers']
# coreos-canal:
# stage: deploy-gce-part2
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *coreos_canal_variables
# when: manual
# except: ['triggers']
# only: ['master', /^pr-.*$/]
# coreos-canal-triggers:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *coreos_canal_variables
# when: on_success
# only: ['triggers']
# rhel7-canal-sep:
# stage: deploy-gce-special
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *rhel7_canal_sep_variables
# when: manual
# except: ['triggers']
# only: ['master', /^pr-.*$/,]
# rhel7-canal-sep-triggers:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *rhel7_canal_sep_variables
# when: on_success
# only: ['triggers']
# centos7-calico-ha:
# stage: deploy-gce-special
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *centos7_calico_ha_variables
# when: manual
# except: ['triggers']
# only: ['master', /^pr-.*$/]
# centos7-calico-ha-triggers:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *centos7_calico_ha_variables
# when: on_success
# only: ['triggers']
# # no triggers yet https://github.com/kubernetes-incubator/kargo/issues/613
# coreos-alpha-weave-ha:
# stage: deploy-gce-special
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *coreos_alpha_weave_ha_variables
# when: manual
# except: ['triggers']
# only: ['master', /^pr-.*$/]
# ubuntu-rkt-sep:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *ubuntu_rkt_sep_variables
# when: manual
# except: ['triggers']
# only: ['master', /^pr-.*$/]
# ubuntu-vault-sep:
# stage: deploy-gce-part1
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *ubuntu_vault_sep_variables
# when: manual
# except: ['triggers']
# only: ['master', /^pr-.*$/]
# ubuntu-flannel-sep:
# stage: deploy-gce-special
# <<: *job
# <<: *gce
# variables:
# <<: *gce_variables
# <<: *ubuntu_flannel_variables
# when: manual
# except: ['triggers']
# only: ['master', /^pr-.*$/]
# # Premoderated with manual actions
# ci-authorized:
# <<: *job
# stage: moderator
# before_script:
# - apt-get -y install jq
# script:
# - /bin/sh scripts/premoderator.sh
# except: ['triggers', 'master']
# syntax-check:
# <<: *job
# stage: unit-tests
# script:
# - ansible-playbook -i inventory/local-tests.cfg -u root -e ansible_ssh_user=root -b --become-user=root cluster.yml -vvv --syntax-check
# - ansible-playbook -i inventory/local-tests.cfg -u root -e ansible_ssh_user=root -b --become-user=root upgrade-cluster.yml -vvv --syntax-check
# - ansible-playbook -i inventory/local-tests.cfg -u root -e ansible_ssh_user=root -b --become-user=root reset.yml -vvv --syntax-check
# - ansible-playbook -i inventory/local-tests.cfg -u root -e ansible_ssh_user=root -b --become-user=root extra_playbooks/upgrade-only-k8s.yml -vvv --syntax-check
# except: ['triggers', 'master']
# yamllint:
# <<: *job
# stage: unit-tests
# script:
# - yamllint roles
# except: ['triggers', 'master']
# tox-inventory-builder:
# stage: unit-tests
# <<: *job
# script:
# - pip install tox
# - cd contrib/inventory_builder && tox
# when: manual
# except: ['triggers', 'master']