| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
| |
timeout was reached.
|
|\
| |
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
Update user-data for FQDN instance hostnames
* Remove node service start from bootstrap.yml. We start and enable the node service in user-data and we want the node service to start after NetworkManager so that the instance will have its final hostname.
* Restart systemd-hostnamed before restarting NetworkManager. In testing this has ensured that instances receive an FQDN hostname which is required for kube.
|
| | |
|
|/
|
|
| |
created AMI
|
|\
| |
| | |
add deprovisioning for ELB (and IAM certs)
|
| |
| |
| |
| |
| |
| | |
add playbooks to handle deleting ELBs and any IAM certs that may have been created during provisioning.
redo ELB creation to remove arbitrary wait and just retry until ELB creation succeeds
|
|\ \
| | |
| | | |
parameterize various master/infra scale group bits
|
| |/
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
allow selectively overriding master scale group bits:
instance type: openshift_aws_master_group_instance_type
scale group min size: openshift_aws_master_group_min_size
scale group max size: openshift_aws_master_group_max_size
scale group desired size: openshift_aws_master_group_desired_size
and overriding infra scale group bits:
instance type: openshift_aws_infra_group_instance_type
scale group min size: openshift_aws_infra_group_min_size
scale group max size: openshift_aws_infra_group_max_size
scale group desired size: openshift_aws_infra_group_desired_size
and overriding compute scale group bits:
instance type: openshift_aws_compute_group_instance_type
scale group min size: openshift_aws_compute_group_min_size
scale group max size: openshift_aws_compute_group_max_size
scale group desired size: openshift_aws_compute_group_desired_size
also consolodate the health check for master/infra/compute since they are all the same
|
|/
|
|
|
|
| |
Default to just cleaning out all the objects in the S3 bucket (IFF openshift_aws_create_s3 is 'true').
If you really, trully want to delete the S3 bucket and free up the bucket name, you can set openshift_aws_really_delete_s3_bucket to 'true' ('false' by default).
|
| |
|
| |
|
|\
| |
| |
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
un-hardcode default subnet az
allow defining 'default_az' for one of the subnets in the vpc structure, and pull that default_az and set it to openshift_aws_subnet_az.
this should allow one less variable to be defined/overridded when using non-default (us-east-1) regions
|
| |
| |
| |
| |
| |
| |
| |
| | |
allow defining 'default_az' for one of the subnets in the vpc structure, and pull that default_az and set it to openshift_aws_subnet_az.
this should allow one less variable to have to be defined/overridded (openshift_aws_subnet_az) when using non-default (us-east-1) regions
update provisioning_vars.yml.example to show an example VPC structure
|
|\ \
| | |
| | | |
Fix to master-internal elb scheme
|
| | | |
|
|\ \ \
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
Automatic merge from submit-queue.
allow uninstalling AWS objects created by prerequisite playbook
add deprovisioners/uninstallers for objects created via playbooks/aws/openshift-cluster/prerequisites.yml
specifically: security groups, vpcs, and any ssh keys
introduce openshift_aws_enable_uninstall_shared_objects to protect AWS objects that would be shared in the case of multiple clusters co-existing in one AWS account. right now it protects the ssh keys, but it can/should be used to protect against deleting the shared IAM instance profile as well. default this variable to False to be on the safe side when uninstalling/deprovisioning.
|
| |/ /
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
add deprovisioners/uninstallers for objects created via playbooks/aws/openshift-cluster/prerequisites.yml
specifically: security groups, vpcs, and any ssh keys
introduce openshift_aws_enable_uninstall_shared_objects to protect AWS objects that would be shared in the case of multiple clusters co-existing in one AWS account. right now it protects the ssh keys, but it can/should be used to protect against deleting the shared IAM instance profile as well. default this variable to False to be on the safe side when uninstalling/deprovisioning.
add some documentation on using deprovisioning playbooks
|
|\ \ \
| | | |
| | | | |
Use master group's ansible_ssh_user when accepting nodes.
|
| |/ /
| | |
| | |
| | | |
that we can use the first master's ansible_ssh_user when delegating.
|
|\ \ \
| |/ /
|/| |
| | |
| | | |
Automatic merge from submit-queue.
Use master group ansible_ssh_user to pull bootstrap kubeconfig
|
| | |
| | |
| | |
| | | |
pull bootstrap kubeconfig.
|
|\ \ \
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
Automatic merge from submit-queue.
Update to AWS EC2 root vol size so that Health Check tasks pass
AWS EC2s under the ASG and LaunchConfig are deployed with the default 10gb root vol size. Heath Checks fail because / filesystem and /var directory is too small. Health Check wants a minimum of 40Gb. 10Gb is just too small for server operations in general. Let's increase to 100Gb. Cost increase for additional EBS storage will be minimal and give plenty of elbow room for normal server operations.
|
| | | | |
|
|\ \ \ \
| |_|_|/
|/| | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
Automatic merge from submit-queue.
Separate ELB & S3 from master node group provisioning
Moved s3 and elb provisioning out of `roles/openshift_aws/tasks/provisioning.yml` and into their own playbooks. These playbooks are now included in the provision playbook and in a WIP infrastructure playbook we intend to run up front in place of prerequisites.
@kwoodson what are your thoughts on something like this?
|
| | | |
| | | |
| | | |
| | | | |
outside of the openshift_aws master provisioning tasks.
|
|/ / /
| | |
| | |
| | |
| | |
| | |
| | | |
This commit continues moving plugins into lib_utils.
This commit does not move any plugins for add-on roles
such as logging and metrics.
|
| | |
| | |
| | |
| | |
| | |
| | | |
add host to g_new_node_hosts so that plays run against the AMI instance
update example vars so that overlay2 is used by default for docker storage
|
|\ \ \
| | | |
| | | | |
Adding ability to update ami drive size.
|
| |/ / |
|
|/ /
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
In Ansible 2.2, the include_role directive came into existence as
a Tech Preview. It is still a Tech Preview through Ansible 2.4
(and in current devel branch), but with a noteable change. The
default behavior switched from static: true to static: false
because that functionality moved to the newly introduced
import_role directive (in order to stay consistent with include*
being dynamic in nature and `import* being static in nature).
The dynamic include is considerably more memory intensive as it will
dynamically create a role import for every host in the inventory
list to be used. (Also worth noting, there is at the time of this
writing an object allocation inefficiency in the dynamic include
that can in certain situations amplify this effect considerably)
This change is meant to mitigate the pressure on memory for the
Ansible control host.
We need to evaluate where it makes sense to dynamically include roles
and revert back to dynamic inclusion if and where it makes sense to do
so.
|
|/ |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
|\
| |
| | |
Renaming csr to bootstrap for consistency.
|
| | |
|
|/ |
|
| |
|
|
|
|
|
|
|
| |
Updating provisioning documents.
Also moved some steps from build_ami.yml to a new
play for better reusability.
|
|\
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Automatic merge from submit-queue.
Build provision split
Make provisioning steps more reusable
Reorganizing and making some of the plays more
reusable.
Depends-on: https://github.com/openshift/openshift-ansible/pull/5565
|
| | |
|
| |
| |
| |
| |
| | |
Reorganizing and making some of the plays more
reusable.
|
|\ \
| |/
|/|
| |
| |
| |
| |
| |
| |
| |
| |
| | |
mgugino-upstream-stage/provision-scale-elb-fixes-master
Automatic merge from submit-queue.
Fix provisiong scale group and elb logic
Some changes necessary to support provisioning
and scale group logic
Some of this work is duplicated from an existing PR. That PR is about to merge, so I'll rebase once that hits.
|
| |
| |
| |
| |
| | |
Some changes necessary to support provisioning
and scale group logic
|
|\ \
| | |
| | | |
Fix some provisioning variables
|