| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
|
| |
Some registries are not configured with valid certificates and thus the
check fails with 'http: server gave HTTP response to HTTPS client'.
Since this is not fetching images, but only checking for existence,
trade security for convenience.
|
|
|
|
|
|
| |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.116
|
|\
| |
| | |
Merged by openshift-bot
|
| |
| |
| |
| |
| | |
If we have no master config assume that we're a clean install.
If we're a clean install and we're 3.6 or greater use etcd v3 storage.
|
|\ \
| | |
| | | |
Merged by openshift-bot
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
bug 1460564. Fixes [BZ #1460564](https://bugzilla.redhat.com/show_bug.cgi?id=1460564).
Unfortunately, the defaults for Elasticsearch prior to v5 allow more
than one "node" to access the same configured storage volume(s).
This change forces this value to 1 to ensure we don't have an ES pod
starting up accessing a volume while another ES pod is shutting down
when reploying. This can lead to "1" directories being created in
`/elasticsearch/persistent/${CLUSTER_NAME}/data/${CLUSTER_NAME}/nodes/`.
By default ES uses a "0" directory there when only one node is accessing
it.
|
|\ \ \
| | | |
| | | | |
Rename cockpit-shell -> cockpit-system
|
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
This is a temporary workaround to make sure the action plugin in the
openshift_health_checker role can read the correct value specified in
the OPENSHIFT_IMAGE_TAG environment variable, instead of a jinja2
template expression.
|
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
The package name has changed.
See
https://bugzilla.redhat.com/show_bug.cgi?id=1461689
https://bugzilla.redhat.com/show_bug.cgi?id=1419718
|
|\ \ \ \
| | | | |
| | | | | |
Update CloudForms templates for CF 4.5/CF 4.2
|
| | | | |
| | | | |
| | | | |
| | | | | |
'cloudforms42' for CF 4.2.
|
| | | | |
| | | | |
| | | | |
| | | | | |
'cloudforms45' for CF 4.5.
|
| | | | | |
|
| |/ / /
|/| | |
| | | |
| | | |
| | | |
| | | | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.114
|
| | | |
| | | |
| | | |
| | | |
| | | |
| | | | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.113
|
|\ \ \ \
| | | | |
| | | | | |
Merged by openshift-bot
|
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | |
| | | | | |
port the code that creates the external Elasticsearch routes to the
new logging roles
Have to suppress this error message:
SSL Problem illegal change cipher spec msg, conn state = 6, handshake state = 1
which is coming from the router health check, until
https://github.com/openshift/origin/issues/14515
is fixed - otherwise, the es log is spammed relentlessly
|
|\ \ \ \ \
| | | | | |
| | | | | | |
Merged by openshift-bot
|
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | | |
We cannot rely on the `watch.Until` call in the `rollout status`
subcommand for the time being, so we need to ignore the result of this
call. This will make the rollout status check best-effort, so we need to
follow it with a poll for the actual status of the rollout, which we can
extract from the `openshift.io/deployment.phase` annotation on the
ReplicationControllers. This annotation can have only three values --
`Running`, `Complete` and `Failed`. If we poll on this attribute until
we stop seeing `Running`, we can then inspect the last result for
`Failed`; if it's present, we have failed the deployment.
Signed-off-by: Steve Kuznetsov <skuznets@redhat.com>
|
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | | |
When deploying the hosted router and registry components, we need to
ensure that they correctly roll out. The previous checks were weak in
that they either simply waited for a set amount of time and/or did one
replica check. They would fail if the router or registry took longer to
deploy or if there were un-ready or failing replicas. The `oc rollout`
command group contains the `status` endpoint for internalizing all of
the logic for determining when a rollout has succeeded or failed, so
simply using this client call will ensure that the router and registry
correctly deploy.
Signed-off-by: Steve Kuznetsov <skuznets@redhat.com>
|
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | |
| | | | | | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.112
|
|\ \ \ \ \ \
| | | | | | |
| | | | | | | |
Add the other featured audit-config paramters as example
|
| | | | | | | |
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.111
|
|\ \ \ \ \ \ \
| |_|_|_|_|/ /
|/| | | | | | |
CFME as an OpenShift Pod
|
| | | | | | | |
|
| | | | | | | |
|
| | | | | | | |
|
| | | | | | | |
|
|\ \ \ \ \ \ \
| | | | | | | |
| | | | | | | | |
System container honor provided ansible cfg
|
| | | | | | | | |
|
| | |/ / / / /
| |/| | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
The provided configuration file was not being honored. This change adds
a new variable called ANSIBLE_CONFIG which points to the ansible.cfg
file within the container. By default it is set to the configuration
provided by openshift-ansible within the container at:
/usr/share/ansible/openshift-ansible/ansible.cfg
|
|\ \ \ \ \ \ \
| |/ / / / / /
|/| | | | | | |
Merged by openshift-bot
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
adding
openshift_logging_fluentd_buffer_queue_limit: 1024
openshift_logging_fluentd_buffer_size_limit: 1m
openshift_logging_mux_buffer_queue_limit: 1024
openshift_logging_mux_buffer_size_limit: 1m
and setting the matched environment variables.
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
roles/openshift_logging_fluentd/defaults/main.yml.
Adding the description for openshift_hosted_logging_fluentd_{cpu,memory}_limit to README.md.
Replaced openshift_hosted_logging_mux_.* with openshift_logging_mux_.*.
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
MUX_CPU_LIMIT, MUX_MEMORY_LIMIT.
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
and memory_limit.
Default values:
openshift_logging_fluentd_cpu_limit: 100m
openshift_logging_fluentd_memory_limit: 512Mi
openshift_logging_fluentd_buffer_queue_limit: 1024
openshift_logging_fluentd_buffer_size_limit: 16m
openshift_logging_fluentd_buffer_size_limit: 1m
openshift_logging_mux_cpu_limit: 500m
openshift_logging_mux_memory_limit: 1Gi
openshift_logging_mux_buffer_queue_limit: 1024
openshift_logging_mux_buffer_size_limit: 1m
|
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | |
| | | | | | | |
Created by command:
/usr/bin/tito tag --accept-auto-changelog --use-version=3.6.110
|
|\ \ \ \ \ \ \
| | | | | | | |
| | | | | | | | |
Merged by openshift-bot
|
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | |
| | | | | | | | |
Add the docker and RPM checks to the list that run at install time.
They can be disabled the same as the existing ones.
Removed cockpit-kubernetes RPM requirement as it no longer is.
Fixed up docker_image_availability to handle oreg_url and other nuances.
Switched to using the openshift_image_tag that's set by openshift_version for
both component and infrastructure images.
Fixed a bug where execute_module was being called with incorrect
positional arg "tmp" as a dict which caused errors down the call stack.
|
|\ \ \ \ \ \ \ \
| | | | | | | | |
| | | | | | | | | |
PAPR updates
|
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | | |
Don't duplicate the actual image tag used in the context. Just print it
in the logs instead. That way there's no risk of it being outdated when
we bump the tag in the YAML.
|
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | | |
Add a section in `repo_structure.md`, and rename from `redhat-ci` to
`PAPR` and point to the new upstream repo in `pull_requests.md`.
Closes: #4078
|
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | | |
Closes: #4418
|
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | |
| | | | | | | | | |
The project was renamed to PAPR, see:
https://github.com/projectatomic/papr
|
|\ \ \ \ \ \ \ \ \
| | | | | | | | | |
| | | | | | | | | | |
Merged by openshift-bot
|
| | | | | | | | | | |
|
|\ \ \ \ \ \ \ \ \ \
| |_|_|_|_|/ / / / /
|/| | | | | | | | | |
Merged by openshift-bot
|
| | | | | | | | | | |
|
|\ \ \ \ \ \ \ \ \ \
| | | | | | | | | | |
| | | | | | | | | | | |
move etcd backup to etcd_common role
|