Skip to content
This repository has been archived by the owner on Sep 30, 2022. It is now read-only.

Latest commit

 

History

History
351 lines (245 loc) · 14.7 KB

faq.rst

File metadata and controls

351 lines (245 loc) · 14.7 KB

FAQ (Technical)

Can I run Containers/Pods as root?

This is not possible due to security restrictions. It is also best practice to run as non-root as recommended by Docker themselves: https://docs.docker.com/engine/security/security/#linux-kernel-capabilities

What do we monitor?

The functionality of OpenShift and all involved services are completely monitored and operated by VSHN. Individual projects are not monitored out of the box - but Kubernetes already has health checks integrated and running. Also replication controllers make sure that Pods are running all the time. If you need a more complex monitoring for your project, feel free to contact us at our Customer Portal.

More information can also be found here: Application Health

Route monitoring

Certificates on application routes are monitored for validity. Users or cluster operators may set any of the following annotations on routes:

  • monitoring.appuio.ch/enable: Whether to monitor route at all (boolean, default true)
  • monitoring.appuio.ch/enable-after: Whether to monitor route after specified point in time (time, no default).
  • monitoring.appuio.ch/verify-tls-certificate: Whether to verify X.509 certificate validity (boolean as string, default true).
  • monitoring.appuio.ch/not-after-remaining-warn, monitoring.appuio.ch/not-after-remaining-crit: Amount of time before reporting warning or critical status when the primary route certificate is to expire (duration).

Value formats:

  • Boolean: "false" or "true".

  • Time: Supported formats (see time.ParseInLocation):

    • 2006-01-02T15:04:05Z07:00
    • 2006-01-02T15:04Z07:00
    • 2006-01-02T15:04
    • 2006-01-02
  • Duration: Parsed using Go's time.ParseDuration function, e.g. 168h30m.

What do we backup?

We backup all data relevant to run the OpenShift cluster. Application data itself is not in the default backup and is the responsibility of the user. However, we can provide a backup service for individual projects. Please contact us at our Customer Portal for more information.

What DNS entries should I add to my custom domain?

When creating an application route, the platform automatically generates a URL which is immediately accessible, e.g. http://django-psql-example-my-project.appuioapp.ch due to wildcard DNS entries under *.appuioapp.ch. If you now want to have this application available under your own custom domain, follow these steps:

  1. Edit the route and change the hostname to your desired hostname, e.g. www.myapp.ch
  2. Point your DNS entry using a CNAME resource record type (important!) to cname.appuioapp.ch

Warning

Always create a route before pointing a DNS entry to APPUiO, otherwise someone else could create a matching route and serve content under your domain.

Note that you can't use CNAME records in the apex domain (example.com, e.g. without www in front of it). If you need to use the apex domain for your application you have the following options:

  1. Redirect to a subdomain (e.g. example.com ➜ www.example.com or app.example.com) with your DNS-provider, set up the subdomain with a CNAME
  2. Use ALIAS-records with your DNS-provider if they support them
  3. Enter 5.102.151.2 and 5.102.151.3 as A records

Which IP addresses are being used?

Disclaimer: These addresses may change at any time. We do not recommend whitelisting by IP address. A better option is to use Transport Layer Security (TLS) with client certificates for authentication.

Incoming connections for routes
5.102.151.2, 5.102.151.3
Outgoing connections from pods
5.102.151.22, 2a06:c00:10:bc00::/56

How can I secure the access to my web application?

OpenShift supports secure routes and everything is prepared on APPUiO to have it secured easily. Just edit the route and change the termination type to edge. There is a default trusted certificate in place for *.appuioapp.ch which is used in this case. If you want to use your own certificate, see Routes.

Can I run a database on APPUiO?

Short answer: Yes. But we do discourage it. Use the gluster-database storage class as described in :ref:`persistent-storage` if you do. See :ref:`tutorial-helm-charts` for a convenient way to deploy a database service.

We provide shared persistent storage using GlusterFS. Please make sure that the database intended to use is capable of storing its data on a shared filesystem. We don't recommend running production databases with GlusterFS as storage backend, because there is a risk of data corruption and when that happens, your database will not run/start anymore. For highly-available and high-performance managed databases, please contact us at our Customer Portal.

I get an error like 'Failed Mount: MountVolume.NewMounter initialization failed for volume "gluster-pv123" : endpoints "glusterfs-cluster" not found'

When you received your account there was a service called "glusterfs-cluster" pointing to the persistent storage endpoint. If you delete it by accident you can re-create it with:

oc create -f - <<EOF
apiVersion: template.openshift.io/v1
kind: List
items:
- apiVersion: v1
  kind: Service
  metadata:
    name: glusterfs-cluster
  spec:
    ports:
    - port: 1
      protocol: TCP
      targetPort: 1
- apiVersion: v1
  kind: Endpoints
  metadata:
    name: glusterfs-cluster
  subsets:
  - addresses:
    - ip: 172.17.176.30
    - ip: 172.17.176.31
    - ip: 172.17.176.32
    ports:
    - port: 1
      protocol: TCP
EOF

Or copy the YAML between "oc" and "EOF" in the Web GUI to "Add to project" ➜ "Import YAML/JSON" Or run oc create -f https://raw.githubusercontent.com/appuio/docs/master/glusterfs-cluster.yaml

Please note that the IP addresses above are dependent on which cluster you are on, these are valid for console.appuio.ch

How do I kill a pod/container

If your container is hanging, either because your application is unresponsive or because the pod is in state "Terminating" for a long time, you can manually kill the pod:

oc delete pod/mypod

If it still hangs you can use more force:

oc delete --grace-period=0 --force pod/mypod

The same functionality is available in the Web GUI: Applications ➜ Pods ➜ Actions ➜ Delete, there is a checkbox "Delete pod immediately without waiting for the processes to terminate gracefully" for applying more force

How do I work with a volume if my application crashes because of the data in the volume?

If your application is unhappy with the data in a persistent volume you can connect to the application pod:

oc rsh mypod

to run commands inside the application container, e.g. to fix or delete the data. In the Web GUI this is Applications ➜ Pods ➜ mypod ➜ Terminal.

If your application crashes at startup this does not work as there is no container to connect to – the container exits as soon as your application exits. If there is a shell included in your container image you can use oc debug to clone your deployment config including volumes for a one-off debugging container:

oc debug deploymentconfig/prometheus

If your container image does not include a shell or you need special recovery tools you can start another container image, mount the volume with the data and then use the tools in the other container image to fix the data manually. Unfortunately the oc run command does not support specifying a volume, so we have to create a deployment config with the volume for it to be mounted and make sure our deployed container does not exit:

1. Get the name of the persistent volume claim (pvc) that contains the data. In this example the application and deployment config (dc) name is 'prometheus':

oc volume dc/prometheus

This produces the following output:

deploymentconfigs/prometheus
  configMap/prometheus-config as prometheus-config-1
    mounted at /etc/prometheus
  pvc/prometheus-data (allocated 1GiB) as prometheus-volume-1
    mounted at /prometheus

You can see the pvc/prometheus-data is the persistent volume claim that is mounted at /prometheus for the application prometheus.

  1. Deploy the helper container (e.g. "busybox", minimal container containing a shell) - if you need special tools to fix the data (e.g. to recover a database) you should use another container image containing these tools), patch it not to exit and mount the volume at /mnt:
# create a new deployment with a "busybox" shell container
oc new-app busybox
# patch the new deployment with a while-true-loop so the container keeps on running
oc patch dc/busybox -p '{"spec":{"template":{"spec":{"containers":[{"name":"busybox","command":["sh"],"args":["-c","while [ 1 ]; do echo hello; sleep 1; done"]}]}}}}'
# mount the persistent volume claim into the container at /mnt
oc volume dc/busybox --add -m /mnt -t pvc --claim-name prometheus-data
# wait for the new deployment with the mount to roll out

Warning

The oc patch command above has a problem with escaping on Windows cmd/PowerShell. You can add the "command" and "args" keys and values in the Web GUI.

  1. Connect to your helper container and work in the volume:
oc rsh dc/busybox
cd /mnt/
# congratulations, you are now in the volume you want to fix
# you can now selectively delete/edit/clean the bad data
  1. Clean up the temporary deployment config afterwards:
oc delete all -l app=busybox

How long do we keep application logs?

Application logs are stored in elasticsearch and accessible via Kibana. All container logs are sent there but only kept for 10 days.

Is OpenShift Service Catalog available to be used?

OpenShift Service Catalog is not supported nor available to be used on APPUiO. Template Service Broker and OpenShift Ansible Broker are not supported nor available. It was once available, but because Red Hat is removing the support of the Service Catalog from OpenShift, we decided to remove the Service Catalog from APPUiO.

See :ref:`tutorial-helm-charts` for an alternative.

How to pull an image from a private registry or private docker hub

To pull an image from a private container registry like Docker Hub Private Repositories you need to create a secret to store the credentials and link it to be used for pulls in your project:

oc create secret docker-registry myimagepullingsecretname \
  --docker-server=docker.io \
  --docker-username=myusername \
  --docker-password=mypassword \
  [email protected] \
  --namespace=myproject

oc secrets link default myimagepullingsecretname \
  --for=pull
  --namespace=myproject

I've exhausted the number of ReplicationControllers?

A DeploymentConfig creates a new ReplicationController for each deployment of a new version. By default there is no limit on the number of "old" ReplicationControllers that should be saved for debugging/rollback purposes which may lead to the project hitting the 100 ReplicationControllers project quota limit after some time.

There is a "revisionHistoryLimit" configuration parameter for DeploymentConfig specs (where the default is 0 meaning no limit) that you can set to a sensible number (I usually set it to 2) that automatically cleans up old ReplicationControllers if there are more than the specified number.

You can set the option using the CLI:

oc patch dc/yourdeploymentconfigname -p '{"spec":{"revisionHistoryLimit":2}}'

Or using the Web-GUI

  1. navigate to the DeploymentConfig, choose Actions -> Edit YAML
  2. navigate to the first "spec:" on the top level, usually there is a " replicas: 1" on the line below it
  3. add " revisionHistoryLimit: 2" between spec and replicas, on the same level as replicas

More information about this in the OpenShift documentation: https://docs.openshift.com/container-platform/3.11/dev_guide/deployments/how_deployments_work.html#creating-a-deployment-configuration

I get an error like 'You have reached your pull rate limit. You may increase the limit by authenticating and upgrading'

Docker Inc. has introduced a pull rate limited for container images stored on Docker Hub. If that limit is reached you get an error like:

toomanyrequests: You have reached your pull rate limit. You may increase the limit by authenticating and upgrading: https://www.docker.com/increase-rate-limit

To avoid this issue, you should always use authenticated image pulls for Docker Hub. You can do this by creating an account on https://hub.docker.com and adding a pull secret to your project, see :ref:`faq-pull-secret`.

If you are using a free account on Docker Hub, this will increase your limit to 200 container image requests per 6 hours. If that is not enough and you are still reaching the limit you can upgrade your Docker Hub subscription to get unlimited container image pulls.

Another possibility consists in migrating images to a different public registry like quay.io or Amazon ECR, which currently don't have any limits for public container images.

More information about this from Docker: https://www.docker.com/increase-rate-limit and on our blog: https://vshn.ch/en/blog/assessing-the-impact-of-new-docker-pull-rate-limits-in-appuio