Workspace⚓︎
The Workspace provides protected user resource management that includes dedicated storage and services for resource discovery and access.
Workspace API⚓︎
The Workspace API provides a REST service through which user workspaces can be created, interrogated, managed and deleted.
Prerequisite - Flux⚓︎
Workspaces are created by instantiating the rm-user-workspace
helm chart for each user/group. The Workspace API uses Flux CD as a helper to manage these subordinate helm charts - via flux resources of type HelmRelease
. Thus, it is necessary to deploy within the cluster the aspects of flux that support this helm chart management - namely the flux helm-controller
, source-controller
and the Kubernetes Custom Resource Definitions (CRD) for HelmRelease
and HelmRepository
..
Flux Deployment⚓︎
The flux controllers and CRDs are deployed to the cluster from yaml via kubectl…
kubectl apply -f https://raw.githubusercontent.com/EOEPCA/deployment-guide/main/local-deploy/eoepca/flux.yaml
EOEPCA Helm Chart Repository⚓︎
The flux helm-controller requires configuration which provides the details of the EOEPCA Helm Chart Repository from where the rm-user-workspace
chart is obtained. For flux, this is configured via a Kubernetes resource of type HelmRepository
- which is configured within the cluster by applying the following yaml to the cluster (ref. kubectl apply -f <yaml-file>
)…
apiVersion: source.toolkit.fluxcd.io/v1beta1
kind: HelmRepository
metadata:
name: eoepca
namespace: rm
spec:
interval: 2m
url: https://eoepca.github.io/helm-charts/
Helm Chart⚓︎
The Workspace API is deployed via the rm-workspace-api
helm chart from the EOEPCA Helm Chart Repository.
The chart is configured via values that are fully documented in the README for the um-workspace-api
chart.
helm install --values workspace-api-values.yaml workspace-api eoepca/rm-workspace-api
Values⚓︎
At minimum, values for the following attributes should be specified:
- The fully-qualified public URL for the service
- (optional) Specification of Ingress for reverse-proxy access to the service
Note that this is only required in the case that the Workspace API will not be protected by theresource-guard
component - ref. Resource Protection. Otherwise the ingress will be handled by theresource-guard
- useingress.enabled: false
. - Prefix for user projects in OpenStack
- Details for underlying S3 object storage service
- Identification of secret that provides the client credentials for resource protection
Example workspace-api-values.yaml
…
fullnameOverride: workspace-api
ingress:
enabled: true
hosts:
- host: workspace-api-open.192.168.49.123.nip.io
paths: ["/"]
tls:
- hosts:
- workspace-api-open.192.168.49.123.nip.io
secretName: workspace-api-open-tls
prefixForName: "guide-user"
workspaceSecretName: "bucket"
namespaceForBucketResource: "rm"
gitRepoResourceForHelmChartName: "eoepca"
gitRepoResourceForHelmChartNamespace: "rm"
helmChartStorageClassName: "standard"
s3Endpoint: "https://cf2.cloudferro.com:8080"
s3Region: "RegionOne"
workspaceDomain: 192.168.49.123.nip.io
harborUrl: "https://harbor.192.168.49.123.nip.io"
harborUsername: "admin"
harborPassword: "changeme"
umaClientSecretName: "resman-client"
umaClientSecretNamespace: "rm"
authServerIp: 192.168.49.123
authServerHostname: "auth"
clusterIssuer: letsencrypt-production
resourceCatalogVolumeStorageType: standard
NOTES:
- The Workspace API assumes a deployment of the Harbor Container Regsitry, as configured by the
harborXXX
values above.
See section Container Registry.
Protection⚓︎
As described in section Resource Protection, the resource-guard
component can be inserted into the request path of the Workspace API service to provide access authorization decisions
helm install --values workspace-api-guard-values.yaml workspace-api-guard eoepca/resource-guard
The resource-guard
must be configured with the values applicable to the Workspace API for the Policy Enforcement Point (pep-engine
) and the UMA User Agent (uma-user-agent
)…
Example workspace-api-guard-values.yaml
…
#---------------------------------------------------------------------------
# Global values
#---------------------------------------------------------------------------
global:
context: workspace-api
pep: workspace-api-pep
domain: 192.168.49.123.nip.io
nginxIp: 192.168.49.123
certManager:
clusterIssuer: letsencrypt-production
#---------------------------------------------------------------------------
# PEP values
#---------------------------------------------------------------------------
pep-engine:
configMap:
asHostname: auth
pdpHostname: auth
# customDefaultResources:
# - name: "Eric's workspace"
# description: "Protected Access for eric to his user workspace"
# resource_uri: "/workspaces/guide-user-eric"
# scopes: []
# default_owner: "d3688daa-385d-45b0-8e04-2062e3e2cd86"
# - name: "Bob's workspace"
# description: "Protected Access for bob to his user workspace"
# resource_uri: "/workspaces/guide-user-bob"
# scopes: []
# default_owner: "f12c2592-0332-49f4-a4fb-7063b3c2a889"
volumeClaim:
name: eoepca-resman-pvc
create: false
#---------------------------------------------------------------------------
# UMA User Agent values
#---------------------------------------------------------------------------
uma-user-agent:
fullnameOverride: workspace-api-agent
nginxIntegration:
enabled: true
hosts:
- host: workspace-api
paths:
- path: /(.*)
service:
name: workspace-api
port: http
annotations:
nginx.ingress.kubernetes.io/proxy-read-timeout: "600"
nginx.ingress.kubernetes.io/enable-cors: "true"
nginx.ingress.kubernetes.io/rewrite-target: /$1
client:
credentialsSecretName: "resman-client"
logging:
level: "info"
unauthorizedResponse: 'Bearer realm="https://auth.192.168.49.123.nip.io/oxauth/auth/passport/passportlogin.htm"'
openAccess: false
insecureTlsSkipVerify: true
NOTES:
- TLS is enabled by the specification of
certManager.clusterIssuer
- The
letsencrypt
Cluster Issuer relies upon the deployment being accessible from the public internet via theglobal.domain
DNS name. If this is not the case, e.g. for a local minikube deployment in which this is unlikely to be so. In this case the TLS will fall-back to the self-signed certificate built-in to the nginx ingress controller insecureTlsSkipVerify
may be required in the case that good TLS certificates cannot be established, e.g. if letsencrypt cannot be used for a local deployment. Otherwise the certificates offered by login-service Authorization Server will fail validation in the Resource Guard.customDefaultResources
can be specified to apply initial protection to the endpoint
Client Secret⚓︎
The Resource Guard requires confidential client credentials to be configured through the file client.yaml
, delivered via a kubernetes secret..
Example client.yaml
…
client-id: a98ba66e-e876-46e1-8619-5e130a38d1a4
client-secret: 73914cfc-c7dd-4b54-8807-ce17c3645558
Example Secret
…
kubectl -n rm create secret generic resman-client \
--from-file=client.yaml \
--dry-run=client -o yaml \
> resman-client-secret.yaml
apiVersion: v1
kind: Secret
metadata:
name: resman-client
namespace: rm
data:
client.yaml: Y2xpZW50LWlkOiBhOThiYTY2ZS1lODc2LTQ2ZTEtODYxOS01ZTEzMGEzOGQxYTQKY2xpZW50LXNlY3JldDogNzM5MTRjZmMtYzdkZC00YjU0LTg4MDctY2UxN2MzNjQ1NTU4
The client credentials are obtained by registration of a client at the login service web interface - e.g. https://auth.192.168.49.123.nip.io. In addition there is a helper script that can be used to create a basic client and obtain the credentials, as described in section Resource Protection…
./local-deploy/bin/register-client auth.192.168.49.123.nip.io "Resource Guard" | tee client.yaml
Workspace API Usage⚓︎
The Workspace API provides a REST interface that is accessed at the endpoint https://workspace-api.192.168.49.123.nip.io/.
See the Swagger Docs.
Additional Information⚓︎
Additional information regarding the Workspace API can be found at:
Bucket Operator⚓︎
The Workspace API creates workspaces for individual users. In doing so, dedicated object storage buckets are created associated to each user workspace - for self-contained storage of user owned resources (data, processing applications, etc.).
The bucket creation relies upon the object storage services of the underlying cloud infrastructure. We have created a Bucket
abstraction as a Kubernetes Custom Resource Definition
. This is served by a Bucket Operator
service that deploys into the Kubernetes cluster to satisfy requests for resources of type Bucket
.
We provide a Bucket Operator
implementation that currently supports the creation of buckets in OpenStack object storage - currently tested only on the CREODIAS (Cloudferro).
The Bucket Operator is deployed via the rm-bucket-operator
helm chart from the EOEPCA Helm Chart Repository.
The chart is configured via values that are fully documented in the README for the um-bucket-operator
chart.
helm install --values bucket-operator-values.yaml bucket-operator eoepca/rm-bucket-operator
Values⚓︎
At minimum, values for the following attributes should be specified:
- The fully-qualified public URL for the service
- OpenStack access details
- Cluster Issuer for TLS
Example bucket-operator-values.yaml
…
domain: 192.168.49.123.nip.io
data:
OS_MEMBERROLEID: "9ee2ff9ee4384b1894a90878d3e92bab"
OS_SERVICEPROJECTID: "d21467d0a0414252a79e29d38f03ff98"
USER_EMAIL_PATTERN: "eoepca+<name>@192.168.49.123.nip.io"
ingress:
annotations:
cert-manager.io/cluster-issuer: letsencrypt-production
kubernetes.io/ingress.class: nginx
nginx.ingress.kubernetes.io/enable-cors: "true"
nginx.ingress.kubernetes.io/proxy-read-timeout: "600"
OpenStack Secret⚓︎
The Bucket Operator
requires privileged access to the OpenStack API, for which credentials are required. These are provided via a Kubernetes secret named openstack
created in the namespace of the Bucket Operator.
For example…
kubectl -n rm create secret generic openstack \
--from-literal=username="${OS_USERNAME}" \
--from-literal=password="${OS_PASSWORD}" \
--from-literal=domainname="${OS_DOMAINNAME}"
See the README for the Bucket Operator, which describes the configuration required for integration with your OpenStack account.
For a worked example see our Scripted Example Deployment - in particular: