Skip to content

Commit

Permalink
update: test all functions (#93)
Browse files Browse the repository at this point in the history
  • Loading branch information
codekow authored Jun 6, 2024
1 parent 03b99c1 commit 2064bb1
Show file tree
Hide file tree
Showing 26 changed files with 704 additions and 983 deletions.
2 changes: 1 addition & 1 deletion .yamllint
Original file line number Diff line number Diff line change
Expand Up @@ -23,4 +23,4 @@ rules:
ignore:
- "scratch/*"
- "components/containers/web-terminal/src/*"
- "components/operators/gpu-operator-certified/instance/components/console-plugin-helm/*"
- "components/operators/gpu-operator-certified/operator/components/console-plugin-helm/*"
6 changes: 4 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@
This project is a catalog of configurations used to provision infrastructure, on
OpenShift, that supports machine learning (ML) and artificial intelligence (AI) workloads.

The intention of this repository is to help support practical use of OpenShift for AI / ML workloads.
The intention of this repository is to help support practical use of OpenShift for AI / ML workloads and provide a catalog of configurations / demos / workshops.

Please look at the [GitOps Catalog](https://github.com/redhat-cop/gitops-catalog) if you only need to automate an operator install.

Expand All @@ -16,12 +16,14 @@ In this repo, look at various [kustomized configs](components/configs) and [argo

## Prerequisites

- OpenShift 4.8+
- OpenShift 4.14+
- role: `cluster-admin` - for all [demo](demos) or [cluster](clusters) configs
- role: `self-provisioner` - for namespaced components

[Red Hat Demo Platform](https://demo.redhat.com) Options (Tested)

NOTE: node sizes below are the **recommended minimum** to select for provisioning

- <a href="https://demo.redhat.com/catalog?item=babylon-catalog-prod/sandboxes-gpte.sandbox-ocp.prod&utm_source=webapp&utm_medium=share-link" target="_blank">AWS with OpenShift Open Environment</a>
- 1 x Control Plane - `m6a.2xlarge`
- 0 x Workers - `m6a.2xlarge`
Expand Down
1 change: 1 addition & 0 deletions components/operators/ack-system/base/user-config-cm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -11,3 +11,4 @@ data:
AWS_ENDPOINT_URL: ""
ENABLE_LEADER_ELECTION: "true"
LEADER_ELECTION_NAMESPACE: ""
RECONCILE_DEFAULT_MAX_CONCURRENT_SYNCS: "1"
4 changes: 2 additions & 2 deletions components/operators/ack-system/scripts/setup.sh
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@ ocp_aws_get_key(){
}

# create secrets for ack controllers
aws_setup_ack_system(){
ocp_aws_setup_ack_system(){
NAMESPACE=ack-system

ocp_aws_get_key
Expand All @@ -56,4 +56,4 @@ aws_setup_ack_system(){
done
}

aws_setup_ack_system
ocp_aws_setup_ack_system
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,9 @@ kind: ConfigMap
metadata:
name: console-plugin-nvidia-gpu
namespace: nvidia-gpu-operator
annotations:
meta.helm.sh/release-name: console-plugin-nvidia-gpu
meta.helm.sh/release-namespace: nvidia-gpu-operator
labels:
helm.sh/chart: console-plugin-nvidia-gpu-0.2.4
app.kubernetes.io/name: console-plugin-nvidia-gpu
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,9 @@ kind: ConsolePlugin
metadata:
name: console-plugin-nvidia-gpu
namespace: nvidia-gpu-operator
annotations:
meta.helm.sh/release-name: console-plugin-nvidia-gpu
meta.helm.sh/release-namespace: nvidia-gpu-operator
labels:
helm.sh/chart: console-plugin-nvidia-gpu-0.2.4
app.kubernetes.io/name: console-plugin-nvidia-gpu
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,9 @@ kind: Deployment
metadata:
name: console-plugin-nvidia-gpu
namespace: nvidia-gpu-operator
annotations:
meta.helm.sh/release-name: console-plugin-nvidia-gpu
meta.helm.sh/release-namespace: nvidia-gpu-operator
labels:
helm.sh/chart: console-plugin-nvidia-gpu-0.2.4
app.kubernetes.io/name: console-plugin-nvidia-gpu
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,10 @@ kind: Service
metadata:
name: console-plugin-nvidia-gpu
namespace: nvidia-gpu-operator
annotations:
meta.helm.sh/release-name: console-plugin-nvidia-gpu
meta.helm.sh/release-namespace: nvidia-gpu-operator
service.alpha.openshift.io/serving-cert-secret-name: plugin-serving-cert
labels:
helm.sh/chart: console-plugin-nvidia-gpu-0.2.4
app.kubernetes.io/name: console-plugin-nvidia-gpu
Expand All @@ -13,8 +17,6 @@ metadata:
app.kubernetes.io/component: console-plugin-nvidia-gpu
app.kubernetes.io/instance: console-plugin-nvidia-gpu
app.kubernetes.io/part-of: console-plugin-nvidia-gpu
annotations:
service.alpha.openshift.io/serving-cert-secret-name: plugin-serving-cert
spec:
ports:
- name: 9443-tcp
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -11,18 +11,13 @@ metadata:
spec:
components:
- container:
env:
- name: YOLO_URL
value: https://raw.githubusercontent.com/redhat-na-ssa/demo-ai-gitops-catalog/main/scripts/library/term.sh
args:
- "/bin/bash"
- "-c"
- |
#!/bin/bash
# set -x
. <(curl -s "${YOLO_URL}" | tee /tmp/yolo.sh)
term_init
. /home/user/.bashrc
sleep infinity # just in case
Expand Down
Loading

0 comments on commit 2064bb1

Please sign in to comment.