Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

update: test all functions #93

Merged
merged 33 commits into from
Jun 6, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
33 commits
Select commit Hold shift + click to select a range
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion .yamllint
Original file line number Diff line number Diff line change
Expand Up @@ -23,4 +23,4 @@ rules:
ignore:
- "scratch/*"
- "components/containers/web-terminal/src/*"
- "components/operators/gpu-operator-certified/instance/components/console-plugin-helm/*"
- "components/operators/gpu-operator-certified/operator/components/console-plugin-helm/*"
6 changes: 4 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@
This project is a catalog of configurations used to provision infrastructure, on
OpenShift, that supports machine learning (ML) and artificial intelligence (AI) workloads.

The intention of this repository is to help support practical use of OpenShift for AI / ML workloads.
The intention of this repository is to help support practical use of OpenShift for AI / ML workloads and provide a catalog of configurations / demos / workshops.

Please look at the [GitOps Catalog](https://github.com/redhat-cop/gitops-catalog) if you only need to automate an operator install.

Expand All @@ -16,12 +16,14 @@ In this repo, look at various [kustomized configs](components/configs) and [argo

## Prerequisites

- OpenShift 4.8+
- OpenShift 4.14+
- role: `cluster-admin` - for all [demo](demos) or [cluster](clusters) configs
- role: `self-provisioner` - for namespaced components

[Red Hat Demo Platform](https://demo.redhat.com) Options (Tested)

NOTE: node sizes below are the **recommended minimum** to select for provisioning

- <a href="https://demo.redhat.com/catalog?item=babylon-catalog-prod/sandboxes-gpte.sandbox-ocp.prod&utm_source=webapp&utm_medium=share-link" target="_blank">AWS with OpenShift Open Environment</a>
- 1 x Control Plane - `m6a.2xlarge`
- 0 x Workers - `m6a.2xlarge`
Expand Down
1 change: 1 addition & 0 deletions components/operators/ack-system/base/user-config-cm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -11,3 +11,4 @@ data:
AWS_ENDPOINT_URL: ""
ENABLE_LEADER_ELECTION: "true"
LEADER_ELECTION_NAMESPACE: ""
RECONCILE_DEFAULT_MAX_CONCURRENT_SYNCS: "1"
4 changes: 2 additions & 2 deletions components/operators/ack-system/scripts/setup.sh
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@ ocp_aws_get_key(){
}

# create secrets for ack controllers
aws_setup_ack_system(){
ocp_aws_setup_ack_system(){
NAMESPACE=ack-system

ocp_aws_get_key
Expand All @@ -56,4 +56,4 @@ aws_setup_ack_system(){
done
}

aws_setup_ack_system
ocp_aws_setup_ack_system
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,9 @@ kind: ConfigMap
metadata:
name: console-plugin-nvidia-gpu
namespace: nvidia-gpu-operator
annotations:
meta.helm.sh/release-name: console-plugin-nvidia-gpu
meta.helm.sh/release-namespace: nvidia-gpu-operator
labels:
helm.sh/chart: console-plugin-nvidia-gpu-0.2.4
app.kubernetes.io/name: console-plugin-nvidia-gpu
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,9 @@ kind: ConsolePlugin
metadata:
name: console-plugin-nvidia-gpu
namespace: nvidia-gpu-operator
annotations:
meta.helm.sh/release-name: console-plugin-nvidia-gpu
meta.helm.sh/release-namespace: nvidia-gpu-operator
labels:
helm.sh/chart: console-plugin-nvidia-gpu-0.2.4
app.kubernetes.io/name: console-plugin-nvidia-gpu
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,9 @@ kind: Deployment
metadata:
name: console-plugin-nvidia-gpu
namespace: nvidia-gpu-operator
annotations:
meta.helm.sh/release-name: console-plugin-nvidia-gpu
meta.helm.sh/release-namespace: nvidia-gpu-operator
labels:
helm.sh/chart: console-plugin-nvidia-gpu-0.2.4
app.kubernetes.io/name: console-plugin-nvidia-gpu
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,10 @@ kind: Service
metadata:
name: console-plugin-nvidia-gpu
namespace: nvidia-gpu-operator
annotations:
meta.helm.sh/release-name: console-plugin-nvidia-gpu
meta.helm.sh/release-namespace: nvidia-gpu-operator
service.alpha.openshift.io/serving-cert-secret-name: plugin-serving-cert
labels:
helm.sh/chart: console-plugin-nvidia-gpu-0.2.4
app.kubernetes.io/name: console-plugin-nvidia-gpu
Expand All @@ -13,8 +17,6 @@ metadata:
app.kubernetes.io/component: console-plugin-nvidia-gpu
app.kubernetes.io/instance: console-plugin-nvidia-gpu
app.kubernetes.io/part-of: console-plugin-nvidia-gpu
annotations:
service.alpha.openshift.io/serving-cert-secret-name: plugin-serving-cert
spec:
ports:
- name: 9443-tcp
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -11,18 +11,13 @@ metadata:
spec:
components:
- container:
env:
- name: YOLO_URL
value: https://raw.githubusercontent.com/redhat-na-ssa/demo-ai-gitops-catalog/main/scripts/library/term.sh
args:
- "/bin/bash"
- "-c"
- |
#!/bin/bash
# set -x

. <(curl -s "${YOLO_URL}" | tee /tmp/yolo.sh)
term_init
. /home/user/.bashrc

sleep infinity # just in case

Expand Down
Loading
Loading