Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions skeleton/template.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -553,7 +553,7 @@ spec:
appRunCommand: "${APP_RUN_COMMAND}"
modelServiceContainer: ${MODEL_SERVICE_CONTAINER}
modelServicePort: ${MODEL_SERVICE_PORT}
customModelName: ${{ steps['fetch-model-from-catalog'].output.entity.metadata.name if parameters.modelServer === 'choose-from-the-catalog' else parameters.modelName }}
customModelName: ${{ steps['fetch-model-from-catalog'].output.entity.metadata.annotations | pick('rhdh.modelcatalog.io/model-name') if parameters.modelServer === 'choose-from-the-catalog' else parameters.modelName }}
modelName: ${MODEL_NAME}
modelSrc: ${MODEL_SRC}
modelServerName: ${{ parameters.modelServer }}
Expand Down Expand Up @@ -652,7 +652,7 @@ spec:
# for vllm
vllmSelected: ${{ parameters.modelServer === 'vLLM' }}
vllmModelServiceContainer: ${VLLM_CONTAINER}
modelName: ${{ parameters.modelName if parameters.modelServer === 'Bring you own model server' else (steps['fetch-model-from-catalog'].output.entity.metadata.name if parameters.modelServer === 'choose-from-the-catalog' else '${LLM_MODEL_NAME}') }}
modelName: ${{ parameters.modelName if parameters.modelServer === 'Bring you own model server' else (steps['fetch-model-from-catalog'].output.entity.metadata.annotations | pick('rhdh.modelcatalog.io/model-name') if parameters.modelServer === 'choose-from-the-catalog' else '${LLM_MODEL_NAME}') }}
modelSrc: ${MODEL_SRC}
maxModelLength: ${LLM_MAX_MODEL_LEN}
# SED_LLM_SERVER_END
Expand Down
2 changes: 1 addition & 1 deletion templates/audio-to-text/template.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -330,7 +330,7 @@ spec:
appRunCommand: "streamlit run whisper_client.py"
modelServiceContainer: quay.io/redhat-ai-dev/whispercpp:1.8.0
modelServicePort: 8001
customModelName: ${{ steps['fetch-model-from-catalog'].output.entity.metadata.name if parameters.modelServer === 'choose-from-the-catalog' else parameters.modelName }}
customModelName: ${{ steps['fetch-model-from-catalog'].output.entity.metadata.annotations | pick('rhdh.modelcatalog.io/model-name') if parameters.modelServer === 'choose-from-the-catalog' else parameters.modelName }}
modelName: ggerganov/whisper.cpp
modelSrc: https://huggingface.co/ggerganov/whisper.cpp
modelServerName: ${{ parameters.modelServer }}
Expand Down
4 changes: 2 additions & 2 deletions templates/chatbot/template.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -362,7 +362,7 @@ spec:
appRunCommand: "streamlit run chatbot_ui.py"
modelServiceContainer: quay.io/redhat-ai-dev/llamacpp_python:0.3.16
modelServicePort: 8001
customModelName: ${{ steps['fetch-model-from-catalog'].output.entity.metadata.name if parameters.modelServer === 'choose-from-the-catalog' else parameters.modelName }}
customModelName: ${{ steps['fetch-model-from-catalog'].output.entity.metadata.annotations | pick('rhdh.modelcatalog.io/model-name') if parameters.modelServer === 'choose-from-the-catalog' else parameters.modelName }}
modelName: ibm-granite/granite-3.1-8b-instruct
modelSrc: https://huggingface.co/ibm-granite/granite-3.1-8b-instruct
modelServerName: ${{ parameters.modelServer }}
Expand Down Expand Up @@ -461,7 +461,7 @@ spec:
# for vllm
vllmSelected: ${{ parameters.modelServer === 'vLLM' }}
vllmModelServiceContainer: quay.io/redhat-ai-dev/vllm-openai-ubi9:v0.11.0
modelName: ${{ parameters.modelName if parameters.modelServer === 'Bring you own model server' else (steps['fetch-model-from-catalog'].output.entity.metadata.name if parameters.modelServer === 'choose-from-the-catalog' else 'ibm-granite/granite-3.1-8b-instruct') }}
modelName: ${{ parameters.modelName if parameters.modelServer === 'Bring you own model server' else (steps['fetch-model-from-catalog'].output.entity.metadata.annotations | pick('rhdh.modelcatalog.io/model-name') if parameters.modelServer === 'choose-from-the-catalog' else 'ibm-granite/granite-3.1-8b-instruct') }}
modelSrc: https://huggingface.co/ibm-granite/granite-3.1-8b-instruct
maxModelLength: 4096
# SED_LLM_SERVER_END
Expand Down
4 changes: 2 additions & 2 deletions templates/codegen/template.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -362,7 +362,7 @@ spec:
appRunCommand: "streamlit run codegen-app.py"
modelServiceContainer: quay.io/redhat-ai-dev/llamacpp_python:0.3.16
modelServicePort: 8001
customModelName: ${{ steps['fetch-model-from-catalog'].output.entity.metadata.name if parameters.modelServer === 'choose-from-the-catalog' else parameters.modelName }}
customModelName: ${{ steps['fetch-model-from-catalog'].output.entity.metadata.annotations | pick('rhdh.modelcatalog.io/model-name') if parameters.modelServer === 'choose-from-the-catalog' else parameters.modelName }}
modelName: TheBloke/Mistral-7B-Instruct-v0.2-AWQ
modelSrc: https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-AWQ
modelServerName: ${{ parameters.modelServer }}
Expand Down Expand Up @@ -461,7 +461,7 @@ spec:
# for vllm
vllmSelected: ${{ parameters.modelServer === 'vLLM' }}
vllmModelServiceContainer: quay.io/redhat-ai-dev/vllm-openai-ubi9:v0.11.0
modelName: ${{ parameters.modelName if parameters.modelServer === 'Bring you own model server' else (steps['fetch-model-from-catalog'].output.entity.metadata.name if parameters.modelServer === 'choose-from-the-catalog' else 'TheBloke/Mistral-7B-Instruct-v0.2-AWQ') }}
modelName: ${{ parameters.modelName if parameters.modelServer === 'Bring you own model server' else (steps['fetch-model-from-catalog'].output.entity.metadata.annotations | pick('rhdh.modelcatalog.io/model-name') if parameters.modelServer === 'choose-from-the-catalog' else 'TheBloke/Mistral-7B-Instruct-v0.2-AWQ') }}
modelSrc: https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-AWQ
maxModelLength: 4096
# SED_LLM_SERVER_END
Expand Down
2 changes: 1 addition & 1 deletion templates/model-server/template.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -248,7 +248,7 @@ spec:
# for vllm
vllmSelected: ${{ parameters.modelServer === 'vLLM' }}
vllmModelServiceContainer: quay.io/redhat-ai-dev/vllm-openai-ubi9:v0.11.0
modelName: ${{ parameters.modelName if parameters.modelServer === 'Bring you own model server' else (steps['fetch-model-from-catalog'].output.entity.metadata.name if parameters.modelServer === 'choose-from-the-catalog' else 'ibm-granite/granite-3.1-8b-instruct') }}
modelName: ${{ parameters.modelName if parameters.modelServer === 'Bring you own model server' else (steps['fetch-model-from-catalog'].output.entity.metadata.annotations | pick('rhdh.modelcatalog.io/model-name') if parameters.modelServer === 'choose-from-the-catalog' else 'ibm-granite/granite-3.1-8b-instruct') }}
modelSrc:
maxModelLength: 4096
# SED_LLM_SERVER_END
Expand Down
2 changes: 1 addition & 1 deletion templates/object-detection/template.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -330,7 +330,7 @@ spec:
appRunCommand: "streamlit run object_detection_client.py"
modelServiceContainer: quay.io/redhat-ai-dev/object_detection_python:latest
modelServicePort: 8000
customModelName: ${{ steps['fetch-model-from-catalog'].output.entity.metadata.name if parameters.modelServer === 'choose-from-the-catalog' else parameters.modelName }}
customModelName: ${{ steps['fetch-model-from-catalog'].output.entity.metadata.annotations | pick('rhdh.modelcatalog.io/model-name') if parameters.modelServer === 'choose-from-the-catalog' else parameters.modelName }}
modelName: facebook/detr-resnet-101
modelSrc: https://huggingface.co/facebook/detr-resnet-101
modelServerName: ${{ parameters.modelServer }}
Expand Down
4 changes: 2 additions & 2 deletions templates/rag/template.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -362,7 +362,7 @@ spec:
appRunCommand: "streamlit run rag_app.py"
modelServiceContainer: quay.io/redhat-ai-dev/llamacpp_python:0.3.16
modelServicePort: 8001
customModelName: ${{ steps['fetch-model-from-catalog'].output.entity.metadata.name if parameters.modelServer === 'choose-from-the-catalog' else parameters.modelName }}
customModelName: ${{ steps['fetch-model-from-catalog'].output.entity.metadata.annotations | pick('rhdh.modelcatalog.io/model-name') if parameters.modelServer === 'choose-from-the-catalog' else parameters.modelName }}
modelName: ibm-granite/granite-3.1-8b-instruct
modelSrc: https://huggingface.co/ibm-granite/granite-3.1-8b-instruct
modelServerName: ${{ parameters.modelServer }}
Expand Down Expand Up @@ -461,7 +461,7 @@ spec:
# for vllm
vllmSelected: ${{ parameters.modelServer === 'vLLM' }}
vllmModelServiceContainer: quay.io/redhat-ai-dev/vllm-openai-ubi9:v0.11.0
modelName: ${{ parameters.modelName if parameters.modelServer === 'Bring you own model server' else (steps['fetch-model-from-catalog'].output.entity.metadata.name if parameters.modelServer === 'choose-from-the-catalog' else 'ibm-granite/granite-3.1-8b-instruct') }}
modelName: ${{ parameters.modelName if parameters.modelServer === 'Bring you own model server' else (steps['fetch-model-from-catalog'].output.entity.metadata.annotations | pick('rhdh.modelcatalog.io/model-name') if parameters.modelServer === 'choose-from-the-catalog' else 'ibm-granite/granite-3.1-8b-instruct') }}
modelSrc: https://huggingface.co/ibm-granite/granite-3.1-8b-instruct
maxModelLength: 4096
# SED_LLM_SERVER_END
Expand Down