Skip to content

Deploying vLLM with Audio and LLM Inference on ROSA with GPUs#668

Open
fjcloud wants to merge 2 commits intomainfrom
fjcloud-vllm
Open

Deploying vLLM with Audio and LLM Inference on ROSA with GPUs#668
fjcloud wants to merge 2 commits intomainfrom
fjcloud-vllm

Conversation

@fjcloud
Copy link
Copy Markdown
Contributor

@fjcloud fjcloud commented Nov 12, 2025

Use Case

Automatically transcribe audio conversations (meetings, customer calls) and analyze content with an LLM to extract insights, decisions, and action items.

Technical stack

ROSA + Red Hat vLLM Build + Red Hat validated models

@netlify
Copy link
Copy Markdown

netlify Bot commented Nov 12, 2025

Deploy Preview for rh-cloud-experts ready!

Name Link
🔨 Latest commit 534e5e2
🔍 Latest deploy log https://app.netlify.com/projects/rh-cloud-experts/deploys/6914a3f6b05b3a0008d7222b
😎 Deploy Preview https://deploy-preview-668--rh-cloud-experts.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify project configuration.

@paulczar
Copy link
Copy Markdown
Contributor

paulczar commented May 5, 2026

Hi @fjcloud,

Thanks for this comprehensive vLLM + ROSA GPU guide! The content is well-structured and clear. I have one request before merging:

Missing frontmatter field: Please add a validated_version field to the frontmatter to indicate which ROSA version(s) this guide was tested on.

For example, if you validated this on ROSA 4.18+, add:

validated_version: "4.18"

This allows us to flag the guide if/when that version reaches EOL and ensures readers know the tested baseline.

Once updated, I can approve and merge. Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants