Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Onboard Katio to Kubernetes services hosted by other cloud vendors #452

Open
7 of 12 tasks
Fei-Guo opened this issue May 30, 2024 · 3 comments
Open
7 of 12 tasks

Onboard Katio to Kubernetes services hosted by other cloud vendors #452

Fei-Guo opened this issue May 30, 2024 · 3 comments
Assignees

Comments

@Fei-Guo
Copy link
Collaborator

Fei-Guo commented May 30, 2024

Tasks

@Fei-Guo
Copy link
Collaborator Author

Fei-Guo commented May 30, 2024

We will start from deploying Kaito in AWS since AWS does support Karpenter, making the integration easier.

@Fei-Guo Fei-Guo changed the title Onboard Katio to Kubernetes services hosted by other cloud vendor Onboard Katio to Kubernetes services hosted by other cloud vendors May 30, 2024
@javaniecampbell
Copy link

javaniecampbell commented Aug 2, 2024

Hey all, just quick question does this feature enhancement will it include self-hosted kubernetes, I checked a few places but wasn't sure so I figured maybe this could be right place to see if this is considered?

The consideration is that some will need for self-host community, home-labs and companies, etc that need the llms to be ran locally.

@Fei-Guo
Copy link
Collaborator Author

Fei-Guo commented Aug 14, 2024

Hey all, just quick question does this feature enhancement will it include self-hosted kubernetes, I checked a few places but wasn't sure so I figured maybe this could be right place to see if this is considered?

The consideration is that some will need for self-host community, home-labs and companies, etc that need the llms to be ran locally.

You can run Kaito in selfmanaged k8s if you already add GPU nodes in the cluster (with proper gpu driver and k8s plugin installed). In this case, you can just add those nodes in the Kaito workspace CR as preferrednodes in the Resource spec. Kaito will skip provisioning gpu nodes and just run inference workload in the existing nodes.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
Status: In Progress
Development

No branches or pull requests

3 participants