K3ai - An opinionated tool to deploy AI platformes at ease
https://docs.k3ai.in - Fully written and mantained by Me
K3ai (keɪ3ai)
K3ai is a lightweight infrastructure-in-a-box specifically built to install and configure AI tools and platforms to quickly experiment and/or run in production over edge devices.
Ready to experiment?
All you have to do is, download the binary for your Operating System, move it to your path (if you like easy things), and use it.
Linux (including Microsoft WSL)
curl -sfL "https://get.k3ai.in" -o k3ai.tar.gz
once downloaded untar the file and move it to your path
tar -xvzf k3ai.tar.gz \
&& chmod +x ./k3ai \
&& sudo mv ./k3ai /usr/local/bin
Windows
Invoke-WebRequest -Uri "https://get-win.k3ai.in" -Ouput k3ai.zip
once downloaded unzip the file and move it to your path or execute it from a folder of your choice (i.e.: k3ai.exe -h)
Expand-Archive -Path .\k3ai.zip
Mac
curl -sfL "https://get-mac.k3ai.in" -o k3ai.tar.gz
once downloaded untar the file and move it to your path
tar -xvzf k3ai.tar.gz \
&& chmod +x ./k3ai \
&& sudo mv ./k3ai /usr/local/bin
Arm64
curl -sfL "https://get-arm.k3ai.in" -o k3ai.tar.gz
once downloaded untar the file and move it to your path
tar -xvzf k3ai.tar.gz \
&& chmod +x ./k3ai \
&& sudo mv ./k3ai /usr/local/bin
Alternative method for Linux
If for any reason it fails just go straight away to https://github.com/kf5i/k3ai-core/releases and download the binary. Place it in your path and that’s it.
or use the following
#Set a variable to grab latest version
Version=$(curl -s "https://api.github.com/repos/kf5i/k3ai-core/releases/latest" | awk -F '"' '/tag_name/{print $4}' | cut -c 2-6)
# get the binaries
wget https://github.com/kf5i/k3ai-core/releases/download/v$Version/k3ai-core_${Version}_linux_amd64.tar.gz
Looking for more interaction? join our Slack channel here****
What we do support:
- Windows
- Linux
- Mac
- ARM
NOTE: Unfortunately not all plugins work with ARM. We will take care of this and make a way to let you know before installing them
Components of K3ai
Currently, we install the following components (the list is changing and growing):
- Kubernetes based on K3s from Rancher: https://k3s.io/
- Kubernetes based on K0s from Mirantis: https://k0sproject.io
- Kubernetes KinD: https://kind.sigs.k8s.io/
- Kubeflow pipelines: https://github.com/kubeflow/pipelines
- Argo Workflows: https://github.com/argoproj/argo
- H2O Community: https://h20.ai
- Kubeflow: https://www.kubeflow.org/ - (coming soon)
- NVIDIA GPU support: https://docs.nvidia.com/datacenter/cloud-native/index.html
- NVIDIA Triton inference server: https://github.com/triton-inference-server/server/tree/master/deploy/single_server (coming soon)
- Tensorflow Serving: https://www.tensorflow.org/tfx/serving/serving_kubernetes:
- ResNet
- Mnist (coming soon)
- and many many others…