* Started work on provider * WIP Adding batch provider * Working basic call into pool client. Need to parameterize the baseurl * Fixed job creation by manipulating the content-type * WIP Kicking off containers. Dirty * [wip] More meat around scheduling simple containers. * Working on basic task wrapper to co-schedule pods * WIP on task wrapper * WIP * Working pod minimal wrapper for batch * Integrate pod template code into provider * Cleaning up * Move to docker without gpu * WIP batch integration * partially working * Working logs * Tidy code * WIP: Testing and readme * Added readme and terraform deployment for GPU Azure Batch pool. * Update to enable low priority nodes for gpu * Fix log formatting bug. Return node logs when container not yet started * Moved to golang v1.10 * Fix cri test * Fix up minor docs Issue. Add provider to readme. Add var for vk image.
21 lines
389 B
YAML
21 lines
389 B
YAML
apiVersion: v1
|
|
kind: Pod
|
|
metadata:
|
|
name: exampegpujob
|
|
spec:
|
|
containers:
|
|
- image: nvidia/cuda
|
|
command: ["nvidia-smi"]
|
|
imagePullPolicy: Always
|
|
name: nvidia
|
|
resources:
|
|
requests:
|
|
memory: 1G
|
|
cpu: 1
|
|
limits:
|
|
nvidia.com/gpu: 1 # requesting 1 GPU
|
|
nodeName: virtual-kubelet
|
|
tolerations:
|
|
- key: azure.com/batch
|
|
effect: NoSchedule
|