diff --git a/.circleci/config.yml b/.circleci/config.yml index 2420fe113..47047e282 100644 --- a/.circleci/config.yml +++ b/.circleci/config.yml @@ -3,7 +3,7 @@ jobs: build: docker: # specify the version - - image: circleci/golang:1.9 + - image: circleci/golang:1.10 working_directory: /go/src/github.com/virtual-kubelet/virtual-kubelet steps: diff --git a/.dockerignore b/.dockerignore new file mode 100644 index 000000000..935cb2a78 --- /dev/null +++ b/.dockerignore @@ -0,0 +1,4 @@ +.vscode +private.env +*.private.* +providers/azurebatch/deployment/ \ No newline at end of file diff --git a/.gitignore b/.gitignore index d871221bc..c69176432 100644 --- a/.gitignore +++ b/.gitignore @@ -5,6 +5,9 @@ *.dylib bin/ +# Certificates +*.pem + # Test binary, build with `go test -c` *.test @@ -24,4 +27,10 @@ bin/ credentials.json # VS Code files -.vscode/ \ No newline at end of file +.vscode/ + +# Terraform ignores +**/.terraform/** +**/terraform-provider-kubernetes +**/*.tfstate* +debug diff --git a/Gopkg.lock b/Gopkg.lock index 9ea1fb35d..7d6746289 100644 --- a/Gopkg.lock +++ b/Gopkg.lock @@ -1,6 +1,15 @@ # This file is autogenerated, do not edit; changes may be undone by the next 'dep ensure'. +[[projects]] + name = "github.com/Azure/azure-sdk-for-go" + packages = [ + "services/batch/2017-09-01.6.0/batch", + "version" + ] + revision = "51228ee60b238eaa5a9d53adc82ebd8b321bca4f" + version = "v15.1.1" + [[projects]] branch = "master" name = "github.com/Azure/go-ansiterm" @@ -13,8 +22,12 @@ [[projects]] name = "github.com/Azure/go-autorest" packages = [ + "autorest", "autorest/adal", - "autorest/date" + "autorest/azure", + "autorest/date", + "autorest/to", + "autorest/validation" ] revision = "eaa7994b2278094c904d31993d26f56324db3052" version = "v10.8.1" @@ -502,6 +515,12 @@ packages = ["."] revision = "7cafcd837844e784b526369c9bce262804aebc60" +[[projects]] + name = "github.com/lawrencegripper/pod2docker" + packages = ["."] + revision = "9a9182c0c682798fe214fa11df7bfca83048d433" + version = "v0.5.1" + [[projects]] name = "github.com/magiconair/properties" packages = ["."] @@ -576,6 +595,12 @@ revision = "792786c7400a136282c1664665ae0a8db921c6c2" version = "v1.0.0" +[[projects]] + name = "github.com/satori/go.uuid" + packages = ["."] + revision = "f58768cc1a7a7e77a3bd49e98cdd21419399b6a3" + version = "v1.2.0" + [[projects]] name = "github.com/spf13/afero" packages = [ @@ -1049,6 +1074,6 @@ [solve-meta] analyzer-name = "dep" analyzer-version = 1 - inputs-digest = "6125e24314e26b6b5fb9a7cd98e4f23fa2e551d882485d1ff5f481abdfd792aa" + inputs-digest = "8f3ca80656c580339bc7105912693cf126c4952c476e7052cbd9de4ef73d9e75" solver-name = "gps-cdcl" solver-version = 1 diff --git a/Gopkg.toml b/Gopkg.toml index b4c742831..d0dd2053a 100644 --- a/Gopkg.toml +++ b/Gopkg.toml @@ -55,6 +55,14 @@ name = "github.com/hyperhq/hypercli" revision = "29217d318cab52815518a1126d57ca010de83e4d" +[[constraint]] + name = "github.com/Azure/azure-sdk-for-go" + version = "15.1.1" + +[[constraint]] + name = "github.com/lawrencegripper/pod2docker" + version = "0.5.1" + [[constraint]] name = "github.com/aws/aws-sdk-go" version = "1.13.38" diff --git a/README.md b/README.md index 195a83036..f03fedab2 100644 --- a/README.md +++ b/README.md @@ -20,6 +20,7 @@ The best description is "Kubernetes API on top, programmable back." * [Usage](#usage) * [Providers](#providers) + [Azure Container Instances Provider](#azure-container-instances-provider) + + [Azure Batch GPU Provider](./providers/azurebatch/README.md) + [AWS Fargate Provider](#aws-fargate-provider) + [Hyper.sh Provider](#hypersh-provider) + [Adding a New Provider via the Provider Interface](#adding-a-new-provider-via-the-provider-interface) diff --git a/charts/virtual-kubelet-0.1.2.tgz b/charts/virtual-kubelet-0.1.2.tgz new file mode 100644 index 000000000..87a37df24 Binary files /dev/null and b/charts/virtual-kubelet-0.1.2.tgz differ diff --git a/charts/virtual-kubelet-for-aks-0.1.4.tgz b/charts/virtual-kubelet-for-aks-0.1.4.tgz new file mode 100644 index 000000000..16d1008d8 Binary files /dev/null and b/charts/virtual-kubelet-for-aks-0.1.4.tgz differ diff --git a/charts/virtual-kubelet-for-aks-latest.tgz b/charts/virtual-kubelet-for-aks-latest.tgz new file mode 100644 index 000000000..16d1008d8 Binary files /dev/null and b/charts/virtual-kubelet-for-aks-latest.tgz differ diff --git a/charts/virtual-kubelet-for-aks/Chart.yaml b/charts/virtual-kubelet-for-aks/Chart.yaml index c92909014..e88317184 100644 --- a/charts/virtual-kubelet-for-aks/Chart.yaml +++ b/charts/virtual-kubelet-for-aks/Chart.yaml @@ -1,5 +1,5 @@ name: virtual-kubelet-for-aks -version: 0.1.3 +version: 0.1.4 description: a Helm chart to install virtual kubelet in an AKS or ACS cluster. sources: - https://github.com/virtual-kubelet/virtual-kubelet diff --git a/charts/virtual-kubelet-for-aks/templates/deployment.yaml b/charts/virtual-kubelet-for-aks/templates/deployment.yaml index fbc9d1727..e7c2269bc 100644 --- a/charts/virtual-kubelet-for-aks/templates/deployment.yaml +++ b/charts/virtual-kubelet-for-aks/templates/deployment.yaml @@ -47,7 +47,7 @@ spec: - name: acs-credential mountPath: "/etc/acs/azure.json" command: ["virtual-kubelet"] - args: ["--provider", "azure", "--namespace", "default", "--nodename", {{ default "virtual-kubelet" .Values.env.nodeName | quote }} , "--os", {{ default "Linux" .Values.env.nodeOsType | quote }}, "--taint", {{ default "azure.com/aci" .Values.env.nodeTaint | quote }}] + args: ["--provider", "azure", "--namespace", {{ default "" .Values.env.monitoredNamespace | quote }}, "--nodename", {{ default "virtual-kubelet" .Values.env.nodeName | quote }} , "--os", {{ default "Linux" .Values.env.nodeOsType | quote }}, "--taint", {{ default "azure.com/aci" .Values.env.nodeTaint | quote }}] volumes: - name: credentials secret: diff --git a/charts/virtual-kubelet-for-aks/values.yaml b/charts/virtual-kubelet-for-aks/values.yaml index f2b3bfc2a..7dd131fcb 100644 --- a/charts/virtual-kubelet-for-aks/values.yaml +++ b/charts/virtual-kubelet-for-aks/values.yaml @@ -14,3 +14,4 @@ env: nodeOsType: apiserverCert: apiserverKey: + monitoredNamespace: diff --git a/charts/virtual-kubelet-latest.tgz b/charts/virtual-kubelet-latest.tgz new file mode 100644 index 000000000..87a37df24 Binary files /dev/null and b/charts/virtual-kubelet-latest.tgz differ diff --git a/charts/virtual-kubelet/Chart.yaml b/charts/virtual-kubelet/Chart.yaml index e6b645e3c..1542a2e5f 100644 --- a/charts/virtual-kubelet/Chart.yaml +++ b/charts/virtual-kubelet/Chart.yaml @@ -1,5 +1,5 @@ name: virtual-kubelet -version: 0.1.1 +version: 0.1.2 description: a Helm chart to install virtual kubelet inside a Kubernetes cluster. sources: - https://github.com/virtual-kubelet/virtual-kubelet diff --git a/charts/virtual-kubelet/templates/deployment.yaml b/charts/virtual-kubelet/templates/deployment.yaml index 0206d4bbb..41ba5b766 100644 --- a/charts/virtual-kubelet/templates/deployment.yaml +++ b/charts/virtual-kubelet/templates/deployment.yaml @@ -34,7 +34,7 @@ spec: - name: credentials mountPath: "/etc/virtual-kubelet" command: ["virtual-kubelet"] - args: ["--provider", "azure", "--namespace", "default", "--nodename", {{ default "virtual-kubelet" .Values.env.nodeName | quote }} , "--os", {{ default "Linux" .Values.env.nodeOsType | quote }}, "--taint", {{ default "azure.com/aci" .Values.env.nodeTaint | quote }}] + args: ["--provider", "azure", "--namespace", {{ default "" .Values.env.monitoredNamespace | quote }}, "--nodename", {{ default "virtual-kubelet" .Values.env.nodeName | quote }} , "--os", {{ default "Linux" .Values.env.nodeOsType | quote }}, "--taint", {{ default "azure.com/aci" .Values.env.nodeTaint | quote }}] volumes: - name: credentials secret: diff --git a/charts/virtual-kubelet/values.yaml b/charts/virtual-kubelet/values.yaml index 48dba14d0..b0426297a 100644 --- a/charts/virtual-kubelet/values.yaml +++ b/charts/virtual-kubelet/values.yaml @@ -14,6 +14,7 @@ env: nodeOsType: apiserverCert: apiserverKey: + monitoredNamespace: # Install Default RBAC roles and bindings rbac: diff --git a/examples/busyecho.yaml b/examples/busyecho.yaml new file mode 100644 index 000000000..0c400830a --- /dev/null +++ b/examples/busyecho.yaml @@ -0,0 +1,29 @@ +apiVersion: v1 +kind: Pod +metadata: + name: busyecho +spec: + containers: + - image: busybox + command: + - echo "Hello from container" + imagePullPolicy: Always + name: nginx + resources: + requests: + memory: 1G + cpu: 1 + ports: + - containerPort: 80 + name: http + protocol: TCP + - containerPort: 443 + name: https + dnsPolicy: ClusterFirst + nodeSelector: + kubernetes.io/role: agent + beta.kubernetes.io/os: linux + type: virtual-kubelet + tolerations: + - key: azure.com/aci + effect: NoSchedule diff --git a/examples/iis-pod.yaml b/examples/iis-pod.yaml index da91337f5..230249120 100644 --- a/examples/iis-pod.yaml +++ b/examples/iis-pod.yaml @@ -22,7 +22,9 @@ spec: name: https dnsPolicy: Default nodeSelector: - kubernetes.io/hostname: virtual-kubelet-myaciconnector-win + kubernetes.io/role: agent + beta.kubernetes.io/os: windows + type: virtual-kubelet automountServiceAccountToken: false tolerations: - key: azure.com/aci diff --git a/examples/nanoserver.yaml b/examples/nanoserver.yaml index 6cafed4c6..2097c50f5 100644 --- a/examples/nanoserver.yaml +++ b/examples/nanoserver.yaml @@ -11,7 +11,9 @@ spec: dnsPolicy: Default restartPolicy: OnFailure nodeSelector: - kubernetes.io/hostname: virtual-kubelet-myaciconnector-win + kubernetes.io/role: agent + beta.kubernetes.io/os: windows + type: virtual-kubelet automountServiceAccountToken: false tolerations: - key: azure.com/aci diff --git a/examples/nginx-pod.yaml b/examples/nginx-pod.yaml index 4e2a6bc4e..c4bf7631e 100644 --- a/examples/nginx-pod.yaml +++ b/examples/nginx-pod.yaml @@ -15,7 +15,9 @@ spec: name: https dnsPolicy: ClusterFirst nodeSelector: - kubernetes.io/hostname: virtual-kubelet-myaciconnector-linux + kubernetes.io/role: agent + beta.kubernetes.io/os: linux + type: virtual-kubelet tolerations: - key: azure.com/aci effect: NoSchedule diff --git a/examples/pause.yaml b/examples/pause.yaml new file mode 100644 index 000000000..19e1372e8 --- /dev/null +++ b/examples/pause.yaml @@ -0,0 +1,27 @@ +apiVersion: v1 +kind: Pod +metadata: + name: pauseexample1 +spec: + containers: + - name: pauseexample1 + imagePullPolicy: Always + image: gcr.io/google_containers/pause:1.0 + resources: + requests: + memory: 1G + cpu: 1 + ports: + - containerPort: 80 + name: http + protocol: TCP + - containerPort: 443 + name: https + dnsPolicy: ClusterFirst + nodeSelector: + kubernetes.io/role: agent + beta.kubernetes.io/os: linux + type: virtual-kubelet + tolerations: + - key: azure.com/aci + effect: NoSchedule diff --git a/providers/azure/aci.go b/providers/azure/aci.go index 6bdf5d55a..f01c4188f 100644 --- a/providers/azure/aci.go +++ b/providers/azure/aci.go @@ -173,10 +173,22 @@ func NewACIProvider(config string, rm *manager.ResourceManager, nodeName, operat } // Set sane defaults for Capacity in case config is not supplied - p.cpu = "20" - p.memory = "100Gi" - p.pods = "20" + p.cpu = "1000" + p.memory = "4Ti" + p.pods = "1000" + if cpuQuota := os.Getenv("ACI_QUOTA_CPU"); cpuQuota != "" { + p.cpu = cpuQuota + } + + if memoryQuota := os.Getenv("ACI_QUOTA_MEMORY"); memoryQuota != "" { + p.memory = memoryQuota + } + + if podsQuota := os.Getenv("ACI_QUOTA_POD"); podsQuota != "" { + p.pods = podsQuota + } + p.operatingSystem = operatingSystem p.nodeName = nodeName p.internalIP = internalIP diff --git a/providers/azurebatch/README.md b/providers/azurebatch/README.md new file mode 100644 index 000000000..02e5cbc29 --- /dev/null +++ b/providers/azurebatch/README.md @@ -0,0 +1,79 @@ +# Kubernetes Virtual Kubelet with Azure Batch + +[Azure Batch](https://docs.microsoft.com/en-us/azure/batch/) provides a HPC Computing environment in Azure for distributed tasks. Azure Batch handles scheduling decrete jobs and tasks accross pools of VM's. It is commonly used for batch processing tasks such as rendering. + +The Virtual kubelet integration allows you to take advantage of this from within Kubernetes. The primary usecase for the provider is to make it easy to use GPU based workload from normal Kubernetes clusters. For example, creating Kubernetes Jobs which train or execute ML models using Nvidia GPU's or using FFMPEG. + +Azure Batch allows for [low priority nodes](https://docs.microsoft.com/en-us/azure/batch/batch-low-pri-vms) which can also help to reduce cost for non-time sensitive workloads. + +__The [ACI provider](../azure/README.MD) is the best option unless you're looking to utilise some specific features of Azure Batch__. + +## Status: Experimental + +This provider is currently in the exterimental stages. Contributions welcome! + +## Quick Start + +The following Terraform template deploys an AKS cluster with the Virtual Kubelet, Azure Batch Account and GPU enabled Azure Batch pool. The Batch pool contains 1 Dedicated NC6 Node and 2 Low Priority NC6 Nodes. + +1. Setup Terraform for Azure following [this guide here](https://docs.microsoft.com/en-us/azure/virtual-machines/linux/terraform-install-configure) +2. From the commandline move to the deployment folder `cd ./providers/azurebatch/deployment` then edit `vars.example.tfvars` adding in your Service Principal details +3. Download the latest version of the Community Kubernetes Provider for Terraform. Get the correct link [from here](https://github.com/sl1pm4t/terraform-provider-kubernetes/releases) and use it as follows: (Current official Terraform K8s provider doesn't support `Deployments`) + +```shell +curl -L -o - PUT_RELASE_BINARY_LINK_YOU_FOUND_HERE | gunzip > terraform-provider-kubernetes +chmod +x ./terraform-provider-kubernetes +``` + +4. Use `terraform init` to initialize the template +5. Use `terraform plan -var-file=./vars.example.tfvars` and `terraform apply -var-file=./vars.example.tfvars` to deploy the template +6. Run `kubectl describe deployment/vkdeployment` to check the virtual kubelet is running correctly. +7. Run `kubectl create -f examplegpupod.yaml` +8. Run `pods=$(kubectl get pods --selector=app=examplegpupod --show-all --output=jsonpath={.items..metadata.name})` then `kubectl logs $pods` to view the logs. Should see: + +```text + [Vector addition of 50000 elements] + Copy input data from the host memory to the CUDA device + CUDA kernel launch with 196 blocks of 256 threads + Copy output data from the CUDA device to the host memory + Test PASSED + Done +``` + +### Tweaking the Quickstart + +You can update [main.tf](./main.tf) to increase the number of nodes allocated to the Azure Batch pool or update [./aks/main.tf](./aks/main.tf) to increase the number of agent nodes allocated to your AKS cluster. + +## Advanced Setup + +## Prerequistes + +1. An Azure Batch Account configurated +2. An Azure Batch Pool created with necessary VM spec. VM's in the pool must have: + - `docker` installed and correctly configured + - `nvidia-docker` and `cuda` drivers installed +3. K8s cluster +4. Azure Service Principal with access to the Azure Batch Account + +## Setup + +The provider expects the following environment variables to be configured: + +``` + ClientID: AZURE_CLIENT_ID + ClientSecret: AZURE_CLIENT_SECRET + ResourceGroup: AZURE_RESOURCE_GROUP + SubscriptionID: AZURE_SUBSCRIPTION_ID + TenantID: AZURE_TENANT_ID + PoolID: AZURE_BATCH_POOLID + JobID (optional):AZURE_BATCH_JOBID + AccountLocation: AZURE_BATCH_ACCOUNT_LOCATION + AccountName: AZURE_BATCH_ACCOUNT_NAME +``` + +## Running + +The provider will assign pods to machines in the Azure Batch Pool. Each machine can, by default, process only one pod at a time +running more than 1 pod per machine isn't currently supported and will result in errors. + +Azure Batch queues tasks when no machines are available so pods will sit in `podPending` state while waiting for a VM to become available. \ No newline at end of file diff --git a/providers/azurebatch/batch.go b/providers/azurebatch/batch.go new file mode 100644 index 000000000..336d3dc61 --- /dev/null +++ b/providers/azurebatch/batch.go @@ -0,0 +1,409 @@ +package azurebatch + +import ( + "context" + "encoding/json" + "fmt" + "github.com/Azure/go-autorest/autorest" + "os" + "strings" + + "io/ioutil" + "log" + "net/http" + + "github.com/Azure/go-autorest/autorest/azure" + + "github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch" + "github.com/Azure/go-autorest/autorest/to" + "github.com/lawrencegripper/pod2docker" + "github.com/virtual-kubelet/virtual-kubelet/manager" + azureCreds "github.com/virtual-kubelet/virtual-kubelet/providers/azure" + "k8s.io/api/core/v1" + "k8s.io/apimachinery/pkg/api/resource" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" +) + +const ( + podJSONKey string = "virtualkubelet_pod" +) + +// Provider the base struct for the Azure Batch provider +type Provider struct { + batchConfig *Config + ctx context.Context + cancelCtx context.CancelFunc + fileClient *batch.FileClient + resourceManager *manager.ResourceManager + listTasks func() (*[]batch.CloudTask, error) + addTask func(batch.TaskAddParameter) (autorest.Response, error) + getTask func(taskID string) (batch.CloudTask, error) + deleteTask func(taskID string) (autorest.Response, error) + getFileFromTask func(taskID, path string) (result batch.ReadCloser, err error) + nodeName string + operatingSystem string + cpu string + memory string + pods string + internalIP string + daemonEndpointPort int32 +} + +// Config - Basic azure config used to interact with ARM resources. +type Config struct { + ClientID string + ClientSecret string + SubscriptionID string + TenantID string + ResourceGroup string + PoolID string + JobID string + AccountName string + AccountLocation string +} + +// NewBatchProvider Creates a batch provider +func NewBatchProvider(configString string, rm *manager.ResourceManager, nodeName, operatingSystem string, internalIP string, daemonEndpointPort int32) (*Provider, error) { + fmt.Println("Starting create provider") + + config := &Config{} + if azureCredsFilepath := os.Getenv("AZURE_CREDENTIALS_LOCATION"); azureCredsFilepath != "" { + creds, err := azureCreds.NewAcsCredential(azureCredsFilepath) + if err != nil { + return nil, err + } + config.ClientID = creds.ClientID + config.ClientSecret = creds.ClientSecret + config.SubscriptionID = creds.SubscriptionID + config.TenantID = creds.TenantID + } + + err := getAzureConfigFromEnv(config) + if err != nil { + log.Println("Failed to get auth information") + } + + return NewBatchProviderFromConfig(config, rm, nodeName, operatingSystem, internalIP, daemonEndpointPort) +} + +// NewBatchProviderFromConfig Creates a batch provider +func NewBatchProviderFromConfig(config *Config, rm *manager.ResourceManager, nodeName, operatingSystem string, internalIP string, daemonEndpointPort int32) (*Provider, error) { + p := Provider{} + p.batchConfig = config + // Set sane defaults for Capacity in case config is not supplied + p.cpu = "20" + p.memory = "100Gi" + p.pods = "20" + p.resourceManager = rm + p.operatingSystem = operatingSystem + p.nodeName = nodeName + p.internalIP = internalIP + p.daemonEndpointPort = daemonEndpointPort + p.ctx, p.cancelCtx = context.WithCancel(context.Background()) + + auth := getAzureADAuthorizer(config, azure.PublicCloud.BatchManagementEndpoint) + + batchBaseURL := getBatchBaseURL(config.AccountName, config.AccountLocation) + _, err := getPool(p.ctx, batchBaseURL, config.PoolID, auth) + if err != nil { + log.Panicf("Error retreiving Azure Batch pool: %v", err) + } + _, err = createOrGetJob(p.ctx, batchBaseURL, config.JobID, config.PoolID, auth) + if err != nil { + log.Panicf("Error retreiving/creating Azure Batch job: %v", err) + } + taskClient := batch.NewTaskClientWithBaseURI(batchBaseURL) + taskClient.Authorizer = auth + p.listTasks = func() (*[]batch.CloudTask, error) { + res, err := taskClient.List(p.ctx, config.JobID, "", "", "", nil, nil, nil, nil, nil) + if err != nil { + return &[]batch.CloudTask{}, err + } + currentTasks := res.Values() + for res.NotDone() { + err = res.Next() + if err != nil { + return &[]batch.CloudTask{}, err + } + pageTasks := res.Values() + if pageTasks != nil || len(pageTasks) != 0 { + currentTasks = append(currentTasks, pageTasks...) + } + } + + return ¤tTasks, nil + } + p.addTask = func(task batch.TaskAddParameter) (autorest.Response, error) { + return taskClient.Add(p.ctx, config.JobID, task, nil, nil, nil, nil) + } + p.getTask = func(taskID string) (batch.CloudTask, error) { + return taskClient.Get(p.ctx, config.JobID, taskID, "", "", nil, nil, nil, nil, "", "", nil, nil) + } + p.deleteTask = func(taskID string) (autorest.Response, error) { + return taskClient.Delete(p.ctx, config.JobID, taskID, nil, nil, nil, nil, "", "", nil, nil) + } + p.getFileFromTask = func(taskID, path string) (batch.ReadCloser, error) { + return p.fileClient.GetFromTask(p.ctx, config.JobID, taskID, path, nil, nil, nil, nil, "", nil, nil) + } + + fileClient := batch.NewFileClientWithBaseURI(batchBaseURL) + fileClient.Authorizer = auth + p.fileClient = &fileClient + + return &p, nil +} + +// CreatePod accepts a Pod definition +func (p *Provider) CreatePod(pod *v1.Pod) error { + log.Println("Creating pod...") + podCommand, err := pod2docker.GetBashCommand(pod2docker.PodComponents{ + InitContainers: pod.Spec.InitContainers, + Containers: pod.Spec.Containers, + PodName: pod.Name, + Volumes: pod.Spec.Volumes, + }) + if err != nil { + return err + } + + bytes, err := json.Marshal(pod) + if err != nil { + panic(err) + } + + task := batch.TaskAddParameter{ + DisplayName: to.StringPtr(string(pod.UID)), + ID: to.StringPtr(getTaskIDForPod(pod.Namespace, pod.Name)), + CommandLine: to.StringPtr(fmt.Sprintf(`/bin/bash -c "%s"`, podCommand)), + UserIdentity: &batch.UserIdentity{ + AutoUser: &batch.AutoUserSpecification{ + ElevationLevel: batch.Admin, + Scope: batch.Pool, + }, + }, + EnvironmentSettings: &[]batch.EnvironmentSetting{ + { + Name: to.StringPtr(podJSONKey), + Value: to.StringPtr(string(bytes)), + }, + }, + } + _, err = p.addTask(task) + if err != nil { + return err + } + + return nil +} + +// GetPodStatus retrieves the status of a given pod by name. +func (p *Provider) GetPodStatus(namespace, name string) (*v1.PodStatus, error) { + log.Println("Getting pod status ....") + pod, err := p.GetPod(namespace, name) + + if err != nil { + return nil, err + } + if pod == nil { + return nil, nil + } + return &pod.Status, nil +} + +// UpdatePod accepts a Pod definition +func (p *Provider) UpdatePod(pod *v1.Pod) error { + log.Println("Pod Update called: No-op as not implemented") + return nil +} + +// DeletePod accepts a Pod definition +func (p *Provider) DeletePod(pod *v1.Pod) error { + taskID := getTaskIDForPod(pod.Namespace, pod.Name) + task, err := p.deleteTask(taskID) + if err != nil { + log.Println(task) + log.Println(err) + return err + } + + log.Printf(fmt.Sprintf("Deleting task: %v", taskID)) + return nil +} + +// GetPod returns a pod by name +func (p *Provider) GetPod(namespace, name string) (*v1.Pod, error) { + log.Println("Getting Pod ...") + task, err := p.getTask(getTaskIDForPod(namespace, name)) + if err != nil { + if task.Response.StatusCode == http.StatusNotFound { + return nil, nil + } + log.Println(err) + return nil, err + } + + pod, err := getPodFromTask(&task) + if err != nil { + panic(err) + } + + status, _ := convertTaskToPodStatus(&task) + pod.Status = *status + + return pod, nil +} + +// GetContainerLogs returns the logs of a container running in a pod by name. +func (p *Provider) GetContainerLogs(namespace, podName, containerName string, tail int) (string, error) { + log.Println("Getting pod logs ....") + + taskID := getTaskIDForPod(namespace, podName) + logFileLocation := fmt.Sprintf("wd/%s.log", containerName) + containerLogReader, err := p.getFileFromTask(taskID, logFileLocation) + + if containerLogReader.Response.Response != nil && containerLogReader.StatusCode == http.StatusNotFound { + stdoutReader, err := p.getFileFromTask(taskID, "stdout.txt") + if err != nil { + return "", err + } + stderrReader, err := p.getFileFromTask(taskID, "stderr.txt") + if err != nil { + return "", err + } + + var builder strings.Builder + builderPtr := &builder + mustWriteString(builderPtr, "Container still starting....\n") + mustWriteString(builderPtr, "Showing startup logs from Azure Batch node instead:\n") + mustWriteString(builderPtr, "----- STDOUT -----\n") + stdoutBytes, _ := ioutil.ReadAll(*stdoutReader.Value) + mustWrite(builderPtr, stdoutBytes) + mustWriteString(builderPtr, "\n") + + mustWriteString(builderPtr, "----- STDERR -----\n") + stderrBytes, _ := ioutil.ReadAll(*stderrReader.Value) + mustWrite(builderPtr, stderrBytes) + mustWriteString(builderPtr, "\n") + + return builder.String(), nil + } + + if err != nil { + return "", err + } + + result, err := formatLogJSON(containerLogReader) + if err != nil { + return "", fmt.Errorf("Container log formating failed err: %v", err) + } + + return result, nil +} + +// GetPods retrieves a list of all pods scheduled to run. +func (p *Provider) GetPods() ([]*v1.Pod, error) { + log.Println("Getting pods...") + tasksPtr, err := p.listTasks() + if err != nil { + panic(err) + } + if tasksPtr == nil { + return []*v1.Pod{}, nil + } + + tasks := *tasksPtr + + pods := make([]*v1.Pod, len(tasks), len(tasks)) + for i, t := range tasks { + pod, err := getPodFromTask(&t) + if err != nil { + panic(err) + } + pods[i] = pod + } + return pods, nil +} + +// Capacity returns a resource list containing the capacity limits +func (p *Provider) Capacity() v1.ResourceList { + return v1.ResourceList{ + "cpu": resource.MustParse(p.cpu), + "memory": resource.MustParse(p.memory), + "pods": resource.MustParse(p.pods), + "nvidia.com/gpu": resource.MustParse("1"), + } +} + +// NodeConditions returns a list of conditions (Ready, OutOfDisk, etc), for updates to the node status +// within Kubernetes. +func (p *Provider) NodeConditions() []v1.NodeCondition { + return []v1.NodeCondition{ + { + Type: "Ready", + Status: v1.ConditionTrue, + LastHeartbeatTime: metav1.Now(), + LastTransitionTime: metav1.Now(), + Reason: "KubeletReady", + Message: "kubelet is ready.", + }, + { + Type: "OutOfDisk", + Status: v1.ConditionFalse, + LastHeartbeatTime: metav1.Now(), + LastTransitionTime: metav1.Now(), + Reason: "KubeletHasSufficientDisk", + Message: "kubelet has sufficient disk space available", + }, + { + Type: "MemoryPressure", + Status: v1.ConditionFalse, + LastHeartbeatTime: metav1.Now(), + LastTransitionTime: metav1.Now(), + Reason: "KubeletHasSufficientMemory", + Message: "kubelet has sufficient memory available", + }, + { + Type: "DiskPressure", + Status: v1.ConditionFalse, + LastHeartbeatTime: metav1.Now(), + LastTransitionTime: metav1.Now(), + Reason: "KubeletHasNoDiskPressure", + Message: "kubelet has no disk pressure", + }, + { + Type: "NetworkUnavailable", + Status: v1.ConditionFalse, + LastHeartbeatTime: metav1.Now(), + LastTransitionTime: metav1.Now(), + Reason: "RouteCreated", + Message: "RouteController created a route", + }, + } +} + +// NodeAddresses returns a list of addresses for the node status +// within Kubernetes. +func (p *Provider) NodeAddresses() []v1.NodeAddress { + // TODO: Make these dynamic and augment with custom ACI specific conditions of interest + return []v1.NodeAddress{ + { + Type: "InternalIP", + Address: p.internalIP, + }, + } +} + +// NodeDaemonEndpoints returns NodeDaemonEndpoints for the node status +// within Kubernetes. +func (p *Provider) NodeDaemonEndpoints() *v1.NodeDaemonEndpoints { + return &v1.NodeDaemonEndpoints{ + KubeletEndpoint: v1.DaemonEndpoint{ + Port: p.daemonEndpointPort, + }, + } +} + +// OperatingSystem returns the operating system for this provider. +func (p *Provider) OperatingSystem() string { + return p.operatingSystem +} diff --git a/providers/azurebatch/batch_helpers.go b/providers/azurebatch/batch_helpers.go new file mode 100644 index 000000000..e8843f779 --- /dev/null +++ b/providers/azurebatch/batch_helpers.go @@ -0,0 +1,216 @@ +package azurebatch + +import ( + "bufio" + "context" + "crypto/md5" + "encoding/json" + "fmt" + "log" + "os" + "strings" + "time" + + "github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/adal" + "github.com/Azure/go-autorest/autorest/azure" +) + +func mustWriteString(builder *strings.Builder, s string) { + _, err := builder.WriteString(s) + if err != nil { + panic(err) + } +} + +func mustWrite(builder *strings.Builder, b []byte) { + _, err := builder.Write(b) + if err != nil { + panic(err) + } +} + +// NewServicePrincipalTokenFromCredentials creates a new ServicePrincipalToken using values of the +// passed credentials map. +func newServicePrincipalTokenFromCredentials(c *Config, scope string) (*adal.ServicePrincipalToken, error) { + oauthConfig, err := adal.NewOAuthConfig(azure.PublicCloud.ActiveDirectoryEndpoint, c.TenantID) + if err != nil { + panic(err) + } + return adal.NewServicePrincipalToken(*oauthConfig, c.ClientID, c.ClientSecret, scope) +} + +// GetAzureADAuthorizer return an authorizor for Azure SP +func getAzureADAuthorizer(c *Config, azureEndpoint string) autorest.Authorizer { + spt, err := newServicePrincipalTokenFromCredentials(c, azureEndpoint) + if err != nil { + panic(fmt.Sprintf("Failed to create authorizer: %v", err)) + } + auth := autorest.NewBearerAuthorizer(spt) + return auth +} + +func getPool(ctx context.Context, batchBaseURL, poolID string, auth autorest.Authorizer) (*batch.PoolClient, error) { + poolClient := batch.NewPoolClientWithBaseURI(batchBaseURL) + poolClient.Authorizer = auth + poolClient.RetryAttempts = 0 + + pool, err := poolClient.Get(ctx, poolID, "*", "", nil, nil, nil, nil, "", "", nil, nil) + + // If we observe an error which isn't related to the pool not existing panic. + // 404 is expected if this is first run. + if err != nil && pool.Response.Response == nil { + log.Printf("Failed to get pool. nil response %v", poolID) + return nil, err + } else if err != nil && pool.StatusCode == 404 { + log.Printf("Pool doesn't exist 404 received Error: %v PoolID: %v", err, poolID) + return nil, err + } else if err != nil { + log.Printf("Failed to get pool. Response:%v", pool.Response) + return nil, err + } + + if pool.State == batch.PoolStateActive { + log.Println("Pool active and running...") + return &poolClient, nil + } + return nil, fmt.Errorf("Pool not in active state: %v", pool.State) +} + +func createOrGetJob(ctx context.Context, batchBaseURL, jobID, poolID string, auth autorest.Authorizer) (*batch.JobClient, error) { + jobClient := batch.NewJobClientWithBaseURI(batchBaseURL) + jobClient.Authorizer = auth + // check if job exists already + currentJob, err := jobClient.Get(ctx, jobID, "", "", nil, nil, nil, nil, "", "", nil, nil) + + if err == nil && currentJob.State == batch.JobStateActive { + log.Println("Wrapper job already exists...") + return &jobClient, nil + } else if currentJob.Response.StatusCode == 404 { + + log.Println("Wrapper job missing... creating...") + wrapperJob := batch.JobAddParameter{ + ID: &jobID, + PoolInfo: &batch.PoolInformation{ + PoolID: &poolID, + }, + } + + _, err := jobClient.Add(ctx, wrapperJob, nil, nil, nil, nil) + + if err != nil { + return nil, err + } + return &jobClient, nil + + } else if currentJob.State == batch.JobStateDeleting { + log.Printf("Job is being deleted... Waiting then will retry") + time.Sleep(time.Minute) + return createOrGetJob(ctx, batchBaseURL, jobID, poolID, auth) + } + + return nil, err +} + +func getBatchBaseURL(batchAccountName, batchAccountLocation string) string { + return fmt.Sprintf("https://%s.%s.batch.azure.com", batchAccountName, batchAccountLocation) +} + +func envHasValue(env string) bool { + val := os.Getenv(env) + if val == "" { + return false + } + return true +} + +// GetConfigFromEnv - Retreives the azure configuration from environment variables. +func getAzureConfigFromEnv(config *Config) error { + if envHasValue("AZURE_CLIENT_ID") { + config.ClientID = os.Getenv("AZURE_CLIENT_ID") + } + if envHasValue("AZURE_CLIENT_SECRET") { + config.ClientSecret = os.Getenv("AZURE_CLIENT_SECRET") + } + if envHasValue("AZURE_RESOURCE_GROUP") { + config.ResourceGroup = os.Getenv("AZURE_RESOURCE_GROUP") + } + if envHasValue("AZURE_SUBSCRIPTION_ID") { + config.SubscriptionID = os.Getenv("AZURE_SUBSCRIPTION_ID") + } + if envHasValue("AZURE_TENANT_ID") { + config.TenantID = os.Getenv("AZURE_TENANT_ID") + } + if envHasValue("AZURE_BATCH_POOLID") { + config.PoolID = os.Getenv("AZURE_BATCH_POOLID") + } + if envHasValue("AZURE_BATCH_JOBID") { + config.JobID = os.Getenv("AZURE_BATCH_JOBID") + } + if envHasValue("AZURE_BATCH_ACCOUNT_LOCATION") { + config.AccountLocation = os.Getenv("AZURE_BATCH_ACCOUNT_LOCATION") + } + if envHasValue("AZURE_BATCH_ACCOUNT_NAME") { + config.AccountName = os.Getenv("AZURE_BATCH_ACCOUNT_NAME") + } + + if config.JobID == "" { + hostname, err := os.Hostname() + if err != nil { + log.Panic(err) + } + config.JobID = hostname + } + + if config.ClientID == "" || + config.ClientSecret == "" || + config.ResourceGroup == "" || + config.SubscriptionID == "" || + config.PoolID == "" || + config.TenantID == "" { + return &ConfigError{CurrentConfig: config, ErrorDetails: "Missing configuration"} + } + + return nil +} + +func getTaskIDForPod(namespace, name string) string { + ID := []byte(fmt.Sprintf("%s-%s", namespace, name)) + return string(fmt.Sprintf("%x", md5.Sum(ID))) +} + +type jsonLog struct { + Log string `json:"log"` +} + +func formatLogJSON(readCloser batch.ReadCloser) (string, error) { + //Read line by line as file isn't valid json. Each line is a single valid json object. + scanner := bufio.NewScanner(*readCloser.Value) + var b strings.Builder + for scanner.Scan() { + result := jsonLog{} + err := json.Unmarshal(scanner.Bytes(), &result) + if err != nil { + return "", err + } + mustWriteString(&b, result.Log) + } + + return b.String(), nil +} + +// ConfigError - Error when reading configuration values. +type ConfigError struct { + CurrentConfig *Config + ErrorDetails string +} + +func (e *ConfigError) Error() string { + configJSON, err := json.Marshal(e.CurrentConfig) + if err != nil { + return e.ErrorDetails + } + + return e.ErrorDetails + ": " + string(configJSON) +} diff --git a/providers/azurebatch/batch_status.go b/providers/azurebatch/batch_status.go new file mode 100644 index 000000000..2d14284ad --- /dev/null +++ b/providers/azurebatch/batch_status.go @@ -0,0 +1,141 @@ +package azurebatch + +import ( + "encoding/json" + "fmt" + + "github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch" + apiv1 "k8s.io/api/core/v1" + metav1 "k8s.io/apimachinery/pkg/apis/meta/v1" +) + +func getPodFromTask(task *batch.CloudTask) (pod *apiv1.Pod, err error) { + if task == nil || task.EnvironmentSettings == nil { + return nil, fmt.Errorf("invalid input task: %v", task) + } + + ok := false + jsonData := "" + settings := *task.EnvironmentSettings + for _, s := range settings { + if *s.Name == podJSONKey && s.Value != nil { + ok = true + jsonData = *s.Value + } + } + if !ok { + return nil, fmt.Errorf("task doesn't have pod json stored in it: %v", task.EnvironmentSettings) + } + + pod = &apiv1.Pod{} + err = json.Unmarshal([]byte(jsonData), pod) + if err != nil { + return nil, err + } + return +} + +func convertTaskToPodStatus(task *batch.CloudTask) (status *apiv1.PodStatus, err error) { + + pod, err := getPodFromTask(task) + if err != nil { + return + } + + // Todo: Review indivudal container status response + status = &apiv1.PodStatus{ + Phase: convertTaskStatusToPodPhase(task), + Conditions: []apiv1.PodCondition{}, + Message: "", + Reason: "", + HostIP: "", + PodIP: "127.0.0.1", + StartTime: &pod.CreationTimestamp, + } + + for _, container := range pod.Spec.Containers { + containerStatus := apiv1.ContainerStatus{ + Name: container.Name, + State: convertTaskStatusToContainerState(task), + Ready: true, + RestartCount: 0, + Image: container.Image, + ImageID: "", + ContainerID: "", + } + status.ContainerStatuses = append(status.ContainerStatuses, containerStatus) + } + + return +} + +func convertTaskStatusToPodPhase(t *batch.CloudTask) (podPhase apiv1.PodPhase) { + switch t.State { + case batch.TaskStatePreparing: + podPhase = apiv1.PodPending + case batch.TaskStateActive: + podPhase = apiv1.PodPending + case batch.TaskStateRunning: + podPhase = apiv1.PodRunning + case batch.TaskStateCompleted: + podPhase = apiv1.PodFailed + + if t.ExecutionInfo != nil && t.ExecutionInfo.ExitCode != nil && *t.ExecutionInfo.ExitCode == 0 { + podPhase = apiv1.PodSucceeded + } + } + + return +} + +func convertTaskStatusToContainerState(t *batch.CloudTask) (containerState apiv1.ContainerState) { + + startTime := metav1.Time{} + if t.ExecutionInfo != nil { + if t.ExecutionInfo.StartTime != nil { + startTime.Time = t.ExecutionInfo.StartTime.Time + } + } + + switch t.State { + case batch.TaskStatePreparing: + containerState = apiv1.ContainerState{ + Waiting: &apiv1.ContainerStateWaiting{ + Message: "Waiting for machine in AzureBatch", + Reason: "Preparing", + }, + } + case batch.TaskStateActive: + containerState = apiv1.ContainerState{ + Waiting: &apiv1.ContainerStateWaiting{ + Message: "Waiting for machine in AzureBatch", + Reason: "Queued", + }, + } + case batch.TaskStateRunning: + containerState = apiv1.ContainerState{ + Running: &apiv1.ContainerStateRunning{ + StartedAt: startTime, + }, + } + case batch.TaskStateCompleted: + termStatus := apiv1.ContainerState{ + Terminated: &apiv1.ContainerStateTerminated{ + FinishedAt: metav1.Time{ + Time: t.StateTransitionTime.Time, + }, + StartedAt: startTime, + }, + } + + if t.ExecutionInfo != nil && t.ExecutionInfo.ExitCode != nil { + exitCode := *t.ExecutionInfo.ExitCode + termStatus.Terminated.ExitCode = exitCode + if exitCode != 0 { + termStatus.Terminated.Message = *t.ExecutionInfo.FailureInfo.Message + } + } + } + + return +} diff --git a/providers/azurebatch/batch_status_test.go b/providers/azurebatch/batch_status_test.go new file mode 100644 index 000000000..205c81cac --- /dev/null +++ b/providers/azurebatch/batch_status_test.go @@ -0,0 +1,166 @@ +package azurebatch + +import ( + "github.com/Azure/go-autorest/autorest/to" + "reflect" + "testing" + + "github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch" + apiv1 "k8s.io/api/core/v1" +) + +func Test_getPodFromTask(t *testing.T) { + type args struct { + task *batch.CloudTask + } + tests := []struct { + name string + task batch.CloudTask + wantPod *apiv1.Pod + wantErr bool + }{ + { + name: "SimplePod", + task: batch.CloudTask{ + EnvironmentSettings: &[]batch.EnvironmentSetting{ + { + Name: to.StringPtr(podJSONKey), + Value: to.StringPtr(`{"metadata":{"creationTimestamp":null},"spec":{"containers":[{"name":"web","image":"nginx:1.12","ports":[{"name":"http","containerPort":80,"protocol":"TCP"}],"resources":{}}]},"status":{}}`), + }, + }, + }, + wantPod: &apiv1.Pod{ + Spec: apiv1.PodSpec{ + Containers: []apiv1.Container{ + { + Name: "web", + Image: "nginx:1.12", + Ports: []apiv1.ContainerPort{ + { + Name: "http", + Protocol: apiv1.ProtocolTCP, + ContainerPort: 80, + }, + }, + }, + }, + }, + }, + }, + { + name: "InvalidJson", + task: batch.CloudTask{ + EnvironmentSettings: &[]batch.EnvironmentSetting{ + { + Name: to.StringPtr(podJSONKey), + Value: to.StringPtr("---notjson--"), + }, + }, + }, + wantErr: true, + }, + { + name: "NilEnvironment", + task: batch.CloudTask{ + EnvironmentSettings: nil, + }, + wantErr: true, + }, + { + name: "NilString", + task: batch.CloudTask{ + EnvironmentSettings: &[]batch.EnvironmentSetting{ + { + Name: to.StringPtr(podJSONKey), + Value: nil, + }, + }, + }, + wantErr: true, + }, + } + for _, tt := range tests { + t.Run(tt.name, func(t *testing.T) { + gotPod, err := getPodFromTask(&tt.task) + if (err != nil) != tt.wantErr { + t.Errorf("getPodFromTask() error = %v, wantErr %v", err, tt.wantErr) + return + } + + if !reflect.DeepEqual(gotPod, tt.wantPod) { + + t.Errorf("getPodFromTask() = %v, want %v", gotPod, tt.wantPod) + } + }) + } +} + +func Test_convertTaskStatusToPodPhase(t *testing.T) { + type args struct { + t *batch.CloudTask + } + tests := []struct { + name string + task batch.CloudTask + wantPodPhase apiv1.PodPhase + }{ + { + name: "PreparingTask", + task: batch.CloudTask{ + State: batch.TaskStatePreparing, + }, + wantPodPhase: apiv1.PodPending, + }, + { + //Active tasks are sitting in a queue waiting for a node + // so maps best to pending state + name: "ActiveTask", + task: batch.CloudTask{ + State: batch.TaskStateActive, + }, + wantPodPhase: apiv1.PodPending, + }, + { + name: "RunningTask", + task: batch.CloudTask{ + State: batch.TaskStateRunning, + }, + wantPodPhase: apiv1.PodRunning, + }, + { + name: "CompletedTask_ExitCode0", + task: batch.CloudTask{ + State: batch.TaskStateCompleted, + ExecutionInfo: &batch.TaskExecutionInformation{ + ExitCode: to.Int32Ptr(0), + }, + }, + wantPodPhase: apiv1.PodSucceeded, + }, + { + name: "CompletedTask_ExitCode127", + task: batch.CloudTask{ + State: batch.TaskStateCompleted, + ExecutionInfo: &batch.TaskExecutionInformation{ + ExitCode: to.Int32Ptr(127), + }, + }, + wantPodPhase: apiv1.PodFailed, + }, + { + name: "CompletedTask_nilExecInfo", + task: batch.CloudTask{ + State: batch.TaskStateCompleted, + ExecutionInfo: nil, + }, + wantPodPhase: apiv1.PodFailed, + }, + } + for _, tt := range tests { + t.Run(tt.name, func(t *testing.T) { + if gotPodPhase := convertTaskStatusToPodPhase(&tt.task); !reflect.DeepEqual(gotPodPhase, tt.wantPodPhase) { + t.Errorf("convertTaskStatusToPodPhase() = %v, want %v", gotPodPhase, tt.wantPodPhase) + } + }) + } +} diff --git a/providers/azurebatch/batch_test.go b/providers/azurebatch/batch_test.go new file mode 100644 index 000000000..b919b1c7a --- /dev/null +++ b/providers/azurebatch/batch_test.go @@ -0,0 +1,165 @@ +package azurebatch + +import ( + "crypto/md5" + "fmt" + "github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch" + "github.com/Azure/go-autorest/autorest" + "io/ioutil" + "net/http" + "os" + "strings" + "testing" + + apiv1 "k8s.io/api/core/v1" +) + +func Test_deletePod(t *testing.T) { + podNamespace := "bob" + podName := "marley" + concatName := []byte(fmt.Sprintf("%s-%s", podNamespace, podName)) + expectedDeleteTaskID := fmt.Sprintf("%x", md5.Sum(concatName)) + + provider := Provider{} + provider.deleteTask = func(taskID string) (autorest.Response, error) { + if taskID != expectedDeleteTaskID { + t.Errorf("Deleted wrong task! Expected delete: %v Actual: %v", taskID, expectedDeleteTaskID) + } + return autorest.Response{}, nil + } + + pod := &apiv1.Pod{} + pod.Name = podName + pod.Namespace = podNamespace + + err := provider.DeletePod(pod) + if err != nil { + t.Error(err) + } +} + +func Test_deletePod_doesntExist(t *testing.T) { + pod := &apiv1.Pod{} + pod.Namespace = "bob" + pod.Name = "marley" + + provider := Provider{} + provider.deleteTask = func(taskID string) (autorest.Response, error) { + return autorest.Response{}, fmt.Errorf("Task doesn't exist") + } + + err := provider.DeletePod(pod) + if err == nil { + t.Error("Expected error but none seen") + } +} + +func Test_createPod(t *testing.T) { + pod := &apiv1.Pod{} + pod.Namespace = "bob" + pod.Name = "marley" + + provider := Provider{} + provider.addTask = func(task batch.TaskAddParameter) (autorest.Response, error) { + if task.CommandLine == nil || *task.CommandLine == "" { + t.Error("Missing commandline args") + } + + derefVars := *task.EnvironmentSettings + if len(derefVars) != 1 || *derefVars[0].Name != podJSONKey { + t.Error("Missing pod json") + } + return autorest.Response{}, nil + } + + err := provider.CreatePod(pod) + if err != nil { + t.Errorf("Unexpected error creating pod %v", err) + } +} + +func Test_createPod_errorResponse(t *testing.T) { + pod := &apiv1.Pod{} + pod.Namespace = "bob" + pod.Name = "marley" + + provider := Provider{} + provider.addTask = func(task batch.TaskAddParameter) (autorest.Response, error) { + return autorest.Response{}, fmt.Errorf("Failed creating task") + } + + err := provider.CreatePod(pod) + if err == nil { + t.Error("Expected error but none seen") + } +} + +func Test_readLogs_404Response_expectReturnStartupLogs(t *testing.T) { + pod := &apiv1.Pod{} + pod.Namespace = "bob" + pod.Name = "marley" + containerName := "sam" + + provider := Provider{} + provider.getFileFromTask = func(taskID, path string) (batch.ReadCloser, error) { + if path == "wd/sam.log" { + // Autorest - Seriously? Can't find a better way to make a 404 :( + return batch.ReadCloser{Response: autorest.Response{Response: &http.Response{StatusCode: 404}}}, nil + } else if path == "stderr.txt" { + response := ioutil.NopCloser(strings.NewReader("stderrResponse")) + return batch.ReadCloser{Value: &response}, nil + } else if path == "stdout.txt" { + response := ioutil.NopCloser(strings.NewReader("stdoutResponse")) + return batch.ReadCloser{Value: &response}, nil + } else { + t.Errorf("Unexpected Filepath: %v", path) + } + + return batch.ReadCloser{}, fmt.Errorf("Failed in test mock of getFileFromTask") + } + + result, err := provider.GetContainerLogs(pod.Namespace, pod.Name, containerName, 0) + if err != nil { + t.Errorf("GetContainerLogs return error: %v", err) + } + + fmt.Print(result) + + if !strings.Contains(result, "stderrResponse") || !strings.Contains(result, "stdoutResponse") { + t.Errorf("Result didn't contain expected content have: %v", result) + } + +} + +func Test_readLogs_JsonResponse_expectFormattedLogs(t *testing.T) { + pod := &apiv1.Pod{} + pod.Namespace = "bob" + pod.Name = "marley" + containerName := "sam" + + provider := Provider{} + provider.getFileFromTask = func(taskID, path string) (batch.ReadCloser, error) { + if path == "wd/sam.log" { + fileReader, err := os.Open("./testdata/logresponse.json") + if err != nil { + t.Error(err) + } + readCloser := ioutil.NopCloser(fileReader) + return batch.ReadCloser{Value: &readCloser, Response: autorest.Response{Response: &http.Response{StatusCode: 200}}}, nil + } + + t.Errorf("Unexpected Filepath: %v", path) + return batch.ReadCloser{}, fmt.Errorf("Failed in test mock of getFileFromTask") + } + + result, err := provider.GetContainerLogs(pod.Namespace, pod.Name, containerName, 0) + if err != nil { + t.Errorf("GetContainerLogs return error: %v", err) + } + + fmt.Print(result) + if !strings.Contains(result, "Copy output data from the CUDA device to the host memory") || strings.Contains(result, "{") { + t.Errorf("Result didn't contain expected content have or had json: %v", result) + } + +} diff --git a/providers/azurebatch/deployment/aks/main.tf b/providers/azurebatch/deployment/aks/main.tf new file mode 100644 index 000000000..bd2b9004f --- /dev/null +++ b/providers/azurebatch/deployment/aks/main.tf @@ -0,0 +1,59 @@ +resource "random_id" "workspace" { + keepers = { + # Generate a new id each time we switch to a new resource group + group_name = "${var.resource_group_name}" + } + + byte_length = 8 +} + +#an attempt to keep the AKS name (and dns label) somewhat unique +resource "random_integer" "random_int" { + min = 100 + max = 999 +} + +resource "azurerm_kubernetes_cluster" "aks" { + name = "aks-${random_integer.random_int.result}" + location = "${var.resource_group_location}" + dns_prefix = "aks-${random_integer.random_int.result}" + + resource_group_name = "${var.resource_group_name}" + kubernetes_version = "1.9.2" + + linux_profile { + admin_username = "${var.linux_admin_username}" + + ssh_key { + key_data = "${var.linux_admin_ssh_publickey}" + } + } + + agent_pool_profile { + name = "agentpool" + count = "2" + vm_size = "Standard_DS2_v2" + os_type = "Linux" + } + + service_principal { + client_id = "${var.client_id}" + client_secret = "${var.client_secret}" + } +} + +output "cluster_client_certificate" { + value = "${base64decode(azurerm_kubernetes_cluster.aks.kube_config.0.client_certificate)}" +} + +output "cluster_client_key" { + value = "${base64decode(azurerm_kubernetes_cluster.aks.kube_config.0.client_key)}" +} + +output "cluster_ca" { + value = "${base64decode(azurerm_kubernetes_cluster.aks.kube_config.0.cluster_ca_certificate)}" +} + +output "host" { + value = "${azurerm_kubernetes_cluster.aks.kube_config.0.host}" +} diff --git a/providers/azurebatch/deployment/aks/variables.tf b/providers/azurebatch/deployment/aks/variables.tf new file mode 100644 index 000000000..17f2974a4 --- /dev/null +++ b/providers/azurebatch/deployment/aks/variables.tf @@ -0,0 +1,31 @@ +variable "client_id" { + type = "string" + description = "Client ID" +} + +variable "client_secret" { + type = "string" + description = "Client secret." +} + +variable "resource_group_name" { + type = "string" + description = "Name of the azure resource group." + default = "akc-rg" +} + +variable "resource_group_location" { + type = "string" + description = "Location of the azure resource group." + default = "eastus" +} + +variable "linux_admin_username" { + type = "string" + description = "User name for authentication to the Kubernetes linux agent virtual machines in the cluster." +} + +variable "linux_admin_ssh_publickey" { + type = "string" + description = "Configure all the linux virtual machines in the cluster with the SSH RSA public key string. The key should include three parts, for example 'ssh-rsa AAAAB...snip...UcyupgH azureuser@linuxvm'" +} diff --git a/providers/azurebatch/deployment/azurebatch/main.tf b/providers/azurebatch/deployment/azurebatch/main.tf new file mode 100644 index 000000000..d4060d3fc --- /dev/null +++ b/providers/azurebatch/deployment/azurebatch/main.tf @@ -0,0 +1,146 @@ +resource "random_string" "batchname" { + keepers = { + # Generate a new id each time we switch to a new resource group + group_name = "${var.resource_group_name}" + } + + length = 8 + upper = false + special = false + number = false +} + +resource "azurerm_template_deployment" "test" { + name = "tfdeployment" + resource_group_name = "${var.resource_group_name}" + + # these key-value pairs are passed into the ARM Template's `parameters` block + parameters { + "batchAccountName" = "${random_string.batchname.result}" + "storageAccountID" = "${var.storage_account_id}" + "poolBoostrapScriptUrl" = "${var.pool_bootstrap_script_url}" + "location" = "${var.resource_group_location}" + "poolID" = "${var.pool_id}" + "vmSku" = "${var.vm_sku}" + "lowPriorityNodeCount" = "${var.low_priority_node_count}" + "dedicatedNodeCount" = "${var.dedicated_node_count}" + } + + deployment_mode = "Incremental" + + template_body = < /etc/docker/daemon.json.merge +python -c "import json;a=json.load(open('/etc/docker/daemon.json.merge'));b=json.load(open('/etc/docker/daemon.json'));a.update(b);f=open('/etc/docker/daemon.json','w');json.dump(a,f);f.close();" +rm -f /etc/docker/daemon.json.merge +sed -i 's|^ExecStart=/usr/bin/dockerd.*|ExecStart=/usr/bin/dockerd|' /lib/systemd/system/docker.service +systemctl daemon-reload +systemctl start docker.service + + + diff --git a/providers/azurebatch/deployment/storage/main.tf b/providers/azurebatch/deployment/storage/main.tf new file mode 100644 index 000000000..69f6a1d6f --- /dev/null +++ b/providers/azurebatch/deployment/storage/main.tf @@ -0,0 +1,77 @@ +resource "random_string" "storage" { + keepers = { + # Generate a new id each time we switch to a new resource group + group_name = "${var.resource_group_name}" + } + + length = 8 + upper = false + special = false + number = false +} + +resource "azurerm_storage_account" "batchstorage" { + name = "${lower(random_string.storage.result)}" + resource_group_name = "${var.resource_group_name}" + location = "${var.resource_group_location}" + account_tier = "Standard" + account_replication_type = "LRS" +} + +resource "azurerm_storage_container" "boostrapscript" { + name = "scripts" + resource_group_name = "${var.resource_group_name}" + storage_account_name = "${azurerm_storage_account.batchstorage.name}" + container_access_type = "private" +} + +resource "azurerm_storage_blob" "initscript" { + name = "init.sh" + + resource_group_name = "${var.resource_group_name}" + storage_account_name = "${azurerm_storage_account.batchstorage.name}" + storage_container_name = "${azurerm_storage_container.boostrapscript.name}" + + type = "block" + source = "${var.pool_bootstrap_script_path}" +} + +data "azurerm_storage_account_sas" "scriptaccess" { + connection_string = "${azurerm_storage_account.batchstorage.primary_connection_string}" + https_only = true + + resource_types { + service = false + container = false + object = true + } + + services { + blob = true + queue = false + table = false + file = false + } + + start = "${timestamp()}" + expiry = "${timeadd(timestamp(), "8776h")}" + + permissions { + read = true + write = false + delete = false + list = false + add = false + create = false + update = false + process = false + } +} + +output "pool_boostrap_script_url" { + value = "${azurerm_storage_blob.initscript.url}${data.azurerm_storage_account_sas.scriptaccess.sas}" +} + +output "id" { + value = "${azurerm_storage_account.batchstorage.id}" +} diff --git a/providers/azurebatch/deployment/storage/variables.tf b/providers/azurebatch/deployment/storage/variables.tf new file mode 100644 index 000000000..a15560ea6 --- /dev/null +++ b/providers/azurebatch/deployment/storage/variables.tf @@ -0,0 +1,14 @@ +variable "resource_group_name" { + description = "Resource group name" + type = "string" +} + +variable "resource_group_location" { + description = "Resource group location" + type = "string" +} + +variable "pool_bootstrap_script_path" { + description = "The filepath of the pool boostrapping script" + type = "string" +} diff --git a/providers/azurebatch/deployment/variables.tf b/providers/azurebatch/deployment/variables.tf new file mode 100644 index 000000000..193cfd973 --- /dev/null +++ b/providers/azurebatch/deployment/variables.tf @@ -0,0 +1,23 @@ +variable "client_id" { + type = "string" + description = "Client ID" +} + +variable "client_secret" { + type = "string" + description = "Client secret." +} + +variable "resource_group_name" { + description = "Resource group name" + type = "string" +} + +variable "resource_group_location" { + description = "Resource group location" + type = "string" +} + +variable "virtualkubelet_docker_image" { + type = "string" +} diff --git a/providers/azurebatch/deployment/vars.example.tfvars b/providers/azurebatch/deployment/vars.example.tfvars new file mode 100644 index 000000000..daf33d231 --- /dev/null +++ b/providers/azurebatch/deployment/vars.example.tfvars @@ -0,0 +1,14 @@ +// Provide the Client ID of a service principal for use by AKS +client_id = "00000000-0000-0000-0000-000000000000" + +// Provide the Client Secret of a service principal for use by AKS +client_secret = "00000000-0000-0000-0000-000000000000" + +// The resource group you would like to deploy too +resource_group_name = "vkgpu" + +// The location of all resources +resource_group_location = "westeurope" + +// Virtual Kubelet docker image +virtualkubelet_docker_image = "microsoft/virtual-kubelet" diff --git a/providers/azurebatch/deployment/virtualkubelet/main.tf b/providers/azurebatch/deployment/virtualkubelet/main.tf new file mode 100644 index 000000000..096527fee --- /dev/null +++ b/providers/azurebatch/deployment/virtualkubelet/main.tf @@ -0,0 +1,126 @@ +provider "kubernetes" { + host = "${var.cluster_host}" + + client_certificate = "${var.cluster_client_certificate}" + client_key = "${var.cluster_client_key}" + cluster_ca_certificate = "${var.cluster_ca}" +} + +resource "kubernetes_secret" "vkcredentials" { + metadata { + name = "vkcredentials" + } + + data { + cert.pem = "${var.cluster_client_certificate}" + key.pem = "${var.cluster_client_key}" + } +} + +resource "kubernetes_deployment" "vkdeployment" { + metadata { + name = "vkdeployment" + } + + spec { + selector { + app = "virtualkubelet" + } + + template { + metadata { + labels { + app = "virtualkubelet" + } + } + + spec { + container { + name = "vk" + image = "${var.virtualkubelet_docker_image}" + + args = [ + "--provider", + "azurebatch", + "--taint", + "azure.com/batch", + "--namespace", + "default", + ] + + port { + container_port = 10250 + protocol = "TCP" + name = "kubeletport" + } + + volume_mount { + name = "azure-credentials" + mount_path = "/etc/aks/azure.json" + } + + volume_mount { + name = "credentials" + mount_path = "/etc/virtual-kubelet" + } + + env = [ + { + name = "AZURE_BATCH_ACCOUNT_LOCATION" + value = "${var.resource_group_location}" + }, + { + name = "AZURE_BATCH_ACCOUNT_NAME" + value = "${var.azure_batch_account_name}" + }, + { + name = "AZURE_BATCH_POOLID" + value = "${var.azure_batch_pool_id}" + }, + { + name = "KUBELET_PORT" + value = "10250" + }, + { + name = "AZURE_CREDENTIALS_LOCATION" + value = "/etc/aks/azure.json" + }, + { + name = "APISERVER_CERT_LOCATION" + value = "/etc/virtual-kubelet/cert.pem" + }, + { + name = "APISERVER_KEY_LOCATION" + value = "/etc/virtual-kubelet/key.pem" + }, + { + name = "VKUBELET_POD_IP" + + value_from { + field_ref { + field_path = "status.podIP" + } + } + }, + ] + } + + volume { + name = "azure-credentials" + + host_path { + path = "/etc/kubernetes/azure.json" + } + } + + volume { + name = "credentials" + + secret { + secret_name = "vkcredentials" + } + } + } + } + } +} diff --git a/providers/azurebatch/deployment/virtualkubelet/variables.tf b/providers/azurebatch/deployment/virtualkubelet/variables.tf new file mode 100644 index 000000000..e8d73875e --- /dev/null +++ b/providers/azurebatch/deployment/virtualkubelet/variables.tf @@ -0,0 +1,41 @@ +variable "cluster_client_certificate" { + type = "string" + description = "Cluster client Certificate" + default = "eastus" +} + +variable "cluster_client_key" { + type = "string" + description = "Cluster client Certificate Key" +} + +variable "cluster_ca" { + type = "string" + description = "Cluster Certificate Authority" +} + +variable "cluster_host" { + type = "string" + description = "Cluster Admin API host" +} + +variable "virtualkubelet_docker_image" { + type = "string" + description = "The docker image to use for deploying the virtual kubelet" +} + +variable "azure_batch_account_name" { + type = "string" + description = "The name of the Azure Batch account to use" +} + +variable "azure_batch_pool_id" { + type = "string" + description = "The PoolID to use in Azure batch" + default = "pool1" +} + +variable "resource_group_location" { + description = "Resource group location" + type = "string" +} diff --git a/providers/azurebatch/testdata/logresponse.json b/providers/azurebatch/testdata/logresponse.json new file mode 100644 index 000000000..05aaef27c --- /dev/null +++ b/providers/azurebatch/testdata/logresponse.json @@ -0,0 +1,6 @@ +{"log":"[Vector addition of 50000 elements]\n","stream":"stdout","time":"2018-05-30T17:02:49.967357287Z"} +{"log":"Copy input data from the host memory to the CUDA device\n","stream":"stdout","time":"2018-05-30T17:02:49.967417086Z"} +{"log":"CUDA kernel launch with 196 blocks of 256 threads\n","stream":"stdout","time":"2018-05-30T17:02:49.967423286Z"} +{"log":"Copy output data from the CUDA device to the host memory\n","stream":"stdout","time":"2018-05-30T17:02:49.967427386Z"} +{"log":"Test PASSED\n","stream":"stdout","time":"2018-05-30T17:02:49.967431286Z"} +{"log":"Done\n","stream":"stdout","time":"2018-05-30T17:02:49.967435286Z"} \ No newline at end of file diff --git a/providers/cri/cri.go b/providers/cri/cri.go index 7151d790e..a4d1d1afb 100644 --- a/providers/cri/cri.go +++ b/providers/cri/cri.go @@ -616,7 +616,7 @@ func readLogFile(filename string, tail int) (string, error) { lines = append(lines, scanner.Text()) } if tail > 0 && tail < len(lines) { - lines = lines[len(lines)-tail : len(lines)] + lines = lines[len(lines)-tail:] } return strings.Join(lines, ""), nil } @@ -636,7 +636,7 @@ func (p *CRIProvider) GetContainerLogs(namespace, podName, containerName string, } container := pod.containers[containerName] if container == nil { - return "", fmt.Errorf("Cannot find container %s in pod %s namespace %s", containerName, pod, namespace) + return "", fmt.Errorf("Cannot find container %s in pod %s namespace %s", containerName, podName, namespace) } return readLogFile(container.LogPath, tail) diff --git a/scripts/createCertAndKey.sh b/scripts/createCertAndKey.sh old mode 100644 new mode 100755 index 4c185161b..ef992c6c4 --- a/scripts/createCertAndKey.sh +++ b/scripts/createCertAndKey.sh @@ -1,5 +1,10 @@ #!/bin/bash # Generate cert and key for chart openssl req -newkey rsa:4096 -new -nodes -x509 -days 3650 -keyout key.pem -out cert.pem -subj "/C=US/ST=CA/L=virtualkubelet/O=virtualkubelet/OU=virtualkubelet/CN=virtualkubelet" -cert=$(base64 cert.pem -w0) -key=$(base64 key.pem -w0) +if [[ "$OSTYPE" == "darwin"* ]]; then + cert=$(base64 cert.pem) + key=$(base64 key.pem) +else + cert=$(base64 cert.pem -w0) + key=$(base64 key.pem -w0) +fi diff --git a/vendor/github.com/Azure/azure-sdk-for-go/LICENSE b/vendor/github.com/Azure/azure-sdk-for-go/LICENSE new file mode 100644 index 000000000..af39a91e7 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/LICENSE @@ -0,0 +1,202 @@ + + Apache License + Version 2.0, January 2004 + http://www.apache.org/licenses/ + + TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + + 1. Definitions. + + "License" shall mean the terms and conditions for use, reproduction, + and distribution as defined by Sections 1 through 9 of this document. + + "Licensor" shall mean the copyright owner or entity authorized by + the copyright owner that is granting the License. + + "Legal Entity" shall mean the union of the acting entity and all + other entities that control, are controlled by, or are under common + control with that entity. For the purposes of this definition, + "control" means (i) the power, direct or indirect, to cause the + direction or management of such entity, whether by contract or + otherwise, or (ii) ownership of fifty percent (50%) or more of the + outstanding shares, or (iii) beneficial ownership of such entity. + + "You" (or "Your") shall mean an individual or Legal Entity + exercising permissions granted by this License. + + "Source" form shall mean the preferred form for making modifications, + including but not limited to software source code, documentation + source, and configuration files. + + "Object" form shall mean any form resulting from mechanical + transformation or translation of a Source form, including but + not limited to compiled object code, generated documentation, + and conversions to other media types. + + "Work" shall mean the work of authorship, whether in Source or + Object form, made available under the License, as indicated by a + copyright notice that is included in or attached to the work + (an example is provided in the Appendix below). + + "Derivative Works" shall mean any work, whether in Source or Object + form, that is based on (or derived from) the Work and for which the + editorial revisions, annotations, elaborations, or other modifications + represent, as a whole, an original work of authorship. For the purposes + of this License, Derivative Works shall not include works that remain + separable from, or merely link (or bind by name) to the interfaces of, + the Work and Derivative Works thereof. + + "Contribution" shall mean any work of authorship, including + the original version of the Work and any modifications or additions + to that Work or Derivative Works thereof, that is intentionally + submitted to Licensor for inclusion in the Work by the copyright owner + or by an individual or Legal Entity authorized to submit on behalf of + the copyright owner. For the purposes of this definition, "submitted" + means any form of electronic, verbal, or written communication sent + to the Licensor or its representatives, including but not limited to + communication on electronic mailing lists, source code control systems, + and issue tracking systems that are managed by, or on behalf of, the + Licensor for the purpose of discussing and improving the Work, but + excluding communication that is conspicuously marked or otherwise + designated in writing by the copyright owner as "Not a Contribution." + + "Contributor" shall mean Licensor and any individual or Legal Entity + on behalf of whom a Contribution has been received by Licensor and + subsequently incorporated within the Work. + + 2. Grant of Copyright License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + copyright license to reproduce, prepare Derivative Works of, + publicly display, publicly perform, sublicense, and distribute the + Work and such Derivative Works in Source or Object form. + + 3. Grant of Patent License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + (except as stated in this section) patent license to make, have made, + use, offer to sell, sell, import, and otherwise transfer the Work, + where such license applies only to those patent claims licensable + by such Contributor that are necessarily infringed by their + Contribution(s) alone or by combination of their Contribution(s) + with the Work to which such Contribution(s) was submitted. If You + institute patent litigation against any entity (including a + cross-claim or counterclaim in a lawsuit) alleging that the Work + or a Contribution incorporated within the Work constitutes direct + or contributory patent infringement, then any patent licenses + granted to You under this License for that Work shall terminate + as of the date such litigation is filed. + + 4. Redistribution. You may reproduce and distribute copies of the + Work or Derivative Works thereof in any medium, with or without + modifications, and in Source or Object form, provided that You + meet the following conditions: + + (a) You must give any other recipients of the Work or + Derivative Works a copy of this License; and + + (b) You must cause any modified files to carry prominent notices + stating that You changed the files; and + + (c) You must retain, in the Source form of any Derivative Works + that You distribute, all copyright, patent, trademark, and + attribution notices from the Source form of the Work, + excluding those notices that do not pertain to any part of + the Derivative Works; and + + (d) If the Work includes a "NOTICE" text file as part of its + distribution, then any Derivative Works that You distribute must + include a readable copy of the attribution notices contained + within such NOTICE file, excluding those notices that do not + pertain to any part of the Derivative Works, in at least one + of the following places: within a NOTICE text file distributed + as part of the Derivative Works; within the Source form or + documentation, if provided along with the Derivative Works; or, + within a display generated by the Derivative Works, if and + wherever such third-party notices normally appear. The contents + of the NOTICE file are for informational purposes only and + do not modify the License. You may add Your own attribution + notices within Derivative Works that You distribute, alongside + or as an addendum to the NOTICE text from the Work, provided + that such additional attribution notices cannot be construed + as modifying the License. + + You may add Your own copyright statement to Your modifications and + may provide additional or different license terms and conditions + for use, reproduction, or distribution of Your modifications, or + for any such Derivative Works as a whole, provided Your use, + reproduction, and distribution of the Work otherwise complies with + the conditions stated in this License. + + 5. Submission of Contributions. Unless You explicitly state otherwise, + any Contribution intentionally submitted for inclusion in the Work + by You to the Licensor shall be under the terms and conditions of + this License, without any additional terms or conditions. + Notwithstanding the above, nothing herein shall supersede or modify + the terms of any separate license agreement you may have executed + with Licensor regarding such Contributions. + + 6. Trademarks. This License does not grant permission to use the trade + names, trademarks, service marks, or product names of the Licensor, + except as required for reasonable and customary use in describing the + origin of the Work and reproducing the content of the NOTICE file. + + 7. Disclaimer of Warranty. Unless required by applicable law or + agreed to in writing, Licensor provides the Work (and each + Contributor provides its Contributions) on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or + implied, including, without limitation, any warranties or conditions + of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A + PARTICULAR PURPOSE. You are solely responsible for determining the + appropriateness of using or redistributing the Work and assume any + risks associated with Your exercise of permissions under this License. + + 8. Limitation of Liability. In no event and under no legal theory, + whether in tort (including negligence), contract, or otherwise, + unless required by applicable law (such as deliberate and grossly + negligent acts) or agreed to in writing, shall any Contributor be + liable to You for damages, including any direct, indirect, special, + incidental, or consequential damages of any character arising as a + result of this License or out of the use or inability to use the + Work (including but not limited to damages for loss of goodwill, + work stoppage, computer failure or malfunction, or any and all + other commercial damages or losses), even if such Contributor + has been advised of the possibility of such damages. + + 9. Accepting Warranty or Additional Liability. While redistributing + the Work or Derivative Works thereof, You may choose to offer, + and charge a fee for, acceptance of support, warranty, indemnity, + or other liability obligations and/or rights consistent with this + License. However, in accepting such obligations, You may act only + on Your own behalf and on Your sole responsibility, not on behalf + of any other Contributor, and only if You agree to indemnify, + defend, and hold each Contributor harmless for any liability + incurred by, or claims asserted against, such Contributor by reason + of your accepting any such warranty or additional liability. + + END OF TERMS AND CONDITIONS + + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "[]" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright 2016 Microsoft Corporation + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. diff --git a/vendor/github.com/Azure/azure-sdk-for-go/NOTICE b/vendor/github.com/Azure/azure-sdk-for-go/NOTICE new file mode 100644 index 000000000..2d1d72608 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/NOTICE @@ -0,0 +1,5 @@ +Microsoft Azure-SDK-for-Go +Copyright 2014-2017 Microsoft + +This product includes software developed at +the Microsoft Corporation (https://www.microsoft.com). diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/account.go b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/account.go new file mode 100644 index 000000000..1e4fba835 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/account.go @@ -0,0 +1,175 @@ +package batch + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "github.com/Azure/go-autorest/autorest/date" + "github.com/Azure/go-autorest/autorest/validation" + "github.com/satori/go.uuid" + "net/http" +) + +// AccountClient is the a client for issuing REST requests to the Azure Batch service. +type AccountClient struct { + BaseClient +} + +// NewAccountClient creates an instance of the AccountClient client. +func NewAccountClient() AccountClient { + return NewAccountClientWithBaseURI(DefaultBaseURI) +} + +// NewAccountClientWithBaseURI creates an instance of the AccountClient client. +func NewAccountClientWithBaseURI(baseURI string) AccountClient { + return AccountClient{NewWithBaseURI(baseURI)} +} + +// ListNodeAgentSkus sends the list node agent skus request. +// +// filter is an OData $filter clause. For more information on constructing this filter, see +// https://docs.microsoft.com/en-us/rest/api/batchservice/odata-filters-in-batch#list-node-agent-skus. maxResults +// is the maximum number of items to return in the response. A maximum of 1000 results will be returned. timeout is +// the maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client AccountClient) ListNodeAgentSkus(ctx context.Context, filter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result AccountListNodeAgentSkusResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: maxResults, + Constraints: []validation.Constraint{{Target: "maxResults", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "maxResults", Name: validation.InclusiveMaximum, Rule: 1000, Chain: nil}, + {Target: "maxResults", Name: validation.InclusiveMinimum, Rule: 1, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("batch.AccountClient", "ListNodeAgentSkus", err.Error()) + } + + result.fn = client.listNodeAgentSkusNextResults + req, err := client.ListNodeAgentSkusPreparer(ctx, filter, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.AccountClient", "ListNodeAgentSkus", nil, "Failure preparing request") + return + } + + resp, err := client.ListNodeAgentSkusSender(req) + if err != nil { + result.alnasr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.AccountClient", "ListNodeAgentSkus", resp, "Failure sending request") + return + } + + result.alnasr, err = client.ListNodeAgentSkusResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.AccountClient", "ListNodeAgentSkus", resp, "Failure responding to request") + } + + return +} + +// ListNodeAgentSkusPreparer prepares the ListNodeAgentSkus request. +func (client AccountClient) ListNodeAgentSkusPreparer(ctx context.Context, filter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(filter) > 0 { + queryParameters["$filter"] = autorest.Encode("query", filter) + } + if maxResults != nil { + queryParameters["maxresults"] = autorest.Encode("query", *maxResults) + } else { + queryParameters["maxresults"] = autorest.Encode("query", 1000) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPath("/nodeagentskus"), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListNodeAgentSkusSender sends the ListNodeAgentSkus request. The method will close the +// http.Response Body if it receives an error. +func (client AccountClient) ListNodeAgentSkusSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListNodeAgentSkusResponder handles the response to the ListNodeAgentSkus request. The method always +// closes the http.Response Body. +func (client AccountClient) ListNodeAgentSkusResponder(resp *http.Response) (result AccountListNodeAgentSkusResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNodeAgentSkusNextResults retrieves the next set of results, if any. +func (client AccountClient) listNodeAgentSkusNextResults(lastResults AccountListNodeAgentSkusResult) (result AccountListNodeAgentSkusResult, err error) { + req, err := lastResults.accountListNodeAgentSkusResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "batch.AccountClient", "listNodeAgentSkusNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListNodeAgentSkusSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "batch.AccountClient", "listNodeAgentSkusNextResults", resp, "Failure sending next results request") + } + result, err = client.ListNodeAgentSkusResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.AccountClient", "listNodeAgentSkusNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListNodeAgentSkusComplete enumerates all values, automatically crossing page boundaries as required. +func (client AccountClient) ListNodeAgentSkusComplete(ctx context.Context, filter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result AccountListNodeAgentSkusResultIterator, err error) { + result.page, err = client.ListNodeAgentSkus(ctx, filter, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/application.go b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/application.go new file mode 100644 index 000000000..de66c6d66 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/application.go @@ -0,0 +1,264 @@ +package batch + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "github.com/Azure/go-autorest/autorest/date" + "github.com/Azure/go-autorest/autorest/validation" + "github.com/satori/go.uuid" + "net/http" +) + +// ApplicationClient is the a client for issuing REST requests to the Azure Batch service. +type ApplicationClient struct { + BaseClient +} + +// NewApplicationClient creates an instance of the ApplicationClient client. +func NewApplicationClient() ApplicationClient { + return NewApplicationClientWithBaseURI(DefaultBaseURI) +} + +// NewApplicationClientWithBaseURI creates an instance of the ApplicationClient client. +func NewApplicationClientWithBaseURI(baseURI string) ApplicationClient { + return ApplicationClient{NewWithBaseURI(baseURI)} +} + +// Get this operation returns only applications and versions that are available for use on compute nodes; that is, that +// can be used in an application package reference. For administrator information about applications and versions that +// are not yet available to compute nodes, use the Azure portal or the Azure Resource Manager API. +// +// applicationID is the ID of the application. timeout is the maximum time that the server can spend processing the +// request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the +// form of a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. +// returnClientRequestID is whether the server should return the client-request-id in the response. ocpDate is the +// time the request was issued. Client libraries typically set this to the current system clock time; set it +// explicitly if you are calling the REST API directly. +func (client ApplicationClient) Get(ctx context.Context, applicationID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result ApplicationSummary, err error) { + req, err := client.GetPreparer(ctx, applicationID, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ApplicationClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.ApplicationClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ApplicationClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client ApplicationClient) GetPreparer(ctx context.Context, applicationID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "applicationId": autorest.Encode("path", applicationID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/applications/{applicationId}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client ApplicationClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client ApplicationClient) GetResponder(resp *http.Response) (result ApplicationSummary, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List this operation returns only applications and versions that are available for use on compute nodes; that is, +// that can be used in an application package reference. For administrator information about applications and versions +// that are not yet available to compute nodes, use the Azure portal or the Azure Resource Manager API. +// +// maxResults is the maximum number of items to return in the response. A maximum of 1000 applications can be +// returned. timeout is the maximum time that the server can spend processing the request, in seconds. The default +// is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a GUID with no +// decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the +// server should return the client-request-id in the response. ocpDate is the time the request was issued. Client +// libraries typically set this to the current system clock time; set it explicitly if you are calling the REST API +// directly. +func (client ApplicationClient) List(ctx context.Context, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result ApplicationListResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: maxResults, + Constraints: []validation.Constraint{{Target: "maxResults", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "maxResults", Name: validation.InclusiveMaximum, Rule: 1000, Chain: nil}, + {Target: "maxResults", Name: validation.InclusiveMinimum, Rule: 1, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("batch.ApplicationClient", "List", err.Error()) + } + + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ApplicationClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.alr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.ApplicationClient", "List", resp, "Failure sending request") + return + } + + result.alr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ApplicationClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client ApplicationClient) ListPreparer(ctx context.Context, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if maxResults != nil { + queryParameters["maxresults"] = autorest.Encode("query", *maxResults) + } else { + queryParameters["maxresults"] = autorest.Encode("query", 1000) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPath("/applications"), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client ApplicationClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client ApplicationClient) ListResponder(resp *http.Response) (result ApplicationListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client ApplicationClient) listNextResults(lastResults ApplicationListResult) (result ApplicationListResult, err error) { + req, err := lastResults.applicationListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "batch.ApplicationClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "batch.ApplicationClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ApplicationClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client ApplicationClient) ListComplete(ctx context.Context, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result ApplicationListResultIterator, err error) { + result.page, err = client.List(ctx, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/certificate.go b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/certificate.go new file mode 100644 index 000000000..28e2724ee --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/certificate.go @@ -0,0 +1,557 @@ +package batch + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "github.com/Azure/go-autorest/autorest/date" + "github.com/Azure/go-autorest/autorest/validation" + "github.com/satori/go.uuid" + "net/http" +) + +// CertificateClient is the a client for issuing REST requests to the Azure Batch service. +type CertificateClient struct { + BaseClient +} + +// NewCertificateClient creates an instance of the CertificateClient client. +func NewCertificateClient() CertificateClient { + return NewCertificateClientWithBaseURI(DefaultBaseURI) +} + +// NewCertificateClientWithBaseURI creates an instance of the CertificateClient client. +func NewCertificateClientWithBaseURI(baseURI string) CertificateClient { + return CertificateClient{NewWithBaseURI(baseURI)} +} + +// Add sends the add request. +// +// certificate is the certificate to be added. timeout is the maximum time that the server can spend processing the +// request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the +// form of a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. +// returnClientRequestID is whether the server should return the client-request-id in the response. ocpDate is the +// time the request was issued. Client libraries typically set this to the current system clock time; set it +// explicitly if you are calling the REST API directly. +func (client CertificateClient) Add(ctx context.Context, certificate CertificateAddParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: certificate, + Constraints: []validation.Constraint{{Target: "certificate.Thumbprint", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "certificate.ThumbprintAlgorithm", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "certificate.Data", Name: validation.Null, Rule: true, Chain: nil}}}}); err != nil { + return result, validation.NewError("batch.CertificateClient", "Add", err.Error()) + } + + req, err := client.AddPreparer(ctx, certificate, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "Add", nil, "Failure preparing request") + return + } + + resp, err := client.AddSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "Add", resp, "Failure sending request") + return + } + + result, err = client.AddResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "Add", resp, "Failure responding to request") + } + + return +} + +// AddPreparer prepares the Add request. +func (client CertificateClient) AddPreparer(ctx context.Context, certificate CertificateAddParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPath("/certificates"), + autorest.WithJSON(certificate), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// AddSender sends the Add request. The method will close the +// http.Response Body if it receives an error. +func (client CertificateClient) AddSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// AddResponder handles the response to the Add request. The method always +// closes the http.Response Body. +func (client CertificateClient) AddResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByClosing()) + result.Response = resp + return +} + +// CancelDeletion if you try to delete a certificate that is being used by a pool or compute node, the status of the +// certificate changes to deleteFailed. If you decide that you want to continue using the certificate, you can use this +// operation to set the status of the certificate back to active. If you intend to delete the certificate, you do not +// need to run this operation after the deletion failed. You must make sure that the certificate is not being used by +// any resources, and then you can try again to delete the certificate. +// +// thumbprintAlgorithm is the algorithm used to derive the thumbprint parameter. This must be sha1. thumbprint is +// the thumbprint of the certificate being deleted. timeout is the maximum time that the server can spend +// processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request +// identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client CertificateClient) CancelDeletion(ctx context.Context, thumbprintAlgorithm string, thumbprint string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.CancelDeletionPreparer(ctx, thumbprintAlgorithm, thumbprint, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "CancelDeletion", nil, "Failure preparing request") + return + } + + resp, err := client.CancelDeletionSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "CancelDeletion", resp, "Failure sending request") + return + } + + result, err = client.CancelDeletionResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "CancelDeletion", resp, "Failure responding to request") + } + + return +} + +// CancelDeletionPreparer prepares the CancelDeletion request. +func (client CertificateClient) CancelDeletionPreparer(ctx context.Context, thumbprintAlgorithm string, thumbprint string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "thumbprint": autorest.Encode("path", thumbprint), + "thumbprintAlgorithm": autorest.Encode("path", thumbprintAlgorithm), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/certificates(thumbprintAlgorithm={thumbprintAlgorithm},thumbprint={thumbprint})/canceldelete", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// CancelDeletionSender sends the CancelDeletion request. The method will close the +// http.Response Body if it receives an error. +func (client CertificateClient) CancelDeletionSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// CancelDeletionResponder handles the response to the CancelDeletion request. The method always +// closes the http.Response Body. +func (client CertificateClient) CancelDeletionResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Delete you cannot delete a certificate if a resource (pool or compute node) is using it. Before you can delete a +// certificate, you must therefore make sure that the certificate is not associated with any existing pools, the +// certificate is not installed on any compute nodes (even if you remove a certificate from a pool, it is not removed +// from existing compute nodes in that pool until they restart), and no running tasks depend on the certificate. If you +// try to delete a certificate that is in use, the deletion fails. The certificate status changes to deleteFailed. You +// can use Cancel Delete Certificate to set the status back to active if you decide that you want to continue using the +// certificate. +// +// thumbprintAlgorithm is the algorithm used to derive the thumbprint parameter. This must be sha1. thumbprint is +// the thumbprint of the certificate to be deleted. timeout is the maximum time that the server can spend +// processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request +// identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client CertificateClient) Delete(ctx context.Context, thumbprintAlgorithm string, thumbprint string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.DeletePreparer(ctx, thumbprintAlgorithm, thumbprint, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "Delete", nil, "Failure preparing request") + return + } + + resp, err := client.DeleteSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "Delete", resp, "Failure sending request") + return + } + + result, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "Delete", resp, "Failure responding to request") + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client CertificateClient) DeletePreparer(ctx context.Context, thumbprintAlgorithm string, thumbprint string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "thumbprint": autorest.Encode("path", thumbprint), + "thumbprintAlgorithm": autorest.Encode("path", thumbprintAlgorithm), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/certificates(thumbprintAlgorithm={thumbprintAlgorithm},thumbprint={thumbprint})", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client CertificateClient) DeleteSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client CertificateClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets information about the specified certificate. +// +// thumbprintAlgorithm is the algorithm used to derive the thumbprint parameter. This must be sha1. thumbprint is +// the thumbprint of the certificate to get. selectParameter is an OData $select clause. timeout is the maximum +// time that the server can spend processing the request, in seconds. The default is 30 seconds. clientRequestID is +// the caller-generated request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client CertificateClient) Get(ctx context.Context, thumbprintAlgorithm string, thumbprint string, selectParameter string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result Certificate, err error) { + req, err := client.GetPreparer(ctx, thumbprintAlgorithm, thumbprint, selectParameter, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client CertificateClient) GetPreparer(ctx context.Context, thumbprintAlgorithm string, thumbprint string, selectParameter string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "thumbprint": autorest.Encode("path", thumbprint), + "thumbprintAlgorithm": autorest.Encode("path", thumbprintAlgorithm), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/certificates(thumbprintAlgorithm={thumbprintAlgorithm},thumbprint={thumbprint})", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client CertificateClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client CertificateClient) GetResponder(resp *http.Response) (result Certificate, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List sends the list request. +// +// filter is an OData $filter clause. For more information on constructing this filter, see +// https://docs.microsoft.com/en-us/rest/api/batchservice/odata-filters-in-batch#list-certificates. selectParameter +// is an OData $select clause. maxResults is the maximum number of items to return in the response. A maximum of +// 1000 certificates can be returned. timeout is the maximum time that the server can spend processing the request, +// in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the form of +// a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID +// is whether the server should return the client-request-id in the response. ocpDate is the time the request was +// issued. Client libraries typically set this to the current system clock time; set it explicitly if you are +// calling the REST API directly. +func (client CertificateClient) List(ctx context.Context, filter string, selectParameter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CertificateListResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: maxResults, + Constraints: []validation.Constraint{{Target: "maxResults", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "maxResults", Name: validation.InclusiveMaximum, Rule: 1000, Chain: nil}, + {Target: "maxResults", Name: validation.InclusiveMinimum, Rule: 1, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("batch.CertificateClient", "List", err.Error()) + } + + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, filter, selectParameter, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.clr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "List", resp, "Failure sending request") + return + } + + result.clr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client CertificateClient) ListPreparer(ctx context.Context, filter string, selectParameter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(filter) > 0 { + queryParameters["$filter"] = autorest.Encode("query", filter) + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if maxResults != nil { + queryParameters["maxresults"] = autorest.Encode("query", *maxResults) + } else { + queryParameters["maxresults"] = autorest.Encode("query", 1000) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPath("/certificates"), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client CertificateClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client CertificateClient) ListResponder(resp *http.Response) (result CertificateListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client CertificateClient) listNextResults(lastResults CertificateListResult) (result CertificateListResult, err error) { + req, err := lastResults.certificateListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "batch.CertificateClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "batch.CertificateClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.CertificateClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client CertificateClient) ListComplete(ctx context.Context, filter string, selectParameter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CertificateListResultIterator, err error) { + result.page, err = client.List(ctx, filter, selectParameter, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/client.go b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/client.go new file mode 100644 index 000000000..87ffc1ecd --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/client.go @@ -0,0 +1,49 @@ +// Package batch implements the Azure ARM Batch service API version 2017-09-01.6.0. +// +// A client for issuing REST requests to the Azure Batch service. +package batch + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "github.com/Azure/go-autorest/autorest" +) + +const ( + // DefaultBaseURI is the default URI used for the service Batch + DefaultBaseURI = "https://batch.core.windows.net" +) + +// BaseClient is the base client for Batch. +type BaseClient struct { + autorest.Client + BaseURI string +} + +// New creates an instance of the BaseClient client. +func New() BaseClient { + return NewWithBaseURI(DefaultBaseURI) +} + +// NewWithBaseURI creates an instance of the BaseClient client. +func NewWithBaseURI(baseURI string) BaseClient { + return BaseClient{ + Client: autorest.NewClientWithUserAgent(UserAgent()), + BaseURI: baseURI, + } +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/computenode.go b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/computenode.go new file mode 100644 index 000000000..f28823c04 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/computenode.go @@ -0,0 +1,1124 @@ +package batch + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "github.com/Azure/go-autorest/autorest/date" + "github.com/Azure/go-autorest/autorest/validation" + "github.com/satori/go.uuid" + "net/http" +) + +// ComputeNodeClient is the a client for issuing REST requests to the Azure Batch service. +type ComputeNodeClient struct { + BaseClient +} + +// NewComputeNodeClient creates an instance of the ComputeNodeClient client. +func NewComputeNodeClient() ComputeNodeClient { + return NewComputeNodeClientWithBaseURI(DefaultBaseURI) +} + +// NewComputeNodeClientWithBaseURI creates an instance of the ComputeNodeClient client. +func NewComputeNodeClientWithBaseURI(baseURI string) ComputeNodeClient { + return ComputeNodeClient{NewWithBaseURI(baseURI)} +} + +// AddUser you can add a user account to a node only when it is in the idle or running state. +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the machine on which you want +// to create a user account. userParameter is the user account to be created. timeout is the maximum time that the +// server can spend processing the request, in seconds. The default is 30 seconds. clientRequestID is the +// caller-generated request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client ComputeNodeClient) AddUser(ctx context.Context, poolID string, nodeID string, userParameter ComputeNodeUser, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: userParameter, + Constraints: []validation.Constraint{{Target: "userParameter.Name", Name: validation.Null, Rule: true, Chain: nil}}}}); err != nil { + return result, validation.NewError("batch.ComputeNodeClient", "AddUser", err.Error()) + } + + req, err := client.AddUserPreparer(ctx, poolID, nodeID, userParameter, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "AddUser", nil, "Failure preparing request") + return + } + + resp, err := client.AddUserSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "AddUser", resp, "Failure sending request") + return + } + + result, err = client.AddUserResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "AddUser", resp, "Failure responding to request") + } + + return +} + +// AddUserPreparer prepares the AddUser request. +func (client ComputeNodeClient) AddUserPreparer(ctx context.Context, poolID string, nodeID string, userParameter ComputeNodeUser, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}/users", pathParameters), + autorest.WithJSON(userParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// AddUserSender sends the AddUser request. The method will close the +// http.Response Body if it receives an error. +func (client ComputeNodeClient) AddUserSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// AddUserResponder handles the response to the AddUser request. The method always +// closes the http.Response Body. +func (client ComputeNodeClient) AddUserResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByClosing()) + result.Response = resp + return +} + +// DeleteUser you can delete a user account to a node only when it is in the idle or running state. +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the machine on which you want +// to delete a user account. userName is the name of the user account to delete. timeout is the maximum time that +// the server can spend processing the request, in seconds. The default is 30 seconds. clientRequestID is the +// caller-generated request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client ComputeNodeClient) DeleteUser(ctx context.Context, poolID string, nodeID string, userName string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.DeleteUserPreparer(ctx, poolID, nodeID, userName, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "DeleteUser", nil, "Failure preparing request") + return + } + + resp, err := client.DeleteUserSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "DeleteUser", resp, "Failure sending request") + return + } + + result, err = client.DeleteUserResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "DeleteUser", resp, "Failure responding to request") + } + + return +} + +// DeleteUserPreparer prepares the DeleteUser request. +func (client ComputeNodeClient) DeleteUserPreparer(ctx context.Context, poolID string, nodeID string, userName string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + "userName": autorest.Encode("path", userName), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}/users/{userName}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteUserSender sends the DeleteUser request. The method will close the +// http.Response Body if it receives an error. +func (client ComputeNodeClient) DeleteUserSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// DeleteUserResponder handles the response to the DeleteUser request. The method always +// closes the http.Response Body. +func (client ComputeNodeClient) DeleteUserResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} + +// DisableScheduling you can disable task scheduling on a node only if its current scheduling state is enabled. +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the compute node on which you +// want to disable task scheduling. nodeDisableSchedulingParameter is the parameters for the request. timeout is +// the maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client ComputeNodeClient) DisableScheduling(ctx context.Context, poolID string, nodeID string, nodeDisableSchedulingParameter *NodeDisableSchedulingParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.DisableSchedulingPreparer(ctx, poolID, nodeID, nodeDisableSchedulingParameter, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "DisableScheduling", nil, "Failure preparing request") + return + } + + resp, err := client.DisableSchedulingSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "DisableScheduling", resp, "Failure sending request") + return + } + + result, err = client.DisableSchedulingResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "DisableScheduling", resp, "Failure responding to request") + } + + return +} + +// DisableSchedulingPreparer prepares the DisableScheduling request. +func (client ComputeNodeClient) DisableSchedulingPreparer(ctx context.Context, poolID string, nodeID string, nodeDisableSchedulingParameter *NodeDisableSchedulingParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}/disablescheduling", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if nodeDisableSchedulingParameter != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithJSON(nodeDisableSchedulingParameter)) + } + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DisableSchedulingSender sends the DisableScheduling request. The method will close the +// http.Response Body if it receives an error. +func (client ComputeNodeClient) DisableSchedulingSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// DisableSchedulingResponder handles the response to the DisableScheduling request. The method always +// closes the http.Response Body. +func (client ComputeNodeClient) DisableSchedulingResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} + +// EnableScheduling you can enable task scheduling on a node only if its current scheduling state is disabled +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the compute node on which you +// want to enable task scheduling. timeout is the maximum time that the server can spend processing the request, in +// seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a +// GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID +// is whether the server should return the client-request-id in the response. ocpDate is the time the request was +// issued. Client libraries typically set this to the current system clock time; set it explicitly if you are +// calling the REST API directly. +func (client ComputeNodeClient) EnableScheduling(ctx context.Context, poolID string, nodeID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.EnableSchedulingPreparer(ctx, poolID, nodeID, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "EnableScheduling", nil, "Failure preparing request") + return + } + + resp, err := client.EnableSchedulingSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "EnableScheduling", resp, "Failure sending request") + return + } + + result, err = client.EnableSchedulingResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "EnableScheduling", resp, "Failure responding to request") + } + + return +} + +// EnableSchedulingPreparer prepares the EnableScheduling request. +func (client ComputeNodeClient) EnableSchedulingPreparer(ctx context.Context, poolID string, nodeID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}/enablescheduling", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// EnableSchedulingSender sends the EnableScheduling request. The method will close the +// http.Response Body if it receives an error. +func (client ComputeNodeClient) EnableSchedulingSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// EnableSchedulingResponder handles the response to the EnableScheduling request. The method always +// closes the http.Response Body. +func (client ComputeNodeClient) EnableSchedulingResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get sends the get request. +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the compute node that you want +// to get information about. selectParameter is an OData $select clause. timeout is the maximum time that the +// server can spend processing the request, in seconds. The default is 30 seconds. clientRequestID is the +// caller-generated request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client ComputeNodeClient) Get(ctx context.Context, poolID string, nodeID string, selectParameter string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result ComputeNode, err error) { + req, err := client.GetPreparer(ctx, poolID, nodeID, selectParameter, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client ComputeNodeClient) GetPreparer(ctx context.Context, poolID string, nodeID string, selectParameter string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client ComputeNodeClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client ComputeNodeClient) GetResponder(resp *http.Response) (result ComputeNode, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// GetRemoteDesktop before you can access a node by using the RDP file, you must create a user account on the node. +// This API can only be invoked on pools created with a cloud service configuration. For pools created with a virtual +// machine configuration, see the GetRemoteLoginSettings API. +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the compute node for which you +// want to get the Remote Desktop Protocol file. timeout is the maximum time that the server can spend processing +// the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in +// the form of a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. +// returnClientRequestID is whether the server should return the client-request-id in the response. ocpDate is the +// time the request was issued. Client libraries typically set this to the current system clock time; set it +// explicitly if you are calling the REST API directly. +func (client ComputeNodeClient) GetRemoteDesktop(ctx context.Context, poolID string, nodeID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result ReadCloser, err error) { + req, err := client.GetRemoteDesktopPreparer(ctx, poolID, nodeID, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "GetRemoteDesktop", nil, "Failure preparing request") + return + } + + resp, err := client.GetRemoteDesktopSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "GetRemoteDesktop", resp, "Failure sending request") + return + } + + result, err = client.GetRemoteDesktopResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "GetRemoteDesktop", resp, "Failure responding to request") + } + + return +} + +// GetRemoteDesktopPreparer prepares the GetRemoteDesktop request. +func (client ComputeNodeClient) GetRemoteDesktopPreparer(ctx context.Context, poolID string, nodeID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}/rdp", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetRemoteDesktopSender sends the GetRemoteDesktop request. The method will close the +// http.Response Body if it receives an error. +func (client ComputeNodeClient) GetRemoteDesktopSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetRemoteDesktopResponder handles the response to the GetRemoteDesktop request. The method always +// closes the http.Response Body. +func (client ComputeNodeClient) GetRemoteDesktopResponder(resp *http.Response) (result ReadCloser, err error) { + result.Value = &resp.Body + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK)) + result.Response = autorest.Response{Response: resp} + return +} + +// GetRemoteLoginSettings before you can remotely login to a node using the remote login settings, you must create a +// user account on the node. This API can be invoked only on pools created with the virtual machine configuration +// property. For pools created with a cloud service configuration, see the GetRemoteDesktop API. +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the compute node for which to +// obtain the remote login settings. timeout is the maximum time that the server can spend processing the request, +// in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the form of +// a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID +// is whether the server should return the client-request-id in the response. ocpDate is the time the request was +// issued. Client libraries typically set this to the current system clock time; set it explicitly if you are +// calling the REST API directly. +func (client ComputeNodeClient) GetRemoteLoginSettings(ctx context.Context, poolID string, nodeID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result ComputeNodeGetRemoteLoginSettingsResult, err error) { + req, err := client.GetRemoteLoginSettingsPreparer(ctx, poolID, nodeID, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "GetRemoteLoginSettings", nil, "Failure preparing request") + return + } + + resp, err := client.GetRemoteLoginSettingsSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "GetRemoteLoginSettings", resp, "Failure sending request") + return + } + + result, err = client.GetRemoteLoginSettingsResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "GetRemoteLoginSettings", resp, "Failure responding to request") + } + + return +} + +// GetRemoteLoginSettingsPreparer prepares the GetRemoteLoginSettings request. +func (client ComputeNodeClient) GetRemoteLoginSettingsPreparer(ctx context.Context, poolID string, nodeID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}/remoteloginsettings", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetRemoteLoginSettingsSender sends the GetRemoteLoginSettings request. The method will close the +// http.Response Body if it receives an error. +func (client ComputeNodeClient) GetRemoteLoginSettingsSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetRemoteLoginSettingsResponder handles the response to the GetRemoteLoginSettings request. The method always +// closes the http.Response Body. +func (client ComputeNodeClient) GetRemoteLoginSettingsResponder(resp *http.Response) (result ComputeNodeGetRemoteLoginSettingsResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List sends the list request. +// +// poolID is the ID of the pool from which you want to list nodes. filter is an OData $filter clause. For more +// information on constructing this filter, see +// https://docs.microsoft.com/en-us/rest/api/batchservice/odata-filters-in-batch#list-nodes-in-a-pool. +// selectParameter is an OData $select clause. maxResults is the maximum number of items to return in the response. +// A maximum of 1000 nodes can be returned. timeout is the maximum time that the server can spend processing the +// request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the +// form of a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. +// returnClientRequestID is whether the server should return the client-request-id in the response. ocpDate is the +// time the request was issued. Client libraries typically set this to the current system clock time; set it +// explicitly if you are calling the REST API directly. +func (client ComputeNodeClient) List(ctx context.Context, poolID string, filter string, selectParameter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result ComputeNodeListResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: maxResults, + Constraints: []validation.Constraint{{Target: "maxResults", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "maxResults", Name: validation.InclusiveMaximum, Rule: 1000, Chain: nil}, + {Target: "maxResults", Name: validation.InclusiveMinimum, Rule: 1, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("batch.ComputeNodeClient", "List", err.Error()) + } + + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, poolID, filter, selectParameter, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.cnlr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "List", resp, "Failure sending request") + return + } + + result.cnlr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client ComputeNodeClient) ListPreparer(ctx context.Context, poolID string, filter string, selectParameter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(filter) > 0 { + queryParameters["$filter"] = autorest.Encode("query", filter) + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if maxResults != nil { + queryParameters["maxresults"] = autorest.Encode("query", *maxResults) + } else { + queryParameters["maxresults"] = autorest.Encode("query", 1000) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client ComputeNodeClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client ComputeNodeClient) ListResponder(resp *http.Response) (result ComputeNodeListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client ComputeNodeClient) listNextResults(lastResults ComputeNodeListResult) (result ComputeNodeListResult, err error) { + req, err := lastResults.computeNodeListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client ComputeNodeClient) ListComplete(ctx context.Context, poolID string, filter string, selectParameter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result ComputeNodeListResultIterator, err error) { + result.page, err = client.List(ctx, poolID, filter, selectParameter, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + return +} + +// Reboot you can restart a node only if it is in an idle or running state. +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the compute node that you want +// to restart. nodeRebootParameter is the parameters for the request. timeout is the maximum time that the server +// can spend processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated +// request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client ComputeNodeClient) Reboot(ctx context.Context, poolID string, nodeID string, nodeRebootParameter *NodeRebootParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.RebootPreparer(ctx, poolID, nodeID, nodeRebootParameter, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "Reboot", nil, "Failure preparing request") + return + } + + resp, err := client.RebootSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "Reboot", resp, "Failure sending request") + return + } + + result, err = client.RebootResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "Reboot", resp, "Failure responding to request") + } + + return +} + +// RebootPreparer prepares the Reboot request. +func (client ComputeNodeClient) RebootPreparer(ctx context.Context, poolID string, nodeID string, nodeRebootParameter *NodeRebootParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}/reboot", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if nodeRebootParameter != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithJSON(nodeRebootParameter)) + } + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// RebootSender sends the Reboot request. The method will close the +// http.Response Body if it receives an error. +func (client ComputeNodeClient) RebootSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// RebootResponder handles the response to the Reboot request. The method always +// closes the http.Response Body. +func (client ComputeNodeClient) RebootResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// Reimage you can reinstall the operating system on a node only if it is in an idle or running state. This API can be +// invoked only on pools created with the cloud service configuration property. +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the compute node that you want +// to restart. nodeReimageParameter is the parameters for the request. timeout is the maximum time that the server +// can spend processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated +// request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client ComputeNodeClient) Reimage(ctx context.Context, poolID string, nodeID string, nodeReimageParameter *NodeReimageParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.ReimagePreparer(ctx, poolID, nodeID, nodeReimageParameter, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "Reimage", nil, "Failure preparing request") + return + } + + resp, err := client.ReimageSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "Reimage", resp, "Failure sending request") + return + } + + result, err = client.ReimageResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "Reimage", resp, "Failure responding to request") + } + + return +} + +// ReimagePreparer prepares the Reimage request. +func (client ComputeNodeClient) ReimagePreparer(ctx context.Context, poolID string, nodeID string, nodeReimageParameter *NodeReimageParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}/reimage", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if nodeReimageParameter != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithJSON(nodeReimageParameter)) + } + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ReimageSender sends the Reimage request. The method will close the +// http.Response Body if it receives an error. +func (client ComputeNodeClient) ReimageSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ReimageResponder handles the response to the Reimage request. The method always +// closes the http.Response Body. +func (client ComputeNodeClient) ReimageResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// UpdateUser this operation replaces of all the updateable properties of the account. For example, if the expiryTime +// element is not specified, the current value is replaced with the default value, not left unmodified. You can update +// a user account on a node only when it is in the idle or running state. +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the machine on which you want +// to update a user account. userName is the name of the user account to update. nodeUpdateUserParameter is the +// parameters for the request. timeout is the maximum time that the server can spend processing the request, in +// seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a +// GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID +// is whether the server should return the client-request-id in the response. ocpDate is the time the request was +// issued. Client libraries typically set this to the current system clock time; set it explicitly if you are +// calling the REST API directly. +func (client ComputeNodeClient) UpdateUser(ctx context.Context, poolID string, nodeID string, userName string, nodeUpdateUserParameter NodeUpdateUserParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.UpdateUserPreparer(ctx, poolID, nodeID, userName, nodeUpdateUserParameter, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "UpdateUser", nil, "Failure preparing request") + return + } + + resp, err := client.UpdateUserSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "UpdateUser", resp, "Failure sending request") + return + } + + result, err = client.UpdateUserResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.ComputeNodeClient", "UpdateUser", resp, "Failure responding to request") + } + + return +} + +// UpdateUserPreparer prepares the UpdateUser request. +func (client ComputeNodeClient) UpdateUserPreparer(ctx context.Context, poolID string, nodeID string, userName string, nodeUpdateUserParameter NodeUpdateUserParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + "userName": autorest.Encode("path", userName), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}/users/{userName}", pathParameters), + autorest.WithJSON(nodeUpdateUserParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// UpdateUserSender sends the UpdateUser request. The method will close the +// http.Response Body if it receives an error. +func (client ComputeNodeClient) UpdateUserSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// UpdateUserResponder handles the response to the UpdateUser request. The method always +// closes the http.Response Body. +func (client ComputeNodeClient) UpdateUserResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/file.go b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/file.go new file mode 100644 index 000000000..d05993078 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/file.go @@ -0,0 +1,946 @@ +package batch + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "github.com/Azure/go-autorest/autorest/date" + "github.com/Azure/go-autorest/autorest/validation" + "github.com/satori/go.uuid" + "net/http" +) + +// FileClient is the a client for issuing REST requests to the Azure Batch service. +type FileClient struct { + BaseClient +} + +// NewFileClient creates an instance of the FileClient client. +func NewFileClient() FileClient { + return NewFileClientWithBaseURI(DefaultBaseURI) +} + +// NewFileClientWithBaseURI creates an instance of the FileClient client. +func NewFileClientWithBaseURI(baseURI string) FileClient { + return FileClient{NewWithBaseURI(baseURI)} +} + +// DeleteFromComputeNode sends the delete from compute node request. +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the compute node from which you +// want to delete the file. filePath is the path to the file or directory that you want to delete. recursive is +// whether to delete children of a directory. If the filePath parameter represents a directory instead of a file, +// you can set recursive to true to delete the directory and all of the files and subdirectories in it. If +// recursive is false then the directory must be empty or deletion will fail. timeout is the maximum time that the +// server can spend processing the request, in seconds. The default is 30 seconds. clientRequestID is the +// caller-generated request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client FileClient) DeleteFromComputeNode(ctx context.Context, poolID string, nodeID string, filePath string, recursive *bool, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.DeleteFromComputeNodePreparer(ctx, poolID, nodeID, filePath, recursive, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "DeleteFromComputeNode", nil, "Failure preparing request") + return + } + + resp, err := client.DeleteFromComputeNodeSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.FileClient", "DeleteFromComputeNode", resp, "Failure sending request") + return + } + + result, err = client.DeleteFromComputeNodeResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "DeleteFromComputeNode", resp, "Failure responding to request") + } + + return +} + +// DeleteFromComputeNodePreparer prepares the DeleteFromComputeNode request. +func (client FileClient) DeleteFromComputeNodePreparer(ctx context.Context, poolID string, nodeID string, filePath string, recursive *bool, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "filePath": autorest.Encode("path", filePath), + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if recursive != nil { + queryParameters["recursive"] = autorest.Encode("query", *recursive) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}/files/{filePath}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteFromComputeNodeSender sends the DeleteFromComputeNode request. The method will close the +// http.Response Body if it receives an error. +func (client FileClient) DeleteFromComputeNodeSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// DeleteFromComputeNodeResponder handles the response to the DeleteFromComputeNode request. The method always +// closes the http.Response Body. +func (client FileClient) DeleteFromComputeNodeResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} + +// DeleteFromTask sends the delete from task request. +// +// jobID is the ID of the job that contains the task. taskID is the ID of the task whose file you want to delete. +// filePath is the path to the task file or directory that you want to delete. recursive is whether to delete +// children of a directory. If the filePath parameter represents a directory instead of a file, you can set +// recursive to true to delete the directory and all of the files and subdirectories in it. If recursive is false +// then the directory must be empty or deletion will fail. timeout is the maximum time that the server can spend +// processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request +// identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client FileClient) DeleteFromTask(ctx context.Context, jobID string, taskID string, filePath string, recursive *bool, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.DeleteFromTaskPreparer(ctx, jobID, taskID, filePath, recursive, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "DeleteFromTask", nil, "Failure preparing request") + return + } + + resp, err := client.DeleteFromTaskSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.FileClient", "DeleteFromTask", resp, "Failure sending request") + return + } + + result, err = client.DeleteFromTaskResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "DeleteFromTask", resp, "Failure responding to request") + } + + return +} + +// DeleteFromTaskPreparer prepares the DeleteFromTask request. +func (client FileClient) DeleteFromTaskPreparer(ctx context.Context, jobID string, taskID string, filePath string, recursive *bool, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "filePath": autorest.Encode("path", filePath), + "jobId": autorest.Encode("path", jobID), + "taskId": autorest.Encode("path", taskID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if recursive != nil { + queryParameters["recursive"] = autorest.Encode("query", *recursive) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/tasks/{taskId}/files/{filePath}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteFromTaskSender sends the DeleteFromTask request. The method will close the +// http.Response Body if it receives an error. +func (client FileClient) DeleteFromTaskSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// DeleteFromTaskResponder handles the response to the DeleteFromTask request. The method always +// closes the http.Response Body. +func (client FileClient) DeleteFromTaskResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} + +// GetFromComputeNode returns the content of the specified compute node file. +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the compute node that contains +// the file. filePath is the path to the compute node file that you want to get the content of. timeout is the +// maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ocpRange is +// the byte range to be retrieved. The default is to retrieve the entire file. The format is +// bytes=startRange-endRange. ifModifiedSince is a timestamp indicating the last modified time of the resource +// known to the client. The operation will be performed only if the resource on the service has been modified since +// the specified time. ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to +// the client. The operation will be performed only if the resource on the service has not been modified since the +// specified time. +func (client FileClient) GetFromComputeNode(ctx context.Context, poolID string, nodeID string, filePath string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ocpRange string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result ReadCloser, err error) { + req, err := client.GetFromComputeNodePreparer(ctx, poolID, nodeID, filePath, timeout, clientRequestID, returnClientRequestID, ocpDate, ocpRange, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "GetFromComputeNode", nil, "Failure preparing request") + return + } + + resp, err := client.GetFromComputeNodeSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.FileClient", "GetFromComputeNode", resp, "Failure sending request") + return + } + + result, err = client.GetFromComputeNodeResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "GetFromComputeNode", resp, "Failure responding to request") + } + + return +} + +// GetFromComputeNodePreparer prepares the GetFromComputeNode request. +func (client FileClient) GetFromComputeNodePreparer(ctx context.Context, poolID string, nodeID string, filePath string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ocpRange string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "filePath": autorest.Encode("path", filePath), + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}/files/{filePath}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ocpRange) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-range", autorest.String(ocpRange))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetFromComputeNodeSender sends the GetFromComputeNode request. The method will close the +// http.Response Body if it receives an error. +func (client FileClient) GetFromComputeNodeSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetFromComputeNodeResponder handles the response to the GetFromComputeNode request. The method always +// closes the http.Response Body. +func (client FileClient) GetFromComputeNodeResponder(resp *http.Response) (result ReadCloser, err error) { + result.Value = &resp.Body + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK)) + result.Response = autorest.Response{Response: resp} + return +} + +// GetFromTask returns the content of the specified task file. +// +// jobID is the ID of the job that contains the task. taskID is the ID of the task whose file you want to retrieve. +// filePath is the path to the task file that you want to get the content of. timeout is the maximum time that the +// server can spend processing the request, in seconds. The default is 30 seconds. clientRequestID is the +// caller-generated request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ocpRange is +// the byte range to be retrieved. The default is to retrieve the entire file. The format is +// bytes=startRange-endRange. ifModifiedSince is a timestamp indicating the last modified time of the resource +// known to the client. The operation will be performed only if the resource on the service has been modified since +// the specified time. ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to +// the client. The operation will be performed only if the resource on the service has not been modified since the +// specified time. +func (client FileClient) GetFromTask(ctx context.Context, jobID string, taskID string, filePath string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ocpRange string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result ReadCloser, err error) { + req, err := client.GetFromTaskPreparer(ctx, jobID, taskID, filePath, timeout, clientRequestID, returnClientRequestID, ocpDate, ocpRange, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "GetFromTask", nil, "Failure preparing request") + return + } + + resp, err := client.GetFromTaskSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.FileClient", "GetFromTask", resp, "Failure sending request") + return + } + + result, err = client.GetFromTaskResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "GetFromTask", resp, "Failure responding to request") + } + + return +} + +// GetFromTaskPreparer prepares the GetFromTask request. +func (client FileClient) GetFromTaskPreparer(ctx context.Context, jobID string, taskID string, filePath string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ocpRange string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "filePath": autorest.Encode("path", filePath), + "jobId": autorest.Encode("path", jobID), + "taskId": autorest.Encode("path", taskID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/tasks/{taskId}/files/{filePath}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ocpRange) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-range", autorest.String(ocpRange))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetFromTaskSender sends the GetFromTask request. The method will close the +// http.Response Body if it receives an error. +func (client FileClient) GetFromTaskSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetFromTaskResponder handles the response to the GetFromTask request. The method always +// closes the http.Response Body. +func (client FileClient) GetFromTaskResponder(resp *http.Response) (result ReadCloser, err error) { + result.Value = &resp.Body + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK)) + result.Response = autorest.Response{Response: resp} + return +} + +// GetPropertiesFromComputeNode gets the properties of the specified compute node file. +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the compute node that contains +// the file. filePath is the path to the compute node file that you want to get the properties of. timeout is the +// maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client FileClient) GetPropertiesFromComputeNode(ctx context.Context, poolID string, nodeID string, filePath string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.GetPropertiesFromComputeNodePreparer(ctx, poolID, nodeID, filePath, timeout, clientRequestID, returnClientRequestID, ocpDate, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "GetPropertiesFromComputeNode", nil, "Failure preparing request") + return + } + + resp, err := client.GetPropertiesFromComputeNodeSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.FileClient", "GetPropertiesFromComputeNode", resp, "Failure sending request") + return + } + + result, err = client.GetPropertiesFromComputeNodeResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "GetPropertiesFromComputeNode", resp, "Failure responding to request") + } + + return +} + +// GetPropertiesFromComputeNodePreparer prepares the GetPropertiesFromComputeNode request. +func (client FileClient) GetPropertiesFromComputeNodePreparer(ctx context.Context, poolID string, nodeID string, filePath string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "filePath": autorest.Encode("path", filePath), + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsHead(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}/files/{filePath}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetPropertiesFromComputeNodeSender sends the GetPropertiesFromComputeNode request. The method will close the +// http.Response Body if it receives an error. +func (client FileClient) GetPropertiesFromComputeNodeSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetPropertiesFromComputeNodeResponder handles the response to the GetPropertiesFromComputeNode request. The method always +// closes the http.Response Body. +func (client FileClient) GetPropertiesFromComputeNodeResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} + +// GetPropertiesFromTask gets the properties of the specified task file. +// +// jobID is the ID of the job that contains the task. taskID is the ID of the task whose file you want to get the +// properties of. filePath is the path to the task file that you want to get the properties of. timeout is the +// maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client FileClient) GetPropertiesFromTask(ctx context.Context, jobID string, taskID string, filePath string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.GetPropertiesFromTaskPreparer(ctx, jobID, taskID, filePath, timeout, clientRequestID, returnClientRequestID, ocpDate, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "GetPropertiesFromTask", nil, "Failure preparing request") + return + } + + resp, err := client.GetPropertiesFromTaskSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.FileClient", "GetPropertiesFromTask", resp, "Failure sending request") + return + } + + result, err = client.GetPropertiesFromTaskResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "GetPropertiesFromTask", resp, "Failure responding to request") + } + + return +} + +// GetPropertiesFromTaskPreparer prepares the GetPropertiesFromTask request. +func (client FileClient) GetPropertiesFromTaskPreparer(ctx context.Context, jobID string, taskID string, filePath string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "filePath": autorest.Encode("path", filePath), + "jobId": autorest.Encode("path", jobID), + "taskId": autorest.Encode("path", taskID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsHead(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/tasks/{taskId}/files/{filePath}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetPropertiesFromTaskSender sends the GetPropertiesFromTask request. The method will close the +// http.Response Body if it receives an error. +func (client FileClient) GetPropertiesFromTaskSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetPropertiesFromTaskResponder handles the response to the GetPropertiesFromTask request. The method always +// closes the http.Response Body. +func (client FileClient) GetPropertiesFromTaskResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} + +// ListFromComputeNode sends the list from compute node request. +// +// poolID is the ID of the pool that contains the compute node. nodeID is the ID of the compute node whose files +// you want to list. filter is an OData $filter clause. For more information on constructing this filter, see +// https://docs.microsoft.com/en-us/rest/api/batchservice/odata-filters-in-batch#list-compute-node-files. recursive +// is whether to list children of a directory. maxResults is the maximum number of items to return in the response. +// A maximum of 1000 files can be returned. timeout is the maximum time that the server can spend processing the +// request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the +// form of a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. +// returnClientRequestID is whether the server should return the client-request-id in the response. ocpDate is the +// time the request was issued. Client libraries typically set this to the current system clock time; set it +// explicitly if you are calling the REST API directly. +func (client FileClient) ListFromComputeNode(ctx context.Context, poolID string, nodeID string, filter string, recursive *bool, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result NodeFileListResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: maxResults, + Constraints: []validation.Constraint{{Target: "maxResults", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "maxResults", Name: validation.InclusiveMaximum, Rule: 1000, Chain: nil}, + {Target: "maxResults", Name: validation.InclusiveMinimum, Rule: 1, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("batch.FileClient", "ListFromComputeNode", err.Error()) + } + + result.fn = client.listFromComputeNodeNextResults + req, err := client.ListFromComputeNodePreparer(ctx, poolID, nodeID, filter, recursive, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "ListFromComputeNode", nil, "Failure preparing request") + return + } + + resp, err := client.ListFromComputeNodeSender(req) + if err != nil { + result.nflr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.FileClient", "ListFromComputeNode", resp, "Failure sending request") + return + } + + result.nflr, err = client.ListFromComputeNodeResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "ListFromComputeNode", resp, "Failure responding to request") + } + + return +} + +// ListFromComputeNodePreparer prepares the ListFromComputeNode request. +func (client FileClient) ListFromComputeNodePreparer(ctx context.Context, poolID string, nodeID string, filter string, recursive *bool, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "nodeId": autorest.Encode("path", nodeID), + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(filter) > 0 { + queryParameters["$filter"] = autorest.Encode("query", filter) + } + if recursive != nil { + queryParameters["recursive"] = autorest.Encode("query", *recursive) + } + if maxResults != nil { + queryParameters["maxresults"] = autorest.Encode("query", *maxResults) + } else { + queryParameters["maxresults"] = autorest.Encode("query", 1000) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/nodes/{nodeId}/files", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListFromComputeNodeSender sends the ListFromComputeNode request. The method will close the +// http.Response Body if it receives an error. +func (client FileClient) ListFromComputeNodeSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListFromComputeNodeResponder handles the response to the ListFromComputeNode request. The method always +// closes the http.Response Body. +func (client FileClient) ListFromComputeNodeResponder(resp *http.Response) (result NodeFileListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listFromComputeNodeNextResults retrieves the next set of results, if any. +func (client FileClient) listFromComputeNodeNextResults(lastResults NodeFileListResult) (result NodeFileListResult, err error) { + req, err := lastResults.nodeFileListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "batch.FileClient", "listFromComputeNodeNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListFromComputeNodeSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "batch.FileClient", "listFromComputeNodeNextResults", resp, "Failure sending next results request") + } + result, err = client.ListFromComputeNodeResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "listFromComputeNodeNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListFromComputeNodeComplete enumerates all values, automatically crossing page boundaries as required. +func (client FileClient) ListFromComputeNodeComplete(ctx context.Context, poolID string, nodeID string, filter string, recursive *bool, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result NodeFileListResultIterator, err error) { + result.page, err = client.ListFromComputeNode(ctx, poolID, nodeID, filter, recursive, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + return +} + +// ListFromTask sends the list from task request. +// +// jobID is the ID of the job that contains the task. taskID is the ID of the task whose files you want to list. +// filter is an OData $filter clause. For more information on constructing this filter, see +// https://docs.microsoft.com/en-us/rest/api/batchservice/odata-filters-in-batch#list-task-files. recursive is +// whether to list children of the task directory. This parameter can be used in combination with the filter +// parameter to list specific type of files. maxResults is the maximum number of items to return in the response. A +// maximum of 1000 files can be returned. timeout is the maximum time that the server can spend processing the +// request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the +// form of a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. +// returnClientRequestID is whether the server should return the client-request-id in the response. ocpDate is the +// time the request was issued. Client libraries typically set this to the current system clock time; set it +// explicitly if you are calling the REST API directly. +func (client FileClient) ListFromTask(ctx context.Context, jobID string, taskID string, filter string, recursive *bool, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result NodeFileListResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: maxResults, + Constraints: []validation.Constraint{{Target: "maxResults", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "maxResults", Name: validation.InclusiveMaximum, Rule: 1000, Chain: nil}, + {Target: "maxResults", Name: validation.InclusiveMinimum, Rule: 1, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("batch.FileClient", "ListFromTask", err.Error()) + } + + result.fn = client.listFromTaskNextResults + req, err := client.ListFromTaskPreparer(ctx, jobID, taskID, filter, recursive, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "ListFromTask", nil, "Failure preparing request") + return + } + + resp, err := client.ListFromTaskSender(req) + if err != nil { + result.nflr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.FileClient", "ListFromTask", resp, "Failure sending request") + return + } + + result.nflr, err = client.ListFromTaskResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "ListFromTask", resp, "Failure responding to request") + } + + return +} + +// ListFromTaskPreparer prepares the ListFromTask request. +func (client FileClient) ListFromTaskPreparer(ctx context.Context, jobID string, taskID string, filter string, recursive *bool, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + "taskId": autorest.Encode("path", taskID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(filter) > 0 { + queryParameters["$filter"] = autorest.Encode("query", filter) + } + if recursive != nil { + queryParameters["recursive"] = autorest.Encode("query", *recursive) + } + if maxResults != nil { + queryParameters["maxresults"] = autorest.Encode("query", *maxResults) + } else { + queryParameters["maxresults"] = autorest.Encode("query", 1000) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/tasks/{taskId}/files", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListFromTaskSender sends the ListFromTask request. The method will close the +// http.Response Body if it receives an error. +func (client FileClient) ListFromTaskSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListFromTaskResponder handles the response to the ListFromTask request. The method always +// closes the http.Response Body. +func (client FileClient) ListFromTaskResponder(resp *http.Response) (result NodeFileListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listFromTaskNextResults retrieves the next set of results, if any. +func (client FileClient) listFromTaskNextResults(lastResults NodeFileListResult) (result NodeFileListResult, err error) { + req, err := lastResults.nodeFileListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "batch.FileClient", "listFromTaskNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListFromTaskSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "batch.FileClient", "listFromTaskNextResults", resp, "Failure sending next results request") + } + result, err = client.ListFromTaskResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.FileClient", "listFromTaskNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListFromTaskComplete enumerates all values, automatically crossing page boundaries as required. +func (client FileClient) ListFromTaskComplete(ctx context.Context, jobID string, taskID string, filter string, recursive *bool, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result NodeFileListResultIterator, err error) { + result.page, err = client.ListFromTask(ctx, jobID, taskID, filter, recursive, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/job.go b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/job.go new file mode 100644 index 000000000..a1ce513e8 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/job.go @@ -0,0 +1,1660 @@ +package batch + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "github.com/Azure/go-autorest/autorest/date" + "github.com/Azure/go-autorest/autorest/validation" + "github.com/satori/go.uuid" + "net/http" +) + +// JobClient is the a client for issuing REST requests to the Azure Batch service. +type JobClient struct { + BaseClient +} + +// NewJobClient creates an instance of the JobClient client. +func NewJobClient() JobClient { + return NewJobClientWithBaseURI(DefaultBaseURI) +} + +// NewJobClientWithBaseURI creates an instance of the JobClient client. +func NewJobClientWithBaseURI(baseURI string) JobClient { + return JobClient{NewWithBaseURI(baseURI)} +} + +// Add the Batch service supports two ways to control the work done as part of a job. In the first approach, the user +// specifies a Job Manager task. The Batch service launches this task when it is ready to start the job. The Job +// Manager task controls all other tasks that run under this job, by using the Task APIs. In the second approach, the +// user directly controls the execution of tasks under an active job, by using the Task APIs. Also note: when naming +// jobs, avoid including sensitive information such as user names or secret project names. This information may appear +// in telemetry logs accessible to Microsoft Support engineers. +// +// job is the job to be added. timeout is the maximum time that the server can spend processing the request, in +// seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a +// GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID +// is whether the server should return the client-request-id in the response. ocpDate is the time the request was +// issued. Client libraries typically set this to the current system clock time; set it explicitly if you are +// calling the REST API directly. +func (client JobClient) Add(ctx context.Context, job JobAddParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: job, + Constraints: []validation.Constraint{{Target: "job.ID", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.JobManagerTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.JobManagerTask.ID", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.JobManagerTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.JobManagerTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.JobManagerTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.JobManagerTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.JobManagerTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.JobManagerTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + {Target: "job.JobPreparationTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.JobPreparationTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.JobPreparationTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.JobPreparationTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.JobPreparationTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.JobPreparationTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.JobPreparationTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + {Target: "job.JobReleaseTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.JobReleaseTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.JobReleaseTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.JobReleaseTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.JobReleaseTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.JobReleaseTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.JobReleaseTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + {Target: "job.PoolInfo", Name: validation.Null, Rule: true, + Chain: []validation.Constraint{{Target: "job.PoolInfo.AutoPoolSpecification", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.PoolInfo.AutoPoolSpecification.Pool", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.PoolInfo.AutoPoolSpecification.Pool.VMSize", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.PoolInfo.AutoPoolSpecification.Pool.CloudServiceConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.PoolInfo.AutoPoolSpecification.Pool.CloudServiceConfiguration.OsFamily", Name: validation.Null, Rule: true, Chain: nil}}}, + {Target: "job.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.ImageReference", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.NodeAgentSKUID", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.ContainerConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.ContainerConfiguration.Type", Name: validation.Null, Rule: true, Chain: nil}}}, + }}, + {Target: "job.PoolInfo.AutoPoolSpecification.Pool.NetworkConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.PoolInfo.AutoPoolSpecification.Pool.NetworkConfiguration.EndpointConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.PoolInfo.AutoPoolSpecification.Pool.NetworkConfiguration.EndpointConfiguration.InboundNATPools", Name: validation.Null, Rule: true, Chain: nil}}}, + }}, + {Target: "job.PoolInfo.AutoPoolSpecification.Pool.StartTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.PoolInfo.AutoPoolSpecification.Pool.StartTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "job.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "job.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + }}, + }}, + }}}}}); err != nil { + return result, validation.NewError("batch.JobClient", "Add", err.Error()) + } + + req, err := client.AddPreparer(ctx, job, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Add", nil, "Failure preparing request") + return + } + + resp, err := client.AddSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobClient", "Add", resp, "Failure sending request") + return + } + + result, err = client.AddResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Add", resp, "Failure responding to request") + } + + return +} + +// AddPreparer prepares the Add request. +func (client JobClient) AddPreparer(ctx context.Context, job JobAddParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPath("/jobs"), + autorest.WithJSON(job), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// AddSender sends the Add request. The method will close the +// http.Response Body if it receives an error. +func (client JobClient) AddSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// AddResponder handles the response to the Add request. The method always +// closes the http.Response Body. +func (client JobClient) AddResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByClosing()) + result.Response = resp + return +} + +// Delete deleting a job also deletes all tasks that are part of that job, and all job statistics. This also overrides +// the retention period for task data; that is, if the job contains tasks which are still retained on compute nodes, +// the Batch services deletes those tasks' working directories and all their contents. When a Delete Job request is +// received, the Batch service sets the job to the deleting state. All update operations on a job that is in deleting +// state will fail with status code 409 (Conflict), with additional information indicating that the job is being +// deleted. +// +// jobID is the ID of the job to delete. timeout is the maximum time that the server can spend processing the +// request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the +// form of a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. +// returnClientRequestID is whether the server should return the client-request-id in the response. ocpDate is the +// time the request was issued. Client libraries typically set this to the current system clock time; set it +// explicitly if you are calling the REST API directly. ifMatch is an ETag value associated with the version of the +// resource known to the client. The operation will be performed only if the resource's current ETag on the service +// exactly matches the value specified by the client. ifNoneMatch is an ETag value associated with the version of +// the resource known to the client. The operation will be performed only if the resource's current ETag on the +// service does not match the value specified by the client. ifModifiedSince is a timestamp indicating the last +// modified time of the resource known to the client. The operation will be performed only if the resource on the +// service has been modified since the specified time. ifUnmodifiedSince is a timestamp indicating the last +// modified time of the resource known to the client. The operation will be performed only if the resource on the +// service has not been modified since the specified time. +func (client JobClient) Delete(ctx context.Context, jobID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.DeletePreparer(ctx, jobID, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Delete", nil, "Failure preparing request") + return + } + + resp, err := client.DeleteSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobClient", "Delete", resp, "Failure sending request") + return + } + + result, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Delete", resp, "Failure responding to request") + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client JobClient) DeletePreparer(ctx context.Context, jobID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client JobClient) DeleteSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client JobClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// Disable the Batch Service immediately moves the job to the disabling state. Batch then uses the disableTasks +// parameter to determine what to do with the currently running tasks of the job. The job remains in the disabling +// state until the disable operation is completed and all tasks have been dealt with according to the disableTasks +// option; the job then moves to the disabled state. No new tasks are started under the job until it moves back to +// active state. If you try to disable a job that is in any state other than active, disabling, or disabled, the +// request fails with status code 409. +// +// jobID is the ID of the job to disable. jobDisableParameter is the parameters for the request. timeout is the +// maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client JobClient) Disable(ctx context.Context, jobID string, jobDisableParameter JobDisableParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.DisablePreparer(ctx, jobID, jobDisableParameter, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Disable", nil, "Failure preparing request") + return + } + + resp, err := client.DisableSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobClient", "Disable", resp, "Failure sending request") + return + } + + result, err = client.DisableResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Disable", resp, "Failure responding to request") + } + + return +} + +// DisablePreparer prepares the Disable request. +func (client JobClient) DisablePreparer(ctx context.Context, jobID string, jobDisableParameter JobDisableParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/disable", pathParameters), + autorest.WithJSON(jobDisableParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DisableSender sends the Disable request. The method will close the +// http.Response Body if it receives an error. +func (client JobClient) DisableSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// DisableResponder handles the response to the Disable request. The method always +// closes the http.Response Body. +func (client JobClient) DisableResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// Enable when you call this API, the Batch service sets a disabled job to the enabling state. After the this operation +// is completed, the job moves to the active state, and scheduling of new tasks under the job resumes. The Batch +// service does not allow a task to remain in the active state for more than 7 days. Therefore, if you enable a job +// containing active tasks which were added more than 7 days ago, those tasks will not run. +// +// jobID is the ID of the job to enable. timeout is the maximum time that the server can spend processing the +// request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the +// form of a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. +// returnClientRequestID is whether the server should return the client-request-id in the response. ocpDate is the +// time the request was issued. Client libraries typically set this to the current system clock time; set it +// explicitly if you are calling the REST API directly. ifMatch is an ETag value associated with the version of the +// resource known to the client. The operation will be performed only if the resource's current ETag on the service +// exactly matches the value specified by the client. ifNoneMatch is an ETag value associated with the version of +// the resource known to the client. The operation will be performed only if the resource's current ETag on the +// service does not match the value specified by the client. ifModifiedSince is a timestamp indicating the last +// modified time of the resource known to the client. The operation will be performed only if the resource on the +// service has been modified since the specified time. ifUnmodifiedSince is a timestamp indicating the last +// modified time of the resource known to the client. The operation will be performed only if the resource on the +// service has not been modified since the specified time. +func (client JobClient) Enable(ctx context.Context, jobID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.EnablePreparer(ctx, jobID, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Enable", nil, "Failure preparing request") + return + } + + resp, err := client.EnableSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobClient", "Enable", resp, "Failure sending request") + return + } + + result, err = client.EnableResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Enable", resp, "Failure responding to request") + } + + return +} + +// EnablePreparer prepares the Enable request. +func (client JobClient) EnablePreparer(ctx context.Context, jobID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/enable", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// EnableSender sends the Enable request. The method will close the +// http.Response Body if it receives an error. +func (client JobClient) EnableSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// EnableResponder handles the response to the Enable request. The method always +// closes the http.Response Body. +func (client JobClient) EnableResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get sends the get request. +// +// jobID is the ID of the job. selectParameter is an OData $select clause. expand is an OData $expand clause. +// timeout is the maximum time that the server can spend processing the request, in seconds. The default is 30 +// seconds. clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such +// as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should +// return the client-request-id in the response. ocpDate is the time the request was issued. Client libraries +// typically set this to the current system clock time; set it explicitly if you are calling the REST API directly. +// ifMatch is an ETag value associated with the version of the resource known to the client. The operation will be +// performed only if the resource's current ETag on the service exactly matches the value specified by the client. +// ifNoneMatch is an ETag value associated with the version of the resource known to the client. The operation will +// be performed only if the resource's current ETag on the service does not match the value specified by the +// client. ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. +// The operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client JobClient) Get(ctx context.Context, jobID string, selectParameter string, expand string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result CloudJob, err error) { + req, err := client.GetPreparer(ctx, jobID, selectParameter, expand, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.JobClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client JobClient) GetPreparer(ctx context.Context, jobID string, selectParameter string, expand string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client JobClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client JobClient) GetResponder(resp *http.Response) (result CloudJob, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// GetAllLifetimeStatistics statistics are aggregated across all jobs that have ever existed in the account, from +// account creation to the last update time of the statistics. +// +// timeout is the maximum time that the server can spend processing the request, in seconds. The default is 30 +// seconds. clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such +// as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should +// return the client-request-id in the response. ocpDate is the time the request was issued. Client libraries +// typically set this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client JobClient) GetAllLifetimeStatistics(ctx context.Context, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result JobStatistics, err error) { + req, err := client.GetAllLifetimeStatisticsPreparer(ctx, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "GetAllLifetimeStatistics", nil, "Failure preparing request") + return + } + + resp, err := client.GetAllLifetimeStatisticsSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.JobClient", "GetAllLifetimeStatistics", resp, "Failure sending request") + return + } + + result, err = client.GetAllLifetimeStatisticsResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "GetAllLifetimeStatistics", resp, "Failure responding to request") + } + + return +} + +// GetAllLifetimeStatisticsPreparer prepares the GetAllLifetimeStatistics request. +func (client JobClient) GetAllLifetimeStatisticsPreparer(ctx context.Context, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPath("/lifetimejobstats"), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetAllLifetimeStatisticsSender sends the GetAllLifetimeStatistics request. The method will close the +// http.Response Body if it receives an error. +func (client JobClient) GetAllLifetimeStatisticsSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetAllLifetimeStatisticsResponder handles the response to the GetAllLifetimeStatistics request. The method always +// closes the http.Response Body. +func (client JobClient) GetAllLifetimeStatisticsResponder(resp *http.Response) (result JobStatistics, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// GetTaskCounts task counts provide a count of the tasks by active, running or completed task state, and a count of +// tasks which succeeded or failed. Tasks in the preparing state are counted as running. If the validationStatus is +// unvalidated, then the Batch service has not been able to check state counts against the task states as reported in +// the List Tasks API. The validationStatus may be unvalidated if the job contains more than 200,000 tasks. +// +// jobID is the ID of the job. timeout is the maximum time that the server can spend processing the request, in +// seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a +// GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID +// is whether the server should return the client-request-id in the response. ocpDate is the time the request was +// issued. Client libraries typically set this to the current system clock time; set it explicitly if you are +// calling the REST API directly. +func (client JobClient) GetTaskCounts(ctx context.Context, jobID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result TaskCounts, err error) { + req, err := client.GetTaskCountsPreparer(ctx, jobID, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "GetTaskCounts", nil, "Failure preparing request") + return + } + + resp, err := client.GetTaskCountsSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.JobClient", "GetTaskCounts", resp, "Failure sending request") + return + } + + result, err = client.GetTaskCountsResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "GetTaskCounts", resp, "Failure responding to request") + } + + return +} + +// GetTaskCountsPreparer prepares the GetTaskCounts request. +func (client JobClient) GetTaskCountsPreparer(ctx context.Context, jobID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/taskcounts", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetTaskCountsSender sends the GetTaskCounts request. The method will close the +// http.Response Body if it receives an error. +func (client JobClient) GetTaskCountsSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetTaskCountsResponder handles the response to the GetTaskCounts request. The method always +// closes the http.Response Body. +func (client JobClient) GetTaskCountsResponder(resp *http.Response) (result TaskCounts, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List sends the list request. +// +// filter is an OData $filter clause. For more information on constructing this filter, see +// https://docs.microsoft.com/en-us/rest/api/batchservice/odata-filters-in-batch#list-jobs. selectParameter is an +// OData $select clause. expand is an OData $expand clause. maxResults is the maximum number of items to return in +// the response. A maximum of 1000 jobs can be returned. timeout is the maximum time that the server can spend +// processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request +// identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client JobClient) List(ctx context.Context, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CloudJobListResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: maxResults, + Constraints: []validation.Constraint{{Target: "maxResults", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "maxResults", Name: validation.InclusiveMaximum, Rule: 1000, Chain: nil}, + {Target: "maxResults", Name: validation.InclusiveMinimum, Rule: 1, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("batch.JobClient", "List", err.Error()) + } + + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, filter, selectParameter, expand, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.cjlr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.JobClient", "List", resp, "Failure sending request") + return + } + + result.cjlr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client JobClient) ListPreparer(ctx context.Context, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(filter) > 0 { + queryParameters["$filter"] = autorest.Encode("query", filter) + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + if maxResults != nil { + queryParameters["maxresults"] = autorest.Encode("query", *maxResults) + } else { + queryParameters["maxresults"] = autorest.Encode("query", 1000) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPath("/jobs"), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client JobClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client JobClient) ListResponder(resp *http.Response) (result CloudJobListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client JobClient) listNextResults(lastResults CloudJobListResult) (result CloudJobListResult, err error) { + req, err := lastResults.cloudJobListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "batch.JobClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "batch.JobClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client JobClient) ListComplete(ctx context.Context, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CloudJobListResultIterator, err error) { + result.page, err = client.List(ctx, filter, selectParameter, expand, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + return +} + +// ListFromJobSchedule sends the list from job schedule request. +// +// jobScheduleID is the ID of the job schedule from which you want to get a list of jobs. filter is an OData +// $filter clause. For more information on constructing this filter, see +// https://docs.microsoft.com/en-us/rest/api/batchservice/odata-filters-in-batch#list-jobs-in-a-job-schedule. +// selectParameter is an OData $select clause. expand is an OData $expand clause. maxResults is the maximum number +// of items to return in the response. A maximum of 1000 jobs can be returned. timeout is the maximum time that the +// server can spend processing the request, in seconds. The default is 30 seconds. clientRequestID is the +// caller-generated request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client JobClient) ListFromJobSchedule(ctx context.Context, jobScheduleID string, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CloudJobListResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: maxResults, + Constraints: []validation.Constraint{{Target: "maxResults", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "maxResults", Name: validation.InclusiveMaximum, Rule: 1000, Chain: nil}, + {Target: "maxResults", Name: validation.InclusiveMinimum, Rule: 1, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("batch.JobClient", "ListFromJobSchedule", err.Error()) + } + + result.fn = client.listFromJobScheduleNextResults + req, err := client.ListFromJobSchedulePreparer(ctx, jobScheduleID, filter, selectParameter, expand, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "ListFromJobSchedule", nil, "Failure preparing request") + return + } + + resp, err := client.ListFromJobScheduleSender(req) + if err != nil { + result.cjlr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.JobClient", "ListFromJobSchedule", resp, "Failure sending request") + return + } + + result.cjlr, err = client.ListFromJobScheduleResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "ListFromJobSchedule", resp, "Failure responding to request") + } + + return +} + +// ListFromJobSchedulePreparer prepares the ListFromJobSchedule request. +func (client JobClient) ListFromJobSchedulePreparer(ctx context.Context, jobScheduleID string, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobScheduleId": autorest.Encode("path", jobScheduleID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(filter) > 0 { + queryParameters["$filter"] = autorest.Encode("query", filter) + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + if maxResults != nil { + queryParameters["maxresults"] = autorest.Encode("query", *maxResults) + } else { + queryParameters["maxresults"] = autorest.Encode("query", 1000) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobschedules/{jobScheduleId}/jobs", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListFromJobScheduleSender sends the ListFromJobSchedule request. The method will close the +// http.Response Body if it receives an error. +func (client JobClient) ListFromJobScheduleSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListFromJobScheduleResponder handles the response to the ListFromJobSchedule request. The method always +// closes the http.Response Body. +func (client JobClient) ListFromJobScheduleResponder(resp *http.Response) (result CloudJobListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listFromJobScheduleNextResults retrieves the next set of results, if any. +func (client JobClient) listFromJobScheduleNextResults(lastResults CloudJobListResult) (result CloudJobListResult, err error) { + req, err := lastResults.cloudJobListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "batch.JobClient", "listFromJobScheduleNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListFromJobScheduleSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "batch.JobClient", "listFromJobScheduleNextResults", resp, "Failure sending next results request") + } + result, err = client.ListFromJobScheduleResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "listFromJobScheduleNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListFromJobScheduleComplete enumerates all values, automatically crossing page boundaries as required. +func (client JobClient) ListFromJobScheduleComplete(ctx context.Context, jobScheduleID string, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CloudJobListResultIterator, err error) { + result.page, err = client.ListFromJobSchedule(ctx, jobScheduleID, filter, selectParameter, expand, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + return +} + +// ListPreparationAndReleaseTaskStatus this API returns the Job Preparation and Job Release task status on all compute +// nodes that have run the Job Preparation or Job Release task. This includes nodes which have since been removed from +// the pool. If this API is invoked on a job which has no Job Preparation or Job Release task, the Batch service +// returns HTTP status code 409 (Conflict) with an error code of JobPreparationTaskNotSpecified. +// +// jobID is the ID of the job. filter is an OData $filter clause. For more information on constructing this filter, +// see +// https://docs.microsoft.com/en-us/rest/api/batchservice/odata-filters-in-batch#list-job-preparation-and-release-status. +// selectParameter is an OData $select clause. maxResults is the maximum number of items to return in the response. +// A maximum of 1000 tasks can be returned. timeout is the maximum time that the server can spend processing the +// request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the +// form of a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. +// returnClientRequestID is whether the server should return the client-request-id in the response. ocpDate is the +// time the request was issued. Client libraries typically set this to the current system clock time; set it +// explicitly if you are calling the REST API directly. +func (client JobClient) ListPreparationAndReleaseTaskStatus(ctx context.Context, jobID string, filter string, selectParameter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CloudJobListPreparationAndReleaseTaskStatusResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: maxResults, + Constraints: []validation.Constraint{{Target: "maxResults", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "maxResults", Name: validation.InclusiveMaximum, Rule: 1000, Chain: nil}, + {Target: "maxResults", Name: validation.InclusiveMinimum, Rule: 1, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("batch.JobClient", "ListPreparationAndReleaseTaskStatus", err.Error()) + } + + result.fn = client.listPreparationAndReleaseTaskStatusNextResults + req, err := client.ListPreparationAndReleaseTaskStatusPreparer(ctx, jobID, filter, selectParameter, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "ListPreparationAndReleaseTaskStatus", nil, "Failure preparing request") + return + } + + resp, err := client.ListPreparationAndReleaseTaskStatusSender(req) + if err != nil { + result.cjlpartsr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.JobClient", "ListPreparationAndReleaseTaskStatus", resp, "Failure sending request") + return + } + + result.cjlpartsr, err = client.ListPreparationAndReleaseTaskStatusResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "ListPreparationAndReleaseTaskStatus", resp, "Failure responding to request") + } + + return +} + +// ListPreparationAndReleaseTaskStatusPreparer prepares the ListPreparationAndReleaseTaskStatus request. +func (client JobClient) ListPreparationAndReleaseTaskStatusPreparer(ctx context.Context, jobID string, filter string, selectParameter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(filter) > 0 { + queryParameters["$filter"] = autorest.Encode("query", filter) + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if maxResults != nil { + queryParameters["maxresults"] = autorest.Encode("query", *maxResults) + } else { + queryParameters["maxresults"] = autorest.Encode("query", 1000) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/jobpreparationandreleasetaskstatus", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListPreparationAndReleaseTaskStatusSender sends the ListPreparationAndReleaseTaskStatus request. The method will close the +// http.Response Body if it receives an error. +func (client JobClient) ListPreparationAndReleaseTaskStatusSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListPreparationAndReleaseTaskStatusResponder handles the response to the ListPreparationAndReleaseTaskStatus request. The method always +// closes the http.Response Body. +func (client JobClient) ListPreparationAndReleaseTaskStatusResponder(resp *http.Response) (result CloudJobListPreparationAndReleaseTaskStatusResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listPreparationAndReleaseTaskStatusNextResults retrieves the next set of results, if any. +func (client JobClient) listPreparationAndReleaseTaskStatusNextResults(lastResults CloudJobListPreparationAndReleaseTaskStatusResult) (result CloudJobListPreparationAndReleaseTaskStatusResult, err error) { + req, err := lastResults.cloudJobListPreparationAndReleaseTaskStatusResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "batch.JobClient", "listPreparationAndReleaseTaskStatusNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListPreparationAndReleaseTaskStatusSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "batch.JobClient", "listPreparationAndReleaseTaskStatusNextResults", resp, "Failure sending next results request") + } + result, err = client.ListPreparationAndReleaseTaskStatusResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "listPreparationAndReleaseTaskStatusNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListPreparationAndReleaseTaskStatusComplete enumerates all values, automatically crossing page boundaries as required. +func (client JobClient) ListPreparationAndReleaseTaskStatusComplete(ctx context.Context, jobID string, filter string, selectParameter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CloudJobListPreparationAndReleaseTaskStatusResultIterator, err error) { + result.page, err = client.ListPreparationAndReleaseTaskStatus(ctx, jobID, filter, selectParameter, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + return +} + +// Patch this replaces only the job properties specified in the request. For example, if the job has constraints, and a +// request does not specify the constraints element, then the job keeps the existing constraints. +// +// jobID is the ID of the job whose properties you want to update. jobPatchParameter is the parameters for the +// request. timeout is the maximum time that the server can spend processing the request, in seconds. The default +// is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a GUID with no +// decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the +// server should return the client-request-id in the response. ocpDate is the time the request was issued. Client +// libraries typically set this to the current system clock time; set it explicitly if you are calling the REST API +// directly. ifMatch is an ETag value associated with the version of the resource known to the client. The +// operation will be performed only if the resource's current ETag on the service exactly matches the value +// specified by the client. ifNoneMatch is an ETag value associated with the version of the resource known to the +// client. The operation will be performed only if the resource's current ETag on the service does not match the +// value specified by the client. ifModifiedSince is a timestamp indicating the last modified time of the resource +// known to the client. The operation will be performed only if the resource on the service has been modified since +// the specified time. ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to +// the client. The operation will be performed only if the resource on the service has not been modified since the +// specified time. +func (client JobClient) Patch(ctx context.Context, jobID string, jobPatchParameter JobPatchParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.PatchPreparer(ctx, jobID, jobPatchParameter, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Patch", nil, "Failure preparing request") + return + } + + resp, err := client.PatchSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobClient", "Patch", resp, "Failure sending request") + return + } + + result, err = client.PatchResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Patch", resp, "Failure responding to request") + } + + return +} + +// PatchPreparer prepares the Patch request. +func (client JobClient) PatchPreparer(ctx context.Context, jobID string, jobPatchParameter JobPatchParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPatch(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}", pathParameters), + autorest.WithJSON(jobPatchParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// PatchSender sends the Patch request. The method will close the +// http.Response Body if it receives an error. +func (client JobClient) PatchSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// PatchResponder handles the response to the Patch request. The method always +// closes the http.Response Body. +func (client JobClient) PatchResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} + +// Terminate when a Terminate Job request is received, the Batch service sets the job to the terminating state. The +// Batch service then terminates any active or running tasks associated with the job, and runs any required Job Release +// tasks. The job then moves into the completed state. +// +// jobID is the ID of the job to terminate. jobTerminateParameter is the parameters for the request. timeout is the +// maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client JobClient) Terminate(ctx context.Context, jobID string, jobTerminateParameter *JobTerminateParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.TerminatePreparer(ctx, jobID, jobTerminateParameter, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Terminate", nil, "Failure preparing request") + return + } + + resp, err := client.TerminateSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobClient", "Terminate", resp, "Failure sending request") + return + } + + result, err = client.TerminateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Terminate", resp, "Failure responding to request") + } + + return +} + +// TerminatePreparer prepares the Terminate request. +func (client JobClient) TerminatePreparer(ctx context.Context, jobID string, jobTerminateParameter *JobTerminateParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/terminate", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if jobTerminateParameter != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithJSON(jobTerminateParameter)) + } + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// TerminateSender sends the Terminate request. The method will close the +// http.Response Body if it receives an error. +func (client JobClient) TerminateSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// TerminateResponder handles the response to the Terminate request. The method always +// closes the http.Response Body. +func (client JobClient) TerminateResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// Update this fully replaces all the updateable properties of the job. For example, if the job has constraints +// associated with it and if constraints is not specified with this request, then the Batch service will remove the +// existing constraints. +// +// jobID is the ID of the job whose properties you want to update. jobUpdateParameter is the parameters for the +// request. timeout is the maximum time that the server can spend processing the request, in seconds. The default +// is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a GUID with no +// decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the +// server should return the client-request-id in the response. ocpDate is the time the request was issued. Client +// libraries typically set this to the current system clock time; set it explicitly if you are calling the REST API +// directly. ifMatch is an ETag value associated with the version of the resource known to the client. The +// operation will be performed only if the resource's current ETag on the service exactly matches the value +// specified by the client. ifNoneMatch is an ETag value associated with the version of the resource known to the +// client. The operation will be performed only if the resource's current ETag on the service does not match the +// value specified by the client. ifModifiedSince is a timestamp indicating the last modified time of the resource +// known to the client. The operation will be performed only if the resource on the service has been modified since +// the specified time. ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to +// the client. The operation will be performed only if the resource on the service has not been modified since the +// specified time. +func (client JobClient) Update(ctx context.Context, jobID string, jobUpdateParameter JobUpdateParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: jobUpdateParameter, + Constraints: []validation.Constraint{{Target: "jobUpdateParameter.PoolInfo", Name: validation.Null, Rule: true, + Chain: []validation.Constraint{{Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.VMSize", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.CloudServiceConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.CloudServiceConfiguration.OsFamily", Name: validation.Null, Rule: true, Chain: nil}}}, + {Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.ImageReference", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.NodeAgentSKUID", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.ContainerConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.ContainerConfiguration.Type", Name: validation.Null, Rule: true, Chain: nil}}}, + }}, + {Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.NetworkConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.NetworkConfiguration.EndpointConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.NetworkConfiguration.EndpointConfiguration.InboundNATPools", Name: validation.Null, Rule: true, Chain: nil}}}, + }}, + {Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.StartTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.StartTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobUpdateParameter.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + }}, + }}, + }}}}}); err != nil { + return result, validation.NewError("batch.JobClient", "Update", err.Error()) + } + + req, err := client.UpdatePreparer(ctx, jobID, jobUpdateParameter, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Update", nil, "Failure preparing request") + return + } + + resp, err := client.UpdateSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobClient", "Update", resp, "Failure sending request") + return + } + + result, err = client.UpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobClient", "Update", resp, "Failure responding to request") + } + + return +} + +// UpdatePreparer prepares the Update request. +func (client JobClient) UpdatePreparer(ctx context.Context, jobID string, jobUpdateParameter JobUpdateParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}", pathParameters), + autorest.WithJSON(jobUpdateParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// UpdateSender sends the Update request. The method will close the +// http.Response Body if it receives an error. +func (client JobClient) UpdateSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// UpdateResponder handles the response to the Update request. The method always +// closes the http.Response Body. +func (client JobClient) UpdateResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/jobschedule.go b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/jobschedule.go new file mode 100644 index 000000000..f6e83fb33 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/jobschedule.go @@ -0,0 +1,1324 @@ +package batch + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "github.com/Azure/go-autorest/autorest/date" + "github.com/Azure/go-autorest/autorest/validation" + "github.com/satori/go.uuid" + "net/http" +) + +// JobScheduleClient is the a client for issuing REST requests to the Azure Batch service. +type JobScheduleClient struct { + BaseClient +} + +// NewJobScheduleClient creates an instance of the JobScheduleClient client. +func NewJobScheduleClient() JobScheduleClient { + return NewJobScheduleClientWithBaseURI(DefaultBaseURI) +} + +// NewJobScheduleClientWithBaseURI creates an instance of the JobScheduleClient client. +func NewJobScheduleClientWithBaseURI(baseURI string) JobScheduleClient { + return JobScheduleClient{NewWithBaseURI(baseURI)} +} + +// Add sends the add request. +// +// cloudJobSchedule is the job schedule to be added. timeout is the maximum time that the server can spend +// processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request +// identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client JobScheduleClient) Add(ctx context.Context, cloudJobSchedule JobScheduleAddParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: cloudJobSchedule, + Constraints: []validation.Constraint{{Target: "cloudJobSchedule.ID", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.Schedule", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification", Name: validation.Null, Rule: true, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.JobManagerTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.JobManagerTask.ID", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.JobManagerTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.JobManagerTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.JobManagerTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.JobManagerTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.JobManagerTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.JobManagerTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + {Target: "cloudJobSchedule.JobSpecification.JobPreparationTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.JobPreparationTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.JobPreparationTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.JobPreparationTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.JobPreparationTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.JobPreparationTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.JobPreparationTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + {Target: "cloudJobSchedule.JobSpecification.JobReleaseTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.JobReleaseTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.JobReleaseTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.JobReleaseTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.JobReleaseTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.JobReleaseTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.JobReleaseTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + {Target: "cloudJobSchedule.JobSpecification.PoolInfo", Name: validation.Null, Rule: true, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.VMSize", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.CloudServiceConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.CloudServiceConfiguration.OsFamily", Name: validation.Null, Rule: true, Chain: nil}}}, + {Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.ImageReference", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.NodeAgentSKUID", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.ContainerConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.ContainerConfiguration.Type", Name: validation.Null, Rule: true, Chain: nil}}}, + }}, + {Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.NetworkConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.NetworkConfiguration.EndpointConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.NetworkConfiguration.EndpointConfiguration.InboundNATPools", Name: validation.Null, Rule: true, Chain: nil}}}, + }}, + {Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "cloudJobSchedule.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + }}, + }}, + }}, + }}}}}); err != nil { + return result, validation.NewError("batch.JobScheduleClient", "Add", err.Error()) + } + + req, err := client.AddPreparer(ctx, cloudJobSchedule, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Add", nil, "Failure preparing request") + return + } + + resp, err := client.AddSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Add", resp, "Failure sending request") + return + } + + result, err = client.AddResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Add", resp, "Failure responding to request") + } + + return +} + +// AddPreparer prepares the Add request. +func (client JobScheduleClient) AddPreparer(ctx context.Context, cloudJobSchedule JobScheduleAddParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPath("/jobschedules"), + autorest.WithJSON(cloudJobSchedule), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// AddSender sends the Add request. The method will close the +// http.Response Body if it receives an error. +func (client JobScheduleClient) AddSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// AddResponder handles the response to the Add request. The method always +// closes the http.Response Body. +func (client JobScheduleClient) AddResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByClosing()) + result.Response = resp + return +} + +// Delete when you delete a job schedule, this also deletes all jobs and tasks under that schedule. When tasks are +// deleted, all the files in their working directories on the compute nodes are also deleted (the retention period is +// ignored). The job schedule statistics are no longer accessible once the job schedule is deleted, though they are +// still counted towards account lifetime statistics. +// +// jobScheduleID is the ID of the job schedule to delete. timeout is the maximum time that the server can spend +// processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request +// identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client JobScheduleClient) Delete(ctx context.Context, jobScheduleID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.DeletePreparer(ctx, jobScheduleID, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Delete", nil, "Failure preparing request") + return + } + + resp, err := client.DeleteSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Delete", resp, "Failure sending request") + return + } + + result, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Delete", resp, "Failure responding to request") + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client JobScheduleClient) DeletePreparer(ctx context.Context, jobScheduleID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobScheduleId": autorest.Encode("path", jobScheduleID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobschedules/{jobScheduleId}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client JobScheduleClient) DeleteSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client JobScheduleClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// Disable no new jobs will be created until the job schedule is enabled again. +// +// jobScheduleID is the ID of the job schedule to disable. timeout is the maximum time that the server can spend +// processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request +// identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client JobScheduleClient) Disable(ctx context.Context, jobScheduleID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.DisablePreparer(ctx, jobScheduleID, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Disable", nil, "Failure preparing request") + return + } + + resp, err := client.DisableSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Disable", resp, "Failure sending request") + return + } + + result, err = client.DisableResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Disable", resp, "Failure responding to request") + } + + return +} + +// DisablePreparer prepares the Disable request. +func (client JobScheduleClient) DisablePreparer(ctx context.Context, jobScheduleID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobScheduleId": autorest.Encode("path", jobScheduleID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobschedules/{jobScheduleId}/disable", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DisableSender sends the Disable request. The method will close the +// http.Response Body if it receives an error. +func (client JobScheduleClient) DisableSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// DisableResponder handles the response to the Disable request. The method always +// closes the http.Response Body. +func (client JobScheduleClient) DisableResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Enable sends the enable request. +// +// jobScheduleID is the ID of the job schedule to enable. timeout is the maximum time that the server can spend +// processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request +// identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client JobScheduleClient) Enable(ctx context.Context, jobScheduleID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.EnablePreparer(ctx, jobScheduleID, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Enable", nil, "Failure preparing request") + return + } + + resp, err := client.EnableSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Enable", resp, "Failure sending request") + return + } + + result, err = client.EnableResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Enable", resp, "Failure responding to request") + } + + return +} + +// EnablePreparer prepares the Enable request. +func (client JobScheduleClient) EnablePreparer(ctx context.Context, jobScheduleID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobScheduleId": autorest.Encode("path", jobScheduleID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobschedules/{jobScheduleId}/enable", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// EnableSender sends the Enable request. The method will close the +// http.Response Body if it receives an error. +func (client JobScheduleClient) EnableSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// EnableResponder handles the response to the Enable request. The method always +// closes the http.Response Body. +func (client JobScheduleClient) EnableResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Exists sends the exists request. +// +// jobScheduleID is the ID of the job schedule which you want to check. timeout is the maximum time that the server +// can spend processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated +// request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client JobScheduleClient) Exists(ctx context.Context, jobScheduleID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.ExistsPreparer(ctx, jobScheduleID, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Exists", nil, "Failure preparing request") + return + } + + resp, err := client.ExistsSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Exists", resp, "Failure sending request") + return + } + + result, err = client.ExistsResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Exists", resp, "Failure responding to request") + } + + return +} + +// ExistsPreparer prepares the Exists request. +func (client JobScheduleClient) ExistsPreparer(ctx context.Context, jobScheduleID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobScheduleId": autorest.Encode("path", jobScheduleID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsHead(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobschedules/{jobScheduleId}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ExistsSender sends the Exists request. The method will close the +// http.Response Body if it receives an error. +func (client JobScheduleClient) ExistsSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ExistsResponder handles the response to the Exists request. The method always +// closes the http.Response Body. +func (client JobScheduleClient) ExistsResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusNotFound), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets information about the specified job schedule. +// +// jobScheduleID is the ID of the job schedule to get. selectParameter is an OData $select clause. expand is an +// OData $expand clause. timeout is the maximum time that the server can spend processing the request, in seconds. +// The default is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a GUID with +// no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether +// the server should return the client-request-id in the response. ocpDate is the time the request was issued. +// Client libraries typically set this to the current system clock time; set it explicitly if you are calling the +// REST API directly. ifMatch is an ETag value associated with the version of the resource known to the client. The +// operation will be performed only if the resource's current ETag on the service exactly matches the value +// specified by the client. ifNoneMatch is an ETag value associated with the version of the resource known to the +// client. The operation will be performed only if the resource's current ETag on the service does not match the +// value specified by the client. ifModifiedSince is a timestamp indicating the last modified time of the resource +// known to the client. The operation will be performed only if the resource on the service has been modified since +// the specified time. ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to +// the client. The operation will be performed only if the resource on the service has not been modified since the +// specified time. +func (client JobScheduleClient) Get(ctx context.Context, jobScheduleID string, selectParameter string, expand string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result CloudJobSchedule, err error) { + req, err := client.GetPreparer(ctx, jobScheduleID, selectParameter, expand, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client JobScheduleClient) GetPreparer(ctx context.Context, jobScheduleID string, selectParameter string, expand string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobScheduleId": autorest.Encode("path", jobScheduleID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobschedules/{jobScheduleId}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client JobScheduleClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client JobScheduleClient) GetResponder(resp *http.Response) (result CloudJobSchedule, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List sends the list request. +// +// filter is an OData $filter clause. For more information on constructing this filter, see +// https://docs.microsoft.com/en-us/rest/api/batchservice/odata-filters-in-batch#list-job-schedules. +// selectParameter is an OData $select clause. expand is an OData $expand clause. maxResults is the maximum number +// of items to return in the response. A maximum of 1000 job schedules can be returned. timeout is the maximum time +// that the server can spend processing the request, in seconds. The default is 30 seconds. clientRequestID is the +// caller-generated request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client JobScheduleClient) List(ctx context.Context, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CloudJobScheduleListResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: maxResults, + Constraints: []validation.Constraint{{Target: "maxResults", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "maxResults", Name: validation.InclusiveMaximum, Rule: 1000, Chain: nil}, + {Target: "maxResults", Name: validation.InclusiveMinimum, Rule: 1, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("batch.JobScheduleClient", "List", err.Error()) + } + + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, filter, selectParameter, expand, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.cjslr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "List", resp, "Failure sending request") + return + } + + result.cjslr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client JobScheduleClient) ListPreparer(ctx context.Context, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(filter) > 0 { + queryParameters["$filter"] = autorest.Encode("query", filter) + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + if maxResults != nil { + queryParameters["maxresults"] = autorest.Encode("query", *maxResults) + } else { + queryParameters["maxresults"] = autorest.Encode("query", 1000) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPath("/jobschedules"), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client JobScheduleClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client JobScheduleClient) ListResponder(resp *http.Response) (result CloudJobScheduleListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client JobScheduleClient) listNextResults(lastResults CloudJobScheduleListResult) (result CloudJobScheduleListResult, err error) { + req, err := lastResults.cloudJobScheduleListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "batch.JobScheduleClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "batch.JobScheduleClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client JobScheduleClient) ListComplete(ctx context.Context, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CloudJobScheduleListResultIterator, err error) { + result.page, err = client.List(ctx, filter, selectParameter, expand, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + return +} + +// Patch this replaces only the job schedule properties specified in the request. For example, if the schedule property +// is not specified with this request, then the Batch service will keep the existing schedule. Changes to a job +// schedule only impact jobs created by the schedule after the update has taken place; currently running jobs are +// unaffected. +// +// jobScheduleID is the ID of the job schedule to update. jobSchedulePatchParameter is the parameters for the +// request. timeout is the maximum time that the server can spend processing the request, in seconds. The default +// is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a GUID with no +// decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the +// server should return the client-request-id in the response. ocpDate is the time the request was issued. Client +// libraries typically set this to the current system clock time; set it explicitly if you are calling the REST API +// directly. ifMatch is an ETag value associated with the version of the resource known to the client. The +// operation will be performed only if the resource's current ETag on the service exactly matches the value +// specified by the client. ifNoneMatch is an ETag value associated with the version of the resource known to the +// client. The operation will be performed only if the resource's current ETag on the service does not match the +// value specified by the client. ifModifiedSince is a timestamp indicating the last modified time of the resource +// known to the client. The operation will be performed only if the resource on the service has been modified since +// the specified time. ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to +// the client. The operation will be performed only if the resource on the service has not been modified since the +// specified time. +func (client JobScheduleClient) Patch(ctx context.Context, jobScheduleID string, jobSchedulePatchParameter JobSchedulePatchParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.PatchPreparer(ctx, jobScheduleID, jobSchedulePatchParameter, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Patch", nil, "Failure preparing request") + return + } + + resp, err := client.PatchSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Patch", resp, "Failure sending request") + return + } + + result, err = client.PatchResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Patch", resp, "Failure responding to request") + } + + return +} + +// PatchPreparer prepares the Patch request. +func (client JobScheduleClient) PatchPreparer(ctx context.Context, jobScheduleID string, jobSchedulePatchParameter JobSchedulePatchParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobScheduleId": autorest.Encode("path", jobScheduleID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPatch(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobschedules/{jobScheduleId}", pathParameters), + autorest.WithJSON(jobSchedulePatchParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// PatchSender sends the Patch request. The method will close the +// http.Response Body if it receives an error. +func (client JobScheduleClient) PatchSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// PatchResponder handles the response to the Patch request. The method always +// closes the http.Response Body. +func (client JobScheduleClient) PatchResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} + +// Terminate sends the terminate request. +// +// jobScheduleID is the ID of the job schedule to terminates. timeout is the maximum time that the server can spend +// processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request +// identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client JobScheduleClient) Terminate(ctx context.Context, jobScheduleID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.TerminatePreparer(ctx, jobScheduleID, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Terminate", nil, "Failure preparing request") + return + } + + resp, err := client.TerminateSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Terminate", resp, "Failure sending request") + return + } + + result, err = client.TerminateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Terminate", resp, "Failure responding to request") + } + + return +} + +// TerminatePreparer prepares the Terminate request. +func (client JobScheduleClient) TerminatePreparer(ctx context.Context, jobScheduleID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobScheduleId": autorest.Encode("path", jobScheduleID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobschedules/{jobScheduleId}/terminate", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// TerminateSender sends the Terminate request. The method will close the +// http.Response Body if it receives an error. +func (client JobScheduleClient) TerminateSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// TerminateResponder handles the response to the Terminate request. The method always +// closes the http.Response Body. +func (client JobScheduleClient) TerminateResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// Update this fully replaces all the updateable properties of the job schedule. For example, if the schedule property +// is not specified with this request, then the Batch service will remove the existing schedule. Changes to a job +// schedule only impact jobs created by the schedule after the update has taken place; currently running jobs are +// unaffected. +// +// jobScheduleID is the ID of the job schedule to update. jobScheduleUpdateParameter is the parameters for the +// request. timeout is the maximum time that the server can spend processing the request, in seconds. The default +// is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a GUID with no +// decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the +// server should return the client-request-id in the response. ocpDate is the time the request was issued. Client +// libraries typically set this to the current system clock time; set it explicitly if you are calling the REST API +// directly. ifMatch is an ETag value associated with the version of the resource known to the client. The +// operation will be performed only if the resource's current ETag on the service exactly matches the value +// specified by the client. ifNoneMatch is an ETag value associated with the version of the resource known to the +// client. The operation will be performed only if the resource's current ETag on the service does not match the +// value specified by the client. ifModifiedSince is a timestamp indicating the last modified time of the resource +// known to the client. The operation will be performed only if the resource on the service has been modified since +// the specified time. ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to +// the client. The operation will be performed only if the resource on the service has not been modified since the +// specified time. +func (client JobScheduleClient) Update(ctx context.Context, jobScheduleID string, jobScheduleUpdateParameter JobScheduleUpdateParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: jobScheduleUpdateParameter, + Constraints: []validation.Constraint{{Target: "jobScheduleUpdateParameter.Schedule", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification", Name: validation.Null, Rule: true, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.JobManagerTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.JobManagerTask.ID", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.JobManagerTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.JobManagerTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.JobManagerTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.JobManagerTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.JobManagerTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.JobManagerTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + {Target: "jobScheduleUpdateParameter.JobSpecification.JobPreparationTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.JobPreparationTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.JobPreparationTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.JobPreparationTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.JobPreparationTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.JobPreparationTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.JobPreparationTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + {Target: "jobScheduleUpdateParameter.JobSpecification.JobReleaseTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.JobReleaseTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.JobReleaseTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.JobReleaseTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.JobReleaseTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.JobReleaseTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.JobReleaseTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + {Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo", Name: validation.Null, Rule: true, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.VMSize", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.CloudServiceConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.CloudServiceConfiguration.OsFamily", Name: validation.Null, Rule: true, Chain: nil}}}, + {Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.ImageReference", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.NodeAgentSKUID", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.ContainerConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.VirtualMachineConfiguration.ContainerConfiguration.Type", Name: validation.Null, Rule: true, Chain: nil}}}, + }}, + {Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.NetworkConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.NetworkConfiguration.EndpointConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.NetworkConfiguration.EndpointConfiguration.InboundNATPools", Name: validation.Null, Rule: true, Chain: nil}}}, + }}, + {Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "jobScheduleUpdateParameter.JobSpecification.PoolInfo.AutoPoolSpecification.Pool.StartTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + }}, + }}, + }}, + }}}}}); err != nil { + return result, validation.NewError("batch.JobScheduleClient", "Update", err.Error()) + } + + req, err := client.UpdatePreparer(ctx, jobScheduleID, jobScheduleUpdateParameter, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Update", nil, "Failure preparing request") + return + } + + resp, err := client.UpdateSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Update", resp, "Failure sending request") + return + } + + result, err = client.UpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.JobScheduleClient", "Update", resp, "Failure responding to request") + } + + return +} + +// UpdatePreparer prepares the Update request. +func (client JobScheduleClient) UpdatePreparer(ctx context.Context, jobScheduleID string, jobScheduleUpdateParameter JobScheduleUpdateParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobScheduleId": autorest.Encode("path", jobScheduleID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobschedules/{jobScheduleId}", pathParameters), + autorest.WithJSON(jobScheduleUpdateParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// UpdateSender sends the Update request. The method will close the +// http.Response Body if it receives an error. +func (client JobScheduleClient) UpdateSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// UpdateResponder handles the response to the Update request. The method always +// closes the http.Response Body. +func (client JobScheduleClient) UpdateResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/models.go b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/models.go new file mode 100644 index 000000000..d75557cf0 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/models.go @@ -0,0 +1,3418 @@ +package batch + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/date" + "github.com/Azure/go-autorest/autorest/to" + "io" + "net/http" +) + +// AccessScope enumerates the values for access scope. +type AccessScope string + +const ( + // Job Grants access to perform all operations on the job containing the task. + Job AccessScope = "job" +) + +// PossibleAccessScopeValues returns an array of possible values for the AccessScope const type. +func PossibleAccessScopeValues() []AccessScope { + return []AccessScope{Job} +} + +// AllocationState enumerates the values for allocation state. +type AllocationState string + +const ( + // Resizing The pool is resizing; that is, compute nodes are being added to or removed from the pool. + Resizing AllocationState = "resizing" + // Steady The pool is not resizing. There are no changes to the number of nodes in the pool in progress. A + // pool enters this state when it is created and when no operations are being performed on the pool to + // change the number of nodes. + Steady AllocationState = "steady" + // Stopping The pool was resizing, but the user has requested that the resize be stopped, but the stop + // request has not yet been completed. + Stopping AllocationState = "stopping" +) + +// PossibleAllocationStateValues returns an array of possible values for the AllocationState const type. +func PossibleAllocationStateValues() []AllocationState { + return []AllocationState{Resizing, Steady, Stopping} +} + +// AutoUserScope enumerates the values for auto user scope. +type AutoUserScope string + +const ( + // Pool Specifies that the task runs as the common auto user account which is created on every node in a + // pool. + Pool AutoUserScope = "pool" + // Task Specifies that the service should create a new user for the task. + Task AutoUserScope = "task" +) + +// PossibleAutoUserScopeValues returns an array of possible values for the AutoUserScope const type. +func PossibleAutoUserScopeValues() []AutoUserScope { + return []AutoUserScope{Pool, Task} +} + +// CachingType enumerates the values for caching type. +type CachingType string + +const ( + // None The caching mode for the disk is not enabled. + None CachingType = "none" + // ReadOnly The caching mode for the disk is read only. + ReadOnly CachingType = "readonly" + // ReadWrite The caching mode for the disk is read and write. + ReadWrite CachingType = "readwrite" +) + +// PossibleCachingTypeValues returns an array of possible values for the CachingType const type. +func PossibleCachingTypeValues() []CachingType { + return []CachingType{None, ReadOnly, ReadWrite} +} + +// CertificateFormat enumerates the values for certificate format. +type CertificateFormat string + +const ( + // Cer The certificate is a base64-encoded X.509 certificate. + Cer CertificateFormat = "cer" + // Pfx The certificate is a PFX (PKCS#12) formatted certificate or certificate chain. + Pfx CertificateFormat = "pfx" +) + +// PossibleCertificateFormatValues returns an array of possible values for the CertificateFormat const type. +func PossibleCertificateFormatValues() []CertificateFormat { + return []CertificateFormat{Cer, Pfx} +} + +// CertificateState enumerates the values for certificate state. +type CertificateState string + +const ( + // Active The certificate is available for use in pools. + Active CertificateState = "active" + // DeleteFailed The user requested that the certificate be deleted, but there are pools that still have + // references to the certificate, or it is still installed on one or more compute nodes. (The latter can + // occur if the certificate has been removed from the pool, but the node has not yet restarted. Nodes + // refresh their certificates only when they restart.) You may use the cancel certificate delete operation + // to cancel the delete, or the delete certificate operation to retry the delete. + DeleteFailed CertificateState = "deletefailed" + // Deleting The user has requested that the certificate be deleted, but the delete operation has not yet + // completed. You may not reference the certificate when creating or updating pools. + Deleting CertificateState = "deleting" +) + +// PossibleCertificateStateValues returns an array of possible values for the CertificateState const type. +func PossibleCertificateStateValues() []CertificateState { + return []CertificateState{Active, DeleteFailed, Deleting} +} + +// CertificateStoreLocation enumerates the values for certificate store location. +type CertificateStoreLocation string + +const ( + // CurrentUser Certificates should be installed to the CurrentUser certificate store. + CurrentUser CertificateStoreLocation = "currentuser" + // LocalMachine Certificates should be installed to the LocalMachine certificate store. + LocalMachine CertificateStoreLocation = "localmachine" +) + +// PossibleCertificateStoreLocationValues returns an array of possible values for the CertificateStoreLocation const type. +func PossibleCertificateStoreLocationValues() []CertificateStoreLocation { + return []CertificateStoreLocation{CurrentUser, LocalMachine} +} + +// CertificateVisibility enumerates the values for certificate visibility. +type CertificateVisibility string + +const ( + // CertificateVisibilityRemoteUser The certificate should be visibile to the user accounts under which + // users remotely access the node. + CertificateVisibilityRemoteUser CertificateVisibility = "remoteuser" + // CertificateVisibilityStartTask The certificate should be visible to the user account under which the + // start task is run. + CertificateVisibilityStartTask CertificateVisibility = "starttask" + // CertificateVisibilityTask The certificate should be visibile to the user accounts under which job tasks + // are run. + CertificateVisibilityTask CertificateVisibility = "task" +) + +// PossibleCertificateVisibilityValues returns an array of possible values for the CertificateVisibility const type. +func PossibleCertificateVisibilityValues() []CertificateVisibility { + return []CertificateVisibility{CertificateVisibilityRemoteUser, CertificateVisibilityStartTask, CertificateVisibilityTask} +} + +// ComputeNodeDeallocationOption enumerates the values for compute node deallocation option. +type ComputeNodeDeallocationOption string + +const ( + // Requeue Terminate running task processes and requeue the tasks. The tasks will run again when a node is + // available. Remove nodes as soon as tasks have been terminated. + Requeue ComputeNodeDeallocationOption = "requeue" + // RetainedData Allow currently running tasks to complete, then wait for all task data retention periods to + // expire. Schedule no new tasks while waiting. Remove nodes when all task retention periods have expired. + RetainedData ComputeNodeDeallocationOption = "retaineddata" + // TaskCompletion Allow currently running tasks to complete. Schedule no new tasks while waiting. Remove + // nodes when all tasks have completed. + TaskCompletion ComputeNodeDeallocationOption = "taskcompletion" + // Terminate Terminate running tasks. The tasks will be completed with failureInfo indicating that they + // were terminated, and will not run again. Remove nodes as soon as tasks have been terminated. + Terminate ComputeNodeDeallocationOption = "terminate" +) + +// PossibleComputeNodeDeallocationOptionValues returns an array of possible values for the ComputeNodeDeallocationOption const type. +func PossibleComputeNodeDeallocationOptionValues() []ComputeNodeDeallocationOption { + return []ComputeNodeDeallocationOption{Requeue, RetainedData, TaskCompletion, Terminate} +} + +// ComputeNodeFillType enumerates the values for compute node fill type. +type ComputeNodeFillType string + +const ( + // Pack As many tasks as possible (maxTasksPerNode) should be assigned to each node in the pool before any + // tasks are assigned to the next node in the pool. + Pack ComputeNodeFillType = "pack" + // Spread Tasks should be assigned evenly across all nodes in the pool. + Spread ComputeNodeFillType = "spread" +) + +// PossibleComputeNodeFillTypeValues returns an array of possible values for the ComputeNodeFillType const type. +func PossibleComputeNodeFillTypeValues() []ComputeNodeFillType { + return []ComputeNodeFillType{Pack, Spread} +} + +// ComputeNodeRebootOption enumerates the values for compute node reboot option. +type ComputeNodeRebootOption string + +const ( + // ComputeNodeRebootOptionRequeue Terminate running task processes and requeue the tasks. The tasks will + // run again when a node is available. Restart the node as soon as tasks have been terminated. + ComputeNodeRebootOptionRequeue ComputeNodeRebootOption = "requeue" + // ComputeNodeRebootOptionRetainedData Allow currently running tasks to complete, then wait for all task + // data retention periods to expire. Schedule no new tasks while waiting. Restart the node when all task + // retention periods have expired. + ComputeNodeRebootOptionRetainedData ComputeNodeRebootOption = "retaineddata" + // ComputeNodeRebootOptionTaskCompletion Allow currently running tasks to complete. Schedule no new tasks + // while waiting. Restart the node when all tasks have completed. + ComputeNodeRebootOptionTaskCompletion ComputeNodeRebootOption = "taskcompletion" + // ComputeNodeRebootOptionTerminate Terminate running tasks. The tasks will be completed with failureInfo + // indicating that they were terminated, and will not run again. Restart the node as soon as tasks have + // been terminated. + ComputeNodeRebootOptionTerminate ComputeNodeRebootOption = "terminate" +) + +// PossibleComputeNodeRebootOptionValues returns an array of possible values for the ComputeNodeRebootOption const type. +func PossibleComputeNodeRebootOptionValues() []ComputeNodeRebootOption { + return []ComputeNodeRebootOption{ComputeNodeRebootOptionRequeue, ComputeNodeRebootOptionRetainedData, ComputeNodeRebootOptionTaskCompletion, ComputeNodeRebootOptionTerminate} +} + +// ComputeNodeReimageOption enumerates the values for compute node reimage option. +type ComputeNodeReimageOption string + +const ( + // ComputeNodeReimageOptionRequeue Terminate running task processes and requeue the tasks. The tasks will + // run again when a node is available. Reimage the node as soon as tasks have been terminated. + ComputeNodeReimageOptionRequeue ComputeNodeReimageOption = "requeue" + // ComputeNodeReimageOptionRetainedData Allow currently running tasks to complete, then wait for all task + // data retention periods to expire. Schedule no new tasks while waiting. Reimage the node when all task + // retention periods have expired. + ComputeNodeReimageOptionRetainedData ComputeNodeReimageOption = "retaineddata" + // ComputeNodeReimageOptionTaskCompletion Allow currently running tasks to complete. Schedule no new tasks + // while waiting. Reimage the node when all tasks have completed. + ComputeNodeReimageOptionTaskCompletion ComputeNodeReimageOption = "taskcompletion" + // ComputeNodeReimageOptionTerminate Terminate running tasks. The tasks will be completed with failureInfo + // indicating that they were terminated, and will not run again. Reimage the node as soon as tasks have + // been terminated. + ComputeNodeReimageOptionTerminate ComputeNodeReimageOption = "terminate" +) + +// PossibleComputeNodeReimageOptionValues returns an array of possible values for the ComputeNodeReimageOption const type. +func PossibleComputeNodeReimageOptionValues() []ComputeNodeReimageOption { + return []ComputeNodeReimageOption{ComputeNodeReimageOptionRequeue, ComputeNodeReimageOptionRetainedData, ComputeNodeReimageOptionTaskCompletion, ComputeNodeReimageOptionTerminate} +} + +// ComputeNodeState enumerates the values for compute node state. +type ComputeNodeState string + +const ( + // Creating The Batch service has obtained the underlying virtual machine from Azure Compute, but it has + // not yet started to join the pool. + Creating ComputeNodeState = "creating" + // Idle The node is not currently running a task. + Idle ComputeNodeState = "idle" + // LeavingPool The node is leaving the pool, either because the user explicitly removed it or because the + // pool is resizing or autoscaling down. + LeavingPool ComputeNodeState = "leavingpool" + // Offline The node is not currently running a task, and scheduling of new tasks to the node is disabled. + Offline ComputeNodeState = "offline" + // Preempted The low-priority node has been preempted. Tasks which were running on the node when it was + // pre-empted will be rescheduled when another node becomes available. + Preempted ComputeNodeState = "preempted" + // Rebooting The node is rebooting. + Rebooting ComputeNodeState = "rebooting" + // Reimaging The node is reimaging. + Reimaging ComputeNodeState = "reimaging" + // Running The node is running one or more tasks (other than a start task). + Running ComputeNodeState = "running" + // Starting The Batch service is starting on the underlying virtual machine. + Starting ComputeNodeState = "starting" + // StartTaskFailed The start task has failed on the compute node (and exhausted all retries), and + // waitForSuccess is set. The node is not usable for running tasks. + StartTaskFailed ComputeNodeState = "starttaskfailed" + // Unknown The Batch service has lost contact with the node, and does not know its true state. + Unknown ComputeNodeState = "unknown" + // Unusable The node cannot be used for task execution due to errors. + Unusable ComputeNodeState = "unusable" + // WaitingForStartTask The start task has started running on the compute node, but waitForSuccess is set + // and the start task has not yet completed. + WaitingForStartTask ComputeNodeState = "waitingforstarttask" +) + +// PossibleComputeNodeStateValues returns an array of possible values for the ComputeNodeState const type. +func PossibleComputeNodeStateValues() []ComputeNodeState { + return []ComputeNodeState{Creating, Idle, LeavingPool, Offline, Preempted, Rebooting, Reimaging, Running, Starting, StartTaskFailed, Unknown, Unusable, WaitingForStartTask} +} + +// DependencyAction enumerates the values for dependency action. +type DependencyAction string + +const ( + // Block Block the task's dependencies. + Block DependencyAction = "block" + // Satisfy Satisfy the task's dependencies. + Satisfy DependencyAction = "satisfy" +) + +// PossibleDependencyActionValues returns an array of possible values for the DependencyAction const type. +func PossibleDependencyActionValues() []DependencyAction { + return []DependencyAction{Block, Satisfy} +} + +// DisableComputeNodeSchedulingOption enumerates the values for disable compute node scheduling option. +type DisableComputeNodeSchedulingOption string + +const ( + // DisableComputeNodeSchedulingOptionRequeue Terminate running task processes and requeue the tasks. The + // tasks may run again on other compute nodes, or when task scheduling is re-enabled on this node. Enter + // offline state as soon as tasks have been terminated. + DisableComputeNodeSchedulingOptionRequeue DisableComputeNodeSchedulingOption = "requeue" + // DisableComputeNodeSchedulingOptionTaskCompletion Allow currently running tasks to complete. Schedule no + // new tasks while waiting. Enter offline state when all tasks have completed. + DisableComputeNodeSchedulingOptionTaskCompletion DisableComputeNodeSchedulingOption = "taskcompletion" + // DisableComputeNodeSchedulingOptionTerminate Terminate running tasks. The tasks will be completed with + // failureInfo indicating that they were terminated, and will not run again. Enter offline state as soon as + // tasks have been terminated. + DisableComputeNodeSchedulingOptionTerminate DisableComputeNodeSchedulingOption = "terminate" +) + +// PossibleDisableComputeNodeSchedulingOptionValues returns an array of possible values for the DisableComputeNodeSchedulingOption const type. +func PossibleDisableComputeNodeSchedulingOptionValues() []DisableComputeNodeSchedulingOption { + return []DisableComputeNodeSchedulingOption{DisableComputeNodeSchedulingOptionRequeue, DisableComputeNodeSchedulingOptionTaskCompletion, DisableComputeNodeSchedulingOptionTerminate} +} + +// DisableJobOption enumerates the values for disable job option. +type DisableJobOption string + +const ( + // DisableJobOptionRequeue Terminate running tasks and requeue them. The tasks will run again when the job + // is enabled. + DisableJobOptionRequeue DisableJobOption = "requeue" + // DisableJobOptionTerminate Terminate running tasks. The tasks will be completed with failureInfo + // indicating that they were terminated, and will not run again. + DisableJobOptionTerminate DisableJobOption = "terminate" + // DisableJobOptionWait Allow currently running tasks to complete. + DisableJobOptionWait DisableJobOption = "wait" +) + +// PossibleDisableJobOptionValues returns an array of possible values for the DisableJobOption const type. +func PossibleDisableJobOptionValues() []DisableJobOption { + return []DisableJobOption{DisableJobOptionRequeue, DisableJobOptionTerminate, DisableJobOptionWait} +} + +// ElevationLevel enumerates the values for elevation level. +type ElevationLevel string + +const ( + // Admin The user is a user with elevated access and operates with full Administrator permissions. + Admin ElevationLevel = "admin" + // NonAdmin The user is a standard user without elevated access. + NonAdmin ElevationLevel = "nonadmin" +) + +// PossibleElevationLevelValues returns an array of possible values for the ElevationLevel const type. +func PossibleElevationLevelValues() []ElevationLevel { + return []ElevationLevel{Admin, NonAdmin} +} + +// ErrorCategory enumerates the values for error category. +type ErrorCategory string + +const ( + // ServerError The error is due to an internal server issue. + ServerError ErrorCategory = "servererror" + // UserError The error is due to a user issue, such as misconfiguration. + UserError ErrorCategory = "usererror" +) + +// PossibleErrorCategoryValues returns an array of possible values for the ErrorCategory const type. +func PossibleErrorCategoryValues() []ErrorCategory { + return []ErrorCategory{ServerError, UserError} +} + +// InboundEndpointProtocol enumerates the values for inbound endpoint protocol. +type InboundEndpointProtocol string + +const ( + // TCP Use TCP for the endpoint. + TCP InboundEndpointProtocol = "tcp" + // UDP Use UDP for the endpoint. + UDP InboundEndpointProtocol = "udp" +) + +// PossibleInboundEndpointProtocolValues returns an array of possible values for the InboundEndpointProtocol const type. +func PossibleInboundEndpointProtocolValues() []InboundEndpointProtocol { + return []InboundEndpointProtocol{TCP, UDP} +} + +// JobAction enumerates the values for job action. +type JobAction string + +const ( + // JobActionDisable Disable the job. This is equivalent to calling the disable job API, with a disableTasks + // value of requeue. + JobActionDisable JobAction = "disable" + // JobActionNone Take no action. + JobActionNone JobAction = "none" + // JobActionTerminate Terminate the job. The terminateReason in the job's executionInfo is set to + // "TaskFailed". + JobActionTerminate JobAction = "terminate" +) + +// PossibleJobActionValues returns an array of possible values for the JobAction const type. +func PossibleJobActionValues() []JobAction { + return []JobAction{JobActionDisable, JobActionNone, JobActionTerminate} +} + +// JobPreparationTaskState enumerates the values for job preparation task state. +type JobPreparationTaskState string + +const ( + // JobPreparationTaskStateCompleted The task has exited with exit code 0, or the task has exhausted its + // retry limit, or the Batch service was unable to start the task due to task preparation errors (such as + // resource file download failures). + JobPreparationTaskStateCompleted JobPreparationTaskState = "completed" + // JobPreparationTaskStateRunning The task is currently running (including retrying). + JobPreparationTaskStateRunning JobPreparationTaskState = "running" +) + +// PossibleJobPreparationTaskStateValues returns an array of possible values for the JobPreparationTaskState const type. +func PossibleJobPreparationTaskStateValues() []JobPreparationTaskState { + return []JobPreparationTaskState{JobPreparationTaskStateCompleted, JobPreparationTaskStateRunning} +} + +// JobReleaseTaskState enumerates the values for job release task state. +type JobReleaseTaskState string + +const ( + // JobReleaseTaskStateCompleted The task has exited with exit code 0, or the task has exhausted its retry + // limit, or the Batch service was unable to start the task due to task preparation errors (such as + // resource file download failures). + JobReleaseTaskStateCompleted JobReleaseTaskState = "completed" + // JobReleaseTaskStateRunning The task is currently running (including retrying). + JobReleaseTaskStateRunning JobReleaseTaskState = "running" +) + +// PossibleJobReleaseTaskStateValues returns an array of possible values for the JobReleaseTaskState const type. +func PossibleJobReleaseTaskStateValues() []JobReleaseTaskState { + return []JobReleaseTaskState{JobReleaseTaskStateCompleted, JobReleaseTaskStateRunning} +} + +// JobScheduleState enumerates the values for job schedule state. +type JobScheduleState string + +const ( + // JobScheduleStateActive The job schedule is active and will create jobs as per its schedule. + JobScheduleStateActive JobScheduleState = "active" + // JobScheduleStateCompleted The schedule has terminated, either by reaching its end time or by the user + // terminating it explicitly. + JobScheduleStateCompleted JobScheduleState = "completed" + // JobScheduleStateDeleting The user has requested that the schedule be deleted, but the delete operation + // is still in progress. The scheduler will not initiate any new jobs for this schedule, and will delete + // any existing jobs and tasks under the schedule, including any active job. The schedule will be deleted + // when all jobs and tasks under the schedule have been deleted. + JobScheduleStateDeleting JobScheduleState = "deleting" + // JobScheduleStateDisabled The user has disabled the schedule. The scheduler will not initiate any new + // jobs will on this schedule, but any existing active job will continue to run. + JobScheduleStateDisabled JobScheduleState = "disabled" + // JobScheduleStateTerminating The schedule has no more work to do, or has been explicitly terminated by + // the user, but the termination operation is still in progress. The scheduler will not initiate any new + // jobs for this schedule, nor is any existing job active. + JobScheduleStateTerminating JobScheduleState = "terminating" +) + +// PossibleJobScheduleStateValues returns an array of possible values for the JobScheduleState const type. +func PossibleJobScheduleStateValues() []JobScheduleState { + return []JobScheduleState{JobScheduleStateActive, JobScheduleStateCompleted, JobScheduleStateDeleting, JobScheduleStateDisabled, JobScheduleStateTerminating} +} + +// JobState enumerates the values for job state. +type JobState string + +const ( + // JobStateActive The job is available to have tasks scheduled. + JobStateActive JobState = "active" + // JobStateCompleted All tasks have terminated, and the system will not accept any more tasks or any + // further changes to the job. + JobStateCompleted JobState = "completed" + // JobStateDeleting A user has requested that the job be deleted, but the delete operation is still in + // progress (for example, because the system is still terminating running tasks). + JobStateDeleting JobState = "deleting" + // JobStateDisabled A user has disabled the job. No tasks are running, and no new tasks will be scheduled. + JobStateDisabled JobState = "disabled" + // JobStateDisabling A user has requested that the job be disabled, but the disable operation is still in + // progress (for example, waiting for tasks to terminate). + JobStateDisabling JobState = "disabling" + // JobStateEnabling A user has requested that the job be enabled, but the enable operation is still in + // progress. + JobStateEnabling JobState = "enabling" + // JobStateTerminating The job is about to complete, either because a Job Manager task has completed or + // because the user has terminated the job, but the terminate operation is still in progress (for example, + // because Job Release tasks are running). + JobStateTerminating JobState = "terminating" +) + +// PossibleJobStateValues returns an array of possible values for the JobState const type. +func PossibleJobStateValues() []JobState { + return []JobState{JobStateActive, JobStateCompleted, JobStateDeleting, JobStateDisabled, JobStateDisabling, JobStateEnabling, JobStateTerminating} +} + +// NetworkSecurityGroupRuleAccess enumerates the values for network security group rule access. +type NetworkSecurityGroupRuleAccess string + +const ( + // Allow Allow access. + Allow NetworkSecurityGroupRuleAccess = "allow" + // Deny Deny access. + Deny NetworkSecurityGroupRuleAccess = "deny" +) + +// PossibleNetworkSecurityGroupRuleAccessValues returns an array of possible values for the NetworkSecurityGroupRuleAccess const type. +func PossibleNetworkSecurityGroupRuleAccessValues() []NetworkSecurityGroupRuleAccess { + return []NetworkSecurityGroupRuleAccess{Allow, Deny} +} + +// OnAllTasksComplete enumerates the values for on all tasks complete. +type OnAllTasksComplete string + +const ( + // NoAction Do nothing. The job remains active unless terminated or disabled by some other means. + NoAction OnAllTasksComplete = "noaction" + // TerminateJob Terminate the job. The job's terminateReason is set to 'AllTasksComplete'. + TerminateJob OnAllTasksComplete = "terminatejob" +) + +// PossibleOnAllTasksCompleteValues returns an array of possible values for the OnAllTasksComplete const type. +func PossibleOnAllTasksCompleteValues() []OnAllTasksComplete { + return []OnAllTasksComplete{NoAction, TerminateJob} +} + +// OnTaskFailure enumerates the values for on task failure. +type OnTaskFailure string + +const ( + // OnTaskFailureNoAction Do nothing. The job remains active unless terminated or disabled by some other + // means. + OnTaskFailureNoAction OnTaskFailure = "noaction" + // OnTaskFailurePerformExitOptionsJobAction Take the action associated with the task exit condition in the + // task's exitConditions collection. (This may still result in no action being taken, if that is what the + // task specifies.) + OnTaskFailurePerformExitOptionsJobAction OnTaskFailure = "performexitoptionsjobaction" +) + +// PossibleOnTaskFailureValues returns an array of possible values for the OnTaskFailure const type. +func PossibleOnTaskFailureValues() []OnTaskFailure { + return []OnTaskFailure{OnTaskFailureNoAction, OnTaskFailurePerformExitOptionsJobAction} +} + +// OSType enumerates the values for os type. +type OSType string + +const ( + // Linux The Linux operating system. + Linux OSType = "linux" + // Windows The Windows operating system. + Windows OSType = "windows" +) + +// PossibleOSTypeValues returns an array of possible values for the OSType const type. +func PossibleOSTypeValues() []OSType { + return []OSType{Linux, Windows} +} + +// OutputFileUploadCondition enumerates the values for output file upload condition. +type OutputFileUploadCondition string + +const ( + // OutputFileUploadConditionTaskCompletion Upload the file(s) after the task process exits, no matter what + // the exit code was. + OutputFileUploadConditionTaskCompletion OutputFileUploadCondition = "taskcompletion" + // OutputFileUploadConditionTaskFailure Upload the file(s) only after the task process exits with a nonzero + // exit code. + OutputFileUploadConditionTaskFailure OutputFileUploadCondition = "taskfailure" + // OutputFileUploadConditionTaskSuccess Upload the file(s) only after the task process exits with an exit + // code of 0. + OutputFileUploadConditionTaskSuccess OutputFileUploadCondition = "tasksuccess" +) + +// PossibleOutputFileUploadConditionValues returns an array of possible values for the OutputFileUploadCondition const type. +func PossibleOutputFileUploadConditionValues() []OutputFileUploadCondition { + return []OutputFileUploadCondition{OutputFileUploadConditionTaskCompletion, OutputFileUploadConditionTaskFailure, OutputFileUploadConditionTaskSuccess} +} + +// PoolLifetimeOption enumerates the values for pool lifetime option. +type PoolLifetimeOption string + +const ( + // PoolLifetimeOptionJob The pool exists for the lifetime of the job to which it is dedicated. The Batch + // service creates the pool when it creates the job. If the 'job' option is applied to a job schedule, the + // Batch service creates a new auto pool for every job created on the schedule. + PoolLifetimeOptionJob PoolLifetimeOption = "job" + // PoolLifetimeOptionJobSchedule The pool exists for the lifetime of the job schedule. The Batch Service + // creates the pool when it creates the first job on the schedule. You may apply this option only to job + // schedules, not to jobs. + PoolLifetimeOptionJobSchedule PoolLifetimeOption = "jobschedule" +) + +// PossiblePoolLifetimeOptionValues returns an array of possible values for the PoolLifetimeOption const type. +func PossiblePoolLifetimeOptionValues() []PoolLifetimeOption { + return []PoolLifetimeOption{PoolLifetimeOptionJob, PoolLifetimeOptionJobSchedule} +} + +// PoolState enumerates the values for pool state. +type PoolState string + +const ( + // PoolStateActive The pool is available to run tasks subject to the availability of compute nodes. + PoolStateActive PoolState = "active" + // PoolStateDeleting The user has requested that the pool be deleted, but the delete operation has not yet + // completed. + PoolStateDeleting PoolState = "deleting" + // PoolStateUpgrading The user has requested that the operating system of the pool's nodes be upgraded, but + // the upgrade operation has not yet completed (that is, some nodes in the pool have not yet been + // upgraded). While upgrading, the pool may be able to run tasks (with reduced capacity) but this is not + // guaranteed. + PoolStateUpgrading PoolState = "upgrading" +) + +// PossiblePoolStateValues returns an array of possible values for the PoolState const type. +func PossiblePoolStateValues() []PoolState { + return []PoolState{PoolStateActive, PoolStateDeleting, PoolStateUpgrading} +} + +// SchedulingState enumerates the values for scheduling state. +type SchedulingState string + +const ( + // Disabled No new tasks will be scheduled on the node. Tasks already running on the node may still run to + // completion. All nodes start with scheduling enabled. + Disabled SchedulingState = "disabled" + // Enabled Tasks can be scheduled on the node. + Enabled SchedulingState = "enabled" +) + +// PossibleSchedulingStateValues returns an array of possible values for the SchedulingState const type. +func PossibleSchedulingStateValues() []SchedulingState { + return []SchedulingState{Disabled, Enabled} +} + +// StartTaskState enumerates the values for start task state. +type StartTaskState string + +const ( + // StartTaskStateCompleted The start task has exited with exit code 0, or the start task has failed and the + // retry limit has reached, or the start task process did not run due to task preparation errors (such as + // resource file download failures). + StartTaskStateCompleted StartTaskState = "completed" + // StartTaskStateRunning The start task is currently running. + StartTaskStateRunning StartTaskState = "running" +) + +// PossibleStartTaskStateValues returns an array of possible values for the StartTaskState const type. +func PossibleStartTaskStateValues() []StartTaskState { + return []StartTaskState{StartTaskStateCompleted, StartTaskStateRunning} +} + +// StorageAccountType enumerates the values for storage account type. +type StorageAccountType string + +const ( + // PremiumLRS The data disk should use premium locally redundant storage. + PremiumLRS StorageAccountType = "premium_lrs" + // StandardLRS The data disk should use standard locally redundant storage. + StandardLRS StorageAccountType = "standard_lrs" +) + +// PossibleStorageAccountTypeValues returns an array of possible values for the StorageAccountType const type. +func PossibleStorageAccountTypeValues() []StorageAccountType { + return []StorageAccountType{PremiumLRS, StandardLRS} +} + +// SubtaskState enumerates the values for subtask state. +type SubtaskState string + +const ( + // SubtaskStateCompleted The task is no longer eligible to run, usually because the task has finished + // successfully, or the task has finished unsuccessfully and has exhausted its retry limit. A task is also + // marked as completed if an error occurred launching the task, or when the task has been terminated. + SubtaskStateCompleted SubtaskState = "completed" + // SubtaskStatePreparing The task has been assigned to a compute node, but is waiting for a required Job + // Preparation task to complete on the node. If the Job Preparation task succeeds, the task will move to + // running. If the Job Preparation task fails, the task will return to active and will be eligible to be + // assigned to a different node. + SubtaskStatePreparing SubtaskState = "preparing" + // SubtaskStateRunning The task is running on a compute node. This includes task-level preparation such as + // downloading resource files or deploying application packages specified on the task - it does not + // necessarily mean that the task command line has started executing. + SubtaskStateRunning SubtaskState = "running" +) + +// PossibleSubtaskStateValues returns an array of possible values for the SubtaskState const type. +func PossibleSubtaskStateValues() []SubtaskState { + return []SubtaskState{SubtaskStateCompleted, SubtaskStatePreparing, SubtaskStateRunning} +} + +// TaskAddStatus enumerates the values for task add status. +type TaskAddStatus string + +const ( + // TaskAddStatusClientError The task failed to add due to a client error and should not be retried without + // modifying the request as appropriate. + TaskAddStatusClientError TaskAddStatus = "clienterror" + // TaskAddStatusServerError Task failed to add due to a server error and can be retried without + // modification. + TaskAddStatusServerError TaskAddStatus = "servererror" + // TaskAddStatusSuccess The task was added successfully. + TaskAddStatusSuccess TaskAddStatus = "success" +) + +// PossibleTaskAddStatusValues returns an array of possible values for the TaskAddStatus const type. +func PossibleTaskAddStatusValues() []TaskAddStatus { + return []TaskAddStatus{TaskAddStatusClientError, TaskAddStatusServerError, TaskAddStatusSuccess} +} + +// TaskCountValidationStatus enumerates the values for task count validation status. +type TaskCountValidationStatus string + +const ( + // Unvalidated The Batch service has not been able to check state counts against the task states as + // reported in the List Tasks API. The validationStatus may be unvalidated if the job contains more than + // 200,000 tasks. + Unvalidated TaskCountValidationStatus = "unvalidated" + // Validated The Batch service has validated the state counts against the task states as reported in the + // List Tasks API. + Validated TaskCountValidationStatus = "validated" +) + +// PossibleTaskCountValidationStatusValues returns an array of possible values for the TaskCountValidationStatus const type. +func PossibleTaskCountValidationStatusValues() []TaskCountValidationStatus { + return []TaskCountValidationStatus{Unvalidated, Validated} +} + +// TaskExecutionResult enumerates the values for task execution result. +type TaskExecutionResult string + +const ( + // Failure There was an error during processing of the task. The failure may have occurred before the task + // process was launched, while the task process was executing, or after the task process exited. + Failure TaskExecutionResult = "failure" + // Success The task ran successfully. + Success TaskExecutionResult = "success" +) + +// PossibleTaskExecutionResultValues returns an array of possible values for the TaskExecutionResult const type. +func PossibleTaskExecutionResultValues() []TaskExecutionResult { + return []TaskExecutionResult{Failure, Success} +} + +// TaskState enumerates the values for task state. +type TaskState string + +const ( + // TaskStateActive The task is queued and able to run, but is not currently assigned to a compute node. A + // task enters this state when it is created, when it is enabled after being disabled, or when it is + // awaiting a retry after a failed run. + TaskStateActive TaskState = "active" + // TaskStateCompleted The task is no longer eligible to run, usually because the task has finished + // successfully, or the task has finished unsuccessfully and has exhausted its retry limit. A task is also + // marked as completed if an error occurred launching the task, or when the task has been terminated. + TaskStateCompleted TaskState = "completed" + // TaskStatePreparing The task has been assigned to a compute node, but is waiting for a required Job + // Preparation task to complete on the node. If the Job Preparation task succeeds, the task will move to + // running. If the Job Preparation task fails, the task will return to active and will be eligible to be + // assigned to a different node. + TaskStatePreparing TaskState = "preparing" + // TaskStateRunning The task is running on a compute node. This includes task-level preparation such as + // downloading resource files or deploying application packages specified on the task - it does not + // necessarily mean that the task command line has started executing. + TaskStateRunning TaskState = "running" +) + +// PossibleTaskStateValues returns an array of possible values for the TaskState const type. +func PossibleTaskStateValues() []TaskState { + return []TaskState{TaskStateActive, TaskStateCompleted, TaskStatePreparing, TaskStateRunning} +} + +// AccountListNodeAgentSkusResult ... +type AccountListNodeAgentSkusResult struct { + autorest.Response `json:"-"` + Value *[]NodeAgentSku `json:"value,omitempty"` + OdataNextLink *string `json:"odata.nextLink,omitempty"` +} + +// AccountListNodeAgentSkusResultIterator provides access to a complete listing of NodeAgentSku values. +type AccountListNodeAgentSkusResultIterator struct { + i int + page AccountListNodeAgentSkusResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *AccountListNodeAgentSkusResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter AccountListNodeAgentSkusResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter AccountListNodeAgentSkusResultIterator) Response() AccountListNodeAgentSkusResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter AccountListNodeAgentSkusResultIterator) Value() NodeAgentSku { + if !iter.page.NotDone() { + return NodeAgentSku{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (alnasr AccountListNodeAgentSkusResult) IsEmpty() bool { + return alnasr.Value == nil || len(*alnasr.Value) == 0 +} + +// accountListNodeAgentSkusResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (alnasr AccountListNodeAgentSkusResult) accountListNodeAgentSkusResultPreparer() (*http.Request, error) { + if alnasr.OdataNextLink == nil || len(to.String(alnasr.OdataNextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(alnasr.OdataNextLink))) +} + +// AccountListNodeAgentSkusResultPage contains a page of NodeAgentSku values. +type AccountListNodeAgentSkusResultPage struct { + fn func(AccountListNodeAgentSkusResult) (AccountListNodeAgentSkusResult, error) + alnasr AccountListNodeAgentSkusResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *AccountListNodeAgentSkusResultPage) Next() error { + next, err := page.fn(page.alnasr) + if err != nil { + return err + } + page.alnasr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page AccountListNodeAgentSkusResultPage) NotDone() bool { + return !page.alnasr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page AccountListNodeAgentSkusResultPage) Response() AccountListNodeAgentSkusResult { + return page.alnasr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page AccountListNodeAgentSkusResultPage) Values() []NodeAgentSku { + if page.alnasr.IsEmpty() { + return nil + } + return *page.alnasr.Value +} + +// AffinityInformation ... +type AffinityInformation struct { + // AffinityID - You can pass the affinityId of a compute node to indicate that this task needs to run on that compute node. Note that this is just a soft affinity. If the target node is busy or unavailable at the time the task is scheduled, then the task will be scheduled elsewhere. + AffinityID *string `json:"affinityId,omitempty"` +} + +// ApplicationListResult ... +type ApplicationListResult struct { + autorest.Response `json:"-"` + Value *[]ApplicationSummary `json:"value,omitempty"` + OdataNextLink *string `json:"odata.nextLink,omitempty"` +} + +// ApplicationListResultIterator provides access to a complete listing of ApplicationSummary values. +type ApplicationListResultIterator struct { + i int + page ApplicationListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *ApplicationListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter ApplicationListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter ApplicationListResultIterator) Response() ApplicationListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter ApplicationListResultIterator) Value() ApplicationSummary { + if !iter.page.NotDone() { + return ApplicationSummary{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (alr ApplicationListResult) IsEmpty() bool { + return alr.Value == nil || len(*alr.Value) == 0 +} + +// applicationListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (alr ApplicationListResult) applicationListResultPreparer() (*http.Request, error) { + if alr.OdataNextLink == nil || len(to.String(alr.OdataNextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(alr.OdataNextLink))) +} + +// ApplicationListResultPage contains a page of ApplicationSummary values. +type ApplicationListResultPage struct { + fn func(ApplicationListResult) (ApplicationListResult, error) + alr ApplicationListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *ApplicationListResultPage) Next() error { + next, err := page.fn(page.alr) + if err != nil { + return err + } + page.alr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page ApplicationListResultPage) NotDone() bool { + return !page.alr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page ApplicationListResultPage) Response() ApplicationListResult { + return page.alr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page ApplicationListResultPage) Values() []ApplicationSummary { + if page.alr.IsEmpty() { + return nil + } + return *page.alr.Value +} + +// ApplicationPackageReference ... +type ApplicationPackageReference struct { + ApplicationID *string `json:"applicationId,omitempty"` + // Version - If this is omitted on a pool, and no default version is specified for this application, the request fails with the error code InvalidApplicationPackageReferences and HTTP status code 409. If this is omitted on a task, and no default version is specified for this application, the task fails with a pre-processing error. + Version *string `json:"version,omitempty"` +} + +// ApplicationSummary ... +type ApplicationSummary struct { + autorest.Response `json:"-"` + ID *string `json:"id,omitempty"` + DisplayName *string `json:"displayName,omitempty"` + Versions *[]string `json:"versions,omitempty"` +} + +// AuthenticationTokenSettings ... +type AuthenticationTokenSettings struct { + // Access - The authentication token grants access to a limited set of Batch service operations. Currently the only supported value for the access property is 'job', which grants access to all operations related to the job which contains the task. + Access *[]AccessScope `json:"access,omitempty"` +} + +// AutoPoolSpecification ... +type AutoPoolSpecification struct { + // AutoPoolIDPrefix - The Batch service assigns each auto pool a unique identifier on creation. To distinguish between pools created for different purposes, you can specify this element to add a prefix to the ID that is assigned. The prefix can be up to 20 characters long. + AutoPoolIDPrefix *string `json:"autoPoolIdPrefix,omitempty"` + // PoolLifetimeOption - Possible values include: 'PoolLifetimeOptionJobSchedule', 'PoolLifetimeOptionJob' + PoolLifetimeOption PoolLifetimeOption `json:"poolLifetimeOption,omitempty"` + // KeepAlive - If false, the Batch service deletes the pool once its lifetime (as determined by the poolLifetimeOption setting) expires; that is, when the job or job schedule completes. If true, the Batch service does not delete the pool automatically. It is up to the user to delete auto pools created with this option. + KeepAlive *bool `json:"keepAlive,omitempty"` + Pool *PoolSpecification `json:"pool,omitempty"` +} + +// AutoScaleRun ... +type AutoScaleRun struct { + autorest.Response `json:"-"` + Timestamp *date.Time `json:"timestamp,omitempty"` + // Results - Each variable value is returned in the form $variable=value, and variables are separated by semicolons. + Results *string `json:"results,omitempty"` + Error *AutoScaleRunError `json:"error,omitempty"` +} + +// AutoScaleRunError ... +type AutoScaleRunError struct { + Code *string `json:"code,omitempty"` + Message *string `json:"message,omitempty"` + Values *[]NameValuePair `json:"values,omitempty"` +} + +// AutoUserSpecification ... +type AutoUserSpecification struct { + // Scope - The default value is task. Possible values include: 'Task', 'Pool' + Scope AutoUserScope `json:"scope,omitempty"` + // ElevationLevel - nonAdmin - The auto user is a standard user without elevated access. admin - The auto user is a user with elevated access and operates with full Administrator permissions. The default value is nonAdmin. Possible values include: 'NonAdmin', 'Admin' + ElevationLevel ElevationLevel `json:"elevationLevel,omitempty"` +} + +// Certificate a certificate that can be installed on compute nodes and can be used to authenticate operations on +// the machine. +type Certificate struct { + autorest.Response `json:"-"` + Thumbprint *string `json:"thumbprint,omitempty"` + ThumbprintAlgorithm *string `json:"thumbprintAlgorithm,omitempty"` + URL *string `json:"url,omitempty"` + // State - Possible values include: 'Active', 'Deleting', 'DeleteFailed' + State CertificateState `json:"state,omitempty"` + StateTransitionTime *date.Time `json:"stateTransitionTime,omitempty"` + // PreviousState - This property is not set if the certificate is in its initial active state. Possible values include: 'Active', 'Deleting', 'DeleteFailed' + PreviousState CertificateState `json:"previousState,omitempty"` + // PreviousStateTransitionTime - This property is not set if the certificate is in its initial Active state. + PreviousStateTransitionTime *date.Time `json:"previousStateTransitionTime,omitempty"` + PublicData *string `json:"publicData,omitempty"` + // DeleteCertificateError - This property is set only if the certificate is in the DeleteFailed state. + DeleteCertificateError *DeleteCertificateError `json:"deleteCertificateError,omitempty"` +} + +// CertificateAddParameter ... +type CertificateAddParameter struct { + Thumbprint *string `json:"thumbprint,omitempty"` + ThumbprintAlgorithm *string `json:"thumbprintAlgorithm,omitempty"` + Data *string `json:"data,omitempty"` + // CertificateFormat - Possible values include: 'Pfx', 'Cer' + CertificateFormat CertificateFormat `json:"certificateFormat,omitempty"` + // Password - This is required if the certificate format is pfx. It should be omitted if the certificate format is cer. + Password *string `json:"password,omitempty"` +} + +// CertificateListResult ... +type CertificateListResult struct { + autorest.Response `json:"-"` + Value *[]Certificate `json:"value,omitempty"` + OdataNextLink *string `json:"odata.nextLink,omitempty"` +} + +// CertificateListResultIterator provides access to a complete listing of Certificate values. +type CertificateListResultIterator struct { + i int + page CertificateListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *CertificateListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter CertificateListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter CertificateListResultIterator) Response() CertificateListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter CertificateListResultIterator) Value() Certificate { + if !iter.page.NotDone() { + return Certificate{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (clr CertificateListResult) IsEmpty() bool { + return clr.Value == nil || len(*clr.Value) == 0 +} + +// certificateListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (clr CertificateListResult) certificateListResultPreparer() (*http.Request, error) { + if clr.OdataNextLink == nil || len(to.String(clr.OdataNextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(clr.OdataNextLink))) +} + +// CertificateListResultPage contains a page of Certificate values. +type CertificateListResultPage struct { + fn func(CertificateListResult) (CertificateListResult, error) + clr CertificateListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *CertificateListResultPage) Next() error { + next, err := page.fn(page.clr) + if err != nil { + return err + } + page.clr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page CertificateListResultPage) NotDone() bool { + return !page.clr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page CertificateListResultPage) Response() CertificateListResult { + return page.clr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page CertificateListResultPage) Values() []Certificate { + if page.clr.IsEmpty() { + return nil + } + return *page.clr.Value +} + +// CertificateReference ... +type CertificateReference struct { + Thumbprint *string `json:"thumbprint,omitempty"` + ThumbprintAlgorithm *string `json:"thumbprintAlgorithm,omitempty"` + // StoreLocation - The default value is currentuser. This property is applicable only for pools configured with Windows nodes (that is, created with cloudServiceConfiguration, or with virtualMachineConfiguration using a Windows image reference). For Linux compute nodes, the certificates are stored in a directory inside the task working directory and an environment variable AZ_BATCH_CERTIFICATES_DIR is supplied to the task to query for this location. For certificates with visibility of 'remoteUser', a 'certs' directory is created in the user's home directory (e.g., /home/{user-name}/certs) and certificates are placed in that directory. Possible values include: 'CurrentUser', 'LocalMachine' + StoreLocation CertificateStoreLocation `json:"storeLocation,omitempty"` + // StoreName - This property is applicable only for pools configured with Windows nodes (that is, created with cloudServiceConfiguration, or with virtualMachineConfiguration using a Windows image reference). Common store names include: My, Root, CA, Trust, Disallowed, TrustedPeople, TrustedPublisher, AuthRoot, AddressBook, but any custom store name can also be used. The default value is My. + StoreName *string `json:"storeName,omitempty"` + // Visibility - You can specify more than one visibility in this collection. The default is all accounts. + Visibility *[]CertificateVisibility `json:"visibility,omitempty"` +} + +// CloudJob ... +type CloudJob struct { + autorest.Response `json:"-"` + // ID - The ID is case-preserving and case-insensitive (that is, you may not have two IDs within an account that differ only by case). + ID *string `json:"id,omitempty"` + DisplayName *string `json:"displayName,omitempty"` + UsesTaskDependencies *bool `json:"usesTaskDependencies,omitempty"` + URL *string `json:"url,omitempty"` + // ETag - This is an opaque string. You can use it to detect whether the job has changed between requests. In particular, you can be pass the ETag when updating a job to specify that your changes should take effect only if nobody else has modified the job in the meantime. + ETag *string `json:"eTag,omitempty"` + // LastModified - This is the last time at which the job level data, such as the job state or priority, changed. It does not factor in task-level changes such as adding new tasks or tasks changing state. + LastModified *date.Time `json:"lastModified,omitempty"` + CreationTime *date.Time `json:"creationTime,omitempty"` + // State - Possible values include: 'JobStateActive', 'JobStateDisabling', 'JobStateDisabled', 'JobStateEnabling', 'JobStateTerminating', 'JobStateCompleted', 'JobStateDeleting' + State JobState `json:"state,omitempty"` + StateTransitionTime *date.Time `json:"stateTransitionTime,omitempty"` + // PreviousState - This property is not set if the job is in its initial Active state. Possible values include: 'JobStateActive', 'JobStateDisabling', 'JobStateDisabled', 'JobStateEnabling', 'JobStateTerminating', 'JobStateCompleted', 'JobStateDeleting' + PreviousState JobState `json:"previousState,omitempty"` + // PreviousStateTransitionTime - This property is not set if the job is in its initial Active state. + PreviousStateTransitionTime *date.Time `json:"previousStateTransitionTime,omitempty"` + // Priority - Priority values can range from -1000 to 1000, with -1000 being the lowest priority and 1000 being the highest priority. The default value is 0. + Priority *int32 `json:"priority,omitempty"` + Constraints *JobConstraints `json:"constraints,omitempty"` + JobManagerTask *JobManagerTask `json:"jobManagerTask,omitempty"` + // JobPreparationTask - The Job Preparation task is a special task run on each node before any other task of the job. + JobPreparationTask *JobPreparationTask `json:"jobPreparationTask,omitempty"` + // JobReleaseTask - The Job Release task is a special task run at the end of the job on each node that has run any other task of the job. + JobReleaseTask *JobReleaseTask `json:"jobReleaseTask,omitempty"` + // CommonEnvironmentSettings - Individual tasks can override an environment setting specified here by specifying the same setting name with a different value. + CommonEnvironmentSettings *[]EnvironmentSetting `json:"commonEnvironmentSettings,omitempty"` + PoolInfo *PoolInformation `json:"poolInfo,omitempty"` + // OnAllTasksComplete - The default is noaction. Possible values include: 'NoAction', 'TerminateJob' + OnAllTasksComplete OnAllTasksComplete `json:"onAllTasksComplete,omitempty"` + // OnTaskFailure - A task is considered to have failed if has a failureInfo. A failureInfo is set if the task completes with a non-zero exit code after exhausting its retry count, or if there was an error starting the task, for example due to a resource file download error. The default is noaction. Possible values include: 'OnTaskFailureNoAction', 'OnTaskFailurePerformExitOptionsJobAction' + OnTaskFailure OnTaskFailure `json:"onTaskFailure,omitempty"` + // Metadata - The Batch service does not assign any meaning to metadata; it is solely for the use of user code. + Metadata *[]MetadataItem `json:"metadata,omitempty"` + ExecutionInfo *JobExecutionInformation `json:"executionInfo,omitempty"` + Stats *JobStatistics `json:"stats,omitempty"` +} + +// CloudJobListPreparationAndReleaseTaskStatusResult ... +type CloudJobListPreparationAndReleaseTaskStatusResult struct { + autorest.Response `json:"-"` + Value *[]JobPreparationAndReleaseTaskExecutionInformation `json:"value,omitempty"` + OdataNextLink *string `json:"odata.nextLink,omitempty"` +} + +// CloudJobListPreparationAndReleaseTaskStatusResultIterator provides access to a complete listing of +// JobPreparationAndReleaseTaskExecutionInformation values. +type CloudJobListPreparationAndReleaseTaskStatusResultIterator struct { + i int + page CloudJobListPreparationAndReleaseTaskStatusResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *CloudJobListPreparationAndReleaseTaskStatusResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter CloudJobListPreparationAndReleaseTaskStatusResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter CloudJobListPreparationAndReleaseTaskStatusResultIterator) Response() CloudJobListPreparationAndReleaseTaskStatusResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter CloudJobListPreparationAndReleaseTaskStatusResultIterator) Value() JobPreparationAndReleaseTaskExecutionInformation { + if !iter.page.NotDone() { + return JobPreparationAndReleaseTaskExecutionInformation{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (cjlpartsr CloudJobListPreparationAndReleaseTaskStatusResult) IsEmpty() bool { + return cjlpartsr.Value == nil || len(*cjlpartsr.Value) == 0 +} + +// cloudJobListPreparationAndReleaseTaskStatusResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (cjlpartsr CloudJobListPreparationAndReleaseTaskStatusResult) cloudJobListPreparationAndReleaseTaskStatusResultPreparer() (*http.Request, error) { + if cjlpartsr.OdataNextLink == nil || len(to.String(cjlpartsr.OdataNextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(cjlpartsr.OdataNextLink))) +} + +// CloudJobListPreparationAndReleaseTaskStatusResultPage contains a page of +// JobPreparationAndReleaseTaskExecutionInformation values. +type CloudJobListPreparationAndReleaseTaskStatusResultPage struct { + fn func(CloudJobListPreparationAndReleaseTaskStatusResult) (CloudJobListPreparationAndReleaseTaskStatusResult, error) + cjlpartsr CloudJobListPreparationAndReleaseTaskStatusResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *CloudJobListPreparationAndReleaseTaskStatusResultPage) Next() error { + next, err := page.fn(page.cjlpartsr) + if err != nil { + return err + } + page.cjlpartsr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page CloudJobListPreparationAndReleaseTaskStatusResultPage) NotDone() bool { + return !page.cjlpartsr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page CloudJobListPreparationAndReleaseTaskStatusResultPage) Response() CloudJobListPreparationAndReleaseTaskStatusResult { + return page.cjlpartsr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page CloudJobListPreparationAndReleaseTaskStatusResultPage) Values() []JobPreparationAndReleaseTaskExecutionInformation { + if page.cjlpartsr.IsEmpty() { + return nil + } + return *page.cjlpartsr.Value +} + +// CloudJobListResult ... +type CloudJobListResult struct { + autorest.Response `json:"-"` + Value *[]CloudJob `json:"value,omitempty"` + OdataNextLink *string `json:"odata.nextLink,omitempty"` +} + +// CloudJobListResultIterator provides access to a complete listing of CloudJob values. +type CloudJobListResultIterator struct { + i int + page CloudJobListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *CloudJobListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter CloudJobListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter CloudJobListResultIterator) Response() CloudJobListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter CloudJobListResultIterator) Value() CloudJob { + if !iter.page.NotDone() { + return CloudJob{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (cjlr CloudJobListResult) IsEmpty() bool { + return cjlr.Value == nil || len(*cjlr.Value) == 0 +} + +// cloudJobListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (cjlr CloudJobListResult) cloudJobListResultPreparer() (*http.Request, error) { + if cjlr.OdataNextLink == nil || len(to.String(cjlr.OdataNextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(cjlr.OdataNextLink))) +} + +// CloudJobListResultPage contains a page of CloudJob values. +type CloudJobListResultPage struct { + fn func(CloudJobListResult) (CloudJobListResult, error) + cjlr CloudJobListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *CloudJobListResultPage) Next() error { + next, err := page.fn(page.cjlr) + if err != nil { + return err + } + page.cjlr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page CloudJobListResultPage) NotDone() bool { + return !page.cjlr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page CloudJobListResultPage) Response() CloudJobListResult { + return page.cjlr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page CloudJobListResultPage) Values() []CloudJob { + if page.cjlr.IsEmpty() { + return nil + } + return *page.cjlr.Value +} + +// CloudJobSchedule ... +type CloudJobSchedule struct { + autorest.Response `json:"-"` + ID *string `json:"id,omitempty"` + DisplayName *string `json:"displayName,omitempty"` + URL *string `json:"url,omitempty"` + // ETag - This is an opaque string. You can use it to detect whether the job schedule has changed between requests. In particular, you can be pass the ETag with an Update Job Schedule request to specify that your changes should take effect only if nobody else has modified the schedule in the meantime. + ETag *string `json:"eTag,omitempty"` + // LastModified - This is the last time at which the schedule level data, such as the job specification or recurrence information, changed. It does not factor in job-level changes such as new jobs being created or jobs changing state. + LastModified *date.Time `json:"lastModified,omitempty"` + CreationTime *date.Time `json:"creationTime,omitempty"` + // State - Possible values include: 'JobScheduleStateActive', 'JobScheduleStateCompleted', 'JobScheduleStateDisabled', 'JobScheduleStateTerminating', 'JobScheduleStateDeleting' + State JobScheduleState `json:"state,omitempty"` + StateTransitionTime *date.Time `json:"stateTransitionTime,omitempty"` + // PreviousState - This property is not present if the job schedule is in its initial active state. Possible values include: 'JobScheduleStateActive', 'JobScheduleStateCompleted', 'JobScheduleStateDisabled', 'JobScheduleStateTerminating', 'JobScheduleStateDeleting' + PreviousState JobScheduleState `json:"previousState,omitempty"` + // PreviousStateTransitionTime - This property is not present if the job schedule is in its initial active state. + PreviousStateTransitionTime *date.Time `json:"previousStateTransitionTime,omitempty"` + Schedule *Schedule `json:"schedule,omitempty"` + JobSpecification *JobSpecification `json:"jobSpecification,omitempty"` + ExecutionInfo *JobScheduleExecutionInformation `json:"executionInfo,omitempty"` + // Metadata - The Batch service does not assign any meaning to metadata; it is solely for the use of user code. + Metadata *[]MetadataItem `json:"metadata,omitempty"` + Stats *JobScheduleStatistics `json:"stats,omitempty"` +} + +// CloudJobScheduleListResult ... +type CloudJobScheduleListResult struct { + autorest.Response `json:"-"` + Value *[]CloudJobSchedule `json:"value,omitempty"` + OdataNextLink *string `json:"odata.nextLink,omitempty"` +} + +// CloudJobScheduleListResultIterator provides access to a complete listing of CloudJobSchedule values. +type CloudJobScheduleListResultIterator struct { + i int + page CloudJobScheduleListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *CloudJobScheduleListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter CloudJobScheduleListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter CloudJobScheduleListResultIterator) Response() CloudJobScheduleListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter CloudJobScheduleListResultIterator) Value() CloudJobSchedule { + if !iter.page.NotDone() { + return CloudJobSchedule{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (cjslr CloudJobScheduleListResult) IsEmpty() bool { + return cjslr.Value == nil || len(*cjslr.Value) == 0 +} + +// cloudJobScheduleListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (cjslr CloudJobScheduleListResult) cloudJobScheduleListResultPreparer() (*http.Request, error) { + if cjslr.OdataNextLink == nil || len(to.String(cjslr.OdataNextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(cjslr.OdataNextLink))) +} + +// CloudJobScheduleListResultPage contains a page of CloudJobSchedule values. +type CloudJobScheduleListResultPage struct { + fn func(CloudJobScheduleListResult) (CloudJobScheduleListResult, error) + cjslr CloudJobScheduleListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *CloudJobScheduleListResultPage) Next() error { + next, err := page.fn(page.cjslr) + if err != nil { + return err + } + page.cjslr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page CloudJobScheduleListResultPage) NotDone() bool { + return !page.cjslr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page CloudJobScheduleListResultPage) Response() CloudJobScheduleListResult { + return page.cjslr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page CloudJobScheduleListResultPage) Values() []CloudJobSchedule { + if page.cjslr.IsEmpty() { + return nil + } + return *page.cjslr.Value +} + +// CloudPool ... +type CloudPool struct { + autorest.Response `json:"-"` + // ID - The ID can contain any combination of alphanumeric characters including hyphens and underscores, and cannot contain more than 64 characters. The ID is case-preserving and case-insensitive (that is, you may not have two IDs within an account that differ only by case). + ID *string `json:"id,omitempty"` + // DisplayName - The display name need not be unique and can contain any Unicode characters up to a maximum length of 1024. + DisplayName *string `json:"displayName,omitempty"` + URL *string `json:"url,omitempty"` + // ETag - This is an opaque string. You can use it to detect whether the pool has changed between requests. In particular, you can be pass the ETag when updating a pool to specify that your changes should take effect only if nobody else has modified the pool in the meantime. + ETag *string `json:"eTag,omitempty"` + // LastModified - This is the last time at which the pool level data, such as the targetDedicatedNodes or enableAutoscale settings, changed. It does not factor in node-level changes such as a compute node changing state. + LastModified *date.Time `json:"lastModified,omitempty"` + CreationTime *date.Time `json:"creationTime,omitempty"` + // State - Possible values include: 'PoolStateActive', 'PoolStateDeleting', 'PoolStateUpgrading' + State PoolState `json:"state,omitempty"` + StateTransitionTime *date.Time `json:"stateTransitionTime,omitempty"` + // AllocationState - Possible values include: 'Steady', 'Resizing', 'Stopping' + AllocationState AllocationState `json:"allocationState,omitempty"` + AllocationStateTransitionTime *date.Time `json:"allocationStateTransitionTime,omitempty"` + // VMSize - For information about available sizes of virtual machines for Cloud Services pools (pools created with cloudServiceConfiguration), see Sizes for Cloud Services (http://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). Batch supports all Cloud Services VM sizes except ExtraSmall, A1V2 and A2V2. For information about available VM sizes for pools using images from the Virtual Machines Marketplace (pools created with virtualMachineConfiguration) see Sizes for Virtual Machines (Linux) (https://azure.microsoft.com/documentation/articles/virtual-machines-linux-sizes/) or Sizes for Virtual Machines (Windows) (https://azure.microsoft.com/documentation/articles/virtual-machines-windows-sizes/). Batch supports all Azure VM sizes except STANDARD_A0 and those with premium storage (STANDARD_GS, STANDARD_DS, and STANDARD_DSV2 series). + VMSize *string `json:"vmSize,omitempty"` + // CloudServiceConfiguration - This property and virtualMachineConfiguration are mutually exclusive and one of the properties must be specified. This property cannot be specified if the Batch account was created with its poolAllocationMode property set to 'UserSubscription'. + CloudServiceConfiguration *CloudServiceConfiguration `json:"cloudServiceConfiguration,omitempty"` + // VirtualMachineConfiguration - This property and cloudServiceConfiguration are mutually exclusive and one of the properties must be specified. + VirtualMachineConfiguration *VirtualMachineConfiguration `json:"virtualMachineConfiguration,omitempty"` + // ResizeTimeout - This is the timeout for the most recent resize operation. (The initial sizing when the pool is created counts as a resize.) The default value is 15 minutes. + ResizeTimeout *string `json:"resizeTimeout,omitempty"` + // ResizeErrors - This property is set only if one or more errors occurred during the last pool resize, and only when the pool allocationState is Steady. + ResizeErrors *[]ResizeError `json:"resizeErrors,omitempty"` + CurrentDedicatedNodes *int32 `json:"currentDedicatedNodes,omitempty"` + // CurrentLowPriorityNodes - Low-priority compute nodes which have been preempted are included in this count. + CurrentLowPriorityNodes *int32 `json:"currentLowPriorityNodes,omitempty"` + TargetDedicatedNodes *int32 `json:"targetDedicatedNodes,omitempty"` + TargetLowPriorityNodes *int32 `json:"targetLowPriorityNodes,omitempty"` + // EnableAutoScale - If false, at least one of targetDedicateNodes and targetLowPriorityNodes must be specified. If true, the autoScaleFormula property is required and the pool automatically resizes according to the formula. The default value is false. + EnableAutoScale *bool `json:"enableAutoScale,omitempty"` + // AutoScaleFormula - This property is set only if the pool automatically scales, i.e. enableAutoScale is true. + AutoScaleFormula *string `json:"autoScaleFormula,omitempty"` + // AutoScaleEvaluationInterval - This property is set only if the pool automatically scales, i.e. enableAutoScale is true. + AutoScaleEvaluationInterval *string `json:"autoScaleEvaluationInterval,omitempty"` + // AutoScaleRun - This property is set only if the pool automatically scales, i.e. enableAutoScale is true. + AutoScaleRun *AutoScaleRun `json:"autoScaleRun,omitempty"` + // EnableInterNodeCommunication - This imposes restrictions on which nodes can be assigned to the pool. Specifying this value can reduce the chance of the requested number of nodes to be allocated in the pool. + EnableInterNodeCommunication *bool `json:"enableInterNodeCommunication,omitempty"` + NetworkConfiguration *NetworkConfiguration `json:"networkConfiguration,omitempty"` + StartTask *StartTask `json:"startTask,omitempty"` + // CertificateReferences - For Windows compute nodes, the Batch service installs the certificates to the specified certificate store and location. For Linux compute nodes, the certificates are stored in a directory inside the task working directory and an environment variable AZ_BATCH_CERTIFICATES_DIR is supplied to the task to query for this location. For certificates with visibility of 'remoteUser', a 'certs' directory is created in the user's home directory (e.g., /home/{user-name}/certs) and certificates are placed in that directory. + CertificateReferences *[]CertificateReference `json:"certificateReferences,omitempty"` + ApplicationPackageReferences *[]ApplicationPackageReference `json:"applicationPackageReferences,omitempty"` + // ApplicationLicenses - The list of application licenses must be a subset of available Batch service application licenses. If a license is requested which is not supported, pool creation will fail. + ApplicationLicenses *[]string `json:"applicationLicenses,omitempty"` + MaxTasksPerNode *int32 `json:"maxTasksPerNode,omitempty"` + TaskSchedulingPolicy *TaskSchedulingPolicy `json:"taskSchedulingPolicy,omitempty"` + UserAccounts *[]UserAccount `json:"userAccounts,omitempty"` + Metadata *[]MetadataItem `json:"metadata,omitempty"` + Stats *PoolStatistics `json:"stats,omitempty"` +} + +// CloudPoolListResult ... +type CloudPoolListResult struct { + autorest.Response `json:"-"` + Value *[]CloudPool `json:"value,omitempty"` + OdataNextLink *string `json:"odata.nextLink,omitempty"` +} + +// CloudPoolListResultIterator provides access to a complete listing of CloudPool values. +type CloudPoolListResultIterator struct { + i int + page CloudPoolListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *CloudPoolListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter CloudPoolListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter CloudPoolListResultIterator) Response() CloudPoolListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter CloudPoolListResultIterator) Value() CloudPool { + if !iter.page.NotDone() { + return CloudPool{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (cplr CloudPoolListResult) IsEmpty() bool { + return cplr.Value == nil || len(*cplr.Value) == 0 +} + +// cloudPoolListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (cplr CloudPoolListResult) cloudPoolListResultPreparer() (*http.Request, error) { + if cplr.OdataNextLink == nil || len(to.String(cplr.OdataNextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(cplr.OdataNextLink))) +} + +// CloudPoolListResultPage contains a page of CloudPool values. +type CloudPoolListResultPage struct { + fn func(CloudPoolListResult) (CloudPoolListResult, error) + cplr CloudPoolListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *CloudPoolListResultPage) Next() error { + next, err := page.fn(page.cplr) + if err != nil { + return err + } + page.cplr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page CloudPoolListResultPage) NotDone() bool { + return !page.cplr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page CloudPoolListResultPage) Response() CloudPoolListResult { + return page.cplr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page CloudPoolListResultPage) Values() []CloudPool { + if page.cplr.IsEmpty() { + return nil + } + return *page.cplr.Value +} + +// CloudServiceConfiguration ... +type CloudServiceConfiguration struct { + // OsFamily - Possible values are: 2 - OS Family 2, equivalent to Windows Server 2008 R2 SP1. 3 - OS Family 3, equivalent to Windows Server 2012. 4 - OS Family 4, equivalent to Windows Server 2012 R2. 5 - OS Family 5, equivalent to Windows Server 2016. For more information, see Azure Guest OS Releases (https://azure.microsoft.com/documentation/articles/cloud-services-guestos-update-matrix/#releases). + OsFamily *string `json:"osFamily,omitempty"` + // TargetOSVersion - The default value is * which specifies the latest operating system version for the specified OS family. + TargetOSVersion *string `json:"targetOSVersion,omitempty"` + // CurrentOSVersion - This may differ from targetOSVersion if the pool state is Upgrading. In this case some virtual machines may be on the targetOSVersion and some may be on the currentOSVersion during the upgrade process. Once all virtual machines have upgraded, currentOSVersion is updated to be the same as targetOSVersion. + CurrentOSVersion *string `json:"currentOSVersion,omitempty"` +} + +// CloudTask ... +type CloudTask struct { + autorest.Response `json:"-"` + // ID - The ID can contain any combination of alphanumeric characters including hyphens and underscores, and cannot contain more than 64 characters. + ID *string `json:"id,omitempty"` + // DisplayName - The display name need not be unique and can contain any Unicode characters up to a maximum length of 1024. + DisplayName *string `json:"displayName,omitempty"` + URL *string `json:"url,omitempty"` + // ETag - This is an opaque string. You can use it to detect whether the task has changed between requests. In particular, you can be pass the ETag when updating a task to specify that your changes should take effect only if nobody else has modified the task in the meantime. + ETag *string `json:"eTag,omitempty"` + LastModified *date.Time `json:"lastModified,omitempty"` + CreationTime *date.Time `json:"creationTime,omitempty"` + // ExitConditions - How the Batch service should respond when the task completes. + ExitConditions *ExitConditions `json:"exitConditions,omitempty"` + // State - Possible values include: 'TaskStateActive', 'TaskStatePreparing', 'TaskStateRunning', 'TaskStateCompleted' + State TaskState `json:"state,omitempty"` + StateTransitionTime *date.Time `json:"stateTransitionTime,omitempty"` + // PreviousState - This property is not set if the task is in its initial Active state. Possible values include: 'TaskStateActive', 'TaskStatePreparing', 'TaskStateRunning', 'TaskStateCompleted' + PreviousState TaskState `json:"previousState,omitempty"` + // PreviousStateTransitionTime - This property is not set if the task is in its initial Active state. + PreviousStateTransitionTime *date.Time `json:"previousStateTransitionTime,omitempty"` + // CommandLine - For multi-instance tasks, the command line is executed as the primary task, after the primary task and all subtasks have finished executing the coordination command line. The command line does not run under a shell, and therefore cannot take advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows or "/bin/sh -c MyCommand" in Linux. + CommandLine *string `json:"commandLine,omitempty"` + // ContainerSettings - If the pool that will run this task has containerConfiguration set, this must be set as well. If the pool that will run this task doesn't have containerConfiguration set, this must not be set. When this is specified, all directories recursively below the AZ_BATCH_NODE_ROOT_DIR (the root of Azure Batch directories on the node) are mapped into the container, all task environment variables are mapped into the container, and the task command line is executed in the container. + ContainerSettings *TaskContainerSettings `json:"containerSettings,omitempty"` + // ResourceFiles - For multi-instance tasks, the resource files will only be downloaded to the compute node on which the primary task is executed. + ResourceFiles *[]ResourceFile `json:"resourceFiles,omitempty"` + // OutputFiles - For multi-instance tasks, the files will only be uploaded from the compute node on which the primary task is executed. + OutputFiles *[]OutputFile `json:"outputFiles,omitempty"` + EnvironmentSettings *[]EnvironmentSetting `json:"environmentSettings,omitempty"` + AffinityInfo *AffinityInformation `json:"affinityInfo,omitempty"` + Constraints *TaskConstraints `json:"constraints,omitempty"` + // UserIdentity - If omitted, the task runs as a non-administrative user unique to the task. + UserIdentity *UserIdentity `json:"userIdentity,omitempty"` + ExecutionInfo *TaskExecutionInformation `json:"executionInfo,omitempty"` + NodeInfo *ComputeNodeInformation `json:"nodeInfo,omitempty"` + MultiInstanceSettings *MultiInstanceSettings `json:"multiInstanceSettings,omitempty"` + Stats *TaskStatistics `json:"stats,omitempty"` + // DependsOn - This task will not be scheduled until all tasks that it depends on have completed successfully. If any of those tasks fail and exhaust their retry counts, this task will never be scheduled. + DependsOn *TaskDependencies `json:"dependsOn,omitempty"` + // ApplicationPackageReferences - Application packages are downloaded and deployed to a shared directory, not the task working directory. Therefore, if a referenced package is already on the compute node, and is up to date, then it is not re-downloaded; the existing copy on the compute node is used. If a referenced application package cannot be installed, for example because the package has been deleted or because download failed, the task fails. + ApplicationPackageReferences *[]ApplicationPackageReference `json:"applicationPackageReferences,omitempty"` + // AuthenticationTokenSettings - If this property is set, the Batch service provides the task with an authentication token which can be used to authenticate Batch service operations without requiring an account access key. The token is provided via the AZ_BATCH_AUTHENTICATION_TOKEN environment variable. The operations that the task can carry out using the token depend on the settings. For example, a task can request job permissions in order to add other tasks to the job, or check the status of the job or of other tasks under the job. + AuthenticationTokenSettings *AuthenticationTokenSettings `json:"authenticationTokenSettings,omitempty"` +} + +// CloudTaskListResult ... +type CloudTaskListResult struct { + autorest.Response `json:"-"` + Value *[]CloudTask `json:"value,omitempty"` + OdataNextLink *string `json:"odata.nextLink,omitempty"` +} + +// CloudTaskListResultIterator provides access to a complete listing of CloudTask values. +type CloudTaskListResultIterator struct { + i int + page CloudTaskListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *CloudTaskListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter CloudTaskListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter CloudTaskListResultIterator) Response() CloudTaskListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter CloudTaskListResultIterator) Value() CloudTask { + if !iter.page.NotDone() { + return CloudTask{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (ctlr CloudTaskListResult) IsEmpty() bool { + return ctlr.Value == nil || len(*ctlr.Value) == 0 +} + +// cloudTaskListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (ctlr CloudTaskListResult) cloudTaskListResultPreparer() (*http.Request, error) { + if ctlr.OdataNextLink == nil || len(to.String(ctlr.OdataNextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(ctlr.OdataNextLink))) +} + +// CloudTaskListResultPage contains a page of CloudTask values. +type CloudTaskListResultPage struct { + fn func(CloudTaskListResult) (CloudTaskListResult, error) + ctlr CloudTaskListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *CloudTaskListResultPage) Next() error { + next, err := page.fn(page.ctlr) + if err != nil { + return err + } + page.ctlr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page CloudTaskListResultPage) NotDone() bool { + return !page.ctlr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page CloudTaskListResultPage) Response() CloudTaskListResult { + return page.ctlr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page CloudTaskListResultPage) Values() []CloudTask { + if page.ctlr.IsEmpty() { + return nil + } + return *page.ctlr.Value +} + +// CloudTaskListSubtasksResult ... +type CloudTaskListSubtasksResult struct { + autorest.Response `json:"-"` + Value *[]SubtaskInformation `json:"value,omitempty"` +} + +// ComputeNode ... +type ComputeNode struct { + autorest.Response `json:"-"` + // ID - Every node that is added to a pool is assigned a unique ID. Whenever a node is removed from a pool, all of its local files are deleted, and the ID is reclaimed and could be reused for new nodes. + ID *string `json:"id,omitempty"` + URL *string `json:"url,omitempty"` + // State - The low-priority node has been preempted. Tasks which were running on the node when it was pre-empted will be rescheduled when another node becomes available. Possible values include: 'Idle', 'Rebooting', 'Reimaging', 'Running', 'Unusable', 'Creating', 'Starting', 'WaitingForStartTask', 'StartTaskFailed', 'Unknown', 'LeavingPool', 'Offline', 'Preempted' + State ComputeNodeState `json:"state,omitempty"` + // SchedulingState - Possible values include: 'Enabled', 'Disabled' + SchedulingState SchedulingState `json:"schedulingState,omitempty"` + StateTransitionTime *date.Time `json:"stateTransitionTime,omitempty"` + // LastBootTime - This property may not be present if the node state is unusable. + LastBootTime *date.Time `json:"lastBootTime,omitempty"` + AllocationTime *date.Time `json:"allocationTime,omitempty"` + // IPAddress - Every node that is added to a pool is assigned a unique IP address. Whenever a node is removed from a pool, all of its local files are deleted, and the IP address is reclaimed and could be reused for new nodes. + IPAddress *string `json:"ipAddress,omitempty"` + // AffinityID - Note that this is just a soft affinity. If the target node is busy or unavailable at the time the task is scheduled, then the task will be scheduled elsewhere. + AffinityID *string `json:"affinityId,omitempty"` + // VMSize - For information about available sizes of virtual machines for Cloud Services pools (pools created with cloudServiceConfiguration), see Sizes for Cloud Services (http://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). Batch supports all Cloud Services VM sizes except ExtraSmall, A1V2 and A2V2. For information about available VM sizes for pools using images from the Virtual Machines Marketplace (pools created with virtualMachineConfiguration) see Sizes for Virtual Machines (Linux) (https://azure.microsoft.com/documentation/articles/virtual-machines-linux-sizes/) or Sizes for Virtual Machines (Windows) (https://azure.microsoft.com/documentation/articles/virtual-machines-windows-sizes/). Batch supports all Azure VM sizes except STANDARD_A0 and those with premium storage (STANDARD_GS, STANDARD_DS, and STANDARD_DSV2 series). + VMSize *string `json:"vmSize,omitempty"` + TotalTasksRun *int32 `json:"totalTasksRun,omitempty"` + RunningTasksCount *int32 `json:"runningTasksCount,omitempty"` + TotalTasksSucceeded *int32 `json:"totalTasksSucceeded,omitempty"` + // RecentTasks - This property is present only if at least one task has run on this node since it was assigned to the pool. + RecentTasks *[]TaskInformation `json:"recentTasks,omitempty"` + StartTask *StartTask `json:"startTask,omitempty"` + StartTaskInfo *StartTaskInformation `json:"startTaskInfo,omitempty"` + // CertificateReferences - For Windows compute nodes, the Batch service installs the certificates to the specified certificate store and location. For Linux compute nodes, the certificates are stored in a directory inside the task working directory and an environment variable AZ_BATCH_CERTIFICATES_DIR is supplied to the task to query for this location. For certificates with visibility of 'remoteUser', a 'certs' directory is created in the user's home directory (e.g., /home/{user-name}/certs) and certificates are placed in that directory. + CertificateReferences *[]CertificateReference `json:"certificateReferences,omitempty"` + Errors *[]ComputeNodeError `json:"errors,omitempty"` + IsDedicated *bool `json:"isDedicated,omitempty"` + EndpointConfiguration *ComputeNodeEndpointConfiguration `json:"endpointConfiguration,omitempty"` +} + +// ComputeNodeEndpointConfiguration ... +type ComputeNodeEndpointConfiguration struct { + InboundEndpoints *[]InboundEndpoint `json:"inboundEndpoints,omitempty"` +} + +// ComputeNodeError ... +type ComputeNodeError struct { + Code *string `json:"code,omitempty"` + Message *string `json:"message,omitempty"` + ErrorDetails *[]NameValuePair `json:"errorDetails,omitempty"` +} + +// ComputeNodeGetRemoteLoginSettingsResult ... +type ComputeNodeGetRemoteLoginSettingsResult struct { + autorest.Response `json:"-"` + RemoteLoginIPAddress *string `json:"remoteLoginIPAddress,omitempty"` + RemoteLoginPort *int32 `json:"remoteLoginPort,omitempty"` +} + +// ComputeNodeInformation ... +type ComputeNodeInformation struct { + AffinityID *string `json:"affinityId,omitempty"` + NodeURL *string `json:"nodeUrl,omitempty"` + PoolID *string `json:"poolId,omitempty"` + NodeID *string `json:"nodeId,omitempty"` + TaskRootDirectory *string `json:"taskRootDirectory,omitempty"` + TaskRootDirectoryURL *string `json:"taskRootDirectoryUrl,omitempty"` +} + +// ComputeNodeListResult ... +type ComputeNodeListResult struct { + autorest.Response `json:"-"` + Value *[]ComputeNode `json:"value,omitempty"` + OdataNextLink *string `json:"odata.nextLink,omitempty"` +} + +// ComputeNodeListResultIterator provides access to a complete listing of ComputeNode values. +type ComputeNodeListResultIterator struct { + i int + page ComputeNodeListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *ComputeNodeListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter ComputeNodeListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter ComputeNodeListResultIterator) Response() ComputeNodeListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter ComputeNodeListResultIterator) Value() ComputeNode { + if !iter.page.NotDone() { + return ComputeNode{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (cnlr ComputeNodeListResult) IsEmpty() bool { + return cnlr.Value == nil || len(*cnlr.Value) == 0 +} + +// computeNodeListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (cnlr ComputeNodeListResult) computeNodeListResultPreparer() (*http.Request, error) { + if cnlr.OdataNextLink == nil || len(to.String(cnlr.OdataNextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(cnlr.OdataNextLink))) +} + +// ComputeNodeListResultPage contains a page of ComputeNode values. +type ComputeNodeListResultPage struct { + fn func(ComputeNodeListResult) (ComputeNodeListResult, error) + cnlr ComputeNodeListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *ComputeNodeListResultPage) Next() error { + next, err := page.fn(page.cnlr) + if err != nil { + return err + } + page.cnlr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page ComputeNodeListResultPage) NotDone() bool { + return !page.cnlr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page ComputeNodeListResultPage) Response() ComputeNodeListResult { + return page.cnlr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page ComputeNodeListResultPage) Values() []ComputeNode { + if page.cnlr.IsEmpty() { + return nil + } + return *page.cnlr.Value +} + +// ComputeNodeUser ... +type ComputeNodeUser struct { + Name *string `json:"name,omitempty"` + // IsAdmin - The default value is false. + IsAdmin *bool `json:"isAdmin,omitempty"` + // ExpiryTime - If omitted, the default is 1 day from the current time. For Linux compute nodes, the expiryTime has a precision up to a day. + ExpiryTime *date.Time `json:"expiryTime,omitempty"` + // Password - The password is required for Windows nodes (those created with 'cloudServiceConfiguration', or created with 'virtualMachineConfiguration' using a Windows image reference). For Linux compute nodes, the password can optionally be specified along with the sshPublicKey property. + Password *string `json:"password,omitempty"` + // SSHPublicKey - The public key should be compatible with OpenSSH encoding and should be base 64 encoded. This property can be specified only for Linux nodes. If this is specified for a Windows node, then the Batch service rejects the request; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). + SSHPublicKey *string `json:"sshPublicKey,omitempty"` +} + +// ContainerConfiguration ... +type ContainerConfiguration struct { + Type *string `json:"type,omitempty"` + // ContainerImageNames - This is the full image reference, as would be specified to "docker pull". An image will be sourced from the default Docker registry unless the image is fully qualified with an alternative registry. + ContainerImageNames *[]string `json:"containerImageNames,omitempty"` + // ContainerRegistries - If any images must be downloaded from a private registry which requires credentials, then those credentials must be provided here. + ContainerRegistries *[]ContainerRegistry `json:"containerRegistries,omitempty"` +} + +// ContainerRegistry ... +type ContainerRegistry struct { + // RegistryServer - If omitted, the default is "docker.io". + RegistryServer *string `json:"registryServer,omitempty"` + UserName *string `json:"username,omitempty"` + Password *string `json:"password,omitempty"` +} + +// DataDisk ... +type DataDisk struct { + // Lun - The lun is used to uniquely identify each data disk. If attaching multiple disks, each should have a distinct lun. + Lun *int32 `json:"lun,omitempty"` + // Caching - The default value for caching is none. For information about the caching options see: https://blogs.msdn.microsoft.com/windowsazurestorage/2012/06/27/exploring-windows-azure-drives-disks-and-images/. Possible values include: 'None', 'ReadOnly', 'ReadWrite' + Caching CachingType `json:"caching,omitempty"` + DiskSizeGB *int32 `json:"diskSizeGB,omitempty"` + // StorageAccountType - If omitted, the default is "standard_lrs". Possible values include: 'StandardLRS', 'PremiumLRS' + StorageAccountType StorageAccountType `json:"storageAccountType,omitempty"` +} + +// DeleteCertificateError ... +type DeleteCertificateError struct { + Code *string `json:"code,omitempty"` + Message *string `json:"message,omitempty"` + // Values - This list includes details such as the active pools and nodes referencing this certificate. However, if a large number of resources reference the certificate, the list contains only about the first hundred. + Values *[]NameValuePair `json:"values,omitempty"` +} + +// EnvironmentSetting ... +type EnvironmentSetting struct { + Name *string `json:"name,omitempty"` + Value *string `json:"value,omitempty"` +} + +// Error ... +type Error struct { + Code *string `json:"code,omitempty"` + Message *ErrorMessage `json:"message,omitempty"` + Values *[]ErrorDetail `json:"values,omitempty"` +} + +// ErrorDetail ... +type ErrorDetail struct { + Key *string `json:"key,omitempty"` + Value *string `json:"value,omitempty"` +} + +// ErrorMessage ... +type ErrorMessage struct { + Lang *string `json:"lang,omitempty"` + Value *string `json:"value,omitempty"` +} + +// ExitCodeMapping ... +type ExitCodeMapping struct { + Code *int32 `json:"code,omitempty"` + ExitOptions *ExitOptions `json:"exitOptions,omitempty"` +} + +// ExitCodeRangeMapping ... +type ExitCodeRangeMapping struct { + Start *int32 `json:"start,omitempty"` + End *int32 `json:"end,omitempty"` + ExitOptions *ExitOptions `json:"exitOptions,omitempty"` +} + +// ExitConditions ... +type ExitConditions struct { + ExitCodes *[]ExitCodeMapping `json:"exitCodes,omitempty"` + ExitCodeRanges *[]ExitCodeRangeMapping `json:"exitCodeRanges,omitempty"` + PreProcessingError *ExitOptions `json:"preProcessingError,omitempty"` + // FileUploadError - If the task exited with an exit code that was specified via exitCodes or exitCodeRanges, and then encountered a file upload error, then the action specified by the exit code takes precedence. + FileUploadError *ExitOptions `json:"fileUploadError,omitempty"` + // Default - This value is used if the task exits with any nonzero exit code not listed in the exitCodes or exitCodeRanges collection, with a pre-processing error if the preProcessingError property is not present, or with a file upload error if the fileUploadError property is not present. If you want non-default behaviour on exit code 0, you must list it explicitly using the exitCodes or exitCodeRanges collection. + Default *ExitOptions `json:"default,omitempty"` +} + +// ExitOptions ... +type ExitOptions struct { + // JobAction - The default is none for exit code 0 and terminate for all other exit conditions. If the job's onTaskFailed property is noaction, then specifying this property returns an error and the add task request fails with an invalid property value error; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). Possible values include: 'JobActionNone', 'JobActionDisable', 'JobActionTerminate' + JobAction JobAction `json:"jobAction,omitempty"` + // DependencyAction - The default is 'satisfy' for exit code 0, and 'block' for all other exit conditions. If the job's usesTaskDependencies property is set to false, then specifying the dependencyAction property returns an error and the add task request fails with an invalid property value error; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). Possible values include: 'Satisfy', 'Block' + DependencyAction DependencyAction `json:"dependencyAction,omitempty"` +} + +// FileProperties ... +type FileProperties struct { + // CreationTime - The creation time is not returned for files on Linux compute nodes. + CreationTime *date.Time `json:"creationTime,omitempty"` + LastModified *date.Time `json:"lastModified,omitempty"` + ContentLength *int64 `json:"contentLength,omitempty"` + ContentType *string `json:"contentType,omitempty"` + // FileMode - The file mode is returned only for files on Linux compute nodes. + FileMode *string `json:"fileMode,omitempty"` +} + +// ImageReference ... +type ImageReference struct { + // Publisher - For example, Canonical or MicrosoftWindowsServer. + Publisher *string `json:"publisher,omitempty"` + // Offer - For example, UbuntuServer or WindowsServer. + Offer *string `json:"offer,omitempty"` + // Sku - For example, 14.04.0-LTS or 2012-R2-Datacenter. + Sku *string `json:"sku,omitempty"` + // Version - A value of 'latest' can be specified to select the latest version of an image. If omitted, the default is 'latest'. + Version *string `json:"version,omitempty"` + // VirtualMachineImageID - This property is mutually exclusive with other ImageReference properties. The virtual machine image must be in the same region and subscription as the Azure Batch account. For information about the firewall settings for the Batch node agent to communicate with the Batch service see https://docs.microsoft.com/en-us/azure/batch/batch-api-basics#virtual-network-vnet-and-firewall-configuration. + VirtualMachineImageID *string `json:"virtualMachineImageId,omitempty"` +} + +// InboundEndpoint ... +type InboundEndpoint struct { + Name *string `json:"name,omitempty"` + // Protocol - Possible values include: 'TCP', 'UDP' + Protocol InboundEndpointProtocol `json:"protocol,omitempty"` + PublicIPAddress *string `json:"publicIPAddress,omitempty"` + PublicFQDN *string `json:"publicFQDN,omitempty"` + FrontendPort *int32 `json:"frontendPort,omitempty"` + BackendPort *int32 `json:"backendPort,omitempty"` +} + +// InboundNATPool ... +type InboundNATPool struct { + // Name - The name must be unique within a Batch pool, can contain letters, numbers, underscores, periods, and hyphens. Names must start with a letter or number, must end with a letter, number, or underscore, and cannot exceed 77 characters. If any invalid values are provided the request fails with HTTP status code 400. + Name *string `json:"name,omitempty"` + // Protocol - Possible values include: 'TCP', 'UDP' + Protocol InboundEndpointProtocol `json:"protocol,omitempty"` + // BackendPort - This must be unique within a Batch pool. Acceptable values are between 1 and 65535 except for 22, 3389, 29876 and 29877 as these are reserved. If any reserved values are provided the request fails with HTTP status code 400. + BackendPort *int32 `json:"backendPort,omitempty"` + // FrontendPortRangeStart - Acceptable values range between 1 and 65534 except ports from 50000 to 55000 which are reserved. All ranges within a pool must be distinct and cannot overlap. Each range must contain at least 40 ports. If any reserved or overlapping values are provided the request fails with HTTP status code 400. + FrontendPortRangeStart *int32 `json:"frontendPortRangeStart,omitempty"` + // FrontendPortRangeEnd - Acceptable values range between 1 and 65534 except ports from 50000 to 55000 which are reserved by the Batch service. All ranges within a pool must be distinct and cannot overlap. Each range must contain at least 40 ports. If any reserved or overlapping values are provided the request fails with HTTP status code 400. + FrontendPortRangeEnd *int32 `json:"frontendPortRangeEnd,omitempty"` + // NetworkSecurityGroupRules - The maximum number of rules that can be specified across all the endpoints on a Batch pool is 25. If no network security group rules are specified, a default rule will be created to allow inbound access to the specified backendPort. If the maximum number of network security group rules is exceeded the request fails with HTTP status code 400. + NetworkSecurityGroupRules *[]NetworkSecurityGroupRule `json:"networkSecurityGroupRules,omitempty"` +} + +// JobAddParameter ... +type JobAddParameter struct { + // ID - The ID can contain any combination of alphanumeric characters including hyphens and underscores, and cannot contain more than 64 characters. The ID is case-preserving and case-insensitive (that is, you may not have two IDs within an account that differ only by case). + ID *string `json:"id,omitempty"` + // DisplayName - The display name need not be unique and can contain any Unicode characters up to a maximum length of 1024. + DisplayName *string `json:"displayName,omitempty"` + // Priority - Priority values can range from -1000 to 1000, with -1000 being the lowest priority and 1000 being the highest priority. The default value is 0. + Priority *int32 `json:"priority,omitempty"` + // Constraints - The execution constraints for the job. + Constraints *JobConstraints `json:"constraints,omitempty"` + // JobManagerTask - If the job does not specify a Job Manager task, the user must explicitly add tasks to the job. If the job does specify a Job Manager task, the Batch service creates the Job Manager task when the job is created, and will try to schedule the Job Manager task before scheduling other tasks in the job. The Job Manager task's typical purpose is to control and/or monitor job execution, for example by deciding what additional tasks to run, determining when the work is complete, etc. (However, a Job Manager task is not restricted to these activities - it is a fully-fledged task in the system and perform whatever actions are required for the job.) For example, a Job Manager task might download a file specified as a parameter, analyze the contents of that file and submit additional tasks based on those contents. + JobManagerTask *JobManagerTask `json:"jobManagerTask,omitempty"` + // JobPreparationTask - If a job has a Job Preparation task, the Batch service will run the Job Preparation task on a compute node before starting any tasks of that job on that compute node. + JobPreparationTask *JobPreparationTask `json:"jobPreparationTask,omitempty"` + // JobReleaseTask - A Job Release task cannot be specified without also specifying a Job Preparation task for the job. The Batch service runs the Job Release task on the compute nodes that have run the Job Preparation task. The primary purpose of the Job Release task is to undo changes to compute nodes made by the Job Preparation task. Example activities include deleting local files, or shutting down services that were started as part of job preparation. + JobReleaseTask *JobReleaseTask `json:"jobReleaseTask,omitempty"` + // CommonEnvironmentSettings - Individual tasks can override an environment setting specified here by specifying the same setting name with a different value. + CommonEnvironmentSettings *[]EnvironmentSetting `json:"commonEnvironmentSettings,omitempty"` + PoolInfo *PoolInformation `json:"poolInfo,omitempty"` + // OnAllTasksComplete - Note that if a job contains no tasks, then all tasks are considered complete. This option is therefore most commonly used with a Job Manager task; if you want to use automatic job termination without a Job Manager, you should initially set onAllTasksComplete to noaction and update the job properties to set onAllTasksComplete to terminatejob once you have finished adding tasks. The default is noaction. Possible values include: 'NoAction', 'TerminateJob' + OnAllTasksComplete OnAllTasksComplete `json:"onAllTasksComplete,omitempty"` + // OnTaskFailure - A task is considered to have failed if has a failureInfo. A failureInfo is set if the task completes with a non-zero exit code after exhausting its retry count, or if there was an error starting the task, for example due to a resource file download error. The default is noaction. Possible values include: 'OnTaskFailureNoAction', 'OnTaskFailurePerformExitOptionsJobAction' + OnTaskFailure OnTaskFailure `json:"onTaskFailure,omitempty"` + // Metadata - The Batch service does not assign any meaning to metadata; it is solely for the use of user code. + Metadata *[]MetadataItem `json:"metadata,omitempty"` + UsesTaskDependencies *bool `json:"usesTaskDependencies,omitempty"` +} + +// JobConstraints ... +type JobConstraints struct { + // MaxWallClockTime - If the job does not complete within the time limit, the Batch service terminates it and any tasks that are still running. In this case, the termination reason will be MaxWallClockTimeExpiry. If this property is not specified, there is no time limit on how long the job may run. + MaxWallClockTime *string `json:"maxWallClockTime,omitempty"` + // MaxTaskRetryCount - Note that this value specifically controls the number of retries. The Batch service will try each task once, and may then retry up to this limit. For example, if the maximum retry count is 3, Batch tries a task up to 4 times (one initial try and 3 retries). If the maximum retry count is 0, the Batch service does not retry tasks. If the maximum retry count is -1, the Batch service retries tasks without limit. The default value is 0 (no retries). + MaxTaskRetryCount *int32 `json:"maxTaskRetryCount,omitempty"` +} + +// JobDisableParameter ... +type JobDisableParameter struct { + // DisableTasks - Possible values include: 'DisableJobOptionRequeue', 'DisableJobOptionTerminate', 'DisableJobOptionWait' + DisableTasks DisableJobOption `json:"disableTasks,omitempty"` +} + +// JobExecutionInformation ... +type JobExecutionInformation struct { + // StartTime - This is the time at which the job was created. + StartTime *date.Time `json:"startTime,omitempty"` + // EndTime - This property is set only if the job is in the completed state. + EndTime *date.Time `json:"endTime,omitempty"` + // PoolID - This element contains the actual pool where the job is assigned. When you get job details from the service, they also contain a poolInfo element, which contains the pool configuration data from when the job was added or updated. That poolInfo element may also contain a poolId element. If it does, the two IDs are the same. If it does not, it means the job ran on an auto pool, and this property contains the ID of that auto pool. + PoolID *string `json:"poolId,omitempty"` + // SchedulingError - This property is not set if there was no error starting the job. + SchedulingError *JobSchedulingError `json:"schedulingError,omitempty"` + // TerminateReason - This property is set only if the job is in the completed state. If the Batch service terminates the job, it sets the reason as follows: JMComplete - the Job Manager task completed, and killJobOnCompletion was set to true. MaxWallClockTimeExpiry - the job reached its maxWallClockTime constraint. TerminateJobSchedule - the job ran as part of a schedule, and the schedule terminated. AllTasksComplete - the job's onAllTasksComplete attribute is set to terminatejob, and all tasks in the job are complete. TaskFailed - the job's onTaskFailure attribute is set to performExitOptionsJobAction, and a task in the job failed with an exit condition that specified a jobAction of terminatejob. Any other string is a user-defined reason specified in a call to the 'Terminate a job' operation. + TerminateReason *string `json:"terminateReason,omitempty"` +} + +// JobManagerTask the Job Manager task is automatically started when the job is created. The Batch service tries to +// schedule the Job Manager task before any other tasks in the job. When shrinking a pool, the Batch service tries +// to preserve compute nodes where Job Manager tasks are running for as long as possible (that is, nodes running +// 'normal' tasks are removed before nodes running Job Manager tasks). When a Job Manager task fails and needs to +// be restarted, the system tries to schedule it at the highest priority. If there are no idle nodes available, the +// system may terminate one of the running tasks in the pool and return it to the queue in order to make room for +// the Job Manager task to restart. Note that a Job Manager task in one job does not have priority over tasks in +// other jobs. Across jobs, only job level priorities are observed. For example, if a Job Manager in a priority 0 +// job needs to be restarted, it will not displace tasks of a priority 1 job. +type JobManagerTask struct { + // ID - The ID can contain any combination of alphanumeric characters including hyphens and underscores and cannot contain more than 64 characters. + ID *string `json:"id,omitempty"` + // DisplayName - It need not be unique and can contain any Unicode characters up to a maximum length of 1024. + DisplayName *string `json:"displayName,omitempty"` + // CommandLine - The command line does not run under a shell, and therefore cannot take advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows or "/bin/sh -c MyCommand" in Linux. + CommandLine *string `json:"commandLine,omitempty"` + // ContainerSettings - If the pool that will run this task has containerConfiguration set, this must be set as well. If the pool that will run this task doesn't have containerConfiguration set, this must not be set. When this is specified, all directories recursively below the AZ_BATCH_NODE_ROOT_DIR (the root of Azure Batch directories on the node) are mapped into the container, all task environment variables are mapped into the container, and the task command line is executed in the container. + ContainerSettings *TaskContainerSettings `json:"containerSettings,omitempty"` + // ResourceFiles - Files listed under this element are located in the task's working directory. + ResourceFiles *[]ResourceFile `json:"resourceFiles,omitempty"` + // OutputFiles - For multi-instance tasks, the files will only be uploaded from the compute node on which the primary task is executed. + OutputFiles *[]OutputFile `json:"outputFiles,omitempty"` + EnvironmentSettings *[]EnvironmentSetting `json:"environmentSettings,omitempty"` + Constraints *TaskConstraints `json:"constraints,omitempty"` + // KillJobOnCompletion - If true, when the Job Manager task completes, the Batch service marks the job as complete. If any tasks are still running at this time (other than Job Release), those tasks are terminated. If false, the completion of the Job Manager task does not affect the job status. In this case, you should either use the onAllTasksComplete attribute to terminate the job, or have a client or user terminate the job explicitly. An example of this is if the Job Manager creates a set of tasks but then takes no further role in their execution. The default value is true. If you are using the onAllTasksComplete and onTaskFailure attributes to control job lifetime, and using the Job Manager task only to create the tasks for the job (not to monitor progress), then it is important to set killJobOnCompletion to false. + KillJobOnCompletion *bool `json:"killJobOnCompletion,omitempty"` + // UserIdentity - If omitted, the task runs as a non-administrative user unique to the task. + UserIdentity *UserIdentity `json:"userIdentity,omitempty"` + // RunExclusive - If true, no other tasks will run on the same compute node for as long as the Job Manager is running. If false, other tasks can run simultaneously with the Job Manager on a compute node. The Job Manager task counts normally against the node's concurrent task limit, so this is only relevant if the node allows multiple concurrent tasks. The default value is true. + RunExclusive *bool `json:"runExclusive,omitempty"` + // ApplicationPackageReferences - Application packages are downloaded and deployed to a shared directory, not the task working directory. Therefore, if a referenced package is already on the compute node, and is up to date, then it is not re-downloaded; the existing copy on the compute node is used. If a referenced application package cannot be installed, for example because the package has been deleted or because download failed, the task fails. + ApplicationPackageReferences *[]ApplicationPackageReference `json:"applicationPackageReferences,omitempty"` + // AuthenticationTokenSettings - If this property is set, the Batch service provides the task with an authentication token which can be used to authenticate Batch service operations without requiring an account access key. The token is provided via the AZ_BATCH_AUTHENTICATION_TOKEN environment variable. The operations that the task can carry out using the token depend on the settings. For example, a task can request job permissions in order to add other tasks to the job, or check the status of the job or of other tasks under the job. + AuthenticationTokenSettings *AuthenticationTokenSettings `json:"authenticationTokenSettings,omitempty"` + // AllowLowPriorityNode - The default value is false. + AllowLowPriorityNode *bool `json:"allowLowPriorityNode,omitempty"` +} + +// JobPatchParameter ... +type JobPatchParameter struct { + // Priority - Priority values can range from -1000 to 1000, with -1000 being the lowest priority and 1000 being the highest priority. If omitted, the priority of the job is left unchanged. + Priority *int32 `json:"priority,omitempty"` + // OnAllTasksComplete - If omitted, the completion behavior is left unchanged. You may not change the value from terminatejob to noaction - that is, once you have engaged automatic job termination, you cannot turn it off again. If you try to do this, the request fails with an 'invalid property value' error response; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). Possible values include: 'NoAction', 'TerminateJob' + OnAllTasksComplete OnAllTasksComplete `json:"onAllTasksComplete,omitempty"` + // Constraints - If omitted, the existing execution constraints are left unchanged. + Constraints *JobConstraints `json:"constraints,omitempty"` + // PoolInfo - You may change the pool for a job only when the job is disabled. The Patch Job call will fail if you include the poolInfo element and the job is not disabled. If you specify an autoPoolSpecification specification in the poolInfo, only the keepAlive property can be updated, and then only if the auto pool has a poolLifetimeOption of job. If omitted, the job continues to run on its current pool. + PoolInfo *PoolInformation `json:"poolInfo,omitempty"` + // Metadata - If omitted, the existing job metadata is left unchanged. + Metadata *[]MetadataItem `json:"metadata,omitempty"` +} + +// JobPreparationAndReleaseTaskExecutionInformation ... +type JobPreparationAndReleaseTaskExecutionInformation struct { + PoolID *string `json:"poolId,omitempty"` + NodeID *string `json:"nodeId,omitempty"` + NodeURL *string `json:"nodeUrl,omitempty"` + JobPreparationTaskExecutionInfo *JobPreparationTaskExecutionInformation `json:"jobPreparationTaskExecutionInfo,omitempty"` + // JobReleaseTaskExecutionInfo - This property is set only if the Job Release task has run on the node. + JobReleaseTaskExecutionInfo *JobReleaseTaskExecutionInformation `json:"jobReleaseTaskExecutionInfo,omitempty"` +} + +// JobPreparationTask you can use Job Preparation to prepare a compute node to run tasks for the job. Activities +// commonly performed in Job Preparation include: Downloading common resource files used by all the tasks in the +// job. The Job Preparation task can download these common resource files to the shared location on the compute +// node. (AZ_BATCH_NODE_ROOT_DIR\shared), or starting a local service on the compute node so that all tasks of that +// job can communicate with it. If the Job Preparation task fails (that is, exhausts its retry count before exiting +// with exit code 0), Batch will not run tasks of this job on the compute node. The node remains ineligible to run +// tasks of this job until it is reimaged. The node remains active and can be used for other jobs. The Job +// Preparation task can run multiple times on the same compute node. Therefore, you should write the Job +// Preparation task to handle re-execution. If the compute node is rebooted, the Job Preparation task is run again +// on the node before scheduling any other task of the job, if rerunOnNodeRebootAfterSuccess is true or if the Job +// Preparation task did not previously complete. If the compute node is reimaged, the Job Preparation task is run +// again before scheduling any task of the job. +type JobPreparationTask struct { + // ID - The ID can contain any combination of alphanumeric characters including hyphens and underscores and cannot contain more than 64 characters. If you do not specify this property, the Batch service assigns a default value of 'jobpreparation'. No other task in the job can have the same ID as the Job Preparation task. If you try to submit a task with the same id, the Batch service rejects the request with error code TaskIdSameAsJobPreparationTask; if you are calling the REST API directly, the HTTP status code is 409 (Conflict). + ID *string `json:"id,omitempty"` + // CommandLine - The command line does not run under a shell, and therefore cannot take advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows or "/bin/sh -c MyCommand" in Linux. + CommandLine *string `json:"commandLine,omitempty"` + // ContainerSettings - When this is specified, all directories recursively below the AZ_BATCH_NODE_ROOT_DIR (the root of Azure Batch directories on the node) are mapped into the container, all task environment variables are mapped into the container, and the task command line is executed in the container. + ContainerSettings *TaskContainerSettings `json:"containerSettings,omitempty"` + // ResourceFiles - Files listed under this element are located in the task's working directory. + ResourceFiles *[]ResourceFile `json:"resourceFiles,omitempty"` + EnvironmentSettings *[]EnvironmentSetting `json:"environmentSettings,omitempty"` + Constraints *TaskConstraints `json:"constraints,omitempty"` + // WaitForSuccess - If true and the Job Preparation task fails on a compute node, the Batch service retries the Job Preparation task up to its maximum retry count (as specified in the constraints element). If the task has still not completed successfully after all retries, then the Batch service will not schedule tasks of the job to the compute node. The compute node remains active and eligible to run tasks of other jobs. If false, the Batch service will not wait for the Job Preparation task to complete. In this case, other tasks of the job can start executing on the compute node while the Job Preparation task is still running; and even if the Job Preparation task fails, new tasks will continue to be scheduled on the node. The default value is true. + WaitForSuccess *bool `json:"waitForSuccess,omitempty"` + // UserIdentity - If omitted, the task runs as a non-administrative user unique to the task on Windows nodes, or a a non-administrative user unique to the pool on Linux nodes. + UserIdentity *UserIdentity `json:"userIdentity,omitempty"` + // RerunOnNodeRebootAfterSuccess - The Job Preparation task is always rerun if a compute node is reimaged, or if the Job Preparation task did not complete (e.g. because the reboot occurred while the task was running). Therefore, you should always write a Job Preparation task to be idempotent and to behave correctly if run multiple times. The default value is true. + RerunOnNodeRebootAfterSuccess *bool `json:"rerunOnNodeRebootAfterSuccess,omitempty"` +} + +// JobPreparationTaskExecutionInformation ... +type JobPreparationTaskExecutionInformation struct { + // StartTime - If the task has been restarted or retried, this is the most recent time at which the task started running. + StartTime *date.Time `json:"startTime,omitempty"` + // EndTime - This property is set only if the task is in the Completed state. + EndTime *date.Time `json:"endTime,omitempty"` + // State - Possible values include: 'JobPreparationTaskStateRunning', 'JobPreparationTaskStateCompleted' + State JobPreparationTaskState `json:"state,omitempty"` + TaskRootDirectory *string `json:"taskRootDirectory,omitempty"` + TaskRootDirectoryURL *string `json:"taskRootDirectoryUrl,omitempty"` + // ExitCode - This parameter is returned only if the task is in the completed state. The exit code for a process reflects the specific convention implemented by the application developer for that process. If you use the exit code value to make decisions in your code, be sure that you know the exit code convention used by the application process. Note that the exit code may also be generated by the compute node operating system, such as when a process is forcibly terminated. + ExitCode *int32 `json:"exitCode,omitempty"` + // ContainerInfo - This property is set only if the task runs in a container context. + ContainerInfo *TaskContainerExecutionInformation `json:"containerInfo,omitempty"` + // FailureInfo - This property is set only if the task is in the completed state and encountered a failure. + FailureInfo *TaskFailureInformation `json:"failureInfo,omitempty"` + // RetryCount - Task application failures (non-zero exit code) are retried, pre-processing errors (the task could not be run) and file upload errors are not retried. The Batch service will retry the task up to the limit specified by the constraints. + RetryCount *int32 `json:"retryCount,omitempty"` + // LastRetryTime - This property is set only if the task was retried (i.e. retryCount is nonzero). If present, this is typically the same as startTime, but may be different if the task has been restarted for reasons other than retry; for example, if the compute node was rebooted during a retry, then the startTime is updated but the lastRetryTime is not. + LastRetryTime *date.Time `json:"lastRetryTime,omitempty"` + // Result - If the value is 'failed', then the details of the failure can be found in the failureInfo property. Possible values include: 'Success', 'Failure' + Result TaskExecutionResult `json:"result,omitempty"` +} + +// JobReleaseTask the Job Release task runs when the job ends, because of one of the following: The user calls the +// Terminate Job API, or the Delete Job API while the job is still active, the job's maximum wall clock time +// constraint is reached, and the job is still active, or the job's Job Manager task completed, and the job is +// configured to terminate when the Job Manager completes. The Job Release task runs on each compute node where +// tasks of the job have run and the Job Preparation task ran and completed. If you reimage a compute node after it +// has run the Job Preparation task, and the job ends without any further tasks of the job running on that compute +// node (and hence the Job Preparation task does not re-run), then the Job Release task does not run on that node. +// If a compute node reboots while the Job Release task is still running, the Job Release task runs again when the +// compute node starts up. The job is not marked as complete until all Job Release tasks have completed. The Job +// Release task runs in the background. It does not occupy a scheduling slot; that is, it does not count towards +// the maxTasksPerNode limit specified on the pool. +type JobReleaseTask struct { + // ID - The ID can contain any combination of alphanumeric characters including hyphens and underscores and cannot contain more than 64 characters. If you do not specify this property, the Batch service assigns a default value of 'jobrelease'. No other task in the job can have the same ID as the Job Release task. If you try to submit a task with the same id, the Batch service rejects the request with error code TaskIdSameAsJobReleaseTask; if you are calling the REST API directly, the HTTP status code is 409 (Conflict). + ID *string `json:"id,omitempty"` + // CommandLine - The command line does not run under a shell, and therefore cannot take advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows or "/bin/sh -c MyCommand" in Linux. + CommandLine *string `json:"commandLine,omitempty"` + // ContainerSettings - When this is specified, all directories recursively below the AZ_BATCH_NODE_ROOT_DIR (the root of Azure Batch directories on the node) are mapped into the container, all task environment variables are mapped into the container, and the task command line is executed in the container. + ContainerSettings *TaskContainerSettings `json:"containerSettings,omitempty"` + // ResourceFiles - Files listed under this element are located in the task's working directory. + ResourceFiles *[]ResourceFile `json:"resourceFiles,omitempty"` + EnvironmentSettings *[]EnvironmentSetting `json:"environmentSettings,omitempty"` + MaxWallClockTime *string `json:"maxWallClockTime,omitempty"` + // RetentionTime - The default is infinite, i.e. the task directory will be retained until the compute node is removed or reimaged. + RetentionTime *string `json:"retentionTime,omitempty"` + // UserIdentity - If omitted, the task runs as a non-administrative user unique to the task. + UserIdentity *UserIdentity `json:"userIdentity,omitempty"` +} + +// JobReleaseTaskExecutionInformation ... +type JobReleaseTaskExecutionInformation struct { + // StartTime - If the task has been restarted or retried, this is the most recent time at which the task started running. + StartTime *date.Time `json:"startTime,omitempty"` + // EndTime - This property is set only if the task is in the Completed state. + EndTime *date.Time `json:"endTime,omitempty"` + // State - Possible values include: 'JobReleaseTaskStateRunning', 'JobReleaseTaskStateCompleted' + State JobReleaseTaskState `json:"state,omitempty"` + TaskRootDirectory *string `json:"taskRootDirectory,omitempty"` + TaskRootDirectoryURL *string `json:"taskRootDirectoryUrl,omitempty"` + // ExitCode - This parameter is returned only if the task is in the completed state. The exit code for a process reflects the specific convention implemented by the application developer for that process. If you use the exit code value to make decisions in your code, be sure that you know the exit code convention used by the application process. Note that the exit code may also be generated by the compute node operating system, such as when a process is forcibly terminated. + ExitCode *int32 `json:"exitCode,omitempty"` + // ContainerInfo - This property is set only if the task runs in a container context. + ContainerInfo *TaskContainerExecutionInformation `json:"containerInfo,omitempty"` + // FailureInfo - This property is set only if the task is in the completed state and encountered a failure. + FailureInfo *TaskFailureInformation `json:"failureInfo,omitempty"` + // Result - If the value is 'failed', then the details of the failure can be found in the failureInfo property. Possible values include: 'Success', 'Failure' + Result TaskExecutionResult `json:"result,omitempty"` +} + +// JobScheduleAddParameter ... +type JobScheduleAddParameter struct { + // ID - The ID can contain any combination of alphanumeric characters including hyphens and underscores, and cannot contain more than 64 characters. The ID is case-preserving and case-insensitive (that is, you may not have two IDs within an account that differ only by case). + ID *string `json:"id,omitempty"` + // DisplayName - The display name need not be unique and can contain any Unicode characters up to a maximum length of 1024. + DisplayName *string `json:"displayName,omitempty"` + Schedule *Schedule `json:"schedule,omitempty"` + JobSpecification *JobSpecification `json:"jobSpecification,omitempty"` + // Metadata - The Batch service does not assign any meaning to metadata; it is solely for the use of user code. + Metadata *[]MetadataItem `json:"metadata,omitempty"` +} + +// JobScheduleExecutionInformation ... +type JobScheduleExecutionInformation struct { + // NextRunTime - This property is meaningful only if the schedule is in the active state when the time comes around. For example, if the schedule is disabled, no job will be created at nextRunTime unless the job is enabled before then. + NextRunTime *date.Time `json:"nextRunTime,omitempty"` + // RecentJob - This property is present only if the at least one job has run under the schedule. + RecentJob *RecentJob `json:"recentJob,omitempty"` + // EndTime - This property is set only if the job schedule is in the completed state. + EndTime *date.Time `json:"endTime,omitempty"` +} + +// JobSchedulePatchParameter ... +type JobSchedulePatchParameter struct { + // Schedule - If you do not specify this element, the existing schedule is left unchanged. + Schedule *Schedule `json:"schedule,omitempty"` + // JobSpecification - Updates affect only jobs that are started after the update has taken place. Any currently active job continues with the older specification. + JobSpecification *JobSpecification `json:"jobSpecification,omitempty"` + // Metadata - If you do not specify this element, existing metadata is left unchanged. + Metadata *[]MetadataItem `json:"metadata,omitempty"` +} + +// JobScheduleStatistics ... +type JobScheduleStatistics struct { + URL *string `json:"url,omitempty"` + StartTime *date.Time `json:"startTime,omitempty"` + LastUpdateTime *date.Time `json:"lastUpdateTime,omitempty"` + UserCPUTime *string `json:"userCPUTime,omitempty"` + KernelCPUTime *string `json:"kernelCPUTime,omitempty"` + // WallClockTime - The wall clock time is the elapsed time from when the task started running on a compute node to when it finished (or to the last time the statistics were updated, if the task had not finished by then). If a task was retried, this includes the wall clock time of all the task retries. + WallClockTime *string `json:"wallClockTime,omitempty"` + ReadIOps *int64 `json:"readIOps,omitempty"` + WriteIOps *int64 `json:"writeIOps,omitempty"` + ReadIOGiB *float64 `json:"readIOGiB,omitempty"` + WriteIOGiB *float64 `json:"writeIOGiB,omitempty"` + NumSucceededTasks *int64 `json:"numSucceededTasks,omitempty"` + NumFailedTasks *int64 `json:"numFailedTasks,omitempty"` + NumTaskRetries *int64 `json:"numTaskRetries,omitempty"` + // WaitTime - This value is only reported in the account lifetime statistics; it is not included in the job statistics. + WaitTime *string `json:"waitTime,omitempty"` +} + +// JobScheduleUpdateParameter ... +type JobScheduleUpdateParameter struct { + // Schedule - If you do not specify this element, it is equivalent to passing the default schedule: that is, a single job scheduled to run immediately. + Schedule *Schedule `json:"schedule,omitempty"` + // JobSpecification - Updates affect only jobs that are started after the update has taken place. Any currently active job continues with the older specification. + JobSpecification *JobSpecification `json:"jobSpecification,omitempty"` + // Metadata - If you do not specify this element, it takes the default value of an empty list; in effect, any existing metadata is deleted. + Metadata *[]MetadataItem `json:"metadata,omitempty"` +} + +// JobSchedulingError ... +type JobSchedulingError struct { + // Category - Possible values include: 'UserError', 'ServerError' + Category ErrorCategory `json:"category,omitempty"` + Code *string `json:"code,omitempty"` + Message *string `json:"message,omitempty"` + Details *[]NameValuePair `json:"details,omitempty"` +} + +// JobSpecification ... +type JobSpecification struct { + // Priority - Priority values can range from -1000 to 1000, with -1000 being the lowest priority and 1000 being the highest priority. The default value is 0. This priority is used as the default for all jobs under the job schedule. You can update a job's priority after it has been created using by using the update job API. + Priority *int32 `json:"priority,omitempty"` + // DisplayName - The name need not be unique and can contain any Unicode characters up to a maximum length of 1024. + DisplayName *string `json:"displayName,omitempty"` + UsesTaskDependencies *bool `json:"usesTaskDependencies,omitempty"` + // OnAllTasksComplete - Note that if a job contains no tasks, then all tasks are considered complete. This option is therefore most commonly used with a Job Manager task; if you want to use automatic job termination without a Job Manager, you should initially set onAllTasksComplete to noaction and update the job properties to set onAllTasksComplete to terminatejob once you have finished adding tasks. The default is noaction. Possible values include: 'NoAction', 'TerminateJob' + OnAllTasksComplete OnAllTasksComplete `json:"onAllTasksComplete,omitempty"` + // OnTaskFailure - The default is noaction. Possible values include: 'OnTaskFailureNoAction', 'OnTaskFailurePerformExitOptionsJobAction' + OnTaskFailure OnTaskFailure `json:"onTaskFailure,omitempty"` + Constraints *JobConstraints `json:"constraints,omitempty"` + // JobManagerTask - If the job does not specify a Job Manager task, the user must explicitly add tasks to the job using the Task API. If the job does specify a Job Manager task, the Batch service creates the Job Manager task when the job is created, and will try to schedule the Job Manager task before scheduling other tasks in the job. + JobManagerTask *JobManagerTask `json:"jobManagerTask,omitempty"` + // JobPreparationTask - If a job has a Job Preparation task, the Batch service will run the Job Preparation task on a compute node before starting any tasks of that job on that compute node. + JobPreparationTask *JobPreparationTask `json:"jobPreparationTask,omitempty"` + // JobReleaseTask - The primary purpose of the Job Release task is to undo changes to compute nodes made by the Job Preparation task. Example activities include deleting local files, or shutting down services that were started as part of job preparation. A Job Release task cannot be specified without also specifying a Job Preparation task for the job. The Batch service runs the Job Release task on the compute nodes that have run the Job Preparation task. + JobReleaseTask *JobReleaseTask `json:"jobReleaseTask,omitempty"` + // CommonEnvironmentSettings - Individual tasks can override an environment setting specified here by specifying the same setting name with a different value. + CommonEnvironmentSettings *[]EnvironmentSetting `json:"commonEnvironmentSettings,omitempty"` + PoolInfo *PoolInformation `json:"poolInfo,omitempty"` + // Metadata - The Batch service does not assign any meaning to metadata; it is solely for the use of user code. + Metadata *[]MetadataItem `json:"metadata,omitempty"` +} + +// JobStatistics ... +type JobStatistics struct { + autorest.Response `json:"-"` + URL *string `json:"url,omitempty"` + StartTime *date.Time `json:"startTime,omitempty"` + LastUpdateTime *date.Time `json:"lastUpdateTime,omitempty"` + UserCPUTime *string `json:"userCPUTime,omitempty"` + KernelCPUTime *string `json:"kernelCPUTime,omitempty"` + // WallClockTime - The wall clock time is the elapsed time from when the task started running on a compute node to when it finished (or to the last time the statistics were updated, if the task had not finished by then). If a task was retried, this includes the wall clock time of all the task retries. + WallClockTime *string `json:"wallClockTime,omitempty"` + ReadIOps *int64 `json:"readIOps,omitempty"` + WriteIOps *int64 `json:"writeIOps,omitempty"` + ReadIOGiB *float64 `json:"readIOGiB,omitempty"` + WriteIOGiB *float64 `json:"writeIOGiB,omitempty"` + // NumSucceededTasks - A task completes successfully if it returns exit code 0. + NumSucceededTasks *int64 `json:"numSucceededTasks,omitempty"` + // NumFailedTasks - A task fails if it exhausts its maximum retry count without returning exit code 0. + NumFailedTasks *int64 `json:"numFailedTasks,omitempty"` + NumTaskRetries *int64 `json:"numTaskRetries,omitempty"` + // WaitTime - The wait time for a task is defined as the elapsed time between the creation of the task and the start of task execution. (If the task is retried due to failures, the wait time is the time to the most recent task execution.) This value is only reported in the account lifetime statistics; it is not included in the job statistics. + WaitTime *string `json:"waitTime,omitempty"` +} + +// JobTerminateParameter ... +type JobTerminateParameter struct { + TerminateReason *string `json:"terminateReason,omitempty"` +} + +// JobUpdateParameter ... +type JobUpdateParameter struct { + // Priority - Priority values can range from -1000 to 1000, with -1000 being the lowest priority and 1000 being the highest priority. If omitted, it is set to the default value 0. + Priority *int32 `json:"priority,omitempty"` + // Constraints - If omitted, the constraints are cleared. + Constraints *JobConstraints `json:"constraints,omitempty"` + // PoolInfo - You may change the pool for a job only when the job is disabled. The Update Job call will fail if you include the poolInfo element and the job is not disabled. If you specify an autoPoolSpecification specification in the poolInfo, only the keepAlive property can be updated, and then only if the auto pool has a poolLifetimeOption of job. + PoolInfo *PoolInformation `json:"poolInfo,omitempty"` + // Metadata - If omitted, it takes the default value of an empty list; in effect, any existing metadata is deleted. + Metadata *[]MetadataItem `json:"metadata,omitempty"` + // OnAllTasksComplete - If omitted, the completion behavior is set to noaction. If the current value is terminatejob, this is an error because a job's completion behavior may not be changed from terminatejob to noaction. You may not change the value from terminatejob to noaction - that is, once you have engaged automatic job termination, you cannot turn it off again. If you try to do this, the request fails and Batch returns status code 400 (Bad Request) and an 'invalid property value' error response. If you do not specify this element in a PUT request, it is equivalent to passing noaction. This is an error if the current value is terminatejob. Possible values include: 'NoAction', 'TerminateJob' + OnAllTasksComplete OnAllTasksComplete `json:"onAllTasksComplete,omitempty"` +} + +// LinuxUserConfiguration ... +type LinuxUserConfiguration struct { + // UID - The uid and gid properties must be specified together or not at all. If not specified the underlying operating system picks the uid. + UID *int32 `json:"uid,omitempty"` + // Gid - The uid and gid properties must be specified together or not at all. If not specified the underlying operating system picks the gid. + Gid *int32 `json:"gid,omitempty"` + // SSHPrivateKey - The private key must not be password protected. The private key is used to automatically configure asymmetric-key based authentication for SSH between nodes in a Linux pool when the pool's enableInterNodeCommunication property is true (it is ignored if enableInterNodeCommunication is false). It does this by placing the key pair into the user's .ssh directory. If not specified, password-less SSH is not configured between nodes (no modification of the user's .ssh directory is done). + SSHPrivateKey *string `json:"sshPrivateKey,omitempty"` +} + +// MetadataItem the Batch service does not assign any meaning to this metadata; it is solely for the use of user +// code. +type MetadataItem struct { + Name *string `json:"name,omitempty"` + Value *string `json:"value,omitempty"` +} + +// MultiInstanceSettings multi-instance tasks are commonly used to support MPI tasks. +type MultiInstanceSettings struct { + // NumberOfInstances - If omitted, the default is 1. + NumberOfInstances *int32 `json:"numberOfInstances,omitempty"` + // CoordinationCommandLine - A typical coordination command line launches a background service and verifies that the service is ready to process inter-node messages. + CoordinationCommandLine *string `json:"coordinationCommandLine,omitempty"` + // CommonResourceFiles - The difference between common resource files and task resource files is that common resource files are downloaded for all subtasks including the primary, whereas task resource files are downloaded only for the primary. Also note that these resource files are not downloaded to the task working directory, but instead are downloaded to the task root directory (one directory above the working directory). + CommonResourceFiles *[]ResourceFile `json:"commonResourceFiles,omitempty"` +} + +// NameValuePair ... +type NameValuePair struct { + Name *string `json:"name,omitempty"` + Value *string `json:"value,omitempty"` +} + +// NetworkConfiguration the network configuration for a pool. +type NetworkConfiguration struct { + // SubnetID - The virtual network must be in the same region and subscription as the Azure Batch account. The specified subnet should have enough free IP addresses to accommodate the number of nodes in the pool. If the subnet doesn't have enough free IP addresses, the pool will partially allocate compute nodes, and a resize error will occur. The 'MicrosoftAzureBatch' service principal must have the 'Classic Virtual Machine Contributor' Role-Based Access Control (RBAC) role for the specified VNet. The specified subnet must allow communication from the Azure Batch service to be able to schedule tasks on the compute nodes. This can be verified by checking if the specified VNet has any associated Network Security Groups (NSG). If communication to the compute nodes in the specified subnet is denied by an NSG, then the Batch service will set the state of the compute nodes to unusable. For pools created with virtualMachineConfiguration only ARM virtual networks ('Microsoft.Network/virtualNetworks') are supported, but for pools created with cloudServiceConfiguration both ARM and classic virtual networks are supported. If the specified VNet has any associated Network Security Groups (NSG), then a few reserved system ports must be enabled for inbound communication. For pools created with a virtual machine configuration, enable ports 29876 and 29877, as well as port 22 for Linux and port 3389 for Windows. For pools created with a cloud service configuration, enable ports 10100, 20100, and 30100. Also enable outbound connections to Azure Storage on port 443. For more details see: https://docs.microsoft.com/en-us/azure/batch/batch-api-basics#virtual-network-vnet-and-firewall-configuration + SubnetID *string `json:"subnetId,omitempty"` + // EndpointConfiguration - Pool endpoint configuration is only supported on pools with the virtualMachineConfiguration property. + EndpointConfiguration *PoolEndpointConfiguration `json:"endpointConfiguration,omitempty"` +} + +// NetworkSecurityGroupRule ... +type NetworkSecurityGroupRule struct { + // Priority - Priorities within a pool must be unique and are evaluated in order of priority. The lower the number the higher the priority. For example, rules could be specified with order numbers of 150, 250, and 350. The rule with the order number of 150 takes precedence over the rule that has an order of 250. Allowed priorities are 150 to 3500. If any reserved or duplicate values are provided the request fails with HTTP status code 400. + Priority *int32 `json:"priority,omitempty"` + // Access - Possible values include: 'Allow', 'Deny' + Access NetworkSecurityGroupRuleAccess `json:"access,omitempty"` + // SourceAddressPrefix - Valid values are a single IP address (i.e. 10.10.10.10), IP subnet (i.e. 192.168.1.0/24), default tag, or * (for all addresses). If any other values are provided the request fails with HTTP status code 400. + SourceAddressPrefix *string `json:"sourceAddressPrefix,omitempty"` +} + +// NodeAgentSku the Batch node agent is a program that runs on each node in the pool, and provides the +// command-and-control interface between the node and the Batch service. There are different implementations of the +// node agent, known as SKUs, for different operating systems. +type NodeAgentSku struct { + ID *string `json:"id,omitempty"` + // VerifiedImageReferences - This collection is not exhaustive (the node agent may be compatible with other images). + VerifiedImageReferences *[]ImageReference `json:"verifiedImageReferences,omitempty"` + // OsType - Possible values include: 'Linux', 'Windows' + OsType OSType `json:"osType,omitempty"` +} + +// NodeDisableSchedulingParameter ... +type NodeDisableSchedulingParameter struct { + // NodeDisableSchedulingOption - The default value is requeue. Possible values include: 'DisableComputeNodeSchedulingOptionRequeue', 'DisableComputeNodeSchedulingOptionTerminate', 'DisableComputeNodeSchedulingOptionTaskCompletion' + NodeDisableSchedulingOption DisableComputeNodeSchedulingOption `json:"nodeDisableSchedulingOption,omitempty"` +} + +// NodeFile ... +type NodeFile struct { + Name *string `json:"name,omitempty"` + URL *string `json:"url,omitempty"` + IsDirectory *bool `json:"isDirectory,omitempty"` + Properties *FileProperties `json:"properties,omitempty"` +} + +// NodeFileListResult ... +type NodeFileListResult struct { + autorest.Response `json:"-"` + Value *[]NodeFile `json:"value,omitempty"` + OdataNextLink *string `json:"odata.nextLink,omitempty"` +} + +// NodeFileListResultIterator provides access to a complete listing of NodeFile values. +type NodeFileListResultIterator struct { + i int + page NodeFileListResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *NodeFileListResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter NodeFileListResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter NodeFileListResultIterator) Response() NodeFileListResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter NodeFileListResultIterator) Value() NodeFile { + if !iter.page.NotDone() { + return NodeFile{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (nflr NodeFileListResult) IsEmpty() bool { + return nflr.Value == nil || len(*nflr.Value) == 0 +} + +// nodeFileListResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (nflr NodeFileListResult) nodeFileListResultPreparer() (*http.Request, error) { + if nflr.OdataNextLink == nil || len(to.String(nflr.OdataNextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(nflr.OdataNextLink))) +} + +// NodeFileListResultPage contains a page of NodeFile values. +type NodeFileListResultPage struct { + fn func(NodeFileListResult) (NodeFileListResult, error) + nflr NodeFileListResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *NodeFileListResultPage) Next() error { + next, err := page.fn(page.nflr) + if err != nil { + return err + } + page.nflr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page NodeFileListResultPage) NotDone() bool { + return !page.nflr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page NodeFileListResultPage) Response() NodeFileListResult { + return page.nflr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page NodeFileListResultPage) Values() []NodeFile { + if page.nflr.IsEmpty() { + return nil + } + return *page.nflr.Value +} + +// NodeRebootParameter ... +type NodeRebootParameter struct { + // NodeRebootOption - The default value is requeue. Possible values include: 'ComputeNodeRebootOptionRequeue', 'ComputeNodeRebootOptionTerminate', 'ComputeNodeRebootOptionTaskCompletion', 'ComputeNodeRebootOptionRetainedData' + NodeRebootOption ComputeNodeRebootOption `json:"nodeRebootOption,omitempty"` +} + +// NodeReimageParameter ... +type NodeReimageParameter struct { + // NodeReimageOption - The default value is requeue. Possible values include: 'ComputeNodeReimageOptionRequeue', 'ComputeNodeReimageOptionTerminate', 'ComputeNodeReimageOptionTaskCompletion', 'ComputeNodeReimageOptionRetainedData' + NodeReimageOption ComputeNodeReimageOption `json:"nodeReimageOption,omitempty"` +} + +// NodeRemoveParameter ... +type NodeRemoveParameter struct { + NodeList *[]string `json:"nodeList,omitempty"` + // ResizeTimeout - The default value is 15 minutes. The minimum value is 5 minutes. If you specify a value less than 5 minutes, the Batch service returns an error; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). + ResizeTimeout *string `json:"resizeTimeout,omitempty"` + // NodeDeallocationOption - The default value is requeue. Possible values include: 'Requeue', 'Terminate', 'TaskCompletion', 'RetainedData' + NodeDeallocationOption ComputeNodeDeallocationOption `json:"nodeDeallocationOption,omitempty"` +} + +// NodeUpdateUserParameter ... +type NodeUpdateUserParameter struct { + // Password - The password is required for Windows nodes (those created with 'cloudServiceConfiguration', or created with 'virtualMachineConfiguration' using a Windows image reference). For Linux compute nodes, the password can optionally be specified along with the sshPublicKey property. If omitted, any existing password is removed. + Password *string `json:"password,omitempty"` + // ExpiryTime - If omitted, the default is 1 day from the current time. For Linux compute nodes, the expiryTime has a precision up to a day. + ExpiryTime *date.Time `json:"expiryTime,omitempty"` + // SSHPublicKey - The public key should be compatible with OpenSSH encoding and should be base 64 encoded. This property can be specified only for Linux nodes. If this is specified for a Windows node, then the Batch service rejects the request; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). If omitted, any existing SSH public key is removed. + SSHPublicKey *string `json:"sshPublicKey,omitempty"` +} + +// OSDisk ... +type OSDisk struct { + // Caching - The default value for caching is none. For information about the caching options see: https://blogs.msdn.microsoft.com/windowsazurestorage/2012/06/27/exploring-windows-azure-drives-disks-and-images/. Possible values include: 'None', 'ReadOnly', 'ReadWrite' + Caching CachingType `json:"caching,omitempty"` +} + +// OutputFile ... +type OutputFile struct { + // FilePattern - Both relative and absolute paths are supported. Relative paths are relative to the task working directory. The following wildcards are supported: * matches 0 or more characters (for example pattern abc* would match abc or abcdef), ** matches any directory, ? matches any single character, [abc] matches one character in the brackets, and [a-c] matches one character in the range. Brackets can include a negation to match any character not specified (for example [!abc] matches any character but a, b, or c). If a file name starts with "." it is ignored by default but may be matched by specifying it explicitly (for example *.gif will not match .a.gif, but .*.gif will). A simple example: **\*.txt matches any file that does not start in '.' and ends with .txt in the task working directory or any subdirectory. If the filename contains a wildcard character it can be escaped using brackets (for example abc[*] would match a file named abc*). Note that both \ and / are treated as directory separators on Windows, but only / is on Linux. Environment variables (%var% on Windows or $var on Linux) are expanded prior to the pattern being applied. + FilePattern *string `json:"filePattern,omitempty"` + Destination *OutputFileDestination `json:"destination,omitempty"` + UploadOptions *OutputFileUploadOptions `json:"uploadOptions,omitempty"` +} + +// OutputFileBlobContainerDestination ... +type OutputFileBlobContainerDestination struct { + // Path - If filePattern refers to a specific file (i.e. contains no wildcards), then path is the name of the blob to which to upload that file. If filePattern contains one or more wildcards (and therefore may match multiple files), then path is the name of the blob virtual directory (which is prepended to each blob name) to which to upload the file(s). If omitted, file(s) are uploaded to the root of the container with a blob name matching their file name. + Path *string `json:"path,omitempty"` + // ContainerURL - The URL must include a Shared Access Signature (SAS) granting write permissions to the container. + ContainerURL *string `json:"containerUrl,omitempty"` +} + +// OutputFileDestination ... +type OutputFileDestination struct { + Container *OutputFileBlobContainerDestination `json:"container,omitempty"` +} + +// OutputFileUploadOptions ... +type OutputFileUploadOptions struct { + // UploadCondition - The default is taskcompletion. Possible values include: 'OutputFileUploadConditionTaskSuccess', 'OutputFileUploadConditionTaskFailure', 'OutputFileUploadConditionTaskCompletion' + UploadCondition OutputFileUploadCondition `json:"uploadCondition,omitempty"` +} + +// PoolAddParameter ... +type PoolAddParameter struct { + // ID - The ID can contain any combination of alphanumeric characters including hyphens and underscores, and cannot contain more than 64 characters. The ID is case-preserving and case-insensitive (that is, you may not have two pool IDs within an account that differ only by case). + ID *string `json:"id,omitempty"` + // DisplayName - The display name need not be unique and can contain any Unicode characters up to a maximum length of 1024. + DisplayName *string `json:"displayName,omitempty"` + // VMSize - For information about available sizes of virtual machines for Cloud Services pools (pools created with cloudServiceConfiguration), see Sizes for Cloud Services (http://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). Batch supports all Cloud Services VM sizes except ExtraSmall, A1V2 and A2V2. For information about available VM sizes for pools using images from the Virtual Machines Marketplace (pools created with virtualMachineConfiguration) see Sizes for Virtual Machines (Linux) (https://azure.microsoft.com/documentation/articles/virtual-machines-linux-sizes/) or Sizes for Virtual Machines (Windows) (https://azure.microsoft.com/documentation/articles/virtual-machines-windows-sizes/). Batch supports all Azure VM sizes except STANDARD_A0 and those with premium storage (STANDARD_GS, STANDARD_DS, and STANDARD_DSV2 series). + VMSize *string `json:"vmSize,omitempty"` + // CloudServiceConfiguration - This property and virtualMachineConfiguration are mutually exclusive and one of the properties must be specified. This property cannot be specified if the Batch account was created with its poolAllocationMode property set to 'UserSubscription'. + CloudServiceConfiguration *CloudServiceConfiguration `json:"cloudServiceConfiguration,omitempty"` + // VirtualMachineConfiguration - This property and cloudServiceConfiguration are mutually exclusive and one of the properties must be specified. + VirtualMachineConfiguration *VirtualMachineConfiguration `json:"virtualMachineConfiguration,omitempty"` + // ResizeTimeout - This timeout applies only to manual scaling; it has no effect when enableAutoScale is set to true. The default value is 15 minutes. The minimum value is 5 minutes. If you specify a value less than 5 minutes, the Batch service returns an error; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). + ResizeTimeout *string `json:"resizeTimeout,omitempty"` + // TargetDedicatedNodes - This property must not be specified if enableAutoScale is set to true. If enableAutoScale is set to false, then you must set either targetDedicatedNodes, targetLowPriorityNodes, or both. + TargetDedicatedNodes *int32 `json:"targetDedicatedNodes,omitempty"` + // TargetLowPriorityNodes - This property must not be specified if enableAutoScale is set to true. If enableAutoScale is set to false, then you must set either targetDedicatedNodes, targetLowPriorityNodes, or both. + TargetLowPriorityNodes *int32 `json:"targetLowPriorityNodes,omitempty"` + // EnableAutoScale - If false, at least one of targetDedicateNodes and targetLowPriorityNodes must be specified. If true, the autoScaleFormula property is required and the pool automatically resizes according to the formula. The default value is false. + EnableAutoScale *bool `json:"enableAutoScale,omitempty"` + // AutoScaleFormula - This property must not be specified if enableAutoScale is set to false. It is required if enableAutoScale is set to true. The formula is checked for validity before the pool is created. If the formula is not valid, the Batch service rejects the request with detailed error information. For more information about specifying this formula, see 'Automatically scale compute nodes in an Azure Batch pool' (https://azure.microsoft.com/documentation/articles/batch-automatic-scaling/). + AutoScaleFormula *string `json:"autoScaleFormula,omitempty"` + // AutoScaleEvaluationInterval - The default value is 15 minutes. The minimum and maximum value are 5 minutes and 168 hours respectively. If you specify a value less than 5 minutes or greater than 168 hours, the Batch service returns an error; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). + AutoScaleEvaluationInterval *string `json:"autoScaleEvaluationInterval,omitempty"` + // EnableInterNodeCommunication - Enabling inter-node communication limits the maximum size of the pool due to deployment restrictions on the nodes of the pool. This may result in the pool not reaching its desired size. The default value is false. + EnableInterNodeCommunication *bool `json:"enableInterNodeCommunication,omitempty"` + NetworkConfiguration *NetworkConfiguration `json:"networkConfiguration,omitempty"` + // StartTask - The task runs when the node is added to the pool or when the node is restarted. + StartTask *StartTask `json:"startTask,omitempty"` + // CertificateReferences - For Windows compute nodes, the Batch service installs the certificates to the specified certificate store and location. For Linux compute nodes, the certificates are stored in a directory inside the task working directory and an environment variable AZ_BATCH_CERTIFICATES_DIR is supplied to the task to query for this location. For certificates with visibility of 'remoteUser', a 'certs' directory is created in the user's home directory (e.g., /home/{user-name}/certs) and certificates are placed in that directory. + CertificateReferences *[]CertificateReference `json:"certificateReferences,omitempty"` + ApplicationPackageReferences *[]ApplicationPackageReference `json:"applicationPackageReferences,omitempty"` + // ApplicationLicenses - The list of application licenses must be a subset of available Batch service application licenses. If a license is requested which is not supported, pool creation will fail. + ApplicationLicenses *[]string `json:"applicationLicenses,omitempty"` + // MaxTasksPerNode - The default value is 1. The maximum value of this setting depends on the size of the compute nodes in the pool (the vmSize setting). + MaxTasksPerNode *int32 `json:"maxTasksPerNode,omitempty"` + TaskSchedulingPolicy *TaskSchedulingPolicy `json:"taskSchedulingPolicy,omitempty"` + UserAccounts *[]UserAccount `json:"userAccounts,omitempty"` + // Metadata - The Batch service does not assign any meaning to metadata; it is solely for the use of user code. + Metadata *[]MetadataItem `json:"metadata,omitempty"` +} + +// PoolEnableAutoScaleParameter ... +type PoolEnableAutoScaleParameter struct { + // AutoScaleFormula - The formula is checked for validity before it is applied to the pool. If the formula is not valid, the Batch service rejects the request with detailed error information. For more information about specifying this formula, see Automatically scale compute nodes in an Azure Batch pool (https://azure.microsoft.com/en-us/documentation/articles/batch-automatic-scaling). + AutoScaleFormula *string `json:"autoScaleFormula,omitempty"` + // AutoScaleEvaluationInterval - The default value is 15 minutes. The minimum and maximum value are 5 minutes and 168 hours respectively. If you specify a value less than 5 minutes or greater than 168 hours, the Batch service rejects the request with an invalid property value error; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). If you specify a new interval, then the existing autoscale evaluation schedule will be stopped and a new autoscale evaluation schedule will be started, with its starting time being the time when this request was issued. + AutoScaleEvaluationInterval *string `json:"autoScaleEvaluationInterval,omitempty"` +} + +// PoolEndpointConfiguration ... +type PoolEndpointConfiguration struct { + // InboundNATPools - The maximum number of inbound NAT pools per Batch pool is 5. If the maximum number of inbound NAT pools is exceeded the request fails with HTTP status code 400. + InboundNATPools *[]InboundNATPool `json:"inboundNATPools,omitempty"` +} + +// PoolEvaluateAutoScaleParameter ... +type PoolEvaluateAutoScaleParameter struct { + // AutoScaleFormula - The formula is validated and its results calculated, but it is not applied to the pool. To apply the formula to the pool, 'Enable automatic scaling on a pool'. For more information about specifying this formula, see Automatically scale compute nodes in an Azure Batch pool (https://azure.microsoft.com/en-us/documentation/articles/batch-automatic-scaling). + AutoScaleFormula *string `json:"autoScaleFormula,omitempty"` +} + +// PoolInformation ... +type PoolInformation struct { + // PoolID - You must ensure that the pool referenced by this property exists. If the pool does not exist at the time the Batch service tries to schedule a job, no tasks for the job will run until you create a pool with that id. Note that the Batch service will not reject the job request; it will simply not run tasks until the pool exists. You must specify either the pool ID or the auto pool specification, but not both. + PoolID *string `json:"poolId,omitempty"` + // AutoPoolSpecification - If auto pool creation fails, the Batch service moves the job to a completed state, and the pool creation error is set in the job's scheduling error property. The Batch service manages the lifetime (both creation and, unless keepAlive is specified, deletion) of the auto pool. Any user actions that affect the lifetime of the auto pool while the job is active will result in unexpected behavior. You must specify either the pool ID or the auto pool specification, but not both. + AutoPoolSpecification *AutoPoolSpecification `json:"autoPoolSpecification,omitempty"` +} + +// PoolListUsageMetricsResult ... +type PoolListUsageMetricsResult struct { + autorest.Response `json:"-"` + Value *[]PoolUsageMetrics `json:"value,omitempty"` + OdataNextLink *string `json:"odata.nextLink,omitempty"` +} + +// PoolListUsageMetricsResultIterator provides access to a complete listing of PoolUsageMetrics values. +type PoolListUsageMetricsResultIterator struct { + i int + page PoolListUsageMetricsResultPage +} + +// Next advances to the next value. If there was an error making +// the request the iterator does not advance and the error is returned. +func (iter *PoolListUsageMetricsResultIterator) Next() error { + iter.i++ + if iter.i < len(iter.page.Values()) { + return nil + } + err := iter.page.Next() + if err != nil { + iter.i-- + return err + } + iter.i = 0 + return nil +} + +// NotDone returns true if the enumeration should be started or is not yet complete. +func (iter PoolListUsageMetricsResultIterator) NotDone() bool { + return iter.page.NotDone() && iter.i < len(iter.page.Values()) +} + +// Response returns the raw server response from the last page request. +func (iter PoolListUsageMetricsResultIterator) Response() PoolListUsageMetricsResult { + return iter.page.Response() +} + +// Value returns the current value or a zero-initialized value if the +// iterator has advanced beyond the end of the collection. +func (iter PoolListUsageMetricsResultIterator) Value() PoolUsageMetrics { + if !iter.page.NotDone() { + return PoolUsageMetrics{} + } + return iter.page.Values()[iter.i] +} + +// IsEmpty returns true if the ListResult contains no values. +func (plumr PoolListUsageMetricsResult) IsEmpty() bool { + return plumr.Value == nil || len(*plumr.Value) == 0 +} + +// poolListUsageMetricsResultPreparer prepares a request to retrieve the next set of results. +// It returns nil if no more results exist. +func (plumr PoolListUsageMetricsResult) poolListUsageMetricsResultPreparer() (*http.Request, error) { + if plumr.OdataNextLink == nil || len(to.String(plumr.OdataNextLink)) < 1 { + return nil, nil + } + return autorest.Prepare(&http.Request{}, + autorest.AsJSON(), + autorest.AsGet(), + autorest.WithBaseURL(to.String(plumr.OdataNextLink))) +} + +// PoolListUsageMetricsResultPage contains a page of PoolUsageMetrics values. +type PoolListUsageMetricsResultPage struct { + fn func(PoolListUsageMetricsResult) (PoolListUsageMetricsResult, error) + plumr PoolListUsageMetricsResult +} + +// Next advances to the next page of values. If there was an error making +// the request the page does not advance and the error is returned. +func (page *PoolListUsageMetricsResultPage) Next() error { + next, err := page.fn(page.plumr) + if err != nil { + return err + } + page.plumr = next + return nil +} + +// NotDone returns true if the page enumeration should be started or is not yet complete. +func (page PoolListUsageMetricsResultPage) NotDone() bool { + return !page.plumr.IsEmpty() +} + +// Response returns the raw server response from the last page request. +func (page PoolListUsageMetricsResultPage) Response() PoolListUsageMetricsResult { + return page.plumr +} + +// Values returns the slice of values for the current page or nil if there are no values. +func (page PoolListUsageMetricsResultPage) Values() []PoolUsageMetrics { + if page.plumr.IsEmpty() { + return nil + } + return *page.plumr.Value +} + +// PoolPatchParameter ... +type PoolPatchParameter struct { + // StartTask - If this element is present, it overwrites any existing start task. If omitted, any existing start task is left unchanged. + StartTask *StartTask `json:"startTask,omitempty"` + // CertificateReferences - If this element is present, it replaces any existing certificate references configured on the pool. If omitted, any existing certificate references are left unchanged. For Windows compute nodes, the Batch service installs the certificates to the specified certificate store and location. For Linux compute nodes, the certificates are stored in a directory inside the task working directory and an environment variable AZ_BATCH_CERTIFICATES_DIR is supplied to the task to query for this location. For certificates with visibility of 'remoteUser', a 'certs' directory is created in the user's home directory (e.g., /home/{user-name}/certs) and certificates are placed in that directory. + CertificateReferences *[]CertificateReference `json:"certificateReferences,omitempty"` + // ApplicationPackageReferences - Changes to application package references affect all new compute nodes joining the pool, but do not affect compute nodes that are already in the pool until they are rebooted or reimaged. If this element is present, it replaces any existing application package references. If you specify an empty collection, then all application package references are removed from the pool. If omitted, any existing application package references are left unchanged. + ApplicationPackageReferences *[]ApplicationPackageReference `json:"applicationPackageReferences,omitempty"` + // Metadata - If this element is present, it replaces any existing metadata configured on the pool. If you specify an empty collection, any metadata is removed from the pool. If omitted, any existing metadata is left unchanged. + Metadata *[]MetadataItem `json:"metadata,omitempty"` +} + +// PoolResizeParameter ... +type PoolResizeParameter struct { + TargetDedicatedNodes *int32 `json:"targetDedicatedNodes,omitempty"` + TargetLowPriorityNodes *int32 `json:"targetLowPriorityNodes,omitempty"` + // ResizeTimeout - The default value is 15 minutes. The minimum value is 5 minutes. If you specify a value less than 5 minutes, the Batch service returns an error; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). + ResizeTimeout *string `json:"resizeTimeout,omitempty"` + // NodeDeallocationOption - The default value is requeue. Possible values include: 'Requeue', 'Terminate', 'TaskCompletion', 'RetainedData' + NodeDeallocationOption ComputeNodeDeallocationOption `json:"nodeDeallocationOption,omitempty"` +} + +// PoolSpecification ... +type PoolSpecification struct { + // DisplayName - The display name need not be unique and can contain any Unicode characters up to a maximum length of 1024. + DisplayName *string `json:"displayName,omitempty"` + // VMSize - For information about available sizes of virtual machines for Cloud Services pools (pools created with cloudServiceConfiguration), see Sizes for Cloud Services (http://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). Batch supports all Cloud Services VM sizes except ExtraSmall, A1V2 and A2V2. For information about available VM sizes for pools using images from the Virtual Machines Marketplace (pools created with virtualMachineConfiguration) see Sizes for Virtual Machines (Linux) (https://azure.microsoft.com/documentation/articles/virtual-machines-linux-sizes/) or Sizes for Virtual Machines (Windows) (https://azure.microsoft.com/documentation/articles/virtual-machines-windows-sizes/). Batch supports all Azure VM sizes except STANDARD_A0 and those with premium storage (STANDARD_GS, STANDARD_DS, and STANDARD_DSV2 series). + VMSize *string `json:"vmSize,omitempty"` + // CloudServiceConfiguration - This property must be specified if the pool needs to be created with Azure PaaS VMs. This property and virtualMachineConfiguration are mutually exclusive and one of the properties must be specified. If neither is specified then the Batch service returns an error; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). This property cannot be specified if the Batch account was created with its poolAllocationMode property set to 'UserSubscription'. + CloudServiceConfiguration *CloudServiceConfiguration `json:"cloudServiceConfiguration,omitempty"` + // VirtualMachineConfiguration - This property must be specified if the pool needs to be created with Azure IaaS VMs. This property and cloudServiceConfiguration are mutually exclusive and one of the properties must be specified. If neither is specified then the Batch service returns an error; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). + VirtualMachineConfiguration *VirtualMachineConfiguration `json:"virtualMachineConfiguration,omitempty"` + // MaxTasksPerNode - The default value is 1. The maximum value of this setting depends on the size of the compute nodes in the pool (the vmSize setting). + MaxTasksPerNode *int32 `json:"maxTasksPerNode,omitempty"` + TaskSchedulingPolicy *TaskSchedulingPolicy `json:"taskSchedulingPolicy,omitempty"` + // ResizeTimeout - This timeout applies only to manual scaling; it has no effect when enableAutoScale is set to true. The default value is 15 minutes. The minimum value is 5 minutes. If you specify a value less than 5 minutes, the Batch service rejects the request with an error; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). + ResizeTimeout *string `json:"resizeTimeout,omitempty"` + // TargetDedicatedNodes - This property must not be specified if enableAutoScale is set to true. If enableAutoScale is set to false, then you must set either targetDedicatedNodes, targetLowPriorityNodes, or both. + TargetDedicatedNodes *int32 `json:"targetDedicatedNodes,omitempty"` + // TargetLowPriorityNodes - This property must not be specified if enableAutoScale is set to true. If enableAutoScale is set to false, then you must set either targetDedicatedNodes, targetLowPriorityNodes, or both. + TargetLowPriorityNodes *int32 `json:"targetLowPriorityNodes,omitempty"` + // EnableAutoScale - If false, at least one of targetDedicateNodes and targetLowPriorityNodes must be specified. If true, the autoScaleFormula element is required. The pool automatically resizes according to the formula. The default value is false. + EnableAutoScale *bool `json:"enableAutoScale,omitempty"` + // AutoScaleFormula - This property must not be specified if enableAutoScale is set to false. It is required if enableAutoScale is set to true. The formula is checked for validity before the pool is created. If the formula is not valid, the Batch service rejects the request with detailed error information. + AutoScaleFormula *string `json:"autoScaleFormula,omitempty"` + // AutoScaleEvaluationInterval - The default value is 15 minutes. The minimum and maximum value are 5 minutes and 168 hours respectively. If you specify a value less than 5 minutes or greater than 168 hours, the Batch service rejects the request with an invalid property value error; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). + AutoScaleEvaluationInterval *string `json:"autoScaleEvaluationInterval,omitempty"` + // EnableInterNodeCommunication - Enabling inter-node communication limits the maximum size of the pool due to deployment restrictions on the nodes of the pool. This may result in the pool not reaching its desired size. The default value is false. + EnableInterNodeCommunication *bool `json:"enableInterNodeCommunication,omitempty"` + NetworkConfiguration *NetworkConfiguration `json:"networkConfiguration,omitempty"` + StartTask *StartTask `json:"startTask,omitempty"` + // CertificateReferences - For Windows compute nodes, the Batch service installs the certificates to the specified certificate store and location. For Linux compute nodes, the certificates are stored in a directory inside the task working directory and an environment variable AZ_BATCH_CERTIFICATES_DIR is supplied to the task to query for this location. For certificates with visibility of 'remoteUser', a 'certs' directory is created in the user's home directory (e.g., /home/{user-name}/certs) and certificates are placed in that directory. + CertificateReferences *[]CertificateReference `json:"certificateReferences,omitempty"` + ApplicationPackageReferences *[]ApplicationPackageReference `json:"applicationPackageReferences,omitempty"` + // ApplicationLicenses - The list of application licenses must be a subset of available Batch service application licenses. If a license is requested which is not supported, pool creation will fail. + ApplicationLicenses *[]string `json:"applicationLicenses,omitempty"` + UserAccounts *[]UserAccount `json:"userAccounts,omitempty"` + // Metadata - The Batch service does not assign any meaning to metadata; it is solely for the use of user code. + Metadata *[]MetadataItem `json:"metadata,omitempty"` +} + +// PoolStatistics ... +type PoolStatistics struct { + autorest.Response `json:"-"` + URL *string `json:"url,omitempty"` + StartTime *date.Time `json:"startTime,omitempty"` + LastUpdateTime *date.Time `json:"lastUpdateTime,omitempty"` + UsageStats *UsageStatistics `json:"usageStats,omitempty"` + ResourceStats *ResourceStatistics `json:"resourceStats,omitempty"` +} + +// PoolUpdatePropertiesParameter ... +type PoolUpdatePropertiesParameter struct { + // StartTask - If this element is present, it overwrites any existing start task. If omitted, any existing start task is removed from the pool. + StartTask *StartTask `json:"startTask,omitempty"` + // CertificateReferences - This list replaces any existing certificate references configured on the pool. If you specify an empty collection, any existing certificate references are removed from the pool. For Windows compute nodes, the Batch service installs the certificates to the specified certificate store and location. For Linux compute nodes, the certificates are stored in a directory inside the task working directory and an environment variable AZ_BATCH_CERTIFICATES_DIR is supplied to the task to query for this location. For certificates with visibility of 'remoteUser', a 'certs' directory is created in the user's home directory (e.g., /home/{user-name}/certs) and certificates are placed in that directory. + CertificateReferences *[]CertificateReference `json:"certificateReferences,omitempty"` + // ApplicationPackageReferences - The list replaces any existing application package references on the pool. Changes to application package references affect all new compute nodes joining the pool, but do not affect compute nodes that are already in the pool until they are rebooted or reimaged. If omitted, or if you specify an empty collection, any existing application packages references are removed from the pool. + ApplicationPackageReferences *[]ApplicationPackageReference `json:"applicationPackageReferences,omitempty"` + // Metadata - This list replaces any existing metadata configured on the pool. If omitted, or if you specify an empty collection, any existing metadata is removed from the pool. + Metadata *[]MetadataItem `json:"metadata,omitempty"` +} + +// PoolUpgradeOSParameter ... +type PoolUpgradeOSParameter struct { + TargetOSVersion *string `json:"targetOSVersion,omitempty"` +} + +// PoolUsageMetrics ... +type PoolUsageMetrics struct { + PoolID *string `json:"poolId,omitempty"` + StartTime *date.Time `json:"startTime,omitempty"` + EndTime *date.Time `json:"endTime,omitempty"` + // VMSize - For information about available sizes of virtual machines for Cloud Services pools (pools created with cloudServiceConfiguration), see Sizes for Cloud Services (http://azure.microsoft.com/documentation/articles/cloud-services-sizes-specs/). Batch supports all Cloud Services VM sizes except ExtraSmall, STANDARD_A1_V2 and STANDARD_A2_V2. For information about available VM sizes for pools using images from the Virtual Machines Marketplace (pools created with virtualMachineConfiguration) see Sizes for Virtual Machines (Linux) (https://azure.microsoft.com/documentation/articles/virtual-machines-linux-sizes/) or Sizes for Virtual Machines (Windows) (https://azure.microsoft.com/documentation/articles/virtual-machines-windows-sizes/). Batch supports all Azure VM sizes except STANDARD_A0 and those with premium storage (STANDARD_GS, STANDARD_DS, and STANDARD_DSV2 series). + VMSize *string `json:"vmSize,omitempty"` + TotalCoreHours *float64 `json:"totalCoreHours,omitempty"` + DataIngressGiB *float64 `json:"dataIngressGiB,omitempty"` + DataEgressGiB *float64 `json:"dataEgressGiB,omitempty"` +} + +// ReadCloser ... +type ReadCloser struct { + autorest.Response `json:"-"` + Value *io.ReadCloser `json:"value,omitempty"` +} + +// RecentJob ... +type RecentJob struct { + ID *string `json:"id,omitempty"` + URL *string `json:"url,omitempty"` +} + +// ResizeError ... +type ResizeError struct { + Code *string `json:"code,omitempty"` + Message *string `json:"message,omitempty"` + Values *[]NameValuePair `json:"values,omitempty"` +} + +// ResourceFile ... +type ResourceFile struct { + // BlobSource - This URL must be readable using anonymous access; that is, the Batch service does not present any credentials when downloading the blob. There are two ways to get such a URL for a blob in Azure storage: include a Shared Access Signature (SAS) granting read permissions on the blob, or set the ACL for the blob or its container to allow public access. + BlobSource *string `json:"blobSource,omitempty"` + FilePath *string `json:"filePath,omitempty"` + // FileMode - This property applies only to files being downloaded to Linux compute nodes. It will be ignored if it is specified for a resourceFile which will be downloaded to a Windows node. If this property is not specified for a Linux node, then a default value of 0770 is applied to the file. + FileMode *string `json:"fileMode,omitempty"` +} + +// ResourceStatistics ... +type ResourceStatistics struct { + StartTime *date.Time `json:"startTime,omitempty"` + LastUpdateTime *date.Time `json:"lastUpdateTime,omitempty"` + AvgCPUPercentage *float64 `json:"avgCPUPercentage,omitempty"` + AvgMemoryGiB *float64 `json:"avgMemoryGiB,omitempty"` + PeakMemoryGiB *float64 `json:"peakMemoryGiB,omitempty"` + AvgDiskGiB *float64 `json:"avgDiskGiB,omitempty"` + PeakDiskGiB *float64 `json:"peakDiskGiB,omitempty"` + DiskReadIOps *int64 `json:"diskReadIOps,omitempty"` + DiskWriteIOps *int64 `json:"diskWriteIOps,omitempty"` + DiskReadGiB *float64 `json:"diskReadGiB,omitempty"` + DiskWriteGiB *float64 `json:"diskWriteGiB,omitempty"` + NetworkReadGiB *float64 `json:"networkReadGiB,omitempty"` + NetworkWriteGiB *float64 `json:"networkWriteGiB,omitempty"` +} + +// Schedule ... +type Schedule struct { + // DoNotRunUntil - If you do not specify a doNotRunUntil time, the schedule becomes ready to create jobs immediately. + DoNotRunUntil *date.Time `json:"doNotRunUntil,omitempty"` + // DoNotRunAfter - If you do not specify a doNotRunAfter time, and you are creating a recurring job schedule, the job schedule will remain active until you explicitly terminate it. + DoNotRunAfter *date.Time `json:"doNotRunAfter,omitempty"` + // StartWindow - If a job is not created within the startWindow interval, then the 'opportunity' is lost; no job will be created until the next recurrence of the schedule. If the schedule is recurring, and the startWindow is longer than the recurrence interval, then this is equivalent to an infinite startWindow, because the job that is 'due' in one recurrenceInterval is not carried forward into the next recurrence interval. The default is infinite. The minimum value is 1 minute. If you specify a lower value, the Batch service rejects the schedule with an error; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). + StartWindow *string `json:"startWindow,omitempty"` + // RecurrenceInterval - Because a job schedule can have at most one active job under it at any given time, if it is time to create a new job under a job schedule, but the previous job is still running, the Batch service will not create the new job until the previous job finishes. If the previous job does not finish within the startWindow period of the new recurrenceInterval, then no new job will be scheduled for that interval. For recurring jobs, you should normally specify a jobManagerTask in the jobSpecification. If you do not use jobManagerTask, you will need an external process to monitor when jobs are created, add tasks to the jobs and terminate the jobs ready for the next recurrence. The default is that the schedule does not recur: one job is created, within the startWindow after the doNotRunUntil time, and the schedule is complete as soon as that job finishes. The minimum value is 1 minute. If you specify a lower value, the Batch service rejects the schedule with an error; if you are calling the REST API directly, the HTTP status code is 400 (Bad Request). + RecurrenceInterval *string `json:"recurrenceInterval,omitempty"` +} + +// StartTask ... +type StartTask struct { + // CommandLine - The command line does not run under a shell, and therefore cannot take advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows or "/bin/sh -c MyCommand" in Linux. + CommandLine *string `json:"commandLine,omitempty"` + // ContainerSettings - When this is specified, all directories recursively below the AZ_BATCH_NODE_ROOT_DIR (the root of Azure Batch directories on the node) are mapped into the container, all task environment variables are mapped into the container, and the task command line is executed in the container. + ContainerSettings *TaskContainerSettings `json:"containerSettings,omitempty"` + // ResourceFiles - Files listed under this element are located in the task's working directory. + ResourceFiles *[]ResourceFile `json:"resourceFiles,omitempty"` + EnvironmentSettings *[]EnvironmentSetting `json:"environmentSettings,omitempty"` + // UserIdentity - If omitted, the task runs as a non-administrative user unique to the task. + UserIdentity *UserIdentity `json:"userIdentity,omitempty"` + // MaxTaskRetryCount - The Batch service retries a task if its exit code is nonzero. Note that this value specifically controls the number of retries. The Batch service will try the task once, and may then retry up to this limit. For example, if the maximum retry count is 3, Batch tries the task up to 4 times (one initial try and 3 retries). If the maximum retry count is 0, the Batch service does not retry the task. If the maximum retry count is -1, the Batch service retries the task without limit. + MaxTaskRetryCount *int32 `json:"maxTaskRetryCount,omitempty"` + // WaitForSuccess - If true and the start task fails on a compute node, the Batch service retries the start task up to its maximum retry count (maxTaskRetryCount). If the task has still not completed successfully after all retries, then the Batch service marks the compute node unusable, and will not schedule tasks to it. This condition can be detected via the node state and failure info details. If false, the Batch service will not wait for the start task to complete. In this case, other tasks can start executing on the compute node while the start task is still running; and even if the start task fails, new tasks will continue to be scheduled on the node. The default is false. + WaitForSuccess *bool `json:"waitForSuccess,omitempty"` +} + +// StartTaskInformation ... +type StartTaskInformation struct { + // State - Possible values include: 'StartTaskStateRunning', 'StartTaskStateCompleted' + State StartTaskState `json:"state,omitempty"` + // StartTime - This value is reset every time the task is restarted or retried (that is, this is the most recent time at which the start task started running). + StartTime *date.Time `json:"startTime,omitempty"` + // EndTime - This is the end time of the most recent run of the start task, if that run has completed (even if that run failed and a retry is pending). This element is not present if the start task is currently running. + EndTime *date.Time `json:"endTime,omitempty"` + // ExitCode - This property is set only if the start task is in the completed state. In general, the exit code for a process reflects the specific convention implemented by the application developer for that process. If you use the exit code value to make decisions in your code, be sure that you know the exit code convention used by the application process. However, if the Batch service terminates the start task (due to timeout, or user termination via the API) you may see an operating system-defined exit code. + ExitCode *int32 `json:"exitCode,omitempty"` + // ContainerInfo - This property is set only if the task runs in a container context. + ContainerInfo *TaskContainerExecutionInformation `json:"containerInfo,omitempty"` + // FailureInfo - This property is set only if the task is in the completed state and encountered a failure. + FailureInfo *TaskFailureInformation `json:"failureInfo,omitempty"` + // RetryCount - Task application failures (non-zero exit code) are retried, pre-processing errors (the task could not be run) and file upload errors are not retried. The Batch service will retry the task up to the limit specified by the constraints. + RetryCount *int32 `json:"retryCount,omitempty"` + // LastRetryTime - This element is present only if the task was retried (i.e. retryCount is nonzero). If present, this is typically the same as startTime, but may be different if the task has been restarted for reasons other than retry; for example, if the compute node was rebooted during a retry, then the startTime is updated but the lastRetryTime is not. + LastRetryTime *date.Time `json:"lastRetryTime,omitempty"` + // Result - If the value is 'failed', then the details of the failure can be found in the failureInfo property. Possible values include: 'Success', 'Failure' + Result TaskExecutionResult `json:"result,omitempty"` +} + +// SubtaskInformation ... +type SubtaskInformation struct { + ID *int32 `json:"id,omitempty"` + NodeInfo *ComputeNodeInformation `json:"nodeInfo,omitempty"` + StartTime *date.Time `json:"startTime,omitempty"` + // EndTime - This property is set only if the subtask is in the Completed state. + EndTime *date.Time `json:"endTime,omitempty"` + // ExitCode - This property is set only if the subtask is in the completed state. In general, the exit code for a process reflects the specific convention implemented by the application developer for that process. If you use the exit code value to make decisions in your code, be sure that you know the exit code convention used by the application process. However, if the Batch service terminates the subtask (due to timeout, or user termination via the API) you may see an operating system-defined exit code. + ExitCode *int32 `json:"exitCode,omitempty"` + // ContainerInfo - This property is set only if the task runs in a container context. + ContainerInfo *TaskContainerExecutionInformation `json:"containerInfo,omitempty"` + // FailureInfo - This property is set only if the task is in the completed state and encountered a failure. + FailureInfo *TaskFailureInformation `json:"failureInfo,omitempty"` + // State - Possible values include: 'SubtaskStatePreparing', 'SubtaskStateRunning', 'SubtaskStateCompleted' + State SubtaskState `json:"state,omitempty"` + StateTransitionTime *date.Time `json:"stateTransitionTime,omitempty"` + // PreviousState - This property is not set if the subtask is in its initial running state. Possible values include: 'SubtaskStatePreparing', 'SubtaskStateRunning', 'SubtaskStateCompleted' + PreviousState SubtaskState `json:"previousState,omitempty"` + // PreviousStateTransitionTime - This property is not set if the subtask is in its initial running state. + PreviousStateTransitionTime *date.Time `json:"previousStateTransitionTime,omitempty"` + // Result - If the value is 'failed', then the details of the failure can be found in the failureInfo property. Possible values include: 'Success', 'Failure' + Result TaskExecutionResult `json:"result,omitempty"` +} + +// TaskAddCollectionParameter ... +type TaskAddCollectionParameter struct { + // Value - The total serialized size of this collection must be less than 4MB. If it is greater than 4MB (for example if each task has 100's of resource files or environment variables), the request will fail with code 'RequestBodyTooLarge' and should be retried again with fewer tasks. + Value *[]TaskAddParameter `json:"value,omitempty"` +} + +// TaskAddCollectionResult ... +type TaskAddCollectionResult struct { + autorest.Response `json:"-"` + Value *[]TaskAddResult `json:"value,omitempty"` +} + +// TaskAddParameter ... +type TaskAddParameter struct { + // ID - The ID can contain any combination of alphanumeric characters including hyphens and underscores, and cannot contain more than 64 characters. The ID is case-preserving and case-insensitive (that is, you may not have two IDs within a job that differ only by case). + ID *string `json:"id,omitempty"` + // DisplayName - The display name need not be unique and can contain any Unicode characters up to a maximum length of 1024. + DisplayName *string `json:"displayName,omitempty"` + // CommandLine - For multi-instance tasks, the command line is executed as the primary task, after the primary task and all subtasks have finished executing the coordination command line. The command line does not run under a shell, and therefore cannot take advantage of shell features such as environment variable expansion. If you want to take advantage of such features, you should invoke the shell in the command line, for example using "cmd /c MyCommand" in Windows or "/bin/sh -c MyCommand" in Linux. + CommandLine *string `json:"commandLine,omitempty"` + // ContainerSettings - If the pool that will run this task has containerConfiguration set, this must be set as well. If the pool that will run this task doesn't have containerConfiguration set, this must not be set. When this is specified, all directories recursively below the AZ_BATCH_NODE_ROOT_DIR (the root of Azure Batch directories on the node) are mapped into the container, all task environment variables are mapped into the container, and the task command line is executed in the container. + ContainerSettings *TaskContainerSettings `json:"containerSettings,omitempty"` + // ExitConditions - How the Batch service should respond when the task completes. + ExitConditions *ExitConditions `json:"exitConditions,omitempty"` + // ResourceFiles - For multi-instance tasks, the resource files will only be downloaded to the compute node on which the primary task is executed. + ResourceFiles *[]ResourceFile `json:"resourceFiles,omitempty"` + // OutputFiles - For multi-instance tasks, the files will only be uploaded from the compute node on which the primary task is executed. + OutputFiles *[]OutputFile `json:"outputFiles,omitempty"` + EnvironmentSettings *[]EnvironmentSetting `json:"environmentSettings,omitempty"` + AffinityInfo *AffinityInformation `json:"affinityInfo,omitempty"` + // Constraints - If you do not specify constraints, the maxTaskRetryCount is the maxTaskRetryCount specified for the job, and the maxWallClockTime and retentionTime are infinite. + Constraints *TaskConstraints `json:"constraints,omitempty"` + // UserIdentity - If omitted, the task runs as a non-administrative user unique to the task. + UserIdentity *UserIdentity `json:"userIdentity,omitempty"` + MultiInstanceSettings *MultiInstanceSettings `json:"multiInstanceSettings,omitempty"` + // DependsOn - This task will not be scheduled until all tasks that it depends on have completed successfully. If any of those tasks fail and exhaust their retry counts, this task will never be scheduled. If the job does not have usesTaskDependencies set to true, and this element is present, the request fails with error code TaskDependenciesNotSpecifiedOnJob. + DependsOn *TaskDependencies `json:"dependsOn,omitempty"` + // ApplicationPackageReferences - Application packages are downloaded and deployed to a shared directory, not the task working directory. Therefore, if a referenced package is already on the compute node, and is up to date, then it is not re-downloaded; the existing copy on the compute node is used. If a referenced application package cannot be installed, for example because the package has been deleted or because download failed, the task fails. + ApplicationPackageReferences *[]ApplicationPackageReference `json:"applicationPackageReferences,omitempty"` + // AuthenticationTokenSettings - If this property is set, the Batch service provides the task with an authentication token which can be used to authenticate Batch service operations without requiring an account access key. The token is provided via the AZ_BATCH_AUTHENTICATION_TOKEN environment variable. The operations that the task can carry out using the token depend on the settings. For example, a task can request job permissions in order to add other tasks to the job, or check the status of the job or of other tasks under the job. + AuthenticationTokenSettings *AuthenticationTokenSettings `json:"authenticationTokenSettings,omitempty"` +} + +// TaskAddResult ... +type TaskAddResult struct { + // Status - Possible values include: 'TaskAddStatusSuccess', 'TaskAddStatusClientError', 'TaskAddStatusServerError' + Status TaskAddStatus `json:"status,omitempty"` + TaskID *string `json:"taskId,omitempty"` + // ETag - You can use this to detect whether the task has changed between requests. In particular, you can be pass the ETag with an Update Task request to specify that your changes should take effect only if nobody else has modified the job in the meantime. + ETag *string `json:"eTag,omitempty"` + LastModified *date.Time `json:"lastModified,omitempty"` + Location *string `json:"location,omitempty"` + Error *Error `json:"error,omitempty"` +} + +// TaskConstraints ... +type TaskConstraints struct { + // MaxWallClockTime - If this is not specified, there is no time limit on how long the task may run. + MaxWallClockTime *string `json:"maxWallClockTime,omitempty"` + // RetentionTime - The default is infinite, i.e. the task directory will be retained until the compute node is removed or reimaged. + RetentionTime *string `json:"retentionTime,omitempty"` + // MaxTaskRetryCount - Note that this value specifically controls the number of retries. The Batch service will try the task once, and may then retry up to this limit. For example, if the maximum retry count is 3, Batch tries the task up to 4 times (one initial try and 3 retries). If the maximum retry count is 0, the Batch service does not retry the task. If the maximum retry count is -1, the Batch service retries the task without limit. + MaxTaskRetryCount *int32 `json:"maxTaskRetryCount,omitempty"` +} + +// TaskContainerExecutionInformation ... +type TaskContainerExecutionInformation struct { + ContainerID *string `json:"containerId,omitempty"` + // State - This is the state of the container according to the Docker service. It is equivalent to the status field returned by "docker inspect". + State *string `json:"state,omitempty"` + // Error - This is the detailed error string from the Docker service, if available. It is equivalent to the error field returned by "docker inspect". + Error *string `json:"error,omitempty"` +} + +// TaskContainerSettings ... +type TaskContainerSettings struct { + // ContainerRunOptions - These additional options are supplied as arguments to the "docker create" command, in addition to those controlled by the Batch Service. + ContainerRunOptions *string `json:"containerRunOptions,omitempty"` + // ImageName - This is the full image reference, as would be specified to "docker pull". If no tag is provided as part of the image name, the tag ":latest" is used as a default. + ImageName *string `json:"imageName,omitempty"` + // Registry - This setting can be omitted if was already provided at pool creation. + Registry *ContainerRegistry `json:"registry,omitempty"` +} + +// TaskCounts ... +type TaskCounts struct { + autorest.Response `json:"-"` + Active *int32 `json:"active,omitempty"` + Running *int32 `json:"running,omitempty"` + Completed *int32 `json:"completed,omitempty"` + Succeeded *int32 `json:"succeeded,omitempty"` + Failed *int32 `json:"failed,omitempty"` + // ValidationStatus - Possible values include: 'Validated', 'Unvalidated' + ValidationStatus TaskCountValidationStatus `json:"validationStatus,omitempty"` +} + +// TaskDependencies ... +type TaskDependencies struct { + // TaskIds - The taskIds collection is limited to 64000 characters total (i.e. the combined length of all task IDs). If the taskIds collection exceeds the maximum length, the Add Task request fails with error code TaskDependencyListTooLong. In this case consider using task ID ranges instead. + TaskIds *[]string `json:"taskIds,omitempty"` + TaskIDRanges *[]TaskIDRange `json:"taskIdRanges,omitempty"` +} + +// TaskExecutionInformation ... +type TaskExecutionInformation struct { + // StartTime - 'Running' corresponds to the running state, so if the task specifies resource files or application packages, then the start time reflects the time at which the task started downloading or deploying these. If the task has been restarted or retried, this is the most recent time at which the task started running. This property is present only for tasks that are in the running or completed state. + StartTime *date.Time `json:"startTime,omitempty"` + // EndTime - This property is set only if the task is in the Completed state. + EndTime *date.Time `json:"endTime,omitempty"` + // ExitCode - This property is set only if the task is in the completed state. In general, the exit code for a process reflects the specific convention implemented by the application developer for that process. If you use the exit code value to make decisions in your code, be sure that you know the exit code convention used by the application process. However, if the Batch service terminates the task (due to timeout, or user termination via the API) you may see an operating system-defined exit code. + ExitCode *int32 `json:"exitCode,omitempty"` + // ContainerInfo - This property is set only if the task runs in a container context. + ContainerInfo *TaskContainerExecutionInformation `json:"containerInfo,omitempty"` + // FailureInfo - This property is set only if the task is in the completed state and encountered a failure. + FailureInfo *TaskFailureInformation `json:"failureInfo,omitempty"` + // RetryCount - Task application failures (non-zero exit code) are retried, pre-processing errors (the task could not be run) and file upload errors are not retried. The Batch service will retry the task up to the limit specified by the constraints. + RetryCount *int32 `json:"retryCount,omitempty"` + // LastRetryTime - This element is present only if the task was retried (i.e. retryCount is nonzero). If present, this is typically the same as startTime, but may be different if the task has been restarted for reasons other than retry; for example, if the compute node was rebooted during a retry, then the startTime is updated but the lastRetryTime is not. + LastRetryTime *date.Time `json:"lastRetryTime,omitempty"` + // RequeueCount - When the user removes nodes from a pool (by resizing/shrinking the pool) or when the job is being disabled, the user can specify that running tasks on the nodes be requeued for execution. This count tracks how many times the task has been requeued for these reasons. + RequeueCount *int32 `json:"requeueCount,omitempty"` + // LastRequeueTime - This property is set only if the requeueCount is nonzero. + LastRequeueTime *date.Time `json:"lastRequeueTime,omitempty"` + // Result - If the value is 'failed', then the details of the failure can be found in the failureInfo property. Possible values include: 'Success', 'Failure' + Result TaskExecutionResult `json:"result,omitempty"` +} + +// TaskFailureInformation ... +type TaskFailureInformation struct { + // Category - Possible values include: 'UserError', 'ServerError' + Category ErrorCategory `json:"category,omitempty"` + Code *string `json:"code,omitempty"` + Message *string `json:"message,omitempty"` + Details *[]NameValuePair `json:"details,omitempty"` +} + +// TaskIDRange the start and end of the range are inclusive. For example, if a range has start 9 and end 12, then +// it represents tasks '9', '10', '11' and '12'. +type TaskIDRange struct { + Start *int32 `json:"start,omitempty"` + End *int32 `json:"end,omitempty"` +} + +// TaskInformation ... +type TaskInformation struct { + TaskURL *string `json:"taskUrl,omitempty"` + JobID *string `json:"jobId,omitempty"` + TaskID *string `json:"taskId,omitempty"` + SubtaskID *int32 `json:"subtaskId,omitempty"` + // TaskState - Possible values include: 'TaskStateActive', 'TaskStatePreparing', 'TaskStateRunning', 'TaskStateCompleted' + TaskState TaskState `json:"taskState,omitempty"` + ExecutionInfo *TaskExecutionInformation `json:"executionInfo,omitempty"` +} + +// TaskSchedulingPolicy ... +type TaskSchedulingPolicy struct { + // NodeFillType - Possible values include: 'Spread', 'Pack' + NodeFillType ComputeNodeFillType `json:"nodeFillType,omitempty"` +} + +// TaskStatistics ... +type TaskStatistics struct { + URL *string `json:"url,omitempty"` + StartTime *date.Time `json:"startTime,omitempty"` + LastUpdateTime *date.Time `json:"lastUpdateTime,omitempty"` + UserCPUTime *string `json:"userCPUTime,omitempty"` + KernelCPUTime *string `json:"kernelCPUTime,omitempty"` + // WallClockTime - The wall clock time is the elapsed time from when the task started running on a compute node to when it finished (or to the last time the statistics were updated, if the task had not finished by then). If the task was retried, this includes the wall clock time of all the task retries. + WallClockTime *string `json:"wallClockTime,omitempty"` + ReadIOps *int64 `json:"readIOps,omitempty"` + WriteIOps *int64 `json:"writeIOps,omitempty"` + ReadIOGiB *float64 `json:"readIOGiB,omitempty"` + WriteIOGiB *float64 `json:"writeIOGiB,omitempty"` + WaitTime *string `json:"waitTime,omitempty"` +} + +// TaskUpdateParameter ... +type TaskUpdateParameter struct { + // Constraints - If omitted, the task is given the default constraints. For multi-instance tasks, updating the retention time applies only to the primary task and not subtasks. + Constraints *TaskConstraints `json:"constraints,omitempty"` +} + +// UsageStatistics ... +type UsageStatistics struct { + StartTime *date.Time `json:"startTime,omitempty"` + LastUpdateTime *date.Time `json:"lastUpdateTime,omitempty"` + DedicatedCoreTime *string `json:"dedicatedCoreTime,omitempty"` +} + +// UserAccount ... +type UserAccount struct { + Name *string `json:"name,omitempty"` + Password *string `json:"password,omitempty"` + // ElevationLevel - nonAdmin - The auto user is a standard user without elevated access. admin - The auto user is a user with elevated access and operates with full Administrator permissions. The default value is nonAdmin. Possible values include: 'NonAdmin', 'Admin' + ElevationLevel ElevationLevel `json:"elevationLevel,omitempty"` + // LinuxUserConfiguration - This property is ignored if specified on a Windows pool. If not specified, the user is created with the default options. + LinuxUserConfiguration *LinuxUserConfiguration `json:"linuxUserConfiguration,omitempty"` +} + +// UserIdentity specify either the userName or autoUser property, but not both. +type UserIdentity struct { + // UserName - The userName and autoUser properties are mutually exclusive; you must specify one but not both. + UserName *string `json:"username,omitempty"` + // AutoUser - The userName and autoUser properties are mutually exclusive; you must specify one but not both. + AutoUser *AutoUserSpecification `json:"autoUser,omitempty"` +} + +// VirtualMachineConfiguration ... +type VirtualMachineConfiguration struct { + ImageReference *ImageReference `json:"imageReference,omitempty"` + OsDisk *OSDisk `json:"osDisk,omitempty"` + // NodeAgentSKUID - The Batch node agent is a program that runs on each node in the pool, and provides the command-and-control interface between the node and the Batch service. There are different implementations of the node agent, known as SKUs, for different operating systems. You must specify a node agent SKU which matches the selected image reference. To get the list of supported node agent SKUs along with their list of verified image references, see the 'List supported node agent SKUs' operation. + NodeAgentSKUID *string `json:"nodeAgentSKUId,omitempty"` + // WindowsConfiguration - This property must not be specified if the imageReference or osDisk property specifies a Linux OS image. + WindowsConfiguration *WindowsConfiguration `json:"windowsConfiguration,omitempty"` + // DataDisks - This property must be specified if the compute nodes in the pool need to have empty data disks attached to them. This cannot be updated. Each node gets its own disk (the disk is not a file share). Existing disks cannot be attached, each attached disk is empty. When the node is removed from the pool, the disk and all data associated with it is also deleted. The disk is not formatted after being attached, it must be formatted before use - for more information see https://docs.microsoft.com/en-us/azure/virtual-machines/linux/classic/attach-disk#initialize-a-new-data-disk-in-linux and https://docs.microsoft.com/en-us/azure/virtual-machines/windows/attach-disk-ps#add-an-empty-data-disk-to-a-virtual-machine. + DataDisks *[]DataDisk `json:"dataDisks,omitempty"` + // LicenseType - This only applies to images that contain the Windows operating system, and should only be used when you hold valid on-premises licenses for the nodes which will be deployed. If omitted, no on-premises licensing discount is applied. Values are: + // Windows_Server - The on-premises license is for Windows Server. + // Windows_Client - The on-premises license is for Windows Client. + LicenseType *string `json:"licenseType,omitempty"` + // ContainerConfiguration - If specified, setup is performed on each node in the pool to allow tasks to run in containers. All regular tasks and job manager tasks run on this pool must specify the containerSettings property, and all other tasks may specify it. + ContainerConfiguration *ContainerConfiguration `json:"containerConfiguration,omitempty"` +} + +// WindowsConfiguration ... +type WindowsConfiguration struct { + // EnableAutomaticUpdates - If omitted, the default value is true. + EnableAutomaticUpdates *bool `json:"enableAutomaticUpdates,omitempty"` +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/pool.go b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/pool.go new file mode 100644 index 000000000..b9846a345 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/pool.go @@ -0,0 +1,1897 @@ +package batch + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "github.com/Azure/go-autorest/autorest/date" + "github.com/Azure/go-autorest/autorest/validation" + "github.com/satori/go.uuid" + "net/http" +) + +// PoolClient is the a client for issuing REST requests to the Azure Batch service. +type PoolClient struct { + BaseClient +} + +// NewPoolClient creates an instance of the PoolClient client. +func NewPoolClient() PoolClient { + return NewPoolClientWithBaseURI(DefaultBaseURI) +} + +// NewPoolClientWithBaseURI creates an instance of the PoolClient client. +func NewPoolClientWithBaseURI(baseURI string) PoolClient { + return PoolClient{NewWithBaseURI(baseURI)} +} + +// Add when naming pools, avoid including sensitive information such as user names or secret project names. This +// information may appear in telemetry logs accessible to Microsoft Support engineers. +// +// pool is the pool to be added. timeout is the maximum time that the server can spend processing the request, in +// seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a +// GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID +// is whether the server should return the client-request-id in the response. ocpDate is the time the request was +// issued. Client libraries typically set this to the current system clock time; set it explicitly if you are +// calling the REST API directly. +func (client PoolClient) Add(ctx context.Context, pool PoolAddParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: pool, + Constraints: []validation.Constraint{{Target: "pool.ID", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "pool.VMSize", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "pool.CloudServiceConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "pool.CloudServiceConfiguration.OsFamily", Name: validation.Null, Rule: true, Chain: nil}}}, + {Target: "pool.VirtualMachineConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "pool.VirtualMachineConfiguration.ImageReference", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "pool.VirtualMachineConfiguration.NodeAgentSKUID", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "pool.VirtualMachineConfiguration.ContainerConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "pool.VirtualMachineConfiguration.ContainerConfiguration.Type", Name: validation.Null, Rule: true, Chain: nil}}}, + }}, + {Target: "pool.NetworkConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "pool.NetworkConfiguration.EndpointConfiguration", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "pool.NetworkConfiguration.EndpointConfiguration.InboundNATPools", Name: validation.Null, Rule: true, Chain: nil}}}, + }}, + {Target: "pool.StartTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "pool.StartTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "pool.StartTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "pool.StartTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "pool.StartTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "pool.StartTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "pool.StartTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}}}}); err != nil { + return result, validation.NewError("batch.PoolClient", "Add", err.Error()) + } + + req, err := client.AddPreparer(ctx, pool, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Add", nil, "Failure preparing request") + return + } + + resp, err := client.AddSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Add", resp, "Failure sending request") + return + } + + result, err = client.AddResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Add", resp, "Failure responding to request") + } + + return +} + +// AddPreparer prepares the Add request. +func (client PoolClient) AddPreparer(ctx context.Context, pool PoolAddParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPath("/pools"), + autorest.WithJSON(pool), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// AddSender sends the Add request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) AddSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// AddResponder handles the response to the Add request. The method always +// closes the http.Response Body. +func (client PoolClient) AddResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByClosing()) + result.Response = resp + return +} + +// Delete when you request that a pool be deleted, the following actions occur: the pool state is set to deleting; any +// ongoing resize operation on the pool are stopped; the Batch service starts resizing the pool to zero nodes; any +// tasks running on existing nodes are terminated and requeued (as if a resize pool operation had been requested with +// the default requeue option); finally, the pool is removed from the system. Because running tasks are requeued, the +// user can rerun these tasks by updating their job to target a different pool. The tasks can then run on the new pool. +// If you want to override the requeue behavior, then you should call resize pool explicitly to shrink the pool to zero +// size before deleting the pool. If you call an Update, Patch or Delete API on a pool in the deleting state, it will +// fail with HTTP status code 409 with error code PoolBeingDeleted. +// +// poolID is the ID of the pool to delete. timeout is the maximum time that the server can spend processing the +// request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the +// form of a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. +// returnClientRequestID is whether the server should return the client-request-id in the response. ocpDate is the +// time the request was issued. Client libraries typically set this to the current system clock time; set it +// explicitly if you are calling the REST API directly. ifMatch is an ETag value associated with the version of the +// resource known to the client. The operation will be performed only if the resource's current ETag on the service +// exactly matches the value specified by the client. ifNoneMatch is an ETag value associated with the version of +// the resource known to the client. The operation will be performed only if the resource's current ETag on the +// service does not match the value specified by the client. ifModifiedSince is a timestamp indicating the last +// modified time of the resource known to the client. The operation will be performed only if the resource on the +// service has been modified since the specified time. ifUnmodifiedSince is a timestamp indicating the last +// modified time of the resource known to the client. The operation will be performed only if the resource on the +// service has not been modified since the specified time. +func (client PoolClient) Delete(ctx context.Context, poolID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.DeletePreparer(ctx, poolID, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Delete", nil, "Failure preparing request") + return + } + + resp, err := client.DeleteSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Delete", resp, "Failure sending request") + return + } + + result, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Delete", resp, "Failure responding to request") + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client PoolClient) DeletePreparer(ctx context.Context, poolID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) DeleteSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client PoolClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// DisableAutoScale sends the disable auto scale request. +// +// poolID is the ID of the pool on which to disable automatic scaling. timeout is the maximum time that the server +// can spend processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated +// request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client PoolClient) DisableAutoScale(ctx context.Context, poolID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.DisableAutoScalePreparer(ctx, poolID, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "DisableAutoScale", nil, "Failure preparing request") + return + } + + resp, err := client.DisableAutoScaleSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.PoolClient", "DisableAutoScale", resp, "Failure sending request") + return + } + + result, err = client.DisableAutoScaleResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "DisableAutoScale", resp, "Failure responding to request") + } + + return +} + +// DisableAutoScalePreparer prepares the DisableAutoScale request. +func (client PoolClient) DisableAutoScalePreparer(ctx context.Context, poolID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/disableautoscale", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DisableAutoScaleSender sends the DisableAutoScale request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) DisableAutoScaleSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// DisableAutoScaleResponder handles the response to the DisableAutoScale request. The method always +// closes the http.Response Body. +func (client PoolClient) DisableAutoScaleResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} + +// EnableAutoScale you cannot enable automatic scaling on a pool if a resize operation is in progress on the pool. If +// automatic scaling of the pool is currently disabled, you must specify a valid autoscale formula as part of the +// request. If automatic scaling of the pool is already enabled, you may specify a new autoscale formula and/or a new +// evaluation interval. You cannot call this API for the same pool more than once every 30 seconds. +// +// poolID is the ID of the pool on which to enable automatic scaling. poolEnableAutoScaleParameter is the +// parameters for the request. timeout is the maximum time that the server can spend processing the request, in +// seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a +// GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID +// is whether the server should return the client-request-id in the response. ocpDate is the time the request was +// issued. Client libraries typically set this to the current system clock time; set it explicitly if you are +// calling the REST API directly. ifMatch is an ETag value associated with the version of the resource known to the +// client. The operation will be performed only if the resource's current ETag on the service exactly matches the +// value specified by the client. ifNoneMatch is an ETag value associated with the version of the resource known to +// the client. The operation will be performed only if the resource's current ETag on the service does not match +// the value specified by the client. ifModifiedSince is a timestamp indicating the last modified time of the +// resource known to the client. The operation will be performed only if the resource on the service has been +// modified since the specified time. ifUnmodifiedSince is a timestamp indicating the last modified time of the +// resource known to the client. The operation will be performed only if the resource on the service has not been +// modified since the specified time. +func (client PoolClient) EnableAutoScale(ctx context.Context, poolID string, poolEnableAutoScaleParameter PoolEnableAutoScaleParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.EnableAutoScalePreparer(ctx, poolID, poolEnableAutoScaleParameter, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "EnableAutoScale", nil, "Failure preparing request") + return + } + + resp, err := client.EnableAutoScaleSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.PoolClient", "EnableAutoScale", resp, "Failure sending request") + return + } + + result, err = client.EnableAutoScaleResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "EnableAutoScale", resp, "Failure responding to request") + } + + return +} + +// EnableAutoScalePreparer prepares the EnableAutoScale request. +func (client PoolClient) EnableAutoScalePreparer(ctx context.Context, poolID string, poolEnableAutoScaleParameter PoolEnableAutoScaleParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/enableautoscale", pathParameters), + autorest.WithJSON(poolEnableAutoScaleParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// EnableAutoScaleSender sends the EnableAutoScale request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) EnableAutoScaleSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// EnableAutoScaleResponder handles the response to the EnableAutoScale request. The method always +// closes the http.Response Body. +func (client PoolClient) EnableAutoScaleResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} + +// EvaluateAutoScale this API is primarily for validating an autoscale formula, as it simply returns the result without +// applying the formula to the pool. The pool must have auto scaling enabled in order to evaluate a formula. +// +// poolID is the ID of the pool on which to evaluate the automatic scaling formula. poolEvaluateAutoScaleParameter +// is the parameters for the request. timeout is the maximum time that the server can spend processing the request, +// in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the form of +// a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID +// is whether the server should return the client-request-id in the response. ocpDate is the time the request was +// issued. Client libraries typically set this to the current system clock time; set it explicitly if you are +// calling the REST API directly. +func (client PoolClient) EvaluateAutoScale(ctx context.Context, poolID string, poolEvaluateAutoScaleParameter PoolEvaluateAutoScaleParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result AutoScaleRun, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: poolEvaluateAutoScaleParameter, + Constraints: []validation.Constraint{{Target: "poolEvaluateAutoScaleParameter.AutoScaleFormula", Name: validation.Null, Rule: true, Chain: nil}}}}); err != nil { + return result, validation.NewError("batch.PoolClient", "EvaluateAutoScale", err.Error()) + } + + req, err := client.EvaluateAutoScalePreparer(ctx, poolID, poolEvaluateAutoScaleParameter, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "EvaluateAutoScale", nil, "Failure preparing request") + return + } + + resp, err := client.EvaluateAutoScaleSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.PoolClient", "EvaluateAutoScale", resp, "Failure sending request") + return + } + + result, err = client.EvaluateAutoScaleResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "EvaluateAutoScale", resp, "Failure responding to request") + } + + return +} + +// EvaluateAutoScalePreparer prepares the EvaluateAutoScale request. +func (client PoolClient) EvaluateAutoScalePreparer(ctx context.Context, poolID string, poolEvaluateAutoScaleParameter PoolEvaluateAutoScaleParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/evaluateautoscale", pathParameters), + autorest.WithJSON(poolEvaluateAutoScaleParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// EvaluateAutoScaleSender sends the EvaluateAutoScale request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) EvaluateAutoScaleSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// EvaluateAutoScaleResponder handles the response to the EvaluateAutoScale request. The method always +// closes the http.Response Body. +func (client PoolClient) EvaluateAutoScaleResponder(resp *http.Response) (result AutoScaleRun, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Exists gets basic properties of a pool. +// +// poolID is the ID of the pool to get. timeout is the maximum time that the server can spend processing the +// request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the +// form of a GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. +// returnClientRequestID is whether the server should return the client-request-id in the response. ocpDate is the +// time the request was issued. Client libraries typically set this to the current system clock time; set it +// explicitly if you are calling the REST API directly. ifMatch is an ETag value associated with the version of the +// resource known to the client. The operation will be performed only if the resource's current ETag on the service +// exactly matches the value specified by the client. ifNoneMatch is an ETag value associated with the version of +// the resource known to the client. The operation will be performed only if the resource's current ETag on the +// service does not match the value specified by the client. ifModifiedSince is a timestamp indicating the last +// modified time of the resource known to the client. The operation will be performed only if the resource on the +// service has been modified since the specified time. ifUnmodifiedSince is a timestamp indicating the last +// modified time of the resource known to the client. The operation will be performed only if the resource on the +// service has not been modified since the specified time. +func (client PoolClient) Exists(ctx context.Context, poolID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.ExistsPreparer(ctx, poolID, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Exists", nil, "Failure preparing request") + return + } + + resp, err := client.ExistsSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Exists", resp, "Failure sending request") + return + } + + result, err = client.ExistsResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Exists", resp, "Failure responding to request") + } + + return +} + +// ExistsPreparer prepares the Exists request. +func (client PoolClient) ExistsPreparer(ctx context.Context, poolID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsHead(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ExistsSender sends the Exists request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) ExistsSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ExistsResponder handles the response to the Exists request. The method always +// closes the http.Response Body. +func (client PoolClient) ExistsResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusNotFound), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get gets information about the specified pool. +// +// poolID is the ID of the pool to get. selectParameter is an OData $select clause. expand is an OData $expand +// clause. timeout is the maximum time that the server can spend processing the request, in seconds. The default is +// 30 seconds. clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration +// such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server +// should return the client-request-id in the response. ocpDate is the time the request was issued. Client +// libraries typically set this to the current system clock time; set it explicitly if you are calling the REST API +// directly. ifMatch is an ETag value associated with the version of the resource known to the client. The +// operation will be performed only if the resource's current ETag on the service exactly matches the value +// specified by the client. ifNoneMatch is an ETag value associated with the version of the resource known to the +// client. The operation will be performed only if the resource's current ETag on the service does not match the +// value specified by the client. ifModifiedSince is a timestamp indicating the last modified time of the resource +// known to the client. The operation will be performed only if the resource on the service has been modified since +// the specified time. ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to +// the client. The operation will be performed only if the resource on the service has not been modified since the +// specified time. +func (client PoolClient) Get(ctx context.Context, poolID string, selectParameter string, expand string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result CloudPool, err error) { + req, err := client.GetPreparer(ctx, poolID, selectParameter, expand, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client PoolClient) GetPreparer(ctx context.Context, poolID string, selectParameter string, expand string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client PoolClient) GetResponder(resp *http.Response) (result CloudPool, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// GetAllLifetimeStatistics statistics are aggregated across all pools that have ever existed in the account, from +// account creation to the last update time of the statistics. +// +// timeout is the maximum time that the server can spend processing the request, in seconds. The default is 30 +// seconds. clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such +// as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should +// return the client-request-id in the response. ocpDate is the time the request was issued. Client libraries +// typically set this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client PoolClient) GetAllLifetimeStatistics(ctx context.Context, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result PoolStatistics, err error) { + req, err := client.GetAllLifetimeStatisticsPreparer(ctx, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "GetAllLifetimeStatistics", nil, "Failure preparing request") + return + } + + resp, err := client.GetAllLifetimeStatisticsSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.PoolClient", "GetAllLifetimeStatistics", resp, "Failure sending request") + return + } + + result, err = client.GetAllLifetimeStatisticsResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "GetAllLifetimeStatistics", resp, "Failure responding to request") + } + + return +} + +// GetAllLifetimeStatisticsPreparer prepares the GetAllLifetimeStatistics request. +func (client PoolClient) GetAllLifetimeStatisticsPreparer(ctx context.Context, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPath("/lifetimepoolstats"), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetAllLifetimeStatisticsSender sends the GetAllLifetimeStatistics request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) GetAllLifetimeStatisticsSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetAllLifetimeStatisticsResponder handles the response to the GetAllLifetimeStatistics request. The method always +// closes the http.Response Body. +func (client PoolClient) GetAllLifetimeStatisticsResponder(resp *http.Response) (result PoolStatistics, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List sends the list request. +// +// filter is an OData $filter clause. For more information on constructing this filter, see +// https://docs.microsoft.com/en-us/rest/api/batchservice/odata-filters-in-batch#list-pools. selectParameter is an +// OData $select clause. expand is an OData $expand clause. maxResults is the maximum number of items to return in +// the response. A maximum of 1000 pools can be returned. timeout is the maximum time that the server can spend +// processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request +// identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client PoolClient) List(ctx context.Context, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CloudPoolListResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: maxResults, + Constraints: []validation.Constraint{{Target: "maxResults", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "maxResults", Name: validation.InclusiveMaximum, Rule: 1000, Chain: nil}, + {Target: "maxResults", Name: validation.InclusiveMinimum, Rule: 1, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("batch.PoolClient", "List", err.Error()) + } + + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, filter, selectParameter, expand, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.cplr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.PoolClient", "List", resp, "Failure sending request") + return + } + + result.cplr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client PoolClient) ListPreparer(ctx context.Context, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(filter) > 0 { + queryParameters["$filter"] = autorest.Encode("query", filter) + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + if maxResults != nil { + queryParameters["maxresults"] = autorest.Encode("query", *maxResults) + } else { + queryParameters["maxresults"] = autorest.Encode("query", 1000) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPath("/pools"), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client PoolClient) ListResponder(resp *http.Response) (result CloudPoolListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client PoolClient) listNextResults(lastResults CloudPoolListResult) (result CloudPoolListResult, err error) { + req, err := lastResults.cloudPoolListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "batch.PoolClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "batch.PoolClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client PoolClient) ListComplete(ctx context.Context, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CloudPoolListResultIterator, err error) { + result.page, err = client.List(ctx, filter, selectParameter, expand, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + return +} + +// ListUsageMetrics if you do not specify a $filter clause including a poolId, the response includes all pools that +// existed in the account in the time range of the returned aggregation intervals. If you do not specify a $filter +// clause including a startTime or endTime these filters default to the start and end times of the last aggregation +// interval currently available; that is, only the last aggregation interval is returned. +// +// startTime is the earliest time from which to include metrics. This must be at least two and a half hours before +// the current time. If not specified this defaults to the start time of the last aggregation interval currently +// available. endTime is the latest time from which to include metrics. This must be at least two hours before the +// current time. If not specified this defaults to the end time of the last aggregation interval currently +// available. filter is an OData $filter clause. For more information on constructing this filter, see +// https://docs.microsoft.com/en-us/rest/api/batchservice/odata-filters-in-batch#list-account-usage-metrics. +// maxResults is the maximum number of items to return in the response. A maximum of 1000 results will be returned. +// timeout is the maximum time that the server can spend processing the request, in seconds. The default is 30 +// seconds. clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such +// as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should +// return the client-request-id in the response. ocpDate is the time the request was issued. Client libraries +// typically set this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client PoolClient) ListUsageMetrics(ctx context.Context, startTime *date.Time, endTime *date.Time, filter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result PoolListUsageMetricsResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: maxResults, + Constraints: []validation.Constraint{{Target: "maxResults", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "maxResults", Name: validation.InclusiveMaximum, Rule: 1000, Chain: nil}, + {Target: "maxResults", Name: validation.InclusiveMinimum, Rule: 1, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("batch.PoolClient", "ListUsageMetrics", err.Error()) + } + + result.fn = client.listUsageMetricsNextResults + req, err := client.ListUsageMetricsPreparer(ctx, startTime, endTime, filter, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "ListUsageMetrics", nil, "Failure preparing request") + return + } + + resp, err := client.ListUsageMetricsSender(req) + if err != nil { + result.plumr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.PoolClient", "ListUsageMetrics", resp, "Failure sending request") + return + } + + result.plumr, err = client.ListUsageMetricsResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "ListUsageMetrics", resp, "Failure responding to request") + } + + return +} + +// ListUsageMetricsPreparer prepares the ListUsageMetrics request. +func (client PoolClient) ListUsageMetricsPreparer(ctx context.Context, startTime *date.Time, endTime *date.Time, filter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if startTime != nil { + queryParameters["starttime"] = autorest.Encode("query", *startTime) + } + if endTime != nil { + queryParameters["endtime"] = autorest.Encode("query", *endTime) + } + if len(filter) > 0 { + queryParameters["$filter"] = autorest.Encode("query", filter) + } + if maxResults != nil { + queryParameters["maxresults"] = autorest.Encode("query", *maxResults) + } else { + queryParameters["maxresults"] = autorest.Encode("query", 1000) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPath("/poolusagemetrics"), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListUsageMetricsSender sends the ListUsageMetrics request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) ListUsageMetricsSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListUsageMetricsResponder handles the response to the ListUsageMetrics request. The method always +// closes the http.Response Body. +func (client PoolClient) ListUsageMetricsResponder(resp *http.Response) (result PoolListUsageMetricsResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listUsageMetricsNextResults retrieves the next set of results, if any. +func (client PoolClient) listUsageMetricsNextResults(lastResults PoolListUsageMetricsResult) (result PoolListUsageMetricsResult, err error) { + req, err := lastResults.poolListUsageMetricsResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "batch.PoolClient", "listUsageMetricsNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListUsageMetricsSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "batch.PoolClient", "listUsageMetricsNextResults", resp, "Failure sending next results request") + } + result, err = client.ListUsageMetricsResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "listUsageMetricsNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListUsageMetricsComplete enumerates all values, automatically crossing page boundaries as required. +func (client PoolClient) ListUsageMetricsComplete(ctx context.Context, startTime *date.Time, endTime *date.Time, filter string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result PoolListUsageMetricsResultIterator, err error) { + result.page, err = client.ListUsageMetrics(ctx, startTime, endTime, filter, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + return +} + +// Patch this only replaces the pool properties specified in the request. For example, if the pool has a start task +// associated with it, and a request does not specify a start task element, then the pool keeps the existing start +// task. +// +// poolID is the ID of the pool to update. poolPatchParameter is the parameters for the request. timeout is the +// maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client PoolClient) Patch(ctx context.Context, poolID string, poolPatchParameter PoolPatchParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.PatchPreparer(ctx, poolID, poolPatchParameter, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Patch", nil, "Failure preparing request") + return + } + + resp, err := client.PatchSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Patch", resp, "Failure sending request") + return + } + + result, err = client.PatchResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Patch", resp, "Failure responding to request") + } + + return +} + +// PatchPreparer prepares the Patch request. +func (client PoolClient) PatchPreparer(ctx context.Context, poolID string, poolPatchParameter PoolPatchParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPatch(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}", pathParameters), + autorest.WithJSON(poolPatchParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// PatchSender sends the Patch request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) PatchSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// PatchResponder handles the response to the Patch request. The method always +// closes the http.Response Body. +func (client PoolClient) PatchResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} + +// RemoveNodes this operation can only run when the allocation state of the pool is steady. When this operation runs, +// the allocation state changes from steady to resizing. +// +// poolID is the ID of the pool from which you want to remove nodes. nodeRemoveParameter is the parameters for the +// request. timeout is the maximum time that the server can spend processing the request, in seconds. The default +// is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a GUID with no +// decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the +// server should return the client-request-id in the response. ocpDate is the time the request was issued. Client +// libraries typically set this to the current system clock time; set it explicitly if you are calling the REST API +// directly. ifMatch is an ETag value associated with the version of the resource known to the client. The +// operation will be performed only if the resource's current ETag on the service exactly matches the value +// specified by the client. ifNoneMatch is an ETag value associated with the version of the resource known to the +// client. The operation will be performed only if the resource's current ETag on the service does not match the +// value specified by the client. ifModifiedSince is a timestamp indicating the last modified time of the resource +// known to the client. The operation will be performed only if the resource on the service has been modified since +// the specified time. ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to +// the client. The operation will be performed only if the resource on the service has not been modified since the +// specified time. +func (client PoolClient) RemoveNodes(ctx context.Context, poolID string, nodeRemoveParameter NodeRemoveParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: nodeRemoveParameter, + Constraints: []validation.Constraint{{Target: "nodeRemoveParameter.NodeList", Name: validation.Null, Rule: true, + Chain: []validation.Constraint{{Target: "nodeRemoveParameter.NodeList", Name: validation.MaxItems, Rule: 100, Chain: nil}}}}}}); err != nil { + return result, validation.NewError("batch.PoolClient", "RemoveNodes", err.Error()) + } + + req, err := client.RemoveNodesPreparer(ctx, poolID, nodeRemoveParameter, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "RemoveNodes", nil, "Failure preparing request") + return + } + + resp, err := client.RemoveNodesSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.PoolClient", "RemoveNodes", resp, "Failure sending request") + return + } + + result, err = client.RemoveNodesResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "RemoveNodes", resp, "Failure responding to request") + } + + return +} + +// RemoveNodesPreparer prepares the RemoveNodes request. +func (client PoolClient) RemoveNodesPreparer(ctx context.Context, poolID string, nodeRemoveParameter NodeRemoveParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/removenodes", pathParameters), + autorest.WithJSON(nodeRemoveParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// RemoveNodesSender sends the RemoveNodes request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) RemoveNodesSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// RemoveNodesResponder handles the response to the RemoveNodes request. The method always +// closes the http.Response Body. +func (client PoolClient) RemoveNodesResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// Resize you can only resize a pool when its allocation state is steady. If the pool is already resizing, the request +// fails with status code 409. When you resize a pool, the pool's allocation state changes from steady to resizing. You +// cannot resize pools which are configured for automatic scaling. If you try to do this, the Batch service returns an +// error 409. If you resize a pool downwards, the Batch service chooses which nodes to remove. To remove specific +// nodes, use the pool remove nodes API instead. +// +// poolID is the ID of the pool to resize. poolResizeParameter is the parameters for the request. timeout is the +// maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client PoolClient) Resize(ctx context.Context, poolID string, poolResizeParameter PoolResizeParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.ResizePreparer(ctx, poolID, poolResizeParameter, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Resize", nil, "Failure preparing request") + return + } + + resp, err := client.ResizeSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Resize", resp, "Failure sending request") + return + } + + result, err = client.ResizeResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "Resize", resp, "Failure responding to request") + } + + return +} + +// ResizePreparer prepares the Resize request. +func (client PoolClient) ResizePreparer(ctx context.Context, poolID string, poolResizeParameter PoolResizeParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/resize", pathParameters), + autorest.WithJSON(poolResizeParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ResizeSender sends the Resize request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) ResizeSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ResizeResponder handles the response to the Resize request. The method always +// closes the http.Response Body. +func (client PoolClient) ResizeResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// StopResize this does not restore the pool to its previous state before the resize operation: it only stops any +// further changes being made, and the pool maintains its current state. After stopping, the pool stabilizes at the +// number of nodes it was at when the stop operation was done. During the stop operation, the pool allocation state +// changes first to stopping and then to steady. A resize operation need not be an explicit resize pool request; this +// API can also be used to halt the initial sizing of the pool when it is created. +// +// poolID is the ID of the pool whose resizing you want to stop. timeout is the maximum time that the server can +// spend processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated +// request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client PoolClient) StopResize(ctx context.Context, poolID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.StopResizePreparer(ctx, poolID, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "StopResize", nil, "Failure preparing request") + return + } + + resp, err := client.StopResizeSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.PoolClient", "StopResize", resp, "Failure sending request") + return + } + + result, err = client.StopResizeResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "StopResize", resp, "Failure responding to request") + } + + return +} + +// StopResizePreparer prepares the StopResize request. +func (client PoolClient) StopResizePreparer(ctx context.Context, poolID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/stopresize", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// StopResizeSender sends the StopResize request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) StopResizeSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// StopResizeResponder handles the response to the StopResize request. The method always +// closes the http.Response Body. +func (client PoolClient) StopResizeResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} + +// UpdateProperties this fully replaces all the updateable properties of the pool. For example, if the pool has a start +// task associated with it and if start task is not specified with this request, then the Batch service will remove the +// existing start task. +// +// poolID is the ID of the pool to update. poolUpdatePropertiesParameter is the parameters for the request. timeout +// is the maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client PoolClient) UpdateProperties(ctx context.Context, poolID string, poolUpdatePropertiesParameter PoolUpdatePropertiesParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: poolUpdatePropertiesParameter, + Constraints: []validation.Constraint{{Target: "poolUpdatePropertiesParameter.StartTask", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "poolUpdatePropertiesParameter.StartTask.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "poolUpdatePropertiesParameter.StartTask.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "poolUpdatePropertiesParameter.StartTask.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "poolUpdatePropertiesParameter.StartTask.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "poolUpdatePropertiesParameter.StartTask.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "poolUpdatePropertiesParameter.StartTask.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + }}, + {Target: "poolUpdatePropertiesParameter.CertificateReferences", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "poolUpdatePropertiesParameter.ApplicationPackageReferences", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "poolUpdatePropertiesParameter.Metadata", Name: validation.Null, Rule: true, Chain: nil}}}}); err != nil { + return result, validation.NewError("batch.PoolClient", "UpdateProperties", err.Error()) + } + + req, err := client.UpdatePropertiesPreparer(ctx, poolID, poolUpdatePropertiesParameter, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "UpdateProperties", nil, "Failure preparing request") + return + } + + resp, err := client.UpdatePropertiesSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.PoolClient", "UpdateProperties", resp, "Failure sending request") + return + } + + result, err = client.UpdatePropertiesResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "UpdateProperties", resp, "Failure responding to request") + } + + return +} + +// UpdatePropertiesPreparer prepares the UpdateProperties request. +func (client PoolClient) UpdatePropertiesPreparer(ctx context.Context, poolID string, poolUpdatePropertiesParameter PoolUpdatePropertiesParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/updateproperties", pathParameters), + autorest.WithJSON(poolUpdatePropertiesParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// UpdatePropertiesSender sends the UpdateProperties request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) UpdatePropertiesSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// UpdatePropertiesResponder handles the response to the UpdateProperties request. The method always +// closes the http.Response Body. +func (client PoolClient) UpdatePropertiesResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// UpgradeOS during an upgrade, the Batch service upgrades each compute node in the pool. When a compute node is chosen +// for upgrade, any tasks running on that node are removed from the node and returned to the queue to be rerun later +// (or on a different compute node). The node will be unavailable until the upgrade is complete. This operation results +// in temporarily reduced pool capacity as nodes are taken out of service to be upgraded. Although the Batch service +// tries to avoid upgrading all compute nodes at the same time, it does not guarantee to do this (particularly on small +// pools); therefore, the pool may be temporarily unavailable to run tasks. When this operation runs, the pool state +// changes to upgrading. When all compute nodes have finished upgrading, the pool state returns to active. While the +// upgrade is in progress, the pool's currentOSVersion reflects the OS version that nodes are upgrading from, and +// targetOSVersion reflects the OS version that nodes are upgrading to. Once the upgrade is complete, currentOSVersion +// is updated to reflect the OS version now running on all nodes. This operation can only be invoked on pools created +// with the cloudServiceConfiguration property. +// +// poolID is the ID of the pool to upgrade. poolUpgradeOSParameter is the parameters for the request. timeout is +// the maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client PoolClient) UpgradeOS(ctx context.Context, poolID string, poolUpgradeOSParameter PoolUpgradeOSParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: poolUpgradeOSParameter, + Constraints: []validation.Constraint{{Target: "poolUpgradeOSParameter.TargetOSVersion", Name: validation.Null, Rule: true, Chain: nil}}}}); err != nil { + return result, validation.NewError("batch.PoolClient", "UpgradeOS", err.Error()) + } + + req, err := client.UpgradeOSPreparer(ctx, poolID, poolUpgradeOSParameter, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "UpgradeOS", nil, "Failure preparing request") + return + } + + resp, err := client.UpgradeOSSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.PoolClient", "UpgradeOS", resp, "Failure sending request") + return + } + + result, err = client.UpgradeOSResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.PoolClient", "UpgradeOS", resp, "Failure responding to request") + } + + return +} + +// UpgradeOSPreparer prepares the UpgradeOS request. +func (client PoolClient) UpgradeOSPreparer(ctx context.Context, poolID string, poolUpgradeOSParameter PoolUpgradeOSParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "poolId": autorest.Encode("path", poolID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/pools/{poolId}/upgradeos", pathParameters), + autorest.WithJSON(poolUpgradeOSParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// UpgradeOSSender sends the UpgradeOS request. The method will close the +// http.Response Body if it receives an error. +func (client PoolClient) UpgradeOSSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// UpgradeOSResponder handles the response to the UpgradeOS request. The method always +// closes the http.Response Body. +func (client PoolClient) UpgradeOSResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted), + autorest.ByClosing()) + result.Response = resp + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/task.go b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/task.go new file mode 100644 index 000000000..f2f82a81f --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/task.go @@ -0,0 +1,1081 @@ +package batch + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +import ( + "context" + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/azure" + "github.com/Azure/go-autorest/autorest/date" + "github.com/Azure/go-autorest/autorest/validation" + "github.com/satori/go.uuid" + "net/http" +) + +// TaskClient is the a client for issuing REST requests to the Azure Batch service. +type TaskClient struct { + BaseClient +} + +// NewTaskClient creates an instance of the TaskClient client. +func NewTaskClient() TaskClient { + return NewTaskClientWithBaseURI(DefaultBaseURI) +} + +// NewTaskClientWithBaseURI creates an instance of the TaskClient client. +func NewTaskClientWithBaseURI(baseURI string) TaskClient { + return TaskClient{NewWithBaseURI(baseURI)} +} + +// Add the maximum lifetime of a task from addition to completion is 7 days. If a task has not completed within 7 days +// of being added it will be terminated by the Batch service and left in whatever state it was in at that time. +// +// jobID is the ID of the job to which the task is to be added. task is the task to be added. timeout is the +// maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client TaskClient) Add(ctx context.Context, jobID string, task TaskAddParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result autorest.Response, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: task, + Constraints: []validation.Constraint{{Target: "task.ID", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "task.CommandLine", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "task.ContainerSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "task.ContainerSettings.ImageName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "task.ContainerSettings.Registry", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "task.ContainerSettings.Registry.UserName", Name: validation.Null, Rule: true, Chain: nil}, + {Target: "task.ContainerSettings.Registry.Password", Name: validation.Null, Rule: true, Chain: nil}, + }}, + }}, + {Target: "task.AffinityInfo", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "task.AffinityInfo.AffinityID", Name: validation.Null, Rule: true, Chain: nil}}}, + {Target: "task.MultiInstanceSettings", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "task.MultiInstanceSettings.CoordinationCommandLine", Name: validation.Null, Rule: true, Chain: nil}}}}}}); err != nil { + return result, validation.NewError("batch.TaskClient", "Add", err.Error()) + } + + req, err := client.AddPreparer(ctx, jobID, task, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Add", nil, "Failure preparing request") + return + } + + resp, err := client.AddSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Add", resp, "Failure sending request") + return + } + + result, err = client.AddResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Add", resp, "Failure responding to request") + } + + return +} + +// AddPreparer prepares the Add request. +func (client TaskClient) AddPreparer(ctx context.Context, jobID string, task TaskAddParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/tasks", pathParameters), + autorest.WithJSON(task), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// AddSender sends the Add request. The method will close the +// http.Response Body if it receives an error. +func (client TaskClient) AddSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// AddResponder handles the response to the Add request. The method always +// closes the http.Response Body. +func (client TaskClient) AddResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusCreated), + autorest.ByClosing()) + result.Response = resp + return +} + +// AddCollection note that each task must have a unique ID. The Batch service may not return the results for each task +// in the same order the tasks were submitted in this request. If the server times out or the connection is closed +// during the request, the request may have been partially or fully processed, or not at all. In such cases, the user +// should re-issue the request. Note that it is up to the user to correctly handle failures when re-issuing a request. +// For example, you should use the same task IDs during a retry so that if the prior operation succeeded, the retry +// will not create extra tasks unexpectedly. If the response contains any tasks which failed to add, a client can retry +// the request. In a retry, it is most efficient to resubmit only tasks that failed to add, and to omit tasks that were +// successfully added on the first attempt. The maximum lifetime of a task from addition to completion is 7 days. If a +// task has not completed within 7 days of being added it will be terminated by the Batch service and left in whatever +// state it was in at that time. +// +// jobID is the ID of the job to which the task collection is to be added. taskCollection is the tasks to be added. +// timeout is the maximum time that the server can spend processing the request, in seconds. The default is 30 +// seconds. clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such +// as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should +// return the client-request-id in the response. ocpDate is the time the request was issued. Client libraries +// typically set this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client TaskClient) AddCollection(ctx context.Context, jobID string, taskCollection TaskAddCollectionParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result TaskAddCollectionResult, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: taskCollection, + Constraints: []validation.Constraint{{Target: "taskCollection.Value", Name: validation.Null, Rule: true, + Chain: []validation.Constraint{{Target: "taskCollection.Value", Name: validation.MaxItems, Rule: 100, Chain: nil}}}}}}); err != nil { + return result, validation.NewError("batch.TaskClient", "AddCollection", err.Error()) + } + + req, err := client.AddCollectionPreparer(ctx, jobID, taskCollection, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "AddCollection", nil, "Failure preparing request") + return + } + + resp, err := client.AddCollectionSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.TaskClient", "AddCollection", resp, "Failure sending request") + return + } + + result, err = client.AddCollectionResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "AddCollection", resp, "Failure responding to request") + } + + return +} + +// AddCollectionPreparer prepares the AddCollection request. +func (client TaskClient) AddCollectionPreparer(ctx context.Context, jobID string, taskCollection TaskAddCollectionParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/addtaskcollection", pathParameters), + autorest.WithJSON(taskCollection), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// AddCollectionSender sends the AddCollection request. The method will close the +// http.Response Body if it receives an error. +func (client TaskClient) AddCollectionSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// AddCollectionResponder handles the response to the AddCollection request. The method always +// closes the http.Response Body. +func (client TaskClient) AddCollectionResponder(resp *http.Response) (result TaskAddCollectionResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Delete when a task is deleted, all of the files in its directory on the compute node where it ran are also deleted +// (regardless of the retention time). For multi-instance tasks, the delete task operation applies synchronously to the +// primary task; subtasks and their files are then deleted asynchronously in the background. +// +// jobID is the ID of the job from which to delete the task. taskID is the ID of the task to delete. timeout is the +// maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client TaskClient) Delete(ctx context.Context, jobID string, taskID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.DeletePreparer(ctx, jobID, taskID, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Delete", nil, "Failure preparing request") + return + } + + resp, err := client.DeleteSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Delete", resp, "Failure sending request") + return + } + + result, err = client.DeleteResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Delete", resp, "Failure responding to request") + } + + return +} + +// DeletePreparer prepares the Delete request. +func (client TaskClient) DeletePreparer(ctx context.Context, jobID string, taskID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + "taskId": autorest.Encode("path", taskID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsDelete(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/tasks/{taskId}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// DeleteSender sends the Delete request. The method will close the +// http.Response Body if it receives an error. +func (client TaskClient) DeleteSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// DeleteResponder handles the response to the Delete request. The method always +// closes the http.Response Body. +func (client TaskClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} + +// Get for multi-instance tasks, information such as affinityId, executionInfo and nodeInfo refer to the primary task. +// Use the list subtasks API to retrieve information about subtasks. +// +// jobID is the ID of the job that contains the task. taskID is the ID of the task to get information about. +// selectParameter is an OData $select clause. expand is an OData $expand clause. timeout is the maximum time that +// the server can spend processing the request, in seconds. The default is 30 seconds. clientRequestID is the +// caller-generated request identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client TaskClient) Get(ctx context.Context, jobID string, taskID string, selectParameter string, expand string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result CloudTask, err error) { + req, err := client.GetPreparer(ctx, jobID, taskID, selectParameter, expand, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Get", nil, "Failure preparing request") + return + } + + resp, err := client.GetSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Get", resp, "Failure sending request") + return + } + + result, err = client.GetResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Get", resp, "Failure responding to request") + } + + return +} + +// GetPreparer prepares the Get request. +func (client TaskClient) GetPreparer(ctx context.Context, jobID string, taskID string, selectParameter string, expand string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + "taskId": autorest.Encode("path", taskID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/tasks/{taskId}", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// GetSender sends the Get request. The method will close the +// http.Response Body if it receives an error. +func (client TaskClient) GetSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// GetResponder handles the response to the Get request. The method always +// closes the http.Response Body. +func (client TaskClient) GetResponder(resp *http.Response) (result CloudTask, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// List for multi-instance tasks, information such as affinityId, executionInfo and nodeInfo refer to the primary task. +// Use the list subtasks API to retrieve information about subtasks. +// +// jobID is the ID of the job. filter is an OData $filter clause. For more information on constructing this filter, +// see https://docs.microsoft.com/en-us/rest/api/batchservice/odata-filters-in-batch#list-tasks. selectParameter is +// an OData $select clause. expand is an OData $expand clause. maxResults is the maximum number of items to return +// in the response. A maximum of 1000 tasks can be returned. timeout is the maximum time that the server can spend +// processing the request, in seconds. The default is 30 seconds. clientRequestID is the caller-generated request +// identity, in the form of a GUID with no decoration such as curly braces, e.g. +// 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client TaskClient) List(ctx context.Context, jobID string, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CloudTaskListResultPage, err error) { + if err := validation.Validate([]validation.Validation{ + {TargetValue: maxResults, + Constraints: []validation.Constraint{{Target: "maxResults", Name: validation.Null, Rule: false, + Chain: []validation.Constraint{{Target: "maxResults", Name: validation.InclusiveMaximum, Rule: 1000, Chain: nil}, + {Target: "maxResults", Name: validation.InclusiveMinimum, Rule: 1, Chain: nil}, + }}}}}); err != nil { + return result, validation.NewError("batch.TaskClient", "List", err.Error()) + } + + result.fn = client.listNextResults + req, err := client.ListPreparer(ctx, jobID, filter, selectParameter, expand, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "List", nil, "Failure preparing request") + return + } + + resp, err := client.ListSender(req) + if err != nil { + result.ctlr.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.TaskClient", "List", resp, "Failure sending request") + return + } + + result.ctlr, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "List", resp, "Failure responding to request") + } + + return +} + +// ListPreparer prepares the List request. +func (client TaskClient) ListPreparer(ctx context.Context, jobID string, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(filter) > 0 { + queryParameters["$filter"] = autorest.Encode("query", filter) + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if len(expand) > 0 { + queryParameters["$expand"] = autorest.Encode("query", expand) + } + if maxResults != nil { + queryParameters["maxresults"] = autorest.Encode("query", *maxResults) + } else { + queryParameters["maxresults"] = autorest.Encode("query", 1000) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/tasks", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSender sends the List request. The method will close the +// http.Response Body if it receives an error. +func (client TaskClient) ListSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListResponder handles the response to the List request. The method always +// closes the http.Response Body. +func (client TaskClient) ListResponder(resp *http.Response) (result CloudTaskListResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// listNextResults retrieves the next set of results, if any. +func (client TaskClient) listNextResults(lastResults CloudTaskListResult) (result CloudTaskListResult, err error) { + req, err := lastResults.cloudTaskListResultPreparer() + if err != nil { + return result, autorest.NewErrorWithError(err, "batch.TaskClient", "listNextResults", nil, "Failure preparing next results request") + } + if req == nil { + return + } + resp, err := client.ListSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + return result, autorest.NewErrorWithError(err, "batch.TaskClient", "listNextResults", resp, "Failure sending next results request") + } + result, err = client.ListResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "listNextResults", resp, "Failure responding to next results request") + } + return +} + +// ListComplete enumerates all values, automatically crossing page boundaries as required. +func (client TaskClient) ListComplete(ctx context.Context, jobID string, filter string, selectParameter string, expand string, maxResults *int32, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CloudTaskListResultIterator, err error) { + result.page, err = client.List(ctx, jobID, filter, selectParameter, expand, maxResults, timeout, clientRequestID, returnClientRequestID, ocpDate) + return +} + +// ListSubtasks if the task is not a multi-instance task then this returns an empty collection. +// +// jobID is the ID of the job. taskID is the ID of the task. selectParameter is an OData $select clause. timeout is +// the maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. +func (client TaskClient) ListSubtasks(ctx context.Context, jobID string, taskID string, selectParameter string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (result CloudTaskListSubtasksResult, err error) { + req, err := client.ListSubtasksPreparer(ctx, jobID, taskID, selectParameter, timeout, clientRequestID, returnClientRequestID, ocpDate) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "ListSubtasks", nil, "Failure preparing request") + return + } + + resp, err := client.ListSubtasksSender(req) + if err != nil { + result.Response = autorest.Response{Response: resp} + err = autorest.NewErrorWithError(err, "batch.TaskClient", "ListSubtasks", resp, "Failure sending request") + return + } + + result, err = client.ListSubtasksResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "ListSubtasks", resp, "Failure responding to request") + } + + return +} + +// ListSubtasksPreparer prepares the ListSubtasks request. +func (client TaskClient) ListSubtasksPreparer(ctx context.Context, jobID string, taskID string, selectParameter string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + "taskId": autorest.Encode("path", taskID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if len(selectParameter) > 0 { + queryParameters["$select"] = autorest.Encode("query", selectParameter) + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/tasks/{taskId}/subtasksinfo", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ListSubtasksSender sends the ListSubtasks request. The method will close the +// http.Response Body if it receives an error. +func (client TaskClient) ListSubtasksSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ListSubtasksResponder handles the response to the ListSubtasks request. The method always +// closes the http.Response Body. +func (client TaskClient) ListSubtasksResponder(resp *http.Response) (result CloudTaskListSubtasksResult, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&result), + autorest.ByClosing()) + result.Response = autorest.Response{Response: resp} + return +} + +// Reactivate reactivation makes a task eligible to be retried again up to its maximum retry count. The task's state is +// changed to active. As the task is no longer in the completed state, any previous exit code or failure information is +// no longer available after reactivation. Each time a task is reactivated, its retry count is reset to 0. Reactivation +// will fail for tasks that are not completed or that previously completed successfully (with an exit code of 0). +// Additionally, it will fail if the job has completed (or is terminating or deleting). +// +// jobID is the ID of the job containing the task. taskID is the ID of the task to reactivate. timeout is the +// maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client TaskClient) Reactivate(ctx context.Context, jobID string, taskID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.ReactivatePreparer(ctx, jobID, taskID, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Reactivate", nil, "Failure preparing request") + return + } + + resp, err := client.ReactivateSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Reactivate", resp, "Failure sending request") + return + } + + result, err = client.ReactivateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Reactivate", resp, "Failure responding to request") + } + + return +} + +// ReactivatePreparer prepares the Reactivate request. +func (client TaskClient) ReactivatePreparer(ctx context.Context, jobID string, taskID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + "taskId": autorest.Encode("path", taskID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/tasks/{taskId}/reactivate", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// ReactivateSender sends the Reactivate request. The method will close the +// http.Response Body if it receives an error. +func (client TaskClient) ReactivateSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// ReactivateResponder handles the response to the Reactivate request. The method always +// closes the http.Response Body. +func (client TaskClient) ReactivateResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Terminate when the task has been terminated, it moves to the completed state. For multi-instance tasks, the +// terminate task operation applies synchronously to the primary task; subtasks are then terminated asynchronously in +// the background. +// +// jobID is the ID of the job containing the task. taskID is the ID of the task to terminate. timeout is the +// maximum time that the server can spend processing the request, in seconds. The default is 30 seconds. +// clientRequestID is the caller-generated request identity, in the form of a GUID with no decoration such as curly +// braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID is whether the server should return the +// client-request-id in the response. ocpDate is the time the request was issued. Client libraries typically set +// this to the current system clock time; set it explicitly if you are calling the REST API directly. ifMatch is an +// ETag value associated with the version of the resource known to the client. The operation will be performed only +// if the resource's current ETag on the service exactly matches the value specified by the client. ifNoneMatch is +// an ETag value associated with the version of the resource known to the client. The operation will be performed +// only if the resource's current ETag on the service does not match the value specified by the client. +// ifModifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has been modified since the specified time. +// ifUnmodifiedSince is a timestamp indicating the last modified time of the resource known to the client. The +// operation will be performed only if the resource on the service has not been modified since the specified time. +func (client TaskClient) Terminate(ctx context.Context, jobID string, taskID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.TerminatePreparer(ctx, jobID, taskID, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Terminate", nil, "Failure preparing request") + return + } + + resp, err := client.TerminateSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Terminate", resp, "Failure sending request") + return + } + + result, err = client.TerminateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Terminate", resp, "Failure responding to request") + } + + return +} + +// TerminatePreparer prepares the Terminate request. +func (client TaskClient) TerminatePreparer(ctx context.Context, jobID string, taskID string, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + "taskId": autorest.Encode("path", taskID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsPost(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/tasks/{taskId}/terminate", pathParameters), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// TerminateSender sends the Terminate request. The method will close the +// http.Response Body if it receives an error. +func (client TaskClient) TerminateSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// TerminateResponder handles the response to the Terminate request. The method always +// closes the http.Response Body. +func (client TaskClient) TerminateResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusNoContent), + autorest.ByClosing()) + result.Response = resp + return +} + +// Update updates the properties of the specified task. +// +// jobID is the ID of the job containing the task. taskID is the ID of the task to update. taskUpdateParameter is +// the parameters for the request. timeout is the maximum time that the server can spend processing the request, in +// seconds. The default is 30 seconds. clientRequestID is the caller-generated request identity, in the form of a +// GUID with no decoration such as curly braces, e.g. 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0. returnClientRequestID +// is whether the server should return the client-request-id in the response. ocpDate is the time the request was +// issued. Client libraries typically set this to the current system clock time; set it explicitly if you are +// calling the REST API directly. ifMatch is an ETag value associated with the version of the resource known to the +// client. The operation will be performed only if the resource's current ETag on the service exactly matches the +// value specified by the client. ifNoneMatch is an ETag value associated with the version of the resource known to +// the client. The operation will be performed only if the resource's current ETag on the service does not match +// the value specified by the client. ifModifiedSince is a timestamp indicating the last modified time of the +// resource known to the client. The operation will be performed only if the resource on the service has been +// modified since the specified time. ifUnmodifiedSince is a timestamp indicating the last modified time of the +// resource known to the client. The operation will be performed only if the resource on the service has not been +// modified since the specified time. +func (client TaskClient) Update(ctx context.Context, jobID string, taskID string, taskUpdateParameter TaskUpdateParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (result autorest.Response, err error) { + req, err := client.UpdatePreparer(ctx, jobID, taskID, taskUpdateParameter, timeout, clientRequestID, returnClientRequestID, ocpDate, ifMatch, ifNoneMatch, ifModifiedSince, ifUnmodifiedSince) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Update", nil, "Failure preparing request") + return + } + + resp, err := client.UpdateSender(req) + if err != nil { + result.Response = resp + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Update", resp, "Failure sending request") + return + } + + result, err = client.UpdateResponder(resp) + if err != nil { + err = autorest.NewErrorWithError(err, "batch.TaskClient", "Update", resp, "Failure responding to request") + } + + return +} + +// UpdatePreparer prepares the Update request. +func (client TaskClient) UpdatePreparer(ctx context.Context, jobID string, taskID string, taskUpdateParameter TaskUpdateParameter, timeout *int32, clientRequestID *uuid.UUID, returnClientRequestID *bool, ocpDate *date.TimeRFC1123, ifMatch string, ifNoneMatch string, ifModifiedSince *date.TimeRFC1123, ifUnmodifiedSince *date.TimeRFC1123) (*http.Request, error) { + pathParameters := map[string]interface{}{ + "jobId": autorest.Encode("path", jobID), + "taskId": autorest.Encode("path", taskID), + } + + const APIVersion = "2017-09-01.6.0" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + if timeout != nil { + queryParameters["timeout"] = autorest.Encode("query", *timeout) + } else { + queryParameters["timeout"] = autorest.Encode("query", 30) + } + + preparer := autorest.CreatePreparer( + autorest.AsContentType("application/json; odata=minimalmetadata; charset=utf-8"), + autorest.AsPut(), + autorest.WithBaseURL(client.BaseURI), + autorest.WithPathParameters("/jobs/{jobId}/tasks/{taskId}", pathParameters), + autorest.WithJSON(taskUpdateParameter), + autorest.WithQueryParameters(queryParameters)) + if clientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("client-request-id", autorest.String(clientRequestID))) + } + if returnClientRequestID != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(returnClientRequestID))) + } else { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("return-client-request-id", autorest.String(false))) + } + if ocpDate != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("ocp-date", autorest.String(ocpDate))) + } + if len(ifMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Match", autorest.String(ifMatch))) + } + if len(ifNoneMatch) > 0 { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-None-Match", autorest.String(ifNoneMatch))) + } + if ifModifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Modified-Since", autorest.String(ifModifiedSince))) + } + if ifUnmodifiedSince != nil { + preparer = autorest.DecoratePreparer(preparer, + autorest.WithHeader("If-Unmodified-Since", autorest.String(ifUnmodifiedSince))) + } + return preparer.Prepare((&http.Request{}).WithContext(ctx)) +} + +// UpdateSender sends the Update request. The method will close the +// http.Response Body if it receives an error. +func (client TaskClient) UpdateSender(req *http.Request) (*http.Response, error) { + return autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...)) +} + +// UpdateResponder handles the response to the Update request. The method always +// closes the http.Response Body. +func (client TaskClient) UpdateResponder(resp *http.Response) (result autorest.Response, err error) { + err = autorest.Respond( + resp, + client.ByInspecting(), + azure.WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByClosing()) + result.Response = resp + return +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/version.go b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/version.go new file mode 100644 index 000000000..836672bf5 --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/services/batch/2017-09-01.6.0/batch/version.go @@ -0,0 +1,30 @@ +package batch + +import "github.com/Azure/azure-sdk-for-go/version" + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +// UserAgent returns the UserAgent string to use when sending http.Requests. +func UserAgent() string { + return "Azure-SDK-For-Go/" + version.Number + " batch/2017-09-01.6.0" +} + +// Version returns the semantic version (see http://semver.org) of the client. +func Version() string { + return version.Number +} diff --git a/vendor/github.com/Azure/azure-sdk-for-go/version/version.go b/vendor/github.com/Azure/azure-sdk-for-go/version/version.go new file mode 100644 index 000000000..3f8c8a8df --- /dev/null +++ b/vendor/github.com/Azure/azure-sdk-for-go/version/version.go @@ -0,0 +1,21 @@ +package version + +// Copyright (c) Microsoft and contributors. All rights reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// +// See the License for the specific language governing permissions and +// limitations under the License. +// +// Code generated by Microsoft (R) AutoRest Code Generator. +// Changes may cause incorrect behavior and will be lost if the code is regenerated. + +// Number contains the semantic version of this SDK. +const Number = "v15.1.1" diff --git a/vendor/github.com/Azure/go-autorest/autorest/authorization.go b/vendor/github.com/Azure/go-autorest/autorest/authorization.go new file mode 100644 index 000000000..77eff45bd --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/authorization.go @@ -0,0 +1,259 @@ +package autorest + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "fmt" + "net/http" + "net/url" + "strings" + + "github.com/Azure/go-autorest/autorest/adal" +) + +const ( + bearerChallengeHeader = "Www-Authenticate" + bearer = "Bearer" + tenantID = "tenantID" + apiKeyAuthorizerHeader = "Ocp-Apim-Subscription-Key" + bingAPISdkHeader = "X-BingApis-SDK-Client" + golangBingAPISdkHeaderValue = "Go-SDK" +) + +// Authorizer is the interface that provides a PrepareDecorator used to supply request +// authorization. Most often, the Authorizer decorator runs last so it has access to the full +// state of the formed HTTP request. +type Authorizer interface { + WithAuthorization() PrepareDecorator +} + +// NullAuthorizer implements a default, "do nothing" Authorizer. +type NullAuthorizer struct{} + +// WithAuthorization returns a PrepareDecorator that does nothing. +func (na NullAuthorizer) WithAuthorization() PrepareDecorator { + return WithNothing() +} + +// APIKeyAuthorizer implements API Key authorization. +type APIKeyAuthorizer struct { + headers map[string]interface{} + queryParameters map[string]interface{} +} + +// NewAPIKeyAuthorizerWithHeaders creates an ApiKeyAuthorizer with headers. +func NewAPIKeyAuthorizerWithHeaders(headers map[string]interface{}) *APIKeyAuthorizer { + return NewAPIKeyAuthorizer(headers, nil) +} + +// NewAPIKeyAuthorizerWithQueryParameters creates an ApiKeyAuthorizer with query parameters. +func NewAPIKeyAuthorizerWithQueryParameters(queryParameters map[string]interface{}) *APIKeyAuthorizer { + return NewAPIKeyAuthorizer(nil, queryParameters) +} + +// NewAPIKeyAuthorizer creates an ApiKeyAuthorizer with headers. +func NewAPIKeyAuthorizer(headers map[string]interface{}, queryParameters map[string]interface{}) *APIKeyAuthorizer { + return &APIKeyAuthorizer{headers: headers, queryParameters: queryParameters} +} + +// WithAuthorization returns a PrepareDecorator that adds an HTTP headers and Query Paramaters +func (aka *APIKeyAuthorizer) WithAuthorization() PrepareDecorator { + return func(p Preparer) Preparer { + return DecoratePreparer(p, WithHeaders(aka.headers), WithQueryParameters(aka.queryParameters)) + } +} + +// CognitiveServicesAuthorizer implements authorization for Cognitive Services. +type CognitiveServicesAuthorizer struct { + subscriptionKey string +} + +// NewCognitiveServicesAuthorizer is +func NewCognitiveServicesAuthorizer(subscriptionKey string) *CognitiveServicesAuthorizer { + return &CognitiveServicesAuthorizer{subscriptionKey: subscriptionKey} +} + +// WithAuthorization is +func (csa *CognitiveServicesAuthorizer) WithAuthorization() PrepareDecorator { + headers := make(map[string]interface{}) + headers[apiKeyAuthorizerHeader] = csa.subscriptionKey + headers[bingAPISdkHeader] = golangBingAPISdkHeaderValue + + return NewAPIKeyAuthorizerWithHeaders(headers).WithAuthorization() +} + +// BearerAuthorizer implements the bearer authorization +type BearerAuthorizer struct { + tokenProvider adal.OAuthTokenProvider +} + +// NewBearerAuthorizer crates a BearerAuthorizer using the given token provider +func NewBearerAuthorizer(tp adal.OAuthTokenProvider) *BearerAuthorizer { + return &BearerAuthorizer{tokenProvider: tp} +} + +// WithAuthorization returns a PrepareDecorator that adds an HTTP Authorization header whose +// value is "Bearer " followed by the token. +// +// By default, the token will be automatically refreshed through the Refresher interface. +func (ba *BearerAuthorizer) WithAuthorization() PrepareDecorator { + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + // the ordering is important here, prefer RefresherWithContext if available + if refresher, ok := ba.tokenProvider.(adal.RefresherWithContext); ok { + err = refresher.EnsureFreshWithContext(r.Context()) + } else if refresher, ok := ba.tokenProvider.(adal.Refresher); ok { + err = refresher.EnsureFresh() + } + if err != nil { + var resp *http.Response + if tokError, ok := err.(adal.TokenRefreshError); ok { + resp = tokError.Response() + } + return r, NewErrorWithError(err, "azure.BearerAuthorizer", "WithAuthorization", resp, + "Failed to refresh the Token for request to %s", r.URL) + } + return Prepare(r, WithHeader(headerAuthorization, fmt.Sprintf("Bearer %s", ba.tokenProvider.OAuthToken()))) + } + return r, err + }) + } +} + +// BearerAuthorizerCallbackFunc is the authentication callback signature. +type BearerAuthorizerCallbackFunc func(tenantID, resource string) (*BearerAuthorizer, error) + +// BearerAuthorizerCallback implements bearer authorization via a callback. +type BearerAuthorizerCallback struct { + sender Sender + callback BearerAuthorizerCallbackFunc +} + +// NewBearerAuthorizerCallback creates a bearer authorization callback. The callback +// is invoked when the HTTP request is submitted. +func NewBearerAuthorizerCallback(sender Sender, callback BearerAuthorizerCallbackFunc) *BearerAuthorizerCallback { + if sender == nil { + sender = &http.Client{} + } + return &BearerAuthorizerCallback{sender: sender, callback: callback} +} + +// WithAuthorization returns a PrepareDecorator that adds an HTTP Authorization header whose value +// is "Bearer " followed by the token. The BearerAuthorizer is obtained via a user-supplied callback. +// +// By default, the token will be automatically refreshed through the Refresher interface. +func (bacb *BearerAuthorizerCallback) WithAuthorization() PrepareDecorator { + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + // make a copy of the request and remove the body as it's not + // required and avoids us having to create a copy of it. + rCopy := *r + removeRequestBody(&rCopy) + + resp, err := bacb.sender.Do(&rCopy) + if err == nil && resp.StatusCode == 401 { + defer resp.Body.Close() + if hasBearerChallenge(resp) { + bc, err := newBearerChallenge(resp) + if err != nil { + return r, err + } + if bacb.callback != nil { + ba, err := bacb.callback(bc.values[tenantID], bc.values["resource"]) + if err != nil { + return r, err + } + return Prepare(r, ba.WithAuthorization()) + } + } + } + } + return r, err + }) + } +} + +// returns true if the HTTP response contains a bearer challenge +func hasBearerChallenge(resp *http.Response) bool { + authHeader := resp.Header.Get(bearerChallengeHeader) + if len(authHeader) == 0 || strings.Index(authHeader, bearer) < 0 { + return false + } + return true +} + +type bearerChallenge struct { + values map[string]string +} + +func newBearerChallenge(resp *http.Response) (bc bearerChallenge, err error) { + challenge := strings.TrimSpace(resp.Header.Get(bearerChallengeHeader)) + trimmedChallenge := challenge[len(bearer)+1:] + + // challenge is a set of key=value pairs that are comma delimited + pairs := strings.Split(trimmedChallenge, ",") + if len(pairs) < 1 { + err = fmt.Errorf("challenge '%s' contains no pairs", challenge) + return bc, err + } + + bc.values = make(map[string]string) + for i := range pairs { + trimmedPair := strings.TrimSpace(pairs[i]) + pair := strings.Split(trimmedPair, "=") + if len(pair) == 2 { + // remove the enclosing quotes + key := strings.Trim(pair[0], "\"") + value := strings.Trim(pair[1], "\"") + + switch key { + case "authorization", "authorization_uri": + // strip the tenant ID from the authorization URL + asURL, err := url.Parse(value) + if err != nil { + return bc, err + } + bc.values[tenantID] = asURL.Path[1:] + default: + bc.values[key] = value + } + } + } + + return bc, err +} + +// EventGridKeyAuthorizer implements authorization for event grid using key authentication. +type EventGridKeyAuthorizer struct { + topicKey string +} + +// NewEventGridKeyAuthorizer creates a new EventGridKeyAuthorizer +// with the specified topic key. +func NewEventGridKeyAuthorizer(topicKey string) EventGridKeyAuthorizer { + return EventGridKeyAuthorizer{topicKey: topicKey} +} + +// WithAuthorization returns a PrepareDecorator that adds the aeg-sas-key authentication header. +func (egta EventGridKeyAuthorizer) WithAuthorization() PrepareDecorator { + headers := map[string]interface{}{ + "aeg-sas-key": egta.topicKey, + } + return NewAPIKeyAuthorizerWithHeaders(headers).WithAuthorization() +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/autorest.go b/vendor/github.com/Azure/go-autorest/autorest/autorest.go new file mode 100644 index 000000000..aafdf021f --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/autorest.go @@ -0,0 +1,150 @@ +/* +Package autorest implements an HTTP request pipeline suitable for use across multiple go-routines +and provides the shared routines relied on by AutoRest (see https://github.com/Azure/autorest/) +generated Go code. + +The package breaks sending and responding to HTTP requests into three phases: Preparing, Sending, +and Responding. A typical pattern is: + + req, err := Prepare(&http.Request{}, + token.WithAuthorization()) + + resp, err := Send(req, + WithLogging(logger), + DoErrorIfStatusCode(http.StatusInternalServerError), + DoCloseIfError(), + DoRetryForAttempts(5, time.Second)) + + err = Respond(resp, + ByDiscardingBody(), + ByClosing()) + +Each phase relies on decorators to modify and / or manage processing. Decorators may first modify +and then pass the data along, pass the data first and then modify the result, or wrap themselves +around passing the data (such as a logger might do). Decorators run in the order provided. For +example, the following: + + req, err := Prepare(&http.Request{}, + WithBaseURL("https://microsoft.com/"), + WithPath("a"), + WithPath("b"), + WithPath("c")) + +will set the URL to: + + https://microsoft.com/a/b/c + +Preparers and Responders may be shared and re-used (assuming the underlying decorators support +sharing and re-use). Performant use is obtained by creating one or more Preparers and Responders +shared among multiple go-routines, and a single Sender shared among multiple sending go-routines, +all bound together by means of input / output channels. + +Decorators hold their passed state within a closure (such as the path components in the example +above). Be careful to share Preparers and Responders only in a context where such held state +applies. For example, it may not make sense to share a Preparer that applies a query string from a +fixed set of values. Similarly, sharing a Responder that reads the response body into a passed +struct (e.g., ByUnmarshallingJson) is likely incorrect. + +Lastly, the Swagger specification (https://swagger.io) that drives AutoRest +(https://github.com/Azure/autorest/) precisely defines two date forms: date and date-time. The +github.com/Azure/go-autorest/autorest/date package provides time.Time derivations to ensure +correct parsing and formatting. + +Errors raised by autorest objects and methods will conform to the autorest.Error interface. + +See the included examples for more detail. For details on the suggested use of this package by +generated clients, see the Client described below. +*/ +package autorest + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "context" + "net/http" + "time" +) + +const ( + // HeaderLocation specifies the HTTP Location header. + HeaderLocation = "Location" + + // HeaderRetryAfter specifies the HTTP Retry-After header. + HeaderRetryAfter = "Retry-After" +) + +// ResponseHasStatusCode returns true if the status code in the HTTP Response is in the passed set +// and false otherwise. +func ResponseHasStatusCode(resp *http.Response, codes ...int) bool { + if resp == nil { + return false + } + return containsInt(codes, resp.StatusCode) +} + +// GetLocation retrieves the URL from the Location header of the passed response. +func GetLocation(resp *http.Response) string { + return resp.Header.Get(HeaderLocation) +} + +// GetRetryAfter extracts the retry delay from the Retry-After header of the passed response. If +// the header is absent or is malformed, it will return the supplied default delay time.Duration. +func GetRetryAfter(resp *http.Response, defaultDelay time.Duration) time.Duration { + retry := resp.Header.Get(HeaderRetryAfter) + if retry == "" { + return defaultDelay + } + + d, err := time.ParseDuration(retry + "s") + if err != nil { + return defaultDelay + } + + return d +} + +// NewPollingRequest allocates and returns a new http.Request to poll for the passed response. +func NewPollingRequest(resp *http.Response, cancel <-chan struct{}) (*http.Request, error) { + location := GetLocation(resp) + if location == "" { + return nil, NewErrorWithResponse("autorest", "NewPollingRequest", resp, "Location header missing from response that requires polling") + } + + req, err := Prepare(&http.Request{Cancel: cancel}, + AsGet(), + WithBaseURL(location)) + if err != nil { + return nil, NewErrorWithError(err, "autorest", "NewPollingRequest", nil, "Failure creating poll request to %s", location) + } + + return req, nil +} + +// NewPollingRequestWithContext allocates and returns a new http.Request with the specified context to poll for the passed response. +func NewPollingRequestWithContext(ctx context.Context, resp *http.Response) (*http.Request, error) { + location := GetLocation(resp) + if location == "" { + return nil, NewErrorWithResponse("autorest", "NewPollingRequestWithContext", resp, "Location header missing from response that requires polling") + } + + req, err := Prepare((&http.Request{}).WithContext(ctx), + AsGet(), + WithBaseURL(location)) + if err != nil { + return nil, NewErrorWithError(err, "autorest", "NewPollingRequestWithContext", nil, "Failure creating poll request to %s", location) + } + + return req, nil +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/azure/async.go b/vendor/github.com/Azure/go-autorest/autorest/azure/async.go new file mode 100644 index 000000000..a58e5ef3f --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/azure/async.go @@ -0,0 +1,511 @@ +package azure + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "bytes" + "context" + "encoding/json" + "fmt" + "io/ioutil" + "net/http" + "strings" + "time" + + "github.com/Azure/go-autorest/autorest" + "github.com/Azure/go-autorest/autorest/date" +) + +const ( + headerAsyncOperation = "Azure-AsyncOperation" +) + +const ( + operationInProgress string = "InProgress" + operationCanceled string = "Canceled" + operationFailed string = "Failed" + operationSucceeded string = "Succeeded" +) + +var pollingCodes = [...]int{http.StatusNoContent, http.StatusAccepted, http.StatusCreated, http.StatusOK} + +// Future provides a mechanism to access the status and results of an asynchronous request. +// Since futures are stateful they should be passed by value to avoid race conditions. +type Future struct { + req *http.Request + resp *http.Response + ps pollingState +} + +// NewFuture returns a new Future object initialized with the specified request. +func NewFuture(req *http.Request) Future { + return Future{req: req} +} + +// Response returns the last HTTP response or nil if there isn't one. +func (f Future) Response() *http.Response { + return f.resp +} + +// Status returns the last status message of the operation. +func (f Future) Status() string { + if f.ps.State == "" { + return "Unknown" + } + return f.ps.State +} + +// PollingMethod returns the method used to monitor the status of the asynchronous operation. +func (f Future) PollingMethod() PollingMethodType { + return f.ps.PollingMethod +} + +// Done queries the service to see if the operation has completed. +func (f *Future) Done(sender autorest.Sender) (bool, error) { + // exit early if this future has terminated + if f.ps.hasTerminated() { + return true, f.errorInfo() + } + resp, err := sender.Do(f.req) + f.resp = resp + if err != nil { + return false, err + } + + if !autorest.ResponseHasStatusCode(resp, pollingCodes[:]...) { + // check response body for error content + if resp.Body != nil { + type respErr struct { + ServiceError ServiceError `json:"error"` + } + re := respErr{} + + defer resp.Body.Close() + b, err := ioutil.ReadAll(resp.Body) + if err != nil { + return false, err + } + err = json.Unmarshal(b, &re) + if err != nil { + return false, err + } + return false, re.ServiceError + } + + // try to return something meaningful + return false, ServiceError{ + Code: fmt.Sprintf("%v", resp.StatusCode), + Message: resp.Status, + } + } + + err = updatePollingState(resp, &f.ps) + if err != nil { + return false, err + } + + if f.ps.hasTerminated() { + return true, f.errorInfo() + } + + f.req, err = newPollingRequest(f.ps) + return false, err +} + +// GetPollingDelay returns a duration the application should wait before checking +// the status of the asynchronous request and true; this value is returned from +// the service via the Retry-After response header. If the header wasn't returned +// then the function returns the zero-value time.Duration and false. +func (f Future) GetPollingDelay() (time.Duration, bool) { + if f.resp == nil { + return 0, false + } + + retry := f.resp.Header.Get(autorest.HeaderRetryAfter) + if retry == "" { + return 0, false + } + + d, err := time.ParseDuration(retry + "s") + if err != nil { + panic(err) + } + + return d, true +} + +// WaitForCompletion will return when one of the following conditions is met: the long +// running operation has completed, the provided context is cancelled, or the client's +// polling duration has been exceeded. It will retry failed polling attempts based on +// the retry value defined in the client up to the maximum retry attempts. +func (f Future) WaitForCompletion(ctx context.Context, client autorest.Client) error { + ctx, cancel := context.WithTimeout(ctx, client.PollingDuration) + defer cancel() + + done, err := f.Done(client) + for attempts := 0; !done; done, err = f.Done(client) { + if attempts >= client.RetryAttempts { + return autorest.NewErrorWithError(err, "azure", "WaitForCompletion", f.resp, "the number of retries has been exceeded") + } + // we want delayAttempt to be zero in the non-error case so + // that DelayForBackoff doesn't perform exponential back-off + var delayAttempt int + var delay time.Duration + if err == nil { + // check for Retry-After delay, if not present use the client's polling delay + var ok bool + delay, ok = f.GetPollingDelay() + if !ok { + delay = client.PollingDelay + } + } else { + // there was an error polling for status so perform exponential + // back-off based on the number of attempts using the client's retry + // duration. update attempts after delayAttempt to avoid off-by-one. + delayAttempt = attempts + delay = client.RetryDuration + attempts++ + } + // wait until the delay elapses or the context is cancelled + delayElapsed := autorest.DelayForBackoff(delay, delayAttempt, ctx.Done()) + if !delayElapsed { + return autorest.NewErrorWithError(ctx.Err(), "azure", "WaitForCompletion", f.resp, "context has been cancelled") + } + } + return err +} + +// if the operation failed the polling state will contain +// error information and implements the error interface +func (f *Future) errorInfo() error { + if !f.ps.hasSucceeded() { + return f.ps + } + return nil +} + +// MarshalJSON implements the json.Marshaler interface. +func (f Future) MarshalJSON() ([]byte, error) { + return json.Marshal(&f.ps) +} + +// UnmarshalJSON implements the json.Unmarshaler interface. +func (f *Future) UnmarshalJSON(data []byte) error { + err := json.Unmarshal(data, &f.ps) + if err != nil { + return err + } + f.req, err = newPollingRequest(f.ps) + return err +} + +// PollingURL returns the URL used for retrieving the status of the long-running operation. +// For LROs that use the Location header the final URL value is used to retrieve the result. +func (f Future) PollingURL() string { + return f.ps.URI +} + +// DoPollForAsynchronous returns a SendDecorator that polls if the http.Response is for an Azure +// long-running operation. It will delay between requests for the duration specified in the +// RetryAfter header or, if the header is absent, the passed delay. Polling may be canceled by +// closing the optional channel on the http.Request. +func DoPollForAsynchronous(delay time.Duration) autorest.SendDecorator { + return func(s autorest.Sender) autorest.Sender { + return autorest.SenderFunc(func(r *http.Request) (resp *http.Response, err error) { + resp, err = s.Do(r) + if err != nil { + return resp, err + } + if !autorest.ResponseHasStatusCode(resp, pollingCodes[:]...) { + return resp, nil + } + + ps := pollingState{} + for err == nil { + err = updatePollingState(resp, &ps) + if err != nil { + break + } + if ps.hasTerminated() { + if !ps.hasSucceeded() { + err = ps + } + break + } + + r, err = newPollingRequest(ps) + if err != nil { + return resp, err + } + r = r.WithContext(resp.Request.Context()) + + delay = autorest.GetRetryAfter(resp, delay) + resp, err = autorest.SendWithSender(s, r, + autorest.AfterDelay(delay)) + } + + return resp, err + }) + } +} + +func getAsyncOperation(resp *http.Response) string { + return resp.Header.Get(http.CanonicalHeaderKey(headerAsyncOperation)) +} + +func hasSucceeded(state string) bool { + return strings.EqualFold(state, operationSucceeded) +} + +func hasTerminated(state string) bool { + return strings.EqualFold(state, operationCanceled) || strings.EqualFold(state, operationFailed) || strings.EqualFold(state, operationSucceeded) +} + +func hasFailed(state string) bool { + return strings.EqualFold(state, operationFailed) +} + +type provisioningTracker interface { + state() string + hasSucceeded() bool + hasTerminated() bool +} + +type operationResource struct { + // Note: + // The specification states services should return the "id" field. However some return it as + // "operationId". + ID string `json:"id"` + OperationID string `json:"operationId"` + Name string `json:"name"` + Status string `json:"status"` + Properties map[string]interface{} `json:"properties"` + OperationError ServiceError `json:"error"` + StartTime date.Time `json:"startTime"` + EndTime date.Time `json:"endTime"` + PercentComplete float64 `json:"percentComplete"` +} + +func (or operationResource) state() string { + return or.Status +} + +func (or operationResource) hasSucceeded() bool { + return hasSucceeded(or.state()) +} + +func (or operationResource) hasTerminated() bool { + return hasTerminated(or.state()) +} + +type provisioningProperties struct { + ProvisioningState string `json:"provisioningState"` +} + +type provisioningStatus struct { + Properties provisioningProperties `json:"properties,omitempty"` + ProvisioningError ServiceError `json:"error,omitempty"` +} + +func (ps provisioningStatus) state() string { + return ps.Properties.ProvisioningState +} + +func (ps provisioningStatus) hasSucceeded() bool { + return hasSucceeded(ps.state()) +} + +func (ps provisioningStatus) hasTerminated() bool { + return hasTerminated(ps.state()) +} + +func (ps provisioningStatus) hasProvisioningError() bool { + // code and message are required fields so only check them + return len(ps.ProvisioningError.Code) > 0 || + len(ps.ProvisioningError.Message) > 0 +} + +// PollingMethodType defines a type used for enumerating polling mechanisms. +type PollingMethodType string + +const ( + // PollingAsyncOperation indicates the polling method uses the Azure-AsyncOperation header. + PollingAsyncOperation PollingMethodType = "AsyncOperation" + + // PollingLocation indicates the polling method uses the Location header. + PollingLocation PollingMethodType = "Location" + + // PollingUnknown indicates an unknown polling method and is the default value. + PollingUnknown PollingMethodType = "" +) + +type pollingState struct { + PollingMethod PollingMethodType `json:"pollingMethod"` + URI string `json:"uri"` + State string `json:"state"` + ServiceError *ServiceError `json:"error,omitempty"` +} + +func (ps pollingState) hasSucceeded() bool { + return hasSucceeded(ps.State) +} + +func (ps pollingState) hasTerminated() bool { + return hasTerminated(ps.State) +} + +func (ps pollingState) hasFailed() bool { + return hasFailed(ps.State) +} + +func (ps pollingState) Error() string { + s := fmt.Sprintf("Long running operation terminated with status '%s'", ps.State) + if ps.ServiceError != nil { + s = fmt.Sprintf("%s: %+v", s, *ps.ServiceError) + } + return s +} + +// updatePollingState maps the operation status -- retrieved from either a provisioningState +// field, the status field of an OperationResource, or inferred from the HTTP status code -- +// into a well-known states. Since the process begins from the initial request, the state +// always comes from either a the provisioningState returned or is inferred from the HTTP +// status code. Subsequent requests will read an Azure OperationResource object if the +// service initially returned the Azure-AsyncOperation header. The responseFormat field notes +// the expected response format. +func updatePollingState(resp *http.Response, ps *pollingState) error { + // Determine the response shape + // -- The first response will always be a provisioningStatus response; only the polling requests, + // depending on the header returned, may be something otherwise. + var pt provisioningTracker + if ps.PollingMethod == PollingAsyncOperation { + pt = &operationResource{} + } else { + pt = &provisioningStatus{} + } + + // If this is the first request (that is, the polling response shape is unknown), determine how + // to poll and what to expect + if ps.PollingMethod == PollingUnknown { + req := resp.Request + if req == nil { + return autorest.NewError("azure", "updatePollingState", "Azure Polling Error - Original HTTP request is missing") + } + + // Prefer the Azure-AsyncOperation header + ps.URI = getAsyncOperation(resp) + if ps.URI != "" { + ps.PollingMethod = PollingAsyncOperation + } else { + ps.PollingMethod = PollingLocation + } + + // Else, use the Location header + if ps.URI == "" { + ps.URI = autorest.GetLocation(resp) + } + + // Lastly, requests against an existing resource, use the last request URI + if ps.URI == "" { + m := strings.ToUpper(req.Method) + if m == http.MethodPatch || m == http.MethodPut || m == http.MethodGet { + ps.URI = req.URL.String() + } + } + } + + // Read and interpret the response (saving the Body in case no polling is necessary) + b := &bytes.Buffer{} + err := autorest.Respond(resp, + autorest.ByCopying(b), + autorest.ByUnmarshallingJSON(pt), + autorest.ByClosing()) + resp.Body = ioutil.NopCloser(b) + if err != nil { + return err + } + + // Interpret the results + // -- Terminal states apply regardless + // -- Unknown states are per-service inprogress states + // -- Otherwise, infer state from HTTP status code + if pt.hasTerminated() { + ps.State = pt.state() + } else if pt.state() != "" { + ps.State = operationInProgress + } else { + switch resp.StatusCode { + case http.StatusAccepted: + ps.State = operationInProgress + + case http.StatusNoContent, http.StatusCreated, http.StatusOK: + ps.State = operationSucceeded + + default: + ps.State = operationFailed + } + } + + if strings.EqualFold(ps.State, operationInProgress) && ps.URI == "" { + return autorest.NewError("azure", "updatePollingState", "Azure Polling Error - Unable to obtain polling URI for %s %s", resp.Request.Method, resp.Request.URL) + } + + // For failed operation, check for error code and message in + // -- Operation resource + // -- Response + // -- Otherwise, Unknown + if ps.hasFailed() { + if or, ok := pt.(*operationResource); ok { + ps.ServiceError = &or.OperationError + } else if p, ok := pt.(*provisioningStatus); ok && p.hasProvisioningError() { + ps.ServiceError = &p.ProvisioningError + } else { + ps.ServiceError = &ServiceError{ + Code: "Unknown", + Message: "None", + } + } + } + return nil +} + +func newPollingRequest(ps pollingState) (*http.Request, error) { + reqPoll, err := autorest.Prepare(&http.Request{}, + autorest.AsGet(), + autorest.WithBaseURL(ps.URI)) + if err != nil { + return nil, autorest.NewErrorWithError(err, "azure", "newPollingRequest", nil, "Failure creating poll request to %s", ps.URI) + } + + return reqPoll, nil +} + +// AsyncOpIncompleteError is the type that's returned from a future that has not completed. +type AsyncOpIncompleteError struct { + // FutureType is the name of the type composed of a azure.Future. + FutureType string +} + +// Error returns an error message including the originating type name of the error. +func (e AsyncOpIncompleteError) Error() string { + return fmt.Sprintf("%s: asynchronous operation has not completed", e.FutureType) +} + +// NewAsyncOpIncompleteError creates a new AsyncOpIncompleteError with the specified parameters. +func NewAsyncOpIncompleteError(futureType string) AsyncOpIncompleteError { + return AsyncOpIncompleteError{ + FutureType: futureType, + } +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/azure/azure.go b/vendor/github.com/Azure/go-autorest/autorest/azure/azure.go new file mode 100644 index 000000000..18d029526 --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/azure/azure.go @@ -0,0 +1,301 @@ +// Package azure provides Azure-specific implementations used with AutoRest. +// See the included examples for more detail. +package azure + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "encoding/json" + "fmt" + "io/ioutil" + "net/http" + "regexp" + "strconv" + "strings" + + "github.com/Azure/go-autorest/autorest" +) + +const ( + // HeaderClientID is the Azure extension header to set a user-specified request ID. + HeaderClientID = "x-ms-client-request-id" + + // HeaderReturnClientID is the Azure extension header to set if the user-specified request ID + // should be included in the response. + HeaderReturnClientID = "x-ms-return-client-request-id" + + // HeaderRequestID is the Azure extension header of the service generated request ID returned + // in the response. + HeaderRequestID = "x-ms-request-id" +) + +// ServiceError encapsulates the error response from an Azure service. +// It adhears to the OData v4 specification for error responses. +type ServiceError struct { + Code string `json:"code"` + Message string `json:"message"` + Target *string `json:"target"` + Details []map[string]interface{} `json:"details"` + InnerError map[string]interface{} `json:"innererror"` +} + +func (se ServiceError) Error() string { + result := fmt.Sprintf("Code=%q Message=%q", se.Code, se.Message) + + if se.Target != nil { + result += fmt.Sprintf(" Target=%q", *se.Target) + } + + if se.Details != nil { + d, err := json.Marshal(se.Details) + if err != nil { + result += fmt.Sprintf(" Details=%v", se.Details) + } + result += fmt.Sprintf(" Details=%v", string(d)) + } + + if se.InnerError != nil { + d, err := json.Marshal(se.InnerError) + if err != nil { + result += fmt.Sprintf(" InnerError=%v", se.InnerError) + } + result += fmt.Sprintf(" InnerError=%v", string(d)) + } + + return result +} + +// UnmarshalJSON implements the json.Unmarshaler interface for the ServiceError type. +func (se *ServiceError) UnmarshalJSON(b []byte) error { + // per the OData v4 spec the details field must be an array of JSON objects. + // unfortunately not all services adhear to the spec and just return a single + // object instead of an array with one object. so we have to perform some + // shenanigans to accommodate both cases. + // http://docs.oasis-open.org/odata/odata-json-format/v4.0/os/odata-json-format-v4.0-os.html#_Toc372793091 + + type serviceError1 struct { + Code string `json:"code"` + Message string `json:"message"` + Target *string `json:"target"` + Details []map[string]interface{} `json:"details"` + InnerError map[string]interface{} `json:"innererror"` + } + + type serviceError2 struct { + Code string `json:"code"` + Message string `json:"message"` + Target *string `json:"target"` + Details map[string]interface{} `json:"details"` + InnerError map[string]interface{} `json:"innererror"` + } + + se1 := serviceError1{} + err := json.Unmarshal(b, &se1) + if err == nil { + se.populate(se1.Code, se1.Message, se1.Target, se1.Details, se1.InnerError) + return nil + } + + se2 := serviceError2{} + err = json.Unmarshal(b, &se2) + if err == nil { + se.populate(se2.Code, se2.Message, se2.Target, nil, se2.InnerError) + se.Details = append(se.Details, se2.Details) + return nil + } + return err +} + +func (se *ServiceError) populate(code, message string, target *string, details []map[string]interface{}, inner map[string]interface{}) { + se.Code = code + se.Message = message + se.Target = target + se.Details = details + se.InnerError = inner +} + +// RequestError describes an error response returned by Azure service. +type RequestError struct { + autorest.DetailedError + + // The error returned by the Azure service. + ServiceError *ServiceError `json:"error"` + + // The request id (from the x-ms-request-id-header) of the request. + RequestID string +} + +// Error returns a human-friendly error message from service error. +func (e RequestError) Error() string { + return fmt.Sprintf("autorest/azure: Service returned an error. Status=%v %v", + e.StatusCode, e.ServiceError) +} + +// IsAzureError returns true if the passed error is an Azure Service error; false otherwise. +func IsAzureError(e error) bool { + _, ok := e.(*RequestError) + return ok +} + +// Resource contains details about an Azure resource. +type Resource struct { + SubscriptionID string + ResourceGroup string + Provider string + ResourceType string + ResourceName string +} + +// ParseResourceID parses a resource ID into a ResourceDetails struct. +// See https://docs.microsoft.com/en-us/azure/azure-resource-manager/resource-group-template-functions-resource#return-value-4. +func ParseResourceID(resourceID string) (Resource, error) { + + const resourceIDPatternText = `(?i)subscriptions/(.+)/resourceGroups/(.+)/providers/(.+?)/(.+?)/(.+)` + resourceIDPattern := regexp.MustCompile(resourceIDPatternText) + match := resourceIDPattern.FindStringSubmatch(resourceID) + + if len(match) == 0 { + return Resource{}, fmt.Errorf("parsing failed for %s. Invalid resource Id format", resourceID) + } + + v := strings.Split(match[5], "/") + resourceName := v[len(v)-1] + + result := Resource{ + SubscriptionID: match[1], + ResourceGroup: match[2], + Provider: match[3], + ResourceType: match[4], + ResourceName: resourceName, + } + + return result, nil +} + +// NewErrorWithError creates a new Error conforming object from the +// passed packageType, method, statusCode of the given resp (UndefinedStatusCode +// if resp is nil), message, and original error. message is treated as a format +// string to which the optional args apply. +func NewErrorWithError(original error, packageType string, method string, resp *http.Response, message string, args ...interface{}) RequestError { + if v, ok := original.(*RequestError); ok { + return *v + } + + statusCode := autorest.UndefinedStatusCode + if resp != nil { + statusCode = resp.StatusCode + } + return RequestError{ + DetailedError: autorest.DetailedError{ + Original: original, + PackageType: packageType, + Method: method, + StatusCode: statusCode, + Message: fmt.Sprintf(message, args...), + }, + } +} + +// WithReturningClientID returns a PrepareDecorator that adds an HTTP extension header of +// x-ms-client-request-id whose value is the passed, undecorated UUID (e.g., +// "0F39878C-5F76-4DB8-A25D-61D2C193C3CA"). It also sets the x-ms-return-client-request-id +// header to true such that UUID accompanies the http.Response. +func WithReturningClientID(uuid string) autorest.PrepareDecorator { + preparer := autorest.CreatePreparer( + WithClientID(uuid), + WithReturnClientID(true)) + + return func(p autorest.Preparer) autorest.Preparer { + return autorest.PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err != nil { + return r, err + } + return preparer.Prepare(r) + }) + } +} + +// WithClientID returns a PrepareDecorator that adds an HTTP extension header of +// x-ms-client-request-id whose value is passed, undecorated UUID (e.g., +// "0F39878C-5F76-4DB8-A25D-61D2C193C3CA"). +func WithClientID(uuid string) autorest.PrepareDecorator { + return autorest.WithHeader(HeaderClientID, uuid) +} + +// WithReturnClientID returns a PrepareDecorator that adds an HTTP extension header of +// x-ms-return-client-request-id whose boolean value indicates if the value of the +// x-ms-client-request-id header should be included in the http.Response. +func WithReturnClientID(b bool) autorest.PrepareDecorator { + return autorest.WithHeader(HeaderReturnClientID, strconv.FormatBool(b)) +} + +// ExtractClientID extracts the client identifier from the x-ms-client-request-id header set on the +// http.Request sent to the service (and returned in the http.Response) +func ExtractClientID(resp *http.Response) string { + return autorest.ExtractHeaderValue(HeaderClientID, resp) +} + +// ExtractRequestID extracts the Azure server generated request identifier from the +// x-ms-request-id header. +func ExtractRequestID(resp *http.Response) string { + return autorest.ExtractHeaderValue(HeaderRequestID, resp) +} + +// WithErrorUnlessStatusCode returns a RespondDecorator that emits an +// azure.RequestError by reading the response body unless the response HTTP status code +// is among the set passed. +// +// If there is a chance service may return responses other than the Azure error +// format and the response cannot be parsed into an error, a decoding error will +// be returned containing the response body. In any case, the Responder will +// return an error if the status code is not satisfied. +// +// If this Responder returns an error, the response body will be replaced with +// an in-memory reader, which needs no further closing. +func WithErrorUnlessStatusCode(codes ...int) autorest.RespondDecorator { + return func(r autorest.Responder) autorest.Responder { + return autorest.ResponderFunc(func(resp *http.Response) error { + err := r.Respond(resp) + if err == nil && !autorest.ResponseHasStatusCode(resp, codes...) { + var e RequestError + defer resp.Body.Close() + + // Copy and replace the Body in case it does not contain an error object. + // This will leave the Body available to the caller. + b, decodeErr := autorest.CopyAndDecode(autorest.EncodedAsJSON, resp.Body, &e) + resp.Body = ioutil.NopCloser(&b) + if decodeErr != nil { + return fmt.Errorf("autorest/azure: error response cannot be parsed: %q error: %v", b.String(), decodeErr) + } else if e.ServiceError == nil { + // Check if error is unwrapped ServiceError + if err := json.Unmarshal(b.Bytes(), &e.ServiceError); err != nil || e.ServiceError.Message == "" { + e.ServiceError = &ServiceError{ + Code: "Unknown", + Message: "Unknown service error", + } + } + } + + e.RequestID = ExtractRequestID(resp) + if e.StatusCode == nil { + e.StatusCode = resp.StatusCode + } + err = &e + } + return err + }) + } +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/azure/environments.go b/vendor/github.com/Azure/go-autorest/autorest/azure/environments.go new file mode 100644 index 000000000..7e41f7fd9 --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/azure/environments.go @@ -0,0 +1,191 @@ +package azure + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "encoding/json" + "fmt" + "io/ioutil" + "os" + "strings" +) + +// EnvironmentFilepathName captures the name of the environment variable containing the path to the file +// to be used while populating the Azure Environment. +const EnvironmentFilepathName = "AZURE_ENVIRONMENT_FILEPATH" + +var environments = map[string]Environment{ + "AZURECHINACLOUD": ChinaCloud, + "AZUREGERMANCLOUD": GermanCloud, + "AZUREPUBLICCLOUD": PublicCloud, + "AZUREUSGOVERNMENTCLOUD": USGovernmentCloud, +} + +// Environment represents a set of endpoints for each of Azure's Clouds. +type Environment struct { + Name string `json:"name"` + ManagementPortalURL string `json:"managementPortalURL"` + PublishSettingsURL string `json:"publishSettingsURL"` + ServiceManagementEndpoint string `json:"serviceManagementEndpoint"` + ResourceManagerEndpoint string `json:"resourceManagerEndpoint"` + ActiveDirectoryEndpoint string `json:"activeDirectoryEndpoint"` + GalleryEndpoint string `json:"galleryEndpoint"` + KeyVaultEndpoint string `json:"keyVaultEndpoint"` + GraphEndpoint string `json:"graphEndpoint"` + ServiceBusEndpoint string `json:"serviceBusEndpoint"` + BatchManagementEndpoint string `json:"batchManagementEndpoint"` + StorageEndpointSuffix string `json:"storageEndpointSuffix"` + SQLDatabaseDNSSuffix string `json:"sqlDatabaseDNSSuffix"` + TrafficManagerDNSSuffix string `json:"trafficManagerDNSSuffix"` + KeyVaultDNSSuffix string `json:"keyVaultDNSSuffix"` + ServiceBusEndpointSuffix string `json:"serviceBusEndpointSuffix"` + ServiceManagementVMDNSSuffix string `json:"serviceManagementVMDNSSuffix"` + ResourceManagerVMDNSSuffix string `json:"resourceManagerVMDNSSuffix"` + ContainerRegistryDNSSuffix string `json:"containerRegistryDNSSuffix"` + TokenAudience string `json:"tokenAudience"` +} + +var ( + // PublicCloud is the default public Azure cloud environment + PublicCloud = Environment{ + Name: "AzurePublicCloud", + ManagementPortalURL: "https://manage.windowsazure.com/", + PublishSettingsURL: "https://manage.windowsazure.com/publishsettings/index", + ServiceManagementEndpoint: "https://management.core.windows.net/", + ResourceManagerEndpoint: "https://management.azure.com/", + ActiveDirectoryEndpoint: "https://login.microsoftonline.com/", + GalleryEndpoint: "https://gallery.azure.com/", + KeyVaultEndpoint: "https://vault.azure.net/", + GraphEndpoint: "https://graph.windows.net/", + ServiceBusEndpoint: "https://servicebus.windows.net/", + BatchManagementEndpoint: "https://batch.core.windows.net/", + StorageEndpointSuffix: "core.windows.net", + SQLDatabaseDNSSuffix: "database.windows.net", + TrafficManagerDNSSuffix: "trafficmanager.net", + KeyVaultDNSSuffix: "vault.azure.net", + ServiceBusEndpointSuffix: "servicebus.windows.net", + ServiceManagementVMDNSSuffix: "cloudapp.net", + ResourceManagerVMDNSSuffix: "cloudapp.azure.com", + ContainerRegistryDNSSuffix: "azurecr.io", + TokenAudience: "https://management.azure.com/", + } + + // USGovernmentCloud is the cloud environment for the US Government + USGovernmentCloud = Environment{ + Name: "AzureUSGovernmentCloud", + ManagementPortalURL: "https://manage.windowsazure.us/", + PublishSettingsURL: "https://manage.windowsazure.us/publishsettings/index", + ServiceManagementEndpoint: "https://management.core.usgovcloudapi.net/", + ResourceManagerEndpoint: "https://management.usgovcloudapi.net/", + ActiveDirectoryEndpoint: "https://login.microsoftonline.us/", + GalleryEndpoint: "https://gallery.usgovcloudapi.net/", + KeyVaultEndpoint: "https://vault.usgovcloudapi.net/", + GraphEndpoint: "https://graph.windows.net/", + ServiceBusEndpoint: "https://servicebus.usgovcloudapi.net/", + BatchManagementEndpoint: "https://batch.core.usgovcloudapi.net/", + StorageEndpointSuffix: "core.usgovcloudapi.net", + SQLDatabaseDNSSuffix: "database.usgovcloudapi.net", + TrafficManagerDNSSuffix: "usgovtrafficmanager.net", + KeyVaultDNSSuffix: "vault.usgovcloudapi.net", + ServiceBusEndpointSuffix: "servicebus.usgovcloudapi.net", + ServiceManagementVMDNSSuffix: "usgovcloudapp.net", + ResourceManagerVMDNSSuffix: "cloudapp.windowsazure.us", + ContainerRegistryDNSSuffix: "azurecr.io", + TokenAudience: "https://management.usgovcloudapi.net/", + } + + // ChinaCloud is the cloud environment operated in China + ChinaCloud = Environment{ + Name: "AzureChinaCloud", + ManagementPortalURL: "https://manage.chinacloudapi.com/", + PublishSettingsURL: "https://manage.chinacloudapi.com/publishsettings/index", + ServiceManagementEndpoint: "https://management.core.chinacloudapi.cn/", + ResourceManagerEndpoint: "https://management.chinacloudapi.cn/", + ActiveDirectoryEndpoint: "https://login.chinacloudapi.cn/", + GalleryEndpoint: "https://gallery.chinacloudapi.cn/", + KeyVaultEndpoint: "https://vault.azure.cn/", + GraphEndpoint: "https://graph.chinacloudapi.cn/", + ServiceBusEndpoint: "https://servicebus.chinacloudapi.cn/", + BatchManagementEndpoint: "https://batch.chinacloudapi.cn/", + StorageEndpointSuffix: "core.chinacloudapi.cn", + SQLDatabaseDNSSuffix: "database.chinacloudapi.cn", + TrafficManagerDNSSuffix: "trafficmanager.cn", + KeyVaultDNSSuffix: "vault.azure.cn", + ServiceBusEndpointSuffix: "servicebus.chinacloudapi.cn", + ServiceManagementVMDNSSuffix: "chinacloudapp.cn", + ResourceManagerVMDNSSuffix: "cloudapp.azure.cn", + ContainerRegistryDNSSuffix: "azurecr.io", + TokenAudience: "https://management.chinacloudapi.cn/", + } + + // GermanCloud is the cloud environment operated in Germany + GermanCloud = Environment{ + Name: "AzureGermanCloud", + ManagementPortalURL: "http://portal.microsoftazure.de/", + PublishSettingsURL: "https://manage.microsoftazure.de/publishsettings/index", + ServiceManagementEndpoint: "https://management.core.cloudapi.de/", + ResourceManagerEndpoint: "https://management.microsoftazure.de/", + ActiveDirectoryEndpoint: "https://login.microsoftonline.de/", + GalleryEndpoint: "https://gallery.cloudapi.de/", + KeyVaultEndpoint: "https://vault.microsoftazure.de/", + GraphEndpoint: "https://graph.cloudapi.de/", + ServiceBusEndpoint: "https://servicebus.cloudapi.de/", + BatchManagementEndpoint: "https://batch.cloudapi.de/", + StorageEndpointSuffix: "core.cloudapi.de", + SQLDatabaseDNSSuffix: "database.cloudapi.de", + TrafficManagerDNSSuffix: "azuretrafficmanager.de", + KeyVaultDNSSuffix: "vault.microsoftazure.de", + ServiceBusEndpointSuffix: "servicebus.cloudapi.de", + ServiceManagementVMDNSSuffix: "azurecloudapp.de", + ResourceManagerVMDNSSuffix: "cloudapp.microsoftazure.de", + ContainerRegistryDNSSuffix: "azurecr.io", + TokenAudience: "https://management.microsoftazure.de/", + } +) + +// EnvironmentFromName returns an Environment based on the common name specified. +func EnvironmentFromName(name string) (Environment, error) { + // IMPORTANT + // As per @radhikagupta5: + // This is technical debt, fundamentally here because Kubernetes is not currently accepting + // contributions to the providers. Once that is an option, the provider should be updated to + // directly call `EnvironmentFromFile`. Until then, we rely on dispatching Azure Stack environment creation + // from this method based on the name that is provided to us. + if strings.EqualFold(name, "AZURESTACKCLOUD") { + return EnvironmentFromFile(os.Getenv(EnvironmentFilepathName)) + } + + name = strings.ToUpper(name) + env, ok := environments[name] + if !ok { + return env, fmt.Errorf("autorest/azure: There is no cloud environment matching the name %q", name) + } + + return env, nil +} + +// EnvironmentFromFile loads an Environment from a configuration file available on disk. +// This function is particularly useful in the Hybrid Cloud model, where one must define their own +// endpoints. +func EnvironmentFromFile(location string) (unmarshaled Environment, err error) { + fileContents, err := ioutil.ReadFile(location) + if err != nil { + return + } + + err = json.Unmarshal(fileContents, &unmarshaled) + + return +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/azure/metadata_environment.go b/vendor/github.com/Azure/go-autorest/autorest/azure/metadata_environment.go new file mode 100644 index 000000000..507f9e95c --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/azure/metadata_environment.go @@ -0,0 +1,245 @@ +package azure + +import ( + "encoding/json" + "fmt" + "io/ioutil" + "net/http" + "strings" + + "github.com/Azure/go-autorest/autorest" +) + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +type audience []string + +type authentication struct { + LoginEndpoint string `json:"loginEndpoint"` + Audiences audience `json:"audiences"` +} + +type environmentMetadataInfo struct { + GalleryEndpoint string `json:"galleryEndpoint"` + GraphEndpoint string `json:"graphEndpoint"` + PortalEndpoint string `json:"portalEndpoint"` + Authentication authentication `json:"authentication"` +} + +// EnvironmentProperty represent property names that clients can override +type EnvironmentProperty string + +const ( + // EnvironmentName ... + EnvironmentName EnvironmentProperty = "name" + // EnvironmentManagementPortalURL .. + EnvironmentManagementPortalURL EnvironmentProperty = "managementPortalURL" + // EnvironmentPublishSettingsURL ... + EnvironmentPublishSettingsURL EnvironmentProperty = "publishSettingsURL" + // EnvironmentServiceManagementEndpoint ... + EnvironmentServiceManagementEndpoint EnvironmentProperty = "serviceManagementEndpoint" + // EnvironmentResourceManagerEndpoint ... + EnvironmentResourceManagerEndpoint EnvironmentProperty = "resourceManagerEndpoint" + // EnvironmentActiveDirectoryEndpoint ... + EnvironmentActiveDirectoryEndpoint EnvironmentProperty = "activeDirectoryEndpoint" + // EnvironmentGalleryEndpoint ... + EnvironmentGalleryEndpoint EnvironmentProperty = "galleryEndpoint" + // EnvironmentKeyVaultEndpoint ... + EnvironmentKeyVaultEndpoint EnvironmentProperty = "keyVaultEndpoint" + // EnvironmentGraphEndpoint ... + EnvironmentGraphEndpoint EnvironmentProperty = "graphEndpoint" + // EnvironmentServiceBusEndpoint ... + EnvironmentServiceBusEndpoint EnvironmentProperty = "serviceBusEndpoint" + // EnvironmentBatchManagementEndpoint ... + EnvironmentBatchManagementEndpoint EnvironmentProperty = "batchManagementEndpoint" + // EnvironmentStorageEndpointSuffix ... + EnvironmentStorageEndpointSuffix EnvironmentProperty = "storageEndpointSuffix" + // EnvironmentSQLDatabaseDNSSuffix ... + EnvironmentSQLDatabaseDNSSuffix EnvironmentProperty = "sqlDatabaseDNSSuffix" + // EnvironmentTrafficManagerDNSSuffix ... + EnvironmentTrafficManagerDNSSuffix EnvironmentProperty = "trafficManagerDNSSuffix" + // EnvironmentKeyVaultDNSSuffix ... + EnvironmentKeyVaultDNSSuffix EnvironmentProperty = "keyVaultDNSSuffix" + // EnvironmentServiceBusEndpointSuffix ... + EnvironmentServiceBusEndpointSuffix EnvironmentProperty = "serviceBusEndpointSuffix" + // EnvironmentServiceManagementVMDNSSuffix ... + EnvironmentServiceManagementVMDNSSuffix EnvironmentProperty = "serviceManagementVMDNSSuffix" + // EnvironmentResourceManagerVMDNSSuffix ... + EnvironmentResourceManagerVMDNSSuffix EnvironmentProperty = "resourceManagerVMDNSSuffix" + // EnvironmentContainerRegistryDNSSuffix ... + EnvironmentContainerRegistryDNSSuffix EnvironmentProperty = "containerRegistryDNSSuffix" + // EnvironmentTokenAudience ... + EnvironmentTokenAudience EnvironmentProperty = "tokenAudience" +) + +// OverrideProperty represents property name and value that clients can override +type OverrideProperty struct { + Key EnvironmentProperty + Value string +} + +// EnvironmentFromURL loads an Environment from a URL +// This function is particularly useful in the Hybrid Cloud model, where one may define their own +// endpoints. +func EnvironmentFromURL(resourceManagerEndpoint string, properties ...OverrideProperty) (environment Environment, err error) { + var metadataEnvProperties environmentMetadataInfo + + if resourceManagerEndpoint == "" { + return environment, fmt.Errorf("Metadata resource manager endpoint is empty") + } + + if metadataEnvProperties, err = retrieveMetadataEnvironment(resourceManagerEndpoint); err != nil { + return environment, err + } + + // Give priority to user's override values + overrideProperties(&environment, properties) + + if environment.Name == "" { + environment.Name = "HybridEnvironment" + } + stampDNSSuffix := environment.StorageEndpointSuffix + if stampDNSSuffix == "" { + stampDNSSuffix = strings.TrimSuffix(strings.TrimPrefix(strings.Replace(resourceManagerEndpoint, strings.Split(resourceManagerEndpoint, ".")[0], "", 1), "."), "/") + environment.StorageEndpointSuffix = stampDNSSuffix + } + if environment.KeyVaultDNSSuffix == "" { + environment.KeyVaultDNSSuffix = fmt.Sprintf("%s.%s", "vault", stampDNSSuffix) + } + if environment.KeyVaultEndpoint == "" { + environment.KeyVaultEndpoint = fmt.Sprintf("%s%s", "https://", environment.KeyVaultDNSSuffix) + } + if environment.TokenAudience == "" { + environment.TokenAudience = metadataEnvProperties.Authentication.Audiences[0] + } + if environment.ActiveDirectoryEndpoint == "" { + environment.ActiveDirectoryEndpoint = metadataEnvProperties.Authentication.LoginEndpoint + } + if environment.ResourceManagerEndpoint == "" { + environment.ResourceManagerEndpoint = resourceManagerEndpoint + } + if environment.GalleryEndpoint == "" { + environment.GalleryEndpoint = metadataEnvProperties.GalleryEndpoint + } + if environment.GraphEndpoint == "" { + environment.GraphEndpoint = metadataEnvProperties.GraphEndpoint + } + + return environment, nil +} + +func overrideProperties(environment *Environment, properties []OverrideProperty) { + for _, property := range properties { + switch property.Key { + case EnvironmentName: + { + environment.Name = property.Value + } + case EnvironmentManagementPortalURL: + { + environment.ManagementPortalURL = property.Value + } + case EnvironmentPublishSettingsURL: + { + environment.PublishSettingsURL = property.Value + } + case EnvironmentServiceManagementEndpoint: + { + environment.ServiceManagementEndpoint = property.Value + } + case EnvironmentResourceManagerEndpoint: + { + environment.ResourceManagerEndpoint = property.Value + } + case EnvironmentActiveDirectoryEndpoint: + { + environment.ActiveDirectoryEndpoint = property.Value + } + case EnvironmentGalleryEndpoint: + { + environment.GalleryEndpoint = property.Value + } + case EnvironmentKeyVaultEndpoint: + { + environment.KeyVaultEndpoint = property.Value + } + case EnvironmentGraphEndpoint: + { + environment.GraphEndpoint = property.Value + } + case EnvironmentServiceBusEndpoint: + { + environment.ServiceBusEndpoint = property.Value + } + case EnvironmentBatchManagementEndpoint: + { + environment.BatchManagementEndpoint = property.Value + } + case EnvironmentStorageEndpointSuffix: + { + environment.StorageEndpointSuffix = property.Value + } + case EnvironmentSQLDatabaseDNSSuffix: + { + environment.SQLDatabaseDNSSuffix = property.Value + } + case EnvironmentTrafficManagerDNSSuffix: + { + environment.TrafficManagerDNSSuffix = property.Value + } + case EnvironmentKeyVaultDNSSuffix: + { + environment.KeyVaultDNSSuffix = property.Value + } + case EnvironmentServiceBusEndpointSuffix: + { + environment.ServiceBusEndpointSuffix = property.Value + } + case EnvironmentServiceManagementVMDNSSuffix: + { + environment.ServiceManagementVMDNSSuffix = property.Value + } + case EnvironmentResourceManagerVMDNSSuffix: + { + environment.ResourceManagerVMDNSSuffix = property.Value + } + case EnvironmentContainerRegistryDNSSuffix: + { + environment.ContainerRegistryDNSSuffix = property.Value + } + case EnvironmentTokenAudience: + { + environment.TokenAudience = property.Value + } + } + } +} + +func retrieveMetadataEnvironment(endpoint string) (environment environmentMetadataInfo, err error) { + client := autorest.NewClientWithUserAgent("") + managementEndpoint := fmt.Sprintf("%s%s", strings.TrimSuffix(endpoint, "/"), "/metadata/endpoints?api-version=1.0") + req, _ := http.NewRequest("GET", managementEndpoint, nil) + response, err := client.Do(req) + if err != nil { + return environment, err + } + defer response.Body.Close() + jsonResponse, err := ioutil.ReadAll(response.Body) + if err != nil { + return environment, err + } + err = json.Unmarshal(jsonResponse, &environment) + return environment, err +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/azure/rp.go b/vendor/github.com/Azure/go-autorest/autorest/azure/rp.go new file mode 100644 index 000000000..65ad0afc8 --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/azure/rp.go @@ -0,0 +1,200 @@ +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package azure + +import ( + "errors" + "fmt" + "net/http" + "net/url" + "strings" + "time" + + "github.com/Azure/go-autorest/autorest" +) + +// DoRetryWithRegistration tries to register the resource provider in case it is unregistered. +// It also handles request retries +func DoRetryWithRegistration(client autorest.Client) autorest.SendDecorator { + return func(s autorest.Sender) autorest.Sender { + return autorest.SenderFunc(func(r *http.Request) (resp *http.Response, err error) { + rr := autorest.NewRetriableRequest(r) + for currentAttempt := 0; currentAttempt < client.RetryAttempts; currentAttempt++ { + err = rr.Prepare() + if err != nil { + return resp, err + } + + resp, err = autorest.SendWithSender(s, rr.Request(), + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...), + ) + if err != nil { + return resp, err + } + + if resp.StatusCode != http.StatusConflict || client.SkipResourceProviderRegistration { + return resp, err + } + var re RequestError + err = autorest.Respond( + resp, + autorest.ByUnmarshallingJSON(&re), + ) + if err != nil { + return resp, err + } + err = re + + if re.ServiceError != nil && re.ServiceError.Code == "MissingSubscriptionRegistration" { + regErr := register(client, r, re) + if regErr != nil { + return resp, fmt.Errorf("failed auto registering Resource Provider: %s. Original error: %s", regErr, err) + } + } + } + return resp, fmt.Errorf("failed request: %s", err) + }) + } +} + +func getProvider(re RequestError) (string, error) { + if re.ServiceError != nil && len(re.ServiceError.Details) > 0 { + return re.ServiceError.Details[0]["target"].(string), nil + } + return "", errors.New("provider was not found in the response") +} + +func register(client autorest.Client, originalReq *http.Request, re RequestError) error { + subID := getSubscription(originalReq.URL.Path) + if subID == "" { + return errors.New("missing parameter subscriptionID to register resource provider") + } + providerName, err := getProvider(re) + if err != nil { + return fmt.Errorf("missing parameter provider to register resource provider: %s", err) + } + newURL := url.URL{ + Scheme: originalReq.URL.Scheme, + Host: originalReq.URL.Host, + } + + // taken from the resources SDK + // with almost identical code, this sections are easier to mantain + // It is also not a good idea to import the SDK here + // https://github.com/Azure/azure-sdk-for-go/blob/9f366792afa3e0ddaecdc860e793ba9d75e76c27/arm/resources/resources/providers.go#L252 + pathParameters := map[string]interface{}{ + "resourceProviderNamespace": autorest.Encode("path", providerName), + "subscriptionId": autorest.Encode("path", subID), + } + + const APIVersion = "2016-09-01" + queryParameters := map[string]interface{}{ + "api-version": APIVersion, + } + + preparer := autorest.CreatePreparer( + autorest.AsPost(), + autorest.WithBaseURL(newURL.String()), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/providers/{resourceProviderNamespace}/register", pathParameters), + autorest.WithQueryParameters(queryParameters), + ) + + req, err := preparer.Prepare(&http.Request{}) + if err != nil { + return err + } + req = req.WithContext(originalReq.Context()) + + resp, err := autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...), + ) + if err != nil { + return err + } + + type Provider struct { + RegistrationState *string `json:"registrationState,omitempty"` + } + var provider Provider + + err = autorest.Respond( + resp, + WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&provider), + autorest.ByClosing(), + ) + if err != nil { + return err + } + + // poll for registered provisioning state + now := time.Now() + for err == nil && time.Since(now) < client.PollingDuration { + // taken from the resources SDK + // https://github.com/Azure/azure-sdk-for-go/blob/9f366792afa3e0ddaecdc860e793ba9d75e76c27/arm/resources/resources/providers.go#L45 + preparer := autorest.CreatePreparer( + autorest.AsGet(), + autorest.WithBaseURL(newURL.String()), + autorest.WithPathParameters("/subscriptions/{subscriptionId}/providers/{resourceProviderNamespace}", pathParameters), + autorest.WithQueryParameters(queryParameters), + ) + req, err = preparer.Prepare(&http.Request{}) + if err != nil { + return err + } + req = req.WithContext(originalReq.Context()) + + resp, err := autorest.SendWithSender(client, req, + autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...), + ) + if err != nil { + return err + } + + err = autorest.Respond( + resp, + WithErrorUnlessStatusCode(http.StatusOK), + autorest.ByUnmarshallingJSON(&provider), + autorest.ByClosing(), + ) + if err != nil { + return err + } + + if provider.RegistrationState != nil && + *provider.RegistrationState == "Registered" { + break + } + + delayed := autorest.DelayWithRetryAfter(resp, originalReq.Context().Done()) + if !delayed && !autorest.DelayForBackoff(client.PollingDelay, 0, originalReq.Context().Done()) { + return originalReq.Context().Err() + } + } + if !(time.Since(now) < client.PollingDuration) { + return errors.New("polling for resource provider registration has exceeded the polling duration") + } + return err +} + +func getSubscription(path string) string { + parts := strings.Split(path, "/") + for i, v := range parts { + if v == "subscriptions" && (i+1) < len(parts) { + return parts[i+1] + } + } + return "" +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/client.go b/vendor/github.com/Azure/go-autorest/autorest/client.go new file mode 100644 index 000000000..4e92dcad0 --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/client.go @@ -0,0 +1,264 @@ +package autorest + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "bytes" + "fmt" + "io" + "io/ioutil" + "log" + "net/http" + "net/http/cookiejar" + "runtime" + "time" +) + +const ( + // DefaultPollingDelay is a reasonable delay between polling requests. + DefaultPollingDelay = 60 * time.Second + + // DefaultPollingDuration is a reasonable total polling duration. + DefaultPollingDuration = 15 * time.Minute + + // DefaultRetryAttempts is number of attempts for retry status codes (5xx). + DefaultRetryAttempts = 3 + + // DefaultRetryDuration is the duration to wait between retries. + DefaultRetryDuration = 30 * time.Second +) + +var ( + // defaultUserAgent builds a string containing the Go version, system archityecture and OS, + // and the go-autorest version. + defaultUserAgent = fmt.Sprintf("Go/%s (%s-%s) go-autorest/%s", + runtime.Version(), + runtime.GOARCH, + runtime.GOOS, + Version(), + ) + + // StatusCodesForRetry are a defined group of status code for which the client will retry + StatusCodesForRetry = []int{ + http.StatusRequestTimeout, // 408 + http.StatusTooManyRequests, // 429 + http.StatusInternalServerError, // 500 + http.StatusBadGateway, // 502 + http.StatusServiceUnavailable, // 503 + http.StatusGatewayTimeout, // 504 + } +) + +const ( + requestFormat = `HTTP Request Begin =================================================== +%s +===================================================== HTTP Request End +` + responseFormat = `HTTP Response Begin =================================================== +%s +===================================================== HTTP Response End +` +) + +// Response serves as the base for all responses from generated clients. It provides access to the +// last http.Response. +type Response struct { + *http.Response `json:"-"` +} + +// LoggingInspector implements request and response inspectors that log the full request and +// response to a supplied log. +type LoggingInspector struct { + Logger *log.Logger +} + +// WithInspection returns a PrepareDecorator that emits the http.Request to the supplied logger. The +// body is restored after being emitted. +// +// Note: Since it reads the entire Body, this decorator should not be used where body streaming is +// important. It is best used to trace JSON or similar body values. +func (li LoggingInspector) WithInspection() PrepareDecorator { + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + var body, b bytes.Buffer + + defer r.Body.Close() + + r.Body = ioutil.NopCloser(io.TeeReader(r.Body, &body)) + if err := r.Write(&b); err != nil { + return nil, fmt.Errorf("Failed to write response: %v", err) + } + + li.Logger.Printf(requestFormat, b.String()) + + r.Body = ioutil.NopCloser(&body) + return p.Prepare(r) + }) + } +} + +// ByInspecting returns a RespondDecorator that emits the http.Response to the supplied logger. The +// body is restored after being emitted. +// +// Note: Since it reads the entire Body, this decorator should not be used where body streaming is +// important. It is best used to trace JSON or similar body values. +func (li LoggingInspector) ByInspecting() RespondDecorator { + return func(r Responder) Responder { + return ResponderFunc(func(resp *http.Response) error { + var body, b bytes.Buffer + defer resp.Body.Close() + resp.Body = ioutil.NopCloser(io.TeeReader(resp.Body, &body)) + if err := resp.Write(&b); err != nil { + return fmt.Errorf("Failed to write response: %v", err) + } + + li.Logger.Printf(responseFormat, b.String()) + + resp.Body = ioutil.NopCloser(&body) + return r.Respond(resp) + }) + } +} + +// Client is the base for autorest generated clients. It provides default, "do nothing" +// implementations of an Authorizer, RequestInspector, and ResponseInspector. It also returns the +// standard, undecorated http.Client as a default Sender. +// +// Generated clients should also use Error (see NewError and NewErrorWithError) for errors and +// return responses that compose with Response. +// +// Most customization of generated clients is best achieved by supplying a custom Authorizer, custom +// RequestInspector, and / or custom ResponseInspector. Users may log requests, implement circuit +// breakers (see https://msdn.microsoft.com/en-us/library/dn589784.aspx) or otherwise influence +// sending the request by providing a decorated Sender. +type Client struct { + Authorizer Authorizer + Sender Sender + RequestInspector PrepareDecorator + ResponseInspector RespondDecorator + + // PollingDelay sets the polling frequency used in absence of a Retry-After HTTP header + PollingDelay time.Duration + + // PollingDuration sets the maximum polling time after which an error is returned. + PollingDuration time.Duration + + // RetryAttempts sets the default number of retry attempts for client. + RetryAttempts int + + // RetryDuration sets the delay duration for retries. + RetryDuration time.Duration + + // UserAgent, if not empty, will be set as the HTTP User-Agent header on all requests sent + // through the Do method. + UserAgent string + + Jar http.CookieJar + + // Set to true to skip attempted registration of resource providers (false by default). + SkipResourceProviderRegistration bool +} + +// NewClientWithUserAgent returns an instance of a Client with the UserAgent set to the passed +// string. +func NewClientWithUserAgent(ua string) Client { + c := Client{ + PollingDelay: DefaultPollingDelay, + PollingDuration: DefaultPollingDuration, + RetryAttempts: DefaultRetryAttempts, + RetryDuration: DefaultRetryDuration, + UserAgent: defaultUserAgent, + } + c.Sender = c.sender() + c.AddToUserAgent(ua) + return c +} + +// AddToUserAgent adds an extension to the current user agent +func (c *Client) AddToUserAgent(extension string) error { + if extension != "" { + c.UserAgent = fmt.Sprintf("%s %s", c.UserAgent, extension) + return nil + } + return fmt.Errorf("Extension was empty, User Agent stayed as %s", c.UserAgent) +} + +// Do implements the Sender interface by invoking the active Sender after applying authorization. +// If Sender is not set, it uses a new instance of http.Client. In both cases it will, if UserAgent +// is set, apply set the User-Agent header. +func (c Client) Do(r *http.Request) (*http.Response, error) { + if r.UserAgent() == "" { + r, _ = Prepare(r, + WithUserAgent(c.UserAgent)) + } + // NOTE: c.WithInspection() must be last in the list so that it can inspect all preceding operations + r, err := Prepare(r, + c.WithAuthorization(), + c.WithInspection()) + if err != nil { + var resp *http.Response + if detErr, ok := err.(DetailedError); ok { + // if the authorization failed (e.g. invalid credentials) there will + // be a response associated with the error, be sure to return it. + resp = detErr.Response + } + return resp, NewErrorWithError(err, "autorest/Client", "Do", nil, "Preparing request failed") + } + + resp, err := SendWithSender(c.sender(), r) + Respond(resp, c.ByInspecting()) + return resp, err +} + +// sender returns the Sender to which to send requests. +func (c Client) sender() Sender { + if c.Sender == nil { + j, _ := cookiejar.New(nil) + return &http.Client{Jar: j} + } + return c.Sender +} + +// WithAuthorization is a convenience method that returns the WithAuthorization PrepareDecorator +// from the current Authorizer. If not Authorizer is set, it uses the NullAuthorizer. +func (c Client) WithAuthorization() PrepareDecorator { + return c.authorizer().WithAuthorization() +} + +// authorizer returns the Authorizer to use. +func (c Client) authorizer() Authorizer { + if c.Authorizer == nil { + return NullAuthorizer{} + } + return c.Authorizer +} + +// WithInspection is a convenience method that passes the request to the supplied RequestInspector, +// if present, or returns the WithNothing PrepareDecorator otherwise. +func (c Client) WithInspection() PrepareDecorator { + if c.RequestInspector == nil { + return WithNothing() + } + return c.RequestInspector +} + +// ByInspecting is a convenience method that passes the response to the supplied ResponseInspector, +// if present, or returns the ByIgnoring RespondDecorator otherwise. +func (c Client) ByInspecting() RespondDecorator { + if c.ResponseInspector == nil { + return ByIgnoring() + } + return c.ResponseInspector +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/error.go b/vendor/github.com/Azure/go-autorest/autorest/error.go new file mode 100644 index 000000000..f724f3332 --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/error.go @@ -0,0 +1,98 @@ +package autorest + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "fmt" + "net/http" +) + +const ( + // UndefinedStatusCode is used when HTTP status code is not available for an error. + UndefinedStatusCode = 0 +) + +// DetailedError encloses a error with details of the package, method, and associated HTTP +// status code (if any). +type DetailedError struct { + Original error + + // PackageType is the package type of the object emitting the error. For types, the value + // matches that produced the the '%T' format specifier of the fmt package. For other elements, + // such as functions, it is just the package name (e.g., "autorest"). + PackageType string + + // Method is the name of the method raising the error. + Method string + + // StatusCode is the HTTP Response StatusCode (if non-zero) that led to the error. + StatusCode interface{} + + // Message is the error message. + Message string + + // Service Error is the response body of failed API in bytes + ServiceError []byte + + // Response is the response object that was returned during failure if applicable. + Response *http.Response +} + +// NewError creates a new Error conforming object from the passed packageType, method, and +// message. message is treated as a format string to which the optional args apply. +func NewError(packageType string, method string, message string, args ...interface{}) DetailedError { + return NewErrorWithError(nil, packageType, method, nil, message, args...) +} + +// NewErrorWithResponse creates a new Error conforming object from the passed +// packageType, method, statusCode of the given resp (UndefinedStatusCode if +// resp is nil), and message. message is treated as a format string to which the +// optional args apply. +func NewErrorWithResponse(packageType string, method string, resp *http.Response, message string, args ...interface{}) DetailedError { + return NewErrorWithError(nil, packageType, method, resp, message, args...) +} + +// NewErrorWithError creates a new Error conforming object from the +// passed packageType, method, statusCode of the given resp (UndefinedStatusCode +// if resp is nil), message, and original error. message is treated as a format +// string to which the optional args apply. +func NewErrorWithError(original error, packageType string, method string, resp *http.Response, message string, args ...interface{}) DetailedError { + if v, ok := original.(DetailedError); ok { + return v + } + + statusCode := UndefinedStatusCode + if resp != nil { + statusCode = resp.StatusCode + } + + return DetailedError{ + Original: original, + PackageType: packageType, + Method: method, + StatusCode: statusCode, + Message: fmt.Sprintf(message, args...), + Response: resp, + } +} + +// Error returns a formatted containing all available details (i.e., PackageType, Method, +// StatusCode, Message, and original error (if any)). +func (e DetailedError) Error() string { + if e.Original == nil { + return fmt.Sprintf("%s#%s: %s: StatusCode=%d", e.PackageType, e.Method, e.Message, e.StatusCode) + } + return fmt.Sprintf("%s#%s: %s: StatusCode=%d -- Original Error: %v", e.PackageType, e.Method, e.Message, e.StatusCode, e.Original) +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/preparer.go b/vendor/github.com/Azure/go-autorest/autorest/preparer.go new file mode 100644 index 000000000..6d67bd733 --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/preparer.go @@ -0,0 +1,480 @@ +package autorest + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "bytes" + "encoding/json" + "fmt" + "io" + "io/ioutil" + "mime/multipart" + "net/http" + "net/url" + "strings" +) + +const ( + mimeTypeJSON = "application/json" + mimeTypeOctetStream = "application/octet-stream" + mimeTypeFormPost = "application/x-www-form-urlencoded" + + headerAuthorization = "Authorization" + headerContentType = "Content-Type" + headerUserAgent = "User-Agent" +) + +// Preparer is the interface that wraps the Prepare method. +// +// Prepare accepts and possibly modifies an http.Request (e.g., adding Headers). Implementations +// must ensure to not share or hold per-invocation state since Preparers may be shared and re-used. +type Preparer interface { + Prepare(*http.Request) (*http.Request, error) +} + +// PreparerFunc is a method that implements the Preparer interface. +type PreparerFunc func(*http.Request) (*http.Request, error) + +// Prepare implements the Preparer interface on PreparerFunc. +func (pf PreparerFunc) Prepare(r *http.Request) (*http.Request, error) { + return pf(r) +} + +// PrepareDecorator takes and possibly decorates, by wrapping, a Preparer. Decorators may affect the +// http.Request and pass it along or, first, pass the http.Request along then affect the result. +type PrepareDecorator func(Preparer) Preparer + +// CreatePreparer creates, decorates, and returns a Preparer. +// Without decorators, the returned Preparer returns the passed http.Request unmodified. +// Preparers are safe to share and re-use. +func CreatePreparer(decorators ...PrepareDecorator) Preparer { + return DecoratePreparer( + Preparer(PreparerFunc(func(r *http.Request) (*http.Request, error) { return r, nil })), + decorators...) +} + +// DecoratePreparer accepts a Preparer and a, possibly empty, set of PrepareDecorators, which it +// applies to the Preparer. Decorators are applied in the order received, but their affect upon the +// request depends on whether they are a pre-decorator (change the http.Request and then pass it +// along) or a post-decorator (pass the http.Request along and alter it on return). +func DecoratePreparer(p Preparer, decorators ...PrepareDecorator) Preparer { + for _, decorate := range decorators { + p = decorate(p) + } + return p +} + +// Prepare accepts an http.Request and a, possibly empty, set of PrepareDecorators. +// It creates a Preparer from the decorators which it then applies to the passed http.Request. +func Prepare(r *http.Request, decorators ...PrepareDecorator) (*http.Request, error) { + if r == nil { + return nil, NewError("autorest", "Prepare", "Invoked without an http.Request") + } + return CreatePreparer(decorators...).Prepare(r) +} + +// WithNothing returns a "do nothing" PrepareDecorator that makes no changes to the passed +// http.Request. +func WithNothing() PrepareDecorator { + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + return p.Prepare(r) + }) + } +} + +// WithHeader returns a PrepareDecorator that sets the specified HTTP header of the http.Request to +// the passed value. It canonicalizes the passed header name (via http.CanonicalHeaderKey) before +// adding the header. +func WithHeader(header string, value string) PrepareDecorator { + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + if r.Header == nil { + r.Header = make(http.Header) + } + r.Header.Set(http.CanonicalHeaderKey(header), value) + } + return r, err + }) + } +} + +// WithHeaders returns a PrepareDecorator that sets the specified HTTP headers of the http.Request to +// the passed value. It canonicalizes the passed headers name (via http.CanonicalHeaderKey) before +// adding them. +func WithHeaders(headers map[string]interface{}) PrepareDecorator { + h := ensureValueStrings(headers) + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + if r.Header == nil { + r.Header = make(http.Header) + } + + for name, value := range h { + r.Header.Set(http.CanonicalHeaderKey(name), value) + } + } + return r, err + }) + } +} + +// WithBearerAuthorization returns a PrepareDecorator that adds an HTTP Authorization header whose +// value is "Bearer " followed by the supplied token. +func WithBearerAuthorization(token string) PrepareDecorator { + return WithHeader(headerAuthorization, fmt.Sprintf("Bearer %s", token)) +} + +// AsContentType returns a PrepareDecorator that adds an HTTP Content-Type header whose value +// is the passed contentType. +func AsContentType(contentType string) PrepareDecorator { + return WithHeader(headerContentType, contentType) +} + +// WithUserAgent returns a PrepareDecorator that adds an HTTP User-Agent header whose value is the +// passed string. +func WithUserAgent(ua string) PrepareDecorator { + return WithHeader(headerUserAgent, ua) +} + +// AsFormURLEncoded returns a PrepareDecorator that adds an HTTP Content-Type header whose value is +// "application/x-www-form-urlencoded". +func AsFormURLEncoded() PrepareDecorator { + return AsContentType(mimeTypeFormPost) +} + +// AsJSON returns a PrepareDecorator that adds an HTTP Content-Type header whose value is +// "application/json". +func AsJSON() PrepareDecorator { + return AsContentType(mimeTypeJSON) +} + +// AsOctetStream returns a PrepareDecorator that adds the "application/octet-stream" Content-Type header. +func AsOctetStream() PrepareDecorator { + return AsContentType(mimeTypeOctetStream) +} + +// WithMethod returns a PrepareDecorator that sets the HTTP method of the passed request. The +// decorator does not validate that the passed method string is a known HTTP method. +func WithMethod(method string) PrepareDecorator { + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r.Method = method + return p.Prepare(r) + }) + } +} + +// AsDelete returns a PrepareDecorator that sets the HTTP method to DELETE. +func AsDelete() PrepareDecorator { return WithMethod("DELETE") } + +// AsGet returns a PrepareDecorator that sets the HTTP method to GET. +func AsGet() PrepareDecorator { return WithMethod("GET") } + +// AsHead returns a PrepareDecorator that sets the HTTP method to HEAD. +func AsHead() PrepareDecorator { return WithMethod("HEAD") } + +// AsOptions returns a PrepareDecorator that sets the HTTP method to OPTIONS. +func AsOptions() PrepareDecorator { return WithMethod("OPTIONS") } + +// AsPatch returns a PrepareDecorator that sets the HTTP method to PATCH. +func AsPatch() PrepareDecorator { return WithMethod("PATCH") } + +// AsPost returns a PrepareDecorator that sets the HTTP method to POST. +func AsPost() PrepareDecorator { return WithMethod("POST") } + +// AsPut returns a PrepareDecorator that sets the HTTP method to PUT. +func AsPut() PrepareDecorator { return WithMethod("PUT") } + +// WithBaseURL returns a PrepareDecorator that populates the http.Request with a url.URL constructed +// from the supplied baseUrl. +func WithBaseURL(baseURL string) PrepareDecorator { + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + var u *url.URL + if u, err = url.Parse(baseURL); err != nil { + return r, err + } + if u.Scheme == "" { + err = fmt.Errorf("autorest: No scheme detected in URL %s", baseURL) + } + if err == nil { + r.URL = u + } + } + return r, err + }) + } +} + +// WithCustomBaseURL returns a PrepareDecorator that replaces brace-enclosed keys within the +// request base URL (i.e., http.Request.URL) with the corresponding values from the passed map. +func WithCustomBaseURL(baseURL string, urlParameters map[string]interface{}) PrepareDecorator { + parameters := ensureValueStrings(urlParameters) + for key, value := range parameters { + baseURL = strings.Replace(baseURL, "{"+key+"}", value, -1) + } + return WithBaseURL(baseURL) +} + +// WithFormData returns a PrepareDecoratore that "URL encodes" (e.g., bar=baz&foo=quux) into the +// http.Request body. +func WithFormData(v url.Values) PrepareDecorator { + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + s := v.Encode() + + if r.Header == nil { + r.Header = make(http.Header) + } + r.Header.Set(http.CanonicalHeaderKey(headerContentType), mimeTypeFormPost) + r.ContentLength = int64(len(s)) + r.Body = ioutil.NopCloser(strings.NewReader(s)) + } + return r, err + }) + } +} + +// WithMultiPartFormData returns a PrepareDecoratore that "URL encodes" (e.g., bar=baz&foo=quux) form parameters +// into the http.Request body. +func WithMultiPartFormData(formDataParameters map[string]interface{}) PrepareDecorator { + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + var body bytes.Buffer + writer := multipart.NewWriter(&body) + for key, value := range formDataParameters { + if rc, ok := value.(io.ReadCloser); ok { + var fd io.Writer + if fd, err = writer.CreateFormFile(key, key); err != nil { + return r, err + } + if _, err = io.Copy(fd, rc); err != nil { + return r, err + } + } else { + if err = writer.WriteField(key, ensureValueString(value)); err != nil { + return r, err + } + } + } + if err = writer.Close(); err != nil { + return r, err + } + if r.Header == nil { + r.Header = make(http.Header) + } + r.Header.Set(http.CanonicalHeaderKey(headerContentType), writer.FormDataContentType()) + r.Body = ioutil.NopCloser(bytes.NewReader(body.Bytes())) + r.ContentLength = int64(body.Len()) + return r, err + } + return r, err + }) + } +} + +// WithFile returns a PrepareDecorator that sends file in request body. +func WithFile(f io.ReadCloser) PrepareDecorator { + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + b, err := ioutil.ReadAll(f) + if err != nil { + return r, err + } + r.Body = ioutil.NopCloser(bytes.NewReader(b)) + r.ContentLength = int64(len(b)) + } + return r, err + }) + } +} + +// WithBool returns a PrepareDecorator that encodes the passed bool into the body of the request +// and sets the Content-Length header. +func WithBool(v bool) PrepareDecorator { + return WithString(fmt.Sprintf("%v", v)) +} + +// WithFloat32 returns a PrepareDecorator that encodes the passed float32 into the body of the +// request and sets the Content-Length header. +func WithFloat32(v float32) PrepareDecorator { + return WithString(fmt.Sprintf("%v", v)) +} + +// WithFloat64 returns a PrepareDecorator that encodes the passed float64 into the body of the +// request and sets the Content-Length header. +func WithFloat64(v float64) PrepareDecorator { + return WithString(fmt.Sprintf("%v", v)) +} + +// WithInt32 returns a PrepareDecorator that encodes the passed int32 into the body of the request +// and sets the Content-Length header. +func WithInt32(v int32) PrepareDecorator { + return WithString(fmt.Sprintf("%v", v)) +} + +// WithInt64 returns a PrepareDecorator that encodes the passed int64 into the body of the request +// and sets the Content-Length header. +func WithInt64(v int64) PrepareDecorator { + return WithString(fmt.Sprintf("%v", v)) +} + +// WithString returns a PrepareDecorator that encodes the passed string into the body of the request +// and sets the Content-Length header. +func WithString(v string) PrepareDecorator { + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + r.ContentLength = int64(len(v)) + r.Body = ioutil.NopCloser(strings.NewReader(v)) + } + return r, err + }) + } +} + +// WithJSON returns a PrepareDecorator that encodes the data passed as JSON into the body of the +// request and sets the Content-Length header. +func WithJSON(v interface{}) PrepareDecorator { + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + b, err := json.Marshal(v) + if err == nil { + r.ContentLength = int64(len(b)) + r.Body = ioutil.NopCloser(bytes.NewReader(b)) + } + } + return r, err + }) + } +} + +// WithPath returns a PrepareDecorator that adds the supplied path to the request URL. If the path +// is absolute (that is, it begins with a "/"), it replaces the existing path. +func WithPath(path string) PrepareDecorator { + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + if r.URL == nil { + return r, NewError("autorest", "WithPath", "Invoked with a nil URL") + } + if r.URL, err = parseURL(r.URL, path); err != nil { + return r, err + } + } + return r, err + }) + } +} + +// WithEscapedPathParameters returns a PrepareDecorator that replaces brace-enclosed keys within the +// request path (i.e., http.Request.URL.Path) with the corresponding values from the passed map. The +// values will be escaped (aka URL encoded) before insertion into the path. +func WithEscapedPathParameters(path string, pathParameters map[string]interface{}) PrepareDecorator { + parameters := escapeValueStrings(ensureValueStrings(pathParameters)) + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + if r.URL == nil { + return r, NewError("autorest", "WithEscapedPathParameters", "Invoked with a nil URL") + } + for key, value := range parameters { + path = strings.Replace(path, "{"+key+"}", value, -1) + } + if r.URL, err = parseURL(r.URL, path); err != nil { + return r, err + } + } + return r, err + }) + } +} + +// WithPathParameters returns a PrepareDecorator that replaces brace-enclosed keys within the +// request path (i.e., http.Request.URL.Path) with the corresponding values from the passed map. +func WithPathParameters(path string, pathParameters map[string]interface{}) PrepareDecorator { + parameters := ensureValueStrings(pathParameters) + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + if r.URL == nil { + return r, NewError("autorest", "WithPathParameters", "Invoked with a nil URL") + } + for key, value := range parameters { + path = strings.Replace(path, "{"+key+"}", value, -1) + } + + if r.URL, err = parseURL(r.URL, path); err != nil { + return r, err + } + } + return r, err + }) + } +} + +func parseURL(u *url.URL, path string) (*url.URL, error) { + p := strings.TrimRight(u.String(), "/") + if !strings.HasPrefix(path, "/") { + path = "/" + path + } + return url.Parse(p + path) +} + +// WithQueryParameters returns a PrepareDecorators that encodes and applies the query parameters +// given in the supplied map (i.e., key=value). +func WithQueryParameters(queryParameters map[string]interface{}) PrepareDecorator { + parameters := ensureValueStrings(queryParameters) + return func(p Preparer) Preparer { + return PreparerFunc(func(r *http.Request) (*http.Request, error) { + r, err := p.Prepare(r) + if err == nil { + if r.URL == nil { + return r, NewError("autorest", "WithQueryParameters", "Invoked with a nil URL") + } + + v := r.URL.Query() + for key, value := range parameters { + d, err := url.QueryUnescape(value) + if err != nil { + return r, err + } + v.Add(key, d) + } + r.URL.RawQuery = v.Encode() + } + return r, err + }) + } +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/responder.go b/vendor/github.com/Azure/go-autorest/autorest/responder.go new file mode 100644 index 000000000..a908a0adb --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/responder.go @@ -0,0 +1,250 @@ +package autorest + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "bytes" + "encoding/json" + "encoding/xml" + "fmt" + "io" + "io/ioutil" + "net/http" + "strings" +) + +// Responder is the interface that wraps the Respond method. +// +// Respond accepts and reacts to an http.Response. Implementations must ensure to not share or hold +// state since Responders may be shared and re-used. +type Responder interface { + Respond(*http.Response) error +} + +// ResponderFunc is a method that implements the Responder interface. +type ResponderFunc func(*http.Response) error + +// Respond implements the Responder interface on ResponderFunc. +func (rf ResponderFunc) Respond(r *http.Response) error { + return rf(r) +} + +// RespondDecorator takes and possibly decorates, by wrapping, a Responder. Decorators may react to +// the http.Response and pass it along or, first, pass the http.Response along then react. +type RespondDecorator func(Responder) Responder + +// CreateResponder creates, decorates, and returns a Responder. Without decorators, the returned +// Responder returns the passed http.Response unmodified. Responders may or may not be safe to share +// and re-used: It depends on the applied decorators. For example, a standard decorator that closes +// the response body is fine to share whereas a decorator that reads the body into a passed struct +// is not. +// +// To prevent memory leaks, ensure that at least one Responder closes the response body. +func CreateResponder(decorators ...RespondDecorator) Responder { + return DecorateResponder( + Responder(ResponderFunc(func(r *http.Response) error { return nil })), + decorators...) +} + +// DecorateResponder accepts a Responder and a, possibly empty, set of RespondDecorators, which it +// applies to the Responder. Decorators are applied in the order received, but their affect upon the +// request depends on whether they are a pre-decorator (react to the http.Response and then pass it +// along) or a post-decorator (pass the http.Response along and then react). +func DecorateResponder(r Responder, decorators ...RespondDecorator) Responder { + for _, decorate := range decorators { + r = decorate(r) + } + return r +} + +// Respond accepts an http.Response and a, possibly empty, set of RespondDecorators. +// It creates a Responder from the decorators it then applies to the passed http.Response. +func Respond(r *http.Response, decorators ...RespondDecorator) error { + if r == nil { + return nil + } + return CreateResponder(decorators...).Respond(r) +} + +// ByIgnoring returns a RespondDecorator that ignores the passed http.Response passing it unexamined +// to the next RespondDecorator. +func ByIgnoring() RespondDecorator { + return func(r Responder) Responder { + return ResponderFunc(func(resp *http.Response) error { + return r.Respond(resp) + }) + } +} + +// ByCopying copies the contents of the http.Response Body into the passed bytes.Buffer as +// the Body is read. +func ByCopying(b *bytes.Buffer) RespondDecorator { + return func(r Responder) Responder { + return ResponderFunc(func(resp *http.Response) error { + err := r.Respond(resp) + if err == nil && resp != nil && resp.Body != nil { + resp.Body = TeeReadCloser(resp.Body, b) + } + return err + }) + } +} + +// ByDiscardingBody returns a RespondDecorator that first invokes the passed Responder after which +// it copies the remaining bytes (if any) in the response body to ioutil.Discard. Since the passed +// Responder is invoked prior to discarding the response body, the decorator may occur anywhere +// within the set. +func ByDiscardingBody() RespondDecorator { + return func(r Responder) Responder { + return ResponderFunc(func(resp *http.Response) error { + err := r.Respond(resp) + if err == nil && resp != nil && resp.Body != nil { + if _, err := io.Copy(ioutil.Discard, resp.Body); err != nil { + return fmt.Errorf("Error discarding the response body: %v", err) + } + } + return err + }) + } +} + +// ByClosing returns a RespondDecorator that first invokes the passed Responder after which it +// closes the response body. Since the passed Responder is invoked prior to closing the response +// body, the decorator may occur anywhere within the set. +func ByClosing() RespondDecorator { + return func(r Responder) Responder { + return ResponderFunc(func(resp *http.Response) error { + err := r.Respond(resp) + if resp != nil && resp.Body != nil { + if err := resp.Body.Close(); err != nil { + return fmt.Errorf("Error closing the response body: %v", err) + } + } + return err + }) + } +} + +// ByClosingIfError returns a RespondDecorator that first invokes the passed Responder after which +// it closes the response if the passed Responder returns an error and the response body exists. +func ByClosingIfError() RespondDecorator { + return func(r Responder) Responder { + return ResponderFunc(func(resp *http.Response) error { + err := r.Respond(resp) + if err != nil && resp != nil && resp.Body != nil { + if err := resp.Body.Close(); err != nil { + return fmt.Errorf("Error closing the response body: %v", err) + } + } + return err + }) + } +} + +// ByUnmarshallingJSON returns a RespondDecorator that decodes a JSON document returned in the +// response Body into the value pointed to by v. +func ByUnmarshallingJSON(v interface{}) RespondDecorator { + return func(r Responder) Responder { + return ResponderFunc(func(resp *http.Response) error { + err := r.Respond(resp) + if err == nil { + b, errInner := ioutil.ReadAll(resp.Body) + // Some responses might include a BOM, remove for successful unmarshalling + b = bytes.TrimPrefix(b, []byte("\xef\xbb\xbf")) + if errInner != nil { + err = fmt.Errorf("Error occurred reading http.Response#Body - Error = '%v'", errInner) + } else if len(strings.Trim(string(b), " ")) > 0 { + errInner = json.Unmarshal(b, v) + if errInner != nil { + err = fmt.Errorf("Error occurred unmarshalling JSON - Error = '%v' JSON = '%s'", errInner, string(b)) + } + } + } + return err + }) + } +} + +// ByUnmarshallingXML returns a RespondDecorator that decodes a XML document returned in the +// response Body into the value pointed to by v. +func ByUnmarshallingXML(v interface{}) RespondDecorator { + return func(r Responder) Responder { + return ResponderFunc(func(resp *http.Response) error { + err := r.Respond(resp) + if err == nil { + b, errInner := ioutil.ReadAll(resp.Body) + if errInner != nil { + err = fmt.Errorf("Error occurred reading http.Response#Body - Error = '%v'", errInner) + } else { + errInner = xml.Unmarshal(b, v) + if errInner != nil { + err = fmt.Errorf("Error occurred unmarshalling Xml - Error = '%v' Xml = '%s'", errInner, string(b)) + } + } + } + return err + }) + } +} + +// WithErrorUnlessStatusCode returns a RespondDecorator that emits an error unless the response +// StatusCode is among the set passed. On error, response body is fully read into a buffer and +// presented in the returned error, as well as in the response body. +func WithErrorUnlessStatusCode(codes ...int) RespondDecorator { + return func(r Responder) Responder { + return ResponderFunc(func(resp *http.Response) error { + err := r.Respond(resp) + if err == nil && !ResponseHasStatusCode(resp, codes...) { + derr := NewErrorWithResponse("autorest", "WithErrorUnlessStatusCode", resp, "%v %v failed with %s", + resp.Request.Method, + resp.Request.URL, + resp.Status) + if resp.Body != nil { + defer resp.Body.Close() + b, _ := ioutil.ReadAll(resp.Body) + derr.ServiceError = b + resp.Body = ioutil.NopCloser(bytes.NewReader(b)) + } + err = derr + } + return err + }) + } +} + +// WithErrorUnlessOK returns a RespondDecorator that emits an error if the response StatusCode is +// anything other than HTTP 200. +func WithErrorUnlessOK() RespondDecorator { + return WithErrorUnlessStatusCode(http.StatusOK) +} + +// ExtractHeader extracts all values of the specified header from the http.Response. It returns an +// empty string slice if the passed http.Response is nil or the header does not exist. +func ExtractHeader(header string, resp *http.Response) []string { + if resp != nil && resp.Header != nil { + return resp.Header[http.CanonicalHeaderKey(header)] + } + return nil +} + +// ExtractHeaderValue extracts the first value of the specified header from the http.Response. It +// returns an empty string if the passed http.Response is nil or the header does not exist. +func ExtractHeaderValue(header string, resp *http.Response) string { + h := ExtractHeader(header, resp) + if len(h) > 0 { + return h[0] + } + return "" +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/retriablerequest.go b/vendor/github.com/Azure/go-autorest/autorest/retriablerequest.go new file mode 100644 index 000000000..fa11dbed7 --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/retriablerequest.go @@ -0,0 +1,52 @@ +package autorest + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "bytes" + "io" + "io/ioutil" + "net/http" +) + +// NewRetriableRequest returns a wrapper around an HTTP request that support retry logic. +func NewRetriableRequest(req *http.Request) *RetriableRequest { + return &RetriableRequest{req: req} +} + +// Request returns the wrapped HTTP request. +func (rr *RetriableRequest) Request() *http.Request { + return rr.req +} + +func (rr *RetriableRequest) prepareFromByteReader() (err error) { + // fall back to making a copy (only do this once) + b := []byte{} + if rr.req.ContentLength > 0 { + b = make([]byte, rr.req.ContentLength) + _, err = io.ReadFull(rr.req.Body, b) + if err != nil { + return err + } + } else { + b, err = ioutil.ReadAll(rr.req.Body) + if err != nil { + return err + } + } + rr.br = bytes.NewReader(b) + rr.req.Body = ioutil.NopCloser(rr.br) + return err +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/retriablerequest_1.7.go b/vendor/github.com/Azure/go-autorest/autorest/retriablerequest_1.7.go new file mode 100644 index 000000000..7143cc61b --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/retriablerequest_1.7.go @@ -0,0 +1,54 @@ +// +build !go1.8 + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package autorest + +import ( + "bytes" + "io/ioutil" + "net/http" +) + +// RetriableRequest provides facilities for retrying an HTTP request. +type RetriableRequest struct { + req *http.Request + br *bytes.Reader +} + +// Prepare signals that the request is about to be sent. +func (rr *RetriableRequest) Prepare() (err error) { + // preserve the request body; this is to support retry logic as + // the underlying transport will always close the reqeust body + if rr.req.Body != nil { + if rr.br != nil { + _, err = rr.br.Seek(0, 0 /*io.SeekStart*/) + rr.req.Body = ioutil.NopCloser(rr.br) + } + if err != nil { + return err + } + if rr.br == nil { + // fall back to making a copy (only do this once) + err = rr.prepareFromByteReader() + } + } + return err +} + +func removeRequestBody(req *http.Request) { + req.Body = nil + req.ContentLength = 0 +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/retriablerequest_1.8.go b/vendor/github.com/Azure/go-autorest/autorest/retriablerequest_1.8.go new file mode 100644 index 000000000..ae15c6bf9 --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/retriablerequest_1.8.go @@ -0,0 +1,66 @@ +// +build go1.8 + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +package autorest + +import ( + "bytes" + "io" + "io/ioutil" + "net/http" +) + +// RetriableRequest provides facilities for retrying an HTTP request. +type RetriableRequest struct { + req *http.Request + rc io.ReadCloser + br *bytes.Reader +} + +// Prepare signals that the request is about to be sent. +func (rr *RetriableRequest) Prepare() (err error) { + // preserve the request body; this is to support retry logic as + // the underlying transport will always close the reqeust body + if rr.req.Body != nil { + if rr.rc != nil { + rr.req.Body = rr.rc + } else if rr.br != nil { + _, err = rr.br.Seek(0, io.SeekStart) + rr.req.Body = ioutil.NopCloser(rr.br) + } + if err != nil { + return err + } + if rr.req.GetBody != nil { + // this will allow us to preserve the body without having to + // make a copy. note we need to do this on each iteration + rr.rc, err = rr.req.GetBody() + if err != nil { + return err + } + } else if rr.br == nil { + // fall back to making a copy (only do this once) + err = rr.prepareFromByteReader() + } + } + return err +} + +func removeRequestBody(req *http.Request) { + req.Body = nil + req.GetBody = nil + req.ContentLength = 0 +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/sender.go b/vendor/github.com/Azure/go-autorest/autorest/sender.go new file mode 100644 index 000000000..cacbd8157 --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/sender.go @@ -0,0 +1,325 @@ +package autorest + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "fmt" + "log" + "math" + "net/http" + "strconv" + "time" +) + +// Sender is the interface that wraps the Do method to send HTTP requests. +// +// The standard http.Client conforms to this interface. +type Sender interface { + Do(*http.Request) (*http.Response, error) +} + +// SenderFunc is a method that implements the Sender interface. +type SenderFunc func(*http.Request) (*http.Response, error) + +// Do implements the Sender interface on SenderFunc. +func (sf SenderFunc) Do(r *http.Request) (*http.Response, error) { + return sf(r) +} + +// SendDecorator takes and possibily decorates, by wrapping, a Sender. Decorators may affect the +// http.Request and pass it along or, first, pass the http.Request along then react to the +// http.Response result. +type SendDecorator func(Sender) Sender + +// CreateSender creates, decorates, and returns, as a Sender, the default http.Client. +func CreateSender(decorators ...SendDecorator) Sender { + return DecorateSender(&http.Client{}, decorators...) +} + +// DecorateSender accepts a Sender and a, possibly empty, set of SendDecorators, which is applies to +// the Sender. Decorators are applied in the order received, but their affect upon the request +// depends on whether they are a pre-decorator (change the http.Request and then pass it along) or a +// post-decorator (pass the http.Request along and react to the results in http.Response). +func DecorateSender(s Sender, decorators ...SendDecorator) Sender { + for _, decorate := range decorators { + s = decorate(s) + } + return s +} + +// Send sends, by means of the default http.Client, the passed http.Request, returning the +// http.Response and possible error. It also accepts a, possibly empty, set of SendDecorators which +// it will apply the http.Client before invoking the Do method. +// +// Send is a convenience method and not recommended for production. Advanced users should use +// SendWithSender, passing and sharing their own Sender (e.g., instance of http.Client). +// +// Send will not poll or retry requests. +func Send(r *http.Request, decorators ...SendDecorator) (*http.Response, error) { + return SendWithSender(&http.Client{}, r, decorators...) +} + +// SendWithSender sends the passed http.Request, through the provided Sender, returning the +// http.Response and possible error. It also accepts a, possibly empty, set of SendDecorators which +// it will apply the http.Client before invoking the Do method. +// +// SendWithSender will not poll or retry requests. +func SendWithSender(s Sender, r *http.Request, decorators ...SendDecorator) (*http.Response, error) { + return DecorateSender(s, decorators...).Do(r) +} + +// AfterDelay returns a SendDecorator that delays for the passed time.Duration before +// invoking the Sender. The delay may be terminated by closing the optional channel on the +// http.Request. If canceled, no further Senders are invoked. +func AfterDelay(d time.Duration) SendDecorator { + return func(s Sender) Sender { + return SenderFunc(func(r *http.Request) (*http.Response, error) { + if !DelayForBackoff(d, 0, r.Context().Done()) { + return nil, fmt.Errorf("autorest: AfterDelay canceled before full delay") + } + return s.Do(r) + }) + } +} + +// AsIs returns a SendDecorator that invokes the passed Sender without modifying the http.Request. +func AsIs() SendDecorator { + return func(s Sender) Sender { + return SenderFunc(func(r *http.Request) (*http.Response, error) { + return s.Do(r) + }) + } +} + +// DoCloseIfError returns a SendDecorator that first invokes the passed Sender after which +// it closes the response if the passed Sender returns an error and the response body exists. +func DoCloseIfError() SendDecorator { + return func(s Sender) Sender { + return SenderFunc(func(r *http.Request) (*http.Response, error) { + resp, err := s.Do(r) + if err != nil { + Respond(resp, ByDiscardingBody(), ByClosing()) + } + return resp, err + }) + } +} + +// DoErrorIfStatusCode returns a SendDecorator that emits an error if the response StatusCode is +// among the set passed. Since these are artificial errors, the response body may still require +// closing. +func DoErrorIfStatusCode(codes ...int) SendDecorator { + return func(s Sender) Sender { + return SenderFunc(func(r *http.Request) (*http.Response, error) { + resp, err := s.Do(r) + if err == nil && ResponseHasStatusCode(resp, codes...) { + err = NewErrorWithResponse("autorest", "DoErrorIfStatusCode", resp, "%v %v failed with %s", + resp.Request.Method, + resp.Request.URL, + resp.Status) + } + return resp, err + }) + } +} + +// DoErrorUnlessStatusCode returns a SendDecorator that emits an error unless the response +// StatusCode is among the set passed. Since these are artificial errors, the response body +// may still require closing. +func DoErrorUnlessStatusCode(codes ...int) SendDecorator { + return func(s Sender) Sender { + return SenderFunc(func(r *http.Request) (*http.Response, error) { + resp, err := s.Do(r) + if err == nil && !ResponseHasStatusCode(resp, codes...) { + err = NewErrorWithResponse("autorest", "DoErrorUnlessStatusCode", resp, "%v %v failed with %s", + resp.Request.Method, + resp.Request.URL, + resp.Status) + } + return resp, err + }) + } +} + +// DoPollForStatusCodes returns a SendDecorator that polls if the http.Response contains one of the +// passed status codes. It expects the http.Response to contain a Location header providing the +// URL at which to poll (using GET) and will poll until the time passed is equal to or greater than +// the supplied duration. It will delay between requests for the duration specified in the +// RetryAfter header or, if the header is absent, the passed delay. Polling may be canceled by +// closing the optional channel on the http.Request. +func DoPollForStatusCodes(duration time.Duration, delay time.Duration, codes ...int) SendDecorator { + return func(s Sender) Sender { + return SenderFunc(func(r *http.Request) (resp *http.Response, err error) { + resp, err = s.Do(r) + + if err == nil && ResponseHasStatusCode(resp, codes...) { + r, err = NewPollingRequestWithContext(r.Context(), resp) + + for err == nil && ResponseHasStatusCode(resp, codes...) { + Respond(resp, + ByDiscardingBody(), + ByClosing()) + resp, err = SendWithSender(s, r, + AfterDelay(GetRetryAfter(resp, delay))) + } + } + + return resp, err + }) + } +} + +// DoRetryForAttempts returns a SendDecorator that retries a failed request for up to the specified +// number of attempts, exponentially backing off between requests using the supplied backoff +// time.Duration (which may be zero). Retrying may be canceled by closing the optional channel on +// the http.Request. +func DoRetryForAttempts(attempts int, backoff time.Duration) SendDecorator { + return func(s Sender) Sender { + return SenderFunc(func(r *http.Request) (resp *http.Response, err error) { + rr := NewRetriableRequest(r) + for attempt := 0; attempt < attempts; attempt++ { + err = rr.Prepare() + if err != nil { + return resp, err + } + resp, err = s.Do(rr.Request()) + if err == nil { + return resp, err + } + if !DelayForBackoff(backoff, attempt, r.Context().Done()) { + return nil, r.Context().Err() + } + } + return resp, err + }) + } +} + +// DoRetryForStatusCodes returns a SendDecorator that retries for specified statusCodes for up to the specified +// number of attempts, exponentially backing off between requests using the supplied backoff +// time.Duration (which may be zero). Retrying may be canceled by closing the optional channel on +// the http.Request. +func DoRetryForStatusCodes(attempts int, backoff time.Duration, codes ...int) SendDecorator { + return func(s Sender) Sender { + return SenderFunc(func(r *http.Request) (resp *http.Response, err error) { + rr := NewRetriableRequest(r) + // Increment to add the first call (attempts denotes number of retries) + attempts++ + for attempt := 0; attempt < attempts; { + err = rr.Prepare() + if err != nil { + return resp, err + } + resp, err = s.Do(rr.Request()) + // if the error isn't temporary don't bother retrying + if err != nil && !IsTemporaryNetworkError(err) { + return nil, err + } + // we want to retry if err is not nil (e.g. transient network failure). note that for failed authentication + // resp and err will both have a value, so in this case we don't want to retry as it will never succeed. + if err == nil && !ResponseHasStatusCode(resp, codes...) || IsTokenRefreshError(err) { + return resp, err + } + delayed := DelayWithRetryAfter(resp, r.Context().Done()) + if !delayed && !DelayForBackoff(backoff, attempt, r.Context().Done()) { + return nil, r.Context().Err() + } + // don't count a 429 against the number of attempts + // so that we continue to retry until it succeeds + if resp == nil || resp.StatusCode != http.StatusTooManyRequests { + attempt++ + } + } + return resp, err + }) + } +} + +// DelayWithRetryAfter invokes time.After for the duration specified in the "Retry-After" header in +// responses with status code 429 +func DelayWithRetryAfter(resp *http.Response, cancel <-chan struct{}) bool { + if resp == nil { + return false + } + retryAfter, _ := strconv.Atoi(resp.Header.Get("Retry-After")) + if resp.StatusCode == http.StatusTooManyRequests && retryAfter > 0 { + select { + case <-time.After(time.Duration(retryAfter) * time.Second): + return true + case <-cancel: + return false + } + } + return false +} + +// DoRetryForDuration returns a SendDecorator that retries the request until the total time is equal +// to or greater than the specified duration, exponentially backing off between requests using the +// supplied backoff time.Duration (which may be zero). Retrying may be canceled by closing the +// optional channel on the http.Request. +func DoRetryForDuration(d time.Duration, backoff time.Duration) SendDecorator { + return func(s Sender) Sender { + return SenderFunc(func(r *http.Request) (resp *http.Response, err error) { + rr := NewRetriableRequest(r) + end := time.Now().Add(d) + for attempt := 0; time.Now().Before(end); attempt++ { + err = rr.Prepare() + if err != nil { + return resp, err + } + resp, err = s.Do(rr.Request()) + if err == nil { + return resp, err + } + if !DelayForBackoff(backoff, attempt, r.Context().Done()) { + return nil, r.Context().Err() + } + } + return resp, err + }) + } +} + +// WithLogging returns a SendDecorator that implements simple before and after logging of the +// request. +func WithLogging(logger *log.Logger) SendDecorator { + return func(s Sender) Sender { + return SenderFunc(func(r *http.Request) (*http.Response, error) { + logger.Printf("Sending %s %s", r.Method, r.URL) + resp, err := s.Do(r) + if err != nil { + logger.Printf("%s %s received error '%v'", r.Method, r.URL, err) + } else { + logger.Printf("%s %s received %s", r.Method, r.URL, resp.Status) + } + return resp, err + }) + } +} + +// DelayForBackoff invokes time.After for the supplied backoff duration raised to the power of +// passed attempt (i.e., an exponential backoff delay). Backoff duration is in seconds and can set +// to zero for no delay. The delay may be canceled by closing the passed channel. If terminated early, +// returns false. +// Note: Passing attempt 1 will result in doubling "backoff" duration. Treat this as a zero-based attempt +// count. +func DelayForBackoff(backoff time.Duration, attempt int, cancel <-chan struct{}) bool { + select { + case <-time.After(time.Duration(backoff.Seconds()*math.Pow(2, float64(attempt))) * time.Second): + return true + case <-cancel: + return false + } +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/to/convert.go b/vendor/github.com/Azure/go-autorest/autorest/to/convert.go new file mode 100644 index 000000000..fdda2ce1a --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/to/convert.go @@ -0,0 +1,147 @@ +/* +Package to provides helpers to ease working with pointer values of marshalled structures. +*/ +package to + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// String returns a string value for the passed string pointer. It returns the empty string if the +// pointer is nil. +func String(s *string) string { + if s != nil { + return *s + } + return "" +} + +// StringPtr returns a pointer to the passed string. +func StringPtr(s string) *string { + return &s +} + +// StringSlice returns a string slice value for the passed string slice pointer. It returns a nil +// slice if the pointer is nil. +func StringSlice(s *[]string) []string { + if s != nil { + return *s + } + return nil +} + +// StringSlicePtr returns a pointer to the passed string slice. +func StringSlicePtr(s []string) *[]string { + return &s +} + +// StringMap returns a map of strings built from the map of string pointers. The empty string is +// used for nil pointers. +func StringMap(msp map[string]*string) map[string]string { + ms := make(map[string]string, len(msp)) + for k, sp := range msp { + if sp != nil { + ms[k] = *sp + } else { + ms[k] = "" + } + } + return ms +} + +// StringMapPtr returns a pointer to a map of string pointers built from the passed map of strings. +func StringMapPtr(ms map[string]string) *map[string]*string { + msp := make(map[string]*string, len(ms)) + for k, s := range ms { + msp[k] = StringPtr(s) + } + return &msp +} + +// Bool returns a bool value for the passed bool pointer. It returns false if the pointer is nil. +func Bool(b *bool) bool { + if b != nil { + return *b + } + return false +} + +// BoolPtr returns a pointer to the passed bool. +func BoolPtr(b bool) *bool { + return &b +} + +// Int returns an int value for the passed int pointer. It returns 0 if the pointer is nil. +func Int(i *int) int { + if i != nil { + return *i + } + return 0 +} + +// IntPtr returns a pointer to the passed int. +func IntPtr(i int) *int { + return &i +} + +// Int32 returns an int value for the passed int pointer. It returns 0 if the pointer is nil. +func Int32(i *int32) int32 { + if i != nil { + return *i + } + return 0 +} + +// Int32Ptr returns a pointer to the passed int32. +func Int32Ptr(i int32) *int32 { + return &i +} + +// Int64 returns an int value for the passed int pointer. It returns 0 if the pointer is nil. +func Int64(i *int64) int64 { + if i != nil { + return *i + } + return 0 +} + +// Int64Ptr returns a pointer to the passed int64. +func Int64Ptr(i int64) *int64 { + return &i +} + +// Float32 returns an int value for the passed int pointer. It returns 0.0 if the pointer is nil. +func Float32(i *float32) float32 { + if i != nil { + return *i + } + return 0.0 +} + +// Float32Ptr returns a pointer to the passed float32. +func Float32Ptr(i float32) *float32 { + return &i +} + +// Float64 returns an int value for the passed int pointer. It returns 0.0 if the pointer is nil. +func Float64(i *float64) float64 { + if i != nil { + return *i + } + return 0.0 +} + +// Float64Ptr returns a pointer to the passed float64. +func Float64Ptr(i float64) *float64 { + return &i +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/utility.go b/vendor/github.com/Azure/go-autorest/autorest/utility.go new file mode 100644 index 000000000..f3a42bfc2 --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/utility.go @@ -0,0 +1,227 @@ +package autorest + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "bytes" + "encoding/json" + "encoding/xml" + "fmt" + "io" + "net" + "net/http" + "net/url" + "reflect" + "strings" + + "github.com/Azure/go-autorest/autorest/adal" +) + +// EncodedAs is a series of constants specifying various data encodings +type EncodedAs string + +const ( + // EncodedAsJSON states that data is encoded as JSON + EncodedAsJSON EncodedAs = "JSON" + + // EncodedAsXML states that data is encoded as Xml + EncodedAsXML EncodedAs = "XML" +) + +// Decoder defines the decoding method json.Decoder and xml.Decoder share +type Decoder interface { + Decode(v interface{}) error +} + +// NewDecoder creates a new decoder appropriate to the passed encoding. +// encodedAs specifies the type of encoding and r supplies the io.Reader containing the +// encoded data. +func NewDecoder(encodedAs EncodedAs, r io.Reader) Decoder { + if encodedAs == EncodedAsJSON { + return json.NewDecoder(r) + } else if encodedAs == EncodedAsXML { + return xml.NewDecoder(r) + } + return nil +} + +// CopyAndDecode decodes the data from the passed io.Reader while making a copy. Having a copy +// is especially useful if there is a chance the data will fail to decode. +// encodedAs specifies the expected encoding, r provides the io.Reader to the data, and v +// is the decoding destination. +func CopyAndDecode(encodedAs EncodedAs, r io.Reader, v interface{}) (bytes.Buffer, error) { + b := bytes.Buffer{} + return b, NewDecoder(encodedAs, io.TeeReader(r, &b)).Decode(v) +} + +// TeeReadCloser returns a ReadCloser that writes to w what it reads from rc. +// It utilizes io.TeeReader to copy the data read and has the same behavior when reading. +// Further, when it is closed, it ensures that rc is closed as well. +func TeeReadCloser(rc io.ReadCloser, w io.Writer) io.ReadCloser { + return &teeReadCloser{rc, io.TeeReader(rc, w)} +} + +type teeReadCloser struct { + rc io.ReadCloser + r io.Reader +} + +func (t *teeReadCloser) Read(p []byte) (int, error) { + return t.r.Read(p) +} + +func (t *teeReadCloser) Close() error { + return t.rc.Close() +} + +func containsInt(ints []int, n int) bool { + for _, i := range ints { + if i == n { + return true + } + } + return false +} + +func escapeValueStrings(m map[string]string) map[string]string { + for key, value := range m { + m[key] = url.QueryEscape(value) + } + return m +} + +func ensureValueStrings(mapOfInterface map[string]interface{}) map[string]string { + mapOfStrings := make(map[string]string) + for key, value := range mapOfInterface { + mapOfStrings[key] = ensureValueString(value) + } + return mapOfStrings +} + +func ensureValueString(value interface{}) string { + if value == nil { + return "" + } + switch v := value.(type) { + case string: + return v + case []byte: + return string(v) + default: + return fmt.Sprintf("%v", v) + } +} + +// MapToValues method converts map[string]interface{} to url.Values. +func MapToValues(m map[string]interface{}) url.Values { + v := url.Values{} + for key, value := range m { + x := reflect.ValueOf(value) + if x.Kind() == reflect.Array || x.Kind() == reflect.Slice { + for i := 0; i < x.Len(); i++ { + v.Add(key, ensureValueString(x.Index(i))) + } + } else { + v.Add(key, ensureValueString(value)) + } + } + return v +} + +// AsStringSlice method converts interface{} to []string. This expects a +//that the parameter passed to be a slice or array of a type that has the underlying +//type a string. +func AsStringSlice(s interface{}) ([]string, error) { + v := reflect.ValueOf(s) + if v.Kind() != reflect.Slice && v.Kind() != reflect.Array { + return nil, NewError("autorest", "AsStringSlice", "the value's type is not an array.") + } + stringSlice := make([]string, 0, v.Len()) + + for i := 0; i < v.Len(); i++ { + stringSlice = append(stringSlice, v.Index(i).String()) + } + return stringSlice, nil +} + +// String method converts interface v to string. If interface is a list, it +// joins list elements using the seperator. Note that only sep[0] will be used for +// joining if any separator is specified. +func String(v interface{}, sep ...string) string { + if len(sep) == 0 { + return ensureValueString(v) + } + stringSlice, ok := v.([]string) + if ok == false { + var err error + stringSlice, err = AsStringSlice(v) + if err != nil { + panic(fmt.Sprintf("autorest: Couldn't convert value to a string %s.", err)) + } + } + return ensureValueString(strings.Join(stringSlice, sep[0])) +} + +// Encode method encodes url path and query parameters. +func Encode(location string, v interface{}, sep ...string) string { + s := String(v, sep...) + switch strings.ToLower(location) { + case "path": + return pathEscape(s) + case "query": + return queryEscape(s) + default: + return s + } +} + +func pathEscape(s string) string { + return strings.Replace(url.QueryEscape(s), "+", "%20", -1) +} + +func queryEscape(s string) string { + return url.QueryEscape(s) +} + +// ChangeToGet turns the specified http.Request into a GET (it assumes it wasn't). +// This is mainly useful for long-running operations that use the Azure-AsyncOperation +// header, so we change the initial PUT into a GET to retrieve the final result. +func ChangeToGet(req *http.Request) *http.Request { + req.Method = "GET" + req.Body = nil + req.ContentLength = 0 + req.Header.Del("Content-Length") + return req +} + +// IsTokenRefreshError returns true if the specified error implements the TokenRefreshError +// interface. If err is a DetailedError it will walk the chain of Original errors. +func IsTokenRefreshError(err error) bool { + if _, ok := err.(adal.TokenRefreshError); ok { + return true + } + if de, ok := err.(DetailedError); ok { + return IsTokenRefreshError(de.Original) + } + return false +} + +// IsTemporaryNetworkError returns true if the specified error is a temporary network error. +func IsTemporaryNetworkError(err error) bool { + if netErr, ok := err.(net.Error); ok && netErr.Temporary() { + return true + } + return false +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/validation/error.go b/vendor/github.com/Azure/go-autorest/autorest/validation/error.go new file mode 100644 index 000000000..fed156dbf --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/validation/error.go @@ -0,0 +1,48 @@ +package validation + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "fmt" +) + +// Error is the type that's returned when the validation of an APIs arguments constraints fails. +type Error struct { + // PackageType is the package type of the object emitting the error. For types, the value + // matches that produced the the '%T' format specifier of the fmt package. For other elements, + // such as functions, it is just the package name (e.g., "autorest"). + PackageType string + + // Method is the name of the method raising the error. + Method string + + // Message is the error message. + Message string +} + +// Error returns a string containing the details of the validation failure. +func (e Error) Error() string { + return fmt.Sprintf("%s#%s: Invalid input: %s", e.PackageType, e.Method, e.Message) +} + +// NewError creates a new Error object with the specified parameters. +// message is treated as a format string to which the optional args apply. +func NewError(packageType string, method string, message string, args ...interface{}) Error { + return Error{ + PackageType: packageType, + Method: method, + Message: fmt.Sprintf(message, args...), + } +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/validation/validation.go b/vendor/github.com/Azure/go-autorest/autorest/validation/validation.go new file mode 100644 index 000000000..d886e0b3f --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/validation/validation.go @@ -0,0 +1,397 @@ +/* +Package validation provides methods for validating parameter value using reflection. +*/ +package validation + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +import ( + "fmt" + "reflect" + "regexp" + "strings" +) + +// Constraint stores constraint name, target field name +// Rule and chain validations. +type Constraint struct { + + // Target field name for validation. + Target string + + // Constraint name e.g. minLength, MaxLength, Pattern, etc. + Name string + + // Rule for constraint e.g. greater than 10, less than 5 etc. + Rule interface{} + + // Chain Validations for struct type + Chain []Constraint +} + +// Validation stores parameter-wise validation. +type Validation struct { + TargetValue interface{} + Constraints []Constraint +} + +// Constraint list +const ( + Empty = "Empty" + Null = "Null" + ReadOnly = "ReadOnly" + Pattern = "Pattern" + MaxLength = "MaxLength" + MinLength = "MinLength" + MaxItems = "MaxItems" + MinItems = "MinItems" + MultipleOf = "MultipleOf" + UniqueItems = "UniqueItems" + InclusiveMaximum = "InclusiveMaximum" + ExclusiveMaximum = "ExclusiveMaximum" + ExclusiveMinimum = "ExclusiveMinimum" + InclusiveMinimum = "InclusiveMinimum" +) + +// Validate method validates constraints on parameter +// passed in validation array. +func Validate(m []Validation) error { + for _, item := range m { + v := reflect.ValueOf(item.TargetValue) + for _, constraint := range item.Constraints { + var err error + switch v.Kind() { + case reflect.Ptr: + err = validatePtr(v, constraint) + case reflect.String: + err = validateString(v, constraint) + case reflect.Struct: + err = validateStruct(v, constraint) + case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64: + err = validateInt(v, constraint) + case reflect.Float32, reflect.Float64: + err = validateFloat(v, constraint) + case reflect.Array, reflect.Slice, reflect.Map: + err = validateArrayMap(v, constraint) + default: + err = createError(v, constraint, fmt.Sprintf("unknown type %v", v.Kind())) + } + + if err != nil { + return err + } + } + } + return nil +} + +func validateStruct(x reflect.Value, v Constraint, name ...string) error { + //Get field name from target name which is in format a.b.c + s := strings.Split(v.Target, ".") + f := x.FieldByName(s[len(s)-1]) + if isZero(f) { + return createError(x, v, fmt.Sprintf("field %q doesn't exist", v.Target)) + } + + return Validate([]Validation{ + { + TargetValue: getInterfaceValue(f), + Constraints: []Constraint{v}, + }, + }) +} + +func validatePtr(x reflect.Value, v Constraint) error { + if v.Name == ReadOnly { + if !x.IsNil() { + return createError(x.Elem(), v, "readonly parameter; must send as nil or empty in request") + } + return nil + } + if x.IsNil() { + return checkNil(x, v) + } + if v.Chain != nil { + return Validate([]Validation{ + { + TargetValue: getInterfaceValue(x.Elem()), + Constraints: v.Chain, + }, + }) + } + return nil +} + +func validateInt(x reflect.Value, v Constraint) error { + i := x.Int() + r, ok := v.Rule.(int) + if !ok { + return createError(x, v, fmt.Sprintf("rule must be integer value for %v constraint; got: %v", v.Name, v.Rule)) + } + switch v.Name { + case MultipleOf: + if i%int64(r) != 0 { + return createError(x, v, fmt.Sprintf("value must be a multiple of %v", r)) + } + case ExclusiveMinimum: + if i <= int64(r) { + return createError(x, v, fmt.Sprintf("value must be greater than %v", r)) + } + case ExclusiveMaximum: + if i >= int64(r) { + return createError(x, v, fmt.Sprintf("value must be less than %v", r)) + } + case InclusiveMinimum: + if i < int64(r) { + return createError(x, v, fmt.Sprintf("value must be greater than or equal to %v", r)) + } + case InclusiveMaximum: + if i > int64(r) { + return createError(x, v, fmt.Sprintf("value must be less than or equal to %v", r)) + } + default: + return createError(x, v, fmt.Sprintf("constraint %v is not applicable for type integer", v.Name)) + } + return nil +} + +func validateFloat(x reflect.Value, v Constraint) error { + f := x.Float() + r, ok := v.Rule.(float64) + if !ok { + return createError(x, v, fmt.Sprintf("rule must be float value for %v constraint; got: %v", v.Name, v.Rule)) + } + switch v.Name { + case ExclusiveMinimum: + if f <= r { + return createError(x, v, fmt.Sprintf("value must be greater than %v", r)) + } + case ExclusiveMaximum: + if f >= r { + return createError(x, v, fmt.Sprintf("value must be less than %v", r)) + } + case InclusiveMinimum: + if f < r { + return createError(x, v, fmt.Sprintf("value must be greater than or equal to %v", r)) + } + case InclusiveMaximum: + if f > r { + return createError(x, v, fmt.Sprintf("value must be less than or equal to %v", r)) + } + default: + return createError(x, v, fmt.Sprintf("constraint %s is not applicable for type float", v.Name)) + } + return nil +} + +func validateString(x reflect.Value, v Constraint) error { + s := x.String() + switch v.Name { + case Empty: + if len(s) == 0 { + return checkEmpty(x, v) + } + case Pattern: + reg, err := regexp.Compile(v.Rule.(string)) + if err != nil { + return createError(x, v, err.Error()) + } + if !reg.MatchString(s) { + return createError(x, v, fmt.Sprintf("value doesn't match pattern %v", v.Rule)) + } + case MaxLength: + if _, ok := v.Rule.(int); !ok { + return createError(x, v, fmt.Sprintf("rule must be integer value for %v constraint; got: %v", v.Name, v.Rule)) + } + if len(s) > v.Rule.(int) { + return createError(x, v, fmt.Sprintf("value length must be less than or equal to %v", v.Rule)) + } + case MinLength: + if _, ok := v.Rule.(int); !ok { + return createError(x, v, fmt.Sprintf("rule must be integer value for %v constraint; got: %v", v.Name, v.Rule)) + } + if len(s) < v.Rule.(int) { + return createError(x, v, fmt.Sprintf("value length must be greater than or equal to %v", v.Rule)) + } + case ReadOnly: + if len(s) > 0 { + return createError(reflect.ValueOf(s), v, "readonly parameter; must send as nil or empty in request") + } + default: + return createError(x, v, fmt.Sprintf("constraint %s is not applicable to string type", v.Name)) + } + + if v.Chain != nil { + return Validate([]Validation{ + { + TargetValue: getInterfaceValue(x), + Constraints: v.Chain, + }, + }) + } + return nil +} + +func validateArrayMap(x reflect.Value, v Constraint) error { + switch v.Name { + case Null: + if x.IsNil() { + return checkNil(x, v) + } + case Empty: + if x.IsNil() || x.Len() == 0 { + return checkEmpty(x, v) + } + case MaxItems: + if _, ok := v.Rule.(int); !ok { + return createError(x, v, fmt.Sprintf("rule must be integer for %v constraint; got: %v", v.Name, v.Rule)) + } + if x.Len() > v.Rule.(int) { + return createError(x, v, fmt.Sprintf("maximum item limit is %v; got: %v", v.Rule, x.Len())) + } + case MinItems: + if _, ok := v.Rule.(int); !ok { + return createError(x, v, fmt.Sprintf("rule must be integer for %v constraint; got: %v", v.Name, v.Rule)) + } + if x.Len() < v.Rule.(int) { + return createError(x, v, fmt.Sprintf("minimum item limit is %v; got: %v", v.Rule, x.Len())) + } + case UniqueItems: + if x.Kind() == reflect.Array || x.Kind() == reflect.Slice { + if !checkForUniqueInArray(x) { + return createError(x, v, fmt.Sprintf("all items in parameter %q must be unique; got:%v", v.Target, x)) + } + } else if x.Kind() == reflect.Map { + if !checkForUniqueInMap(x) { + return createError(x, v, fmt.Sprintf("all items in parameter %q must be unique; got:%v", v.Target, x)) + } + } else { + return createError(x, v, fmt.Sprintf("type must be array, slice or map for constraint %v; got: %v", v.Name, x.Kind())) + } + case ReadOnly: + if x.Len() != 0 { + return createError(x, v, "readonly parameter; must send as nil or empty in request") + } + case Pattern: + reg, err := regexp.Compile(v.Rule.(string)) + if err != nil { + return createError(x, v, err.Error()) + } + keys := x.MapKeys() + for _, k := range keys { + if !reg.MatchString(k.String()) { + return createError(k, v, fmt.Sprintf("map key doesn't match pattern %v", v.Rule)) + } + } + default: + return createError(x, v, fmt.Sprintf("constraint %v is not applicable to array, slice and map type", v.Name)) + } + + if v.Chain != nil { + return Validate([]Validation{ + { + TargetValue: getInterfaceValue(x), + Constraints: v.Chain, + }, + }) + } + return nil +} + +func checkNil(x reflect.Value, v Constraint) error { + if _, ok := v.Rule.(bool); !ok { + return createError(x, v, fmt.Sprintf("rule must be bool value for %v constraint; got: %v", v.Name, v.Rule)) + } + if v.Rule.(bool) { + return createError(x, v, "value can not be null; required parameter") + } + return nil +} + +func checkEmpty(x reflect.Value, v Constraint) error { + if _, ok := v.Rule.(bool); !ok { + return createError(x, v, fmt.Sprintf("rule must be bool value for %v constraint; got: %v", v.Name, v.Rule)) + } + + if v.Rule.(bool) { + return createError(x, v, "value can not be null or empty; required parameter") + } + return nil +} + +func checkForUniqueInArray(x reflect.Value) bool { + if x == reflect.Zero(reflect.TypeOf(x)) || x.Len() == 0 { + return false + } + arrOfInterface := make([]interface{}, x.Len()) + + for i := 0; i < x.Len(); i++ { + arrOfInterface[i] = x.Index(i).Interface() + } + + m := make(map[interface{}]bool) + for _, val := range arrOfInterface { + if m[val] { + return false + } + m[val] = true + } + return true +} + +func checkForUniqueInMap(x reflect.Value) bool { + if x == reflect.Zero(reflect.TypeOf(x)) || x.Len() == 0 { + return false + } + mapOfInterface := make(map[interface{}]interface{}, x.Len()) + + keys := x.MapKeys() + for _, k := range keys { + mapOfInterface[k.Interface()] = x.MapIndex(k).Interface() + } + + m := make(map[interface{}]bool) + for _, val := range mapOfInterface { + if m[val] { + return false + } + m[val] = true + } + return true +} + +func getInterfaceValue(x reflect.Value) interface{} { + if x.Kind() == reflect.Invalid { + return nil + } + return x.Interface() +} + +func isZero(x interface{}) bool { + return x == reflect.Zero(reflect.TypeOf(x)).Interface() +} + +func createError(x reflect.Value, v Constraint, err string) error { + return fmt.Errorf("autorest/validation: validation failed: parameter=%s constraint=%s value=%#v details: %s", + v.Target, v.Name, getInterfaceValue(x), err) +} + +// NewErrorWithValidationError appends package type and method name in +// validation error. +// +// Deprecated: Please use validation.NewError() instead. +func NewErrorWithValidationError(err error, packageType, method string) error { + return NewError(packageType, method, err.Error()) +} diff --git a/vendor/github.com/Azure/go-autorest/autorest/version.go b/vendor/github.com/Azure/go-autorest/autorest/version.go new file mode 100644 index 000000000..d265055f5 --- /dev/null +++ b/vendor/github.com/Azure/go-autorest/autorest/version.go @@ -0,0 +1,20 @@ +package autorest + +// Copyright 2017 Microsoft Corporation +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +// Version returns the semantic version (see http://semver.org). +func Version() string { + return "v10.8.1" +} diff --git a/vendor/github.com/lawrencegripper/pod2docker/LICENSE b/vendor/github.com/lawrencegripper/pod2docker/LICENSE new file mode 100644 index 000000000..b68e1a45c --- /dev/null +++ b/vendor/github.com/lawrencegripper/pod2docker/LICENSE @@ -0,0 +1,21 @@ +MIT License + +Copyright (c) 2018 Lawrence Gripper + +Permission is hereby granted, free of charge, to any person obtaining a copy +of this software and associated documentation files (the "Software"), to deal +in the Software without restriction, including without limitation the rights +to use, copy, modify, merge, publish, distribute, sublicense, and/or sell +copies of the Software, and to permit persons to whom the Software is +furnished to do so, subject to the following conditions: + +The above copyright notice and this permission notice shall be included in all +copies or substantial portions of the Software. + +THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +SOFTWARE. diff --git a/vendor/github.com/lawrencegripper/pod2docker/pod2docker.go b/vendor/github.com/lawrencegripper/pod2docker/pod2docker.go new file mode 100644 index 000000000..007aeecd4 --- /dev/null +++ b/vendor/github.com/lawrencegripper/pod2docker/pod2docker.go @@ -0,0 +1,105 @@ +package pod2docker + +import ( + "bytes" + "strings" + "text/template" + + "k8s.io/api/core/v1" +) + +// ImageRegistryCredential - Used to input a credential used by docker login +type ImageRegistryCredential struct { + Server string `json:"server,omitempty"` + Username string `json:"username,omitempty"` + Password string `json:"password,omitempty"` +} + +// PodComponents provides details to run a pod +type PodComponents struct { + PullCredentials []ImageRegistryCredential + InitContainers []v1.Container + Containers []v1.Container + Volumes []v1.Volume + PodName string +} + +// GetBashCommand generates the bash script to execute the pod +func GetBashCommand(p PodComponents) (string, error) { + template := template.New("run.sh.tmpl").Option("missingkey=error").Funcs(template.FuncMap{ + "getLaunchCommand": getLaunchCommand, + "isHostPathVolume": isHostPathVolume, + "isEmptyDirVolume": isEmptyDirVolume, + "isPullAlways": isPullAlways, + "getValidVolumeMounts": getValidVolumeMounts, + "isNvidiaRuntime": isNvidiaRuntime, + }) + + template, err := template.Parse(azureBatchPodTemplate) + if err != nil { + return "", err + } + var output bytes.Buffer + err = template.Execute(&output, p) + return output.String(), err +} + +func getLaunchCommand(container v1.Container) (cmd string) { + if len(container.Command) > 0 { + cmd += strings.Join(container.Command, " ") + } + if len(cmd) > 0 { + cmd += " " + } + if len(container.Args) > 0 { + cmd += strings.Join(container.Args, " ") + } + return +} + +func isNvidiaRuntime(c v1.Container) bool { + if _, exists := c.Resources.Limits["nvidia.com/gpu"]; exists { + return true + } + return false +} + +func isHostPathVolume(v v1.Volume) bool { + if v.HostPath == nil { + return false + } + return true +} + +func isEmptyDirVolume(v v1.Volume) bool { + if v.EmptyDir == nil { + return false + } + return true +} + +func isPullAlways(c v1.Container) bool { + if c.ImagePullPolicy == v1.PullAlways { + return true + } + return false +} + +func getValidVolumeMounts(container v1.Container, volumes []v1.Volume) []v1.VolumeMount { + volDic := make(map[string]v1.Volume) + for _, vol := range volumes { + volDic[vol.Name] = vol + } + var mounts []v1.VolumeMount + for _, mount := range container.VolumeMounts { + vol, ok := volDic[mount.Name] + if !ok { + continue + } + if vol.EmptyDir == nil && vol.HostPath == nil { + continue + } + mounts = append(mounts, mount) + } + return mounts +} diff --git a/vendor/github.com/lawrencegripper/pod2docker/pod2docker_template.go b/vendor/github.com/lawrencegripper/pod2docker/pod2docker_template.go new file mode 100644 index 000000000..e7bd745f0 --- /dev/null +++ b/vendor/github.com/lawrencegripper/pod2docker/pod2docker_template.go @@ -0,0 +1,167 @@ +package pod2docker + +// Todo: Investigate a better way to inline this template - especially when escaping the backticks. +// Consider: https://mattjibson.com/blog/2014/11/19/esc-embedding-static-assets/ +const azureBatchPodTemplate = ` +#!/bin/bash +set -eE +trap cleanup EXIT + +if ! type 'docker' > /dev/null; then + echo 'Docker not installed... exiting' + exit 1 +fi + +{{/* Vars */}} +{{$podName := .PodName}} +{{$volumes := .Volumes}} + +{{/* Login to required image repositories */}} +{{range .PullCredentials }} +docker login -u {{.Username}} -p {{.Password}} {{.Server}} +{{end}} + +function cleanup(){ + {{/* Take a copy of the container log is removed when container is deleted */}} + echo 'Pod Exited: Copying logs' + {{range $index, $container := .InitContainers}} + if [[ -f ./initcontainer-{{$index}}.cid ]]; then + container_{{$index}}_ID=$(<./initcontainer-{{$index}}.cid) + container_{{$index}}_Log_Path=$(docker inspect --format='{{"{{.LogPath}}"}}' $container_{{$index}}_ID) + cp $container_{{$index}}_Log_Path ./{{$container.Name}}.log + + docker rm -f $container_{{$index}}_ID + rm -f ./initcontainer-{{$index}}.cid + fi + {{end}} + + {{range $index, $container := .Containers}} + if [[ -f ./{{$container.Name}}.log && -f ./container-{{$index}}.cid ]]; then + container_{{$index}}_ID=$(<./container-{{$index}}.cid) + container_{{$index}}_Log_Path=$(docker inspect --format='{{"{{.LogPath}}"}}' $container_{{$index}}_ID) + rm ./{{$container.Name}}.log {{/* Remove the existing symlink */}} + cp $container_{{$index}}_Log_Path ./{{$container.Name}}.log + fi + {{end}} + + {{/* Remove the containers, network and volumes */}} + + echo 'Pod Exited: Removing all containers' + if ls container-* 1> /dev/null 2>&1; then + for line in ` + "`ls container-*`" + ` + do + id=$(cat $line) + echo '-Logs container..' + docker logs $id + echo '-Removing container..' + docker rm -f $id + rm -f $line + done + fi + echo '-Removing pause container..' + docker rm -f {{$podName}} || echo 'Remove pause container failed' + rm -f ./pauseid.cid + echo '-Removing network container..' + docker network rm {{$podName}} || echo 'Remove network failed' + + echo '-Removing volumes..' + {{range .Volumes}} + docker volume rm -f {{$podName}}_{{.Name}} || echo 'Remove volume failed' + {{end}} +} + +{{/* Create Pod network and start it */}} +docker network create {{$podName}} +docker run -d --network {{$podName}} --name {{$podName}} --cidfile="./pauseid.cid" gcr.io/google_containers/pause:1.0 + +{{/* Handle volumes */}} +{{range .Volumes}} +{{if isHostPathVolume .}} +docker volume create --name {{$podName}}_{{.Name}} --opt type=none --opt device={{.VolumeSource.HostPath.Path}} --opt o=bind +{{end}} +{{if isEmptyDirVolume .}} +docker volume create {{$podName}}_{{.Name}} +{{end}} +{{end}} + +{{/* Run the init containers in the Pod. Attaching to shared namespace */}} +{{range $index, $container := .InitContainers}} +echo 'Running init container {{$index}}..' + {{if isPullAlways .}} +docker pull {{$container.Image}} + {{end}} +docker run --network container:{{$podName}} --ipc container:{{$podName}} \ + {{- if isNvidiaRuntime $container}} + --runtime nvidia \ + {{- end}} + {{- range $index, $envs := $container.Env}} +-e "{{$envs.Name}}:{{$envs.Value}}" \ + {{- end}} + {{- range $index, $mount := getValidVolumeMounts $container $volumes}} +-v {{$podName}}_{{$mount.Name}}:{{$mount.MountPath}} \ + {{- end}} +--cidfile=./initcontainer-{{$index}}.cid {{$container.Image}} {{getLaunchCommand $container}} +{{end}} + + +{{/* Run the containers in the Pod. Attaching to shared namespace */}} +{{range $index, $container := .Containers}} + {{if isPullAlways .}} +docker pull {{$container.Image}} + {{end}} +docker run -d --network container:{{$podName}} --ipc container:{{$podName}} \ + {{- if isNvidiaRuntime $container}} + --runtime nvidia \ + {{- end}} + {{- range $index, $envs := $container.Env}} +-e "{{$envs.Name}}:{{$envs.Value}}" \ + {{- end}} + {{- range $index, $mount := getValidVolumeMounts $container $volumes}} +-v {{$podName}}_{{$mount.Name}}:{{$mount.MountPath}} \ + {{- end}} +--cidfile=./container-{{$index}}.cid {{$container.Image}} {{getLaunchCommand $container}} +{{end}} + +{{/* Symlink all container logs files to task directory */}} +{{range $index, $container := .Containers}} +container_{{$index}}_ID=$(<./container-{{$index}}.cid) +container_{{$index}}_Log_Path=$(docker inspect --format='{{"{{.LogPath}}"}}' $container_{{$index}}_ID) +ln -f -s $container_{{$index}}_Log_Path ./{{$container.Name}}.log +{{end}} + +echo 'Running Pod: {{.PodName}}' + +{{/* Wait until any of these containers stop */}} +echo 'Waiting for any of the containers to exit' +for line in ` + "`ls container-*`" + ` +do + id=$(cat $line) + docker wait $id & +done + +while [ $(jobs -p | wc -l) == {{.Containers | len}} ] +do + sleep 2 +done + + +{{/* Get exit codes from containers */}} +echo 'Checking container exit codes' +overallExitCode=0 +for line in ` + "`ls container-*`" + ` +do + id=$(cat $line) + echo 'Getting exitcode' + exitCode=$(docker inspect -f {{"{{.State.ExitCode}}"}} $id) + + echo 'ID: ' $id ' ExitCode: ' $exitCode + echo 'Checking exitcode' + if (($exitCode != 0)) + then + echo 'Assigning exitcode' + overallExitCode=$exitCode + fi +done + +exit $overallExitCode +` diff --git a/vendor/github.com/satori/go.uuid/LICENSE b/vendor/github.com/satori/go.uuid/LICENSE new file mode 100644 index 000000000..926d54987 --- /dev/null +++ b/vendor/github.com/satori/go.uuid/LICENSE @@ -0,0 +1,20 @@ +Copyright (C) 2013-2018 by Maxim Bublis + +Permission is hereby granted, free of charge, to any person obtaining +a copy of this software and associated documentation files (the +"Software"), to deal in the Software without restriction, including +without limitation the rights to use, copy, modify, merge, publish, +distribute, sublicense, and/or sell copies of the Software, and to +permit persons to whom the Software is furnished to do so, subject to +the following conditions: + +The above copyright notice and this permission notice shall be +included in all copies or substantial portions of the Software. + +THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, +EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF +MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND +NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE +LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION +OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION +WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. diff --git a/vendor/github.com/satori/go.uuid/codec.go b/vendor/github.com/satori/go.uuid/codec.go new file mode 100644 index 000000000..656892c53 --- /dev/null +++ b/vendor/github.com/satori/go.uuid/codec.go @@ -0,0 +1,206 @@ +// Copyright (C) 2013-2018 by Maxim Bublis +// +// Permission is hereby granted, free of charge, to any person obtaining +// a copy of this software and associated documentation files (the +// "Software"), to deal in the Software without restriction, including +// without limitation the rights to use, copy, modify, merge, publish, +// distribute, sublicense, and/or sell copies of the Software, and to +// permit persons to whom the Software is furnished to do so, subject to +// the following conditions: +// +// The above copyright notice and this permission notice shall be +// included in all copies or substantial portions of the Software. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, +// EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF +// MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND +// NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE +// LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION +// OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION +// WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. + +package uuid + +import ( + "bytes" + "encoding/hex" + "fmt" +) + +// FromBytes returns UUID converted from raw byte slice input. +// It will return error if the slice isn't 16 bytes long. +func FromBytes(input []byte) (u UUID, err error) { + err = u.UnmarshalBinary(input) + return +} + +// FromBytesOrNil returns UUID converted from raw byte slice input. +// Same behavior as FromBytes, but returns a Nil UUID on error. +func FromBytesOrNil(input []byte) UUID { + uuid, err := FromBytes(input) + if err != nil { + return Nil + } + return uuid +} + +// FromString returns UUID parsed from string input. +// Input is expected in a form accepted by UnmarshalText. +func FromString(input string) (u UUID, err error) { + err = u.UnmarshalText([]byte(input)) + return +} + +// FromStringOrNil returns UUID parsed from string input. +// Same behavior as FromString, but returns a Nil UUID on error. +func FromStringOrNil(input string) UUID { + uuid, err := FromString(input) + if err != nil { + return Nil + } + return uuid +} + +// MarshalText implements the encoding.TextMarshaler interface. +// The encoding is the same as returned by String. +func (u UUID) MarshalText() (text []byte, err error) { + text = []byte(u.String()) + return +} + +// UnmarshalText implements the encoding.TextUnmarshaler interface. +// Following formats are supported: +// "6ba7b810-9dad-11d1-80b4-00c04fd430c8", +// "{6ba7b810-9dad-11d1-80b4-00c04fd430c8}", +// "urn:uuid:6ba7b810-9dad-11d1-80b4-00c04fd430c8" +// "6ba7b8109dad11d180b400c04fd430c8" +// ABNF for supported UUID text representation follows: +// uuid := canonical | hashlike | braced | urn +// plain := canonical | hashlike +// canonical := 4hexoct '-' 2hexoct '-' 2hexoct '-' 6hexoct +// hashlike := 12hexoct +// braced := '{' plain '}' +// urn := URN ':' UUID-NID ':' plain +// URN := 'urn' +// UUID-NID := 'uuid' +// 12hexoct := 6hexoct 6hexoct +// 6hexoct := 4hexoct 2hexoct +// 4hexoct := 2hexoct 2hexoct +// 2hexoct := hexoct hexoct +// hexoct := hexdig hexdig +// hexdig := '0' | '1' | '2' | '3' | '4' | '5' | '6' | '7' | '8' | '9' | +// 'a' | 'b' | 'c' | 'd' | 'e' | 'f' | +// 'A' | 'B' | 'C' | 'D' | 'E' | 'F' +func (u *UUID) UnmarshalText(text []byte) (err error) { + switch len(text) { + case 32: + return u.decodeHashLike(text) + case 36: + return u.decodeCanonical(text) + case 38: + return u.decodeBraced(text) + case 41: + fallthrough + case 45: + return u.decodeURN(text) + default: + return fmt.Errorf("uuid: incorrect UUID length: %s", text) + } +} + +// decodeCanonical decodes UUID string in format +// "6ba7b810-9dad-11d1-80b4-00c04fd430c8". +func (u *UUID) decodeCanonical(t []byte) (err error) { + if t[8] != '-' || t[13] != '-' || t[18] != '-' || t[23] != '-' { + return fmt.Errorf("uuid: incorrect UUID format %s", t) + } + + src := t[:] + dst := u[:] + + for i, byteGroup := range byteGroups { + if i > 0 { + src = src[1:] // skip dash + } + _, err = hex.Decode(dst[:byteGroup/2], src[:byteGroup]) + if err != nil { + return + } + src = src[byteGroup:] + dst = dst[byteGroup/2:] + } + + return +} + +// decodeHashLike decodes UUID string in format +// "6ba7b8109dad11d180b400c04fd430c8". +func (u *UUID) decodeHashLike(t []byte) (err error) { + src := t[:] + dst := u[:] + + if _, err = hex.Decode(dst, src); err != nil { + return err + } + return +} + +// decodeBraced decodes UUID string in format +// "{6ba7b810-9dad-11d1-80b4-00c04fd430c8}" or in format +// "{6ba7b8109dad11d180b400c04fd430c8}". +func (u *UUID) decodeBraced(t []byte) (err error) { + l := len(t) + + if t[0] != '{' || t[l-1] != '}' { + return fmt.Errorf("uuid: incorrect UUID format %s", t) + } + + return u.decodePlain(t[1 : l-1]) +} + +// decodeURN decodes UUID string in format +// "urn:uuid:6ba7b810-9dad-11d1-80b4-00c04fd430c8" or in format +// "urn:uuid:6ba7b8109dad11d180b400c04fd430c8". +func (u *UUID) decodeURN(t []byte) (err error) { + total := len(t) + + urn_uuid_prefix := t[:9] + + if !bytes.Equal(urn_uuid_prefix, urnPrefix) { + return fmt.Errorf("uuid: incorrect UUID format: %s", t) + } + + return u.decodePlain(t[9:total]) +} + +// decodePlain decodes UUID string in canonical format +// "6ba7b810-9dad-11d1-80b4-00c04fd430c8" or in hash-like format +// "6ba7b8109dad11d180b400c04fd430c8". +func (u *UUID) decodePlain(t []byte) (err error) { + switch len(t) { + case 32: + return u.decodeHashLike(t) + case 36: + return u.decodeCanonical(t) + default: + return fmt.Errorf("uuid: incorrrect UUID length: %s", t) + } +} + +// MarshalBinary implements the encoding.BinaryMarshaler interface. +func (u UUID) MarshalBinary() (data []byte, err error) { + data = u.Bytes() + return +} + +// UnmarshalBinary implements the encoding.BinaryUnmarshaler interface. +// It will return error if the slice isn't 16 bytes long. +func (u *UUID) UnmarshalBinary(data []byte) (err error) { + if len(data) != Size { + err = fmt.Errorf("uuid: UUID must be exactly 16 bytes long, got %d bytes", len(data)) + return + } + copy(u[:], data) + + return +} diff --git a/vendor/github.com/satori/go.uuid/generator.go b/vendor/github.com/satori/go.uuid/generator.go new file mode 100644 index 000000000..3f2f1da2d --- /dev/null +++ b/vendor/github.com/satori/go.uuid/generator.go @@ -0,0 +1,239 @@ +// Copyright (C) 2013-2018 by Maxim Bublis +// +// Permission is hereby granted, free of charge, to any person obtaining +// a copy of this software and associated documentation files (the +// "Software"), to deal in the Software without restriction, including +// without limitation the rights to use, copy, modify, merge, publish, +// distribute, sublicense, and/or sell copies of the Software, and to +// permit persons to whom the Software is furnished to do so, subject to +// the following conditions: +// +// The above copyright notice and this permission notice shall be +// included in all copies or substantial portions of the Software. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, +// EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF +// MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND +// NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE +// LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION +// OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION +// WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. + +package uuid + +import ( + "crypto/md5" + "crypto/rand" + "crypto/sha1" + "encoding/binary" + "hash" + "net" + "os" + "sync" + "time" +) + +// Difference in 100-nanosecond intervals between +// UUID epoch (October 15, 1582) and Unix epoch (January 1, 1970). +const epochStart = 122192928000000000 + +var ( + global = newDefaultGenerator() + + epochFunc = unixTimeFunc + posixUID = uint32(os.Getuid()) + posixGID = uint32(os.Getgid()) +) + +// NewV1 returns UUID based on current timestamp and MAC address. +func NewV1() UUID { + return global.NewV1() +} + +// NewV2 returns DCE Security UUID based on POSIX UID/GID. +func NewV2(domain byte) UUID { + return global.NewV2(domain) +} + +// NewV3 returns UUID based on MD5 hash of namespace UUID and name. +func NewV3(ns UUID, name string) UUID { + return global.NewV3(ns, name) +} + +// NewV4 returns random generated UUID. +func NewV4() UUID { + return global.NewV4() +} + +// NewV5 returns UUID based on SHA-1 hash of namespace UUID and name. +func NewV5(ns UUID, name string) UUID { + return global.NewV5(ns, name) +} + +// Generator provides interface for generating UUIDs. +type Generator interface { + NewV1() UUID + NewV2(domain byte) UUID + NewV3(ns UUID, name string) UUID + NewV4() UUID + NewV5(ns UUID, name string) UUID +} + +// Default generator implementation. +type generator struct { + storageOnce sync.Once + storageMutex sync.Mutex + + lastTime uint64 + clockSequence uint16 + hardwareAddr [6]byte +} + +func newDefaultGenerator() Generator { + return &generator{} +} + +// NewV1 returns UUID based on current timestamp and MAC address. +func (g *generator) NewV1() UUID { + u := UUID{} + + timeNow, clockSeq, hardwareAddr := g.getStorage() + + binary.BigEndian.PutUint32(u[0:], uint32(timeNow)) + binary.BigEndian.PutUint16(u[4:], uint16(timeNow>>32)) + binary.BigEndian.PutUint16(u[6:], uint16(timeNow>>48)) + binary.BigEndian.PutUint16(u[8:], clockSeq) + + copy(u[10:], hardwareAddr) + + u.SetVersion(V1) + u.SetVariant(VariantRFC4122) + + return u +} + +// NewV2 returns DCE Security UUID based on POSIX UID/GID. +func (g *generator) NewV2(domain byte) UUID { + u := UUID{} + + timeNow, clockSeq, hardwareAddr := g.getStorage() + + switch domain { + case DomainPerson: + binary.BigEndian.PutUint32(u[0:], posixUID) + case DomainGroup: + binary.BigEndian.PutUint32(u[0:], posixGID) + } + + binary.BigEndian.PutUint16(u[4:], uint16(timeNow>>32)) + binary.BigEndian.PutUint16(u[6:], uint16(timeNow>>48)) + binary.BigEndian.PutUint16(u[8:], clockSeq) + u[9] = domain + + copy(u[10:], hardwareAddr) + + u.SetVersion(V2) + u.SetVariant(VariantRFC4122) + + return u +} + +// NewV3 returns UUID based on MD5 hash of namespace UUID and name. +func (g *generator) NewV3(ns UUID, name string) UUID { + u := newFromHash(md5.New(), ns, name) + u.SetVersion(V3) + u.SetVariant(VariantRFC4122) + + return u +} + +// NewV4 returns random generated UUID. +func (g *generator) NewV4() UUID { + u := UUID{} + g.safeRandom(u[:]) + u.SetVersion(V4) + u.SetVariant(VariantRFC4122) + + return u +} + +// NewV5 returns UUID based on SHA-1 hash of namespace UUID and name. +func (g *generator) NewV5(ns UUID, name string) UUID { + u := newFromHash(sha1.New(), ns, name) + u.SetVersion(V5) + u.SetVariant(VariantRFC4122) + + return u +} + +func (g *generator) initStorage() { + g.initClockSequence() + g.initHardwareAddr() +} + +func (g *generator) initClockSequence() { + buf := make([]byte, 2) + g.safeRandom(buf) + g.clockSequence = binary.BigEndian.Uint16(buf) +} + +func (g *generator) initHardwareAddr() { + interfaces, err := net.Interfaces() + if err == nil { + for _, iface := range interfaces { + if len(iface.HardwareAddr) >= 6 { + copy(g.hardwareAddr[:], iface.HardwareAddr) + return + } + } + } + + // Initialize hardwareAddr randomly in case + // of real network interfaces absence + g.safeRandom(g.hardwareAddr[:]) + + // Set multicast bit as recommended in RFC 4122 + g.hardwareAddr[0] |= 0x01 +} + +func (g *generator) safeRandom(dest []byte) { + if _, err := rand.Read(dest); err != nil { + panic(err) + } +} + +// Returns UUID v1/v2 storage state. +// Returns epoch timestamp, clock sequence, and hardware address. +func (g *generator) getStorage() (uint64, uint16, []byte) { + g.storageOnce.Do(g.initStorage) + + g.storageMutex.Lock() + defer g.storageMutex.Unlock() + + timeNow := epochFunc() + // Clock changed backwards since last UUID generation. + // Should increase clock sequence. + if timeNow <= g.lastTime { + g.clockSequence++ + } + g.lastTime = timeNow + + return timeNow, g.clockSequence, g.hardwareAddr[:] +} + +// Returns difference in 100-nanosecond intervals between +// UUID epoch (October 15, 1582) and current time. +// This is default epoch calculation function. +func unixTimeFunc() uint64 { + return epochStart + uint64(time.Now().UnixNano()/100) +} + +// Returns UUID based on hashing of namespace UUID and name. +func newFromHash(h hash.Hash, ns UUID, name string) UUID { + u := UUID{} + h.Write(ns[:]) + h.Write([]byte(name)) + copy(u[:], h.Sum(nil)) + + return u +} diff --git a/vendor/github.com/satori/go.uuid/sql.go b/vendor/github.com/satori/go.uuid/sql.go new file mode 100644 index 000000000..56759d390 --- /dev/null +++ b/vendor/github.com/satori/go.uuid/sql.go @@ -0,0 +1,78 @@ +// Copyright (C) 2013-2018 by Maxim Bublis +// +// Permission is hereby granted, free of charge, to any person obtaining +// a copy of this software and associated documentation files (the +// "Software"), to deal in the Software without restriction, including +// without limitation the rights to use, copy, modify, merge, publish, +// distribute, sublicense, and/or sell copies of the Software, and to +// permit persons to whom the Software is furnished to do so, subject to +// the following conditions: +// +// The above copyright notice and this permission notice shall be +// included in all copies or substantial portions of the Software. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, +// EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF +// MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND +// NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE +// LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION +// OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION +// WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. + +package uuid + +import ( + "database/sql/driver" + "fmt" +) + +// Value implements the driver.Valuer interface. +func (u UUID) Value() (driver.Value, error) { + return u.String(), nil +} + +// Scan implements the sql.Scanner interface. +// A 16-byte slice is handled by UnmarshalBinary, while +// a longer byte slice or a string is handled by UnmarshalText. +func (u *UUID) Scan(src interface{}) error { + switch src := src.(type) { + case []byte: + if len(src) == Size { + return u.UnmarshalBinary(src) + } + return u.UnmarshalText(src) + + case string: + return u.UnmarshalText([]byte(src)) + } + + return fmt.Errorf("uuid: cannot convert %T to UUID", src) +} + +// NullUUID can be used with the standard sql package to represent a +// UUID value that can be NULL in the database +type NullUUID struct { + UUID UUID + Valid bool +} + +// Value implements the driver.Valuer interface. +func (u NullUUID) Value() (driver.Value, error) { + if !u.Valid { + return nil, nil + } + // Delegate to UUID Value function + return u.UUID.Value() +} + +// Scan implements the sql.Scanner interface. +func (u *NullUUID) Scan(src interface{}) error { + if src == nil { + u.UUID, u.Valid = Nil, false + return nil + } + + // Delegate to UUID Scan function + u.Valid = true + return u.UUID.Scan(src) +} diff --git a/vendor/github.com/satori/go.uuid/uuid.go b/vendor/github.com/satori/go.uuid/uuid.go new file mode 100644 index 000000000..a2b8e2ca2 --- /dev/null +++ b/vendor/github.com/satori/go.uuid/uuid.go @@ -0,0 +1,161 @@ +// Copyright (C) 2013-2018 by Maxim Bublis +// +// Permission is hereby granted, free of charge, to any person obtaining +// a copy of this software and associated documentation files (the +// "Software"), to deal in the Software without restriction, including +// without limitation the rights to use, copy, modify, merge, publish, +// distribute, sublicense, and/or sell copies of the Software, and to +// permit persons to whom the Software is furnished to do so, subject to +// the following conditions: +// +// The above copyright notice and this permission notice shall be +// included in all copies or substantial portions of the Software. +// +// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, +// EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF +// MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND +// NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE +// LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION +// OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION +// WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. + +// Package uuid provides implementation of Universally Unique Identifier (UUID). +// Supported versions are 1, 3, 4 and 5 (as specified in RFC 4122) and +// version 2 (as specified in DCE 1.1). +package uuid + +import ( + "bytes" + "encoding/hex" +) + +// Size of a UUID in bytes. +const Size = 16 + +// UUID representation compliant with specification +// described in RFC 4122. +type UUID [Size]byte + +// UUID versions +const ( + _ byte = iota + V1 + V2 + V3 + V4 + V5 +) + +// UUID layout variants. +const ( + VariantNCS byte = iota + VariantRFC4122 + VariantMicrosoft + VariantFuture +) + +// UUID DCE domains. +const ( + DomainPerson = iota + DomainGroup + DomainOrg +) + +// String parse helpers. +var ( + urnPrefix = []byte("urn:uuid:") + byteGroups = []int{8, 4, 4, 4, 12} +) + +// Nil is special form of UUID that is specified to have all +// 128 bits set to zero. +var Nil = UUID{} + +// Predefined namespace UUIDs. +var ( + NamespaceDNS = Must(FromString("6ba7b810-9dad-11d1-80b4-00c04fd430c8")) + NamespaceURL = Must(FromString("6ba7b811-9dad-11d1-80b4-00c04fd430c8")) + NamespaceOID = Must(FromString("6ba7b812-9dad-11d1-80b4-00c04fd430c8")) + NamespaceX500 = Must(FromString("6ba7b814-9dad-11d1-80b4-00c04fd430c8")) +) + +// Equal returns true if u1 and u2 equals, otherwise returns false. +func Equal(u1 UUID, u2 UUID) bool { + return bytes.Equal(u1[:], u2[:]) +} + +// Version returns algorithm version used to generate UUID. +func (u UUID) Version() byte { + return u[6] >> 4 +} + +// Variant returns UUID layout variant. +func (u UUID) Variant() byte { + switch { + case (u[8] >> 7) == 0x00: + return VariantNCS + case (u[8] >> 6) == 0x02: + return VariantRFC4122 + case (u[8] >> 5) == 0x06: + return VariantMicrosoft + case (u[8] >> 5) == 0x07: + fallthrough + default: + return VariantFuture + } +} + +// Bytes returns bytes slice representation of UUID. +func (u UUID) Bytes() []byte { + return u[:] +} + +// Returns canonical string representation of UUID: +// xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx. +func (u UUID) String() string { + buf := make([]byte, 36) + + hex.Encode(buf[0:8], u[0:4]) + buf[8] = '-' + hex.Encode(buf[9:13], u[4:6]) + buf[13] = '-' + hex.Encode(buf[14:18], u[6:8]) + buf[18] = '-' + hex.Encode(buf[19:23], u[8:10]) + buf[23] = '-' + hex.Encode(buf[24:], u[10:]) + + return string(buf) +} + +// SetVersion sets version bits. +func (u *UUID) SetVersion(v byte) { + u[6] = (u[6] & 0x0f) | (v << 4) +} + +// SetVariant sets variant bits. +func (u *UUID) SetVariant(v byte) { + switch v { + case VariantNCS: + u[8] = (u[8]&(0xff>>1) | (0x00 << 7)) + case VariantRFC4122: + u[8] = (u[8]&(0xff>>2) | (0x02 << 6)) + case VariantMicrosoft: + u[8] = (u[8]&(0xff>>3) | (0x06 << 5)) + case VariantFuture: + fallthrough + default: + u[8] = (u[8]&(0xff>>3) | (0x07 << 5)) + } +} + +// Must is a helper that wraps a call to a function returning (UUID, error) +// and panics if the error is non-nil. It is intended for use in variable +// initializations such as +// var packageUUID = uuid.Must(uuid.FromString("123e4567-e89b-12d3-a456-426655440000")); +func Must(u UUID, err error) UUID { + if err != nil { + panic(err) + } + return u +} diff --git a/vkubelet/provider.go b/vkubelet/provider.go index 58ecb6c4f..d63c76d8d 100644 --- a/vkubelet/provider.go +++ b/vkubelet/provider.go @@ -3,6 +3,7 @@ package vkubelet import ( "github.com/virtual-kubelet/virtual-kubelet/providers/aws" "github.com/virtual-kubelet/virtual-kubelet/providers/azure" + "github.com/virtual-kubelet/virtual-kubelet/providers/azurebatch" "github.com/virtual-kubelet/virtual-kubelet/providers/cri" "github.com/virtual-kubelet/virtual-kubelet/providers/hypersh" "github.com/virtual-kubelet/virtual-kubelet/providers/web" @@ -14,6 +15,7 @@ var _ Provider = (*aws.FargateProvider)(nil) var _ Provider = (*azure.ACIProvider)(nil) var _ Provider = (*hypersh.HyperProvider)(nil) var _ Provider = (*web.BrokerProvider)(nil) +var _ Provider = (*azurebatch.Provider)(nil) var _ Provider = (*cri.CRIProvider)(nil) // Provider contains the methods required to implement a virtual-kubelet provider. diff --git a/vkubelet/vkubelet.go b/vkubelet/vkubelet.go index 5cca600c6..3808ca469 100644 --- a/vkubelet/vkubelet.go +++ b/vkubelet/vkubelet.go @@ -13,6 +13,7 @@ import ( "github.com/virtual-kubelet/virtual-kubelet/manager" "github.com/virtual-kubelet/virtual-kubelet/providers/aws" "github.com/virtual-kubelet/virtual-kubelet/providers/azure" + "github.com/virtual-kubelet/virtual-kubelet/providers/azurebatch" "github.com/virtual-kubelet/virtual-kubelet/providers/cri" "github.com/virtual-kubelet/virtual-kubelet/providers/hypersh" "github.com/virtual-kubelet/virtual-kubelet/providers/mock" @@ -90,6 +91,11 @@ func New(nodeName, operatingSystem, namespace, kubeConfig, taint, provider, prov if err != nil { return nil, err } + case "azurebatch": + p, err = azurebatch.NewBatchProvider(providerConfig, rm, nodeName, operatingSystem, internalIP, daemonEndpointPort) + if err != nil { + return nil, err + } case "hyper": p, err = hypersh.NewHyperProvider(providerConfig, rm, nodeName, operatingSystem) if err != nil { @@ -274,6 +280,13 @@ func (s *Server) updateNode() { n.ResourceVersion = "" // Blank out resource version to prevent object has been modified error n.Status.Conditions = s.provider.NodeConditions() + + capacity := s.provider.Capacity() + n.Status.Capacity = capacity + n.Status.Allocatable = capacity + + n.Status.Addresses = s.provider.NodeAddresses() + n, err = s.k8sClient.CoreV1().Nodes().UpdateStatus(n) if err != nil { log.Println("Failed to update node:", err)