* Add Virtual Kubelet provider for VIC Initial virtual kubelet provider for VMware VIC. This provider currently handles creating and starting of a pod VM via the VIC portlayer and persona server. Image store handling via the VIC persona server. This provider currently requires the feature/wolfpack branch of VIC. * Added pod stop and delete. Also added node capacity. Added the ability to stop and delete pod VMs via VIC. Also retrieve node capacity information from the VCH. * Cleanup and readme file Some file clean up and added a Readme.md markdown file for the VIC provider. * Cleaned up errors, added function comments, moved operation code 1. Cleaned up error handling. Set standard for creating errors. 2. Added method prototype comments for all interface functions. 3. Moved PodCreator, PodStarter, PodStopper, and PodDeleter to a new folder. * Add mocking code and unit tests for podcache, podcreator, and podstarter Used the unit test framework used in VIC to handle assertions in the provider's unit test. Mocking code generated using OSS project mockery, which is compatible with the testify assertion framework. * Vendored packages for the VIC provider Requires feature/wolfpack branch of VIC and a few specific commit sha of projects used within VIC. * Implementation of POD Stopper and Deleter unit tests (#4) * Updated files for initial PR
321 lines
8.3 KiB
Go
321 lines
8.3 KiB
Go
// Copyright 2017 VMware, Inc. All Rights Reserved.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package archive
|
|
|
|
import (
|
|
"archive/tar"
|
|
"context"
|
|
"fmt"
|
|
"io"
|
|
"sync"
|
|
"testing"
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
|
|
"github.com/vmware/vic/pkg/trace"
|
|
"github.com/vmware/vic/pkg/uid"
|
|
)
|
|
|
|
// generateArchive takes a number of files and a file size and generates a tar archive
|
|
// based on that data. It returns:
|
|
// index of entry names in the archive
|
|
// archive byte stream
|
|
func generateArchive(name string, num, size int) ([]string, io.Reader) {
|
|
r, w := io.Pipe()
|
|
tw := tar.NewWriter(w)
|
|
|
|
// stable reference for expected archive entries
|
|
index := make([]string, num)
|
|
for i := 0; i < num; i++ {
|
|
index[i] = string(uid.New())
|
|
}
|
|
|
|
// our file contents are zeros - this is the worst case for stripping trailing headers
|
|
zeros := make([]byte, size)
|
|
|
|
go func() {
|
|
for i := 0; i < num; i++ {
|
|
// we only really care about two things in the header
|
|
hdr := &tar.Header{
|
|
Name: index[i],
|
|
Size: int64(size),
|
|
}
|
|
|
|
fmt.Printf("Writing header for file %s:%d\n", name, i)
|
|
err := tw.WriteHeader(hdr)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
fmt.Printf("Writing data for file %s:%d\n", name, i)
|
|
n, err := tw.Write(zeros)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
if n != size {
|
|
panic(fmt.Sprintf("Failed to write all bytes: %d != %d", n, size))
|
|
}
|
|
}
|
|
|
|
// add the end-of-archive
|
|
tw.Close()
|
|
w.Close()
|
|
}()
|
|
|
|
return index, r
|
|
}
|
|
|
|
// TestSingleStripper ensures that basic function (stripping end-of-archive footer) works as
|
|
// expected. I found no real way, when using the TarReader to actually assert that the footer
|
|
// has been dropped so this is left to assert correct passthrough of archive data and the
|
|
// dropping of the footer is asserted by the multistream cases.
|
|
func TestSingleStripper(t *testing.T) {
|
|
size := 2048
|
|
count := 5
|
|
index, reader := generateArchive("single", count, size)
|
|
|
|
source := tar.NewReader(reader)
|
|
stripper := NewStripper(trace.NewOperation(context.Background(), "strip"), source, nil)
|
|
|
|
pr, pw := io.Pipe()
|
|
tr := tar.NewReader(pr)
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(1)
|
|
go func() {
|
|
n, err := io.Copy(pw, stripper)
|
|
pw.Close()
|
|
|
|
wg.Done()
|
|
fmt.Printf("Done copying from stripper: %d, %s\n", n, err)
|
|
if !assert.NoError(t, err, "Expected nil error from io.Copy on end-of-file") {
|
|
t.FailNow()
|
|
}
|
|
|
|
}()
|
|
|
|
for i := 0; i <= len(index); i++ {
|
|
fmt.Printf("Reading header for file %d\n", i)
|
|
header, err := tr.Next()
|
|
if err == io.EOF {
|
|
fmt.Printf("End-of-file")
|
|
// TODO: is this pass or fail?
|
|
return
|
|
}
|
|
|
|
if err != nil {
|
|
fmt.Printf("Error from archive: %s\n", err)
|
|
t.FailNow()
|
|
}
|
|
|
|
if !assert.NotEqual(t, i, len(index), "Expected EOF after index exhausted") {
|
|
t.FailNow()
|
|
}
|
|
|
|
if !assert.Equal(t, header.Name, index[i], "Expected header name to match index") {
|
|
t.FailNow()
|
|
}
|
|
if !assert.Equal(t, header.Size, int64(size), "Expected file size in header to match target generated size") {
|
|
t.FailNow()
|
|
}
|
|
|
|
// make the buf just that little bit bigger to allow for errrors in the copy if they would occur
|
|
buf := make([]byte, size+10)
|
|
|
|
fmt.Printf("Reading data for file %d\n", i)
|
|
n, err := tr.Read(buf)
|
|
|
|
if !assert.NoError(t, err, "No expected errors from file data copy") {
|
|
t.FailNow()
|
|
}
|
|
if !assert.Equal(t, n, size, "Expected file data size to match target generated size") {
|
|
t.FailNow()
|
|
}
|
|
}
|
|
|
|
wg.Wait()
|
|
}
|
|
|
|
// TestConjoinedStrippers ensures that the footer is correctly dropped from a stripped archive
|
|
// and that a TarReader continues to provide headers from the following conjoined streams.
|
|
func TestConjoinedStrippers(t *testing.T) {
|
|
size := 2048
|
|
count := 3
|
|
multiplicity := 3
|
|
|
|
indices := make([][]string, multiplicity)
|
|
strippers := make([]io.Reader, multiplicity)
|
|
|
|
for m := 0; m < multiplicity; m++ {
|
|
var reader io.Reader
|
|
indices[m], reader = generateArchive(fmt.Sprintf("archive-%d", m), count, size)
|
|
source := tar.NewReader(reader)
|
|
strippers[m] = NewStripper(trace.NewOperation(context.Background(), fmt.Sprintf("strip-%d", m)), source, nil)
|
|
}
|
|
|
|
conjoined := MultiReader(strippers...)
|
|
|
|
pr, pw := io.Pipe()
|
|
tr := tar.NewReader(pr)
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(1)
|
|
go func() {
|
|
n, err := io.Copy(pw, conjoined)
|
|
pw.Close()
|
|
|
|
wg.Done()
|
|
fmt.Printf("Done copying from strippers: %d, %s\n", n, err)
|
|
if !assert.NoError(t, err, "Expected nil error from io.Copy on end-of-file") {
|
|
t.FailNow()
|
|
}
|
|
}()
|
|
|
|
expectedEntries := count * multiplicity
|
|
for i := 0; i <= expectedEntries; i++ {
|
|
fmt.Printf("Reading header for file %d\n", i)
|
|
header, err := tr.Next()
|
|
if err == io.EOF {
|
|
fmt.Printf("End-of-file\n")
|
|
// TODO: is this pass or fail?
|
|
return
|
|
}
|
|
|
|
if err != nil {
|
|
fmt.Printf("Error from archive: %s\n", err)
|
|
t.FailNow()
|
|
}
|
|
|
|
if !assert.NotEqual(t, i, expectedEntries, "Expected EOF after index exhausted") {
|
|
t.FailNow()
|
|
}
|
|
|
|
member := i / count
|
|
entry := i % count
|
|
|
|
if !assert.Equal(t, header.Name, indices[member][entry], "Expected header name to match index") {
|
|
t.FailNow()
|
|
}
|
|
if !assert.Equal(t, header.Size, int64(size), "Expected file size in header to match target generated size") {
|
|
t.FailNow()
|
|
}
|
|
|
|
// make the buf just that little bit bigger to allow for errrors in the copy if they would occur
|
|
buf := make([]byte, size+10)
|
|
|
|
fmt.Printf("Reading data for file %d\n", i)
|
|
n, err := tr.Read(buf)
|
|
|
|
if !assert.NoError(t, err, "No expected errors from file data copy") {
|
|
t.FailNow()
|
|
}
|
|
if !assert.Equal(t, n, size, "Expected file data size to match target generated size") {
|
|
t.FailNow()
|
|
}
|
|
}
|
|
|
|
wg.Wait()
|
|
}
|
|
|
|
// TestConjoinedStrippersWithCloser ensures that we can conjoin readers, multiple strippers and a regular, in order to get
|
|
// the end-of-archive footer as the finale. We have a wait group to ensure that all routines have finished before declaring
|
|
// success.
|
|
func TestConjoinedStrippersWithCloser(t *testing.T) {
|
|
size := 2048
|
|
count := 3
|
|
multiplicity := 3
|
|
|
|
indices := make([][]string, multiplicity)
|
|
readers := make([]io.Reader, multiplicity)
|
|
|
|
for m := 0; m < multiplicity; m++ {
|
|
var reader io.Reader
|
|
indices[m], reader = generateArchive(fmt.Sprintf("archive-%d", m), count, size)
|
|
source := tar.NewReader(reader)
|
|
|
|
if m < multiplicity-1 {
|
|
fmt.Printf("added stripper\n")
|
|
readers[m] = NewStripper(trace.NewOperation(context.Background(), fmt.Sprintf("strip-%d", m)), source, nil)
|
|
} else {
|
|
fmt.Printf("added raw\n")
|
|
readers[m] = reader
|
|
}
|
|
}
|
|
|
|
conjoined := MultiReader(readers...)
|
|
|
|
pr, pw := io.Pipe()
|
|
tr := tar.NewReader(pr)
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(1)
|
|
go func() {
|
|
n, err := io.Copy(pw, conjoined)
|
|
pw.Close()
|
|
|
|
wg.Done()
|
|
fmt.Printf("Done copying from all sources: %d, %s\n", n, err)
|
|
if !assert.NoError(t, err, "Expected nil error from io.Copy on end-of-file") {
|
|
t.FailNow()
|
|
}
|
|
}()
|
|
|
|
expectedEntries := count * multiplicity
|
|
for i := 0; i <= expectedEntries; i++ {
|
|
fmt.Printf("Reading header for file %d\n", i)
|
|
header, err := tr.Next()
|
|
if err == io.EOF {
|
|
fmt.Printf("End-of-file\n")
|
|
|
|
wg.Wait()
|
|
return
|
|
}
|
|
|
|
if err != nil {
|
|
fmt.Printf("Error from archive: %s\n", err)
|
|
t.FailNow()
|
|
}
|
|
|
|
if !assert.NotEqual(t, i, expectedEntries, "Expected EOF after index exhausted") {
|
|
t.FailNow()
|
|
}
|
|
|
|
member := i / count
|
|
entry := i % count
|
|
|
|
if !assert.Equal(t, header.Name, indices[member][entry], "Expected header name to match index") {
|
|
t.FailNow()
|
|
}
|
|
if !assert.Equal(t, header.Size, int64(size), "Expected file size in header to match target generated size") {
|
|
t.FailNow()
|
|
}
|
|
|
|
// make the buf just that little bit bigger to allow for errrors in the copy if they would occur
|
|
buf := make([]byte, size+10)
|
|
|
|
fmt.Printf("Reading data for file %d\n", i)
|
|
n, err := tr.Read(buf)
|
|
|
|
if !assert.NoError(t, err, "No expected errors from file data copy") {
|
|
t.FailNow()
|
|
}
|
|
if !assert.Equal(t, n, size, "Expected file data size to match target generated size") {
|
|
t.FailNow()
|
|
}
|
|
}
|
|
|
|
wg.Wait()
|
|
}
|