Files
virtual-kubelet/vendor/github.com/vmware/vic/lib/archive/stripper_test.go
Loc Nguyen 513cebe7b7 VMware vSphere Integrated Containers provider (#206)
* Add Virtual Kubelet provider for VIC

Initial virtual kubelet provider for VMware VIC.  This provider currently
handles creating and starting of a pod VM via the VIC portlayer and persona
server.  Image store handling via the VIC persona server.  This provider
currently requires the feature/wolfpack branch of VIC.

* Added pod stop and delete.  Also added node capacity.

Added the ability to stop and delete pod VMs via VIC.  Also retrieve
node capacity information from the VCH.

* Cleanup and readme file

Some file clean up and added a Readme.md markdown file for the VIC
provider.

* Cleaned up errors, added function comments, moved operation code

1. Cleaned up error handling.  Set standard for creating errors.
2. Added method prototype comments for all interface functions.
3. Moved PodCreator, PodStarter, PodStopper, and PodDeleter to a new folder.

* Add mocking code and unit tests for podcache, podcreator, and podstarter

Used the unit test framework used in VIC to handle assertions in the provider's
unit test.  Mocking code generated using OSS project mockery, which is compatible
with the testify assertion framework.

* Vendored packages for the VIC provider

Requires feature/wolfpack branch of VIC and a few specific commit sha of
projects used within VIC.

* Implementation of POD Stopper and Deleter unit tests (#4)

* Updated files for initial PR
2018-06-04 15:41:32 -07:00

321 lines
8.3 KiB
Go

// Copyright 2017 VMware, Inc. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package archive
import (
"archive/tar"
"context"
"fmt"
"io"
"sync"
"testing"
"github.com/stretchr/testify/assert"
"github.com/vmware/vic/pkg/trace"
"github.com/vmware/vic/pkg/uid"
)
// generateArchive takes a number of files and a file size and generates a tar archive
// based on that data. It returns:
// index of entry names in the archive
// archive byte stream
func generateArchive(name string, num, size int) ([]string, io.Reader) {
r, w := io.Pipe()
tw := tar.NewWriter(w)
// stable reference for expected archive entries
index := make([]string, num)
for i := 0; i < num; i++ {
index[i] = string(uid.New())
}
// our file contents are zeros - this is the worst case for stripping trailing headers
zeros := make([]byte, size)
go func() {
for i := 0; i < num; i++ {
// we only really care about two things in the header
hdr := &tar.Header{
Name: index[i],
Size: int64(size),
}
fmt.Printf("Writing header for file %s:%d\n", name, i)
err := tw.WriteHeader(hdr)
if err != nil {
panic(err)
}
fmt.Printf("Writing data for file %s:%d\n", name, i)
n, err := tw.Write(zeros)
if err != nil {
panic(err)
}
if n != size {
panic(fmt.Sprintf("Failed to write all bytes: %d != %d", n, size))
}
}
// add the end-of-archive
tw.Close()
w.Close()
}()
return index, r
}
// TestSingleStripper ensures that basic function (stripping end-of-archive footer) works as
// expected. I found no real way, when using the TarReader to actually assert that the footer
// has been dropped so this is left to assert correct passthrough of archive data and the
// dropping of the footer is asserted by the multistream cases.
func TestSingleStripper(t *testing.T) {
size := 2048
count := 5
index, reader := generateArchive("single", count, size)
source := tar.NewReader(reader)
stripper := NewStripper(trace.NewOperation(context.Background(), "strip"), source, nil)
pr, pw := io.Pipe()
tr := tar.NewReader(pr)
var wg sync.WaitGroup
wg.Add(1)
go func() {
n, err := io.Copy(pw, stripper)
pw.Close()
wg.Done()
fmt.Printf("Done copying from stripper: %d, %s\n", n, err)
if !assert.NoError(t, err, "Expected nil error from io.Copy on end-of-file") {
t.FailNow()
}
}()
for i := 0; i <= len(index); i++ {
fmt.Printf("Reading header for file %d\n", i)
header, err := tr.Next()
if err == io.EOF {
fmt.Printf("End-of-file")
// TODO: is this pass or fail?
return
}
if err != nil {
fmt.Printf("Error from archive: %s\n", err)
t.FailNow()
}
if !assert.NotEqual(t, i, len(index), "Expected EOF after index exhausted") {
t.FailNow()
}
if !assert.Equal(t, header.Name, index[i], "Expected header name to match index") {
t.FailNow()
}
if !assert.Equal(t, header.Size, int64(size), "Expected file size in header to match target generated size") {
t.FailNow()
}
// make the buf just that little bit bigger to allow for errrors in the copy if they would occur
buf := make([]byte, size+10)
fmt.Printf("Reading data for file %d\n", i)
n, err := tr.Read(buf)
if !assert.NoError(t, err, "No expected errors from file data copy") {
t.FailNow()
}
if !assert.Equal(t, n, size, "Expected file data size to match target generated size") {
t.FailNow()
}
}
wg.Wait()
}
// TestConjoinedStrippers ensures that the footer is correctly dropped from a stripped archive
// and that a TarReader continues to provide headers from the following conjoined streams.
func TestConjoinedStrippers(t *testing.T) {
size := 2048
count := 3
multiplicity := 3
indices := make([][]string, multiplicity)
strippers := make([]io.Reader, multiplicity)
for m := 0; m < multiplicity; m++ {
var reader io.Reader
indices[m], reader = generateArchive(fmt.Sprintf("archive-%d", m), count, size)
source := tar.NewReader(reader)
strippers[m] = NewStripper(trace.NewOperation(context.Background(), fmt.Sprintf("strip-%d", m)), source, nil)
}
conjoined := MultiReader(strippers...)
pr, pw := io.Pipe()
tr := tar.NewReader(pr)
var wg sync.WaitGroup
wg.Add(1)
go func() {
n, err := io.Copy(pw, conjoined)
pw.Close()
wg.Done()
fmt.Printf("Done copying from strippers: %d, %s\n", n, err)
if !assert.NoError(t, err, "Expected nil error from io.Copy on end-of-file") {
t.FailNow()
}
}()
expectedEntries := count * multiplicity
for i := 0; i <= expectedEntries; i++ {
fmt.Printf("Reading header for file %d\n", i)
header, err := tr.Next()
if err == io.EOF {
fmt.Printf("End-of-file\n")
// TODO: is this pass or fail?
return
}
if err != nil {
fmt.Printf("Error from archive: %s\n", err)
t.FailNow()
}
if !assert.NotEqual(t, i, expectedEntries, "Expected EOF after index exhausted") {
t.FailNow()
}
member := i / count
entry := i % count
if !assert.Equal(t, header.Name, indices[member][entry], "Expected header name to match index") {
t.FailNow()
}
if !assert.Equal(t, header.Size, int64(size), "Expected file size in header to match target generated size") {
t.FailNow()
}
// make the buf just that little bit bigger to allow for errrors in the copy if they would occur
buf := make([]byte, size+10)
fmt.Printf("Reading data for file %d\n", i)
n, err := tr.Read(buf)
if !assert.NoError(t, err, "No expected errors from file data copy") {
t.FailNow()
}
if !assert.Equal(t, n, size, "Expected file data size to match target generated size") {
t.FailNow()
}
}
wg.Wait()
}
// TestConjoinedStrippersWithCloser ensures that we can conjoin readers, multiple strippers and a regular, in order to get
// the end-of-archive footer as the finale. We have a wait group to ensure that all routines have finished before declaring
// success.
func TestConjoinedStrippersWithCloser(t *testing.T) {
size := 2048
count := 3
multiplicity := 3
indices := make([][]string, multiplicity)
readers := make([]io.Reader, multiplicity)
for m := 0; m < multiplicity; m++ {
var reader io.Reader
indices[m], reader = generateArchive(fmt.Sprintf("archive-%d", m), count, size)
source := tar.NewReader(reader)
if m < multiplicity-1 {
fmt.Printf("added stripper\n")
readers[m] = NewStripper(trace.NewOperation(context.Background(), fmt.Sprintf("strip-%d", m)), source, nil)
} else {
fmt.Printf("added raw\n")
readers[m] = reader
}
}
conjoined := MultiReader(readers...)
pr, pw := io.Pipe()
tr := tar.NewReader(pr)
var wg sync.WaitGroup
wg.Add(1)
go func() {
n, err := io.Copy(pw, conjoined)
pw.Close()
wg.Done()
fmt.Printf("Done copying from all sources: %d, %s\n", n, err)
if !assert.NoError(t, err, "Expected nil error from io.Copy on end-of-file") {
t.FailNow()
}
}()
expectedEntries := count * multiplicity
for i := 0; i <= expectedEntries; i++ {
fmt.Printf("Reading header for file %d\n", i)
header, err := tr.Next()
if err == io.EOF {
fmt.Printf("End-of-file\n")
wg.Wait()
return
}
if err != nil {
fmt.Printf("Error from archive: %s\n", err)
t.FailNow()
}
if !assert.NotEqual(t, i, expectedEntries, "Expected EOF after index exhausted") {
t.FailNow()
}
member := i / count
entry := i % count
if !assert.Equal(t, header.Name, indices[member][entry], "Expected header name to match index") {
t.FailNow()
}
if !assert.Equal(t, header.Size, int64(size), "Expected file size in header to match target generated size") {
t.FailNow()
}
// make the buf just that little bit bigger to allow for errrors in the copy if they would occur
buf := make([]byte, size+10)
fmt.Printf("Reading data for file %d\n", i)
n, err := tr.Read(buf)
if !assert.NoError(t, err, "No expected errors from file data copy") {
t.FailNow()
}
if !assert.Equal(t, n, size, "Expected file data size to match target generated size") {
t.FailNow()
}
}
wg.Wait()
}