* Add Virtual Kubelet provider for VIC Initial virtual kubelet provider for VMware VIC. This provider currently handles creating and starting of a pod VM via the VIC portlayer and persona server. Image store handling via the VIC persona server. This provider currently requires the feature/wolfpack branch of VIC. * Added pod stop and delete. Also added node capacity. Added the ability to stop and delete pod VMs via VIC. Also retrieve node capacity information from the VCH. * Cleanup and readme file Some file clean up and added a Readme.md markdown file for the VIC provider. * Cleaned up errors, added function comments, moved operation code 1. Cleaned up error handling. Set standard for creating errors. 2. Added method prototype comments for all interface functions. 3. Moved PodCreator, PodStarter, PodStopper, and PodDeleter to a new folder. * Add mocking code and unit tests for podcache, podcreator, and podstarter Used the unit test framework used in VIC to handle assertions in the provider's unit test. Mocking code generated using OSS project mockery, which is compatible with the testify assertion framework. * Vendored packages for the VIC provider Requires feature/wolfpack branch of VIC and a few specific commit sha of projects used within VIC. * Implementation of POD Stopper and Deleter unit tests (#4) * Updated files for initial PR
321 lines
9.3 KiB
Go
321 lines
9.3 KiB
Go
// Copyright 2016-2017 VMware, Inc. All Rights Reserved.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package main
|
|
|
|
import (
|
|
"crypto/tls"
|
|
"crypto/x509"
|
|
"flag"
|
|
"fmt"
|
|
"os"
|
|
|
|
log "github.com/Sirupsen/logrus"
|
|
apiserver "github.com/docker/docker/api/server"
|
|
"github.com/docker/docker/api/server/middleware"
|
|
"github.com/docker/docker/api/server/router"
|
|
"github.com/docker/docker/api/server/router/checkpoint"
|
|
"github.com/docker/docker/api/server/router/container"
|
|
"github.com/docker/docker/api/server/router/image"
|
|
"github.com/docker/docker/api/server/router/network"
|
|
"github.com/docker/docker/api/server/router/plugin"
|
|
"github.com/docker/docker/api/server/router/swarm"
|
|
"github.com/docker/docker/api/server/router/system"
|
|
"github.com/docker/docker/api/server/router/volume"
|
|
"github.com/docker/docker/daemon/cluster"
|
|
"github.com/docker/docker/pkg/listeners"
|
|
"github.com/docker/docker/pkg/signal"
|
|
"github.com/docker/docker/runconfig"
|
|
"github.com/docker/go-connections/tlsconfig"
|
|
|
|
vicbackends "github.com/vmware/vic/lib/apiservers/engine/backends"
|
|
"github.com/vmware/vic/lib/apiservers/engine/backends/executor"
|
|
"github.com/vmware/vic/lib/config"
|
|
"github.com/vmware/vic/lib/constants"
|
|
"github.com/vmware/vic/lib/portlayer/util"
|
|
"github.com/vmware/vic/lib/pprof"
|
|
viclog "github.com/vmware/vic/pkg/log"
|
|
"github.com/vmware/vic/pkg/log/syslog"
|
|
"github.com/vmware/vic/pkg/trace"
|
|
"github.com/vmware/vic/pkg/version"
|
|
"github.com/vmware/vic/pkg/vsphere/extraconfig"
|
|
"github.com/vmware/vic/pkg/vsphere/tags"
|
|
)
|
|
|
|
type CliOptions struct {
|
|
serverPort *uint
|
|
portLayerAddr *string
|
|
portLayerPort *uint
|
|
debug *bool
|
|
|
|
proto string
|
|
}
|
|
|
|
const (
|
|
productName = "vSphere Integrated Containers"
|
|
)
|
|
|
|
var (
|
|
vchConfig config.VirtualContainerHostConfigSpec
|
|
cli CliOptions
|
|
)
|
|
|
|
func init() {
|
|
// #nosec: Errors unhandled.
|
|
pprof.StartPprof("docker personality", pprof.DockerPort)
|
|
|
|
flag.Usage = Usage
|
|
|
|
_ = flag.String("serveraddr", "127.0.0.1", "Server address to listen") // ignored
|
|
cli.serverPort = flag.Uint("port", 9000, "Port to listen")
|
|
cli.portLayerAddr = flag.String("port-layer-addr", "127.0.0.1", "Port layer server address")
|
|
cli.portLayerPort = flag.Uint("port-layer-port", 9001, "Port Layer server port")
|
|
|
|
cli.debug = flag.Bool("debug", false, "Enable debuglevel logging")
|
|
}
|
|
|
|
func Usage() {
|
|
// #nosec: Errors unhandled.
|
|
fmt.Fprintf(os.Stderr, "\nvSphere Integrated Container Daemon Usage:\n")
|
|
flag.PrintDefaults()
|
|
// #nosec: Errors unhandled.
|
|
fmt.Fprintf(os.Stderr, "\n")
|
|
}
|
|
|
|
func main() {
|
|
// Get flags
|
|
ok := handleFlags()
|
|
|
|
if !ok {
|
|
os.Exit(1)
|
|
}
|
|
|
|
if err := initLogging(); err != nil {
|
|
log.Fatalf("failed to initialize logging: %s", err)
|
|
}
|
|
|
|
if err := vicbackends.Init(*cli.portLayerAddr, productName, *cli.serverPort, &vchConfig); err != nil {
|
|
log.Fatalf("failed to initialize backend: %s", err)
|
|
}
|
|
|
|
plEventMonitor := vicbackends.NewPortlayerEventMonitor(vicbackends.PlEventProxy{}, vicbackends.DockerEventPublisher{})
|
|
// #nosec: Errors unhandled.
|
|
plEventMonitor.Start()
|
|
// Start API server wit options from command line args
|
|
api := startServer()
|
|
|
|
setAPIRoutes(api)
|
|
|
|
serveAPIWait := make(chan error)
|
|
go api.Wait(serveAPIWait)
|
|
|
|
signal.Trap(func() {
|
|
api.Close()
|
|
})
|
|
|
|
<-serveAPIWait
|
|
plEventMonitor.Stop()
|
|
}
|
|
|
|
func handleFlags() bool {
|
|
flag.Parse()
|
|
|
|
// load the vch config
|
|
src, err := extraconfig.GuestInfoSource()
|
|
if err != nil {
|
|
log.Fatalf("Unable to load configuration from guestinfo: %s", err)
|
|
}
|
|
extraconfig.Decode(src, &vchConfig)
|
|
|
|
if vchConfig.Diagnostics.DebugLevel > 2 {
|
|
// use debug port layer mode
|
|
*cli.portLayerPort = constants.DebugPortLayerPort
|
|
clientIP, err := util.ClientIP()
|
|
if err != nil {
|
|
log.Fatalf("Unable to look up %s for portlayer API: %s", constants.ClientHostName, err)
|
|
}
|
|
*cli.portLayerAddr = clientIP.String()
|
|
}
|
|
*cli.portLayerAddr = fmt.Sprintf("%s:%d", *cli.portLayerAddr, *cli.portLayerPort)
|
|
cli.proto = "tcp"
|
|
|
|
return true
|
|
}
|
|
|
|
func initLogging() error {
|
|
logcfg := viclog.NewLoggingConfig()
|
|
if *cli.debug || vchConfig.Diagnostics.DebugLevel > 0 {
|
|
logcfg.Level = log.DebugLevel
|
|
trace.Logger.Level = log.DebugLevel
|
|
syslog.Logger.Level = log.DebugLevel
|
|
}
|
|
|
|
if vchConfig.Diagnostics.SysLogConfig != nil {
|
|
logcfg.Syslog = &viclog.SyslogConfig{
|
|
Network: vchConfig.Diagnostics.SysLogConfig.Network,
|
|
RAddr: vchConfig.Diagnostics.SysLogConfig.RAddr,
|
|
Priority: syslog.Info | syslog.Daemon,
|
|
}
|
|
}
|
|
|
|
tags.Logger = log.StandardLogger()
|
|
|
|
err := viclog.Init(logcfg)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return trace.InitLogger(logcfg)
|
|
}
|
|
|
|
func loadCAPool() *x509.CertPool {
|
|
// If we should verify the server, we need to load a trusted ca
|
|
pool := x509.NewCertPool()
|
|
|
|
pem := vchConfig.CertificateAuthorities
|
|
if len(pem) == 0 {
|
|
return nil
|
|
}
|
|
|
|
if !pool.AppendCertsFromPEM(vchConfig.CertificateAuthorities) {
|
|
log.Fatalf("Unable to load CAs in config")
|
|
}
|
|
|
|
log.Debugf("Loaded %d CAs from config", len(pool.Subjects()))
|
|
return pool
|
|
}
|
|
|
|
func startServer() *apiserver.Server {
|
|
serverConfig := &apiserver.Config{
|
|
Logging: true,
|
|
Version: "1.22", //dockerversion.Version,
|
|
}
|
|
|
|
// FIXME: assignment copies lock value to tlsConfig: crypto/tls.Config contains sync.Once contains sync.Mutex
|
|
// #nosec: TLS InsecureSkipVerify may be true
|
|
tlsConfig := func(c *tls.Config) *tls.Config {
|
|
return &tls.Config{
|
|
Certificates: c.Certificates,
|
|
NameToCertificate: c.NameToCertificate,
|
|
GetCertificate: c.GetCertificate,
|
|
RootCAs: c.RootCAs,
|
|
NextProtos: c.NextProtos,
|
|
ServerName: c.ServerName,
|
|
ClientAuth: c.ClientAuth,
|
|
ClientCAs: c.ClientCAs,
|
|
InsecureSkipVerify: c.InsecureSkipVerify,
|
|
CipherSuites: c.CipherSuites,
|
|
PreferServerCipherSuites: c.PreferServerCipherSuites,
|
|
SessionTicketsDisabled: c.SessionTicketsDisabled,
|
|
SessionTicketKey: c.SessionTicketKey,
|
|
ClientSessionCache: c.ClientSessionCache,
|
|
MinVersion: tls.VersionTLS12,
|
|
MaxVersion: c.MaxVersion,
|
|
CurvePreferences: c.CurvePreferences,
|
|
}
|
|
}(tlsconfig.ServerDefault())
|
|
|
|
if !vchConfig.HostCertificate.IsNil() {
|
|
log.Info("TLS enabled")
|
|
|
|
cert, err := vchConfig.HostCertificate.Certificate()
|
|
if err != nil {
|
|
// This is only viable because we've verified those certificates
|
|
log.Fatalf("Could not load certificate from config and refusing to run without TLS with a host certificate specified: %s", err)
|
|
}
|
|
|
|
tlsConfig.Certificates = []tls.Certificate{*cert}
|
|
serverConfig.TLSConfig = tlsConfig
|
|
|
|
// Set options for TLS
|
|
if len(vchConfig.CertificateAuthorities) > 0 {
|
|
log.Info("Client verification enabled")
|
|
// server requires and verifies client's certificate
|
|
tlsConfig.ClientAuth = tls.RequireAndVerifyClientCert
|
|
tlsConfig.ClientCAs = loadCAPool()
|
|
tlsConfig.InsecureSkipVerify = false
|
|
}
|
|
}
|
|
|
|
addr := "0.0.0.0"
|
|
// exposing this on all interfaces
|
|
if vchConfig.Diagnostics.DebugLevel <= 2 {
|
|
|
|
// determine the address to listen on
|
|
clientIP, err := util.ClientIP()
|
|
if err != nil {
|
|
// TODO: don't want to directly enter this into vchConfig.Sessions[].Started but no
|
|
// structure currently to report back contents otherwise
|
|
log.Fatalf("Unable to look up %s to serve docker API: %s", constants.ClientHostName, err)
|
|
}
|
|
|
|
addr = clientIP.String()
|
|
}
|
|
|
|
api := apiserver.New(serverConfig)
|
|
mw := middleware.NewVersionMiddleware(version.DockerAPIVersion,
|
|
version.DockerDefaultVersion,
|
|
version.DockerMinimumVersion)
|
|
api.UseMiddleware(mw)
|
|
fullserver := fmt.Sprintf("%s:%d", addr, *cli.serverPort)
|
|
l, err := listeners.Init(cli.proto, fullserver, "", serverConfig.TLSConfig)
|
|
if err != nil {
|
|
log.Fatal(err)
|
|
}
|
|
|
|
log.Printf("Listener created for HTTP on %s//%s", addr, cli.proto)
|
|
api.Accept(fullserver, l...)
|
|
|
|
return api
|
|
}
|
|
|
|
func setAPIRoutes(api *apiserver.Server) {
|
|
decoder := runconfig.ContainerDecoder{}
|
|
|
|
swarmBackend := executor.SwarmBackend{}
|
|
|
|
c, err := cluster.New(cluster.Config{
|
|
Root: "",
|
|
Name: "",
|
|
Backend: swarmBackend,
|
|
NetworkSubnetsProvider: nil,
|
|
DefaultAdvertiseAddr: "",
|
|
RuntimeRoot: "",
|
|
})
|
|
if err != nil {
|
|
log.Fatalf("Error creating cluster component: %v", err)
|
|
}
|
|
|
|
routers := []router.Router{
|
|
image.NewRouter(vicbackends.NewImageBackend(), decoder),
|
|
container.NewRouter(vicbackends.NewContainerBackend(), decoder),
|
|
volume.NewRouter(vicbackends.NewVolumeBackend()),
|
|
network.NewRouter(vicbackends.NewNetworkBackend(), c),
|
|
system.NewRouter(vicbackends.NewSystemBackend(), c),
|
|
swarm.NewRouter(vicbackends.NewSwarmBackend()),
|
|
checkpoint.NewRouter(vicbackends.NewCheckpointBackend(), decoder),
|
|
plugin.NewRouter(vicbackends.NewPluginBackend()),
|
|
}
|
|
|
|
for _, r := range routers {
|
|
for _, route := range r.Routes() {
|
|
if experimental, ok := route.(router.ExperimentalRoute); ok {
|
|
experimental.Enable()
|
|
}
|
|
}
|
|
}
|
|
|
|
api.InitRouter(false, routers...)
|
|
}
|