Mark current linting errors as non-errors

This is basically claiming linting bankruptcy. It marks all of the
issues we had up until this point as nolint.
This commit is contained in:
Sargun Dhillon
2019-09-03 10:43:37 -07:00
parent 5949e6279d
commit 7133a372d6
6 changed files with 12 additions and 11 deletions

View File

@@ -69,7 +69,7 @@ func installFlags(flags *pflag.FlagSet, c *Opts) {
flags.StringVar(&c.TaintKey, "taint", c.TaintKey, "Set node taint key") flags.StringVar(&c.TaintKey, "taint", c.TaintKey, "Set node taint key")
flags.BoolVar(&c.DisableTaint, "disable-taint", c.DisableTaint, "disable the virtual-kubelet node taint") flags.BoolVar(&c.DisableTaint, "disable-taint", c.DisableTaint, "disable the virtual-kubelet node taint")
flags.MarkDeprecated("taint", "Taint key should now be configured using the VK_TAINT_KEY environment variable") flags.MarkDeprecated("taint", "Taint key should now be configured using the VK_TAINT_KEY environment variable") //nolint:errcheck
flags.IntVar(&c.PodSyncWorkers, "pod-sync-workers", c.PodSyncWorkers, `set the number of pod synchronization workers`) flags.IntVar(&c.PodSyncWorkers, "pod-sync-workers", c.PodSyncWorkers, `set the number of pod synchronization workers`)
flags.BoolVar(&c.EnableNodeLease, "enable-node-lease", c.EnableNodeLease, `use node leases (1.13) for node heartbeats`) flags.BoolVar(&c.EnableNodeLease, "enable-node-lease", c.EnableNodeLease, `use node leases (1.13) for node heartbeats`)

View File

@@ -42,7 +42,7 @@ var (
*/ */
// MockV0Provider implements the virtual-kubelet provider interface and stores pods in memory. // MockV0Provider implements the virtual-kubelet provider interface and stores pods in memory.
type MockV0Provider struct { type MockV0Provider struct { //nolint:golint
nodeName string nodeName string
operatingSystem string operatingSystem string
internalIP string internalIP string
@@ -54,12 +54,12 @@ type MockV0Provider struct {
} }
// MockProvider is like MockV0Provider, but implements the PodNotifier interface // MockProvider is like MockV0Provider, but implements the PodNotifier interface
type MockProvider struct { type MockProvider struct { //nolint:golint
*MockV0Provider *MockV0Provider
} }
// MockConfig contains a mock virtual-kubelet's configurable parameters. // MockConfig contains a mock virtual-kubelet's configurable parameters.
type MockConfig struct { type MockConfig struct { //nolint:golint
CPU string `json:"cpu,omitempty"` CPU string `json:"cpu,omitempty"`
Memory string `json:"memory,omitempty"` Memory string `json:"memory,omitempty"`
Pods string `json:"pods,omitempty"` Pods string `json:"pods,omitempty"`
@@ -355,7 +355,7 @@ func (p *MockV0Provider) GetPods(ctx context.Context) ([]*v1.Pod, error) {
} }
func (p *MockV0Provider) ConfigureNode(ctx context.Context, n *v1.Node) { func (p *MockV0Provider) ConfigureNode(ctx context.Context, n *v1.Node) {
ctx, span := trace.StartSpan(ctx, "mock.ConfigureNode") ctx, span := trace.StartSpan(ctx, "mock.ConfigureNode") //nolint:ineffassign
defer span.End() defer span.End()
n.Status.Capacity = p.capacity() n.Status.Capacity = p.capacity()
@@ -453,7 +453,8 @@ func (p *MockV0Provider) nodeDaemonEndpoints() v1.NodeDaemonEndpoints {
// GetStatsSummary returns dummy stats for all pods known by this provider. // GetStatsSummary returns dummy stats for all pods known by this provider.
func (p *MockV0Provider) GetStatsSummary(ctx context.Context) (*stats.Summary, error) { func (p *MockV0Provider) GetStatsSummary(ctx context.Context) (*stats.Summary, error) {
ctx, span := trace.StartSpan(ctx, "GetStatsSummary") var span trace.Span
ctx, span = trace.StartSpan(ctx, "GetStatsSummary") //nolint: ineffassign
defer span.End() defer span.End()
// Grab the current timestamp so we can report it as the time the stats were generated. // Grab the current timestamp so we can report it as the time the stats were generated.

View File

@@ -6,7 +6,7 @@ import (
) )
func registerMock(s *provider.Store) { func registerMock(s *provider.Store) {
s.Register("mock", func(cfg provider.InitConfig) (provider.Provider, error) { s.Register("mock", func(cfg provider.InitConfig) (provider.Provider, error) { //nolint:errcheck
return mock.NewMockProvider( return mock.NewMockProvider(
cfg.ConfigPath, cfg.ConfigPath,
cfg.NodeName, cfg.NodeName,
@@ -16,7 +16,7 @@ func registerMock(s *provider.Store) {
) )
}) })
s.Register("mockV0", func(cfg provider.InitConfig) (provider.Provider, error) { s.Register("mockV0", func(cfg provider.InitConfig) (provider.Provider, error) { //nolint:errcheck
return mock.NewMockProvider( return mock.NewMockProvider(
cfg.ConfigPath, cfg.ConfigPath,
cfg.NodeName, cfg.NodeName,

View File

@@ -33,7 +33,7 @@ func handleError(f handlerFunc) http.HandlerFunc {
code := httpStatusCode(err) code := httpStatusCode(err)
w.WriteHeader(code) w.WriteHeader(code)
io.WriteString(w, err.Error()) io.WriteString(w, err.Error()) //nolint:errcheck
logger := log.G(req.Context()).WithError(err).WithField("httpStatusCode", code) logger := log.G(req.Context()).WithError(err).WithField("httpStatusCode", code)
if code >= 500 { if code >= 500 {

View File

@@ -28,7 +28,7 @@ type PodListerFunc func(context.Context) ([]*v1.Pod, error)
func HandleRunningPods(getPods PodListerFunc) http.HandlerFunc { func HandleRunningPods(getPods PodListerFunc) http.HandlerFunc {
scheme := runtime.NewScheme() scheme := runtime.NewScheme()
v1.SchemeBuilder.AddToScheme(scheme) v1.SchemeBuilder.AddToScheme(scheme) //nolint:errcheck
codecs := serializer.NewCodecFactory(scheme) codecs := serializer.NewCodecFactory(scheme)
return handleError(func(w http.ResponseWriter, req *http.Request) error { return handleError(func(w http.ResponseWriter, req *http.Request) error {

View File

@@ -38,7 +38,7 @@ import (
// //
// Note: Implementers can choose to manage a node themselves, in which case // Note: Implementers can choose to manage a node themselves, in which case
// it is not needed to provide an implementation for this interface. // it is not needed to provide an implementation for this interface.
type NodeProvider interface { type NodeProvider interface { //nolint:golint
// Ping checks if the node is still active. // Ping checks if the node is still active.
// This is intended to be lightweight as it will be called periodically as a // This is intended to be lightweight as it will be called periodically as a
// heartbeat to keep the node marked as ready in Kubernetes. // heartbeat to keep the node marked as ready in Kubernetes.