Refactor the way that the that node_ping_controller works

This moves node ping controller to using the new internal lock
API.

The reason for this is twofold:
* The channel approach that was used to notify other
  controllers of changes could only be used once (at startup),
  and couldn't be used in the future to broadcast node
  ping status. The idea idea is here that we could move
  to a sync.Cond style API and only wakeup other controllers
  on change, as opposed to constantly polling each other
* The problem with sync.Cond is that it's not context friendly.
  If we want to do stuff like wait on a sync.cond and use a context
  or a timer or similar, it doesn't work whereas this API allows
  context cancellations on condition change.

The idea is that as we have more controllers that act as centralized
sources of authority, they can broadcast out their state.
This commit is contained in:
Sargun Dhillon
2020-11-05 03:01:46 -08:00
parent d562b71d9a
commit 11c63bca6f
3 changed files with 227 additions and 22 deletions

98
internal/lock/monitor.go Normal file
View File

@@ -0,0 +1,98 @@
package lock
import (
"sync"
)
// NewMonitorVariable instantiates an empty monitor variable
func NewMonitorVariable() MonitorVariable {
mv := &monitorVariable{
versionInvalidationChannel: make(chan struct{}),
}
return mv
}
// MonitorVariable is a specific monitor variable which allows for channel-subscription to changes to
// the internal value of the MonitorVariable.
type MonitorVariable interface {
Set(value interface{})
Subscribe() Subscription
}
// Subscription is not concurrency safe. It must not be shared between multiple goroutines.
type Subscription interface {
// On instantiation, if the value has been set, this will return a closed channel. Otherwise, it will follow the
// standard semantic, which is when the Monitor Variable is updated, this channel will close. The channel is updated
// based on reading Value(). Once a value is read, the channel returned will only be closed if a the Monitor Variable
// is set to a new value.
NewValueReady() <-chan struct{}
// Value returns a value object in a non-blocking fashion. This also means it may return an uninitialized value.
// If the monitor variable has not yet been set, the "Version" of the value will be 0.
Value() Value
}
type Value struct {
Value interface{}
Version int64
}
type monitorVariable struct {
lock sync.Mutex
currentValue interface{}
// 0 indicates uninitialized
currentVersion int64
versionInvalidationChannel chan struct{}
}
func (m *monitorVariable) Set(newValue interface{}) {
m.lock.Lock()
defer m.lock.Unlock()
m.currentValue = newValue
m.currentVersion++
close(m.versionInvalidationChannel)
m.versionInvalidationChannel = make(chan struct{})
}
func (m *monitorVariable) Subscribe() Subscription {
m.lock.Lock()
defer m.lock.Unlock()
sub := &subscription{
mv: m,
}
if m.currentVersion > 0 {
// A value has been set. Set the first versionInvalidationChannel to a closed one.
closedCh := make(chan struct{})
close(closedCh)
sub.lastVersionReadInvalidationChannel = closedCh
} else {
// The value hasn't yet been initialized.
sub.lastVersionReadInvalidationChannel = m.versionInvalidationChannel
}
return sub
}
type subscription struct {
mv *monitorVariable
lastVersionRead int64
lastVersionReadInvalidationChannel chan struct{}
}
func (s *subscription) NewValueReady() <-chan struct{} {
/* This lock could be finer grained (on just the subscription) */
s.mv.lock.Lock()
defer s.mv.lock.Unlock()
return s.lastVersionReadInvalidationChannel
}
func (s *subscription) Value() Value {
s.mv.lock.Lock()
defer s.mv.lock.Unlock()
val := Value{
Value: s.mv.currentValue,
Version: s.mv.currentVersion,
}
s.lastVersionRead = s.mv.currentVersion
s.lastVersionReadInvalidationChannel = s.mv.versionInvalidationChannel
return val
}

View File

@@ -0,0 +1,113 @@
package lock
import (
"sync"
"testing"
"time"
"golang.org/x/sync/errgroup"
"k8s.io/apimachinery/pkg/util/sets"
"gotest.tools/assert"
is "gotest.tools/assert/cmp"
)
func TestMonitorUninitialized(t *testing.T) {
t.Parallel()
mv := NewMonitorVariable()
subscription := mv.Subscribe()
select {
case <-subscription.NewValueReady():
t.Fatalf("Received value update message: %v", subscription.Value())
case <-time.After(time.Second):
}
}
func TestGetUninitialized(t *testing.T) {
mv := NewMonitorVariable()
subscription := mv.Subscribe()
val := subscription.Value()
assert.Assert(t, is.Equal(val.Version, int64(0)))
}
func TestMonitorSetInitialVersionAfterListen(t *testing.T) {
mv := NewMonitorVariable()
subscription := mv.Subscribe()
go mv.Set("test")
<-subscription.NewValueReady()
assert.Assert(t, is.Equal(subscription.Value().Value, "test"))
}
func TestMonitorSetInitialVersionBeforeListen(t *testing.T) {
mv := NewMonitorVariable()
subscription := mv.Subscribe()
mv.Set("test")
<-subscription.NewValueReady()
assert.Assert(t, is.Equal(subscription.Value().Value, "test"))
}
func TestMonitorMultipleVersionsBlock(t *testing.T) {
t.Parallel()
mv := NewMonitorVariable()
subscription := mv.Subscribe()
mv.Set("test")
<-subscription.NewValueReady()
/* This should mark the "current" version as seen */
val := subscription.Value()
assert.Assert(t, is.Equal(val.Version, int64(1)))
select {
case <-subscription.NewValueReady():
t.Fatalf("Received value update message: %v", subscription.Value())
case <-time.After(time.Second):
}
}
func TestMonitorMultipleVersions(t *testing.T) {
t.Parallel()
lock := sync.Mutex{}
lock.Lock()
mv := NewMonitorVariable()
triggers := []int{}
ch := make(chan struct{}, 10)
go func() {
defer lock.Unlock()
subscription := mv.Subscribe()
for {
select {
case <-subscription.NewValueReady():
val := subscription.Value()
triggers = append(triggers, val.Value.(int))
ch <- struct{}{}
if val.Value == 9 {
return
}
}
}
}()
for i := 0; i < 10; i++ {
mv.Set(i)
// Wait for the trigger to occur
<-ch
}
// Wait for the goroutine to finish
lock.Lock()
t.Logf("Saw %v triggers", triggers)
assert.Assert(t, is.Len(triggers, 10))
// Make sure we saw all 10 unique values
assert.Assert(t, is.Equal(sets.NewInt(triggers...).Len(), 10))
}
func TestMonitorMultipleSubscribers(t *testing.T) {
group := &errgroup.Group{}
mv := NewMonitorVariable()
for i := 0; i < 10; i++ {
sub := mv.Subscribe()
group.Go(func() error {
<-sub.NewValueReady()
return nil
})
}
mv.Set(1)
_ = group.Wait()
}

View File

@@ -2,9 +2,9 @@ package node
import (
"context"
"sync"
"time"
"github.com/virtual-kubelet/virtual-kubelet/internal/lock"
"github.com/virtual-kubelet/virtual-kubelet/log"
"github.com/virtual-kubelet/virtual-kubelet/trace"
"golang.org/x/sync/singleflight"
@@ -12,14 +12,10 @@ import (
)
type nodePingController struct {
nodeProvider NodeProvider
pingInterval time.Duration
firstPingCompleted chan struct{}
pingTimeout *time.Duration
// "Results"
sync.Mutex
result *pingResult
nodeProvider NodeProvider
pingInterval time.Duration
pingTimeout *time.Duration
cond lock.MonitorVariable
}
type pingResult struct {
@@ -37,10 +33,10 @@ func newNodePingController(node NodeProvider, pingInterval time.Duration, timeou
}
return &nodePingController{
nodeProvider: node,
pingInterval: pingInterval,
firstPingCompleted: make(chan struct{}),
pingTimeout: timeout,
nodeProvider: node,
pingInterval: pingInterval,
pingTimeout: timeout,
cond: lock.NewMonitorVariable(),
}
}
@@ -87,28 +83,26 @@ func (npc *nodePingController) run(ctx context.Context) {
pingResult.pingTime = result.Val.(time.Time)
}
npc.Lock()
defer npc.Unlock()
npc.result = &pingResult
npc.cond.Set(&pingResult)
span.SetStatus(pingResult.error)
}
// Run the first check manually
checkFunc(ctx)
close(npc.firstPingCompleted)
wait.UntilWithContext(ctx, checkFunc, npc.pingInterval)
}
// getResult returns the current ping result in a non-blocking fashion except for the first ping. It waits for the
// first ping to be successful before returning. If the context is cancelled while waiting for that value, it will
// return immediately.
func (npc *nodePingController) getResult(ctx context.Context) (*pingResult, error) {
sub := npc.cond.Subscribe()
select {
case <-ctx.Done():
return nil, ctx.Err()
case <-npc.firstPingCompleted:
case <-sub.NewValueReady():
}
npc.Lock()
defer npc.Unlock()
return npc.result, nil
return sub.Value().Value.(*pingResult), nil
}