mirror of
https://github.com/gogrlx/nats-server.git
synced 2026-04-02 03:38:42 -07:00
3284 lines
92 KiB
Go
3284 lines
92 KiB
Go
// Copyright 2018-2020 The NATS Authors
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package server
|
|
|
|
import (
|
|
"bytes"
|
|
"encoding/hex"
|
|
"errors"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"math"
|
|
"math/rand"
|
|
"net/http"
|
|
"net/textproto"
|
|
"net/url"
|
|
"reflect"
|
|
"sort"
|
|
"strconv"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/nats-io/jwt/v2"
|
|
"github.com/nats-io/nuid"
|
|
)
|
|
|
|
// For backwards compatibility with NATS < 2.0, users who are not explicitly defined into an
|
|
// account will be grouped in the default global account.
|
|
const globalAccountName = DEFAULT_GLOBAL_ACCOUNT
|
|
|
|
// Account are subject namespace definitions. By default no messages are shared between accounts.
|
|
// You can share via Exports and Imports of Streams and Services.
|
|
type Account struct {
|
|
Name string
|
|
Nkey string
|
|
Issuer string
|
|
claimJWT string
|
|
updated time.Time
|
|
mu sync.RWMutex
|
|
sqmu sync.Mutex
|
|
sl *Sublist
|
|
ic *client
|
|
isid uint64
|
|
etmr *time.Timer
|
|
ctmr *time.Timer
|
|
strack map[string]sconns
|
|
nrclients int32
|
|
sysclients int32
|
|
nleafs int32
|
|
nrleafs int32
|
|
clients map[*client]struct{}
|
|
rm map[string]int32
|
|
lqws map[string]int32
|
|
usersRevoked map[string]int64
|
|
actsRevoked map[string]int64
|
|
lleafs []*client
|
|
imports importMap
|
|
exports exportMap
|
|
js *jsAccount
|
|
jsLimits *JetStreamAccountLimits
|
|
limits
|
|
expired bool
|
|
incomplete bool
|
|
signingKeys []string
|
|
srv *Server // server this account is registered with (possibly nil)
|
|
lds string // loop detection subject for leaf nodes
|
|
siReply []byte // service reply prefix, will form wildcard subscription.
|
|
prand *rand.Rand
|
|
eventIds *nuid.NUID
|
|
eventIdsMu sync.Mutex
|
|
defaultPerms *Permissions
|
|
}
|
|
|
|
// Account based limits.
|
|
type limits struct {
|
|
mpay int32
|
|
msubs int32
|
|
mconns int32
|
|
mleafs int32
|
|
}
|
|
|
|
// Used to track remote clients and leafnodes per remote server.
|
|
type sconns struct {
|
|
conns int32
|
|
leafs int32
|
|
}
|
|
|
|
// Import stream mapping struct
|
|
type streamImport struct {
|
|
acc *Account
|
|
from string
|
|
prefix string
|
|
claim *jwt.Import
|
|
invalid bool
|
|
}
|
|
|
|
// Import service mapping struct
|
|
type serviceImport struct {
|
|
acc *Account
|
|
claim *jwt.Import
|
|
se *serviceExport
|
|
sid []byte
|
|
from string
|
|
to string
|
|
exsub string
|
|
ts int64
|
|
rt ServiceRespType
|
|
latency *serviceLatency
|
|
m1 *ServiceLatency
|
|
rc *client
|
|
hasWC bool
|
|
response bool
|
|
invalid bool
|
|
share bool
|
|
tracking bool
|
|
trackingHdr http.Header // header from request
|
|
}
|
|
|
|
// This is used to record when we create a mapping for implicit service
|
|
// imports. We use this to clean up entries that are not singletons when
|
|
// we detect that interest is no longer present. The key to the map will
|
|
// be the actual interest. We record the mapped subject and the account.
|
|
type serviceRespEntry struct {
|
|
acc *Account
|
|
msub string
|
|
}
|
|
|
|
// ServiceRespType represents the types of service request response types.
|
|
type ServiceRespType uint8
|
|
|
|
// Service response types. Defaults to a singleton.
|
|
const (
|
|
Singleton ServiceRespType = iota
|
|
Streamed
|
|
Chunked
|
|
)
|
|
|
|
// String helper.
|
|
func (rt ServiceRespType) String() string {
|
|
switch rt {
|
|
case Singleton:
|
|
return "Singleton"
|
|
case Streamed:
|
|
return "Streamed"
|
|
case Chunked:
|
|
return "Chunked"
|
|
}
|
|
return "Unknown ServiceResType"
|
|
}
|
|
|
|
// exportAuth holds configured approvals or boolean indicating an
|
|
// auth token is required for import.
|
|
type exportAuth struct {
|
|
tokenReq bool
|
|
approved map[string]*Account
|
|
}
|
|
|
|
// streamExport
|
|
type streamExport struct {
|
|
exportAuth
|
|
}
|
|
|
|
// serviceExport holds additional information for exported services.
|
|
type serviceExport struct {
|
|
exportAuth
|
|
acc *Account
|
|
respType ServiceRespType
|
|
latency *serviceLatency
|
|
rtmr *time.Timer
|
|
respThresh time.Duration
|
|
}
|
|
|
|
// Used to track service latency.
|
|
type serviceLatency struct {
|
|
sampling int8 // percentage from 1-100 or 0 to indicate triggered by header
|
|
subject string
|
|
}
|
|
|
|
// exportMap tracks the exported streams and services.
|
|
type exportMap struct {
|
|
streams map[string]*streamExport
|
|
services map[string]*serviceExport
|
|
responses map[string]*serviceImport
|
|
}
|
|
|
|
// importMap tracks the imported streams and services.
|
|
// For services we will also track the response mappings as well.
|
|
type importMap struct {
|
|
streams []*streamImport
|
|
services map[string]*serviceImport
|
|
rrMap map[string][]*serviceRespEntry
|
|
}
|
|
|
|
// NewAccount creates a new unlimited account with the given name.
|
|
func NewAccount(name string) *Account {
|
|
a := &Account{
|
|
Name: name,
|
|
limits: limits{-1, -1, -1, -1},
|
|
eventIds: nuid.New(),
|
|
}
|
|
|
|
return a
|
|
}
|
|
|
|
// Used to create shallow copies of accounts for transfer
|
|
// from opts to real accounts in server struct.
|
|
func (a *Account) shallowCopy() *Account {
|
|
na := NewAccount(a.Name)
|
|
na.Nkey = a.Nkey
|
|
na.Issuer = a.Issuer
|
|
|
|
if a.imports.streams != nil {
|
|
na.imports.streams = make([]*streamImport, 0, len(a.imports.streams))
|
|
for _, v := range a.imports.streams {
|
|
si := *v
|
|
na.imports.streams = append(na.imports.streams, &si)
|
|
}
|
|
}
|
|
if a.imports.services != nil {
|
|
na.imports.services = make(map[string]*serviceImport)
|
|
for k, v := range a.imports.services {
|
|
si := *v
|
|
na.imports.services[k] = &si
|
|
}
|
|
}
|
|
if a.exports.streams != nil {
|
|
na.exports.streams = make(map[string]*streamExport)
|
|
for k, v := range a.exports.streams {
|
|
if v != nil {
|
|
se := *v
|
|
na.exports.streams[k] = &se
|
|
} else {
|
|
na.exports.streams[k] = nil
|
|
}
|
|
}
|
|
}
|
|
if a.exports.services != nil {
|
|
na.exports.services = make(map[string]*serviceExport)
|
|
for k, v := range a.exports.services {
|
|
if v != nil {
|
|
se := *v
|
|
na.exports.services[k] = &se
|
|
} else {
|
|
na.exports.services[k] = nil
|
|
}
|
|
}
|
|
}
|
|
na.jsLimits = a.jsLimits
|
|
|
|
return na
|
|
}
|
|
|
|
// nextEventID uses its own lock for better concurrency.
|
|
func (a *Account) nextEventID() string {
|
|
a.eventIdsMu.Lock()
|
|
id := a.eventIds.Next()
|
|
a.eventIdsMu.Unlock()
|
|
return id
|
|
}
|
|
|
|
// Called to track a remote server and connections and leafnodes it
|
|
// has for this account.
|
|
func (a *Account) updateRemoteServer(m *AccountNumConns) []*client {
|
|
a.mu.Lock()
|
|
if a.strack == nil {
|
|
a.strack = make(map[string]sconns)
|
|
}
|
|
// This does not depend on receiving all updates since each one is idempotent.
|
|
// FIXME(dlc) - We should cleanup when these both go to zero.
|
|
prev := a.strack[m.Server.ID]
|
|
a.strack[m.Server.ID] = sconns{conns: int32(m.Conns), leafs: int32(m.LeafNodes)}
|
|
a.nrclients += int32(m.Conns) - prev.conns
|
|
a.nrleafs += int32(m.LeafNodes) - prev.leafs
|
|
|
|
mtce := a.mconns != jwt.NoLimit && (len(a.clients)-int(a.sysclients)+int(a.nrclients) > int(a.mconns))
|
|
// If we are over here some have snuck in and we need to rebalance.
|
|
// All others will probably be doing the same thing but better to be
|
|
// conservative and bit harsh here. Clients will reconnect if we over compensate.
|
|
var clients []*client
|
|
if mtce {
|
|
clients = make([]*client, 0, len(a.clients))
|
|
for c := range a.clients {
|
|
clients = append(clients, c)
|
|
}
|
|
sort.Slice(clients, func(i, j int) bool {
|
|
return clients[i].start.After(clients[j].start)
|
|
})
|
|
over := (len(a.clients) - int(a.sysclients) + int(a.nrclients)) - int(a.mconns)
|
|
if over < len(clients) {
|
|
clients = clients[:over]
|
|
}
|
|
}
|
|
// Now check leafnodes.
|
|
mtlce := a.mleafs != jwt.NoLimit && (a.nleafs+a.nrleafs > a.mleafs)
|
|
if mtlce {
|
|
// Take ones from the end.
|
|
leafs := a.lleafs
|
|
over := int(a.nleafs + a.nrleafs - a.mleafs)
|
|
if over < len(leafs) {
|
|
leafs = leafs[len(leafs)-over:]
|
|
}
|
|
clients = append(clients, leafs...)
|
|
}
|
|
a.mu.Unlock()
|
|
|
|
// If we have exceeded our max clients this will be populated.
|
|
return clients
|
|
}
|
|
|
|
// Removes tracking for a remote server that has shutdown.
|
|
func (a *Account) removeRemoteServer(sid string) {
|
|
a.mu.Lock()
|
|
if a.strack != nil {
|
|
prev := a.strack[sid]
|
|
delete(a.strack, sid)
|
|
a.nrclients -= prev.conns
|
|
a.nrleafs -= prev.leafs
|
|
}
|
|
a.mu.Unlock()
|
|
}
|
|
|
|
// When querying for subject interest this is the number of
|
|
// expected responses. We need to actually check that the entry
|
|
// has active connections.
|
|
func (a *Account) expectedRemoteResponses() (expected int32) {
|
|
a.mu.RLock()
|
|
for _, sc := range a.strack {
|
|
if sc.conns > 0 || sc.leafs > 0 {
|
|
expected++
|
|
}
|
|
}
|
|
a.mu.RUnlock()
|
|
return
|
|
}
|
|
|
|
// Clears eventing and tracking for this account.
|
|
func (a *Account) clearEventing() {
|
|
a.mu.Lock()
|
|
a.nrclients = 0
|
|
// Now clear state
|
|
clearTimer(&a.etmr)
|
|
clearTimer(&a.ctmr)
|
|
a.clients = nil
|
|
a.strack = nil
|
|
a.mu.Unlock()
|
|
}
|
|
|
|
// GetName will return the accounts name.
|
|
func (a *Account) GetName() string {
|
|
if a == nil {
|
|
return "n/a"
|
|
}
|
|
a.mu.RLock()
|
|
name := a.Name
|
|
a.mu.RUnlock()
|
|
return name
|
|
}
|
|
|
|
// NumConnections returns active number of clients for this account for
|
|
// all known servers.
|
|
func (a *Account) NumConnections() int {
|
|
a.mu.RLock()
|
|
nc := len(a.clients) + int(a.nrclients)
|
|
a.mu.RUnlock()
|
|
return nc
|
|
}
|
|
|
|
// NumRemoteConnections returns the number of client or leaf connections that
|
|
// are not on this server.
|
|
func (a *Account) NumRemoteConnections() int {
|
|
a.mu.RLock()
|
|
nc := int(a.nrclients + a.nrleafs)
|
|
a.mu.RUnlock()
|
|
return nc
|
|
}
|
|
|
|
// NumLocalConnections returns active number of clients for this account
|
|
// on this server.
|
|
func (a *Account) NumLocalConnections() int {
|
|
a.mu.RLock()
|
|
nlc := a.numLocalConnections()
|
|
a.mu.RUnlock()
|
|
return nlc
|
|
}
|
|
|
|
// Do not account for the system accounts.
|
|
func (a *Account) numLocalConnections() int {
|
|
return len(a.clients) - int(a.sysclients) - int(a.nleafs)
|
|
}
|
|
|
|
// This is for extended local interest.
|
|
// Lock should not be held.
|
|
func (a *Account) numLocalAndLeafConnections() int {
|
|
a.mu.RLock()
|
|
nlc := len(a.clients) - int(a.sysclients)
|
|
a.mu.RUnlock()
|
|
return nlc
|
|
}
|
|
|
|
func (a *Account) numLocalLeafNodes() int {
|
|
return int(a.nleafs)
|
|
}
|
|
|
|
// MaxTotalConnectionsReached returns if we have reached our limit for number of connections.
|
|
func (a *Account) MaxTotalConnectionsReached() bool {
|
|
var mtce bool
|
|
a.mu.RLock()
|
|
if a.mconns != jwt.NoLimit {
|
|
mtce = len(a.clients)-int(a.sysclients)+int(a.nrclients) >= int(a.mconns)
|
|
}
|
|
a.mu.RUnlock()
|
|
return mtce
|
|
}
|
|
|
|
// MaxActiveConnections return the set limit for the account system
|
|
// wide for total number of active connections.
|
|
func (a *Account) MaxActiveConnections() int {
|
|
a.mu.RLock()
|
|
mconns := int(a.mconns)
|
|
a.mu.RUnlock()
|
|
return mconns
|
|
}
|
|
|
|
// MaxTotalLeafNodesReached returns if we have reached our limit for number of leafnodes.
|
|
func (a *Account) MaxTotalLeafNodesReached() bool {
|
|
a.mu.RLock()
|
|
mtc := a.maxTotalLeafNodesReached()
|
|
a.mu.RUnlock()
|
|
return mtc
|
|
}
|
|
|
|
func (a *Account) maxTotalLeafNodesReached() bool {
|
|
if a.mleafs != jwt.NoLimit {
|
|
return a.nleafs+a.nrleafs >= a.mleafs
|
|
}
|
|
return false
|
|
}
|
|
|
|
// NumLeafNodes returns the active number of local and remote
|
|
// leaf node connections.
|
|
func (a *Account) NumLeafNodes() int {
|
|
a.mu.RLock()
|
|
nln := int(a.nleafs + a.nrleafs)
|
|
a.mu.RUnlock()
|
|
return nln
|
|
}
|
|
|
|
// NumRemoteLeafNodes returns the active number of remote
|
|
// leaf node connections.
|
|
func (a *Account) NumRemoteLeafNodes() int {
|
|
a.mu.RLock()
|
|
nrn := int(a.nrleafs)
|
|
a.mu.RUnlock()
|
|
return nrn
|
|
}
|
|
|
|
// MaxActiveLeafNodes return the set limit for the account system
|
|
// wide for total number of leavenode connections.
|
|
// NOTE: these are tracked separately.
|
|
func (a *Account) MaxActiveLeafNodes() int {
|
|
a.mu.RLock()
|
|
mleafs := int(a.mleafs)
|
|
a.mu.RUnlock()
|
|
return mleafs
|
|
}
|
|
|
|
// RoutedSubs returns how many subjects we would send across a route when first
|
|
// connected or expressing interest. Local client subs.
|
|
func (a *Account) RoutedSubs() int {
|
|
a.mu.RLock()
|
|
defer a.mu.RUnlock()
|
|
return len(a.rm)
|
|
}
|
|
|
|
// TotalSubs returns total number of Subscriptions for this account.
|
|
func (a *Account) TotalSubs() int {
|
|
a.mu.RLock()
|
|
defer a.mu.RUnlock()
|
|
return int(a.sl.Count())
|
|
}
|
|
|
|
// SubscriptionInterest returns true if this account has a matching subscription
|
|
// for the given `subject`. Works only for literal subjects.
|
|
// TODO: Add support for wildcards
|
|
func (a *Account) SubscriptionInterest(subject string) bool {
|
|
return a.Interest(subject) > 0
|
|
}
|
|
|
|
// Interest returns the number of subscriptions for a given subject that match.
|
|
func (a *Account) Interest(subject string) int {
|
|
var nms int
|
|
a.mu.RLock()
|
|
if a.sl != nil {
|
|
res := a.sl.Match(subject)
|
|
nms = len(res.psubs) + len(res.qsubs)
|
|
}
|
|
a.mu.RUnlock()
|
|
return nms
|
|
}
|
|
|
|
// addClient keeps our accounting of local active clients or leafnodes updated.
|
|
// Returns previous total.
|
|
func (a *Account) addClient(c *client) int {
|
|
a.mu.Lock()
|
|
n := len(a.clients)
|
|
if a.clients != nil {
|
|
a.clients[c] = struct{}{}
|
|
}
|
|
added := n != len(a.clients)
|
|
if added {
|
|
if c.kind == SYSTEM {
|
|
a.sysclients++
|
|
} else if c.kind == LEAF {
|
|
a.nleafs++
|
|
a.lleafs = append(a.lleafs, c)
|
|
}
|
|
}
|
|
a.mu.Unlock()
|
|
|
|
if c != nil && c.srv != nil && added {
|
|
c.srv.accConnsUpdate(a)
|
|
}
|
|
|
|
return n
|
|
}
|
|
|
|
// Helper function to remove leaf nodes. If number of leafnodes gets large
|
|
// this may need to be optimized out of linear search but believe number
|
|
// of active leafnodes per account scope to be small and therefore cache friendly.
|
|
// Lock should be held on account.
|
|
func (a *Account) removeLeafNode(c *client) {
|
|
ll := len(a.lleafs)
|
|
for i, l := range a.lleafs {
|
|
if l == c {
|
|
a.lleafs[i] = a.lleafs[ll-1]
|
|
if ll == 1 {
|
|
a.lleafs = nil
|
|
} else {
|
|
a.lleafs = a.lleafs[:ll-1]
|
|
}
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// removeClient keeps our accounting of local active clients updated.
|
|
func (a *Account) removeClient(c *client) int {
|
|
a.mu.Lock()
|
|
n := len(a.clients)
|
|
delete(a.clients, c)
|
|
removed := n != len(a.clients)
|
|
if removed {
|
|
if c.kind == SYSTEM {
|
|
a.sysclients--
|
|
} else if c.kind == LEAF {
|
|
a.nleafs--
|
|
a.removeLeafNode(c)
|
|
}
|
|
}
|
|
a.mu.Unlock()
|
|
|
|
if c != nil && c.srv != nil && removed {
|
|
c.srv.mu.Lock()
|
|
doRemove := a != c.srv.gacc
|
|
c.srv.mu.Unlock()
|
|
if doRemove {
|
|
c.srv.accConnsUpdate(a)
|
|
}
|
|
}
|
|
return n
|
|
}
|
|
|
|
func (a *Account) randomClient() *client {
|
|
if a.ic != nil {
|
|
return a.ic
|
|
}
|
|
var c *client
|
|
for c = range a.clients {
|
|
break
|
|
}
|
|
return c
|
|
}
|
|
|
|
// AddServiceExport will configure the account with the defined export.
|
|
func (a *Account) AddServiceExport(subject string, accounts []*Account) error {
|
|
return a.AddServiceExportWithResponse(subject, Singleton, accounts)
|
|
}
|
|
|
|
// AddServiceExportWithResponse will configure the account with the defined export and response type.
|
|
func (a *Account) AddServiceExportWithResponse(subject string, respType ServiceRespType, accounts []*Account) error {
|
|
if a == nil {
|
|
return ErrMissingAccount
|
|
}
|
|
a.mu.Lock()
|
|
defer a.mu.Unlock()
|
|
|
|
if a.exports.services == nil {
|
|
a.exports.services = make(map[string]*serviceExport)
|
|
}
|
|
|
|
se := a.exports.services[subject]
|
|
// Always create a service export
|
|
if se == nil {
|
|
se = &serviceExport{}
|
|
}
|
|
|
|
if respType != Singleton {
|
|
se.respType = respType
|
|
}
|
|
|
|
if accounts != nil {
|
|
// empty means auth required but will be import token.
|
|
if len(accounts) == 0 {
|
|
se.tokenReq = true
|
|
} else {
|
|
if se.approved == nil {
|
|
se.approved = make(map[string]*Account, len(accounts))
|
|
}
|
|
for _, acc := range accounts {
|
|
se.approved[acc.Name] = acc
|
|
}
|
|
}
|
|
}
|
|
lrt := a.lowestServiceExportResponseTime()
|
|
se.acc = a
|
|
se.respThresh = DEFAULT_SERVICE_EXPORT_RESPONSE_THRESHOLD
|
|
a.exports.services[subject] = se
|
|
if nlrt := a.lowestServiceExportResponseTime(); nlrt != lrt {
|
|
a.updateAllClientsServiceExportResponseTime(nlrt)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// TrackServiceExport will enable latency tracking of the named service.
|
|
// Results will be published in this account to the given results subject.
|
|
func (a *Account) TrackServiceExport(service, results string) error {
|
|
return a.TrackServiceExportWithSampling(service, results, DEFAULT_SERVICE_LATENCY_SAMPLING)
|
|
}
|
|
|
|
// TrackServiceExportWithSampling will enable latency tracking of the named service for the given
|
|
// sampling rate (1-100). Results will be published in this account to the given results subject.
|
|
func (a *Account) TrackServiceExportWithSampling(service, results string, sampling int) error {
|
|
if a == nil {
|
|
return ErrMissingAccount
|
|
}
|
|
|
|
if sampling != 0 { // 0 means triggered by header
|
|
if sampling < 1 || sampling > 100 {
|
|
return ErrBadSampling
|
|
}
|
|
}
|
|
if !IsValidPublishSubject(results) {
|
|
return ErrBadPublishSubject
|
|
}
|
|
// Don't loop back on outselves.
|
|
if a.IsExportService(results) {
|
|
return ErrBadPublishSubject
|
|
}
|
|
|
|
if a.srv != nil && !a.srv.EventsEnabled() {
|
|
return ErrNoSysAccount
|
|
}
|
|
|
|
a.mu.Lock()
|
|
if a.exports.services == nil {
|
|
a.mu.Unlock()
|
|
return ErrMissingService
|
|
}
|
|
ea, ok := a.exports.services[service]
|
|
if !ok {
|
|
a.mu.Unlock()
|
|
return ErrMissingService
|
|
}
|
|
if ea == nil {
|
|
ea = &serviceExport{}
|
|
a.exports.services[service] = ea
|
|
} else if ea.respType != Singleton {
|
|
a.mu.Unlock()
|
|
return ErrBadServiceType
|
|
}
|
|
ea.latency = &serviceLatency{
|
|
sampling: int8(sampling),
|
|
subject: results,
|
|
}
|
|
s := a.srv
|
|
a.mu.Unlock()
|
|
|
|
if s == nil {
|
|
return nil
|
|
}
|
|
|
|
// Now track down the imports and add in latency as needed to enable.
|
|
s.accounts.Range(func(k, v interface{}) bool {
|
|
acc := v.(*Account)
|
|
acc.mu.Lock()
|
|
for _, im := range acc.imports.services {
|
|
if im != nil && im.acc.Name == a.Name && subjectIsSubsetMatch(im.to, service) {
|
|
im.latency = ea.latency
|
|
}
|
|
}
|
|
acc.mu.Unlock()
|
|
return true
|
|
})
|
|
|
|
return nil
|
|
}
|
|
|
|
// UnTrackServiceExport will disable latency tracking of the named service.
|
|
func (a *Account) UnTrackServiceExport(service string) {
|
|
if a == nil || (a.srv != nil && !a.srv.EventsEnabled()) {
|
|
return
|
|
}
|
|
|
|
a.mu.Lock()
|
|
if a == nil || a.exports.services == nil {
|
|
a.mu.Unlock()
|
|
return
|
|
}
|
|
ea, ok := a.exports.services[service]
|
|
if !ok || ea == nil || ea.latency == nil {
|
|
a.mu.Unlock()
|
|
return
|
|
}
|
|
// We have latency here.
|
|
ea.latency = nil
|
|
s := a.srv
|
|
a.mu.Unlock()
|
|
|
|
if s == nil {
|
|
return
|
|
}
|
|
|
|
// Now track down the imports and clean them up.
|
|
s.accounts.Range(func(k, v interface{}) bool {
|
|
acc := v.(*Account)
|
|
acc.mu.Lock()
|
|
for _, im := range acc.imports.services {
|
|
if im != nil && im.acc.Name == a.Name {
|
|
if subjectIsSubsetMatch(im.to, service) {
|
|
im.latency, im.m1 = nil, nil
|
|
}
|
|
}
|
|
}
|
|
acc.mu.Unlock()
|
|
return true
|
|
})
|
|
}
|
|
|
|
// IsExportService will indicate if this service exists. Will check wildcard scenarios.
|
|
func (a *Account) IsExportService(service string) bool {
|
|
a.mu.RLock()
|
|
defer a.mu.RUnlock()
|
|
_, ok := a.exports.services[service]
|
|
if ok {
|
|
return true
|
|
}
|
|
tokens := strings.Split(service, tsep)
|
|
for subj := range a.exports.services {
|
|
if isSubsetMatch(tokens, subj) {
|
|
return true
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
// IsExportServiceTracking will indicate if given publish subject is an export service with tracking enabled.
|
|
func (a *Account) IsExportServiceTracking(service string) bool {
|
|
a.mu.RLock()
|
|
ea, ok := a.exports.services[service]
|
|
if ok && ea == nil {
|
|
a.mu.RUnlock()
|
|
return false
|
|
}
|
|
if ok && ea != nil && ea.latency != nil {
|
|
a.mu.RUnlock()
|
|
return true
|
|
}
|
|
// FIXME(dlc) - Might want to cache this is in the hot path checking for latency tracking.
|
|
tokens := strings.Split(service, tsep)
|
|
for subj, ea := range a.exports.services {
|
|
if isSubsetMatch(tokens, subj) && ea != nil && ea.latency != nil {
|
|
a.mu.RUnlock()
|
|
return true
|
|
}
|
|
}
|
|
a.mu.RUnlock()
|
|
return false
|
|
}
|
|
|
|
// ServiceLatency is the JSON message sent out in response to latency tracking for
|
|
// an accounts exported services. Additional client info is available in requestor
|
|
// and responder. Note that for a requestor, the only information shared by default
|
|
// is the RTT used to calculate the total latency. The requestor's account can
|
|
// designate to share the additional information in the service import.
|
|
type ServiceLatency struct {
|
|
TypedEvent
|
|
|
|
Status int `json:"status"`
|
|
Error string `json:"description,omitempty"`
|
|
Requestor LatencyClient `json:"requestor,omitempty"`
|
|
Responder LatencyClient `json:"responder,omitempty"`
|
|
RequestHeader http.Header `json:"header,omitempty"` // only contains header(s) triggering the measurement
|
|
RequestStart time.Time `json:"start"`
|
|
ServiceLatency time.Duration `json:"service"`
|
|
SystemLatency time.Duration `json:"system"`
|
|
TotalLatency time.Duration `json:"total"`
|
|
}
|
|
|
|
// ServiceLatencyType is the NATS Event Type for ServiceLatency
|
|
const ServiceLatencyType = "io.nats.server.metric.v1.service_latency"
|
|
|
|
// LatencyClient is the JSON message structure assigned to requestors and responders.
|
|
// Note that for a requestor, the only information shared by default is the RTT used
|
|
// to calculate the total latency. The requestor's account can designate to share
|
|
// the additional information in the service import.
|
|
type LatencyClient struct {
|
|
Account string `json:"acc"`
|
|
RTT time.Duration `json:"rtt"`
|
|
Start time.Time `json:"start,omitempty"`
|
|
User string `json:"user,omitempty"`
|
|
Name string `json:"name,omitempty"`
|
|
Lang string `json:"lang,omitempty"`
|
|
Version string `json:"ver,omitempty"`
|
|
IP string `json:"ip,omitempty"`
|
|
CID uint64 `json:"cid,omitempty"`
|
|
Server string `json:"server,omitempty"`
|
|
}
|
|
|
|
// NATSTotalTime is a helper function that totals the NATS latencies.
|
|
func (nl *ServiceLatency) NATSTotalTime() time.Duration {
|
|
return nl.Requestor.RTT + nl.Responder.RTT + nl.SystemLatency
|
|
}
|
|
|
|
// Merge function to merge m1 and m2 (requestor and responder) measurements
|
|
// when there are two samples. This happens when the requestor and responder
|
|
// are on different servers.
|
|
//
|
|
// m2 ServiceLatency is correct, so use that.
|
|
// m1 TotalLatency is correct, so use that.
|
|
// Will use those to back into NATS latency.
|
|
func (m1 *ServiceLatency) merge(m2 *ServiceLatency) {
|
|
m1.SystemLatency = m1.ServiceLatency - (m2.ServiceLatency + m2.Responder.RTT)
|
|
m1.ServiceLatency = m2.ServiceLatency
|
|
m1.Responder = m2.Responder
|
|
sanitizeLatencyMetric(m1)
|
|
}
|
|
|
|
// sanitizeLatencyMetric adjusts latency metric values that could go
|
|
// negative in some edge conditions since we estimate client RTT
|
|
// for both requestor and responder.
|
|
// These numbers are never meant to be negative, it just could be
|
|
// how we back into the values based on estimated RTT.
|
|
func sanitizeLatencyMetric(sl *ServiceLatency) {
|
|
if sl.ServiceLatency < 0 {
|
|
sl.ServiceLatency = 0
|
|
}
|
|
if sl.SystemLatency < 0 {
|
|
sl.SystemLatency = 0
|
|
}
|
|
}
|
|
|
|
// Used for transporting remote latency measurements.
|
|
type remoteLatency struct {
|
|
Account string `json:"account"`
|
|
ReqId string `json:"req_id"`
|
|
M2 ServiceLatency `json:"m2"`
|
|
respThresh time.Duration
|
|
}
|
|
|
|
// sendLatencyResult will send a latency result and clear the si of the requestor(rc).
|
|
func (a *Account) sendLatencyResult(si *serviceImport, sl *ServiceLatency) {
|
|
sl.Type = ServiceLatencyType
|
|
sl.ID = a.nextEventID()
|
|
sl.Time = time.Now().UTC()
|
|
a.mu.Lock()
|
|
lsubj := si.latency.subject
|
|
si.rc = nil
|
|
a.mu.Unlock()
|
|
|
|
a.srv.sendInternalAccountMsg(a, lsubj, sl)
|
|
}
|
|
|
|
// Used to send a bad request metric when we do not have a reply subject
|
|
func (a *Account) sendBadRequestTrackingLatency(si *serviceImport, requestor *client, header http.Header) {
|
|
sl := &ServiceLatency{
|
|
Status: 400,
|
|
Error: "Bad Request",
|
|
Requestor: requestor.getClientInfo(si.share),
|
|
}
|
|
sl.RequestHeader = header
|
|
sl.RequestStart = time.Now().Add(-sl.Requestor.RTT).UTC()
|
|
a.sendLatencyResult(si, sl)
|
|
}
|
|
|
|
// Used to send a latency result when the requestor interest was lost before the
|
|
// response could be delivered.
|
|
func (a *Account) sendReplyInterestLostTrackLatency(si *serviceImport) {
|
|
sl := &ServiceLatency{
|
|
Status: 408,
|
|
Error: "Request Timeout",
|
|
}
|
|
a.mu.RLock()
|
|
rc := si.rc
|
|
share := si.share
|
|
ts := si.ts
|
|
sl.RequestHeader = si.trackingHdr
|
|
a.mu.RUnlock()
|
|
if rc != nil {
|
|
sl.Requestor = rc.getClientInfo(share)
|
|
}
|
|
sl.RequestStart = time.Unix(0, ts-int64(sl.Requestor.RTT)).UTC()
|
|
a.sendLatencyResult(si, sl)
|
|
}
|
|
|
|
func (a *Account) sendBackendErrorTrackingLatency(si *serviceImport, reason rsiReason) {
|
|
sl := &ServiceLatency{}
|
|
a.mu.RLock()
|
|
rc := si.rc
|
|
share := si.share
|
|
ts := si.ts
|
|
sl.RequestHeader = si.trackingHdr
|
|
a.mu.RUnlock()
|
|
if rc != nil {
|
|
sl.Requestor = rc.getClientInfo(share)
|
|
}
|
|
sl.RequestStart = time.Unix(0, ts-int64(sl.Requestor.RTT)).UTC()
|
|
if reason == rsiNoDelivery {
|
|
sl.Status = 503
|
|
sl.Error = "Service Unavailable"
|
|
} else if reason == rsiTimeout {
|
|
sl.Status = 504
|
|
sl.Error = "Service Timeout"
|
|
}
|
|
a.sendLatencyResult(si, sl)
|
|
}
|
|
|
|
// sendTrackingMessage will send out the appropriate tracking information for the
|
|
// service request/response latency. This is called when the requestor's server has
|
|
// received the response.
|
|
// TODO(dlc) - holding locks for RTTs may be too much long term. Should revisit.
|
|
func (a *Account) sendTrackingLatency(si *serviceImport, responder *client) bool {
|
|
if si.rc == nil {
|
|
return true
|
|
}
|
|
|
|
ts := time.Now()
|
|
serviceRTT := time.Duration(ts.UnixNano() - si.ts)
|
|
requestor := si.rc
|
|
|
|
sl := &ServiceLatency{
|
|
Status: 200,
|
|
Requestor: requestor.getClientInfo(si.share),
|
|
Responder: responder.getClientInfo(true),
|
|
}
|
|
sl.RequestStart = time.Unix(0, si.ts-int64(sl.Requestor.RTT)).UTC()
|
|
sl.ServiceLatency = serviceRTT - sl.Responder.RTT
|
|
sl.TotalLatency = sl.Requestor.RTT + serviceRTT
|
|
if sl.Responder.RTT > 0 {
|
|
sl.SystemLatency = time.Since(ts)
|
|
sl.TotalLatency += sl.SystemLatency
|
|
}
|
|
sl.RequestHeader = si.trackingHdr
|
|
sanitizeLatencyMetric(sl)
|
|
|
|
sl.Type = ServiceLatencyType
|
|
sl.ID = a.nextEventID()
|
|
sl.Time = time.Now().UTC()
|
|
|
|
// If we are expecting a remote measurement, store our sl here.
|
|
// We need to account for the race between this and us receiving the
|
|
// remote measurement.
|
|
// FIXME(dlc) - We need to clean these up but this should happen
|
|
// already with the auto-expire logic.
|
|
if responder != nil && responder.kind != CLIENT {
|
|
si.acc.mu.Lock()
|
|
if si.m1 != nil {
|
|
m1, m2 := sl, si.m1
|
|
m1.merge(m2)
|
|
si.acc.mu.Unlock()
|
|
a.srv.sendInternalAccountMsg(a, si.latency.subject, m1)
|
|
a.mu.Lock()
|
|
si.rc = nil
|
|
a.mu.Unlock()
|
|
return true
|
|
}
|
|
si.m1 = sl
|
|
si.acc.mu.Unlock()
|
|
return false
|
|
} else {
|
|
a.srv.sendInternalAccountMsg(a, si.latency.subject, sl)
|
|
a.mu.Lock()
|
|
si.rc = nil
|
|
a.mu.Unlock()
|
|
}
|
|
return true
|
|
}
|
|
|
|
// This will check to make sure our response lower threshold is set
|
|
// properly in any clients doing rrTracking.
|
|
// Lock should be held.
|
|
func (a *Account) updateAllClientsServiceExportResponseTime(lrt time.Duration) {
|
|
for c := range a.clients {
|
|
c.mu.Lock()
|
|
if c.rrTracking != nil && lrt != c.rrTracking.lrt {
|
|
c.rrTracking.lrt = lrt
|
|
if c.rrTracking.ptmr.Stop() {
|
|
c.rrTracking.ptmr.Reset(lrt)
|
|
}
|
|
}
|
|
c.mu.Unlock()
|
|
}
|
|
}
|
|
|
|
// Will select the lowest respThresh from all service exports.
|
|
// Read lock should be held.
|
|
func (a *Account) lowestServiceExportResponseTime() time.Duration {
|
|
// Lowest we will allow is 5 minutes. Its an upper bound for this function.
|
|
lrt := time.Duration(5 * time.Minute)
|
|
for _, se := range a.exports.services {
|
|
if se.respThresh < lrt {
|
|
lrt = se.respThresh
|
|
}
|
|
}
|
|
return lrt
|
|
}
|
|
|
|
// AddServiceImportWithClaim will add in the service import via the jwt claim.
|
|
func (a *Account) AddServiceImportWithClaim(destination *Account, from, to string, imClaim *jwt.Import) error {
|
|
if destination == nil {
|
|
return ErrMissingAccount
|
|
}
|
|
// Empty means use from. Also means we can use wildcards since we are not doing remapping.
|
|
if !IsValidSubject(from) || (to != "" && (!IsValidLiteralSubject(from) || !IsValidLiteralSubject(to))) {
|
|
return ErrInvalidSubject
|
|
}
|
|
|
|
// Empty means use from.
|
|
if to == "" {
|
|
to = from
|
|
}
|
|
// First check to see if the account has authorized us to route to the "to" subject.
|
|
if !destination.checkServiceImportAuthorized(a, to, imClaim) {
|
|
return ErrServiceImportAuthorization
|
|
}
|
|
|
|
_, err := a.addServiceImport(destination, from, to, imClaim)
|
|
return err
|
|
}
|
|
|
|
// SetServiceImportSharing will allow sharing of information about requests with the export account.
|
|
// Used for service latency tracking at the moment.
|
|
func (a *Account) SetServiceImportSharing(destination *Account, to string, allow bool) error {
|
|
a.mu.Lock()
|
|
defer a.mu.Unlock()
|
|
if a.isClaimAccount() {
|
|
return fmt.Errorf("claim based accounts can not be updated directly")
|
|
}
|
|
for _, si := range a.imports.services {
|
|
if si.acc == destination && si.to == to {
|
|
si.share = allow
|
|
return nil
|
|
}
|
|
}
|
|
return fmt.Errorf("service import not found")
|
|
}
|
|
|
|
// AddServiceImport will add a route to an account to send published messages / requests
|
|
// to the destination account. From is the local subject to map, To is the
|
|
// subject that will appear on the destination account. Destination will need
|
|
// to have an import rule to allow access via addService.
|
|
func (a *Account) AddServiceImport(destination *Account, from, to string) error {
|
|
return a.AddServiceImportWithClaim(destination, from, to, nil)
|
|
}
|
|
|
|
// NumPendingReverseResponses returns the number of response mappings we have for all outstanding
|
|
// requests for service imports.
|
|
func (a *Account) NumPendingReverseResponses() int {
|
|
a.mu.RLock()
|
|
defer a.mu.RUnlock()
|
|
return len(a.imports.rrMap)
|
|
}
|
|
|
|
// NumPendingAllResponses return the number of all responses outstanding for service exports.
|
|
func (a *Account) NumPendingAllResponses() int {
|
|
return a.NumPendingResponses("")
|
|
}
|
|
|
|
// NumResponsesPending returns the number of responses outstanding for service exports
|
|
// on this account. An empty filter string returns all responses regardless of which export.
|
|
// If you specify the filter we will only return ones that are for that export.
|
|
// NOTE this is only for what this server is tracking.
|
|
func (a *Account) NumPendingResponses(filter string) int {
|
|
a.mu.RLock()
|
|
defer a.mu.RUnlock()
|
|
if filter == "" {
|
|
return len(a.exports.responses)
|
|
}
|
|
se := a.getServiceExport(filter)
|
|
if se == nil {
|
|
return 0
|
|
}
|
|
var nre int
|
|
for _, si := range a.exports.responses {
|
|
if si.se == se {
|
|
nre++
|
|
}
|
|
}
|
|
return nre
|
|
}
|
|
|
|
// NumServiceImports returns the number of service imports we have configured.
|
|
func (a *Account) NumServiceImports() int {
|
|
a.mu.RLock()
|
|
defer a.mu.RUnlock()
|
|
return len(a.imports.services)
|
|
}
|
|
|
|
// Reason why we are removing this response serviceImport.
|
|
type rsiReason int
|
|
|
|
const (
|
|
rsiOk = rsiReason(iota)
|
|
rsiNoDelivery
|
|
rsiTimeout
|
|
)
|
|
|
|
// removeRespServiceImport removes a response si mapping and the reverse entries for interest detection.
|
|
func (a *Account) removeRespServiceImport(si *serviceImport, reason rsiReason) {
|
|
if si == nil {
|
|
return
|
|
}
|
|
|
|
a.mu.Lock()
|
|
delete(a.exports.responses, si.from)
|
|
dest := si.acc
|
|
to := si.to
|
|
tracking := si.tracking
|
|
rc := si.rc
|
|
a.mu.Unlock()
|
|
|
|
if tracking && rc != nil {
|
|
a.sendBackendErrorTrackingLatency(si, reason)
|
|
}
|
|
|
|
dest.checkForReverseEntry(to, si, false)
|
|
}
|
|
|
|
// removeServiceImport will remove the route by subject.
|
|
func (a *Account) removeServiceImport(subject string) {
|
|
a.mu.Lock()
|
|
si, ok := a.imports.services[subject]
|
|
delete(a.imports.services, subject)
|
|
|
|
var sid []byte
|
|
c := a.ic
|
|
|
|
if ok && si != nil {
|
|
if a.ic != nil && si.sid != nil {
|
|
sid = si.sid
|
|
}
|
|
}
|
|
a.mu.Unlock()
|
|
|
|
if sid != nil {
|
|
c.processUnsub(sid)
|
|
}
|
|
}
|
|
|
|
// This tracks responses to service requests mappings. This is used for cleanup.
|
|
func (a *Account) addReverseRespMapEntry(acc *Account, reply, from string) {
|
|
a.mu.Lock()
|
|
if a.imports.rrMap == nil {
|
|
a.imports.rrMap = make(map[string][]*serviceRespEntry)
|
|
}
|
|
sre := &serviceRespEntry{acc, from}
|
|
sra := a.imports.rrMap[reply]
|
|
a.imports.rrMap[reply] = append(sra, sre)
|
|
a.mu.Unlock()
|
|
}
|
|
|
|
// checkForReverseEntries is for when we are trying to match reverse entries to a wildcard.
|
|
// This will be called from checkForReverseEntry when the reply arg is a wildcard subject.
|
|
// This will usually be called in a go routine since we need to walk all the entries.
|
|
func (a *Account) checkForReverseEntries(reply string, checkInterest bool) {
|
|
a.mu.RLock()
|
|
if len(a.imports.rrMap) == 0 {
|
|
a.mu.RUnlock()
|
|
return
|
|
}
|
|
|
|
if subjectIsLiteral(reply) {
|
|
a.mu.RUnlock()
|
|
a.checkForReverseEntry(reply, nil, checkInterest)
|
|
return
|
|
}
|
|
|
|
var _rs [32]string
|
|
rs := _rs[:0]
|
|
|
|
for k := range a.imports.rrMap {
|
|
if subjectIsSubsetMatch(k, reply) {
|
|
rs = append(rs, k)
|
|
}
|
|
}
|
|
a.mu.RUnlock()
|
|
|
|
for _, reply := range rs {
|
|
a.checkForReverseEntry(reply, nil, checkInterest)
|
|
}
|
|
}
|
|
|
|
// This checks for any response map entries. If you specify an si we will only match and
|
|
// clean up for that one, otherwise we remove them all.
|
|
func (a *Account) checkForReverseEntry(reply string, si *serviceImport, checkInterest bool) {
|
|
a.mu.RLock()
|
|
if len(a.imports.rrMap) == 0 {
|
|
a.mu.RUnlock()
|
|
return
|
|
}
|
|
|
|
if subjectHasWildcard(reply) {
|
|
a.mu.RUnlock()
|
|
go a.checkForReverseEntries(reply, checkInterest)
|
|
return
|
|
}
|
|
|
|
sres := a.imports.rrMap[reply]
|
|
if sres == nil {
|
|
a.mu.RUnlock()
|
|
return
|
|
}
|
|
|
|
// If we are here we have an entry we should check.
|
|
// If requested we will first check if there is any
|
|
// interest for this subject for the entire account.
|
|
// If there is we can not delete any entries yet.
|
|
// Note that if we are here reply has to be a literal subject.
|
|
if checkInterest {
|
|
rr := a.sl.Match(reply)
|
|
// If interest still exists we can not clean these up yet.
|
|
if len(rr.psubs)+len(rr.qsubs) > 0 {
|
|
a.mu.RUnlock()
|
|
return
|
|
}
|
|
}
|
|
a.mu.RUnlock()
|
|
|
|
// Delete the appropriate entries here based on optional si.
|
|
a.mu.Lock()
|
|
if si == nil {
|
|
delete(a.imports.rrMap, reply)
|
|
} else {
|
|
// Find the one we are looking for..
|
|
for i, sre := range sres {
|
|
if sre.msub == si.from {
|
|
sres = append(sres[:i], sres[i+1:]...)
|
|
break
|
|
}
|
|
}
|
|
if len(sres) > 0 {
|
|
a.imports.rrMap[si.to] = sres
|
|
} else {
|
|
delete(a.imports.rrMap, si.to)
|
|
}
|
|
}
|
|
a.mu.Unlock()
|
|
|
|
// If we are here we no longer have interest and we have
|
|
// response entries that we should clean up.
|
|
if si == nil {
|
|
for _, sre := range sres {
|
|
acc := sre.acc
|
|
var trackingCleanup bool
|
|
var rsi *serviceImport
|
|
acc.mu.Lock()
|
|
if rsi = acc.exports.responses[sre.msub]; rsi != nil {
|
|
delete(acc.exports.responses, rsi.from)
|
|
trackingCleanup = rsi.tracking && rsi.rc != nil
|
|
}
|
|
acc.mu.Unlock()
|
|
|
|
if trackingCleanup {
|
|
acc.sendReplyInterestLostTrackLatency(rsi)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Internal check to see if a service import exists.
|
|
func (a *Account) serviceImportExists(dest *Account, from string) bool {
|
|
a.mu.RLock()
|
|
dup := a.imports.services[from]
|
|
a.mu.RUnlock()
|
|
return dup != nil
|
|
}
|
|
|
|
// Add a service import.
|
|
// This does no checks and should only be called by the msg processing code. Use
|
|
// AddServiceImport from above if responding to user input or config changes, etc.
|
|
func (a *Account) addServiceImport(dest *Account, from, to string, claim *jwt.Import) (*serviceImport, error) {
|
|
rt := Singleton
|
|
var lat *serviceLatency
|
|
|
|
dest.mu.RLock()
|
|
se := dest.getServiceExport(to)
|
|
if se != nil {
|
|
rt = se.respType
|
|
lat = se.latency
|
|
}
|
|
dest.mu.RUnlock()
|
|
|
|
a.mu.Lock()
|
|
if a.imports.services == nil {
|
|
a.imports.services = make(map[string]*serviceImport)
|
|
} else if dup := a.imports.services[from]; dup != nil {
|
|
a.mu.Unlock()
|
|
return nil, fmt.Errorf("duplicate service import subject %q, previously used in import for account %q, subject %q",
|
|
from, dup.acc.Name, dup.to)
|
|
}
|
|
|
|
if to == "" {
|
|
to = from
|
|
}
|
|
hasWC := subjectHasWildcard(from)
|
|
|
|
si := &serviceImport{dest, claim, se, nil, from, to, "", 0, rt, lat, nil, nil, hasWC, false, false, false, false, nil}
|
|
a.imports.services[from] = si
|
|
a.mu.Unlock()
|
|
|
|
if err := a.addServiceImportSub(si); err != nil {
|
|
a.removeServiceImport(si.from)
|
|
return nil, err
|
|
}
|
|
return si, nil
|
|
}
|
|
|
|
// Returns the internal client, will create one if not present.
|
|
// Lock should be held.
|
|
func (a *Account) internalClient() *client {
|
|
if a.ic == nil && a.srv != nil {
|
|
a.ic = a.srv.createInternalAccountClient()
|
|
a.ic.acc = a
|
|
}
|
|
return a.ic
|
|
}
|
|
|
|
// Internal account scoped subscriptions.
|
|
func (a *Account) subscribeInternal(subject string, cb msgHandler) (*subscription, error) {
|
|
a.mu.Lock()
|
|
c := a.internalClient()
|
|
a.isid++
|
|
sid := strconv.FormatUint(a.isid, 10)
|
|
a.mu.Unlock()
|
|
|
|
// This will happen in parsing when the account has not been properly setup.
|
|
if c == nil {
|
|
return nil, fmt.Errorf("no internal account client")
|
|
}
|
|
|
|
return c.processSub([]byte(subject), nil, []byte(sid), cb, false)
|
|
}
|
|
|
|
// This will add an account subscription that matches the "from" from a service import entry.
|
|
func (a *Account) addServiceImportSub(si *serviceImport) error {
|
|
a.mu.Lock()
|
|
c := a.internalClient()
|
|
// This will happen in parsing when the account has not been properly setup.
|
|
if c == nil {
|
|
a.mu.Unlock()
|
|
return nil
|
|
}
|
|
if si.sid != nil {
|
|
a.mu.Unlock()
|
|
return fmt.Errorf("duplicate call to create subscription for service import")
|
|
}
|
|
a.isid++
|
|
sid := strconv.FormatUint(a.isid, 10)
|
|
si.sid = []byte(sid)
|
|
subject := si.from
|
|
a.mu.Unlock()
|
|
|
|
cb := func(sub *subscription, c *client, subject, reply string, msg []byte) {
|
|
c.processServiceImport(si, a, msg)
|
|
}
|
|
_, err := c.processSub([]byte(subject), nil, []byte(sid), cb, true)
|
|
return err
|
|
}
|
|
|
|
// Remove all the subscriptions associated with service imports.
|
|
func (a *Account) removeAllServiceImportSubs() {
|
|
a.mu.RLock()
|
|
var sids [][]byte
|
|
for _, si := range a.imports.services {
|
|
if si.sid != nil {
|
|
sids = append(sids, si.sid)
|
|
si.sid = nil
|
|
}
|
|
}
|
|
c := a.ic
|
|
a.ic = nil
|
|
a.mu.RUnlock()
|
|
|
|
if c == nil {
|
|
return
|
|
}
|
|
for _, sid := range sids {
|
|
c.processUnsub(sid)
|
|
}
|
|
c.closeConnection(InternalClient)
|
|
}
|
|
|
|
// Add in subscriptions for all registered service imports.
|
|
func (a *Account) addAllServiceImportSubs() {
|
|
for _, si := range a.imports.services {
|
|
a.addServiceImportSub(si)
|
|
}
|
|
}
|
|
|
|
var (
|
|
// header where all information is encoded in one value.
|
|
trcUber = textproto.CanonicalMIMEHeaderKey("Uber-Trace-Id")
|
|
trcCtx = textproto.CanonicalMIMEHeaderKey("Traceparent")
|
|
trcB3 = textproto.CanonicalMIMEHeaderKey("B3")
|
|
// openzipkin header to check
|
|
trcB3Sm = textproto.CanonicalMIMEHeaderKey("X-B3-Sampled")
|
|
trcB3Id = textproto.CanonicalMIMEHeaderKey("X-B3-TraceId")
|
|
// additional header needed to include when present
|
|
trcB3PSId = textproto.CanonicalMIMEHeaderKey("X-B3-ParentSpanId")
|
|
trcB3SId = textproto.CanonicalMIMEHeaderKey("X-B3-SpanId")
|
|
trcCtxSt = textproto.CanonicalMIMEHeaderKey("Tracestate")
|
|
trcUberCtxPrefix = textproto.CanonicalMIMEHeaderKey("Uberctx-")
|
|
)
|
|
|
|
func newB3Header(h http.Header) http.Header {
|
|
retHdr := http.Header{}
|
|
if v, ok := h[trcB3Sm]; ok {
|
|
retHdr[trcB3Sm] = v
|
|
}
|
|
if v, ok := h[trcB3Id]; ok {
|
|
retHdr[trcB3Id] = v
|
|
}
|
|
if v, ok := h[trcB3PSId]; ok {
|
|
retHdr[trcB3PSId] = v
|
|
}
|
|
if v, ok := h[trcB3SId]; ok {
|
|
retHdr[trcB3SId] = v
|
|
}
|
|
return retHdr
|
|
}
|
|
|
|
func newUberHeader(h http.Header, tId []string) http.Header {
|
|
retHdr := http.Header{trcUber: tId}
|
|
for k, v := range h {
|
|
if strings.HasPrefix(k, trcUberCtxPrefix) {
|
|
retHdr[k] = v
|
|
}
|
|
}
|
|
return retHdr
|
|
}
|
|
|
|
func newTraceCtxHeader(h http.Header, tId []string) http.Header {
|
|
retHdr := http.Header{trcCtx: tId}
|
|
if v, ok := h[trcCtxSt]; ok {
|
|
retHdr[trcCtxSt] = v
|
|
}
|
|
return retHdr
|
|
}
|
|
|
|
// Helper to determine when to sample. When header has a value, sampling is driven by header
|
|
func shouldSample(l *serviceLatency, c *client) (bool, http.Header) {
|
|
if l == nil {
|
|
return false, nil
|
|
}
|
|
if l.sampling < 0 {
|
|
return false, nil
|
|
}
|
|
if l.sampling >= 100 {
|
|
return true, nil
|
|
}
|
|
if l.sampling > 0 && rand.Int31n(100) <= int32(l.sampling) {
|
|
return true, nil
|
|
}
|
|
h := c.parseState.getHeader()
|
|
if len(h) == 0 {
|
|
return false, nil
|
|
}
|
|
if tId := h[trcUber]; len(tId) != 0 {
|
|
// sample 479fefe9525eddb:5adb976bfc1f95c1:479fefe9525eddb:1
|
|
tk := strings.Split(tId[0], ":")
|
|
if len(tk) == 4 && len(tk[3]) > 0 && len(tk[3]) <= 2 {
|
|
dst := [2]byte{}
|
|
src := [2]byte{'0', tk[3][0]}
|
|
if len(tk[3]) == 2 {
|
|
src[1] = tk[3][1]
|
|
}
|
|
if _, err := hex.Decode(dst[:], src[:]); err == nil && dst[0]&1 == 1 {
|
|
return true, newUberHeader(h, tId)
|
|
}
|
|
}
|
|
return false, nil
|
|
} else if sampled := h[trcB3Sm]; len(sampled) != 0 && sampled[0] == "1" {
|
|
return true, newB3Header(h) // allowed
|
|
} else if len(sampled) != 0 && sampled[0] == "0" {
|
|
return false, nil // denied
|
|
} else if _, ok := h[trcB3Id]; ok {
|
|
// sample 80f198ee56343ba864fe8b2a57d3eff7
|
|
// presence (with X-B3-Sampled not being 0) means sampling left to recipient
|
|
return true, newB3Header(h)
|
|
} else if b3 := h[trcB3]; len(b3) != 0 {
|
|
// sample 80f198ee56343ba864fe8b2a57d3eff7-e457b5a2e4d86bd1-1-05e3ac9a4f6e3b90
|
|
// sample 0
|
|
tk := strings.Split(b3[0], "-")
|
|
if len(tk) > 2 && tk[2] == "0" {
|
|
return false, nil // denied
|
|
} else if len(tk) == 1 && tk[0] == "0" {
|
|
return false, nil // denied
|
|
}
|
|
return true, http.Header{trcB3: b3} // sampling allowed or left to recipient of header
|
|
} else if tId := h[trcCtx]; len(tId) != 0 {
|
|
// sample 00-4bf92f3577b34da6a3ce929d0e0e4736-00f067aa0ba902b7-01
|
|
tk := strings.Split(tId[0], "-")
|
|
if len(tk) == 4 && len([]byte(tk[3])) == 2 && tk[3] == "01" {
|
|
return true, newTraceCtxHeader(h, tId)
|
|
} else {
|
|
return false, nil
|
|
}
|
|
}
|
|
return false, nil
|
|
}
|
|
|
|
// Used to mimic client like replies.
|
|
const (
|
|
replyPrefix = "_R_."
|
|
trackSuffix = ".T"
|
|
replyPrefixLen = len(replyPrefix)
|
|
baseServerLen = 10
|
|
replyLen = 6
|
|
minReplyLen = 15
|
|
digits = "0123456789ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz"
|
|
base = 62
|
|
)
|
|
|
|
// This is where all service export responses are handled.
|
|
func (a *Account) processServiceImportResponse(sub *subscription, c *client, subject, reply string, msg []byte) {
|
|
a.mu.RLock()
|
|
if a.expired || len(a.exports.responses) == 0 {
|
|
a.mu.RUnlock()
|
|
return
|
|
}
|
|
si := a.exports.responses[subject]
|
|
if si == nil || si.invalid {
|
|
a.mu.RUnlock()
|
|
return
|
|
}
|
|
a.mu.RUnlock()
|
|
|
|
// Send for normal processing.
|
|
c.processServiceImport(si, a, msg)
|
|
}
|
|
|
|
// Will create a wildcard subscription to handle interest graph propagation for all
|
|
// service replies.
|
|
// Lock should not be held.
|
|
func (a *Account) createRespWildcard() []byte {
|
|
a.mu.Lock()
|
|
if a.prand == nil {
|
|
a.prand = rand.New(rand.NewSource(time.Now().UnixNano()))
|
|
}
|
|
var b = [baseServerLen]byte{'_', 'R', '_', '.'}
|
|
rn := a.prand.Int63()
|
|
for i, l := replyPrefixLen, rn; i < len(b); i++ {
|
|
b[i] = digits[l%base]
|
|
l /= base
|
|
}
|
|
a.siReply = append(b[:], '.')
|
|
pre := a.siReply
|
|
wcsub := append(a.siReply, '>')
|
|
c := a.internalClient()
|
|
a.isid++
|
|
sid := strconv.FormatUint(a.isid, 10)
|
|
a.mu.Unlock()
|
|
|
|
// Create subscription and internal callback for all the wildcard response subjects.
|
|
c.processSub(wcsub, nil, []byte(sid), a.processServiceImportResponse, false)
|
|
|
|
return pre
|
|
}
|
|
|
|
// Test whether this is a tracked reply.
|
|
func isTrackedReply(reply []byte) bool {
|
|
lreply := len(reply) - 1
|
|
return lreply > 3 && reply[lreply-1] == '.' && reply[lreply] == 'T'
|
|
}
|
|
|
|
// Generate a new service reply from the wildcard prefix.
|
|
// FIXME(dlc) - probably do not have to use rand here. about 25ns per.
|
|
func (a *Account) newServiceReply(tracking bool) []byte {
|
|
a.mu.RLock()
|
|
replyPre := a.siReply
|
|
s := a.srv
|
|
a.mu.RUnlock()
|
|
|
|
if replyPre == nil {
|
|
replyPre = a.createRespWildcard()
|
|
}
|
|
|
|
var b [replyLen]byte
|
|
rn := a.prand.Int63()
|
|
for i, l := 0, rn; i < len(b); i++ {
|
|
b[i] = digits[l%base]
|
|
l /= base
|
|
}
|
|
// Make sure to copy.
|
|
reply := make([]byte, 0, len(replyPre)+len(b))
|
|
reply = append(reply, replyPre...)
|
|
reply = append(reply, b[:]...)
|
|
|
|
if tracking && s.sys != nil {
|
|
// Add in our tracking identifier. This allows the metrics to get back to only
|
|
// this server without needless SUBS/UNSUBS.
|
|
reply = append(reply, '.')
|
|
reply = append(reply, s.sys.shash...)
|
|
reply = append(reply, '.', 'T')
|
|
}
|
|
return reply
|
|
}
|
|
|
|
// Checks if a serviceImport was created to map responses.
|
|
func (si *serviceImport) isRespServiceImport() bool {
|
|
return si != nil && si.response
|
|
}
|
|
|
|
// Sets the response theshold timer for a service export.
|
|
// Account lock should be held
|
|
func (se *serviceExport) setResponseThresholdTimer() {
|
|
if se.rtmr != nil {
|
|
return // Already set
|
|
}
|
|
se.rtmr = time.AfterFunc(se.respThresh, se.checkExpiredResponses)
|
|
}
|
|
|
|
// Account lock should be held
|
|
func (se *serviceExport) clearResponseThresholdTimer() bool {
|
|
if se.rtmr == nil {
|
|
return true
|
|
}
|
|
stopped := se.rtmr.Stop()
|
|
se.rtmr = nil
|
|
return stopped
|
|
}
|
|
|
|
// checkExpiredResponses will check for any pending responses that need to
|
|
// be cleaned up.
|
|
func (se *serviceExport) checkExpiredResponses() {
|
|
acc := se.acc
|
|
if acc == nil {
|
|
acc.mu.Lock()
|
|
se.clearResponseThresholdTimer()
|
|
acc.mu.Unlock()
|
|
return
|
|
}
|
|
|
|
var expired []*serviceImport
|
|
mints := time.Now().UnixNano() - int64(se.respThresh)
|
|
|
|
// TODO(dlc) - Should we release lock while doing this? Or only do these in batches?
|
|
// Should we break this up for responses only from this service export?
|
|
// Responses live on acc directly for fast inbound processsing for the _R_ wildcard.
|
|
// We could do another indirection at this level but just to get to the service export?
|
|
var totalResponses int
|
|
acc.mu.RLock()
|
|
for _, si := range acc.exports.responses {
|
|
if si.se == se {
|
|
totalResponses++
|
|
if si.ts <= mints {
|
|
expired = append(expired, si)
|
|
}
|
|
}
|
|
}
|
|
acc.mu.RUnlock()
|
|
|
|
for _, si := range expired {
|
|
acc.removeRespServiceImport(si, rsiTimeout)
|
|
}
|
|
|
|
// Pull out expired to determine if we have any left for timer.
|
|
totalResponses -= len(expired)
|
|
|
|
// Redo timer as needed.
|
|
acc.mu.Lock()
|
|
if totalResponses > 0 && se.rtmr != nil {
|
|
se.rtmr.Stop()
|
|
se.rtmr.Reset(se.respThresh)
|
|
} else {
|
|
se.clearResponseThresholdTimer()
|
|
}
|
|
acc.mu.Unlock()
|
|
}
|
|
|
|
// ServiceExportResponseThreshold returns the current threshold.
|
|
func (a *Account) ServiceExportResponseThreshold(export string) (time.Duration, error) {
|
|
a.mu.Lock()
|
|
defer a.mu.Unlock()
|
|
se := a.getServiceExport(export)
|
|
if se == nil {
|
|
return 0, fmt.Errorf("no export defined for %q", export)
|
|
}
|
|
return se.respThresh, nil
|
|
}
|
|
|
|
// SetServiceExportResponseThreshold sets the maximum time the system will a response to be delivered
|
|
// from a service export responder.
|
|
func (a *Account) SetServiceExportResponseThreshold(export string, maxTime time.Duration) error {
|
|
a.mu.Lock()
|
|
defer a.mu.Unlock()
|
|
if a.isClaimAccount() {
|
|
return fmt.Errorf("claim based accounts can not be updated directly")
|
|
}
|
|
lrt := a.lowestServiceExportResponseTime()
|
|
se := a.getServiceExport(export)
|
|
if se == nil {
|
|
return fmt.Errorf("no export defined for %q", export)
|
|
}
|
|
se.respThresh = maxTime
|
|
if nlrt := a.lowestServiceExportResponseTime(); nlrt != lrt {
|
|
a.updateAllClientsServiceExportResponseTime(nlrt)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// This is for internal service import responses.
|
|
func (a *Account) addRespServiceImport(dest *Account, to string, osi *serviceImport, tracking bool, header http.Header) *serviceImport {
|
|
nrr := string(osi.acc.newServiceReply(tracking))
|
|
|
|
a.mu.Lock()
|
|
rt := osi.rt
|
|
|
|
// dest is the requestor's account. a is the service responder with the export.
|
|
// Marked as internal here, that is how we distinguish.
|
|
si := &serviceImport{dest, nil, osi.se, nil, nrr, to, osi.to, 0, rt, nil, nil, nil, false, true, false, osi.share, false, nil}
|
|
|
|
if a.exports.responses == nil {
|
|
a.exports.responses = make(map[string]*serviceImport)
|
|
}
|
|
a.exports.responses[nrr] = si
|
|
|
|
// Always grab time and make sure response threshold timer is running.
|
|
si.ts = time.Now().UnixNano()
|
|
osi.se.setResponseThresholdTimer()
|
|
|
|
if rt == Singleton && tracking {
|
|
si.latency = osi.latency
|
|
si.tracking = true
|
|
si.trackingHdr = header
|
|
}
|
|
a.mu.Unlock()
|
|
|
|
// We do not do individual subscriptions here like we do on configured imports.
|
|
// We have an internal callback for all responses inbound to this account and
|
|
// will process appropriately there. This does not pollute the sublist and the caches.
|
|
|
|
// We do add in the reverse map such that we can detect loss of interest and do proper
|
|
// cleanup of this si as interest goes away.
|
|
dest.addReverseRespMapEntry(a, to, nrr)
|
|
|
|
return si
|
|
}
|
|
|
|
// AddStreamImportWithClaim will add in the stream import from a specific account with optional token.
|
|
func (a *Account) AddStreamImportWithClaim(account *Account, from, prefix string, imClaim *jwt.Import) error {
|
|
if account == nil {
|
|
return ErrMissingAccount
|
|
}
|
|
|
|
// First check to see if the account has authorized export of the subject.
|
|
if !account.checkStreamImportAuthorized(a, from, imClaim) {
|
|
return ErrStreamImportAuthorization
|
|
}
|
|
|
|
// Check prefix if it exists and make sure its a literal.
|
|
// Append token separator if not already present.
|
|
if prefix != "" {
|
|
// Make sure there are no wildcards here, this prefix needs to be a literal
|
|
// since it will be prepended to a publish subject.
|
|
if !subjectIsLiteral(prefix) {
|
|
return ErrStreamImportBadPrefix
|
|
}
|
|
if prefix[len(prefix)-1] != btsep {
|
|
prefix = prefix + string(btsep)
|
|
}
|
|
}
|
|
a.mu.Lock()
|
|
if a.isStreamImportDuplicate(account, from) {
|
|
a.mu.Unlock()
|
|
return ErrStreamImportDuplicate
|
|
}
|
|
a.imports.streams = append(a.imports.streams, &streamImport{account, from, prefix, imClaim, false})
|
|
a.mu.Unlock()
|
|
return nil
|
|
}
|
|
|
|
// isStreamImportDuplicate checks for duplicate.
|
|
// Lock should be held.
|
|
func (a *Account) isStreamImportDuplicate(acc *Account, from string) bool {
|
|
for _, si := range a.imports.streams {
|
|
if si.acc == acc && si.from == from {
|
|
return true
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
// AddStreamImport will add in the stream import from a specific account.
|
|
func (a *Account) AddStreamImport(account *Account, from, prefix string) error {
|
|
return a.AddStreamImportWithClaim(account, from, prefix, nil)
|
|
}
|
|
|
|
// IsPublicExport is a placeholder to denote a public export.
|
|
var IsPublicExport = []*Account(nil)
|
|
|
|
// AddStreamExport will add an export to the account. If accounts is nil
|
|
// it will signify a public export, meaning anyone can impoort.
|
|
func (a *Account) AddStreamExport(subject string, accounts []*Account) error {
|
|
if a == nil {
|
|
return ErrMissingAccount
|
|
}
|
|
|
|
a.mu.Lock()
|
|
defer a.mu.Unlock()
|
|
|
|
if a.exports.streams == nil {
|
|
a.exports.streams = make(map[string]*streamExport)
|
|
}
|
|
ea := a.exports.streams[subject]
|
|
if accounts != nil {
|
|
if ea == nil {
|
|
ea = &streamExport{}
|
|
}
|
|
// empty means auth required but will be import token.
|
|
if len(accounts) == 0 {
|
|
ea.tokenReq = true
|
|
} else {
|
|
if ea.approved == nil {
|
|
ea.approved = make(map[string]*Account, len(accounts))
|
|
}
|
|
for _, acc := range accounts {
|
|
ea.approved[acc.Name] = acc
|
|
}
|
|
}
|
|
}
|
|
a.exports.streams[subject] = ea
|
|
return nil
|
|
}
|
|
|
|
// Check if another account is authorized to import from us.
|
|
func (a *Account) checkStreamImportAuthorized(account *Account, subject string, imClaim *jwt.Import) bool {
|
|
// Find the subject in the exports list.
|
|
a.mu.RLock()
|
|
auth := a.checkStreamImportAuthorizedNoLock(account, subject, imClaim)
|
|
a.mu.RUnlock()
|
|
return auth
|
|
}
|
|
|
|
func (a *Account) checkStreamImportAuthorizedNoLock(account *Account, subject string, imClaim *jwt.Import) bool {
|
|
if a.exports.streams == nil || !IsValidSubject(subject) {
|
|
return false
|
|
}
|
|
return a.checkStreamExportApproved(account, subject, imClaim)
|
|
}
|
|
|
|
func (a *Account) checkAuth(ea *exportAuth, account *Account, imClaim *jwt.Import) bool {
|
|
// if ea is nil or ea.approved is nil, that denotes a public export
|
|
if ea == nil || (ea.approved == nil && !ea.tokenReq) {
|
|
return true
|
|
}
|
|
// Check if token required
|
|
if ea.tokenReq {
|
|
return a.checkActivation(account, imClaim, true)
|
|
}
|
|
// If we have a matching account we are authorized
|
|
_, ok := ea.approved[account.Name]
|
|
return ok
|
|
}
|
|
|
|
func (a *Account) checkStreamExportApproved(account *Account, subject string, imClaim *jwt.Import) bool {
|
|
// Check direct match of subject first
|
|
ea, ok := a.exports.streams[subject]
|
|
if ok {
|
|
if ea == nil {
|
|
return true
|
|
}
|
|
return a.checkAuth(&ea.exportAuth, account, imClaim)
|
|
}
|
|
// ok if we are here we did not match directly so we need to test each one.
|
|
// The import subject arg has to take precedence, meaning the export
|
|
// has to be a true subset of the import claim. We already checked for
|
|
// exact matches above.
|
|
tokens := strings.Split(subject, tsep)
|
|
for subj, ea := range a.exports.streams {
|
|
if isSubsetMatch(tokens, subj) {
|
|
if ea == nil {
|
|
return true
|
|
}
|
|
return a.checkAuth(&ea.exportAuth, account, imClaim)
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
func (a *Account) checkServiceExportApproved(account *Account, subject string, imClaim *jwt.Import) bool {
|
|
// Check direct match of subject first
|
|
se, ok := a.exports.services[subject]
|
|
if ok {
|
|
// if se is nil or eq.approved is nil, that denotes a public export
|
|
if se == nil || (se.approved == nil && !se.tokenReq) {
|
|
return true
|
|
}
|
|
// Check if token required
|
|
if se.tokenReq {
|
|
return a.checkActivation(account, imClaim, true)
|
|
}
|
|
// If we have a matching account we are authorized
|
|
_, ok := se.approved[account.Name]
|
|
return ok
|
|
}
|
|
// ok if we are here we did not match directly so we need to test each one.
|
|
// The import subject arg has to take precedence, meaning the export
|
|
// has to be a true subset of the import claim. We already checked for
|
|
// exact matches above.
|
|
tokens := strings.Split(subject, tsep)
|
|
for subj, se := range a.exports.services {
|
|
if isSubsetMatch(tokens, subj) {
|
|
if se == nil || (se.approved == nil && !se.tokenReq) {
|
|
return true
|
|
}
|
|
// Check if token required
|
|
if se.tokenReq {
|
|
return a.checkActivation(account, imClaim, true)
|
|
}
|
|
_, ok := se.approved[account.Name]
|
|
return ok
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
// Helper function to get a serviceExport.
|
|
// Lock should be held on entry.
|
|
func (a *Account) getServiceExport(subj string) *serviceExport {
|
|
se, ok := a.exports.services[subj]
|
|
// The export probably has a wildcard, so lookup that up.
|
|
if !ok {
|
|
se = a.getWildcardServiceExport(subj)
|
|
}
|
|
return se
|
|
}
|
|
|
|
// This helper is used when trying to match a serviceExport record that is
|
|
// represented by a wildcard.
|
|
// Lock should be held on entry.
|
|
func (a *Account) getWildcardServiceExport(from string) *serviceExport {
|
|
tokens := strings.Split(from, tsep)
|
|
for subj, se := range a.exports.services {
|
|
if isSubsetMatch(tokens, subj) {
|
|
return se
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Will fetch the activation token for an import.
|
|
func fetchActivation(url string) string {
|
|
// FIXME(dlc) - Make configurable.
|
|
c := &http.Client{Timeout: 2 * time.Second}
|
|
resp, err := c.Get(url)
|
|
if err != nil || resp == nil {
|
|
return ""
|
|
}
|
|
defer resp.Body.Close()
|
|
body, err := ioutil.ReadAll(resp.Body)
|
|
if err != nil {
|
|
return ""
|
|
}
|
|
return string(body)
|
|
}
|
|
|
|
// These are import stream specific versions for when an activation expires.
|
|
func (a *Account) streamActivationExpired(exportAcc *Account, subject string) {
|
|
a.mu.RLock()
|
|
if a.expired || a.imports.streams == nil {
|
|
a.mu.RUnlock()
|
|
return
|
|
}
|
|
var si *streamImport
|
|
for _, si = range a.imports.streams {
|
|
if si.acc == exportAcc && si.from == subject {
|
|
break
|
|
}
|
|
}
|
|
|
|
if si == nil || si.invalid {
|
|
a.mu.RUnlock()
|
|
return
|
|
}
|
|
a.mu.RUnlock()
|
|
|
|
if si.acc.checkActivation(a, si.claim, false) {
|
|
// The token has been updated most likely and we are good to go.
|
|
return
|
|
}
|
|
|
|
a.mu.Lock()
|
|
si.invalid = true
|
|
clients := make([]*client, 0, len(a.clients))
|
|
for c := range a.clients {
|
|
clients = append(clients, c)
|
|
}
|
|
awcsti := map[string]struct{}{a.Name: {}}
|
|
a.mu.Unlock()
|
|
for _, c := range clients {
|
|
c.processSubsOnConfigReload(awcsti)
|
|
}
|
|
}
|
|
|
|
// These are import service specific versions for when an activation expires.
|
|
func (a *Account) serviceActivationExpired(subject string) {
|
|
a.mu.RLock()
|
|
if a.expired || a.imports.services == nil {
|
|
a.mu.RUnlock()
|
|
return
|
|
}
|
|
si := a.imports.services[subject]
|
|
if si == nil || si.invalid {
|
|
a.mu.RUnlock()
|
|
return
|
|
}
|
|
a.mu.RUnlock()
|
|
|
|
if si.acc.checkActivation(a, si.claim, false) {
|
|
// The token has been updated most likely and we are good to go.
|
|
return
|
|
}
|
|
|
|
a.mu.Lock()
|
|
si.invalid = true
|
|
a.mu.Unlock()
|
|
}
|
|
|
|
// Fires for expired activation tokens. We could track this with timers etc.
|
|
// Instead we just re-analyze where we are and if we need to act.
|
|
func (a *Account) activationExpired(exportAcc *Account, subject string, kind jwt.ExportType) {
|
|
switch kind {
|
|
case jwt.Stream:
|
|
a.streamActivationExpired(exportAcc, subject)
|
|
case jwt.Service:
|
|
a.serviceActivationExpired(subject)
|
|
}
|
|
}
|
|
|
|
// checkActivation will check the activation token for validity.
|
|
func (a *Account) checkActivation(importAcc *Account, claim *jwt.Import, expTimer bool) bool {
|
|
if claim == nil || claim.Token == "" {
|
|
return false
|
|
}
|
|
// Create a quick clone so we can inline Token JWT.
|
|
clone := *claim
|
|
|
|
// We grab the token from a URL by hand here since we need expiration etc.
|
|
if url, err := url.Parse(clone.Token); err == nil && url.Scheme != "" {
|
|
clone.Token = fetchActivation(url.String())
|
|
}
|
|
vr := jwt.CreateValidationResults()
|
|
clone.Validate(a.Name, vr)
|
|
if vr.IsBlocking(true) {
|
|
return false
|
|
}
|
|
act, err := jwt.DecodeActivationClaims(clone.Token)
|
|
if err != nil {
|
|
return false
|
|
}
|
|
if !a.isIssuerClaimTrusted(act) {
|
|
return false
|
|
}
|
|
vr = jwt.CreateValidationResults()
|
|
act.Validate(vr)
|
|
if vr.IsBlocking(true) {
|
|
return false
|
|
}
|
|
if act.Expires != 0 {
|
|
tn := time.Now().Unix()
|
|
if act.Expires <= tn {
|
|
return false
|
|
}
|
|
if expTimer {
|
|
expiresAt := time.Duration(act.Expires - tn)
|
|
time.AfterFunc(expiresAt*time.Second, func() {
|
|
importAcc.activationExpired(a, string(act.ImportSubject), claim.Type)
|
|
})
|
|
}
|
|
}
|
|
// Check for token revocation..
|
|
if a.actsRevoked != nil {
|
|
if t, ok := a.actsRevoked[act.Subject]; ok && t <= time.Now().Unix() {
|
|
return false
|
|
}
|
|
}
|
|
|
|
return true
|
|
}
|
|
|
|
// Returns true if the activation claim is trusted. That is the issuer matches
|
|
// the account or is an entry in the signing keys.
|
|
func (a *Account) isIssuerClaimTrusted(claims *jwt.ActivationClaims) bool {
|
|
// if no issuer account, issuer is the account
|
|
if claims.IssuerAccount == "" {
|
|
return true
|
|
}
|
|
// If the IssuerAccount is not us, then this is considered an error.
|
|
if a.Name != claims.IssuerAccount {
|
|
if a.srv != nil {
|
|
a.srv.Errorf("Invalid issuer account %q in activation claim (subject: %q - type: %q) for account %q",
|
|
claims.IssuerAccount, claims.Activation.ImportSubject, claims.Activation.ImportType, a.Name)
|
|
}
|
|
return false
|
|
}
|
|
return a.hasIssuerNoLock(claims.Issuer)
|
|
}
|
|
|
|
// Returns true if `a` and `b` stream imports are the same. Note that the
|
|
// check is done with the account's name, not the pointer. This is used
|
|
// during config reload where we are comparing current and new config
|
|
// in which pointers are different.
|
|
// No lock is acquired in this function, so it is assumed that the
|
|
// import maps are not changed while this executes.
|
|
func (a *Account) checkStreamImportsEqual(b *Account) bool {
|
|
if len(a.imports.streams) != len(b.imports.streams) {
|
|
return false
|
|
}
|
|
// Load the b imports into a map index by what we are looking for.
|
|
bm := make(map[string]*streamImport, len(b.imports.streams))
|
|
for _, bim := range b.imports.streams {
|
|
bm[bim.acc.Name+bim.from+bim.prefix] = bim
|
|
}
|
|
for _, aim := range a.imports.streams {
|
|
if _, ok := bm[aim.acc.Name+aim.from+aim.prefix]; !ok {
|
|
return false
|
|
}
|
|
}
|
|
return true
|
|
}
|
|
|
|
func (a *Account) checkStreamExportsEqual(b *Account) bool {
|
|
if len(a.exports.streams) != len(b.exports.streams) {
|
|
return false
|
|
}
|
|
for subj, aea := range a.exports.streams {
|
|
bea, ok := b.exports.streams[subj]
|
|
if !ok {
|
|
return false
|
|
}
|
|
if !reflect.DeepEqual(aea, bea) {
|
|
return false
|
|
}
|
|
}
|
|
return true
|
|
}
|
|
|
|
func (a *Account) checkServiceExportsEqual(b *Account) bool {
|
|
if len(a.exports.services) != len(b.exports.services) {
|
|
return false
|
|
}
|
|
for subj, aea := range a.exports.services {
|
|
bea, ok := b.exports.services[subj]
|
|
if !ok {
|
|
return false
|
|
}
|
|
if !reflect.DeepEqual(aea, bea) {
|
|
return false
|
|
}
|
|
}
|
|
return true
|
|
}
|
|
|
|
// Check if another account is authorized to route requests to this service.
|
|
func (a *Account) checkServiceImportAuthorized(account *Account, subject string, imClaim *jwt.Import) bool {
|
|
a.mu.RLock()
|
|
authorized := a.checkServiceImportAuthorizedNoLock(account, subject, imClaim)
|
|
a.mu.RUnlock()
|
|
return authorized
|
|
}
|
|
|
|
// Check if another account is authorized to route requests to this service.
|
|
func (a *Account) checkServiceImportAuthorizedNoLock(account *Account, subject string, imClaim *jwt.Import) bool {
|
|
// Find the subject in the services list.
|
|
if a.exports.services == nil {
|
|
return false
|
|
}
|
|
return a.checkServiceExportApproved(account, subject, imClaim)
|
|
}
|
|
|
|
// IsExpired returns expiration status.
|
|
func (a *Account) IsExpired() bool {
|
|
a.mu.RLock()
|
|
exp := a.expired
|
|
a.mu.RUnlock()
|
|
return exp
|
|
}
|
|
|
|
// Called when an account has expired.
|
|
func (a *Account) expiredTimeout() {
|
|
// Mark expired first.
|
|
a.mu.Lock()
|
|
a.expired = true
|
|
a.mu.Unlock()
|
|
|
|
// Collect the clients and expire them.
|
|
cs := make([]*client, 0, len(a.clients))
|
|
a.mu.RLock()
|
|
for c := range a.clients {
|
|
cs = append(cs, c)
|
|
}
|
|
a.mu.RUnlock()
|
|
|
|
for _, c := range cs {
|
|
c.accountAuthExpired()
|
|
}
|
|
}
|
|
|
|
// Sets the expiration timer for an account JWT that has it set.
|
|
func (a *Account) setExpirationTimer(d time.Duration) {
|
|
a.etmr = time.AfterFunc(d, a.expiredTimeout)
|
|
}
|
|
|
|
// Lock should be held
|
|
func (a *Account) clearExpirationTimer() bool {
|
|
if a.etmr == nil {
|
|
return true
|
|
}
|
|
stopped := a.etmr.Stop()
|
|
a.etmr = nil
|
|
return stopped
|
|
}
|
|
|
|
// checkUserRevoked will check if a user has been revoked.
|
|
func (a *Account) checkUserRevoked(nkey string, issuedAt int64) bool {
|
|
a.mu.RLock()
|
|
defer a.mu.RUnlock()
|
|
if a.usersRevoked == nil {
|
|
return false
|
|
}
|
|
if t, ok := a.usersRevoked[nkey]; !ok || t < issuedAt {
|
|
return false
|
|
}
|
|
return true
|
|
}
|
|
|
|
// Check expiration and set the proper state as needed.
|
|
func (a *Account) checkExpiration(claims *jwt.ClaimsData) {
|
|
a.mu.Lock()
|
|
defer a.mu.Unlock()
|
|
|
|
a.clearExpirationTimer()
|
|
if claims.Expires == 0 {
|
|
a.expired = false
|
|
return
|
|
}
|
|
tn := time.Now().Unix()
|
|
if claims.Expires <= tn {
|
|
a.expired = true
|
|
return
|
|
}
|
|
expiresAt := time.Duration(claims.Expires - tn)
|
|
a.setExpirationTimer(expiresAt * time.Second)
|
|
a.expired = false
|
|
}
|
|
|
|
// hasIssuer returns true if the issuer matches the account
|
|
// issuer or it is a signing key for the account.
|
|
func (a *Account) hasIssuer(issuer string) bool {
|
|
a.mu.RLock()
|
|
hi := a.hasIssuerNoLock(issuer)
|
|
a.mu.RUnlock()
|
|
return hi
|
|
}
|
|
|
|
// hasIssuerNoLock is the unlocked version of hasIssuer
|
|
func (a *Account) hasIssuerNoLock(issuer string) bool {
|
|
// same issuer
|
|
if a.Issuer == issuer {
|
|
return true
|
|
}
|
|
for i := 0; i < len(a.signingKeys); i++ {
|
|
if a.signingKeys[i] == issuer {
|
|
return true
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
// Returns the loop detection subject used for leafnodes
|
|
func (a *Account) getLDSubject() string {
|
|
a.mu.RLock()
|
|
lds := a.lds
|
|
a.mu.RUnlock()
|
|
return lds
|
|
}
|
|
|
|
// Placeholder for signaling token auth required.
|
|
var tokenAuthReq = []*Account{}
|
|
|
|
func authAccounts(tokenReq bool) []*Account {
|
|
if tokenReq {
|
|
return tokenAuthReq
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// SetAccountResolver will assign the account resolver.
|
|
func (s *Server) SetAccountResolver(ar AccountResolver) {
|
|
s.mu.Lock()
|
|
s.accResolver = ar
|
|
s.mu.Unlock()
|
|
}
|
|
|
|
// AccountResolver returns the registered account resolver.
|
|
func (s *Server) AccountResolver() AccountResolver {
|
|
s.mu.Lock()
|
|
ar := s.accResolver
|
|
s.mu.Unlock()
|
|
return ar
|
|
}
|
|
|
|
// isClaimAccount returns if this account is backed by a JWT claim.
|
|
// Lock should be held.
|
|
func (a *Account) isClaimAccount() bool {
|
|
return a.claimJWT != ""
|
|
}
|
|
|
|
// updateAccountClaims will update an existing account with new claims.
|
|
// This will replace any exports or imports previously defined.
|
|
// Lock MUST NOT be held upon entry.
|
|
func (s *Server) UpdateAccountClaims(a *Account, ac *jwt.AccountClaims) {
|
|
s.updateAccountClaimsWithRefresh(a, ac, true)
|
|
}
|
|
|
|
// updateAccountClaimsWithRefresh will update an existing account with new claims.
|
|
// If refreshImportingAccounts is true it will also update incomplete dependent accounts
|
|
// This will replace any exports or imports previously defined.
|
|
// Lock MUST NOT be held upon entry.
|
|
func (s *Server) updateAccountClaimsWithRefresh(a *Account, ac *jwt.AccountClaims, refreshImportingAccounts bool) {
|
|
if a == nil {
|
|
return
|
|
}
|
|
s.Debugf("Updating account claims: %s", a.Name)
|
|
a.checkExpiration(ac.Claims())
|
|
|
|
a.mu.Lock()
|
|
// Clone to update, only select certain fields.
|
|
old := &Account{Name: a.Name, exports: a.exports, limits: a.limits, signingKeys: a.signingKeys}
|
|
|
|
// Reset exports and imports here.
|
|
|
|
// Exports is creating a whole new map.
|
|
a.exports = exportMap{}
|
|
|
|
// Imports are checked unlocked in processInbound, so we can't change out the struct here. Need to process inline.
|
|
if a.imports.streams != nil {
|
|
old.imports.streams = a.imports.streams
|
|
a.imports.streams = nil
|
|
}
|
|
if a.imports.services != nil {
|
|
old.imports.services = make(map[string]*serviceImport, len(a.imports.services))
|
|
}
|
|
for k, v := range a.imports.services {
|
|
old.imports.services[k] = v
|
|
delete(a.imports.services, k)
|
|
}
|
|
|
|
// Reset any notion of export revocations.
|
|
a.actsRevoked = nil
|
|
|
|
// update account signing keys
|
|
a.signingKeys = nil
|
|
signersChanged := false
|
|
if len(ac.SigningKeys) > 0 {
|
|
// insure copy the new keys and sort
|
|
a.signingKeys = append(a.signingKeys, ac.SigningKeys...)
|
|
sort.Strings(a.signingKeys)
|
|
}
|
|
if len(a.signingKeys) != len(old.signingKeys) {
|
|
signersChanged = true
|
|
} else {
|
|
for i := 0; i < len(old.signingKeys); i++ {
|
|
if a.signingKeys[i] != old.signingKeys[i] {
|
|
signersChanged = true
|
|
break
|
|
}
|
|
}
|
|
}
|
|
a.mu.Unlock()
|
|
|
|
gatherClients := func() []*client {
|
|
a.mu.RLock()
|
|
clients := make([]*client, 0, len(a.clients))
|
|
for c := range a.clients {
|
|
clients = append(clients, c)
|
|
}
|
|
a.mu.RUnlock()
|
|
return clients
|
|
}
|
|
|
|
jsEnabled := s.JetStreamEnabled()
|
|
if jsEnabled && a == s.SystemAccount() {
|
|
for _, export := range allJsExports {
|
|
s.Debugf("Adding jetstream service export %q for %s", export, a.Name)
|
|
if err := a.AddServiceExport(export, nil); err != nil {
|
|
s.Errorf("Error setting up jetstream service exports: %v", err)
|
|
}
|
|
}
|
|
}
|
|
|
|
for _, e := range ac.Exports {
|
|
switch e.Type {
|
|
case jwt.Stream:
|
|
s.Debugf("Adding stream export %q for %s", e.Subject, a.Name)
|
|
if err := a.AddStreamExport(string(e.Subject), authAccounts(e.TokenReq)); err != nil {
|
|
s.Debugf("Error adding stream export to account [%s]: %v", a.Name, err.Error())
|
|
}
|
|
case jwt.Service:
|
|
s.Debugf("Adding service export %q for %s", e.Subject, a.Name)
|
|
rt := Singleton
|
|
switch e.ResponseType {
|
|
case jwt.ResponseTypeStream:
|
|
rt = Streamed
|
|
case jwt.ResponseTypeChunked:
|
|
rt = Chunked
|
|
}
|
|
if err := a.AddServiceExportWithResponse(string(e.Subject), rt, authAccounts(e.TokenReq)); err != nil {
|
|
s.Debugf("Error adding service export to account [%s]: %v", a.Name, err)
|
|
}
|
|
if e.Latency != nil {
|
|
if err := a.TrackServiceExportWithSampling(string(e.Subject), string(e.Latency.Results), e.Latency.Sampling); err != nil {
|
|
s.Debugf("Error adding latency tracking for service export to account [%s]: %v", a.Name, err)
|
|
}
|
|
}
|
|
}
|
|
// We will track these at the account level. Should not have any collisions.
|
|
if e.Revocations != nil {
|
|
a.mu.Lock()
|
|
if a.actsRevoked == nil {
|
|
a.actsRevoked = make(map[string]int64)
|
|
}
|
|
for k, t := range e.Revocations {
|
|
a.actsRevoked[k] = t
|
|
}
|
|
a.mu.Unlock()
|
|
}
|
|
}
|
|
var incompleteImports []*jwt.Import
|
|
for _, i := range ac.Imports {
|
|
// check tmpAccounts with priority
|
|
var acc *Account
|
|
var err error
|
|
if v, ok := s.tmpAccounts.Load(i.Account); ok {
|
|
acc = v.(*Account)
|
|
} else {
|
|
acc, err = s.lookupAccount(i.Account)
|
|
}
|
|
if acc == nil || err != nil {
|
|
s.Errorf("Can't locate account [%s] for import of [%v] %s (err=%v)", i.Account, i.Subject, i.Type, err)
|
|
incompleteImports = append(incompleteImports, i)
|
|
continue
|
|
}
|
|
switch i.Type {
|
|
case jwt.Stream:
|
|
s.Debugf("Adding stream import %s:%q for %s:%q", acc.Name, i.Subject, a.Name, i.To)
|
|
if err := a.AddStreamImportWithClaim(acc, string(i.Subject), string(i.To), i); err != nil {
|
|
s.Debugf("Error adding stream import to account [%s]: %v", a.Name, err.Error())
|
|
incompleteImports = append(incompleteImports, i)
|
|
}
|
|
case jwt.Service:
|
|
// FIXME(dlc) - need to add in respThresh here eventually.
|
|
s.Debugf("Adding service import %s:%q for %s:%q", acc.Name, i.Subject, a.Name, i.To)
|
|
if err := a.AddServiceImportWithClaim(acc, string(i.Subject), string(i.To), i); err != nil {
|
|
s.Debugf("Error adding service import to account [%s]: %v", a.Name, err.Error())
|
|
incompleteImports = append(incompleteImports, i)
|
|
}
|
|
}
|
|
}
|
|
// Now let's apply any needed changes from import/export changes.
|
|
if !a.checkStreamImportsEqual(old) {
|
|
awcsti := map[string]struct{}{a.Name: {}}
|
|
for _, c := range gatherClients() {
|
|
c.processSubsOnConfigReload(awcsti)
|
|
}
|
|
}
|
|
// Now check if stream exports have changed.
|
|
if !a.checkStreamExportsEqual(old) || signersChanged {
|
|
clients := map[*client]struct{}{}
|
|
// We need to check all accounts that have an import claim from this account.
|
|
awcsti := map[string]struct{}{}
|
|
s.accounts.Range(func(k, v interface{}) bool {
|
|
acc := v.(*Account)
|
|
// Move to the next if this account is actually account "a".
|
|
if acc.Name == a.Name {
|
|
return true
|
|
}
|
|
// TODO: checkStreamImportAuthorized() stack should not be trying
|
|
// to lock "acc". If we find that to be needed, we will need to
|
|
// rework this to ensure we don't lock acc.
|
|
acc.mu.Lock()
|
|
for _, im := range acc.imports.streams {
|
|
if im != nil && im.acc.Name == a.Name {
|
|
// Check for if we are still authorized for an import.
|
|
im.invalid = !a.checkStreamImportAuthorized(acc, im.from, im.claim)
|
|
awcsti[acc.Name] = struct{}{}
|
|
for c := range acc.clients {
|
|
clients[c] = struct{}{}
|
|
}
|
|
}
|
|
}
|
|
acc.mu.Unlock()
|
|
return true
|
|
})
|
|
// Now walk clients.
|
|
for c := range clients {
|
|
c.processSubsOnConfigReload(awcsti)
|
|
}
|
|
}
|
|
// Now check if service exports have changed.
|
|
if !a.checkServiceExportsEqual(old) || signersChanged {
|
|
s.accounts.Range(func(k, v interface{}) bool {
|
|
acc := v.(*Account)
|
|
// Move to the next if this account is actually account "a".
|
|
if acc.Name == a.Name {
|
|
return true
|
|
}
|
|
// TODO: checkServiceImportAuthorized() stack should not be trying
|
|
// to lock "acc". If we find that to be needed, we will need to
|
|
// rework this to ensure we don't lock acc.
|
|
acc.mu.Lock()
|
|
for _, si := range acc.imports.services {
|
|
if si != nil && si.acc.Name == a.Name {
|
|
// Check for if we are still authorized for an import.
|
|
si.invalid = !a.checkServiceImportAuthorized(acc, si.to, si.claim)
|
|
if si.latency != nil && !si.response {
|
|
// Make sure we should still be tracking latency.
|
|
if se := a.getServiceExport(si.to); se != nil {
|
|
si.latency = se.latency
|
|
}
|
|
}
|
|
}
|
|
}
|
|
acc.mu.Unlock()
|
|
return true
|
|
})
|
|
}
|
|
|
|
// Now make sure we shutdown the old service import subscriptions.
|
|
var sids [][]byte
|
|
a.mu.RLock()
|
|
c := a.ic
|
|
for _, si := range old.imports.services {
|
|
if c != nil && si.sid != nil {
|
|
sids = append(sids, si.sid)
|
|
}
|
|
}
|
|
a.mu.RUnlock()
|
|
for _, sid := range sids {
|
|
c.processUnsub(sid)
|
|
}
|
|
|
|
// Now do limits if they are present.
|
|
a.mu.Lock()
|
|
a.msubs = int32(ac.Limits.Subs)
|
|
a.mpay = int32(ac.Limits.Payload)
|
|
a.mconns = int32(ac.Limits.Conn)
|
|
a.mleafs = int32(ac.Limits.LeafNodeConn)
|
|
// Check for any revocations
|
|
if len(ac.Revocations) > 0 {
|
|
// We will always replace whatever we had with most current, so no
|
|
// need to look at what we have.
|
|
a.usersRevoked = make(map[string]int64, len(ac.Revocations))
|
|
for pk, t := range ac.Revocations {
|
|
a.usersRevoked[pk] = t
|
|
}
|
|
} else {
|
|
a.usersRevoked = nil
|
|
}
|
|
a.defaultPerms = buildPermissionsFromJwt(&ac.DefaultPermissions)
|
|
a.incomplete = len(incompleteImports) != 0
|
|
for _, i := range incompleteImports {
|
|
s.incompleteAccExporterMap.Store(i.Account, struct{}{})
|
|
}
|
|
if a.srv == nil {
|
|
a.srv = s
|
|
}
|
|
if jsEnabled {
|
|
if ac.Limits.JetStreamLimits.DiskStorage != 0 || ac.Limits.JetStreamLimits.MemoryStorage != 0 {
|
|
// JetStreamAccountLimits and jwt.JetStreamLimits use same value for unlimited
|
|
a.jsLimits = &JetStreamAccountLimits{
|
|
MaxMemory: ac.Limits.JetStreamLimits.MemoryStorage,
|
|
MaxStore: ac.Limits.JetStreamLimits.DiskStorage,
|
|
MaxStreams: int(ac.Limits.JetStreamLimits.Streams),
|
|
MaxConsumers: int(ac.Limits.JetStreamLimits.Consumer),
|
|
}
|
|
} else if a.jsLimits != nil {
|
|
// covers failed update followed by disable
|
|
a.jsLimits = nil
|
|
}
|
|
}
|
|
a.updated = time.Now()
|
|
a.mu.Unlock()
|
|
|
|
clients := gatherClients()
|
|
// Sort if we are over the limit.
|
|
if a.MaxTotalConnectionsReached() {
|
|
sort.Slice(clients, func(i, j int) bool {
|
|
return clients[i].start.After(clients[j].start)
|
|
})
|
|
}
|
|
|
|
if jsEnabled {
|
|
if err := s.configJetStream(a); err != nil {
|
|
s.Errorf("Error configuring jetstream for account [%s]: %v", a.Name, err.Error())
|
|
a.mu.Lock()
|
|
// Absent reload of js server cfg, this is going to be broken until js is disabled
|
|
a.incomplete = true
|
|
a.mu.Unlock()
|
|
}
|
|
}
|
|
|
|
for i, c := range clients {
|
|
a.mu.RLock()
|
|
exceeded := a.mconns != jwt.NoLimit && i >= int(a.mconns)
|
|
a.mu.RUnlock()
|
|
if exceeded {
|
|
c.maxAccountConnExceeded()
|
|
continue
|
|
}
|
|
c.mu.Lock()
|
|
c.applyAccountLimits()
|
|
theJWT := c.opts.JWT
|
|
c.mu.Unlock()
|
|
// Check for being revoked here. We use ac one to avoid the account lock.
|
|
if ac.Revocations != nil && theJWT != "" {
|
|
if juc, err := jwt.DecodeUserClaims(theJWT); err != nil {
|
|
c.Debugf("User JWT not valid: %v", err)
|
|
c.authViolation()
|
|
continue
|
|
} else if ok := ac.IsClaimRevoked(juc); ok {
|
|
c.sendErrAndDebug("User Authentication Revoked")
|
|
c.closeConnection(Revocation)
|
|
continue
|
|
}
|
|
}
|
|
}
|
|
|
|
// Check if the signing keys changed, might have to evict
|
|
if signersChanged {
|
|
for _, c := range clients {
|
|
c.mu.Lock()
|
|
sk := c.user.SigningKey
|
|
c.mu.Unlock()
|
|
if sk != "" && !a.hasIssuer(sk) {
|
|
c.closeConnection(AuthenticationViolation)
|
|
}
|
|
}
|
|
}
|
|
|
|
if _, ok := s.incompleteAccExporterMap.Load(old.Name); ok && refreshImportingAccounts {
|
|
s.incompleteAccExporterMap.Delete(old.Name)
|
|
s.accounts.Range(func(key, value interface{}) bool {
|
|
acc := value.(*Account)
|
|
acc.mu.RLock()
|
|
incomplete := acc.incomplete
|
|
name := acc.Name
|
|
// Must use jwt in account or risk failing on fetch
|
|
// This jwt may not be the same that caused exportingAcc to be in incompleteAccExporterMap
|
|
claimJWT := acc.claimJWT
|
|
acc.mu.RUnlock()
|
|
if incomplete && name != old.Name {
|
|
if accClaims, _, err := s.verifyAccountClaims(claimJWT); err == nil {
|
|
// Since claimJWT has not changed, acc can become complete
|
|
// but it won't alter incomplete for it's dependents accounts.
|
|
s.updateAccountClaimsWithRefresh(acc, accClaims, false)
|
|
// old.Name was deleted before ranging over accounts
|
|
// If it exists again, UpdateAccountClaims set it for failed imports of acc.
|
|
// So there was one import of acc that imported this account and failed again.
|
|
// Since this account just got updated, the import itself may be in error. So trace that.
|
|
if _, ok := s.incompleteAccExporterMap.Load(old.Name); ok {
|
|
s.incompleteAccExporterMap.Delete(old.Name)
|
|
s.Errorf("Account %s has issues importing account %s", name, old.Name)
|
|
}
|
|
}
|
|
}
|
|
return true
|
|
})
|
|
}
|
|
}
|
|
|
|
// Helper to build an internal account structure from a jwt.AccountClaims.
|
|
// Lock MUST NOT be held upon entry.
|
|
func (s *Server) buildInternalAccount(ac *jwt.AccountClaims) *Account {
|
|
acc := NewAccount(ac.Subject)
|
|
acc.Issuer = ac.Issuer
|
|
// Set this here since we are placing in s.tmpAccounts below and may be
|
|
// referenced by an route RS+, etc.
|
|
s.setAccountSublist(acc)
|
|
|
|
// We don't want to register an account that is in the process of
|
|
// being built, however, to solve circular import dependencies, we
|
|
// need to store it here.
|
|
s.tmpAccounts.Store(ac.Subject, acc)
|
|
s.UpdateAccountClaims(acc, ac)
|
|
return acc
|
|
}
|
|
|
|
// Helper to build Permissions from jwt.Permissions
|
|
// or return nil if none were specified
|
|
func buildPermissionsFromJwt(uc *jwt.Permissions) *Permissions {
|
|
if uc == nil {
|
|
return nil
|
|
}
|
|
var p *Permissions
|
|
if len(uc.Pub.Allow) > 0 || len(uc.Pub.Deny) > 0 {
|
|
if p == nil {
|
|
p = &Permissions{}
|
|
}
|
|
p.Publish = &SubjectPermission{}
|
|
p.Publish.Allow = uc.Pub.Allow
|
|
p.Publish.Deny = uc.Pub.Deny
|
|
}
|
|
if len(uc.Sub.Allow) > 0 || len(uc.Sub.Deny) > 0 {
|
|
if p == nil {
|
|
p = &Permissions{}
|
|
}
|
|
p.Subscribe = &SubjectPermission{}
|
|
p.Subscribe.Allow = uc.Sub.Allow
|
|
p.Subscribe.Deny = uc.Sub.Deny
|
|
}
|
|
if uc.Resp != nil {
|
|
if p == nil {
|
|
p = &Permissions{}
|
|
}
|
|
p.Response = &ResponsePermission{
|
|
MaxMsgs: uc.Resp.MaxMsgs,
|
|
Expires: uc.Resp.Expires,
|
|
}
|
|
validateResponsePermissions(p)
|
|
}
|
|
return p
|
|
}
|
|
|
|
// Helper to build internal NKeyUser.
|
|
func buildInternalNkeyUser(uc *jwt.UserClaims, acts map[string]struct{}, acc *Account) *NkeyUser {
|
|
nu := &NkeyUser{Nkey: uc.Subject, Account: acc, AllowedConnectionTypes: acts}
|
|
if uc.IssuerAccount != "" {
|
|
nu.SigningKey = uc.Issuer
|
|
}
|
|
|
|
// Now check for permissions.
|
|
var p = buildPermissionsFromJwt(&uc.Permissions)
|
|
if p == nil && acc.defaultPerms != nil {
|
|
p = acc.defaultPerms.clone()
|
|
}
|
|
nu.Permissions = p
|
|
return nu
|
|
}
|
|
|
|
const fetchTimeout = 2 * time.Second
|
|
|
|
// AccountResolver interface. This is to fetch Account JWTs by public nkeys
|
|
type AccountResolver interface {
|
|
Fetch(name string) (string, error)
|
|
Store(name, jwt string) error
|
|
IsReadOnly() bool
|
|
Start(server *Server) error
|
|
IsTrackingUpdate() bool
|
|
Reload() error
|
|
Close()
|
|
}
|
|
|
|
// Default implementations of IsReadOnly/Start so only need to be written when changed
|
|
type resolverDefaultsOpsImpl struct{}
|
|
|
|
func (*resolverDefaultsOpsImpl) IsReadOnly() bool {
|
|
return true
|
|
}
|
|
|
|
func (*resolverDefaultsOpsImpl) IsTrackingUpdate() bool {
|
|
return false
|
|
}
|
|
|
|
func (*resolverDefaultsOpsImpl) Start(*Server) error {
|
|
return nil
|
|
}
|
|
|
|
func (*resolverDefaultsOpsImpl) Reload() error {
|
|
return nil
|
|
}
|
|
|
|
func (*resolverDefaultsOpsImpl) Close() {
|
|
}
|
|
|
|
func (*resolverDefaultsOpsImpl) Store(_, _ string) error {
|
|
return fmt.Errorf("Store operation not supported for URL Resolver")
|
|
}
|
|
|
|
// MemAccResolver is a memory only resolver.
|
|
// Mostly for testing.
|
|
type MemAccResolver struct {
|
|
sm sync.Map
|
|
resolverDefaultsOpsImpl
|
|
}
|
|
|
|
// Fetch will fetch the account jwt claims from the internal sync.Map.
|
|
func (m *MemAccResolver) Fetch(name string) (string, error) {
|
|
if j, ok := m.sm.Load(name); ok {
|
|
return j.(string), nil
|
|
}
|
|
return _EMPTY_, ErrMissingAccount
|
|
}
|
|
|
|
// Store will store the account jwt claims in the internal sync.Map.
|
|
func (m *MemAccResolver) Store(name, jwt string) error {
|
|
m.sm.Store(name, jwt)
|
|
return nil
|
|
}
|
|
|
|
func (ur *MemAccResolver) IsReadOnly() bool {
|
|
return false
|
|
}
|
|
|
|
// URLAccResolver implements an http fetcher.
|
|
type URLAccResolver struct {
|
|
url string
|
|
c *http.Client
|
|
resolverDefaultsOpsImpl
|
|
}
|
|
|
|
// NewURLAccResolver returns a new resolver for the given base URL.
|
|
func NewURLAccResolver(url string) (*URLAccResolver, error) {
|
|
if !strings.HasSuffix(url, "/") {
|
|
url += "/"
|
|
}
|
|
// FIXME(dlc) - Make timeout and others configurable.
|
|
// We create our own transport to amortize TLS.
|
|
tr := &http.Transport{
|
|
MaxIdleConns: 10,
|
|
IdleConnTimeout: 30 * time.Second,
|
|
}
|
|
ur := &URLAccResolver{
|
|
url: url,
|
|
c: &http.Client{Timeout: fetchTimeout, Transport: tr},
|
|
}
|
|
return ur, nil
|
|
}
|
|
|
|
// Fetch will fetch the account jwt claims from the base url, appending the
|
|
// account name onto the end.
|
|
func (ur *URLAccResolver) Fetch(name string) (string, error) {
|
|
url := ur.url + name
|
|
resp, err := ur.c.Get(url)
|
|
if err != nil {
|
|
return _EMPTY_, fmt.Errorf("could not fetch <%q>: %v", url, err)
|
|
} else if resp == nil {
|
|
return _EMPTY_, fmt.Errorf("could not fetch <%q>: no response", url)
|
|
} else if resp.StatusCode != http.StatusOK {
|
|
return _EMPTY_, fmt.Errorf("could not fetch <%q>: %v", url, resp.Status)
|
|
}
|
|
defer resp.Body.Close()
|
|
body, err := ioutil.ReadAll(resp.Body)
|
|
if err != nil {
|
|
return _EMPTY_, err
|
|
}
|
|
return string(body), nil
|
|
}
|
|
|
|
// Resolver based on nats for synchronization and backing directory for storage.
|
|
type DirAccResolver struct {
|
|
*DirJWTStore
|
|
*Server
|
|
syncInterval time.Duration
|
|
}
|
|
|
|
func (dr *DirAccResolver) IsTrackingUpdate() bool {
|
|
return true
|
|
}
|
|
|
|
func (dr *DirAccResolver) Reload() error {
|
|
return dr.DirJWTStore.Reload()
|
|
}
|
|
|
|
func respondToUpdate(s *Server, respSubj string, acc string, message string, err error) {
|
|
if err == nil {
|
|
if acc == "" {
|
|
s.Debugf("%s", message)
|
|
} else {
|
|
s.Debugf("%s - %s", message, acc)
|
|
}
|
|
} else {
|
|
if acc == "" {
|
|
s.Errorf("%s - %s", message, err)
|
|
} else {
|
|
s.Errorf("%s - %s - %s", message, acc, err)
|
|
}
|
|
}
|
|
if respSubj == "" {
|
|
return
|
|
}
|
|
server := &ServerInfo{}
|
|
response := map[string]interface{}{"server": server}
|
|
m := map[string]interface{}{}
|
|
if acc != "" {
|
|
m["account"] = acc
|
|
}
|
|
if err == nil {
|
|
m["code"] = http.StatusOK
|
|
m["message"] = message
|
|
response["data"] = m
|
|
} else {
|
|
m["code"] = http.StatusInternalServerError
|
|
m["description"] = fmt.Sprintf("%s - %v", message, err)
|
|
response["error"] = m
|
|
}
|
|
s.sendInternalMsgLocked(respSubj, _EMPTY_, server, response)
|
|
}
|
|
|
|
func handleListRequest(store *DirJWTStore, s *Server, reply string) {
|
|
if reply == "" {
|
|
return
|
|
}
|
|
accIds := make([]string, 0, 1024)
|
|
if err := store.PackWalk(1, func(partialPackMsg string) {
|
|
if tk := strings.Split(partialPackMsg, "|"); len(tk) == 2 {
|
|
accIds = append(accIds, tk[0])
|
|
}
|
|
}); err != nil {
|
|
// let them timeout
|
|
s.Errorf("list request error: %v", err)
|
|
} else {
|
|
s.Debugf("list request responded with %d account ids", len(accIds))
|
|
server := &ServerInfo{}
|
|
response := map[string]interface{}{"server": server, "data": accIds}
|
|
s.sendInternalMsgLocked(reply, _EMPTY_, server, response)
|
|
}
|
|
}
|
|
|
|
func handleDeleteRequest(store *DirJWTStore, s *Server, msg []byte, reply string) {
|
|
accIds := strings.Split(strings.Replace(string(msg), "\r\n", "\n", -1), "\n")
|
|
errs := []string{}
|
|
passCnt := 0
|
|
for _, acc := range accIds {
|
|
if acc == "" {
|
|
continue
|
|
}
|
|
if err := store.delete(acc); err != nil {
|
|
errs = append(errs, err.Error())
|
|
} else {
|
|
passCnt++
|
|
}
|
|
}
|
|
if len(errs) == 0 {
|
|
respondToUpdate(s, reply, "", fmt.Sprintf("deleted %d accounts", passCnt), nil)
|
|
} else {
|
|
respondToUpdate(s, reply, "", fmt.Sprintf("deleted %d accounts, failed for %d", passCnt, len(errs)),
|
|
errors.New(strings.Join(errs, "<\n")))
|
|
}
|
|
}
|
|
|
|
func (dr *DirAccResolver) Start(s *Server) error {
|
|
dr.Lock()
|
|
defer dr.Unlock()
|
|
dr.Server = s
|
|
dr.DirJWTStore.changed = func(pubKey string) {
|
|
if v, ok := s.accounts.Load(pubKey); !ok {
|
|
} else if jwt, err := dr.LoadAcc(pubKey); err != nil {
|
|
s.Errorf("update got error on load: %v", err)
|
|
} else if err := s.updateAccountWithClaimJWT(v.(*Account), jwt); err != nil {
|
|
s.Errorf("update resulted in error %v", err)
|
|
}
|
|
}
|
|
packRespIb := s.newRespInbox()
|
|
for _, reqSub := range []string{accUpdateEventSubjOld, accUpdateEventSubjNew} {
|
|
// subscribe to account jwt update requests
|
|
if _, err := s.sysSubscribe(fmt.Sprintf(reqSub, "*"), func(_ *subscription, _ *client, subj, resp string, msg []byte) {
|
|
pubKey := ""
|
|
tk := strings.Split(subj, tsep)
|
|
if len(tk) == accUpdateTokensNew {
|
|
pubKey = tk[accReqAccIndex]
|
|
} else if len(tk) == accUpdateTokensOld {
|
|
pubKey = tk[accUpdateAccIdxOld]
|
|
} else {
|
|
s.Debugf("jwt update skipped due to bad subject %q", subj)
|
|
return
|
|
}
|
|
if claim, err := jwt.DecodeAccountClaims(string(msg)); err != nil {
|
|
respondToUpdate(s, resp, pubKey, "jwt update resulted in error", err)
|
|
} else if claim.Subject != pubKey {
|
|
err := errors.New("subject does not match jwt content")
|
|
respondToUpdate(s, resp, pubKey, "jwt update resulted in error", err)
|
|
} else if err := dr.save(pubKey, string(msg)); err != nil {
|
|
respondToUpdate(s, resp, pubKey, "jwt update resulted in error", err)
|
|
} else {
|
|
respondToUpdate(s, resp, pubKey, "jwt updated", nil)
|
|
}
|
|
}); err != nil {
|
|
return fmt.Errorf("error setting up update handling: %v", err)
|
|
}
|
|
}
|
|
if _, err := s.sysSubscribe(accClaimsReqSubj, func(_ *subscription, _ *client, subj, resp string, msg []byte) {
|
|
if claim, err := jwt.DecodeAccountClaims(string(msg)); err != nil {
|
|
respondToUpdate(s, resp, "n/a", "jwt update resulted in error", err)
|
|
} else if err := dr.save(claim.Subject, string(msg)); err != nil {
|
|
respondToUpdate(s, resp, claim.Subject, "jwt update resulted in error", err)
|
|
} else {
|
|
respondToUpdate(s, resp, claim.Subject, "jwt updated", nil)
|
|
}
|
|
}); err != nil {
|
|
return fmt.Errorf("error setting up update handling: %v", err)
|
|
}
|
|
// respond to lookups with our version
|
|
if _, err := s.sysSubscribe(fmt.Sprintf(accLookupReqSubj, "*"), func(_ *subscription, _ *client, subj, reply string, msg []byte) {
|
|
if reply == "" {
|
|
return
|
|
}
|
|
tk := strings.Split(subj, tsep)
|
|
if len(tk) != accLookupReqTokens {
|
|
return
|
|
}
|
|
if theJWT, err := dr.DirJWTStore.LoadAcc(tk[accReqAccIndex]); err != nil {
|
|
s.Errorf("Merging resulted in error: %v", err)
|
|
} else {
|
|
s.sendInternalMsgLocked(reply, "", nil, []byte(theJWT))
|
|
}
|
|
}); err != nil {
|
|
return fmt.Errorf("error setting up lookup request handling: %v", err)
|
|
}
|
|
// respond to pack requests with one or more pack messages
|
|
// an empty message signifies the end of the response responder
|
|
if _, err := s.sysSubscribeQ(accPackReqSubj, "responder", func(_ *subscription, _ *client, _, reply string, theirHash []byte) {
|
|
if reply == "" {
|
|
return
|
|
}
|
|
ourHash := dr.DirJWTStore.Hash()
|
|
if bytes.Equal(theirHash, ourHash[:]) {
|
|
s.sendInternalMsgLocked(reply, "", nil, []byte{})
|
|
s.Debugf("pack request matches hash %x", ourHash[:])
|
|
} else if err := dr.DirJWTStore.PackWalk(1, func(partialPackMsg string) {
|
|
s.sendInternalMsgLocked(reply, "", nil, []byte(partialPackMsg))
|
|
}); err != nil {
|
|
// let them timeout
|
|
s.Errorf("pack request error: %v", err)
|
|
} else {
|
|
s.Debugf("pack request hash %x - finished responding with hash %x", theirHash, ourHash)
|
|
s.sendInternalMsgLocked(reply, "", nil, []byte{})
|
|
}
|
|
}); err != nil {
|
|
return fmt.Errorf("error setting up pack request handling: %v", err)
|
|
}
|
|
// respond to list requests with one message containing all account ids
|
|
if _, err := s.sysSubscribe(accListReqSubj, func(_ *subscription, _ *client, _, reply string, _ []byte) {
|
|
handleListRequest(dr.DirJWTStore, s, reply)
|
|
}); err != nil {
|
|
return fmt.Errorf("error setting up list request handling: %v", err)
|
|
}
|
|
if _, err := s.sysSubscribe(accDeleteReqSubj, func(_ *subscription, _ *client, _, reply string, msg []byte) {
|
|
handleDeleteRequest(dr.DirJWTStore, s, msg, reply)
|
|
}); err != nil {
|
|
return fmt.Errorf("error setting up delete request handling: %v", err)
|
|
}
|
|
// embed pack responses into store
|
|
if _, err := s.sysSubscribe(packRespIb, func(_ *subscription, _ *client, _, _ string, msg []byte) {
|
|
hash := dr.DirJWTStore.Hash()
|
|
if len(msg) == 0 { // end of response stream
|
|
s.Debugf("Merging Finished and resulting in: %x", dr.DirJWTStore.Hash())
|
|
return
|
|
} else if err := dr.DirJWTStore.Merge(string(msg)); err != nil {
|
|
s.Errorf("Merging resulted in error: %v", err)
|
|
} else {
|
|
s.Debugf("Merging succeeded and changed %x to %x", hash, dr.DirJWTStore.Hash())
|
|
}
|
|
}); err != nil {
|
|
return fmt.Errorf("error setting up pack response handling: %v", err)
|
|
}
|
|
// periodically send out pack message
|
|
quit := s.quitCh
|
|
s.startGoRoutine(func() {
|
|
defer s.grWG.Done()
|
|
ticker := time.NewTicker(dr.syncInterval)
|
|
for {
|
|
select {
|
|
case <-quit:
|
|
ticker.Stop()
|
|
return
|
|
case <-ticker.C:
|
|
}
|
|
ourHash := dr.DirJWTStore.Hash()
|
|
s.Debugf("Checking store state: %x", ourHash)
|
|
s.sendInternalMsgLocked(accPackReqSubj, packRespIb, nil, ourHash[:])
|
|
}
|
|
})
|
|
s.Noticef("Managing all jwt in exclusive directory %s", dr.directory)
|
|
return nil
|
|
}
|
|
|
|
func (dr *DirAccResolver) Fetch(name string) (string, error) {
|
|
if theJWT, err := dr.LoadAcc(name); theJWT != "" {
|
|
return theJWT, nil
|
|
} else {
|
|
dr.Lock()
|
|
srv := dr.Server
|
|
dr.Unlock()
|
|
if srv == nil {
|
|
return "", err
|
|
}
|
|
return srv.fetch(dr, name) // lookup from other server
|
|
}
|
|
}
|
|
|
|
func (dr *DirAccResolver) Store(name, jwt string) error {
|
|
return dr.saveIfNewer(name, jwt)
|
|
}
|
|
|
|
func NewDirAccResolver(path string, limit int64, syncInterval time.Duration) (*DirAccResolver, error) {
|
|
if limit == 0 {
|
|
limit = math.MaxInt64
|
|
}
|
|
if syncInterval <= 0 {
|
|
syncInterval = time.Minute
|
|
}
|
|
store, err := NewExpiringDirJWTStore(path, false, true, 0, limit, false, 0, nil)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return &DirAccResolver{store, nil, syncInterval}, nil
|
|
}
|
|
|
|
// Caching resolver using nats for lookups and making use of a directory for storage
|
|
type CacheDirAccResolver struct {
|
|
DirAccResolver
|
|
ttl time.Duration
|
|
}
|
|
|
|
func (s *Server) fetch(res AccountResolver, name string) (string, error) {
|
|
if s == nil {
|
|
return "", ErrNoAccountResolver
|
|
}
|
|
respC := make(chan []byte, 1)
|
|
accountLookupRequest := fmt.Sprintf(accLookupReqSubj, name)
|
|
s.mu.Lock()
|
|
if s.sys == nil || s.sys.replies == nil {
|
|
s.mu.Unlock()
|
|
return "", fmt.Errorf("eventing shut down")
|
|
}
|
|
replySubj := s.newRespInbox()
|
|
replies := s.sys.replies
|
|
// Store our handler.
|
|
replies[replySubj] = func(sub *subscription, _ *client, subject, _ string, msg []byte) {
|
|
clone := make([]byte, len(msg))
|
|
copy(clone, msg)
|
|
s.mu.Lock()
|
|
if _, ok := replies[replySubj]; ok {
|
|
select {
|
|
case respC <- clone: // only use first response and only if there is still interest
|
|
default:
|
|
}
|
|
}
|
|
s.mu.Unlock()
|
|
}
|
|
s.sendInternalMsg(accountLookupRequest, replySubj, nil, []byte{})
|
|
quit := s.quitCh
|
|
s.mu.Unlock()
|
|
var err error
|
|
var theJWT string
|
|
select {
|
|
case <-quit:
|
|
err = errors.New("fetching jwt failed due to shutdown")
|
|
case <-time.After(fetchTimeout):
|
|
err = errors.New("fetching jwt timed out")
|
|
case m := <-respC:
|
|
if err = res.Store(name, string(m)); err == nil {
|
|
theJWT = string(m)
|
|
}
|
|
}
|
|
s.mu.Lock()
|
|
delete(replies, replySubj)
|
|
s.mu.Unlock()
|
|
close(respC)
|
|
return theJWT, err
|
|
}
|
|
|
|
func NewCacheDirAccResolver(path string, limit int64, ttl time.Duration) (*CacheDirAccResolver, error) {
|
|
if limit <= 0 {
|
|
limit = 1_000
|
|
}
|
|
store, err := NewExpiringDirJWTStore(path, false, true, 0, limit, true, ttl, nil)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return &CacheDirAccResolver{DirAccResolver{store, nil, 0}, ttl}, nil
|
|
}
|
|
|
|
func (dr *CacheDirAccResolver) Start(s *Server) error {
|
|
dr.Lock()
|
|
defer dr.Unlock()
|
|
dr.Server = s
|
|
dr.DirJWTStore.changed = func(pubKey string) {
|
|
if v, ok := s.accounts.Load(pubKey); !ok {
|
|
} else if jwt, err := dr.LoadAcc(pubKey); err != nil {
|
|
s.Errorf("update got error on load: %v", err)
|
|
} else if err := s.updateAccountWithClaimJWT(v.(*Account), jwt); err != nil {
|
|
s.Errorf("update resulted in error %v", err)
|
|
}
|
|
}
|
|
for _, reqSub := range []string{accUpdateEventSubjOld, accUpdateEventSubjNew} {
|
|
// subscribe to account jwt update requests
|
|
if _, err := s.sysSubscribe(fmt.Sprintf(reqSub, "*"), func(_ *subscription, _ *client, subj, resp string, msg []byte) {
|
|
pubKey := ""
|
|
tk := strings.Split(subj, tsep)
|
|
if len(tk) == accUpdateTokensNew {
|
|
pubKey = tk[accReqAccIndex]
|
|
} else if len(tk) == accUpdateTokensOld {
|
|
pubKey = tk[accUpdateAccIdxOld]
|
|
} else {
|
|
s.Debugf("jwt update cache skipped due to bad subject %q", subj)
|
|
return
|
|
}
|
|
if claim, err := jwt.DecodeAccountClaims(string(msg)); err != nil {
|
|
respondToUpdate(s, resp, pubKey, "jwt update cache resulted in error", err)
|
|
} else if claim.Subject != pubKey {
|
|
err := errors.New("subject does not match jwt content")
|
|
respondToUpdate(s, resp, pubKey, "jwt update cache resulted in error", err)
|
|
} else if _, ok := s.accounts.Load(pubKey); !ok {
|
|
respondToUpdate(s, resp, pubKey, "jwt update cache skipped", nil)
|
|
} else if err := dr.save(pubKey, string(msg)); err != nil {
|
|
respondToUpdate(s, resp, pubKey, "jwt update cache resulted in error", err)
|
|
} else {
|
|
respondToUpdate(s, resp, pubKey, "jwt updated cache", nil)
|
|
}
|
|
}); err != nil {
|
|
return fmt.Errorf("error setting up update handling: %v", err)
|
|
}
|
|
}
|
|
if _, err := s.sysSubscribe(accClaimsReqSubj, func(_ *subscription, _ *client, subj, resp string, msg []byte) {
|
|
if claim, err := jwt.DecodeAccountClaims(string(msg)); err != nil {
|
|
respondToUpdate(s, resp, "n/a", "jwt update cache resulted in error", err)
|
|
} else if _, ok := s.accounts.Load(claim.Subject); !ok {
|
|
respondToUpdate(s, resp, claim.Subject, "jwt update cache skipped", nil)
|
|
} else if err := dr.save(claim.Subject, string(msg)); err != nil {
|
|
respondToUpdate(s, resp, claim.Subject, "jwt update cache resulted in error", err)
|
|
} else {
|
|
respondToUpdate(s, resp, claim.Subject, "jwt updated cache", nil)
|
|
}
|
|
}); err != nil {
|
|
return fmt.Errorf("error setting up update handling: %v", err)
|
|
}
|
|
// respond to list requests with one message containing all account ids
|
|
if _, err := s.sysSubscribe(accListReqSubj, func(_ *subscription, _ *client, _, reply string, _ []byte) {
|
|
handleListRequest(dr.DirJWTStore, s, reply)
|
|
}); err != nil {
|
|
return fmt.Errorf("error setting up list request handling: %v", err)
|
|
}
|
|
if _, err := s.sysSubscribe(accDeleteReqSubj, func(_ *subscription, _ *client, _, reply string, msg []byte) {
|
|
handleDeleteRequest(dr.DirJWTStore, s, msg, reply)
|
|
}); err != nil {
|
|
return fmt.Errorf("error setting up list request handling: %v", err)
|
|
}
|
|
s.Noticef("Managing some jwt in exclusive directory %s", dr.directory)
|
|
return nil
|
|
}
|