mirror of
https://github.com/postmannen/ctrl.git
synced 2024-12-14 12:37:31 +00:00
400 lines
14 KiB
Go
400 lines
14 KiB
Go
// Notes:
|
|
package steward
|
|
|
|
import (
|
|
"bytes"
|
|
"encoding/gob"
|
|
"fmt"
|
|
"log"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/nats-io/nats.go"
|
|
"github.com/prometheus/client_golang/prometheus"
|
|
)
|
|
|
|
type processName string
|
|
|
|
func processNameGet(sn subjectName, pk processKind) processName {
|
|
pn := fmt.Sprintf("%s_%s", sn, pk)
|
|
return processName(pn)
|
|
}
|
|
|
|
// server is the structure that will hold the state about spawned
|
|
// processes on a local instance.
|
|
type server struct {
|
|
natsConn *nats.Conn
|
|
// TODO: sessions should probably hold a slice/map of processes ?
|
|
processes map[processName]process
|
|
// The last processID created
|
|
lastProcessID int
|
|
// The name of the node
|
|
nodeName string
|
|
// Mutex for locking when writing to the process map
|
|
mu sync.Mutex
|
|
// The channel where we put new messages read from file,
|
|
// or some other process who wants to send something via the
|
|
// system
|
|
// We can than range this channel for new messages to process.
|
|
newMessagesCh chan []subjectAndMessage
|
|
// errorKernel is doing all the error handling like what to do if
|
|
// an error occurs.
|
|
// TODO: Will also send error messages to cental error subscriber.
|
|
errorKernel *errorKernel
|
|
// used to check if the methods specified in message is valid
|
|
methodsAvailable MethodsAvailable
|
|
// Map who holds the command and event types available.
|
|
// Used to check if the commandOrEvent specified in message is valid
|
|
commandOrEventAvailable CommandOrEventAvailable
|
|
// metric exporter
|
|
metrics *metrics
|
|
// subscriberServices are where we find the services and the API to
|
|
// use services needed by subscriber.
|
|
// For example, this can be a service that knows
|
|
// how to forward the data for a received message of type log to a
|
|
// central logger.
|
|
subscriberServices *subscriberServices
|
|
// Is this the central error logger ?
|
|
centralErrorLogger bool
|
|
// default message timeout in seconds. This can be overridden on the message level
|
|
defaultMessageTimeout int
|
|
// default amount of retries that will be done before a message is thrown away, and out of the system
|
|
defaultMessageRetries int
|
|
}
|
|
|
|
// newServer will prepare and return a server type
|
|
func NewServer(brokerAddress string, nodeName string, promHostAndPort string, centralErrorLogger bool, defaultMessageTimeout int, defaultMessageRetries int) (*server, error) {
|
|
conn, err := nats.Connect(brokerAddress, nil)
|
|
if err != nil {
|
|
log.Printf("error: nats.Connect failed: %v\n", err)
|
|
}
|
|
|
|
var m Method
|
|
var coe CommandOrEvent
|
|
|
|
s := &server{
|
|
nodeName: nodeName,
|
|
natsConn: conn,
|
|
processes: make(map[processName]process),
|
|
newMessagesCh: make(chan []subjectAndMessage),
|
|
methodsAvailable: m.GetMethodsAvailable(),
|
|
commandOrEventAvailable: coe.GetCommandOrEventAvailable(),
|
|
metrics: newMetrics(promHostAndPort),
|
|
subscriberServices: newSubscriberServices(),
|
|
centralErrorLogger: centralErrorLogger,
|
|
defaultMessageTimeout: defaultMessageTimeout,
|
|
defaultMessageRetries: defaultMessageRetries,
|
|
}
|
|
|
|
return s, nil
|
|
|
|
}
|
|
|
|
// Start will spawn up all the predefined subscriber processes.
|
|
// Spawning of publisher processes is done on the fly by checking
|
|
// if there is publisher process for a given message subject, and
|
|
// not exist it will spawn one.
|
|
func (s *server) Start() {
|
|
// Start the error kernel that will do all the error handling
|
|
// not done within a process.
|
|
s.errorKernel = newErrorKernel()
|
|
s.errorKernel.startErrorKernel(s.newMessagesCh)
|
|
|
|
// Start collecting the metrics
|
|
go s.startMetrics()
|
|
|
|
// Start the checking the input file for new messages from operator.
|
|
go s.getMessagesFromFile("./", "inmsg.txt", s.newMessagesCh)
|
|
|
|
// Start the textLogging service that will run on the subscribers
|
|
// TODO: This should only be started if the flag value provided when
|
|
// starting asks to subscribe to TextLogging events.
|
|
go s.subscriberServices.startTextLogging()
|
|
|
|
// Start up the predefined subscribers.
|
|
// TODO: What to subscribe on should be handled via flags, or config
|
|
// files.
|
|
s.subscribersStart()
|
|
|
|
time.Sleep(time.Second * 2)
|
|
s.printProcessesMap()
|
|
|
|
// Start the processing of new messaging from an input channel.
|
|
s.processNewMessages("./incommmingBuffer.db", s.newMessagesCh)
|
|
|
|
select {}
|
|
|
|
}
|
|
|
|
func (s *server) printProcessesMap() {
|
|
fmt.Println("--------------------------------------------------------------------------------------------")
|
|
fmt.Printf("*** Output of processes map :\n")
|
|
for _, v := range s.processes {
|
|
fmt.Printf("*** - : %v\n", v)
|
|
}
|
|
|
|
s.metrics.metricsCh <- metricType{
|
|
metric: prometheus.NewGauge(prometheus.GaugeOpts{
|
|
Name: "total_running_processes",
|
|
Help: "The current number of total running processes",
|
|
}),
|
|
value: float64(len(s.processes)),
|
|
}
|
|
|
|
fmt.Println("--------------------------------------------------------------------------------------------")
|
|
}
|
|
|
|
// processKind are either kindSubscriber or kindPublisher, and are
|
|
// used to distinguish the kind of process to spawn and to know
|
|
// the process kind put in the process map.
|
|
type processKind string
|
|
|
|
const (
|
|
processKindSubscriber processKind = "subscriber"
|
|
processKindPublisher processKind = "publisher"
|
|
)
|
|
|
|
// process are represent the communication to one individual host
|
|
type process struct {
|
|
messageID int
|
|
// the subject used for the specific process. One process
|
|
// can contain only one sender on a message bus, hence
|
|
// also one subject
|
|
subject Subject
|
|
// Put a node here to be able know the node a process is at.
|
|
// NB: Might not be needed later on.
|
|
node node
|
|
// The processID for the current process
|
|
processID int
|
|
// errorCh is used to report errors from a process
|
|
// NB: Implementing this as an int to report for testing
|
|
errorCh chan errProcess
|
|
processKind processKind
|
|
}
|
|
|
|
// prepareNewProcess will set the the provided values and the default
|
|
// values for a process.
|
|
func (s *server) processPrepareNew(subject Subject, errCh chan errProcess, processKind processKind) process {
|
|
// create the initial configuration for a sessions communicating with 1 host process.
|
|
s.lastProcessID++
|
|
proc := process{
|
|
messageID: 0,
|
|
subject: subject,
|
|
node: node(subject.ToNode),
|
|
processID: s.lastProcessID,
|
|
errorCh: errCh,
|
|
processKind: processKind,
|
|
//messageCh: make(chan Message),
|
|
}
|
|
|
|
return proc
|
|
}
|
|
|
|
// spawnWorkerProcess will spawn take care of spawning both publisher
|
|
// and subscriber proesses.
|
|
//It will give the process the next available ID, and also add the
|
|
// process to the processes map.
|
|
func (s *server) spawnWorkerProcess(proc process) {
|
|
s.mu.Lock()
|
|
// We use the full name of the subject to identify a unique
|
|
// process. We can do that since a process can only handle
|
|
// one message queue.
|
|
var pn processName
|
|
if proc.processKind == processKindPublisher {
|
|
pn = processNameGet(proc.subject.name(), processKindPublisher)
|
|
}
|
|
if proc.processKind == processKindSubscriber {
|
|
pn = processNameGet(proc.subject.name(), processKindSubscriber)
|
|
}
|
|
|
|
s.processes[pn] = proc
|
|
s.mu.Unlock()
|
|
|
|
// TODO: I think it makes most sense that the messages would come to
|
|
// here from some other message-pickup-process, and that process will
|
|
// give the message to the correct publisher process. A channel that
|
|
// is listened on in the for loop below could be used to receive the
|
|
// messages from the message-pickup-process.
|
|
//
|
|
// Handle publisher workers
|
|
if proc.processKind == processKindPublisher {
|
|
s.publishMessages(proc)
|
|
}
|
|
|
|
// handle subscriber workers
|
|
if proc.processKind == processKindSubscriber {
|
|
s.subscribeMessages(proc)
|
|
}
|
|
}
|
|
|
|
func (s *server) messageDeliverNats(proc process, message Message) {
|
|
retryAttempts := 0
|
|
|
|
for {
|
|
dataPayload, err := gobEncodeMessage(message)
|
|
if err != nil {
|
|
log.Printf("error: createDataPayload: %v\n", err)
|
|
}
|
|
|
|
msg := &nats.Msg{
|
|
Subject: string(proc.subject.name()),
|
|
// Subject: fmt.Sprintf("%s.%s.%s", proc.node, "command", "shellCommand"),
|
|
// Structure of the reply message are:
|
|
// reply.<nodename>.<message type>.<method>
|
|
Reply: fmt.Sprintf("reply.%s", proc.subject.name()),
|
|
Data: dataPayload,
|
|
}
|
|
|
|
// The SubscribeSync used in the subscriber, will get messages that
|
|
// are sent after it started subscribing, so we start a publisher
|
|
// that sends out a message every second.
|
|
//
|
|
// Create a subscriber for the reply message.
|
|
subReply, err := s.natsConn.SubscribeSync(msg.Reply)
|
|
if err != nil {
|
|
log.Printf("error: nc.SubscribeSync failed: failed to create reply message: %v\n", err)
|
|
continue
|
|
}
|
|
|
|
// Publish message
|
|
err = s.natsConn.PublishMsg(msg)
|
|
if err != nil {
|
|
log.Printf("error: publish failed: %v\n", err)
|
|
continue
|
|
}
|
|
|
|
// If the message is an ACK type of message we must check that a
|
|
// reply, and if it is not we don't wait here at all.
|
|
fmt.Printf("---- MESSAGE : %v\n", message)
|
|
if message.CommandOrEvent == CommandACK || message.CommandOrEvent == EventACK {
|
|
// Wait up until 10 seconds for a reply,
|
|
// continue and resend if to reply received.
|
|
msgReply, err := subReply.NextMsg(time.Second * time.Duration(message.Timeout))
|
|
if err != nil {
|
|
log.Printf("error: subReply.NextMsg failed for node=%v, subject=%v: %v\n", proc.node, proc.subject.name(), err)
|
|
|
|
// did not receive a reply, decide what to do..
|
|
retryAttempts++
|
|
fmt.Printf("Retry attempts:%v, retries: %v, timeout: %v\n", retryAttempts, message.Retries, message.Timeout)
|
|
switch {
|
|
case message.Retries == 0:
|
|
// 0 indicates unlimited retries
|
|
continue
|
|
case retryAttempts >= message.Retries:
|
|
// max retries reached
|
|
log.Printf("info: max retries for message reached, breaking out: %v", retryAttempts)
|
|
return
|
|
default:
|
|
// none of the above matched, so we've not reached max retries yet
|
|
continue
|
|
}
|
|
}
|
|
log.Printf("info: publisher: received ACK for message: %s\n", msgReply.Data)
|
|
}
|
|
return
|
|
}
|
|
}
|
|
|
|
// handler will deserialize the message when a new message is received,
|
|
// check the MessageType field in the message to decide what kind of
|
|
// message it is and then it will check how to handle that message type,
|
|
// and handle it.
|
|
// This handler function should be started in it's own go routine,so
|
|
// one individual handler is started per message received so we can keep
|
|
// the state of the message being processed, and then reply back to the
|
|
// correct sending process's reply, meaning so we ACK back to the correct
|
|
// publisher.
|
|
func (s *server) subscriberHandler(natsConn *nats.Conn, thisNode string, msg *nats.Msg) {
|
|
|
|
message := Message{}
|
|
|
|
// Create a buffer to decode the gob encoded binary data back
|
|
// to it's original structure.
|
|
buf := bytes.NewBuffer(msg.Data)
|
|
gobDec := gob.NewDecoder(buf)
|
|
err := gobDec.Decode(&message)
|
|
if err != nil {
|
|
log.Printf("error: gob decoding failed: %v\n", err)
|
|
}
|
|
|
|
//fmt.Printf("%v\n", msg)
|
|
// TODO: Maybe the handling of the errors within the subscriber
|
|
// should also involve the error-kernel to report back centrally
|
|
// that there was a problem like missing method to handle a specific
|
|
// method etc.
|
|
switch {
|
|
case message.CommandOrEvent == CommandACK || message.CommandOrEvent == EventACK:
|
|
log.Printf("info: subscriberHandler: message.CommandOrEvent received was = %v, preparing to call handler\n", message.CommandOrEvent)
|
|
mf, ok := s.methodsAvailable.CheckIfExists(message.Method)
|
|
if !ok {
|
|
// TODO: Check how errors should be handled here!!!
|
|
log.Printf("error: subscriberHandler: method type not available: %v\n", message.CommandOrEvent)
|
|
}
|
|
fmt.Printf("*** DEBUG: BEFORE CALLING HANDLER: ACK\n")
|
|
out, err := mf.handler(s, message, thisNode)
|
|
|
|
if err != nil {
|
|
// TODO: Send to error kernel ?
|
|
log.Printf("error: subscriberHandler: failed to execute event: %v\n", err)
|
|
}
|
|
|
|
// Send a confirmation message back to the publisher
|
|
natsConn.Publish(msg.Reply, out)
|
|
|
|
// TESTING: Simulate that we also want to send some error that occured
|
|
// to the errorCentral
|
|
{
|
|
err := fmt.Errorf("error: some testing error we want to send out")
|
|
sendErrorLogMessage(s.newMessagesCh, node(thisNode), err)
|
|
}
|
|
case message.CommandOrEvent == CommandNACK || message.CommandOrEvent == EventNACK:
|
|
log.Printf("info: subscriberHandler: message.CommandOrEvent received was = %v, preparing to call handler\n", message.CommandOrEvent)
|
|
mf, ok := s.methodsAvailable.CheckIfExists(message.Method)
|
|
if !ok {
|
|
// TODO: Check how errors should be handled here!!!
|
|
log.Printf("error: subscriberHandler: method type not available: %v\n", message.CommandOrEvent)
|
|
}
|
|
// since we don't send a reply for a NACK message, we don't care about the
|
|
// out return when calling mf.handler
|
|
fmt.Printf("*** DEBUG: BEFORE CALLING HANDLER: NACK\n")
|
|
_, err := mf.handler(s, message, thisNode)
|
|
|
|
if err != nil {
|
|
// TODO: Send to error kernel ?
|
|
log.Printf("error: subscriberHandler: failed to execute event: %v\n", err)
|
|
}
|
|
default:
|
|
log.Printf("info: did not find that specific type of command: %#v\n", message.CommandOrEvent)
|
|
}
|
|
}
|
|
|
|
// sendErrorMessage will put the error message directly on the channel that is
|
|
// read by the nats publishing functions.
|
|
func sendErrorLogMessage(newMessagesCh chan<- []subjectAndMessage, FromNode node, theError error) {
|
|
// --- Testing
|
|
sam := createErrorMsgContent(FromNode, theError)
|
|
newMessagesCh <- []subjectAndMessage{sam}
|
|
}
|
|
|
|
// createErrorMsgContent will prepare a subject and message with the content
|
|
// of the error
|
|
func createErrorMsgContent(FromNode node, theError error) subjectAndMessage {
|
|
// TESTING: Creating an error message to send to errorCentral
|
|
fmt.Printf(" --- Sending error message to central !!!!!!!!!!!!!!!!!!!!!!!!!!!!\n")
|
|
sam := subjectAndMessage{
|
|
Subject: Subject{
|
|
ToNode: "errorCentral",
|
|
CommandOrEvent: EventNACK,
|
|
Method: ErrorLog,
|
|
},
|
|
Message: Message{
|
|
ToNode: "errorCentral",
|
|
FromNode: FromNode,
|
|
Data: []string{theError.Error()},
|
|
CommandOrEvent: EventNACK,
|
|
Method: ErrorLog,
|
|
},
|
|
}
|
|
|
|
return sam
|
|
}
|