Cleanup logs for Allocation recovery
on startup. The changes do not have functional consequences as adding the allocation just overwrites the old one.
This commit is contained in:
@ -150,7 +150,7 @@ func (a *APIClient) LoadRunnerPortMappings(runnerID string) ([]nomadApi.PortMapp
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (a *APIClient) LoadRunnerJobs(environmentID dto.EnvironmentID) ([]*nomadApi.Job, error) {
|
func (a *APIClient) LoadRunnerJobs(environmentID dto.EnvironmentID) ([]*nomadApi.Job, error) {
|
||||||
go a.initializeAllocations()
|
go a.initializeAllocations(environmentID)
|
||||||
|
|
||||||
runnerIDs, err := a.LoadRunnerIDs(RunnerJobID(environmentID, ""))
|
runnerIDs, err := a.LoadRunnerIDs(RunnerJobID(environmentID, ""))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -234,17 +234,19 @@ func dumpNomadEventToInflux(event *nomadApi.Event) {
|
|||||||
monitoring.WriteInfluxPoint(p)
|
monitoring.WriteInfluxPoint(p)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *APIClient) initializeAllocations() {
|
func (a *APIClient) initializeAllocations(environmentID dto.EnvironmentID) {
|
||||||
allocationStubs, err := a.listAllocations()
|
allocationStubs, err := a.listAllocations()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.WithError(err).Warn("Could not initialize allocations")
|
log.WithError(err).Warn("Could not initialize allocations")
|
||||||
} else {
|
} else {
|
||||||
for _, stub := range allocationStubs {
|
for _, stub := range allocationStubs {
|
||||||
if IsEnvironmentTemplateID(stub.JobID) {
|
switch {
|
||||||
|
case IsEnvironmentTemplateID(stub.JobID):
|
||||||
continue
|
continue
|
||||||
} else if stub.ClientStatus == structs.AllocClientStatusPending ||
|
case !strings.HasPrefix(stub.JobID, RunnerJobID(environmentID, "")):
|
||||||
stub.ClientStatus == structs.AllocClientStatusRunning {
|
continue
|
||||||
log.WithField("jobID", stub.JobID).WithField("status", stub.ClientStatus).Debug("Recovered Runner")
|
case stub.ClientStatus == structs.AllocClientStatusPending || stub.ClientStatus == structs.AllocClientStatusRunning:
|
||||||
|
log.WithField("jobID", stub.JobID).WithField("status", stub.ClientStatus).Debug("Recovered Allocation")
|
||||||
a.allocations.Add(stub.ID,
|
a.allocations.Add(stub.ID,
|
||||||
&allocationData{allocClientStatus: stub.ClientStatus, start: time.Unix(0, stub.CreateTime), jobID: stub.JobID})
|
&allocationData{allocClientStatus: stub.ClientStatus, start: time.Unix(0, stub.CreateTime), jobID: stub.JobID})
|
||||||
}
|
}
|
||||||
|
@ -115,6 +115,7 @@ func (m *NomadRunnerManager) loadSingleJob(job *nomadApi.Job, environmentLogger
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
newJob := NewNomadJob(*job.ID, portMappings, m.apiClient, m.Return)
|
newJob := NewNomadJob(*job.ID, portMappings, m.apiClient, m.Return)
|
||||||
|
log.WithField("isUsed", isUsed).WithField("runner_id", newJob.ID()).Debug("Recovered Runner")
|
||||||
if isUsed {
|
if isUsed {
|
||||||
m.usedRunners.Add(newJob.ID(), newJob)
|
m.usedRunners.Add(newJob.ID(), newJob)
|
||||||
timeout, err := strconv.Atoi(configTaskGroup.Meta[nomad.ConfigMetaTimeoutKey])
|
timeout, err := strconv.Atoi(configTaskGroup.Meta[nomad.ConfigMetaTimeoutKey])
|
||||||
|
Reference in New Issue
Block a user