@@ -91,9 +91,8 @@ type Manager struct {
9191
9292 tasksToDeploy chan <- * ResourceOffersDeploymentRequest
9393
94- reviveOffersTrg chan struct {}
95- reviveOffersDone chan struct {}
96- cq * controlcommands.CommandQueue
94+ reviveOffersTrg chan struct {}
95+ cq * controlcommands.CommandQueue
9796
9897 tasksLaunched int
9998 tasksFinished int
@@ -142,7 +141,6 @@ func NewManager(shutdown func(), internalEventCh chan<- event.Event) (taskman *M
142141 taskman .cq = taskman .schedulerState .commandqueue
143142 taskman .tasksToDeploy = taskman .schedulerState .tasksToDeploy
144143 taskman .reviveOffersTrg = taskman .schedulerState .reviveOffersTrg
145- taskman .reviveOffersDone = taskman .schedulerState .reviveOffersDone
146144 taskman .ackKilledTasks = newAcks ()
147145
148146 schedState .setupCli ()
@@ -158,8 +156,7 @@ func (m *Manager) newTaskForMesosOffer(
158156 offer * mesos.Offer ,
159157 descriptor * Descriptor ,
160158 localBindMap channel.BindMap ,
161- executorId mesos.ExecutorID ,
162- ) (t * Task ) {
159+ executorId mesos.ExecutorID ) (t * Task ) {
163160 newId := uid .New ().String ()
164161 t = & Task {
165162 name : fmt .Sprintf ("%s#%s" , descriptor .TaskClassName , newId ),
@@ -200,8 +197,8 @@ func getTaskClassList(taskClassesRequired []string) (taskClassList []*taskclass.
200197 if err != nil {
201198 return
202199 }
203- repo := repoManager .GetAllRepos ()[tempRepo .GetIdentifier ()] // get IRepo pointer from RepoManager
204- if repo == nil { // should never end up here
200+ repo := repoManager .GetAllRepos ()[tempRepo .GetIdentifier ()] //get IRepo pointer from RepoManager
201+ if repo == nil { //should never end up here
205202 return nil , errors .New ("getTaskClassList: repo not found for " + taskClass )
206203 }
207204
@@ -226,6 +223,7 @@ func getTaskClassList(taskClassesRequired []string) (taskClassList []*taskclass.
226223 taskInfo := strings .Split (taskPath , "/tasks/" )
227224 if len (taskInfo ) == 1 {
228225 taskFilename = taskInfo [0 ]
226+
229227 } else {
230228 taskFilename = taskInfo [1 ]
231229 }
@@ -282,7 +280,7 @@ func (m *Manager) removeInactiveClasses() {
282280 return
283281}
284282
285- func (m * Manager ) RemoveReposClasses (repoPath string ) { // Currently unused
283+ func (m * Manager ) RemoveReposClasses (repoPath string ) { //Currently unused
286284 utils .EnsureTrailingSlash (& repoPath )
287285
288286 _ = m .classes .Do (func (classMap * map [string ]* taskclass.Class ) error {
@@ -329,6 +327,7 @@ func (m *Manager) RefreshClasses(taskClassesRequired []string) (err error) {
329327}
330328
331329func (m * Manager ) acquireTasks (envId uid.ID , taskDescriptors Descriptors ) (err error ) {
330+
332331 /*
333332 Here's what's gonna happen:
334333 1) check if any tasks are already in Roster, whether they are already locked
@@ -517,7 +516,7 @@ func (m *Manager) acquireTasks(envId uid.ID, taskDescriptors Descriptors) (err e
517516 timeReviveOffers := time .Now ()
518517 timeDeployMu := time .Now ()
519518 m .reviveOffersTrg <- struct {}{} // signal scheduler to revive offers
520- <- m .reviveOffersDone // we only continue when it's done
519+ <- m .reviveOffersTrg // we only continue when it's done
521520 utils .TimeTrack (timeReviveOffers , "acquireTasks: revive offers" ,
522521 log .WithField ("tasksToRun" , len (tasksToRun )).
523522 WithField ("partition" , envId ))
@@ -598,7 +597,7 @@ func (m *Manager) acquireTasks(envId uid.ID, taskDescriptors Descriptors) (err e
598597 // can't lock some of them, so we must roll back and keep them
599598 // unlocked in the roster.
600599 var deployedTaskIds []string
601- for taskPtr := range deployedTasks {
600+ for taskPtr , _ := range deployedTasks {
602601 taskPtr .SetParent (nil )
603602 deployedTaskIds = append (deployedTaskIds , taskPtr .taskId )
604603 }
@@ -613,11 +612,11 @@ func (m *Manager) acquireTasks(envId uid.ID, taskDescriptors Descriptors) (err e
613612 }
614613
615614 // Finally, we write to the roster. Point of no return!
616- for taskPtr := range deployedTasks {
615+ for taskPtr , _ := range deployedTasks {
617616 m .roster .append (taskPtr )
618617 }
619618 if deploymentSuccess {
620- for taskPtr := range deployedTasks {
619+ for taskPtr , _ := range deployedTasks {
621620 taskPtr .GetParent ().SetTask (taskPtr )
622621 }
623622 for taskPtr , descriptor := range tasksAlreadyRunning {
@@ -630,6 +629,7 @@ func (m *Manager) acquireTasks(envId uid.ID, taskDescriptors Descriptors) (err e
630629}
631630
632631func (m * Manager ) releaseTasks (envId uid.ID , tasks Tasks ) error {
632+
633633 taskReleaseErrors := make (map [string ]error )
634634 taskIdsReleased := make ([]string , 0 )
635635
@@ -686,7 +686,7 @@ func (m *Manager) configureTasks(envId uid.ID, tasks Tasks) error {
686686 taskPath := task .GetParentRolePath ()
687687 for inbChName , endpoint := range task .GetLocalBindMap () {
688688 var bindMapKey string
689- if strings .HasPrefix (inbChName , "::" ) { // global channel alias
689+ if strings .HasPrefix (inbChName , "::" ) { //global channel alias
690690 bindMapKey = inbChName
691691
692692 // deduplication
@@ -785,6 +785,7 @@ func (m *Manager) configureTasks(envId uid.ID, tasks Tasks) error {
785785func (m * Manager ) transitionTasks (envId uid.ID , tasks Tasks , src string , event string , dest string , commonArgs controlcommands.PropertyMap ) error {
786786 notify := make (chan controlcommands.MesosCommandResponse )
787787 receivers , err := tasks .GetMesosCommandTargets ()
788+
788789 if err != nil {
789790 return err
790791 }
@@ -869,6 +870,7 @@ func (m *Manager) TriggerHooks(envId uid.ID, tasks Tasks) error {
869870
870871 notify := make (chan controlcommands.MesosCommandResponse )
871872 receivers , err := tasks .GetMesosCommandTargets ()
873+
872874 if err != nil {
873875 return err
874876 }
@@ -933,6 +935,7 @@ func (m *Manager) GetTask(id string) *Task {
933935}
934936
935937func (m * Manager ) updateTaskState (taskId string , state string ) {
938+
936939 taskPtr := m .roster .getByTaskId (taskId )
937940 if taskPtr == nil {
938941 log .WithField ("taskId" , taskId ).
@@ -986,7 +989,7 @@ func (m *Manager) updateTaskStatus(status *mesos.TaskStatus) {
986989 }
987990 if ack , ok := m .ackKilledTasks .getValue (taskId ); ok {
988991 ack <- struct {}{}
989- // close(ack) // It can even be left open?
992+ //close(ack) // It can even be left open?
990993 }
991994
992995 return
@@ -1027,6 +1030,7 @@ func (m *Manager) updateTaskStatus(status *mesos.TaskStatus) {
10271030
10281031// Kill all tasks outside an environment (all unlocked tasks)
10291032func (m * Manager ) Cleanup () (killed Tasks , running Tasks , err error ) {
1033+
10301034 toKill := m .roster .filtered (func (t * Task ) bool {
10311035 return ! t .IsLocked ()
10321036 })
0 commit comments