@@ -91,8 +91,9 @@ type Manager struct {
9191
9292 tasksToDeploy chan <- * ResourceOffersDeploymentRequest
9393
94- reviveOffersTrg chan struct {}
95- cq * controlcommands.CommandQueue
94+ reviveOffersTrg chan struct {}
95+ reviveOffersDone chan struct {}
96+ cq * controlcommands.CommandQueue
9697
9798 tasksLaunched int
9899 tasksFinished int
@@ -141,6 +142,7 @@ func NewManager(shutdown func(), internalEventCh chan<- event.Event) (taskman *M
141142 taskman .cq = taskman .schedulerState .commandqueue
142143 taskman .tasksToDeploy = taskman .schedulerState .tasksToDeploy
143144 taskman .reviveOffersTrg = taskman .schedulerState .reviveOffersTrg
145+ taskman .reviveOffersDone = taskman .schedulerState .reviveOffersDone
144146 taskman .ackKilledTasks = newAcks ()
145147
146148 schedState .setupCli ()
@@ -156,7 +158,8 @@ func (m *Manager) newTaskForMesosOffer(
156158 offer * mesos.Offer ,
157159 descriptor * Descriptor ,
158160 localBindMap channel.BindMap ,
159- executorId mesos.ExecutorID ) (t * Task ) {
161+ executorId mesos.ExecutorID ,
162+ ) (t * Task ) {
160163 newId := uid .New ().String ()
161164 t = & Task {
162165 name : fmt .Sprintf ("%s#%s" , descriptor .TaskClassName , newId ),
@@ -197,8 +200,8 @@ func getTaskClassList(taskClassesRequired []string) (taskClassList []*taskclass.
197200 if err != nil {
198201 return
199202 }
200- repo := repoManager .GetAllRepos ()[tempRepo .GetIdentifier ()] //get IRepo pointer from RepoManager
201- if repo == nil { //should never end up here
203+ repo := repoManager .GetAllRepos ()[tempRepo .GetIdentifier ()] // get IRepo pointer from RepoManager
204+ if repo == nil { // should never end up here
202205 return nil , errors .New ("getTaskClassList: repo not found for " + taskClass )
203206 }
204207
@@ -223,7 +226,6 @@ func getTaskClassList(taskClassesRequired []string) (taskClassList []*taskclass.
223226 taskInfo := strings .Split (taskPath , "/tasks/" )
224227 if len (taskInfo ) == 1 {
225228 taskFilename = taskInfo [0 ]
226-
227229 } else {
228230 taskFilename = taskInfo [1 ]
229231 }
@@ -280,7 +282,7 @@ func (m *Manager) removeInactiveClasses() {
280282 return
281283}
282284
283- func (m * Manager ) RemoveReposClasses (repoPath string ) { //Currently unused
285+ func (m * Manager ) RemoveReposClasses (repoPath string ) { // Currently unused
284286 utils .EnsureTrailingSlash (& repoPath )
285287
286288 _ = m .classes .Do (func (classMap * map [string ]* taskclass.Class ) error {
@@ -327,7 +329,6 @@ func (m *Manager) RefreshClasses(taskClassesRequired []string) (err error) {
327329}
328330
329331func (m * Manager ) acquireTasks (envId uid.ID , taskDescriptors Descriptors ) (err error ) {
330-
331332 /*
332333 Here's what's gonna happen:
333334 1) check if any tasks are already in Roster, whether they are already locked
@@ -516,7 +517,7 @@ func (m *Manager) acquireTasks(envId uid.ID, taskDescriptors Descriptors) (err e
516517 timeReviveOffers := time .Now ()
517518 timeDeployMu := time .Now ()
518519 m .reviveOffersTrg <- struct {}{} // signal scheduler to revive offers
519- <- m .reviveOffersTrg // we only continue when it's done
520+ <- m .reviveOffersDone // we only continue when it's done
520521 utils .TimeTrack (timeReviveOffers , "acquireTasks: revive offers" ,
521522 log .WithField ("tasksToRun" , len (tasksToRun )).
522523 WithField ("partition" , envId ))
@@ -597,7 +598,7 @@ func (m *Manager) acquireTasks(envId uid.ID, taskDescriptors Descriptors) (err e
597598 // can't lock some of them, so we must roll back and keep them
598599 // unlocked in the roster.
599600 var deployedTaskIds []string
600- for taskPtr , _ := range deployedTasks {
601+ for taskPtr := range deployedTasks {
601602 taskPtr .SetParent (nil )
602603 deployedTaskIds = append (deployedTaskIds , taskPtr .taskId )
603604 }
@@ -612,11 +613,11 @@ func (m *Manager) acquireTasks(envId uid.ID, taskDescriptors Descriptors) (err e
612613 }
613614
614615 // Finally, we write to the roster. Point of no return!
615- for taskPtr , _ := range deployedTasks {
616+ for taskPtr := range deployedTasks {
616617 m .roster .append (taskPtr )
617618 }
618619 if deploymentSuccess {
619- for taskPtr , _ := range deployedTasks {
620+ for taskPtr := range deployedTasks {
620621 taskPtr .GetParent ().SetTask (taskPtr )
621622 }
622623 for taskPtr , descriptor := range tasksAlreadyRunning {
@@ -629,7 +630,6 @@ func (m *Manager) acquireTasks(envId uid.ID, taskDescriptors Descriptors) (err e
629630}
630631
631632func (m * Manager ) releaseTasks (envId uid.ID , tasks Tasks ) error {
632-
633633 taskReleaseErrors := make (map [string ]error )
634634 taskIdsReleased := make ([]string , 0 )
635635
@@ -686,7 +686,7 @@ func (m *Manager) configureTasks(envId uid.ID, tasks Tasks) error {
686686 taskPath := task .GetParentRolePath ()
687687 for inbChName , endpoint := range task .GetLocalBindMap () {
688688 var bindMapKey string
689- if strings .HasPrefix (inbChName , "::" ) { //global channel alias
689+ if strings .HasPrefix (inbChName , "::" ) { // global channel alias
690690 bindMapKey = inbChName
691691
692692 // deduplication
@@ -785,7 +785,6 @@ func (m *Manager) configureTasks(envId uid.ID, tasks Tasks) error {
785785func (m * Manager ) transitionTasks (envId uid.ID , tasks Tasks , src string , event string , dest string , commonArgs controlcommands.PropertyMap ) error {
786786 notify := make (chan controlcommands.MesosCommandResponse )
787787 receivers , err := tasks .GetMesosCommandTargets ()
788-
789788 if err != nil {
790789 return err
791790 }
@@ -870,7 +869,6 @@ func (m *Manager) TriggerHooks(envId uid.ID, tasks Tasks) error {
870869
871870 notify := make (chan controlcommands.MesosCommandResponse )
872871 receivers , err := tasks .GetMesosCommandTargets ()
873-
874872 if err != nil {
875873 return err
876874 }
@@ -935,7 +933,6 @@ func (m *Manager) GetTask(id string) *Task {
935933}
936934
937935func (m * Manager ) updateTaskState (taskId string , state string ) {
938-
939936 taskPtr := m .roster .getByTaskId (taskId )
940937 if taskPtr == nil {
941938 log .WithField ("taskId" , taskId ).
@@ -989,7 +986,7 @@ func (m *Manager) updateTaskStatus(status *mesos.TaskStatus) {
989986 }
990987 if ack , ok := m .ackKilledTasks .getValue (taskId ); ok {
991988 ack <- struct {}{}
992- //close(ack) // It can even be left open?
989+ // close(ack) // It can even be left open?
993990 }
994991
995992 return
@@ -1030,7 +1027,6 @@ func (m *Manager) updateTaskStatus(status *mesos.TaskStatus) {
10301027
10311028// Kill all tasks outside an environment (all unlocked tasks)
10321029func (m * Manager ) Cleanup () (killed Tasks , running Tasks , err error ) {
1033-
10341030 toKill := m .roster .filtered (func (t * Task ) bool {
10351031 return ! t .IsLocked ()
10361032 })
0 commit comments