2025-02-17 10:18:50 +01:00
|
|
|
package controllers
|
|
|
|
|
|
|
|
|
|
import (
|
|
|
|
|
"encoding/json"
|
2025-07-10 12:50:28 +02:00
|
|
|
"fmt"
|
2026-02-23 18:10:47 +01:00
|
|
|
"net/http"
|
2026-01-14 15:15:26 +01:00
|
|
|
"oc-scheduler/infrastructure"
|
2025-02-17 10:18:50 +01:00
|
|
|
|
|
|
|
|
oclib "cloud.o-forge.io/core/oc-lib"
|
|
|
|
|
"cloud.o-forge.io/core/oc-lib/dbs"
|
|
|
|
|
"cloud.o-forge.io/core/oc-lib/tools"
|
|
|
|
|
beego "github.com/beego/beego/v2/server/web"
|
|
|
|
|
"github.com/google/uuid"
|
2026-02-23 18:10:47 +01:00
|
|
|
gorillaws "github.com/gorilla/websocket"
|
2025-02-17 10:18:50 +01:00
|
|
|
)
|
|
|
|
|
|
|
|
|
|
var orderCollection = oclib.LibDataEnum(oclib.ORDER)
|
2025-04-30 17:55:46 +02:00
|
|
|
var logger = oclib.GetLogger()
|
2025-02-17 10:18:50 +01:00
|
|
|
|
|
|
|
|
// Operations about workflow
|
|
|
|
|
type WorkflowSchedulerController struct {
|
|
|
|
|
beego.Controller
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// @Title Schedule
|
|
|
|
|
// @Description schedule workflow
|
|
|
|
|
// @Param id path string true "id execution"
|
|
|
|
|
// @Param body body models.compute true "The compute content"
|
|
|
|
|
// @Success 200 {workspace} models.workspace
|
|
|
|
|
// @router /:id [post]
|
|
|
|
|
func (o *WorkflowSchedulerController) Schedule() {
|
2025-07-10 12:50:28 +02:00
|
|
|
logger := oclib.GetLogger()
|
|
|
|
|
|
2025-02-17 10:18:50 +01:00
|
|
|
code := 200
|
|
|
|
|
e := ""
|
|
|
|
|
user, peerID, groups := oclib.ExtractTokenInfo(*o.Ctx.Request)
|
2025-07-10 12:50:28 +02:00
|
|
|
wfId := o.Ctx.Input.Param(":id")
|
2026-01-14 15:15:26 +01:00
|
|
|
var resp *infrastructure.WorkflowSchedule
|
2025-02-17 10:18:50 +01:00
|
|
|
json.Unmarshal(o.Ctx.Input.CopyBody(100000), &resp)
|
2026-01-14 15:15:26 +01:00
|
|
|
|
2025-02-17 10:18:50 +01:00
|
|
|
caller := tools.NewHTTPCaller(map[tools.DataType]map[tools.METHOD]string{ // paths to call other OC services
|
|
|
|
|
tools.PEER: {
|
|
|
|
|
tools.POST: "/status/",
|
|
|
|
|
},
|
|
|
|
|
tools.BOOKING: {
|
|
|
|
|
tools.GET: "/booking/check/:id/:start_date/:end_date",
|
|
|
|
|
tools.POST: "/booking/",
|
|
|
|
|
},
|
|
|
|
|
})
|
2025-05-27 15:52:59 +02:00
|
|
|
|
2025-07-10 12:50:28 +02:00
|
|
|
logger.Info().Msg("Booking for " + wfId)
|
2026-02-20 10:35:02 +01:00
|
|
|
req := oclib.NewRequestAdmin(collection, caller)
|
|
|
|
|
// req := oclib.NewRequest(collection, user, peerID, groups, caller)
|
2025-02-17 10:18:50 +01:00
|
|
|
resp.UUID = uuid.New().String()
|
2026-02-20 10:35:02 +01:00
|
|
|
fmt.Println(user, peerID, groups)
|
2026-02-25 09:04:48 +01:00
|
|
|
sch, _, _, err := resp.Schedules(wfId, &tools.APIRequest{
|
2026-01-14 15:15:26 +01:00
|
|
|
Username: user,
|
|
|
|
|
PeerID: peerID,
|
|
|
|
|
Groups: groups,
|
|
|
|
|
Caller: caller,
|
2026-02-20 10:35:02 +01:00
|
|
|
Admin: true,
|
2026-01-14 15:15:26 +01:00
|
|
|
})
|
2025-02-17 10:18:50 +01:00
|
|
|
if err != nil {
|
2025-02-18 15:01:10 +01:00
|
|
|
if sch != nil {
|
|
|
|
|
for _, w := range sch.WorkflowExecution {
|
|
|
|
|
req.DeleteOne(w.GetID())
|
|
|
|
|
}
|
2025-02-17 10:18:50 +01:00
|
|
|
}
|
|
|
|
|
o.Data["json"] = map[string]interface{}{
|
|
|
|
|
"data": nil,
|
|
|
|
|
"code": 409,
|
2025-07-10 12:50:28 +02:00
|
|
|
"error": "Error when scheduling your execution(s): " + err.Error(),
|
2025-02-17 10:18:50 +01:00
|
|
|
}
|
|
|
|
|
o.ServeJSON()
|
|
|
|
|
return
|
|
|
|
|
}
|
|
|
|
|
o.Data["json"] = map[string]interface{}{
|
2025-02-18 15:01:10 +01:00
|
|
|
"data": sch.WorkflowExecution,
|
2025-02-17 10:18:50 +01:00
|
|
|
"code": code,
|
|
|
|
|
"error": e,
|
|
|
|
|
}
|
|
|
|
|
o.ServeJSON()
|
|
|
|
|
}
|
|
|
|
|
|
2026-02-23 18:10:47 +01:00
|
|
|
var wsUpgrader = gorillaws.Upgrader{
|
|
|
|
|
CheckOrigin: func(r *http.Request) bool { return true },
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// @Title CheckStream
|
|
|
|
|
// @Description WebSocket stream of slot availability for a workflow.
|
|
|
|
|
// After the handshake the client sends one JSON frame containing the
|
|
|
|
|
// WorkflowSchedule parameters (start, end, booking_mode, duration_s, …).
|
|
|
|
|
// The server responds with a CheckResult frame immediately and again each time
|
|
|
|
|
// a planner for one of the workflow's storage/compute peers is updated.
|
|
|
|
|
// When the stream is interrupted the cache entries for those peers are evicted
|
|
|
|
|
// and a PB_CLOSE_PLANNER event is emitted on NATS.
|
|
|
|
|
// Query params:
|
|
|
|
|
// - as_possible=true ignore start date, search from now
|
|
|
|
|
// - preemption=true validate anyway, raise warnings
|
|
|
|
|
//
|
|
|
|
|
// @Param id path string true "workflow id"
|
|
|
|
|
// @Param as_possible query bool false "find nearest free slot from now"
|
|
|
|
|
// @Param preemption query bool false "validate anyway, raise warnings"
|
|
|
|
|
// @Success 101
|
|
|
|
|
// @router /:id/check [get]
|
|
|
|
|
func (o *WorkflowSchedulerController) CheckStream() {
|
|
|
|
|
wfID := o.Ctx.Input.Param(":id")
|
|
|
|
|
asap, _ := o.GetBool("as_possible", false)
|
|
|
|
|
preemption, _ := o.GetBool("preemption", false)
|
|
|
|
|
|
|
|
|
|
user, peerID, groups := oclib.ExtractTokenInfo(*o.Ctx.Request)
|
|
|
|
|
req := &tools.APIRequest{
|
|
|
|
|
Username: user,
|
|
|
|
|
PeerID: peerID,
|
|
|
|
|
Groups: groups,
|
|
|
|
|
Caller: nil,
|
|
|
|
|
Admin: true,
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Resolve the peer IDs concerned by this workflow before upgrading so we
|
|
|
|
|
// can abort cleanly with a plain HTTP error if the workflow is not found.
|
|
|
|
|
watchedPeers, err := infrastructure.GetWorkflowPeerIDs(wfID, req)
|
|
|
|
|
if err != nil {
|
|
|
|
|
o.Data["json"] = map[string]interface{}{"code": 404, "error": err.Error()}
|
|
|
|
|
o.ServeJSON()
|
|
|
|
|
return
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Upgrade to WebSocket.
|
|
|
|
|
conn, err := wsUpgrader.Upgrade(o.Ctx.ResponseWriter, o.Ctx.Request, nil)
|
|
|
|
|
if err != nil {
|
|
|
|
|
return
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Read the schedule parameters sent by the client as the first message.
|
|
|
|
|
var ws infrastructure.WorkflowSchedule
|
|
|
|
|
if err := conn.ReadJSON(&ws); err != nil {
|
|
|
|
|
conn.Close()
|
|
|
|
|
return
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Subscribe to planner updates for the initially resolved peers and to
|
|
|
|
|
// workflow change notifications (peer list may change on workflow edit).
|
|
|
|
|
plannerCh, plannerUnsub := infrastructure.SubscribePlannerUpdates(watchedPeers)
|
|
|
|
|
wfCh, wfUnsub := infrastructure.SubscribeWorkflowUpdates(wfID)
|
|
|
|
|
|
|
|
|
|
// Cleanup on exit: cancel subscriptions, evict planner cache entries,
|
|
|
|
|
// signal PB_CLOSE_PLANNER on NATS for each peer that was being watched.
|
|
|
|
|
defer func() {
|
|
|
|
|
conn.Close()
|
|
|
|
|
plannerUnsub()
|
|
|
|
|
wfUnsub()
|
|
|
|
|
for _, peer := range watchedPeers {
|
|
|
|
|
if b, err := json.Marshal(map[string]interface{}{"peer_id": peer}); err == nil {
|
|
|
|
|
infrastructure.EmitNATS(peer, tools.PropalgationMessage{
|
|
|
|
|
Action: tools.PB_CLOSE_PLANNER,
|
|
|
|
|
Payload: b,
|
|
|
|
|
})
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}()
|
|
|
|
|
|
|
|
|
|
push := func() error {
|
|
|
|
|
result, checkErr := ws.Check(wfID, asap, preemption, req)
|
|
|
|
|
if checkErr != nil {
|
|
|
|
|
return checkErr
|
|
|
|
|
}
|
|
|
|
|
return conn.WriteJSON(result)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Initial check.
|
|
|
|
|
if err := push(); err != nil {
|
|
|
|
|
return
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// Detect client-side close in a separate goroutine.
|
|
|
|
|
closeCh := make(chan struct{})
|
|
|
|
|
go func() {
|
|
|
|
|
defer close(closeCh)
|
|
|
|
|
for {
|
|
|
|
|
if _, _, err := conn.ReadMessage(); err != nil {
|
|
|
|
|
return
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}()
|
|
|
|
|
|
|
|
|
|
// Stream loop.
|
|
|
|
|
for {
|
|
|
|
|
select {
|
|
|
|
|
case <-wfCh:
|
|
|
|
|
// The workflow was modified: refresh the peer list and re-subscribe
|
|
|
|
|
// so the stream watches the correct set of planners going forward.
|
|
|
|
|
if newPeers, err := infrastructure.GetWorkflowPeerIDs(wfID, req); err == nil {
|
|
|
|
|
plannerUnsub()
|
|
|
|
|
watchedPeers = newPeers
|
|
|
|
|
plannerCh, plannerUnsub = infrastructure.SubscribePlannerUpdates(newPeers)
|
|
|
|
|
}
|
|
|
|
|
if err := push(); err != nil {
|
|
|
|
|
return
|
|
|
|
|
}
|
|
|
|
|
case <-plannerCh:
|
|
|
|
|
if err := push(); err != nil {
|
|
|
|
|
return
|
|
|
|
|
}
|
|
|
|
|
case <-closeCh:
|
|
|
|
|
return
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2025-02-17 10:18:50 +01:00
|
|
|
// @Title UnSchedule
|
|
|
|
|
// @Description schedule workflow
|
|
|
|
|
// @Param id path string true "id execution"
|
|
|
|
|
// @Param body body models.compute true "The compute content"
|
|
|
|
|
// @Success 200 {workspace} models.workspace
|
|
|
|
|
// @router /:id [delete]
|
|
|
|
|
func (o *WorkflowSchedulerController) UnSchedule() {
|
2026-02-20 10:35:02 +01:00
|
|
|
// user, peerID, groups := oclib.ExtractTokenInfo(*o.Ctx.Request)
|
2025-02-17 10:18:50 +01:00
|
|
|
id := o.Ctx.Input.Param(":id")
|
|
|
|
|
// TODO UNSCHEDULER
|
|
|
|
|
filter := &dbs.Filters{
|
|
|
|
|
And: map[string][]dbs.Filter{
|
|
|
|
|
"workflow_id": {{Operator: dbs.EQUAL.String(), Value: id}},
|
|
|
|
|
},
|
|
|
|
|
}
|
2026-02-20 10:35:02 +01:00
|
|
|
o.Data["json"] = oclib.NewRequestAdmin(collection, nil).Search(filter, "", true)
|
|
|
|
|
|
|
|
|
|
// o.Data["json"] = oclib.NewRequest(collection, user, peerID, groups, nil).Search(filter, "", true)
|
2025-02-17 10:18:50 +01:00
|
|
|
o.ServeJSON()
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// @Title SearchScheduledDraftOrder
|
|
|
|
|
// @Description schedule workflow
|
|
|
|
|
// @Param id path string true "id execution"
|
|
|
|
|
// @Success 200 {workspace} models.workspace
|
|
|
|
|
// @router /:id/order [get]
|
|
|
|
|
func (o *WorkflowSchedulerController) SearchScheduledDraftOrder() {
|
|
|
|
|
user, peerID, groups := oclib.ExtractTokenInfo(*o.Ctx.Request)
|
|
|
|
|
id := o.Ctx.Input.Param(":id")
|
2026-02-20 10:35:02 +01:00
|
|
|
fmt.Println(user, peerID, groups)
|
2025-02-17 10:18:50 +01:00
|
|
|
filter := &dbs.Filters{
|
|
|
|
|
And: map[string][]dbs.Filter{
|
|
|
|
|
"workflow_id": {{Operator: dbs.EQUAL.String(), Value: id}},
|
|
|
|
|
"order_by": {{Operator: dbs.EQUAL.String(), Value: peerID}},
|
|
|
|
|
},
|
|
|
|
|
}
|
2026-02-20 10:35:02 +01:00
|
|
|
o.Data["json"] = oclib.NewRequestAdmin(orderCollection, nil).Search(filter, "", true)
|
|
|
|
|
|
|
|
|
|
//o.Data["json"] = oclib.NewRequest(orderCollection, user, peerID, groups, nil).Search(filter, "", true)
|
2025-02-17 10:18:50 +01:00
|
|
|
o.ServeJSON()
|
|
|
|
|
}
|