mirror of
https://github.com/woodpecker-ci/woodpecker.git
synced 2024-12-18 08:26:45 +02:00
315 lines
7.7 KiB
Go
315 lines
7.7 KiB
Go
|
package main
|
||
|
|
||
|
import (
|
||
|
"context"
|
||
|
"encoding/json"
|
||
|
"io"
|
||
|
"io/ioutil"
|
||
|
"log"
|
||
|
"math"
|
||
|
"net/url"
|
||
|
"os"
|
||
|
"strconv"
|
||
|
"sync"
|
||
|
"time"
|
||
|
|
||
|
"github.com/laszlocph/drone-oss-08/cncd/pipeline/pipeline"
|
||
|
"github.com/laszlocph/drone-oss-08/cncd/pipeline/pipeline/backend"
|
||
|
"github.com/laszlocph/drone-oss-08/cncd/pipeline/pipeline/backend/docker"
|
||
|
"github.com/laszlocph/drone-oss-08/cncd/pipeline/pipeline/interrupt"
|
||
|
"github.com/laszlocph/drone-oss-08/cncd/pipeline/pipeline/multipart"
|
||
|
"github.com/laszlocph/drone-oss-08/cncd/pipeline/pipeline/rpc"
|
||
|
|
||
|
_ "github.com/joho/godotenv/autoload"
|
||
|
"github.com/tevino/abool"
|
||
|
"github.com/urfave/cli"
|
||
|
)
|
||
|
|
||
|
const (
|
||
|
maxFileUpload = 5000000
|
||
|
maxLogsUpload = 5000000
|
||
|
)
|
||
|
|
||
|
func main() {
|
||
|
app := cli.NewApp()
|
||
|
app.Name = "piped"
|
||
|
app.Usage = "piped stars a pipeline execution daemon"
|
||
|
app.Action = start
|
||
|
app.Flags = []cli.Flag{
|
||
|
cli.StringFlag{
|
||
|
Name: "endpoint",
|
||
|
EnvVar: "PIPED_ENDPOINT,PIPED_SERVER",
|
||
|
Value: "ws://localhost:9999",
|
||
|
},
|
||
|
cli.StringFlag{
|
||
|
Name: "token",
|
||
|
EnvVar: "PIPED_TOKEN,PIPED_SECRET",
|
||
|
},
|
||
|
cli.DurationFlag{
|
||
|
Name: "backoff",
|
||
|
EnvVar: "PIPED_BACKOFF",
|
||
|
Value: time.Second * 15,
|
||
|
},
|
||
|
cli.IntFlag{
|
||
|
Name: "retry-limit",
|
||
|
EnvVar: "PIPED_RETRY_LIMIT",
|
||
|
Value: math.MaxInt32,
|
||
|
},
|
||
|
cli.StringFlag{
|
||
|
Name: "platform",
|
||
|
EnvVar: "PIPED_PLATFORM",
|
||
|
Value: "linux/amd64",
|
||
|
},
|
||
|
cli.Int64Flag{
|
||
|
Name: "upload-limit",
|
||
|
EnvVar: "PIPED_UPLOAD_LIMIT",
|
||
|
Value: math.MaxInt32,
|
||
|
},
|
||
|
}
|
||
|
app.Commands = []cli.Command{
|
||
|
onceCommand,
|
||
|
}
|
||
|
|
||
|
if err := app.Run(os.Args); err != nil {
|
||
|
log.Fatalln(err)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func start(c *cli.Context) error {
|
||
|
endpoint, err := url.Parse(
|
||
|
c.String("endpoint"),
|
||
|
)
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
filter := rpc.Filter{
|
||
|
Labels: map[string]string{
|
||
|
"platform": c.String("platform"),
|
||
|
},
|
||
|
}
|
||
|
|
||
|
client, err := rpc.NewClient(
|
||
|
endpoint.String(),
|
||
|
rpc.WithRetryLimit(
|
||
|
c.Int("retry-limit"),
|
||
|
),
|
||
|
rpc.WithBackoff(
|
||
|
c.Duration("backoff"),
|
||
|
),
|
||
|
rpc.WithToken(
|
||
|
c.String("token"),
|
||
|
),
|
||
|
)
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
defer client.Close()
|
||
|
|
||
|
sigterm := abool.New()
|
||
|
ctx := context.Background()
|
||
|
ctx = interrupt.WithContextFunc(ctx, func() {
|
||
|
println("ctrl+c received, terminating process")
|
||
|
sigterm.Set()
|
||
|
})
|
||
|
|
||
|
for {
|
||
|
if sigterm.IsSet() {
|
||
|
return nil
|
||
|
}
|
||
|
if err := run(ctx, client, filter); err != nil {
|
||
|
return err
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func run(ctx context.Context, client rpc.Peer, filter rpc.Filter) error {
|
||
|
log.Println("pipeline: request next execution")
|
||
|
|
||
|
// get the next job from the queue
|
||
|
work, err := client.Next(ctx, filter)
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
if work == nil {
|
||
|
return nil
|
||
|
}
|
||
|
log.Printf("pipeline: received next execution: %s", work.ID)
|
||
|
if os.Getenv("SUICIDE_MODE") != "" {
|
||
|
os.Exit(1)
|
||
|
}
|
||
|
// new docker engine
|
||
|
engine, err := docker.NewEnv()
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
timeout := time.Hour
|
||
|
if minutes := work.Timeout; minutes != 0 {
|
||
|
timeout = time.Duration(minutes) * time.Minute
|
||
|
}
|
||
|
|
||
|
ctx, cancel := context.WithTimeout(context.Background(), timeout)
|
||
|
defer cancel()
|
||
|
|
||
|
cancelled := abool.New()
|
||
|
go func() {
|
||
|
werr := client.Wait(ctx, work.ID)
|
||
|
if werr != nil {
|
||
|
cancelled.SetTo(true) // TODO verify error is really an error
|
||
|
log.Printf("pipeline: cancel signal received: %s: %s", work.ID, werr)
|
||
|
cancel()
|
||
|
} else {
|
||
|
log.Printf("pipeline: cancel channel closed: %s", work.ID)
|
||
|
}
|
||
|
}()
|
||
|
|
||
|
go func() {
|
||
|
for {
|
||
|
select {
|
||
|
case <-ctx.Done():
|
||
|
log.Printf("pipeline: cancel ping loop: %s", work.ID)
|
||
|
return
|
||
|
case <-time.After(time.Minute):
|
||
|
log.Printf("pipeline: ping queue: %s", work.ID)
|
||
|
client.Extend(ctx, work.ID)
|
||
|
}
|
||
|
}
|
||
|
}()
|
||
|
|
||
|
state := rpc.State{}
|
||
|
state.Started = time.Now().Unix()
|
||
|
err = client.Init(context.Background(), work.ID, state)
|
||
|
if err != nil {
|
||
|
log.Printf("pipeline: error signaling pipeline init: %s: %s", work.ID, err)
|
||
|
}
|
||
|
|
||
|
var uploads sync.WaitGroup
|
||
|
defaultLogger := pipeline.LogFunc(func(proc *backend.Step, rc multipart.Reader) error {
|
||
|
part, rerr := rc.NextPart()
|
||
|
if rerr != nil {
|
||
|
return rerr
|
||
|
}
|
||
|
uploads.Add(1)
|
||
|
|
||
|
var secrets []string
|
||
|
for _, secret := range work.Config.Secrets {
|
||
|
if secret.Mask {
|
||
|
secrets = append(secrets, secret.Value)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
limitedPart := io.LimitReader(part, maxLogsUpload)
|
||
|
logstream := rpc.NewLineWriter(client, work.ID, proc.Alias, secrets...)
|
||
|
io.Copy(logstream, limitedPart)
|
||
|
|
||
|
file := &rpc.File{}
|
||
|
file.Mime = "application/json+logs"
|
||
|
file.Proc = proc.Alias
|
||
|
file.Name = "logs.json"
|
||
|
file.Data, _ = json.Marshal(logstream.Lines())
|
||
|
file.Size = len(file.Data)
|
||
|
file.Time = time.Now().Unix()
|
||
|
|
||
|
if serr := client.Upload(context.Background(), work.ID, file); serr != nil {
|
||
|
log.Printf("pipeline: cannot upload logs: %s: %s: %s", work.ID, file.Mime, serr)
|
||
|
} else {
|
||
|
log.Printf("pipeline: finish uploading logs: %s: step %s: %s", file.Mime, work.ID, proc.Alias)
|
||
|
}
|
||
|
|
||
|
defer func() {
|
||
|
log.Printf("pipeline: finish uploading logs: %s: step %s", work.ID, proc.Alias)
|
||
|
uploads.Done()
|
||
|
}()
|
||
|
|
||
|
part, rerr = rc.NextPart()
|
||
|
if rerr != nil {
|
||
|
return nil
|
||
|
}
|
||
|
// TODO should be configurable
|
||
|
limitedPart = io.LimitReader(part, maxFileUpload)
|
||
|
file = &rpc.File{}
|
||
|
file.Mime = part.Header().Get("Content-Type")
|
||
|
file.Proc = proc.Alias
|
||
|
file.Name = part.FileName()
|
||
|
file.Data, _ = ioutil.ReadAll(limitedPart)
|
||
|
file.Size = len(file.Data)
|
||
|
file.Time = time.Now().Unix()
|
||
|
|
||
|
if serr := client.Upload(context.Background(), work.ID, file); serr != nil {
|
||
|
log.Printf("pipeline: cannot upload artifact: %s: %s: %s", work.ID, file.Mime, serr)
|
||
|
} else {
|
||
|
log.Printf("pipeline: finish uploading artifact: %s: step %s: %s", file.Mime, work.ID, proc.Alias)
|
||
|
}
|
||
|
return nil
|
||
|
})
|
||
|
|
||
|
defaultTracer := pipeline.TraceFunc(func(state *pipeline.State) error {
|
||
|
procState := rpc.State{
|
||
|
Proc: state.Pipeline.Step.Alias,
|
||
|
Exited: state.Process.Exited,
|
||
|
ExitCode: state.Process.ExitCode,
|
||
|
Started: time.Now().Unix(), // TODO do not do this
|
||
|
Finished: time.Now().Unix(),
|
||
|
}
|
||
|
defer func() {
|
||
|
if uerr := client.Update(context.Background(), work.ID, procState); uerr != nil {
|
||
|
log.Printf("Pipeine: error updating pipeline step status: %s: %s: %s", work.ID, procState.Proc, uerr)
|
||
|
}
|
||
|
}()
|
||
|
if state.Process.Exited {
|
||
|
return nil
|
||
|
}
|
||
|
if state.Pipeline.Step.Environment == nil {
|
||
|
state.Pipeline.Step.Environment = map[string]string{}
|
||
|
}
|
||
|
state.Pipeline.Step.Environment["CI_BUILD_STATUS"] = "success"
|
||
|
state.Pipeline.Step.Environment["CI_BUILD_STARTED"] = strconv.FormatInt(state.Pipeline.Time, 10)
|
||
|
state.Pipeline.Step.Environment["CI_BUILD_FINISHED"] = strconv.FormatInt(time.Now().Unix(), 10)
|
||
|
|
||
|
state.Pipeline.Step.Environment["CI_JOB_STATUS"] = "success"
|
||
|
state.Pipeline.Step.Environment["CI_JOB_STARTED"] = strconv.FormatInt(state.Pipeline.Time, 10)
|
||
|
state.Pipeline.Step.Environment["CI_JOB_FINISHED"] = strconv.FormatInt(time.Now().Unix(), 10)
|
||
|
|
||
|
if state.Pipeline.Error != nil {
|
||
|
state.Pipeline.Step.Environment["CI_BUILD_STATUS"] = "failure"
|
||
|
state.Pipeline.Step.Environment["CI_JOB_STATUS"] = "failure"
|
||
|
}
|
||
|
return nil
|
||
|
})
|
||
|
|
||
|
err = pipeline.New(work.Config,
|
||
|
pipeline.WithContext(ctx),
|
||
|
pipeline.WithLogger(defaultLogger),
|
||
|
pipeline.WithTracer(defaultTracer),
|
||
|
pipeline.WithEngine(engine),
|
||
|
).Run()
|
||
|
|
||
|
state.Finished = time.Now().Unix()
|
||
|
state.Exited = true
|
||
|
if err != nil {
|
||
|
state.Error = err.Error()
|
||
|
if xerr, ok := err.(*pipeline.ExitError); ok {
|
||
|
state.ExitCode = xerr.Code
|
||
|
}
|
||
|
if xerr, ok := err.(*pipeline.OomError); ok {
|
||
|
state.ExitCode = xerr.Code
|
||
|
}
|
||
|
if cancelled.IsSet() {
|
||
|
state.ExitCode = 130
|
||
|
} else if state.ExitCode == 0 {
|
||
|
state.ExitCode = 1
|
||
|
}
|
||
|
}
|
||
|
|
||
|
log.Printf("pipeline: execution complete: %s", work.ID)
|
||
|
|
||
|
uploads.Wait()
|
||
|
|
||
|
err = client.Done(context.Background(), work.ID, state)
|
||
|
if err != nil {
|
||
|
log.Printf("Pipeine: error signaling pipeline done: %s: %s", work.ID, err)
|
||
|
}
|
||
|
|
||
|
return nil
|
||
|
}
|