mirror of
https://gitea.com/Lydanne/buildx.git
synced 2025-05-18 00:47:48 +08:00

Previously, we would fail while trying to use the global "target" field when using a matrix. The contents of the matrix really don't matter for this. What was happening was that we would copy the "target" property into the child evaluation context, so that when it was updated on the parent, it wouldn't propagate to the child. The correct behavior here is to avoid copying variables from the target evaluation context if it is the root. Signed-off-by: Justin Chadwell <me@jedevc.com>
1418 lines
34 KiB
Go
1418 lines
34 KiB
Go
package bake
|
|
|
|
import (
|
|
"context"
|
|
"encoding/csv"
|
|
"io"
|
|
"os"
|
|
"path"
|
|
"path/filepath"
|
|
"regexp"
|
|
"sort"
|
|
"strconv"
|
|
"strings"
|
|
"time"
|
|
|
|
composecli "github.com/compose-spec/compose-go/cli"
|
|
"github.com/docker/buildx/bake/hclparser"
|
|
"github.com/docker/buildx/build"
|
|
controllerapi "github.com/docker/buildx/controller/pb"
|
|
"github.com/docker/buildx/util/buildflags"
|
|
"github.com/docker/buildx/util/platformutil"
|
|
"github.com/docker/buildx/util/progress"
|
|
"github.com/docker/cli/cli/config"
|
|
hcl "github.com/hashicorp/hcl/v2"
|
|
"github.com/moby/buildkit/client"
|
|
"github.com/moby/buildkit/client/llb"
|
|
"github.com/moby/buildkit/session/auth/authprovider"
|
|
"github.com/pkg/errors"
|
|
"github.com/zclconf/go-cty/cty"
|
|
"github.com/zclconf/go-cty/cty/convert"
|
|
)
|
|
|
|
var (
|
|
validTargetNameChars = `[a-zA-Z0-9_-]+`
|
|
targetNamePattern = regexp.MustCompile(`^` + validTargetNameChars + `$`)
|
|
)
|
|
|
|
type File struct {
|
|
Name string
|
|
Data []byte
|
|
}
|
|
|
|
type Override struct {
|
|
Value string
|
|
ArrValue []string
|
|
}
|
|
|
|
func defaultFilenames() []string {
|
|
names := []string{}
|
|
names = append(names, composecli.DefaultFileNames...)
|
|
names = append(names, []string{
|
|
"docker-bake.json",
|
|
"docker-bake.override.json",
|
|
"docker-bake.hcl",
|
|
"docker-bake.override.hcl",
|
|
}...)
|
|
return names
|
|
}
|
|
|
|
func ReadLocalFiles(names []string, stdin io.Reader, l progress.SubLogger) ([]File, error) {
|
|
isDefault := false
|
|
if len(names) == 0 {
|
|
isDefault = true
|
|
names = defaultFilenames()
|
|
}
|
|
out := make([]File, 0, len(names))
|
|
|
|
setStatus := func(st *client.VertexStatus) {
|
|
if l != nil {
|
|
l.SetStatus(st)
|
|
}
|
|
}
|
|
|
|
for _, n := range names {
|
|
var dt []byte
|
|
var err error
|
|
if n == "-" {
|
|
dt, err = readWithProgress(stdin, setStatus)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
} else {
|
|
dt, err = readFileWithProgress(n, isDefault, setStatus)
|
|
if dt == nil && err == nil {
|
|
continue
|
|
}
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
out = append(out, File{Name: n, Data: dt})
|
|
}
|
|
return out, nil
|
|
}
|
|
|
|
func readFileWithProgress(fname string, isDefault bool, setStatus func(st *client.VertexStatus)) (dt []byte, err error) {
|
|
st := &client.VertexStatus{
|
|
ID: "reading " + fname,
|
|
}
|
|
|
|
defer func() {
|
|
now := time.Now()
|
|
st.Completed = &now
|
|
if dt != nil || err != nil {
|
|
setStatus(st)
|
|
}
|
|
}()
|
|
|
|
now := time.Now()
|
|
st.Started = &now
|
|
|
|
f, err := os.Open(fname)
|
|
if err != nil {
|
|
if isDefault && errors.Is(err, os.ErrNotExist) {
|
|
return nil, nil
|
|
}
|
|
return nil, err
|
|
}
|
|
defer f.Close()
|
|
setStatus(st)
|
|
|
|
info, err := f.Stat()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
st.Total = info.Size()
|
|
setStatus(st)
|
|
|
|
buf := make([]byte, 1024)
|
|
for {
|
|
n, err := f.Read(buf)
|
|
if err == io.EOF {
|
|
break
|
|
}
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
dt = append(dt, buf[:n]...)
|
|
st.Current += int64(n)
|
|
setStatus(st)
|
|
}
|
|
|
|
return dt, nil
|
|
}
|
|
|
|
func readWithProgress(r io.Reader, setStatus func(st *client.VertexStatus)) (dt []byte, err error) {
|
|
st := &client.VertexStatus{
|
|
ID: "reading from stdin",
|
|
}
|
|
|
|
defer func() {
|
|
now := time.Now()
|
|
st.Completed = &now
|
|
setStatus(st)
|
|
}()
|
|
|
|
now := time.Now()
|
|
st.Started = &now
|
|
setStatus(st)
|
|
|
|
buf := make([]byte, 1024)
|
|
for {
|
|
n, err := r.Read(buf)
|
|
if err == io.EOF {
|
|
break
|
|
}
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
dt = append(dt, buf[:n]...)
|
|
st.Current += int64(n)
|
|
setStatus(st)
|
|
}
|
|
|
|
return dt, nil
|
|
}
|
|
|
|
func ListTargets(files []File) ([]string, error) {
|
|
c, err := ParseFiles(files, nil)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
var targets []string
|
|
for _, g := range c.Groups {
|
|
targets = append(targets, g.Name)
|
|
}
|
|
for _, t := range c.Targets {
|
|
targets = append(targets, t.Name)
|
|
}
|
|
return dedupSlice(targets), nil
|
|
}
|
|
|
|
func ReadTargets(ctx context.Context, files []File, targets, overrides []string, defaults map[string]string) (map[string]*Target, map[string]*Group, error) {
|
|
c, err := ParseFiles(files, defaults)
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
for i, t := range targets {
|
|
targets[i] = sanitizeTargetName(t)
|
|
}
|
|
|
|
o, err := c.newOverrides(overrides)
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
m := map[string]*Target{}
|
|
n := map[string]*Group{}
|
|
for _, target := range targets {
|
|
ts, gs := c.ResolveGroup(target)
|
|
for _, tname := range ts {
|
|
t, err := c.ResolveTarget(tname, o)
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
if t != nil {
|
|
m[tname] = t
|
|
}
|
|
}
|
|
for _, gname := range gs {
|
|
for _, group := range c.Groups {
|
|
if group.Name == gname {
|
|
n[gname] = group
|
|
break
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
for _, target := range targets {
|
|
if target == "default" {
|
|
continue
|
|
}
|
|
if _, ok := n["default"]; !ok {
|
|
n["default"] = &Group{Name: "default"}
|
|
}
|
|
n["default"].Targets = append(n["default"].Targets, target)
|
|
}
|
|
if g, ok := n["default"]; ok {
|
|
g.Targets = dedupSlice(g.Targets)
|
|
}
|
|
|
|
for name, t := range m {
|
|
if err := c.loadLinks(name, t, m, o, nil); err != nil {
|
|
return nil, nil, err
|
|
}
|
|
}
|
|
|
|
return m, n, nil
|
|
}
|
|
|
|
func dedupSlice(s []string) []string {
|
|
if len(s) == 0 {
|
|
return s
|
|
}
|
|
var res []string
|
|
seen := make(map[string]struct{})
|
|
for _, val := range s {
|
|
if _, ok := seen[val]; !ok {
|
|
res = append(res, val)
|
|
seen[val] = struct{}{}
|
|
}
|
|
}
|
|
return res
|
|
}
|
|
|
|
func dedupMap(ms ...map[string]string) map[string]string {
|
|
if len(ms) == 0 {
|
|
return nil
|
|
}
|
|
res := map[string]string{}
|
|
for _, m := range ms {
|
|
if len(m) == 0 {
|
|
continue
|
|
}
|
|
for k, v := range m {
|
|
if _, ok := res[k]; !ok {
|
|
res[k] = v
|
|
}
|
|
}
|
|
}
|
|
return res
|
|
}
|
|
|
|
func sliceToMap(env []string) (res map[string]string) {
|
|
res = make(map[string]string)
|
|
for _, s := range env {
|
|
kv := strings.SplitN(s, "=", 2)
|
|
key := kv[0]
|
|
switch {
|
|
case len(kv) == 1:
|
|
res[key] = ""
|
|
default:
|
|
res[key] = kv[1]
|
|
}
|
|
}
|
|
return
|
|
}
|
|
|
|
func ParseFiles(files []File, defaults map[string]string) (_ *Config, err error) {
|
|
defer func() {
|
|
err = formatHCLError(err, files)
|
|
}()
|
|
|
|
var c Config
|
|
var composeFiles []File
|
|
var hclFiles []*hcl.File
|
|
for _, f := range files {
|
|
isCompose, composeErr := validateComposeFile(f.Data, f.Name)
|
|
if isCompose {
|
|
if composeErr != nil {
|
|
return nil, composeErr
|
|
}
|
|
composeFiles = append(composeFiles, f)
|
|
}
|
|
if !isCompose {
|
|
hf, isHCL, err := ParseHCLFile(f.Data, f.Name)
|
|
if isHCL {
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
hclFiles = append(hclFiles, hf)
|
|
} else if composeErr != nil {
|
|
return nil, errors.Wrapf(err, "failed to parse %s: parsing yaml: %v, parsing hcl", f.Name, composeErr)
|
|
} else {
|
|
return nil, err
|
|
}
|
|
}
|
|
}
|
|
|
|
if len(composeFiles) > 0 {
|
|
cfg, cmperr := ParseComposeFiles(composeFiles)
|
|
if cmperr != nil {
|
|
return nil, errors.Wrap(cmperr, "failed to parse compose file")
|
|
}
|
|
c = mergeConfig(c, *cfg)
|
|
c = dedupeConfig(c)
|
|
}
|
|
|
|
if len(hclFiles) > 0 {
|
|
renamed, err := hclparser.Parse(hclparser.MergeFiles(hclFiles), hclparser.Opt{
|
|
LookupVar: os.LookupEnv,
|
|
Vars: defaults,
|
|
ValidateLabel: validateTargetName,
|
|
}, &c)
|
|
if err.HasErrors() {
|
|
return nil, err
|
|
}
|
|
|
|
for _, renamed := range renamed {
|
|
for oldName, newNames := range renamed {
|
|
newNames = dedupSlice(newNames)
|
|
if len(newNames) == 1 && oldName == newNames[0] {
|
|
continue
|
|
}
|
|
c.Groups = append(c.Groups, &Group{
|
|
Name: oldName,
|
|
Targets: newNames,
|
|
})
|
|
}
|
|
}
|
|
c = dedupeConfig(c)
|
|
}
|
|
|
|
return &c, nil
|
|
}
|
|
|
|
func dedupeConfig(c Config) Config {
|
|
c2 := c
|
|
c2.Groups = make([]*Group, 0, len(c2.Groups))
|
|
for _, g := range c.Groups {
|
|
g1 := *g
|
|
g1.Targets = dedupSlice(g1.Targets)
|
|
c2.Groups = append(c2.Groups, &g1)
|
|
}
|
|
c2.Targets = make([]*Target, 0, len(c2.Targets))
|
|
mt := map[string]*Target{}
|
|
for _, t := range c.Targets {
|
|
if t2, ok := mt[t.Name]; ok {
|
|
t2.Merge(t)
|
|
} else {
|
|
mt[t.Name] = t
|
|
c2.Targets = append(c2.Targets, t)
|
|
}
|
|
}
|
|
return c2
|
|
}
|
|
|
|
func ParseFile(dt []byte, fn string) (*Config, error) {
|
|
return ParseFiles([]File{{Data: dt, Name: fn}}, nil)
|
|
}
|
|
|
|
type Config struct {
|
|
Groups []*Group `json:"group" hcl:"group,block" cty:"group"`
|
|
Targets []*Target `json:"target" hcl:"target,block" cty:"target"`
|
|
}
|
|
|
|
func mergeConfig(c1, c2 Config) Config {
|
|
if c1.Groups == nil {
|
|
c1.Groups = []*Group{}
|
|
}
|
|
|
|
for _, g2 := range c2.Groups {
|
|
var g1 *Group
|
|
for _, g := range c1.Groups {
|
|
if g2.Name == g.Name {
|
|
g1 = g
|
|
break
|
|
}
|
|
}
|
|
if g1 == nil {
|
|
c1.Groups = append(c1.Groups, g2)
|
|
continue
|
|
}
|
|
|
|
nextTarget:
|
|
for _, t2 := range g2.Targets {
|
|
for _, t1 := range g1.Targets {
|
|
if t1 == t2 {
|
|
continue nextTarget
|
|
}
|
|
}
|
|
g1.Targets = append(g1.Targets, t2)
|
|
}
|
|
c1.Groups = append(c1.Groups, g1)
|
|
}
|
|
|
|
if c1.Targets == nil {
|
|
c1.Targets = []*Target{}
|
|
}
|
|
|
|
for _, t2 := range c2.Targets {
|
|
var t1 *Target
|
|
for _, t := range c1.Targets {
|
|
if t2.Name == t.Name {
|
|
t1 = t
|
|
break
|
|
}
|
|
}
|
|
if t1 != nil {
|
|
t1.Merge(t2)
|
|
t2 = t1
|
|
}
|
|
c1.Targets = append(c1.Targets, t2)
|
|
}
|
|
|
|
return c1
|
|
}
|
|
|
|
func (c Config) expandTargets(pattern string) ([]string, error) {
|
|
for _, target := range c.Targets {
|
|
if target.Name == pattern {
|
|
return []string{pattern}, nil
|
|
}
|
|
}
|
|
|
|
var names []string
|
|
for _, target := range c.Targets {
|
|
ok, err := path.Match(pattern, target.Name)
|
|
if err != nil {
|
|
return nil, errors.Wrapf(err, "could not match targets with '%s'", pattern)
|
|
}
|
|
if ok {
|
|
names = append(names, target.Name)
|
|
}
|
|
}
|
|
if len(names) == 0 {
|
|
return nil, errors.Errorf("could not find any target matching '%s'", pattern)
|
|
}
|
|
return names, nil
|
|
}
|
|
|
|
func (c Config) loadLinks(name string, t *Target, m map[string]*Target, o map[string]map[string]Override, visited []string) error {
|
|
visited = append(visited, name)
|
|
for _, v := range t.Contexts {
|
|
if strings.HasPrefix(v, "target:") {
|
|
target := strings.TrimPrefix(v, "target:")
|
|
if target == t.Name {
|
|
return errors.Errorf("target %s cannot link to itself", target)
|
|
}
|
|
for _, v := range visited {
|
|
if v == target {
|
|
return errors.Errorf("infinite loop from %s to %s", name, target)
|
|
}
|
|
}
|
|
t2, ok := m[target]
|
|
if !ok {
|
|
var err error
|
|
t2, err = c.ResolveTarget(target, o)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
t2.Outputs = nil
|
|
t2.linked = true
|
|
m[target] = t2
|
|
}
|
|
if err := c.loadLinks(target, t2, m, o, visited); err != nil {
|
|
return err
|
|
}
|
|
if len(t.Platforms) > 1 && len(t2.Platforms) > 1 {
|
|
if !sliceEqual(t.Platforms, t2.Platforms) {
|
|
return errors.Errorf("target %s can't be used by %s because it is defined for different platforms %v and %v", target, name, t2.Platforms, t.Platforms)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (c Config) newOverrides(v []string) (map[string]map[string]Override, error) {
|
|
m := map[string]map[string]Override{}
|
|
for _, v := range v {
|
|
parts := strings.SplitN(v, "=", 2)
|
|
keys := strings.SplitN(parts[0], ".", 3)
|
|
if len(keys) < 2 {
|
|
return nil, errors.Errorf("invalid override key %s, expected target.name", parts[0])
|
|
}
|
|
|
|
pattern := keys[0]
|
|
if len(parts) != 2 && keys[1] != "args" {
|
|
return nil, errors.Errorf("invalid override %s, expected target.name=value", v)
|
|
}
|
|
|
|
names, err := c.expandTargets(pattern)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
kk := strings.SplitN(parts[0], ".", 2)
|
|
|
|
for _, name := range names {
|
|
t, ok := m[name]
|
|
if !ok {
|
|
t = map[string]Override{}
|
|
m[name] = t
|
|
}
|
|
|
|
o := t[kk[1]]
|
|
|
|
switch keys[1] {
|
|
case "output", "cache-to", "cache-from", "tags", "platform", "secrets", "ssh", "attest":
|
|
if len(parts) == 2 {
|
|
o.ArrValue = append(o.ArrValue, parts[1])
|
|
}
|
|
case "args":
|
|
if len(keys) != 3 {
|
|
return nil, errors.Errorf("invalid key %s, args requires name", parts[0])
|
|
}
|
|
if len(parts) < 2 {
|
|
v, ok := os.LookupEnv(keys[2])
|
|
if !ok {
|
|
continue
|
|
}
|
|
o.Value = v
|
|
}
|
|
fallthrough
|
|
case "contexts":
|
|
if len(keys) != 3 {
|
|
return nil, errors.Errorf("invalid key %s, contexts requires name", parts[0])
|
|
}
|
|
fallthrough
|
|
default:
|
|
if len(parts) == 2 {
|
|
o.Value = parts[1]
|
|
}
|
|
}
|
|
|
|
t[kk[1]] = o
|
|
}
|
|
}
|
|
return m, nil
|
|
}
|
|
|
|
func (c Config) ResolveGroup(name string) ([]string, []string) {
|
|
targets, groups := c.group(name, map[string]visit{})
|
|
return dedupSlice(targets), dedupSlice(groups)
|
|
}
|
|
|
|
type visit struct {
|
|
target []string
|
|
group []string
|
|
}
|
|
|
|
func (c Config) group(name string, visited map[string]visit) ([]string, []string) {
|
|
if v, ok := visited[name]; ok {
|
|
return v.target, v.group
|
|
}
|
|
var g *Group
|
|
for _, group := range c.Groups {
|
|
if group.Name == name {
|
|
g = group
|
|
break
|
|
}
|
|
}
|
|
if g == nil {
|
|
return []string{name}, nil
|
|
}
|
|
visited[name] = visit{}
|
|
targets := make([]string, 0, len(g.Targets))
|
|
groups := []string{name}
|
|
for _, t := range g.Targets {
|
|
ttarget, tgroup := c.group(t, visited)
|
|
if len(ttarget) > 0 {
|
|
targets = append(targets, ttarget...)
|
|
} else {
|
|
targets = append(targets, t)
|
|
}
|
|
if len(tgroup) > 0 {
|
|
groups = append(groups, tgroup...)
|
|
}
|
|
}
|
|
visited[name] = visit{target: targets, group: groups}
|
|
return targets, groups
|
|
}
|
|
|
|
func (c Config) ResolveTarget(name string, overrides map[string]map[string]Override) (*Target, error) {
|
|
t, err := c.target(name, map[string]*Target{}, overrides)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
t.Inherits = nil
|
|
if t.Context == nil {
|
|
s := "."
|
|
t.Context = &s
|
|
}
|
|
if t.Dockerfile == nil {
|
|
s := "Dockerfile"
|
|
t.Dockerfile = &s
|
|
}
|
|
return t, nil
|
|
}
|
|
|
|
func (c Config) target(name string, visited map[string]*Target, overrides map[string]map[string]Override) (*Target, error) {
|
|
if t, ok := visited[name]; ok {
|
|
return t, nil
|
|
}
|
|
visited[name] = nil
|
|
var t *Target
|
|
for _, target := range c.Targets {
|
|
if target.Name == name {
|
|
t = target
|
|
break
|
|
}
|
|
}
|
|
if t == nil {
|
|
return nil, errors.Errorf("failed to find target %s", name)
|
|
}
|
|
tt := &Target{}
|
|
for _, name := range t.Inherits {
|
|
t, err := c.target(name, visited, overrides)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if t != nil {
|
|
tt.Merge(t)
|
|
}
|
|
}
|
|
m := defaultTarget()
|
|
m.Merge(tt)
|
|
m.Merge(t)
|
|
tt = m
|
|
if err := tt.AddOverrides(overrides[name]); err != nil {
|
|
return nil, err
|
|
}
|
|
tt.normalize()
|
|
visited[name] = tt
|
|
return tt, nil
|
|
}
|
|
|
|
type Group struct {
|
|
Name string `json:"-" hcl:"name,label" cty:"name"`
|
|
Targets []string `json:"targets" hcl:"targets" cty:"targets"`
|
|
// Target // TODO?
|
|
}
|
|
|
|
type Target struct {
|
|
Name string `json:"-" hcl:"name,label" cty:"name"`
|
|
|
|
// Inherits is the only field that cannot be overridden with --set
|
|
Inherits []string `json:"inherits,omitempty" hcl:"inherits,optional" cty:"inherits"`
|
|
|
|
Annotations []string `json:"annotations,omitempty" hcl:"annotations,optional" cty:"annotations"`
|
|
Attest []string `json:"attest,omitempty" hcl:"attest,optional" cty:"attest"`
|
|
Context *string `json:"context,omitempty" hcl:"context,optional" cty:"context"`
|
|
Contexts map[string]string `json:"contexts,omitempty" hcl:"contexts,optional" cty:"contexts"`
|
|
Dockerfile *string `json:"dockerfile,omitempty" hcl:"dockerfile,optional" cty:"dockerfile"`
|
|
DockerfileInline *string `json:"dockerfile-inline,omitempty" hcl:"dockerfile-inline,optional" cty:"dockerfile-inline"`
|
|
Args map[string]*string `json:"args,omitempty" hcl:"args,optional" cty:"args"`
|
|
Labels map[string]*string `json:"labels,omitempty" hcl:"labels,optional" cty:"labels"`
|
|
Tags []string `json:"tags,omitempty" hcl:"tags,optional" cty:"tags"`
|
|
CacheFrom []string `json:"cache-from,omitempty" hcl:"cache-from,optional" cty:"cache-from"`
|
|
CacheTo []string `json:"cache-to,omitempty" hcl:"cache-to,optional" cty:"cache-to"`
|
|
Target *string `json:"target,omitempty" hcl:"target,optional" cty:"target"`
|
|
Secrets []string `json:"secret,omitempty" hcl:"secret,optional" cty:"secret"`
|
|
SSH []string `json:"ssh,omitempty" hcl:"ssh,optional" cty:"ssh"`
|
|
Platforms []string `json:"platforms,omitempty" hcl:"platforms,optional" cty:"platforms"`
|
|
Outputs []string `json:"output,omitempty" hcl:"output,optional" cty:"output"`
|
|
Pull *bool `json:"pull,omitempty" hcl:"pull,optional" cty:"pull"`
|
|
NoCache *bool `json:"no-cache,omitempty" hcl:"no-cache,optional" cty:"no-cache"`
|
|
NetworkMode *string `json:"-" hcl:"-" cty:"-"`
|
|
NoCacheFilter []string `json:"no-cache-filter,omitempty" hcl:"no-cache-filter,optional" cty:"no-cache-filter"`
|
|
// IMPORTANT: if you add more fields here, do not forget to update newOverrides and docs/bake-reference.md.
|
|
|
|
// linked is a private field to mark a target used as a linked one
|
|
linked bool
|
|
}
|
|
|
|
var _ hclparser.WithEvalContexts = &Target{}
|
|
var _ hclparser.WithGetName = &Target{}
|
|
var _ hclparser.WithEvalContexts = &Group{}
|
|
var _ hclparser.WithGetName = &Group{}
|
|
|
|
func (t *Target) normalize() {
|
|
t.Annotations = removeDupes(t.Annotations)
|
|
t.Attest = removeAttestDupes(t.Attest)
|
|
t.Tags = removeDupes(t.Tags)
|
|
t.Secrets = removeDupes(t.Secrets)
|
|
t.SSH = removeDupes(t.SSH)
|
|
t.Platforms = removeDupes(t.Platforms)
|
|
t.CacheFrom = removeDupes(t.CacheFrom)
|
|
t.CacheTo = removeDupes(t.CacheTo)
|
|
t.Outputs = removeDupes(t.Outputs)
|
|
t.NoCacheFilter = removeDupes(t.NoCacheFilter)
|
|
|
|
for k, v := range t.Contexts {
|
|
if v == "" {
|
|
delete(t.Contexts, k)
|
|
}
|
|
}
|
|
if len(t.Contexts) == 0 {
|
|
t.Contexts = nil
|
|
}
|
|
}
|
|
|
|
func (t *Target) Merge(t2 *Target) {
|
|
if t2.Context != nil {
|
|
t.Context = t2.Context
|
|
}
|
|
if t2.Dockerfile != nil {
|
|
t.Dockerfile = t2.Dockerfile
|
|
}
|
|
if t2.DockerfileInline != nil {
|
|
t.DockerfileInline = t2.DockerfileInline
|
|
}
|
|
for k, v := range t2.Args {
|
|
if v == nil {
|
|
continue
|
|
}
|
|
if t.Args == nil {
|
|
t.Args = map[string]*string{}
|
|
}
|
|
t.Args[k] = v
|
|
}
|
|
for k, v := range t2.Contexts {
|
|
if t.Contexts == nil {
|
|
t.Contexts = map[string]string{}
|
|
}
|
|
t.Contexts[k] = v
|
|
}
|
|
for k, v := range t2.Labels {
|
|
if v == nil {
|
|
continue
|
|
}
|
|
if t.Labels == nil {
|
|
t.Labels = map[string]*string{}
|
|
}
|
|
t.Labels[k] = v
|
|
}
|
|
if t2.Tags != nil { // no merge
|
|
t.Tags = t2.Tags
|
|
}
|
|
if t2.Target != nil {
|
|
t.Target = t2.Target
|
|
}
|
|
if t2.Annotations != nil { // merge
|
|
t.Annotations = append(t.Annotations, t2.Annotations...)
|
|
}
|
|
if t2.Attest != nil { // merge
|
|
t.Attest = append(t.Attest, t2.Attest...)
|
|
t.Attest = removeAttestDupes(t.Attest)
|
|
}
|
|
if t2.Secrets != nil { // merge
|
|
t.Secrets = append(t.Secrets, t2.Secrets...)
|
|
}
|
|
if t2.SSH != nil { // merge
|
|
t.SSH = append(t.SSH, t2.SSH...)
|
|
}
|
|
if t2.Platforms != nil { // no merge
|
|
t.Platforms = t2.Platforms
|
|
}
|
|
if t2.CacheFrom != nil { // merge
|
|
t.CacheFrom = append(t.CacheFrom, t2.CacheFrom...)
|
|
}
|
|
if t2.CacheTo != nil { // no merge
|
|
t.CacheTo = t2.CacheTo
|
|
}
|
|
if t2.Outputs != nil { // no merge
|
|
t.Outputs = t2.Outputs
|
|
}
|
|
if t2.Pull != nil {
|
|
t.Pull = t2.Pull
|
|
}
|
|
if t2.NoCache != nil {
|
|
t.NoCache = t2.NoCache
|
|
}
|
|
if t2.NetworkMode != nil {
|
|
t.NetworkMode = t2.NetworkMode
|
|
}
|
|
if t2.NoCacheFilter != nil { // merge
|
|
t.NoCacheFilter = append(t.NoCacheFilter, t2.NoCacheFilter...)
|
|
}
|
|
t.Inherits = append(t.Inherits, t2.Inherits...)
|
|
}
|
|
|
|
func (t *Target) AddOverrides(overrides map[string]Override) error {
|
|
for key, o := range overrides {
|
|
value := o.Value
|
|
keys := strings.SplitN(key, ".", 2)
|
|
switch keys[0] {
|
|
case "context":
|
|
t.Context = &value
|
|
case "dockerfile":
|
|
t.Dockerfile = &value
|
|
case "args":
|
|
if len(keys) != 2 {
|
|
return errors.Errorf("args require name")
|
|
}
|
|
if t.Args == nil {
|
|
t.Args = map[string]*string{}
|
|
}
|
|
t.Args[keys[1]] = &value
|
|
case "contexts":
|
|
if len(keys) != 2 {
|
|
return errors.Errorf("contexts require name")
|
|
}
|
|
if t.Contexts == nil {
|
|
t.Contexts = map[string]string{}
|
|
}
|
|
t.Contexts[keys[1]] = value
|
|
case "labels":
|
|
if len(keys) != 2 {
|
|
return errors.Errorf("labels require name")
|
|
}
|
|
if t.Labels == nil {
|
|
t.Labels = map[string]*string{}
|
|
}
|
|
t.Labels[keys[1]] = &value
|
|
case "tags":
|
|
t.Tags = o.ArrValue
|
|
case "cache-from":
|
|
t.CacheFrom = o.ArrValue
|
|
case "cache-to":
|
|
t.CacheTo = o.ArrValue
|
|
case "target":
|
|
t.Target = &value
|
|
case "secrets":
|
|
t.Secrets = o.ArrValue
|
|
case "ssh":
|
|
t.SSH = o.ArrValue
|
|
case "platform":
|
|
t.Platforms = o.ArrValue
|
|
case "output":
|
|
t.Outputs = o.ArrValue
|
|
case "annotations":
|
|
t.Annotations = append(t.Annotations, o.ArrValue...)
|
|
case "attest":
|
|
t.Attest = append(t.Attest, o.ArrValue...)
|
|
case "no-cache":
|
|
noCache, err := strconv.ParseBool(value)
|
|
if err != nil {
|
|
return errors.Errorf("invalid value %s for boolean key no-cache", value)
|
|
}
|
|
t.NoCache = &noCache
|
|
case "no-cache-filter":
|
|
t.NoCacheFilter = o.ArrValue
|
|
case "pull":
|
|
pull, err := strconv.ParseBool(value)
|
|
if err != nil {
|
|
return errors.Errorf("invalid value %s for boolean key pull", value)
|
|
}
|
|
t.Pull = &pull
|
|
case "push":
|
|
_, err := strconv.ParseBool(value)
|
|
if err != nil {
|
|
return errors.Errorf("invalid value %s for boolean key push", value)
|
|
}
|
|
if len(t.Outputs) == 0 {
|
|
t.Outputs = append(t.Outputs, "type=image,push=true")
|
|
} else {
|
|
for i, output := range t.Outputs {
|
|
if typ := parseOutputType(output); typ == "image" || typ == "registry" {
|
|
t.Outputs[i] = t.Outputs[i] + ",push=" + value
|
|
}
|
|
}
|
|
}
|
|
default:
|
|
return errors.Errorf("unknown key: %s", keys[0])
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (g *Group) GetEvalContexts(ectx *hcl.EvalContext, block *hcl.Block, loadDeps func(hcl.Expression) hcl.Diagnostics) ([]*hcl.EvalContext, error) {
|
|
content, _, err := block.Body.PartialContent(&hcl.BodySchema{
|
|
Attributes: []hcl.AttributeSchema{{Name: "matrix"}},
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if _, ok := content.Attributes["matrix"]; ok {
|
|
return nil, errors.Errorf("matrix is not supported for groups")
|
|
}
|
|
return []*hcl.EvalContext{ectx}, nil
|
|
}
|
|
|
|
func (t *Target) GetEvalContexts(ectx *hcl.EvalContext, block *hcl.Block, loadDeps func(hcl.Expression) hcl.Diagnostics) ([]*hcl.EvalContext, error) {
|
|
content, _, err := block.Body.PartialContent(&hcl.BodySchema{
|
|
Attributes: []hcl.AttributeSchema{{Name: "matrix"}},
|
|
})
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
attr, ok := content.Attributes["matrix"]
|
|
if !ok {
|
|
return []*hcl.EvalContext{ectx}, nil
|
|
}
|
|
if diags := loadDeps(attr.Expr); diags.HasErrors() {
|
|
return nil, diags
|
|
}
|
|
value, err := attr.Expr.Value(ectx)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if !value.Type().IsMapType() && !value.Type().IsObjectType() {
|
|
return nil, errors.Errorf("matrix must be a map")
|
|
}
|
|
matrix := value.AsValueMap()
|
|
|
|
ectxs := []*hcl.EvalContext{ectx}
|
|
for k, expr := range matrix {
|
|
if !expr.CanIterateElements() {
|
|
return nil, errors.Errorf("matrix values must be a list")
|
|
}
|
|
|
|
ectxs2 := []*hcl.EvalContext{}
|
|
for _, v := range expr.AsValueSlice() {
|
|
for _, e := range ectxs {
|
|
e2 := ectx.NewChild()
|
|
e2.Variables = make(map[string]cty.Value)
|
|
if e != ectx {
|
|
for k, v := range e.Variables {
|
|
e2.Variables[k] = v
|
|
}
|
|
}
|
|
e2.Variables[k] = v
|
|
ectxs2 = append(ectxs2, e2)
|
|
}
|
|
}
|
|
ectxs = ectxs2
|
|
}
|
|
return ectxs, nil
|
|
}
|
|
|
|
func (g *Group) GetName(ectx *hcl.EvalContext, block *hcl.Block, loadDeps func(hcl.Expression) hcl.Diagnostics) (string, error) {
|
|
content, _, diags := block.Body.PartialContent(&hcl.BodySchema{
|
|
Attributes: []hcl.AttributeSchema{{Name: "name"}, {Name: "matrix"}},
|
|
})
|
|
if diags != nil {
|
|
return "", diags
|
|
}
|
|
|
|
if _, ok := content.Attributes["name"]; ok {
|
|
return "", errors.Errorf("name is not supported for groups")
|
|
}
|
|
if _, ok := content.Attributes["matrix"]; ok {
|
|
return "", errors.Errorf("matrix is not supported for groups")
|
|
}
|
|
return block.Labels[0], nil
|
|
}
|
|
|
|
func (t *Target) GetName(ectx *hcl.EvalContext, block *hcl.Block, loadDeps func(hcl.Expression) hcl.Diagnostics) (string, error) {
|
|
content, _, diags := block.Body.PartialContent(&hcl.BodySchema{
|
|
Attributes: []hcl.AttributeSchema{{Name: "name"}, {Name: "matrix"}},
|
|
})
|
|
if diags != nil {
|
|
return "", diags
|
|
}
|
|
|
|
attr, ok := content.Attributes["name"]
|
|
if !ok {
|
|
return block.Labels[0], nil
|
|
}
|
|
if _, ok := content.Attributes["matrix"]; !ok {
|
|
return "", errors.Errorf("name requires matrix")
|
|
}
|
|
if diags := loadDeps(attr.Expr); diags.HasErrors() {
|
|
return "", diags
|
|
}
|
|
value, diags := attr.Expr.Value(ectx)
|
|
if diags != nil {
|
|
return "", diags
|
|
}
|
|
|
|
value, err := convert.Convert(value, cty.String)
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
return value.AsString(), nil
|
|
}
|
|
|
|
func TargetsToBuildOpt(m map[string]*Target, inp *Input) (map[string]build.Options, error) {
|
|
m2 := make(map[string]build.Options, len(m))
|
|
for k, v := range m {
|
|
bo, err := toBuildOpt(v, inp)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
m2[k] = *bo
|
|
}
|
|
return m2, nil
|
|
}
|
|
|
|
func updateContext(t *build.Inputs, inp *Input) {
|
|
if inp == nil || inp.State == nil {
|
|
return
|
|
}
|
|
|
|
for k, v := range t.NamedContexts {
|
|
if v.Path == "." {
|
|
t.NamedContexts[k] = build.NamedContext{Path: inp.URL}
|
|
}
|
|
if strings.HasPrefix(v.Path, "cwd://") || strings.HasPrefix(v.Path, "target:") || strings.HasPrefix(v.Path, "docker-image:") {
|
|
continue
|
|
}
|
|
if build.IsRemoteURL(v.Path) {
|
|
continue
|
|
}
|
|
st := llb.Scratch().File(llb.Copy(*inp.State, v.Path, "/"), llb.WithCustomNamef("set context %s to %s", k, v.Path))
|
|
t.NamedContexts[k] = build.NamedContext{State: &st}
|
|
}
|
|
|
|
if t.ContextPath == "." {
|
|
t.ContextPath = inp.URL
|
|
return
|
|
}
|
|
if strings.HasPrefix(t.ContextPath, "cwd://") {
|
|
return
|
|
}
|
|
if build.IsRemoteURL(t.ContextPath) {
|
|
return
|
|
}
|
|
st := llb.Scratch().File(
|
|
llb.Copy(*inp.State, t.ContextPath, "/", &llb.CopyInfo{
|
|
CopyDirContentsOnly: true,
|
|
}),
|
|
llb.WithCustomNamef("set context to %s", t.ContextPath),
|
|
)
|
|
t.ContextState = &st
|
|
}
|
|
|
|
// validateContextsEntitlements is a basic check to ensure contexts do not
|
|
// escape local directories when loaded from remote sources. This is to be
|
|
// replaced with proper entitlements support in the future.
|
|
func validateContextsEntitlements(t build.Inputs, inp *Input) error {
|
|
if inp == nil || inp.State == nil {
|
|
return nil
|
|
}
|
|
if v, ok := os.LookupEnv("BAKE_ALLOW_REMOTE_FS_ACCESS"); ok {
|
|
if vv, _ := strconv.ParseBool(v); vv {
|
|
return nil
|
|
}
|
|
}
|
|
if t.ContextState == nil {
|
|
if err := checkPath(t.ContextPath); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
for _, v := range t.NamedContexts {
|
|
if v.State != nil {
|
|
continue
|
|
}
|
|
if err := checkPath(v.Path); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func checkPath(p string) error {
|
|
if build.IsRemoteURL(p) || strings.HasPrefix(p, "target:") || strings.HasPrefix(p, "docker-image:") {
|
|
return nil
|
|
}
|
|
p, err := filepath.EvalSymlinks(p)
|
|
if err != nil {
|
|
if os.IsNotExist(err) {
|
|
return nil
|
|
}
|
|
return err
|
|
}
|
|
p, err = filepath.Abs(p)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
wd, err := os.Getwd()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
rel, err := filepath.Rel(wd, p)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
parts := strings.Split(rel, string(os.PathSeparator))
|
|
if parts[0] == ".." {
|
|
return errors.Errorf("path %s is outside of the working directory, please set BAKE_ALLOW_REMOTE_FS_ACCESS=1", p)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func toBuildOpt(t *Target, inp *Input) (*build.Options, error) {
|
|
if v := t.Context; v != nil && *v == "-" {
|
|
return nil, errors.Errorf("context from stdin not allowed in bake")
|
|
}
|
|
if v := t.Dockerfile; v != nil && *v == "-" {
|
|
return nil, errors.Errorf("dockerfile from stdin not allowed in bake")
|
|
}
|
|
|
|
contextPath := "."
|
|
if t.Context != nil {
|
|
contextPath = *t.Context
|
|
}
|
|
if !strings.HasPrefix(contextPath, "cwd://") && !build.IsRemoteURL(contextPath) {
|
|
contextPath = path.Clean(contextPath)
|
|
}
|
|
dockerfilePath := "Dockerfile"
|
|
if t.Dockerfile != nil {
|
|
dockerfilePath = *t.Dockerfile
|
|
}
|
|
if !strings.HasPrefix(dockerfilePath, "cwd://") {
|
|
dockerfilePath = path.Clean(dockerfilePath)
|
|
}
|
|
|
|
bi := build.Inputs{
|
|
ContextPath: contextPath,
|
|
DockerfilePath: dockerfilePath,
|
|
NamedContexts: toNamedContexts(t.Contexts),
|
|
}
|
|
if t.DockerfileInline != nil {
|
|
bi.DockerfileInline = *t.DockerfileInline
|
|
}
|
|
updateContext(&bi, inp)
|
|
if strings.HasPrefix(bi.DockerfilePath, "cwd://") {
|
|
// If Dockerfile is local for a remote invocation, we first check if
|
|
// it's not outside the working directory and then resolve it to an
|
|
// absolute path.
|
|
bi.DockerfilePath = path.Clean(strings.TrimPrefix(bi.DockerfilePath, "cwd://"))
|
|
if err := checkPath(bi.DockerfilePath); err != nil {
|
|
return nil, err
|
|
}
|
|
var err error
|
|
bi.DockerfilePath, err = filepath.Abs(bi.DockerfilePath)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
} else if !build.IsRemoteURL(bi.DockerfilePath) && strings.HasPrefix(bi.ContextPath, "cwd://") && (inp != nil && build.IsRemoteURL(inp.URL)) {
|
|
// We don't currently support reading a remote Dockerfile with a local
|
|
// context when doing a remote invocation because we automatically
|
|
// derive the dockerfile from the context atm:
|
|
//
|
|
// target "default" {
|
|
// context = BAKE_CMD_CONTEXT
|
|
// dockerfile = "Dockerfile.app"
|
|
// }
|
|
//
|
|
// > docker buildx bake https://github.com/foo/bar.git
|
|
// failed to solve: failed to read dockerfile: open /var/lib/docker/tmp/buildkit-mount3004544897/Dockerfile.app: no such file or directory
|
|
//
|
|
// To avoid mistakenly reading a local Dockerfile, we check if the
|
|
// Dockerfile exists locally and if so, we error out.
|
|
if _, err := os.Stat(filepath.Join(path.Clean(strings.TrimPrefix(bi.ContextPath, "cwd://")), bi.DockerfilePath)); err == nil {
|
|
return nil, errors.Errorf("reading a dockerfile for a remote build invocation is currently not supported")
|
|
}
|
|
}
|
|
if strings.HasPrefix(bi.ContextPath, "cwd://") {
|
|
bi.ContextPath = path.Clean(strings.TrimPrefix(bi.ContextPath, "cwd://"))
|
|
}
|
|
if !build.IsRemoteURL(bi.ContextPath) && bi.ContextState == nil && !path.IsAbs(bi.DockerfilePath) {
|
|
bi.DockerfilePath = path.Join(bi.ContextPath, bi.DockerfilePath)
|
|
}
|
|
for k, v := range bi.NamedContexts {
|
|
if strings.HasPrefix(v.Path, "cwd://") {
|
|
bi.NamedContexts[k] = build.NamedContext{Path: path.Clean(strings.TrimPrefix(v.Path, "cwd://"))}
|
|
}
|
|
}
|
|
|
|
if err := validateContextsEntitlements(bi, inp); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
t.Context = &bi.ContextPath
|
|
|
|
args := map[string]string{}
|
|
for k, v := range t.Args {
|
|
if v == nil {
|
|
continue
|
|
}
|
|
args[k] = *v
|
|
}
|
|
|
|
labels := map[string]string{}
|
|
for k, v := range t.Labels {
|
|
if v == nil {
|
|
continue
|
|
}
|
|
labels[k] = *v
|
|
}
|
|
|
|
noCache := false
|
|
if t.NoCache != nil {
|
|
noCache = *t.NoCache
|
|
}
|
|
pull := false
|
|
if t.Pull != nil {
|
|
pull = *t.Pull
|
|
}
|
|
networkMode := ""
|
|
if t.NetworkMode != nil {
|
|
networkMode = *t.NetworkMode
|
|
}
|
|
|
|
bo := &build.Options{
|
|
Inputs: bi,
|
|
Tags: t.Tags,
|
|
BuildArgs: args,
|
|
Labels: labels,
|
|
NoCache: noCache,
|
|
NoCacheFilter: t.NoCacheFilter,
|
|
Pull: pull,
|
|
NetworkMode: networkMode,
|
|
Linked: t.linked,
|
|
}
|
|
|
|
platforms, err := platformutil.Parse(t.Platforms)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
bo.Platforms = platforms
|
|
|
|
dockerConfig := config.LoadDefaultConfigFile(os.Stderr)
|
|
bo.Session = append(bo.Session, authprovider.NewDockerAuthProvider(dockerConfig, nil))
|
|
|
|
secrets, err := buildflags.ParseSecretSpecs(t.Secrets)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
secretAttachment, err := controllerapi.CreateSecrets(secrets)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
bo.Session = append(bo.Session, secretAttachment)
|
|
|
|
sshSpecs, err := buildflags.ParseSSHSpecs(t.SSH)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if len(sshSpecs) == 0 && (buildflags.IsGitSSH(bi.ContextPath) || (inp != nil && buildflags.IsGitSSH(inp.URL))) {
|
|
sshSpecs = append(sshSpecs, &controllerapi.SSH{ID: "default"})
|
|
}
|
|
sshAttachment, err := controllerapi.CreateSSH(sshSpecs)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
bo.Session = append(bo.Session, sshAttachment)
|
|
|
|
if t.Target != nil {
|
|
bo.Target = *t.Target
|
|
}
|
|
|
|
cacheImports, err := buildflags.ParseCacheEntry(t.CacheFrom)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
bo.CacheFrom = controllerapi.CreateCaches(cacheImports)
|
|
|
|
cacheExports, err := buildflags.ParseCacheEntry(t.CacheTo)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
bo.CacheTo = controllerapi.CreateCaches(cacheExports)
|
|
|
|
outputs, err := buildflags.ParseExports(t.Outputs)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
bo.Exports, err = controllerapi.CreateExports(outputs)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
annotations, err := buildflags.ParseAnnotations(t.Annotations)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
for _, e := range bo.Exports {
|
|
for k, v := range annotations {
|
|
e.Attrs[k.String()] = v
|
|
}
|
|
}
|
|
|
|
attests, err := buildflags.ParseAttests(t.Attest)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
bo.Attests = controllerapi.CreateAttestations(attests)
|
|
|
|
bo.SourcePolicy, err = build.ReadSourcePolicy()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return bo, nil
|
|
}
|
|
|
|
func defaultTarget() *Target {
|
|
return &Target{}
|
|
}
|
|
|
|
func removeDupes(s []string) []string {
|
|
i := 0
|
|
seen := make(map[string]struct{}, len(s))
|
|
for _, v := range s {
|
|
if _, ok := seen[v]; ok {
|
|
continue
|
|
}
|
|
if v == "" {
|
|
continue
|
|
}
|
|
seen[v] = struct{}{}
|
|
s[i] = v
|
|
i++
|
|
}
|
|
return s[:i]
|
|
}
|
|
|
|
func removeAttestDupes(s []string) []string {
|
|
res := []string{}
|
|
m := map[string]int{}
|
|
for _, v := range s {
|
|
att, err := buildflags.ParseAttest(v)
|
|
if err != nil {
|
|
res = append(res, v)
|
|
continue
|
|
}
|
|
|
|
if i, ok := m[att.Type]; ok {
|
|
res[i] = v
|
|
} else {
|
|
m[att.Type] = len(res)
|
|
res = append(res, v)
|
|
}
|
|
}
|
|
return res
|
|
}
|
|
|
|
func parseOutputType(str string) string {
|
|
csvReader := csv.NewReader(strings.NewReader(str))
|
|
fields, err := csvReader.Read()
|
|
if err != nil {
|
|
return ""
|
|
}
|
|
for _, field := range fields {
|
|
parts := strings.SplitN(field, "=", 2)
|
|
if len(parts) == 2 {
|
|
if parts[0] == "type" {
|
|
return parts[1]
|
|
}
|
|
}
|
|
}
|
|
return ""
|
|
}
|
|
|
|
func validateTargetName(name string) error {
|
|
if !targetNamePattern.MatchString(name) {
|
|
return errors.Errorf("only %q are allowed", validTargetNameChars)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func sanitizeTargetName(target string) string {
|
|
// as stipulated in compose spec, service name can contain a dot so as
|
|
// best-effort and to avoid any potential ambiguity, we replace the dot
|
|
// with an underscore.
|
|
return strings.ReplaceAll(target, ".", "_")
|
|
}
|
|
|
|
func sliceEqual(s1, s2 []string) bool {
|
|
if len(s1) != len(s2) {
|
|
return false
|
|
}
|
|
sort.Strings(s1)
|
|
sort.Strings(s2)
|
|
for i := range s1 {
|
|
if s1[i] != s2[i] {
|
|
return false
|
|
}
|
|
}
|
|
return true
|
|
}
|
|
|
|
func toNamedContexts(m map[string]string) map[string]build.NamedContext {
|
|
m2 := make(map[string]build.NamedContext, len(m))
|
|
for k, v := range m {
|
|
m2[k] = build.NamedContext{Path: v}
|
|
}
|
|
return m2
|
|
}
|