This mega patch primarily introduces a new function engine. The main reasons for this new engine are: 1) Massively improved performance with lock-contended graphs. Certain large function graphs could have very high lock-contention which turned out to be much slower than I would have liked. This new algorithm happens to be basically lock-free, so that's another helpful improvement. 2) Glitch-free function graphs. The function graphs could "glitch" (an FRP term) which could be undesirable in theory. In practice this was never really an issue, and I've not explicitly guaranteed that the new graphs are provably glitch-free, but in practice things are a lot more consistent. 3) Simpler graph shape. The new graphs don't require the private channels. This makes understanding the graphs a lot easier. 4) Branched graphs only run half. Previously we would run two pure side of an if statement, and while this was mostly meant as an early experiment, it stayed in for far too long and now was the right time to remove this. This also means our graphs are much smaller and more efficient too. Note that this changed the function API slightly. Everything has been ported. It's possible that we introduce a new API in the future, but it is unexpected to cause removal of the two current APIs. In addition, we finally split out the "schedule" aspect from world.schedule(). The "pick me" aspects now happen in a separate resource, rather than as a yucky side-effect in the function. This also lets us more precisely choose when we're scheduled, and we can observe without being chosen too. As usual many thanks to Sam for helping through some of the algorithmic graph shape issues!
639 lines
20 KiB
Go
639 lines
20 KiB
Go
// Mgmt
|
|
// Copyright (C) James Shubin and the project contributors
|
|
// Written by James Shubin <james@shubin.ca> and the project contributors
|
|
//
|
|
// This program is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with this program. If not, see <https://www.gnu.org/licenses/>.
|
|
//
|
|
// Additional permission under GNU GPL version 3 section 7
|
|
//
|
|
// If you modify this program, or any covered work, by linking or combining it
|
|
// with embedded mcl code and modules (and that the embedded mcl code and
|
|
// modules which link with this program, contain a copy of their source code in
|
|
// the authoritative form) containing parts covered by the terms of any other
|
|
// license, the licensors of this program grant you additional permission to
|
|
// convey the resulting work. Furthermore, the licensors of this program grant
|
|
// the original author, James Shubin, additional permission to update this
|
|
// additional permission if he deems it necessary to achieve the goals of this
|
|
// additional permission.
|
|
|
|
// Package gapi is the Graph API implementation for the mcl language frontend.
|
|
package gapi
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"fmt"
|
|
"os"
|
|
"path/filepath"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
cliUtil "github.com/purpleidea/mgmt/cli/util"
|
|
"github.com/purpleidea/mgmt/engine"
|
|
"github.com/purpleidea/mgmt/gapi"
|
|
"github.com/purpleidea/mgmt/lang"
|
|
"github.com/purpleidea/mgmt/lang/ast"
|
|
"github.com/purpleidea/mgmt/lang/download"
|
|
"github.com/purpleidea/mgmt/lang/funcs/vars"
|
|
"github.com/purpleidea/mgmt/lang/inputs"
|
|
"github.com/purpleidea/mgmt/lang/interfaces"
|
|
"github.com/purpleidea/mgmt/lang/interpolate"
|
|
"github.com/purpleidea/mgmt/lang/parser"
|
|
"github.com/purpleidea/mgmt/lang/types"
|
|
"github.com/purpleidea/mgmt/lang/unification"
|
|
"github.com/purpleidea/mgmt/pgraph"
|
|
"github.com/purpleidea/mgmt/util"
|
|
"github.com/purpleidea/mgmt/util/errwrap"
|
|
|
|
"github.com/spf13/afero"
|
|
)
|
|
|
|
const (
|
|
// Name is the name of this frontend.
|
|
Name = "lang"
|
|
)
|
|
|
|
func init() {
|
|
gapi.Register(Name, func() gapi.GAPI { return &GAPI{} }) // register
|
|
}
|
|
|
|
// GAPI implements the main lang GAPI interface.
|
|
type GAPI struct {
|
|
InputURI string // input URI of code file system to run
|
|
|
|
// Data is some additional data for the lang struct.
|
|
Data *lang.Data
|
|
|
|
lang *lang.Lang // lang struct
|
|
ctx context.Context
|
|
cancel func()
|
|
reterr error
|
|
|
|
// this data struct is only available *after* Init, so as a result, it
|
|
// can not be used inside the Cli(...) method.
|
|
data *gapi.Data
|
|
initialized bool
|
|
wg *sync.WaitGroup // sync group for tunnel go routines
|
|
err error
|
|
errMutex *sync.Mutex // guards err
|
|
}
|
|
|
|
// Cli takes an *Info struct, and returns our deploy if activated, and if there
|
|
// are any validation problems, you should return an error. If there is no
|
|
// deploy, then you should return a nil deploy and a nil error. This is passed
|
|
// in a functional file system interface. For standalone usage, this will be a
|
|
// temporary memory-backed filesystem so that the same deploy API is used, and
|
|
// for normal clustered usage, this will be the normal implementation which is
|
|
// usually an etcd backed fs. At this point we should be copying the necessary
|
|
// local file system data into our fs for future use when the GAPI is running.
|
|
// IOW, running this Cli function, when activated, produces a deploy object
|
|
// which is run by our main loop. The difference between running from `deploy`
|
|
// or from `run` (both of which can activate this GAPI) is that `deploy` copies
|
|
// to an etcdFs, and `run` copies to a memFs. All GAPI's run off of the fs that
|
|
// is passed in.
|
|
func (obj *GAPI) Cli(info *gapi.Info) (*gapi.Deploy, error) {
|
|
args, ok := info.Args.(*cliUtil.LangArgs)
|
|
if !ok {
|
|
// programming error
|
|
return nil, fmt.Errorf("could not convert to our struct")
|
|
}
|
|
|
|
fs := info.Fs // copy files from local filesystem *into* this fs...
|
|
prefix := "" // TODO: do we need this?
|
|
debug := info.Debug
|
|
logf := func(format string, v ...interface{}) {
|
|
info.Logf(Name+": "+format, v...)
|
|
}
|
|
|
|
// empty by default (don't set for deploy, only download)
|
|
modules := args.ModulePath
|
|
if modules != "" && !strings.HasSuffix(modules, "/") {
|
|
return nil, fmt.Errorf("module path does not end with a slash")
|
|
}
|
|
if modules != "" && !strings.HasPrefix(modules, "/") {
|
|
wd, err := os.Getwd()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
modules = filepath.Join(wd, modules) + "/"
|
|
}
|
|
|
|
// TODO: while reading through trees of metadata files, we could also
|
|
// check the license compatibility of deps...
|
|
|
|
osFs := afero.NewOsFs()
|
|
readOnlyOsFs := afero.NewReadOnlyFs(osFs) // can't be readonly to dl!
|
|
//bp := afero.NewBasePathFs(osFs, base) // TODO: can this prevent parent dir access?
|
|
afs := &afero.Afero{Fs: readOnlyOsFs} // wrap so that we're implementing ioutil
|
|
localFs := &util.AferoFs{Afero: afs} // always the local fs
|
|
downloadAfs := &afero.Afero{Fs: osFs}
|
|
downloadFs := &util.AferoFs{Afero: downloadAfs} // TODO: use with a parent path preventer?
|
|
|
|
// the fs input here is the local fs we're reading to get the files from
|
|
// this is different from the fs variable which is our output dest!!!
|
|
output, err := inputs.ParseInput(args.Input, localFs)
|
|
if err != nil {
|
|
return nil, errwrap.Wrapf(err, "could not activate an input parser")
|
|
}
|
|
|
|
// no need to run recursion detection since this is the beginning
|
|
// TODO: do the paths need to be cleaned for "../" before comparison?
|
|
|
|
logf("lexing/parsing...")
|
|
xast, err := parser.LexParse(bytes.NewReader(output.Main))
|
|
if err != nil {
|
|
return nil, errwrap.Wrapf(err, "could not generate AST")
|
|
}
|
|
if debug {
|
|
logf("behold, the AST: %+v", xast)
|
|
}
|
|
|
|
// This runs the necessary downloads. It passes a downloader in, which
|
|
// can be used to pull down or update any missing imports.
|
|
var downloader interfaces.Downloader
|
|
if args.Download {
|
|
downloadInfo := &interfaces.DownloadInfo{
|
|
Fs: downloadFs, // the local fs!
|
|
|
|
// flags are passed in during Init()
|
|
Noop: info.Flags.Noop,
|
|
Sema: info.Flags.Sema,
|
|
Update: args.Update,
|
|
|
|
Debug: debug,
|
|
Logf: func(format string, v ...interface{}) {
|
|
// TODO: is this a sane prefix to use here?
|
|
logf("get: "+format, v...)
|
|
},
|
|
}
|
|
// this fulfills the interfaces.Downloader interface
|
|
downloader = &download.Downloader{
|
|
Depth: args.Depth, // default of infinite is -1
|
|
Retry: args.Retry, // infinite is -1
|
|
}
|
|
if err := downloader.Init(downloadInfo); err != nil {
|
|
return nil, errwrap.Wrapf(err, "could not initialize downloader")
|
|
}
|
|
}
|
|
|
|
importGraph, err := pgraph.NewGraph("importGraph")
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
importVertex := &pgraph.SelfVertex{
|
|
Name: "", // first node is the empty string
|
|
Graph: importGraph, // store a reference to ourself
|
|
}
|
|
importGraph.AddVertex(importVertex)
|
|
|
|
//logf("init...")
|
|
logf("import: %s", output.Base)
|
|
// init and validate the structure of the AST
|
|
data := &interfaces.Data{
|
|
// TODO: add missing fields here if/when needed
|
|
Fs: output.FS, // formerly: localFs // the local fs!
|
|
FsURI: output.FS.URI(), // formerly: localFs.URI() // TODO: is this right?
|
|
Base: output.Base, // base dir (absolute path) that this is rooted in
|
|
Files: output.Files,
|
|
Imports: importVertex,
|
|
Metadata: output.Metadata,
|
|
Modules: modules,
|
|
|
|
LexParser: parser.LexParse,
|
|
Downloader: downloader,
|
|
StrInterpolater: interpolate.StrInterpolate,
|
|
SourceFinder: os.ReadFile,
|
|
//Local: obj.Local, // TODO: do we need this?
|
|
//World: obj.World, // TODO: do we need this?
|
|
|
|
Prefix: prefix,
|
|
Debug: debug,
|
|
Logf: func(format string, v ...interface{}) {
|
|
// TODO: is this a sane prefix to use here?
|
|
logf("ast: "+format, v...)
|
|
},
|
|
}
|
|
// some of this might happen *after* interpolate in SetScope or Unify...
|
|
if err := xast.Init(data); err != nil {
|
|
return nil, errwrap.Wrapf(err, "could not init and validate AST")
|
|
}
|
|
|
|
if debug {
|
|
logf("interpolating...")
|
|
}
|
|
// interpolate strings and other expansionable nodes in AST
|
|
iast, err := xast.Interpolate()
|
|
if err != nil {
|
|
return nil, errwrap.Wrapf(err, "could not interpolate AST")
|
|
}
|
|
|
|
hostname := ""
|
|
if h := info.Flags.Hostname; h != nil {
|
|
hostname = *h // it's optional, since this value is not used...
|
|
}
|
|
variables := map[string]interfaces.Expr{
|
|
"purpleidea": &ast.ExprStr{V: "hello world!"}, // james says hi
|
|
// TODO: change to a func when we can change hostname dynamically!
|
|
"hostname": &ast.ExprStr{V: hostname}, // NOTE: can be empty b/c not used
|
|
}
|
|
consts := ast.VarPrefixToVariablesScope(vars.ConstNamespace) // strips prefix!
|
|
addback := vars.ConstNamespace + interfaces.ModuleSep // add it back...
|
|
variables, err = ast.MergeExprMaps(variables, consts, addback)
|
|
if err != nil {
|
|
return nil, errwrap.Wrapf(err, "couldn't merge in consts")
|
|
}
|
|
|
|
// top-level, built-in, initial global scope
|
|
scope := &interfaces.Scope{
|
|
Variables: variables,
|
|
// all the built-in top-level, core functions enter here...
|
|
Functions: ast.FuncPrefixToFunctionsScope(""), // runs funcs.LookupPrefix
|
|
}
|
|
|
|
if debug {
|
|
logf("scope building...")
|
|
}
|
|
// propagate the scope down through the AST...
|
|
// We use SetScope because it follows all of the imports through. I did
|
|
// not think we needed to pass in an initial scope because the download
|
|
// operation should not depend on any initial scope values, since those
|
|
// would all be runtime changes, and we do not support dynamic imports,
|
|
// however, we need to since we're doing type unification to err early!
|
|
if err := iast.SetScope(scope); err != nil { // empty initial scope!
|
|
return nil, errwrap.Wrapf(err, "could not set scope")
|
|
}
|
|
|
|
// Previously the `get` command would stop here.
|
|
if args.OnlyDownload {
|
|
return nil, nil // success!
|
|
}
|
|
|
|
unificationStrategy := make(map[string]string)
|
|
if name := args.UnifySolver; name != nil && *name != "" {
|
|
unificationStrategy[unification.StrategyNameKey] = *name
|
|
}
|
|
if len(args.UnifyOptimizations) > 0 {
|
|
// TODO: use a query string parser instead?
|
|
unificationStrategy[unification.StrategyOptimizationsKey] = strings.Join(args.UnifyOptimizations, ",")
|
|
}
|
|
|
|
if !args.SkipUnify {
|
|
// apply type unification
|
|
unificationLogf := func(format string, v ...interface{}) {
|
|
logf("unification: "+format, v...)
|
|
}
|
|
logf("running type unification...")
|
|
|
|
solver, err := unification.LookupDefault()
|
|
if err != nil {
|
|
return nil, errwrap.Wrapf(err, "could not get default solver")
|
|
}
|
|
unifier := &unification.Unifier{
|
|
AST: iast,
|
|
Solver: solver,
|
|
Strategy: unificationStrategy,
|
|
UnifiedState: types.NewUnifiedState(),
|
|
Debug: debug,
|
|
Logf: unificationLogf,
|
|
}
|
|
startTime := time.Now()
|
|
unifyErr := unifier.Unify(context.TODO())
|
|
delta := time.Since(startTime)
|
|
formatted := delta.String()
|
|
if delta.Milliseconds() > 1000 { // 1 second
|
|
formatted = delta.Truncate(time.Millisecond).String()
|
|
}
|
|
if unifyErr != nil {
|
|
if args.OnlyUnify {
|
|
logf("type unification failed after %s", formatted)
|
|
}
|
|
return nil, errwrap.Wrapf(unifyErr, "could not unify types")
|
|
}
|
|
|
|
if args.OnlyUnify {
|
|
logf("type unification succeeded in %s", formatted)
|
|
return nil, nil // we end early
|
|
}
|
|
}
|
|
|
|
// get the list of needed files (this is available after SetScope)
|
|
logf("collecting files...")
|
|
fileList, err := ast.CollectFiles(iast)
|
|
if err != nil {
|
|
return nil, errwrap.Wrapf(err, "could not collect files")
|
|
}
|
|
|
|
// add in our initial files
|
|
|
|
// we can sometimes be missing our top-level metadata.yaml and main.mcl
|
|
files := []string{}
|
|
files = append(files, output.Files...)
|
|
files = append(files, fileList...)
|
|
|
|
writeableFS, ok := fs.(engine.WriteableFS)
|
|
if !ok {
|
|
return nil, fmt.Errorf("the FS was not writeable")
|
|
}
|
|
|
|
// run some copy operations to add data into the filesystem
|
|
for _, fn := range output.Workers {
|
|
if err := fn(writeableFS); err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
// There are duplicates if in our dag we use the same import twice.
|
|
files = util.StrRemoveDuplicatesInList(files)
|
|
|
|
// Add any missing dirs, so that we don't need to use `MkdirAll`...
|
|
// FIXME: It's possible that the dirs get generated upstream, but it's
|
|
// not exactly clear where they'd need to get added into the list. If we
|
|
// figure that out, we can remove this additional step. It's trickier,
|
|
// because adding duplicates isn't desirable either.
|
|
//dirs, err := util.MissingMkdirs(files)
|
|
//if err != nil {
|
|
// // possible programming error
|
|
// return nil, errwrap.Wrapf(err, "unexpected missing mkdirs input")
|
|
//}
|
|
//parents := util.DirParents(output.Base)
|
|
//parents = append(parents, output.Base) // include self
|
|
//
|
|
// And we don't want to include any of the parents above the Base dir...
|
|
//for _, x := range dirs {
|
|
// if util.StrInList(x, parents) {
|
|
// continue
|
|
// }
|
|
// files = append(files, x)
|
|
//}
|
|
|
|
// sort by depth dependency order! (or mkdir -p all the dirs first)
|
|
// TODO: is this natively already in a correctly sorted order?
|
|
util.PathSlice(files).Sort() // sort it
|
|
for _, src := range files { // absolute paths
|
|
// rebase path src to root file system of "/" for etcdfs...
|
|
|
|
// everywhere we expect absolute, but we should use relative :/
|
|
//tree, err := util.FsTree(fs, "/")
|
|
//if err != nil {
|
|
// return nil, err
|
|
//}
|
|
//logf("tree:\n%s", tree)
|
|
|
|
// XXX: Should we have output.Base and output.ModulesBase?
|
|
// XXX: Maybe we should rebase to whichever the src starts with?
|
|
//commonBase := util.CommonPathPrefix(src, output.Base)
|
|
//logf("src:\n%s", src)
|
|
//logf("base:\n%s", output.Base)
|
|
//logf("common:\n%s", commonBase)
|
|
commonBase := output.Base // old method!
|
|
// NOTE: Instead of commonBase, we used to use output.Base here,
|
|
// but it seems this breaks if the modules path is not inside
|
|
// the normal code base. Such as if the src is:
|
|
// /etc/mgmt/modules/github.com/purpleidea/mgmt/modules/misc/main.mcl
|
|
// and the base is: /etc/mgmt/main/ if we run the mgmt binary
|
|
// with: mgmt run lang --module-path '/etc/mgmt/modules/' /etc/mgmt/main/
|
|
// for example.
|
|
// NOTE: We could possibly always rebase onto "/", but we'd like
|
|
// to eliminate the local path structure from our deploys for a
|
|
// weak kind of privacy of that users directory structure.
|
|
dst, err := util.Rebase(src, commonBase, "/")
|
|
if err != nil {
|
|
if modules == "" || !strings.HasPrefix(src, modules) {
|
|
// possible programming error
|
|
logf("src:\n%s", src)
|
|
logf("base:\n%s", output.Base)
|
|
//logf("common:\n%s", commonBase)
|
|
return nil, errwrap.Wrapf(err, "malformed source file path: `%s`", src)
|
|
}
|
|
// HACK: maybe it's a module?
|
|
// If we have a different base path, it might be a
|
|
// module dir. Maybe this hack covers all scenarios.
|
|
|
|
// Remove the actual "modules/" dir from the end...
|
|
m, err := util.RemovePathSuffix(modules)
|
|
if err != nil {
|
|
return nil, errwrap.Wrapf(err, "malformed source module dir: `%s`", modules)
|
|
}
|
|
|
|
// ...so that we keep `/modules/` as the new module dir.
|
|
dst, err = util.Rebase(src, m, "/")
|
|
if err != nil {
|
|
// possible programming error
|
|
logf("src:\n%s", src)
|
|
logf("base:\n%s", m)
|
|
return nil, errwrap.Wrapf(err, "malformed source module path: `%s`", src)
|
|
}
|
|
} else if modules != "" && strings.HasPrefix(src, modules) {
|
|
// Here we've succeeded in the above rebase, but before
|
|
// we accept it completely, let's first check if it's
|
|
// actually a MODULE_PATH situation. If so, let's rebase
|
|
// it only the `/modules/` directory which is used in
|
|
// the deploy.
|
|
dst, err = util.Rebase(src, modules, "/"+interfaces.ModuleDirectory)
|
|
if err != nil {
|
|
return nil, errwrap.Wrapf(err, "malformed src path: `%s`", src)
|
|
}
|
|
}
|
|
|
|
if strings.HasSuffix(src, "/") { // it's a dir
|
|
// FIXME: I think fixing CopyDirToFs might be better...
|
|
if dst != "/" { // XXX: hack, don't nest the copy badly!
|
|
out, err := util.RemovePathSuffix(dst)
|
|
if err != nil {
|
|
// possible programming error
|
|
return nil, errwrap.Wrapf(err, "malformed dst dir path: `%s`", dst)
|
|
}
|
|
dst = out
|
|
}
|
|
// TODO: add more tests to this (it is actually CopyFs)
|
|
// TODO: Used to be: CopyDirToFs, but it had issues...
|
|
if err := gapi.CopyDirToFsForceAll(fs, src, dst); err != nil {
|
|
return nil, errwrap.Wrapf(err, "can't copy dir from `%s` to `%s`", src, dst)
|
|
}
|
|
continue
|
|
}
|
|
// it's a regular file path
|
|
|
|
// Occasionally, we need the dir to exist first or we'll error.
|
|
if err := gapi.MkdirAllOnFs(writeableFS, util.Dirname(dst), 0700); err != nil {
|
|
return nil, errwrap.Wrapf(err, "can't mkdir at `%s`", dst)
|
|
}
|
|
if err := gapi.CopyFileToFs(writeableFS, src, dst); err != nil {
|
|
return nil, errwrap.Wrapf(err, "can't copy file from `%s` to `%s`", src, dst)
|
|
}
|
|
}
|
|
|
|
// display the deploy fs tree
|
|
if debug { // this should only be shown on debug, or `entry` looks messy!
|
|
logf("input: %s", args.Input)
|
|
tree, err := util.FsTree(fs, "/")
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
logf("tree:\n%s", tree)
|
|
}
|
|
|
|
return &gapi.Deploy{
|
|
Name: Name,
|
|
Noop: info.Flags.Noop,
|
|
Sema: info.Flags.Sema,
|
|
|
|
NoAutoEdges: info.Flags.NoAutoEdges,
|
|
|
|
GAPI: &GAPI{
|
|
InputURI: fs.URI(),
|
|
Data: &lang.Data{
|
|
UnificationStrategy: unificationStrategy,
|
|
// TODO: add properties here...
|
|
},
|
|
},
|
|
}, nil
|
|
}
|
|
|
|
// Init initializes the lang GAPI struct.
|
|
func (obj *GAPI) Init(data *gapi.Data) error {
|
|
if obj.initialized {
|
|
return fmt.Errorf("already initialized")
|
|
}
|
|
if obj.InputURI == "" {
|
|
return fmt.Errorf("the InputURI param must be specified")
|
|
}
|
|
obj.data = data // store for later
|
|
obj.wg = &sync.WaitGroup{}
|
|
obj.errMutex = &sync.Mutex{}
|
|
obj.initialized = true
|
|
|
|
if obj.InputURI == "-" {
|
|
return fmt.Errorf("stdin passthrough is not supported at this time")
|
|
}
|
|
|
|
fs, err := obj.data.World.Fs(obj.InputURI) // open the remote file system
|
|
if err != nil {
|
|
return errwrap.Wrapf(err, "can't load code from file system `%s`", obj.InputURI)
|
|
}
|
|
// the lang always tries to load from this standard path: /metadata.yaml
|
|
input := "/" + interfaces.MetadataFilename // path in remote fs
|
|
|
|
lang := &lang.Lang{
|
|
Fs: fs,
|
|
FsURI: obj.InputURI,
|
|
Input: input,
|
|
Data: obj.Data,
|
|
|
|
Hostname: obj.data.Hostname,
|
|
Local: obj.data.Local,
|
|
World: obj.data.World,
|
|
Debug: obj.data.Debug,
|
|
Logf: func(format string, v ...interface{}) {
|
|
// TODO: add the Name prefix in parent logger
|
|
obj.data.Logf(Name+": "+format, v...)
|
|
},
|
|
}
|
|
if err := lang.Init(context.TODO()); err != nil { // XXX: CTX?
|
|
return errwrap.Wrapf(err, "can't init the lang")
|
|
}
|
|
obj.lang = lang // once we can't fail, store the struct...
|
|
|
|
return nil
|
|
}
|
|
|
|
// Info returns some data about the GAPI implementation.
|
|
func (obj *GAPI) Info() *gapi.InfoResult {
|
|
return &gapi.InfoResult{
|
|
URI: obj.InputURI,
|
|
}
|
|
}
|
|
|
|
// Next returns nil errors every time there could be a new graph.
|
|
func (obj *GAPI) Next(ctx context.Context) chan gapi.Next {
|
|
ch := make(chan gapi.Next)
|
|
|
|
obj.wg.Add(1)
|
|
go func() {
|
|
defer obj.lang.Cleanup() // after everyone closes
|
|
defer obj.wg.Wait() // wait before cleanup
|
|
defer obj.wg.Done()
|
|
err := obj.lang.Run(ctx)
|
|
obj.errAppend(err)
|
|
// When run terminates, the "official" error takes precedence.
|
|
if err := obj.lang.Err(); err != nil {
|
|
obj.errAppend(err)
|
|
}
|
|
}()
|
|
|
|
obj.wg.Add(1)
|
|
go func() {
|
|
defer obj.wg.Done()
|
|
defer close(ch) // this will run before the obj.wg.Done()
|
|
if !obj.initialized {
|
|
next := gapi.Next{
|
|
Err: fmt.Errorf("%s: GAPI is not initialized", Name),
|
|
Exit: true, // exit, b/c programming error?
|
|
}
|
|
select {
|
|
case ch <- next:
|
|
case <-ctx.Done():
|
|
obj.errAppend(ctx.Err())
|
|
}
|
|
return
|
|
}
|
|
|
|
streamChan := obj.lang.Stream(ctx)
|
|
|
|
var ok bool
|
|
for {
|
|
var graph *pgraph.Graph
|
|
select {
|
|
case graph, ok = <-streamChan: // a variable changed
|
|
if !ok { // the channel closed!
|
|
return
|
|
}
|
|
|
|
case <-ctx.Done():
|
|
obj.errAppend(ctx.Err())
|
|
return
|
|
}
|
|
obj.data.Logf("generating new graph...")
|
|
|
|
next := gapi.Next{
|
|
Graph: graph,
|
|
}
|
|
select {
|
|
case ch <- next: // trigger a run (send a msg)
|
|
|
|
// unblock if we exit while waiting to send!
|
|
case <-ctx.Done():
|
|
obj.errAppend(ctx.Err())
|
|
return
|
|
}
|
|
}
|
|
}()
|
|
return ch
|
|
}
|
|
|
|
// Err will contain the last error when Next shuts down. It waits for all the
|
|
// running processes to exit before it returns.
|
|
func (obj *GAPI) Err() error {
|
|
obj.wg.Wait()
|
|
return obj.err
|
|
}
|
|
|
|
// errAppend is a simple helper function.
|
|
func (obj *GAPI) errAppend(err error) {
|
|
obj.errMutex.Lock()
|
|
obj.err = errwrap.Append(obj.err, err)
|
|
obj.errMutex.Unlock()
|
|
}
|