mirror of
https://github.com/TECHNOFAB11/jsonnet-bundler.git
synced 2025-12-12 16:10:04 +01:00
Adds a `version` field to the `jsonnetfile.json`, so that `jb` can automatically recognize too old / too new schema versions, instead of panicking.
352 lines
8.7 KiB
Go
352 lines
8.7 KiB
Go
// Copyright 2018 jsonnet-bundler authors
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package pkg
|
|
|
|
import (
|
|
"context"
|
|
"crypto/sha256"
|
|
"encoding/base64"
|
|
"fmt"
|
|
"io"
|
|
"os"
|
|
"path/filepath"
|
|
"strings"
|
|
|
|
"github.com/fatih/color"
|
|
"github.com/pkg/errors"
|
|
|
|
"github.com/jsonnet-bundler/jsonnet-bundler/pkg/jsonnetfile"
|
|
v1 "github.com/jsonnet-bundler/jsonnet-bundler/spec/v1"
|
|
"github.com/jsonnet-bundler/jsonnet-bundler/spec/v1/deps"
|
|
)
|
|
|
|
var (
|
|
VersionMismatch = errors.New("multiple colliding versions specified")
|
|
)
|
|
|
|
// Ensure receives all direct packages, the directory to vendor into and all known locks.
|
|
// It then makes sure all direct and nested dependencies are present in vendor at the correct version:
|
|
//
|
|
// If the package is locked and the files in vendor match the sha256 checksum,
|
|
// nothing needs to be done. Otherwise, the package is retrieved from the
|
|
// upstream source and added into vendor. If previously locked, the sums are
|
|
// checked as well.
|
|
// In case a (nested) package is already present in the lock,
|
|
// the one from the lock takes precedence. This allows the user to set the
|
|
// desired version in case by `jb install`ing it.
|
|
//
|
|
// Finally, all unknown files and directories are removed from vendor/
|
|
// The full list of locked depedencies is returned
|
|
func Ensure(direct v1.JsonnetFile, vendorDir string, oldLocks map[string]deps.Dependency) (map[string]deps.Dependency, error) {
|
|
// ensure all required files are in vendor
|
|
// This is the actual installation
|
|
locks, err := ensure(direct.Dependencies, vendorDir, oldLocks)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// remove unchanged legacyNames
|
|
CleanLegacyName(locks)
|
|
|
|
// find unknown dirs in vendor/
|
|
names := []string{}
|
|
err = filepath.Walk(vendorDir, func(path string, i os.FileInfo, err error) error {
|
|
if path == vendorDir {
|
|
return nil
|
|
}
|
|
if !i.IsDir() {
|
|
return nil
|
|
}
|
|
|
|
names = append(names, path)
|
|
return nil
|
|
})
|
|
|
|
// remove them
|
|
for _, dir := range names {
|
|
name, err := filepath.Rel(vendorDir, dir)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if !known(locks, name) {
|
|
if err := os.RemoveAll(dir); err != nil {
|
|
return nil, err
|
|
}
|
|
if !strings.HasPrefix(name, ".tmp") {
|
|
color.Magenta("CLEAN %s", dir)
|
|
}
|
|
}
|
|
}
|
|
|
|
// remove all symlinks, optionally adding known ones back later if wished
|
|
if err := cleanLegacySymlinks(vendorDir, locks); err != nil {
|
|
return nil, err
|
|
}
|
|
if !direct.LegacyImports {
|
|
return locks, nil
|
|
}
|
|
if err := linkLegacy(vendorDir, locks); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// return the final lockfile contents
|
|
return locks, nil
|
|
}
|
|
|
|
func CleanLegacyName(list map[string]deps.Dependency) {
|
|
for k, d := range list {
|
|
// unset if not changed by user
|
|
if d.LegacyNameCompat == d.Source.LegacyName() {
|
|
dep := list[k]
|
|
dep.LegacyNameCompat = ""
|
|
list[k] = dep
|
|
}
|
|
}
|
|
}
|
|
|
|
func cleanLegacySymlinks(vendorDir string, locks map[string]deps.Dependency) error {
|
|
// local packages need to be ignored
|
|
locals := map[string]bool{}
|
|
for _, d := range locks {
|
|
if d.Source.LocalSource == nil {
|
|
continue
|
|
}
|
|
|
|
locals[filepath.Join(vendorDir, d.Name())] = true
|
|
}
|
|
|
|
// remove all symlinks first
|
|
return filepath.Walk(vendorDir, func(path string, i os.FileInfo, err error) error {
|
|
if locals[path] {
|
|
return nil
|
|
}
|
|
|
|
if i.Mode()&os.ModeSymlink != 0 {
|
|
if err := os.Remove(path); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
})
|
|
}
|
|
|
|
func linkLegacy(vendorDir string, locks map[string]deps.Dependency) error {
|
|
// create only the ones we want
|
|
for _, d := range locks {
|
|
// localSource still uses the relative style
|
|
if d.Source.LocalSource != nil {
|
|
continue
|
|
}
|
|
|
|
legacyName := filepath.Join(vendorDir, d.LegacyName())
|
|
pkgName := d.Name()
|
|
|
|
taken, err := checkLegacyNameTaken(legacyName, pkgName)
|
|
if err != nil {
|
|
fmt.Println(err)
|
|
continue
|
|
}
|
|
if taken {
|
|
continue
|
|
}
|
|
|
|
// create the symlink
|
|
if err := os.Symlink(
|
|
filepath.Join(pkgName),
|
|
filepath.Join(legacyName),
|
|
); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func checkLegacyNameTaken(legacyName string, pkgName string) (bool, error) {
|
|
fi, err := os.Lstat(legacyName)
|
|
if err != nil {
|
|
// does not exist: not taken
|
|
if os.IsNotExist(err) {
|
|
return false, nil
|
|
}
|
|
// a real error
|
|
return false, err
|
|
}
|
|
|
|
// is it a symlink?
|
|
if fi.Mode()&os.ModeSymlink != 0 {
|
|
s, err := os.Readlink(legacyName)
|
|
if err != nil {
|
|
return false, err
|
|
}
|
|
color.Yellow("WARN: cannot link '%s' to '%s', because package '%s' already uses that name. The absolute import still works\n", pkgName, legacyName, s)
|
|
return true, nil
|
|
}
|
|
|
|
// sth else
|
|
color.Yellow("WARN: cannot link '%s' to '%s', because the file/directory already exists. The absolute import still works.\n", pkgName, legacyName)
|
|
return true, nil
|
|
}
|
|
|
|
func known(deps map[string]deps.Dependency, p string) bool {
|
|
for _, d := range deps {
|
|
k := d.Name()
|
|
if strings.HasPrefix(p, k) || strings.HasPrefix(k, p) {
|
|
return true
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
func ensure(direct map[string]deps.Dependency, vendorDir string, locks map[string]deps.Dependency) (map[string]deps.Dependency, error) {
|
|
deps := make(map[string]deps.Dependency)
|
|
|
|
for _, d := range direct {
|
|
l, present := locks[d.Name()]
|
|
|
|
// already locked and the integrity is intact
|
|
if present {
|
|
d.Version = locks[d.Name()].Version
|
|
|
|
if check(l, vendorDir) {
|
|
deps[d.Name()] = l
|
|
continue
|
|
}
|
|
}
|
|
expectedSum := locks[d.Name()].Sum
|
|
|
|
// either not present or not intact: download again
|
|
dir := filepath.Join(vendorDir, d.Name())
|
|
os.RemoveAll(dir)
|
|
|
|
locked, err := download(d, vendorDir)
|
|
if err != nil {
|
|
return nil, errors.Wrap(err, "downloading")
|
|
}
|
|
if expectedSum != "" && locked.Sum != expectedSum {
|
|
return nil, fmt.Errorf("checksum mismatch for %s. Expected %s but got %s", d.Name(), expectedSum, locked.Sum)
|
|
}
|
|
deps[d.Name()] = *locked
|
|
// we settled on a new version, add it to the locks for recursion
|
|
locks[d.Name()] = *locked
|
|
}
|
|
|
|
for _, d := range deps {
|
|
f, err := jsonnetfile.Load(filepath.Join(vendorDir, d.Name(), jsonnetfile.File))
|
|
if err != nil {
|
|
if os.IsNotExist(err) {
|
|
continue
|
|
}
|
|
return nil, err
|
|
}
|
|
|
|
nested, err := ensure(f.Dependencies, vendorDir, locks)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
for _, d := range nested {
|
|
if _, ok := deps[d.Name()]; !ok {
|
|
deps[d.Name()] = d
|
|
}
|
|
}
|
|
}
|
|
|
|
return deps, nil
|
|
}
|
|
|
|
// download retrieves a package from a remote upstream. The checksum of the
|
|
// files is generated afterwards.
|
|
func download(d deps.Dependency, vendorDir string) (*deps.Dependency, error) {
|
|
var p Interface
|
|
switch {
|
|
case d.Source.GitSource != nil:
|
|
p = NewGitPackage(d.Source.GitSource)
|
|
case d.Source.LocalSource != nil:
|
|
p = NewLocalPackage(d.Source.LocalSource)
|
|
}
|
|
|
|
if p == nil {
|
|
return nil, errors.New("either git or local source is required")
|
|
}
|
|
|
|
version, err := p.Install(context.TODO(), d.Name(), vendorDir, d.Version)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
var sum string
|
|
if d.Source.LocalSource == nil {
|
|
sum = hashDir(filepath.Join(vendorDir, d.Name()))
|
|
}
|
|
|
|
d.Version = version
|
|
d.Sum = sum
|
|
return &d, nil
|
|
}
|
|
|
|
// check returns whether the files present at the vendor/ folder match the
|
|
// sha256 sum of the package. local-directory dependencies are not checked as
|
|
// their purpose is to change during development where integrity checking would
|
|
// be a hindrance.
|
|
func check(d deps.Dependency, vendorDir string) bool {
|
|
// assume a local dependency is intact as long as it exists
|
|
if d.Source.LocalSource != nil {
|
|
x, err := jsonnetfile.Exists(filepath.Join(vendorDir, d.Name()))
|
|
if err != nil {
|
|
return false
|
|
}
|
|
return x
|
|
}
|
|
|
|
if d.Sum == "" {
|
|
// no sum available, need to download
|
|
return false
|
|
}
|
|
|
|
dir := filepath.Join(vendorDir, d.Name())
|
|
sum := hashDir(dir)
|
|
return d.Sum == sum
|
|
}
|
|
|
|
// hashDir computes the checksum of a directory by concatenating all files and
|
|
// hashing this data using sha256. This can be memory heavy with lots of data,
|
|
// but jsonnet files should be fairly small
|
|
func hashDir(dir string) string {
|
|
hasher := sha256.New()
|
|
|
|
filepath.Walk(dir, func(path string, info os.FileInfo, err error) error {
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
if info.IsDir() {
|
|
return nil
|
|
}
|
|
|
|
f, err := os.Open(path)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer f.Close()
|
|
|
|
if _, err := io.Copy(hasher, f); err != nil {
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
})
|
|
|
|
return base64.StdEncoding.EncodeToString(hasher.Sum(nil))
|
|
}
|