add template function allowing user to read keys from vault
update dependencies add minimal docs
This commit is contained in:
parent
3ea8ee6cc6
commit
dfaf624f4c
|
@ -13,6 +13,7 @@ import (
|
||||||
consulapi "github.com/hashicorp/consul/api"
|
consulapi "github.com/hashicorp/consul/api"
|
||||||
"github.com/hashicorp/packer/common/uuid"
|
"github.com/hashicorp/packer/common/uuid"
|
||||||
"github.com/hashicorp/packer/version"
|
"github.com/hashicorp/packer/version"
|
||||||
|
vaultapi "github.com/hashicorp/vault/api"
|
||||||
)
|
)
|
||||||
|
|
||||||
// InitTime is the UTC time when this package was initialized. It is
|
// InitTime is the UTC time when this package was initialized. It is
|
||||||
|
@ -38,6 +39,7 @@ var FuncGens = map[string]FuncGenerator{
|
||||||
"user": funcGenUser,
|
"user": funcGenUser,
|
||||||
"packer_version": funcGenPackerVersion,
|
"packer_version": funcGenPackerVersion,
|
||||||
"consul_key": funcGenConsul,
|
"consul_key": funcGenConsul,
|
||||||
|
"vault": funcGenVault,
|
||||||
|
|
||||||
"upper": funcGenPrimitive(strings.ToUpper),
|
"upper": funcGenPrimitive(strings.ToUpper),
|
||||||
"lower": funcGenPrimitive(strings.ToLower),
|
"lower": funcGenPrimitive(strings.ToLower),
|
||||||
|
@ -202,7 +204,43 @@ func funcGenConsul(ctx *Context) interface{} {
|
||||||
if value == "" {
|
if value == "" {
|
||||||
return "", fmt.Errorf("value is empty at path %s", k)
|
return "", fmt.Errorf("value is empty at path %s", k)
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func funcGenVault(ctx *Context) interface{} {
|
||||||
|
return func(path string, key string) (string, error) {
|
||||||
|
// Only allow interpolation from Vault when env vars are being read.
|
||||||
|
if !ctx.EnableEnv {
|
||||||
|
// The error message doesn't have to be that detailed since
|
||||||
|
// semantic checks should catch this.
|
||||||
|
return "", errors.New("Vault vars are only allowed in the variables section")
|
||||||
|
}
|
||||||
|
if token := os.Getenv("VAULT_TOKEN"); token == "" {
|
||||||
|
return "", errors.New("Must set VAULT_TOKEN env var in order to " +
|
||||||
|
"use vault template function")
|
||||||
|
}
|
||||||
|
// const EnvVaultAddress = "VAULT_ADDR"
|
||||||
|
// const EnvVaultToken = "VAULT_TOKEN"
|
||||||
|
vaultConfig := vaultapi.DefaultConfig()
|
||||||
|
cli, err := vaultapi.NewClient(vaultConfig)
|
||||||
|
if err != nil {
|
||||||
|
return "", errors.New(fmt.Sprintf("Error getting Vault client: %s", err))
|
||||||
|
}
|
||||||
|
secret, err := cli.Logical().Read(path)
|
||||||
|
if err != nil {
|
||||||
|
return "", errors.New(fmt.Sprintf("Error reading vault secret: %s", err))
|
||||||
|
}
|
||||||
|
if secret == nil {
|
||||||
|
return "", errors.New(fmt.Sprintf("Vault Secret does not exist at the given path."))
|
||||||
|
}
|
||||||
|
|
||||||
|
data := secret.Data["data"]
|
||||||
|
if data == nil {
|
||||||
|
return "", errors.New(fmt.Sprintf("Vault data was empty at the "+
|
||||||
|
"given path. Warnings: %s", strings.Join(secret.Warnings, "; ")))
|
||||||
|
}
|
||||||
|
|
||||||
|
value := secret.Data["data"].(map[string]interface{})[key].(string)
|
||||||
return value, nil
|
return value, nil
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -0,0 +1,15 @@
|
||||||
|
# This is the official list of Snappy-Go authors for copyright purposes.
|
||||||
|
# This file is distinct from the CONTRIBUTORS files.
|
||||||
|
# See the latter for an explanation.
|
||||||
|
|
||||||
|
# Names should be added to this file as
|
||||||
|
# Name or Organization <email address>
|
||||||
|
# The email address is not required for organizations.
|
||||||
|
|
||||||
|
# Please keep the list sorted.
|
||||||
|
|
||||||
|
Damian Gryski <dgryski@gmail.com>
|
||||||
|
Google Inc.
|
||||||
|
Jan Mercl <0xjnml@gmail.com>
|
||||||
|
Rodolfo Carvalho <rhcarvalho@gmail.com>
|
||||||
|
Sebastien Binet <seb.binet@gmail.com>
|
|
@ -0,0 +1,37 @@
|
||||||
|
# This is the official list of people who can contribute
|
||||||
|
# (and typically have contributed) code to the Snappy-Go repository.
|
||||||
|
# The AUTHORS file lists the copyright holders; this file
|
||||||
|
# lists people. For example, Google employees are listed here
|
||||||
|
# but not in AUTHORS, because Google holds the copyright.
|
||||||
|
#
|
||||||
|
# The submission process automatically checks to make sure
|
||||||
|
# that people submitting code are listed in this file (by email address).
|
||||||
|
#
|
||||||
|
# Names should be added to this file only after verifying that
|
||||||
|
# the individual or the individual's organization has agreed to
|
||||||
|
# the appropriate Contributor License Agreement, found here:
|
||||||
|
#
|
||||||
|
# http://code.google.com/legal/individual-cla-v1.0.html
|
||||||
|
# http://code.google.com/legal/corporate-cla-v1.0.html
|
||||||
|
#
|
||||||
|
# The agreement for individuals can be filled out on the web.
|
||||||
|
#
|
||||||
|
# When adding J Random Contributor's name to this file,
|
||||||
|
# either J's name or J's organization's name should be
|
||||||
|
# added to the AUTHORS file, depending on whether the
|
||||||
|
# individual or corporate CLA was used.
|
||||||
|
|
||||||
|
# Names should be added to this file like so:
|
||||||
|
# Name <email address>
|
||||||
|
|
||||||
|
# Please keep the list sorted.
|
||||||
|
|
||||||
|
Damian Gryski <dgryski@gmail.com>
|
||||||
|
Jan Mercl <0xjnml@gmail.com>
|
||||||
|
Kai Backman <kaib@golang.org>
|
||||||
|
Marc-Antoine Ruel <maruel@chromium.org>
|
||||||
|
Nigel Tao <nigeltao@golang.org>
|
||||||
|
Rob Pike <r@golang.org>
|
||||||
|
Rodolfo Carvalho <rhcarvalho@gmail.com>
|
||||||
|
Russ Cox <rsc@golang.org>
|
||||||
|
Sebastien Binet <seb.binet@gmail.com>
|
|
@ -0,0 +1,27 @@
|
||||||
|
Copyright (c) 2011 The Snappy-Go Authors. All rights reserved.
|
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without
|
||||||
|
modification, are permitted provided that the following conditions are
|
||||||
|
met:
|
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright
|
||||||
|
notice, this list of conditions and the following disclaimer.
|
||||||
|
* Redistributions in binary form must reproduce the above
|
||||||
|
copyright notice, this list of conditions and the following disclaimer
|
||||||
|
in the documentation and/or other materials provided with the
|
||||||
|
distribution.
|
||||||
|
* Neither the name of Google Inc. nor the names of its
|
||||||
|
contributors may be used to endorse or promote products derived from
|
||||||
|
this software without specific prior written permission.
|
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
@ -0,0 +1,107 @@
|
||||||
|
The Snappy compression format in the Go programming language.
|
||||||
|
|
||||||
|
To download and install from source:
|
||||||
|
$ go get github.com/golang/snappy
|
||||||
|
|
||||||
|
Unless otherwise noted, the Snappy-Go source files are distributed
|
||||||
|
under the BSD-style license found in the LICENSE file.
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
Benchmarks.
|
||||||
|
|
||||||
|
The golang/snappy benchmarks include compressing (Z) and decompressing (U) ten
|
||||||
|
or so files, the same set used by the C++ Snappy code (github.com/google/snappy
|
||||||
|
and note the "google", not "golang"). On an "Intel(R) Core(TM) i7-3770 CPU @
|
||||||
|
3.40GHz", Go's GOARCH=amd64 numbers as of 2016-05-29:
|
||||||
|
|
||||||
|
"go test -test.bench=."
|
||||||
|
|
||||||
|
_UFlat0-8 2.19GB/s ± 0% html
|
||||||
|
_UFlat1-8 1.41GB/s ± 0% urls
|
||||||
|
_UFlat2-8 23.5GB/s ± 2% jpg
|
||||||
|
_UFlat3-8 1.91GB/s ± 0% jpg_200
|
||||||
|
_UFlat4-8 14.0GB/s ± 1% pdf
|
||||||
|
_UFlat5-8 1.97GB/s ± 0% html4
|
||||||
|
_UFlat6-8 814MB/s ± 0% txt1
|
||||||
|
_UFlat7-8 785MB/s ± 0% txt2
|
||||||
|
_UFlat8-8 857MB/s ± 0% txt3
|
||||||
|
_UFlat9-8 719MB/s ± 1% txt4
|
||||||
|
_UFlat10-8 2.84GB/s ± 0% pb
|
||||||
|
_UFlat11-8 1.05GB/s ± 0% gaviota
|
||||||
|
|
||||||
|
_ZFlat0-8 1.04GB/s ± 0% html
|
||||||
|
_ZFlat1-8 534MB/s ± 0% urls
|
||||||
|
_ZFlat2-8 15.7GB/s ± 1% jpg
|
||||||
|
_ZFlat3-8 740MB/s ± 3% jpg_200
|
||||||
|
_ZFlat4-8 9.20GB/s ± 1% pdf
|
||||||
|
_ZFlat5-8 991MB/s ± 0% html4
|
||||||
|
_ZFlat6-8 379MB/s ± 0% txt1
|
||||||
|
_ZFlat7-8 352MB/s ± 0% txt2
|
||||||
|
_ZFlat8-8 396MB/s ± 1% txt3
|
||||||
|
_ZFlat9-8 327MB/s ± 1% txt4
|
||||||
|
_ZFlat10-8 1.33GB/s ± 1% pb
|
||||||
|
_ZFlat11-8 605MB/s ± 1% gaviota
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
"go test -test.bench=. -tags=noasm"
|
||||||
|
|
||||||
|
_UFlat0-8 621MB/s ± 2% html
|
||||||
|
_UFlat1-8 494MB/s ± 1% urls
|
||||||
|
_UFlat2-8 23.2GB/s ± 1% jpg
|
||||||
|
_UFlat3-8 1.12GB/s ± 1% jpg_200
|
||||||
|
_UFlat4-8 4.35GB/s ± 1% pdf
|
||||||
|
_UFlat5-8 609MB/s ± 0% html4
|
||||||
|
_UFlat6-8 296MB/s ± 0% txt1
|
||||||
|
_UFlat7-8 288MB/s ± 0% txt2
|
||||||
|
_UFlat8-8 309MB/s ± 1% txt3
|
||||||
|
_UFlat9-8 280MB/s ± 1% txt4
|
||||||
|
_UFlat10-8 753MB/s ± 0% pb
|
||||||
|
_UFlat11-8 400MB/s ± 0% gaviota
|
||||||
|
|
||||||
|
_ZFlat0-8 409MB/s ± 1% html
|
||||||
|
_ZFlat1-8 250MB/s ± 1% urls
|
||||||
|
_ZFlat2-8 12.3GB/s ± 1% jpg
|
||||||
|
_ZFlat3-8 132MB/s ± 0% jpg_200
|
||||||
|
_ZFlat4-8 2.92GB/s ± 0% pdf
|
||||||
|
_ZFlat5-8 405MB/s ± 1% html4
|
||||||
|
_ZFlat6-8 179MB/s ± 1% txt1
|
||||||
|
_ZFlat7-8 170MB/s ± 1% txt2
|
||||||
|
_ZFlat8-8 189MB/s ± 1% txt3
|
||||||
|
_ZFlat9-8 164MB/s ± 1% txt4
|
||||||
|
_ZFlat10-8 479MB/s ± 1% pb
|
||||||
|
_ZFlat11-8 270MB/s ± 1% gaviota
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
For comparison (Go's encoded output is byte-for-byte identical to C++'s), here
|
||||||
|
are the numbers from C++ Snappy's
|
||||||
|
|
||||||
|
make CXXFLAGS="-O2 -DNDEBUG -g" clean snappy_unittest.log && cat snappy_unittest.log
|
||||||
|
|
||||||
|
BM_UFlat/0 2.4GB/s html
|
||||||
|
BM_UFlat/1 1.4GB/s urls
|
||||||
|
BM_UFlat/2 21.8GB/s jpg
|
||||||
|
BM_UFlat/3 1.5GB/s jpg_200
|
||||||
|
BM_UFlat/4 13.3GB/s pdf
|
||||||
|
BM_UFlat/5 2.1GB/s html4
|
||||||
|
BM_UFlat/6 1.0GB/s txt1
|
||||||
|
BM_UFlat/7 959.4MB/s txt2
|
||||||
|
BM_UFlat/8 1.0GB/s txt3
|
||||||
|
BM_UFlat/9 864.5MB/s txt4
|
||||||
|
BM_UFlat/10 2.9GB/s pb
|
||||||
|
BM_UFlat/11 1.2GB/s gaviota
|
||||||
|
|
||||||
|
BM_ZFlat/0 944.3MB/s html (22.31 %)
|
||||||
|
BM_ZFlat/1 501.6MB/s urls (47.78 %)
|
||||||
|
BM_ZFlat/2 14.3GB/s jpg (99.95 %)
|
||||||
|
BM_ZFlat/3 538.3MB/s jpg_200 (73.00 %)
|
||||||
|
BM_ZFlat/4 8.3GB/s pdf (83.30 %)
|
||||||
|
BM_ZFlat/5 903.5MB/s html4 (22.52 %)
|
||||||
|
BM_ZFlat/6 336.0MB/s txt1 (57.88 %)
|
||||||
|
BM_ZFlat/7 312.3MB/s txt2 (61.91 %)
|
||||||
|
BM_ZFlat/8 353.1MB/s txt3 (54.99 %)
|
||||||
|
BM_ZFlat/9 289.9MB/s txt4 (66.26 %)
|
||||||
|
BM_ZFlat/10 1.2GB/s pb (19.68 %)
|
||||||
|
BM_ZFlat/11 527.4MB/s gaviota (37.72 %)
|
|
@ -0,0 +1,237 @@
|
||||||
|
// Copyright 2011 The Snappy-Go Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
package snappy
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/binary"
|
||||||
|
"errors"
|
||||||
|
"io"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
// ErrCorrupt reports that the input is invalid.
|
||||||
|
ErrCorrupt = errors.New("snappy: corrupt input")
|
||||||
|
// ErrTooLarge reports that the uncompressed length is too large.
|
||||||
|
ErrTooLarge = errors.New("snappy: decoded block is too large")
|
||||||
|
// ErrUnsupported reports that the input isn't supported.
|
||||||
|
ErrUnsupported = errors.New("snappy: unsupported input")
|
||||||
|
|
||||||
|
errUnsupportedLiteralLength = errors.New("snappy: unsupported literal length")
|
||||||
|
)
|
||||||
|
|
||||||
|
// DecodedLen returns the length of the decoded block.
|
||||||
|
func DecodedLen(src []byte) (int, error) {
|
||||||
|
v, _, err := decodedLen(src)
|
||||||
|
return v, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// decodedLen returns the length of the decoded block and the number of bytes
|
||||||
|
// that the length header occupied.
|
||||||
|
func decodedLen(src []byte) (blockLen, headerLen int, err error) {
|
||||||
|
v, n := binary.Uvarint(src)
|
||||||
|
if n <= 0 || v > 0xffffffff {
|
||||||
|
return 0, 0, ErrCorrupt
|
||||||
|
}
|
||||||
|
|
||||||
|
const wordSize = 32 << (^uint(0) >> 32 & 1)
|
||||||
|
if wordSize == 32 && v > 0x7fffffff {
|
||||||
|
return 0, 0, ErrTooLarge
|
||||||
|
}
|
||||||
|
return int(v), n, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
const (
|
||||||
|
decodeErrCodeCorrupt = 1
|
||||||
|
decodeErrCodeUnsupportedLiteralLength = 2
|
||||||
|
)
|
||||||
|
|
||||||
|
// Decode returns the decoded form of src. The returned slice may be a sub-
|
||||||
|
// slice of dst if dst was large enough to hold the entire decoded block.
|
||||||
|
// Otherwise, a newly allocated slice will be returned.
|
||||||
|
//
|
||||||
|
// The dst and src must not overlap. It is valid to pass a nil dst.
|
||||||
|
func Decode(dst, src []byte) ([]byte, error) {
|
||||||
|
dLen, s, err := decodedLen(src)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if dLen <= len(dst) {
|
||||||
|
dst = dst[:dLen]
|
||||||
|
} else {
|
||||||
|
dst = make([]byte, dLen)
|
||||||
|
}
|
||||||
|
switch decode(dst, src[s:]) {
|
||||||
|
case 0:
|
||||||
|
return dst, nil
|
||||||
|
case decodeErrCodeUnsupportedLiteralLength:
|
||||||
|
return nil, errUnsupportedLiteralLength
|
||||||
|
}
|
||||||
|
return nil, ErrCorrupt
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewReader returns a new Reader that decompresses from r, using the framing
|
||||||
|
// format described at
|
||||||
|
// https://github.com/google/snappy/blob/master/framing_format.txt
|
||||||
|
func NewReader(r io.Reader) *Reader {
|
||||||
|
return &Reader{
|
||||||
|
r: r,
|
||||||
|
decoded: make([]byte, maxBlockSize),
|
||||||
|
buf: make([]byte, maxEncodedLenOfMaxBlockSize+checksumSize),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Reader is an io.Reader that can read Snappy-compressed bytes.
|
||||||
|
type Reader struct {
|
||||||
|
r io.Reader
|
||||||
|
err error
|
||||||
|
decoded []byte
|
||||||
|
buf []byte
|
||||||
|
// decoded[i:j] contains decoded bytes that have not yet been passed on.
|
||||||
|
i, j int
|
||||||
|
readHeader bool
|
||||||
|
}
|
||||||
|
|
||||||
|
// Reset discards any buffered data, resets all state, and switches the Snappy
|
||||||
|
// reader to read from r. This permits reusing a Reader rather than allocating
|
||||||
|
// a new one.
|
||||||
|
func (r *Reader) Reset(reader io.Reader) {
|
||||||
|
r.r = reader
|
||||||
|
r.err = nil
|
||||||
|
r.i = 0
|
||||||
|
r.j = 0
|
||||||
|
r.readHeader = false
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *Reader) readFull(p []byte, allowEOF bool) (ok bool) {
|
||||||
|
if _, r.err = io.ReadFull(r.r, p); r.err != nil {
|
||||||
|
if r.err == io.ErrUnexpectedEOF || (r.err == io.EOF && !allowEOF) {
|
||||||
|
r.err = ErrCorrupt
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// Read satisfies the io.Reader interface.
|
||||||
|
func (r *Reader) Read(p []byte) (int, error) {
|
||||||
|
if r.err != nil {
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
for {
|
||||||
|
if r.i < r.j {
|
||||||
|
n := copy(p, r.decoded[r.i:r.j])
|
||||||
|
r.i += n
|
||||||
|
return n, nil
|
||||||
|
}
|
||||||
|
if !r.readFull(r.buf[:4], true) {
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
chunkType := r.buf[0]
|
||||||
|
if !r.readHeader {
|
||||||
|
if chunkType != chunkTypeStreamIdentifier {
|
||||||
|
r.err = ErrCorrupt
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
r.readHeader = true
|
||||||
|
}
|
||||||
|
chunkLen := int(r.buf[1]) | int(r.buf[2])<<8 | int(r.buf[3])<<16
|
||||||
|
if chunkLen > len(r.buf) {
|
||||||
|
r.err = ErrUnsupported
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
|
||||||
|
// The chunk types are specified at
|
||||||
|
// https://github.com/google/snappy/blob/master/framing_format.txt
|
||||||
|
switch chunkType {
|
||||||
|
case chunkTypeCompressedData:
|
||||||
|
// Section 4.2. Compressed data (chunk type 0x00).
|
||||||
|
if chunkLen < checksumSize {
|
||||||
|
r.err = ErrCorrupt
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
buf := r.buf[:chunkLen]
|
||||||
|
if !r.readFull(buf, false) {
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
checksum := uint32(buf[0]) | uint32(buf[1])<<8 | uint32(buf[2])<<16 | uint32(buf[3])<<24
|
||||||
|
buf = buf[checksumSize:]
|
||||||
|
|
||||||
|
n, err := DecodedLen(buf)
|
||||||
|
if err != nil {
|
||||||
|
r.err = err
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
if n > len(r.decoded) {
|
||||||
|
r.err = ErrCorrupt
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
if _, err := Decode(r.decoded, buf); err != nil {
|
||||||
|
r.err = err
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
if crc(r.decoded[:n]) != checksum {
|
||||||
|
r.err = ErrCorrupt
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
r.i, r.j = 0, n
|
||||||
|
continue
|
||||||
|
|
||||||
|
case chunkTypeUncompressedData:
|
||||||
|
// Section 4.3. Uncompressed data (chunk type 0x01).
|
||||||
|
if chunkLen < checksumSize {
|
||||||
|
r.err = ErrCorrupt
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
buf := r.buf[:checksumSize]
|
||||||
|
if !r.readFull(buf, false) {
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
checksum := uint32(buf[0]) | uint32(buf[1])<<8 | uint32(buf[2])<<16 | uint32(buf[3])<<24
|
||||||
|
// Read directly into r.decoded instead of via r.buf.
|
||||||
|
n := chunkLen - checksumSize
|
||||||
|
if n > len(r.decoded) {
|
||||||
|
r.err = ErrCorrupt
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
if !r.readFull(r.decoded[:n], false) {
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
if crc(r.decoded[:n]) != checksum {
|
||||||
|
r.err = ErrCorrupt
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
r.i, r.j = 0, n
|
||||||
|
continue
|
||||||
|
|
||||||
|
case chunkTypeStreamIdentifier:
|
||||||
|
// Section 4.1. Stream identifier (chunk type 0xff).
|
||||||
|
if chunkLen != len(magicBody) {
|
||||||
|
r.err = ErrCorrupt
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
if !r.readFull(r.buf[:len(magicBody)], false) {
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
for i := 0; i < len(magicBody); i++ {
|
||||||
|
if r.buf[i] != magicBody[i] {
|
||||||
|
r.err = ErrCorrupt
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
if chunkType <= 0x7f {
|
||||||
|
// Section 4.5. Reserved unskippable chunks (chunk types 0x02-0x7f).
|
||||||
|
r.err = ErrUnsupported
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
// Section 4.4 Padding (chunk type 0xfe).
|
||||||
|
// Section 4.6. Reserved skippable chunks (chunk types 0x80-0xfd).
|
||||||
|
if !r.readFull(r.buf[:chunkLen], false) {
|
||||||
|
return 0, r.err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,14 @@
|
||||||
|
// Copyright 2016 The Snappy-Go Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
// +build !appengine
|
||||||
|
// +build gc
|
||||||
|
// +build !noasm
|
||||||
|
|
||||||
|
package snappy
|
||||||
|
|
||||||
|
// decode has the same semantics as in decode_other.go.
|
||||||
|
//
|
||||||
|
//go:noescape
|
||||||
|
func decode(dst, src []byte) int
|
|
@ -0,0 +1,490 @@
|
||||||
|
// Copyright 2016 The Go Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
// +build !appengine
|
||||||
|
// +build gc
|
||||||
|
// +build !noasm
|
||||||
|
|
||||||
|
#include "textflag.h"
|
||||||
|
|
||||||
|
// The asm code generally follows the pure Go code in decode_other.go, except
|
||||||
|
// where marked with a "!!!".
|
||||||
|
|
||||||
|
// func decode(dst, src []byte) int
|
||||||
|
//
|
||||||
|
// All local variables fit into registers. The non-zero stack size is only to
|
||||||
|
// spill registers and push args when issuing a CALL. The register allocation:
|
||||||
|
// - AX scratch
|
||||||
|
// - BX scratch
|
||||||
|
// - CX length or x
|
||||||
|
// - DX offset
|
||||||
|
// - SI &src[s]
|
||||||
|
// - DI &dst[d]
|
||||||
|
// + R8 dst_base
|
||||||
|
// + R9 dst_len
|
||||||
|
// + R10 dst_base + dst_len
|
||||||
|
// + R11 src_base
|
||||||
|
// + R12 src_len
|
||||||
|
// + R13 src_base + src_len
|
||||||
|
// - R14 used by doCopy
|
||||||
|
// - R15 used by doCopy
|
||||||
|
//
|
||||||
|
// The registers R8-R13 (marked with a "+") are set at the start of the
|
||||||
|
// function, and after a CALL returns, and are not otherwise modified.
|
||||||
|
//
|
||||||
|
// The d variable is implicitly DI - R8, and len(dst)-d is R10 - DI.
|
||||||
|
// The s variable is implicitly SI - R11, and len(src)-s is R13 - SI.
|
||||||
|
TEXT ·decode(SB), NOSPLIT, $48-56
|
||||||
|
// Initialize SI, DI and R8-R13.
|
||||||
|
MOVQ dst_base+0(FP), R8
|
||||||
|
MOVQ dst_len+8(FP), R9
|
||||||
|
MOVQ R8, DI
|
||||||
|
MOVQ R8, R10
|
||||||
|
ADDQ R9, R10
|
||||||
|
MOVQ src_base+24(FP), R11
|
||||||
|
MOVQ src_len+32(FP), R12
|
||||||
|
MOVQ R11, SI
|
||||||
|
MOVQ R11, R13
|
||||||
|
ADDQ R12, R13
|
||||||
|
|
||||||
|
loop:
|
||||||
|
// for s < len(src)
|
||||||
|
CMPQ SI, R13
|
||||||
|
JEQ end
|
||||||
|
|
||||||
|
// CX = uint32(src[s])
|
||||||
|
//
|
||||||
|
// switch src[s] & 0x03
|
||||||
|
MOVBLZX (SI), CX
|
||||||
|
MOVL CX, BX
|
||||||
|
ANDL $3, BX
|
||||||
|
CMPL BX, $1
|
||||||
|
JAE tagCopy
|
||||||
|
|
||||||
|
// ----------------------------------------
|
||||||
|
// The code below handles literal tags.
|
||||||
|
|
||||||
|
// case tagLiteral:
|
||||||
|
// x := uint32(src[s] >> 2)
|
||||||
|
// switch
|
||||||
|
SHRL $2, CX
|
||||||
|
CMPL CX, $60
|
||||||
|
JAE tagLit60Plus
|
||||||
|
|
||||||
|
// case x < 60:
|
||||||
|
// s++
|
||||||
|
INCQ SI
|
||||||
|
|
||||||
|
doLit:
|
||||||
|
// This is the end of the inner "switch", when we have a literal tag.
|
||||||
|
//
|
||||||
|
// We assume that CX == x and x fits in a uint32, where x is the variable
|
||||||
|
// used in the pure Go decode_other.go code.
|
||||||
|
|
||||||
|
// length = int(x) + 1
|
||||||
|
//
|
||||||
|
// Unlike the pure Go code, we don't need to check if length <= 0 because
|
||||||
|
// CX can hold 64 bits, so the increment cannot overflow.
|
||||||
|
INCQ CX
|
||||||
|
|
||||||
|
// Prepare to check if copying length bytes will run past the end of dst or
|
||||||
|
// src.
|
||||||
|
//
|
||||||
|
// AX = len(dst) - d
|
||||||
|
// BX = len(src) - s
|
||||||
|
MOVQ R10, AX
|
||||||
|
SUBQ DI, AX
|
||||||
|
MOVQ R13, BX
|
||||||
|
SUBQ SI, BX
|
||||||
|
|
||||||
|
// !!! Try a faster technique for short (16 or fewer bytes) copies.
|
||||||
|
//
|
||||||
|
// if length > 16 || len(dst)-d < 16 || len(src)-s < 16 {
|
||||||
|
// goto callMemmove // Fall back on calling runtime·memmove.
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// The C++ snappy code calls this TryFastAppend. It also checks len(src)-s
|
||||||
|
// against 21 instead of 16, because it cannot assume that all of its input
|
||||||
|
// is contiguous in memory and so it needs to leave enough source bytes to
|
||||||
|
// read the next tag without refilling buffers, but Go's Decode assumes
|
||||||
|
// contiguousness (the src argument is a []byte).
|
||||||
|
CMPQ CX, $16
|
||||||
|
JGT callMemmove
|
||||||
|
CMPQ AX, $16
|
||||||
|
JLT callMemmove
|
||||||
|
CMPQ BX, $16
|
||||||
|
JLT callMemmove
|
||||||
|
|
||||||
|
// !!! Implement the copy from src to dst as a 16-byte load and store.
|
||||||
|
// (Decode's documentation says that dst and src must not overlap.)
|
||||||
|
//
|
||||||
|
// This always copies 16 bytes, instead of only length bytes, but that's
|
||||||
|
// OK. If the input is a valid Snappy encoding then subsequent iterations
|
||||||
|
// will fix up the overrun. Otherwise, Decode returns a nil []byte (and a
|
||||||
|
// non-nil error), so the overrun will be ignored.
|
||||||
|
//
|
||||||
|
// Note that on amd64, it is legal and cheap to issue unaligned 8-byte or
|
||||||
|
// 16-byte loads and stores. This technique probably wouldn't be as
|
||||||
|
// effective on architectures that are fussier about alignment.
|
||||||
|
MOVOU 0(SI), X0
|
||||||
|
MOVOU X0, 0(DI)
|
||||||
|
|
||||||
|
// d += length
|
||||||
|
// s += length
|
||||||
|
ADDQ CX, DI
|
||||||
|
ADDQ CX, SI
|
||||||
|
JMP loop
|
||||||
|
|
||||||
|
callMemmove:
|
||||||
|
// if length > len(dst)-d || length > len(src)-s { etc }
|
||||||
|
CMPQ CX, AX
|
||||||
|
JGT errCorrupt
|
||||||
|
CMPQ CX, BX
|
||||||
|
JGT errCorrupt
|
||||||
|
|
||||||
|
// copy(dst[d:], src[s:s+length])
|
||||||
|
//
|
||||||
|
// This means calling runtime·memmove(&dst[d], &src[s], length), so we push
|
||||||
|
// DI, SI and CX as arguments. Coincidentally, we also need to spill those
|
||||||
|
// three registers to the stack, to save local variables across the CALL.
|
||||||
|
MOVQ DI, 0(SP)
|
||||||
|
MOVQ SI, 8(SP)
|
||||||
|
MOVQ CX, 16(SP)
|
||||||
|
MOVQ DI, 24(SP)
|
||||||
|
MOVQ SI, 32(SP)
|
||||||
|
MOVQ CX, 40(SP)
|
||||||
|
CALL runtime·memmove(SB)
|
||||||
|
|
||||||
|
// Restore local variables: unspill registers from the stack and
|
||||||
|
// re-calculate R8-R13.
|
||||||
|
MOVQ 24(SP), DI
|
||||||
|
MOVQ 32(SP), SI
|
||||||
|
MOVQ 40(SP), CX
|
||||||
|
MOVQ dst_base+0(FP), R8
|
||||||
|
MOVQ dst_len+8(FP), R9
|
||||||
|
MOVQ R8, R10
|
||||||
|
ADDQ R9, R10
|
||||||
|
MOVQ src_base+24(FP), R11
|
||||||
|
MOVQ src_len+32(FP), R12
|
||||||
|
MOVQ R11, R13
|
||||||
|
ADDQ R12, R13
|
||||||
|
|
||||||
|
// d += length
|
||||||
|
// s += length
|
||||||
|
ADDQ CX, DI
|
||||||
|
ADDQ CX, SI
|
||||||
|
JMP loop
|
||||||
|
|
||||||
|
tagLit60Plus:
|
||||||
|
// !!! This fragment does the
|
||||||
|
//
|
||||||
|
// s += x - 58; if uint(s) > uint(len(src)) { etc }
|
||||||
|
//
|
||||||
|
// checks. In the asm version, we code it once instead of once per switch case.
|
||||||
|
ADDQ CX, SI
|
||||||
|
SUBQ $58, SI
|
||||||
|
MOVQ SI, BX
|
||||||
|
SUBQ R11, BX
|
||||||
|
CMPQ BX, R12
|
||||||
|
JA errCorrupt
|
||||||
|
|
||||||
|
// case x == 60:
|
||||||
|
CMPL CX, $61
|
||||||
|
JEQ tagLit61
|
||||||
|
JA tagLit62Plus
|
||||||
|
|
||||||
|
// x = uint32(src[s-1])
|
||||||
|
MOVBLZX -1(SI), CX
|
||||||
|
JMP doLit
|
||||||
|
|
||||||
|
tagLit61:
|
||||||
|
// case x == 61:
|
||||||
|
// x = uint32(src[s-2]) | uint32(src[s-1])<<8
|
||||||
|
MOVWLZX -2(SI), CX
|
||||||
|
JMP doLit
|
||||||
|
|
||||||
|
tagLit62Plus:
|
||||||
|
CMPL CX, $62
|
||||||
|
JA tagLit63
|
||||||
|
|
||||||
|
// case x == 62:
|
||||||
|
// x = uint32(src[s-3]) | uint32(src[s-2])<<8 | uint32(src[s-1])<<16
|
||||||
|
MOVWLZX -3(SI), CX
|
||||||
|
MOVBLZX -1(SI), BX
|
||||||
|
SHLL $16, BX
|
||||||
|
ORL BX, CX
|
||||||
|
JMP doLit
|
||||||
|
|
||||||
|
tagLit63:
|
||||||
|
// case x == 63:
|
||||||
|
// x = uint32(src[s-4]) | uint32(src[s-3])<<8 | uint32(src[s-2])<<16 | uint32(src[s-1])<<24
|
||||||
|
MOVL -4(SI), CX
|
||||||
|
JMP doLit
|
||||||
|
|
||||||
|
// The code above handles literal tags.
|
||||||
|
// ----------------------------------------
|
||||||
|
// The code below handles copy tags.
|
||||||
|
|
||||||
|
tagCopy4:
|
||||||
|
// case tagCopy4:
|
||||||
|
// s += 5
|
||||||
|
ADDQ $5, SI
|
||||||
|
|
||||||
|
// if uint(s) > uint(len(src)) { etc }
|
||||||
|
MOVQ SI, BX
|
||||||
|
SUBQ R11, BX
|
||||||
|
CMPQ BX, R12
|
||||||
|
JA errCorrupt
|
||||||
|
|
||||||
|
// length = 1 + int(src[s-5])>>2
|
||||||
|
SHRQ $2, CX
|
||||||
|
INCQ CX
|
||||||
|
|
||||||
|
// offset = int(uint32(src[s-4]) | uint32(src[s-3])<<8 | uint32(src[s-2])<<16 | uint32(src[s-1])<<24)
|
||||||
|
MOVLQZX -4(SI), DX
|
||||||
|
JMP doCopy
|
||||||
|
|
||||||
|
tagCopy2:
|
||||||
|
// case tagCopy2:
|
||||||
|
// s += 3
|
||||||
|
ADDQ $3, SI
|
||||||
|
|
||||||
|
// if uint(s) > uint(len(src)) { etc }
|
||||||
|
MOVQ SI, BX
|
||||||
|
SUBQ R11, BX
|
||||||
|
CMPQ BX, R12
|
||||||
|
JA errCorrupt
|
||||||
|
|
||||||
|
// length = 1 + int(src[s-3])>>2
|
||||||
|
SHRQ $2, CX
|
||||||
|
INCQ CX
|
||||||
|
|
||||||
|
// offset = int(uint32(src[s-2]) | uint32(src[s-1])<<8)
|
||||||
|
MOVWQZX -2(SI), DX
|
||||||
|
JMP doCopy
|
||||||
|
|
||||||
|
tagCopy:
|
||||||
|
// We have a copy tag. We assume that:
|
||||||
|
// - BX == src[s] & 0x03
|
||||||
|
// - CX == src[s]
|
||||||
|
CMPQ BX, $2
|
||||||
|
JEQ tagCopy2
|
||||||
|
JA tagCopy4
|
||||||
|
|
||||||
|
// case tagCopy1:
|
||||||
|
// s += 2
|
||||||
|
ADDQ $2, SI
|
||||||
|
|
||||||
|
// if uint(s) > uint(len(src)) { etc }
|
||||||
|
MOVQ SI, BX
|
||||||
|
SUBQ R11, BX
|
||||||
|
CMPQ BX, R12
|
||||||
|
JA errCorrupt
|
||||||
|
|
||||||
|
// offset = int(uint32(src[s-2])&0xe0<<3 | uint32(src[s-1]))
|
||||||
|
MOVQ CX, DX
|
||||||
|
ANDQ $0xe0, DX
|
||||||
|
SHLQ $3, DX
|
||||||
|
MOVBQZX -1(SI), BX
|
||||||
|
ORQ BX, DX
|
||||||
|
|
||||||
|
// length = 4 + int(src[s-2])>>2&0x7
|
||||||
|
SHRQ $2, CX
|
||||||
|
ANDQ $7, CX
|
||||||
|
ADDQ $4, CX
|
||||||
|
|
||||||
|
doCopy:
|
||||||
|
// This is the end of the outer "switch", when we have a copy tag.
|
||||||
|
//
|
||||||
|
// We assume that:
|
||||||
|
// - CX == length && CX > 0
|
||||||
|
// - DX == offset
|
||||||
|
|
||||||
|
// if offset <= 0 { etc }
|
||||||
|
CMPQ DX, $0
|
||||||
|
JLE errCorrupt
|
||||||
|
|
||||||
|
// if d < offset { etc }
|
||||||
|
MOVQ DI, BX
|
||||||
|
SUBQ R8, BX
|
||||||
|
CMPQ BX, DX
|
||||||
|
JLT errCorrupt
|
||||||
|
|
||||||
|
// if length > len(dst)-d { etc }
|
||||||
|
MOVQ R10, BX
|
||||||
|
SUBQ DI, BX
|
||||||
|
CMPQ CX, BX
|
||||||
|
JGT errCorrupt
|
||||||
|
|
||||||
|
// forwardCopy(dst[d:d+length], dst[d-offset:]); d += length
|
||||||
|
//
|
||||||
|
// Set:
|
||||||
|
// - R14 = len(dst)-d
|
||||||
|
// - R15 = &dst[d-offset]
|
||||||
|
MOVQ R10, R14
|
||||||
|
SUBQ DI, R14
|
||||||
|
MOVQ DI, R15
|
||||||
|
SUBQ DX, R15
|
||||||
|
|
||||||
|
// !!! Try a faster technique for short (16 or fewer bytes) forward copies.
|
||||||
|
//
|
||||||
|
// First, try using two 8-byte load/stores, similar to the doLit technique
|
||||||
|
// above. Even if dst[d:d+length] and dst[d-offset:] can overlap, this is
|
||||||
|
// still OK if offset >= 8. Note that this has to be two 8-byte load/stores
|
||||||
|
// and not one 16-byte load/store, and the first store has to be before the
|
||||||
|
// second load, due to the overlap if offset is in the range [8, 16).
|
||||||
|
//
|
||||||
|
// if length > 16 || offset < 8 || len(dst)-d < 16 {
|
||||||
|
// goto slowForwardCopy
|
||||||
|
// }
|
||||||
|
// copy 16 bytes
|
||||||
|
// d += length
|
||||||
|
CMPQ CX, $16
|
||||||
|
JGT slowForwardCopy
|
||||||
|
CMPQ DX, $8
|
||||||
|
JLT slowForwardCopy
|
||||||
|
CMPQ R14, $16
|
||||||
|
JLT slowForwardCopy
|
||||||
|
MOVQ 0(R15), AX
|
||||||
|
MOVQ AX, 0(DI)
|
||||||
|
MOVQ 8(R15), BX
|
||||||
|
MOVQ BX, 8(DI)
|
||||||
|
ADDQ CX, DI
|
||||||
|
JMP loop
|
||||||
|
|
||||||
|
slowForwardCopy:
|
||||||
|
// !!! If the forward copy is longer than 16 bytes, or if offset < 8, we
|
||||||
|
// can still try 8-byte load stores, provided we can overrun up to 10 extra
|
||||||
|
// bytes. As above, the overrun will be fixed up by subsequent iterations
|
||||||
|
// of the outermost loop.
|
||||||
|
//
|
||||||
|
// The C++ snappy code calls this technique IncrementalCopyFastPath. Its
|
||||||
|
// commentary says:
|
||||||
|
//
|
||||||
|
// ----
|
||||||
|
//
|
||||||
|
// The main part of this loop is a simple copy of eight bytes at a time
|
||||||
|
// until we've copied (at least) the requested amount of bytes. However,
|
||||||
|
// if d and d-offset are less than eight bytes apart (indicating a
|
||||||
|
// repeating pattern of length < 8), we first need to expand the pattern in
|
||||||
|
// order to get the correct results. For instance, if the buffer looks like
|
||||||
|
// this, with the eight-byte <d-offset> and <d> patterns marked as
|
||||||
|
// intervals:
|
||||||
|
//
|
||||||
|
// abxxxxxxxxxxxx
|
||||||
|
// [------] d-offset
|
||||||
|
// [------] d
|
||||||
|
//
|
||||||
|
// a single eight-byte copy from <d-offset> to <d> will repeat the pattern
|
||||||
|
// once, after which we can move <d> two bytes without moving <d-offset>:
|
||||||
|
//
|
||||||
|
// ababxxxxxxxxxx
|
||||||
|
// [------] d-offset
|
||||||
|
// [------] d
|
||||||
|
//
|
||||||
|
// and repeat the exercise until the two no longer overlap.
|
||||||
|
//
|
||||||
|
// This allows us to do very well in the special case of one single byte
|
||||||
|
// repeated many times, without taking a big hit for more general cases.
|
||||||
|
//
|
||||||
|
// The worst case of extra writing past the end of the match occurs when
|
||||||
|
// offset == 1 and length == 1; the last copy will read from byte positions
|
||||||
|
// [0..7] and write to [4..11], whereas it was only supposed to write to
|
||||||
|
// position 1. Thus, ten excess bytes.
|
||||||
|
//
|
||||||
|
// ----
|
||||||
|
//
|
||||||
|
// That "10 byte overrun" worst case is confirmed by Go's
|
||||||
|
// TestSlowForwardCopyOverrun, which also tests the fixUpSlowForwardCopy
|
||||||
|
// and finishSlowForwardCopy algorithm.
|
||||||
|
//
|
||||||
|
// if length > len(dst)-d-10 {
|
||||||
|
// goto verySlowForwardCopy
|
||||||
|
// }
|
||||||
|
SUBQ $10, R14
|
||||||
|
CMPQ CX, R14
|
||||||
|
JGT verySlowForwardCopy
|
||||||
|
|
||||||
|
makeOffsetAtLeast8:
|
||||||
|
// !!! As above, expand the pattern so that offset >= 8 and we can use
|
||||||
|
// 8-byte load/stores.
|
||||||
|
//
|
||||||
|
// for offset < 8 {
|
||||||
|
// copy 8 bytes from dst[d-offset:] to dst[d:]
|
||||||
|
// length -= offset
|
||||||
|
// d += offset
|
||||||
|
// offset += offset
|
||||||
|
// // The two previous lines together means that d-offset, and therefore
|
||||||
|
// // R15, is unchanged.
|
||||||
|
// }
|
||||||
|
CMPQ DX, $8
|
||||||
|
JGE fixUpSlowForwardCopy
|
||||||
|
MOVQ (R15), BX
|
||||||
|
MOVQ BX, (DI)
|
||||||
|
SUBQ DX, CX
|
||||||
|
ADDQ DX, DI
|
||||||
|
ADDQ DX, DX
|
||||||
|
JMP makeOffsetAtLeast8
|
||||||
|
|
||||||
|
fixUpSlowForwardCopy:
|
||||||
|
// !!! Add length (which might be negative now) to d (implied by DI being
|
||||||
|
// &dst[d]) so that d ends up at the right place when we jump back to the
|
||||||
|
// top of the loop. Before we do that, though, we save DI to AX so that, if
|
||||||
|
// length is positive, copying the remaining length bytes will write to the
|
||||||
|
// right place.
|
||||||
|
MOVQ DI, AX
|
||||||
|
ADDQ CX, DI
|
||||||
|
|
||||||
|
finishSlowForwardCopy:
|
||||||
|
// !!! Repeat 8-byte load/stores until length <= 0. Ending with a negative
|
||||||
|
// length means that we overrun, but as above, that will be fixed up by
|
||||||
|
// subsequent iterations of the outermost loop.
|
||||||
|
CMPQ CX, $0
|
||||||
|
JLE loop
|
||||||
|
MOVQ (R15), BX
|
||||||
|
MOVQ BX, (AX)
|
||||||
|
ADDQ $8, R15
|
||||||
|
ADDQ $8, AX
|
||||||
|
SUBQ $8, CX
|
||||||
|
JMP finishSlowForwardCopy
|
||||||
|
|
||||||
|
verySlowForwardCopy:
|
||||||
|
// verySlowForwardCopy is a simple implementation of forward copy. In C
|
||||||
|
// parlance, this is a do/while loop instead of a while loop, since we know
|
||||||
|
// that length > 0. In Go syntax:
|
||||||
|
//
|
||||||
|
// for {
|
||||||
|
// dst[d] = dst[d - offset]
|
||||||
|
// d++
|
||||||
|
// length--
|
||||||
|
// if length == 0 {
|
||||||
|
// break
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
MOVB (R15), BX
|
||||||
|
MOVB BX, (DI)
|
||||||
|
INCQ R15
|
||||||
|
INCQ DI
|
||||||
|
DECQ CX
|
||||||
|
JNZ verySlowForwardCopy
|
||||||
|
JMP loop
|
||||||
|
|
||||||
|
// The code above handles copy tags.
|
||||||
|
// ----------------------------------------
|
||||||
|
|
||||||
|
end:
|
||||||
|
// This is the end of the "for s < len(src)".
|
||||||
|
//
|
||||||
|
// if d != len(dst) { etc }
|
||||||
|
CMPQ DI, R10
|
||||||
|
JNE errCorrupt
|
||||||
|
|
||||||
|
// return 0
|
||||||
|
MOVQ $0, ret+48(FP)
|
||||||
|
RET
|
||||||
|
|
||||||
|
errCorrupt:
|
||||||
|
// return decodeErrCodeCorrupt
|
||||||
|
MOVQ $1, ret+48(FP)
|
||||||
|
RET
|
|
@ -0,0 +1,101 @@
|
||||||
|
// Copyright 2016 The Snappy-Go Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
// +build !amd64 appengine !gc noasm
|
||||||
|
|
||||||
|
package snappy
|
||||||
|
|
||||||
|
// decode writes the decoding of src to dst. It assumes that the varint-encoded
|
||||||
|
// length of the decompressed bytes has already been read, and that len(dst)
|
||||||
|
// equals that length.
|
||||||
|
//
|
||||||
|
// It returns 0 on success or a decodeErrCodeXxx error code on failure.
|
||||||
|
func decode(dst, src []byte) int {
|
||||||
|
var d, s, offset, length int
|
||||||
|
for s < len(src) {
|
||||||
|
switch src[s] & 0x03 {
|
||||||
|
case tagLiteral:
|
||||||
|
x := uint32(src[s] >> 2)
|
||||||
|
switch {
|
||||||
|
case x < 60:
|
||||||
|
s++
|
||||||
|
case x == 60:
|
||||||
|
s += 2
|
||||||
|
if uint(s) > uint(len(src)) { // The uint conversions catch overflow from the previous line.
|
||||||
|
return decodeErrCodeCorrupt
|
||||||
|
}
|
||||||
|
x = uint32(src[s-1])
|
||||||
|
case x == 61:
|
||||||
|
s += 3
|
||||||
|
if uint(s) > uint(len(src)) { // The uint conversions catch overflow from the previous line.
|
||||||
|
return decodeErrCodeCorrupt
|
||||||
|
}
|
||||||
|
x = uint32(src[s-2]) | uint32(src[s-1])<<8
|
||||||
|
case x == 62:
|
||||||
|
s += 4
|
||||||
|
if uint(s) > uint(len(src)) { // The uint conversions catch overflow from the previous line.
|
||||||
|
return decodeErrCodeCorrupt
|
||||||
|
}
|
||||||
|
x = uint32(src[s-3]) | uint32(src[s-2])<<8 | uint32(src[s-1])<<16
|
||||||
|
case x == 63:
|
||||||
|
s += 5
|
||||||
|
if uint(s) > uint(len(src)) { // The uint conversions catch overflow from the previous line.
|
||||||
|
return decodeErrCodeCorrupt
|
||||||
|
}
|
||||||
|
x = uint32(src[s-4]) | uint32(src[s-3])<<8 | uint32(src[s-2])<<16 | uint32(src[s-1])<<24
|
||||||
|
}
|
||||||
|
length = int(x) + 1
|
||||||
|
if length <= 0 {
|
||||||
|
return decodeErrCodeUnsupportedLiteralLength
|
||||||
|
}
|
||||||
|
if length > len(dst)-d || length > len(src)-s {
|
||||||
|
return decodeErrCodeCorrupt
|
||||||
|
}
|
||||||
|
copy(dst[d:], src[s:s+length])
|
||||||
|
d += length
|
||||||
|
s += length
|
||||||
|
continue
|
||||||
|
|
||||||
|
case tagCopy1:
|
||||||
|
s += 2
|
||||||
|
if uint(s) > uint(len(src)) { // The uint conversions catch overflow from the previous line.
|
||||||
|
return decodeErrCodeCorrupt
|
||||||
|
}
|
||||||
|
length = 4 + int(src[s-2])>>2&0x7
|
||||||
|
offset = int(uint32(src[s-2])&0xe0<<3 | uint32(src[s-1]))
|
||||||
|
|
||||||
|
case tagCopy2:
|
||||||
|
s += 3
|
||||||
|
if uint(s) > uint(len(src)) { // The uint conversions catch overflow from the previous line.
|
||||||
|
return decodeErrCodeCorrupt
|
||||||
|
}
|
||||||
|
length = 1 + int(src[s-3])>>2
|
||||||
|
offset = int(uint32(src[s-2]) | uint32(src[s-1])<<8)
|
||||||
|
|
||||||
|
case tagCopy4:
|
||||||
|
s += 5
|
||||||
|
if uint(s) > uint(len(src)) { // The uint conversions catch overflow from the previous line.
|
||||||
|
return decodeErrCodeCorrupt
|
||||||
|
}
|
||||||
|
length = 1 + int(src[s-5])>>2
|
||||||
|
offset = int(uint32(src[s-4]) | uint32(src[s-3])<<8 | uint32(src[s-2])<<16 | uint32(src[s-1])<<24)
|
||||||
|
}
|
||||||
|
|
||||||
|
if offset <= 0 || d < offset || length > len(dst)-d {
|
||||||
|
return decodeErrCodeCorrupt
|
||||||
|
}
|
||||||
|
// Copy from an earlier sub-slice of dst to a later sub-slice. Unlike
|
||||||
|
// the built-in copy function, this byte-by-byte copy always runs
|
||||||
|
// forwards, even if the slices overlap. Conceptually, this is:
|
||||||
|
//
|
||||||
|
// d += forwardCopy(dst[d:d+length], dst[d-offset:])
|
||||||
|
for end := d + length; d != end; d++ {
|
||||||
|
dst[d] = dst[d-offset]
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if d != len(dst) {
|
||||||
|
return decodeErrCodeCorrupt
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
|
@ -0,0 +1,285 @@
|
||||||
|
// Copyright 2011 The Snappy-Go Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
package snappy
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/binary"
|
||||||
|
"errors"
|
||||||
|
"io"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Encode returns the encoded form of src. The returned slice may be a sub-
|
||||||
|
// slice of dst if dst was large enough to hold the entire encoded block.
|
||||||
|
// Otherwise, a newly allocated slice will be returned.
|
||||||
|
//
|
||||||
|
// The dst and src must not overlap. It is valid to pass a nil dst.
|
||||||
|
func Encode(dst, src []byte) []byte {
|
||||||
|
if n := MaxEncodedLen(len(src)); n < 0 {
|
||||||
|
panic(ErrTooLarge)
|
||||||
|
} else if len(dst) < n {
|
||||||
|
dst = make([]byte, n)
|
||||||
|
}
|
||||||
|
|
||||||
|
// The block starts with the varint-encoded length of the decompressed bytes.
|
||||||
|
d := binary.PutUvarint(dst, uint64(len(src)))
|
||||||
|
|
||||||
|
for len(src) > 0 {
|
||||||
|
p := src
|
||||||
|
src = nil
|
||||||
|
if len(p) > maxBlockSize {
|
||||||
|
p, src = p[:maxBlockSize], p[maxBlockSize:]
|
||||||
|
}
|
||||||
|
if len(p) < minNonLiteralBlockSize {
|
||||||
|
d += emitLiteral(dst[d:], p)
|
||||||
|
} else {
|
||||||
|
d += encodeBlock(dst[d:], p)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return dst[:d]
|
||||||
|
}
|
||||||
|
|
||||||
|
// inputMargin is the minimum number of extra input bytes to keep, inside
|
||||||
|
// encodeBlock's inner loop. On some architectures, this margin lets us
|
||||||
|
// implement a fast path for emitLiteral, where the copy of short (<= 16 byte)
|
||||||
|
// literals can be implemented as a single load to and store from a 16-byte
|
||||||
|
// register. That literal's actual length can be as short as 1 byte, so this
|
||||||
|
// can copy up to 15 bytes too much, but that's OK as subsequent iterations of
|
||||||
|
// the encoding loop will fix up the copy overrun, and this inputMargin ensures
|
||||||
|
// that we don't overrun the dst and src buffers.
|
||||||
|
const inputMargin = 16 - 1
|
||||||
|
|
||||||
|
// minNonLiteralBlockSize is the minimum size of the input to encodeBlock that
|
||||||
|
// could be encoded with a copy tag. This is the minimum with respect to the
|
||||||
|
// algorithm used by encodeBlock, not a minimum enforced by the file format.
|
||||||
|
//
|
||||||
|
// The encoded output must start with at least a 1 byte literal, as there are
|
||||||
|
// no previous bytes to copy. A minimal (1 byte) copy after that, generated
|
||||||
|
// from an emitCopy call in encodeBlock's main loop, would require at least
|
||||||
|
// another inputMargin bytes, for the reason above: we want any emitLiteral
|
||||||
|
// calls inside encodeBlock's main loop to use the fast path if possible, which
|
||||||
|
// requires being able to overrun by inputMargin bytes. Thus,
|
||||||
|
// minNonLiteralBlockSize equals 1 + 1 + inputMargin.
|
||||||
|
//
|
||||||
|
// The C++ code doesn't use this exact threshold, but it could, as discussed at
|
||||||
|
// https://groups.google.com/d/topic/snappy-compression/oGbhsdIJSJ8/discussion
|
||||||
|
// The difference between Go (2+inputMargin) and C++ (inputMargin) is purely an
|
||||||
|
// optimization. It should not affect the encoded form. This is tested by
|
||||||
|
// TestSameEncodingAsCppShortCopies.
|
||||||
|
const minNonLiteralBlockSize = 1 + 1 + inputMargin
|
||||||
|
|
||||||
|
// MaxEncodedLen returns the maximum length of a snappy block, given its
|
||||||
|
// uncompressed length.
|
||||||
|
//
|
||||||
|
// It will return a negative value if srcLen is too large to encode.
|
||||||
|
func MaxEncodedLen(srcLen int) int {
|
||||||
|
n := uint64(srcLen)
|
||||||
|
if n > 0xffffffff {
|
||||||
|
return -1
|
||||||
|
}
|
||||||
|
// Compressed data can be defined as:
|
||||||
|
// compressed := item* literal*
|
||||||
|
// item := literal* copy
|
||||||
|
//
|
||||||
|
// The trailing literal sequence has a space blowup of at most 62/60
|
||||||
|
// since a literal of length 60 needs one tag byte + one extra byte
|
||||||
|
// for length information.
|
||||||
|
//
|
||||||
|
// Item blowup is trickier to measure. Suppose the "copy" op copies
|
||||||
|
// 4 bytes of data. Because of a special check in the encoding code,
|
||||||
|
// we produce a 4-byte copy only if the offset is < 65536. Therefore
|
||||||
|
// the copy op takes 3 bytes to encode, and this type of item leads
|
||||||
|
// to at most the 62/60 blowup for representing literals.
|
||||||
|
//
|
||||||
|
// Suppose the "copy" op copies 5 bytes of data. If the offset is big
|
||||||
|
// enough, it will take 5 bytes to encode the copy op. Therefore the
|
||||||
|
// worst case here is a one-byte literal followed by a five-byte copy.
|
||||||
|
// That is, 6 bytes of input turn into 7 bytes of "compressed" data.
|
||||||
|
//
|
||||||
|
// This last factor dominates the blowup, so the final estimate is:
|
||||||
|
n = 32 + n + n/6
|
||||||
|
if n > 0xffffffff {
|
||||||
|
return -1
|
||||||
|
}
|
||||||
|
return int(n)
|
||||||
|
}
|
||||||
|
|
||||||
|
var errClosed = errors.New("snappy: Writer is closed")
|
||||||
|
|
||||||
|
// NewWriter returns a new Writer that compresses to w.
|
||||||
|
//
|
||||||
|
// The Writer returned does not buffer writes. There is no need to Flush or
|
||||||
|
// Close such a Writer.
|
||||||
|
//
|
||||||
|
// Deprecated: the Writer returned is not suitable for many small writes, only
|
||||||
|
// for few large writes. Use NewBufferedWriter instead, which is efficient
|
||||||
|
// regardless of the frequency and shape of the writes, and remember to Close
|
||||||
|
// that Writer when done.
|
||||||
|
func NewWriter(w io.Writer) *Writer {
|
||||||
|
return &Writer{
|
||||||
|
w: w,
|
||||||
|
obuf: make([]byte, obufLen),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewBufferedWriter returns a new Writer that compresses to w, using the
|
||||||
|
// framing format described at
|
||||||
|
// https://github.com/google/snappy/blob/master/framing_format.txt
|
||||||
|
//
|
||||||
|
// The Writer returned buffers writes. Users must call Close to guarantee all
|
||||||
|
// data has been forwarded to the underlying io.Writer. They may also call
|
||||||
|
// Flush zero or more times before calling Close.
|
||||||
|
func NewBufferedWriter(w io.Writer) *Writer {
|
||||||
|
return &Writer{
|
||||||
|
w: w,
|
||||||
|
ibuf: make([]byte, 0, maxBlockSize),
|
||||||
|
obuf: make([]byte, obufLen),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Writer is an io.Writer that can write Snappy-compressed bytes.
|
||||||
|
type Writer struct {
|
||||||
|
w io.Writer
|
||||||
|
err error
|
||||||
|
|
||||||
|
// ibuf is a buffer for the incoming (uncompressed) bytes.
|
||||||
|
//
|
||||||
|
// Its use is optional. For backwards compatibility, Writers created by the
|
||||||
|
// NewWriter function have ibuf == nil, do not buffer incoming bytes, and
|
||||||
|
// therefore do not need to be Flush'ed or Close'd.
|
||||||
|
ibuf []byte
|
||||||
|
|
||||||
|
// obuf is a buffer for the outgoing (compressed) bytes.
|
||||||
|
obuf []byte
|
||||||
|
|
||||||
|
// wroteStreamHeader is whether we have written the stream header.
|
||||||
|
wroteStreamHeader bool
|
||||||
|
}
|
||||||
|
|
||||||
|
// Reset discards the writer's state and switches the Snappy writer to write to
|
||||||
|
// w. This permits reusing a Writer rather than allocating a new one.
|
||||||
|
func (w *Writer) Reset(writer io.Writer) {
|
||||||
|
w.w = writer
|
||||||
|
w.err = nil
|
||||||
|
if w.ibuf != nil {
|
||||||
|
w.ibuf = w.ibuf[:0]
|
||||||
|
}
|
||||||
|
w.wroteStreamHeader = false
|
||||||
|
}
|
||||||
|
|
||||||
|
// Write satisfies the io.Writer interface.
|
||||||
|
func (w *Writer) Write(p []byte) (nRet int, errRet error) {
|
||||||
|
if w.ibuf == nil {
|
||||||
|
// Do not buffer incoming bytes. This does not perform or compress well
|
||||||
|
// if the caller of Writer.Write writes many small slices. This
|
||||||
|
// behavior is therefore deprecated, but still supported for backwards
|
||||||
|
// compatibility with code that doesn't explicitly Flush or Close.
|
||||||
|
return w.write(p)
|
||||||
|
}
|
||||||
|
|
||||||
|
// The remainder of this method is based on bufio.Writer.Write from the
|
||||||
|
// standard library.
|
||||||
|
|
||||||
|
for len(p) > (cap(w.ibuf)-len(w.ibuf)) && w.err == nil {
|
||||||
|
var n int
|
||||||
|
if len(w.ibuf) == 0 {
|
||||||
|
// Large write, empty buffer.
|
||||||
|
// Write directly from p to avoid copy.
|
||||||
|
n, _ = w.write(p)
|
||||||
|
} else {
|
||||||
|
n = copy(w.ibuf[len(w.ibuf):cap(w.ibuf)], p)
|
||||||
|
w.ibuf = w.ibuf[:len(w.ibuf)+n]
|
||||||
|
w.Flush()
|
||||||
|
}
|
||||||
|
nRet += n
|
||||||
|
p = p[n:]
|
||||||
|
}
|
||||||
|
if w.err != nil {
|
||||||
|
return nRet, w.err
|
||||||
|
}
|
||||||
|
n := copy(w.ibuf[len(w.ibuf):cap(w.ibuf)], p)
|
||||||
|
w.ibuf = w.ibuf[:len(w.ibuf)+n]
|
||||||
|
nRet += n
|
||||||
|
return nRet, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (w *Writer) write(p []byte) (nRet int, errRet error) {
|
||||||
|
if w.err != nil {
|
||||||
|
return 0, w.err
|
||||||
|
}
|
||||||
|
for len(p) > 0 {
|
||||||
|
obufStart := len(magicChunk)
|
||||||
|
if !w.wroteStreamHeader {
|
||||||
|
w.wroteStreamHeader = true
|
||||||
|
copy(w.obuf, magicChunk)
|
||||||
|
obufStart = 0
|
||||||
|
}
|
||||||
|
|
||||||
|
var uncompressed []byte
|
||||||
|
if len(p) > maxBlockSize {
|
||||||
|
uncompressed, p = p[:maxBlockSize], p[maxBlockSize:]
|
||||||
|
} else {
|
||||||
|
uncompressed, p = p, nil
|
||||||
|
}
|
||||||
|
checksum := crc(uncompressed)
|
||||||
|
|
||||||
|
// Compress the buffer, discarding the result if the improvement
|
||||||
|
// isn't at least 12.5%.
|
||||||
|
compressed := Encode(w.obuf[obufHeaderLen:], uncompressed)
|
||||||
|
chunkType := uint8(chunkTypeCompressedData)
|
||||||
|
chunkLen := 4 + len(compressed)
|
||||||
|
obufEnd := obufHeaderLen + len(compressed)
|
||||||
|
if len(compressed) >= len(uncompressed)-len(uncompressed)/8 {
|
||||||
|
chunkType = chunkTypeUncompressedData
|
||||||
|
chunkLen = 4 + len(uncompressed)
|
||||||
|
obufEnd = obufHeaderLen
|
||||||
|
}
|
||||||
|
|
||||||
|
// Fill in the per-chunk header that comes before the body.
|
||||||
|
w.obuf[len(magicChunk)+0] = chunkType
|
||||||
|
w.obuf[len(magicChunk)+1] = uint8(chunkLen >> 0)
|
||||||
|
w.obuf[len(magicChunk)+2] = uint8(chunkLen >> 8)
|
||||||
|
w.obuf[len(magicChunk)+3] = uint8(chunkLen >> 16)
|
||||||
|
w.obuf[len(magicChunk)+4] = uint8(checksum >> 0)
|
||||||
|
w.obuf[len(magicChunk)+5] = uint8(checksum >> 8)
|
||||||
|
w.obuf[len(magicChunk)+6] = uint8(checksum >> 16)
|
||||||
|
w.obuf[len(magicChunk)+7] = uint8(checksum >> 24)
|
||||||
|
|
||||||
|
if _, err := w.w.Write(w.obuf[obufStart:obufEnd]); err != nil {
|
||||||
|
w.err = err
|
||||||
|
return nRet, err
|
||||||
|
}
|
||||||
|
if chunkType == chunkTypeUncompressedData {
|
||||||
|
if _, err := w.w.Write(uncompressed); err != nil {
|
||||||
|
w.err = err
|
||||||
|
return nRet, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
nRet += len(uncompressed)
|
||||||
|
}
|
||||||
|
return nRet, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Flush flushes the Writer to its underlying io.Writer.
|
||||||
|
func (w *Writer) Flush() error {
|
||||||
|
if w.err != nil {
|
||||||
|
return w.err
|
||||||
|
}
|
||||||
|
if len(w.ibuf) == 0 {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
w.write(w.ibuf)
|
||||||
|
w.ibuf = w.ibuf[:0]
|
||||||
|
return w.err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close calls Flush and then closes the Writer.
|
||||||
|
func (w *Writer) Close() error {
|
||||||
|
w.Flush()
|
||||||
|
ret := w.err
|
||||||
|
if w.err == nil {
|
||||||
|
w.err = errClosed
|
||||||
|
}
|
||||||
|
return ret
|
||||||
|
}
|
|
@ -0,0 +1,29 @@
|
||||||
|
// Copyright 2016 The Snappy-Go Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
// +build !appengine
|
||||||
|
// +build gc
|
||||||
|
// +build !noasm
|
||||||
|
|
||||||
|
package snappy
|
||||||
|
|
||||||
|
// emitLiteral has the same semantics as in encode_other.go.
|
||||||
|
//
|
||||||
|
//go:noescape
|
||||||
|
func emitLiteral(dst, lit []byte) int
|
||||||
|
|
||||||
|
// emitCopy has the same semantics as in encode_other.go.
|
||||||
|
//
|
||||||
|
//go:noescape
|
||||||
|
func emitCopy(dst []byte, offset, length int) int
|
||||||
|
|
||||||
|
// extendMatch has the same semantics as in encode_other.go.
|
||||||
|
//
|
||||||
|
//go:noescape
|
||||||
|
func extendMatch(src []byte, i, j int) int
|
||||||
|
|
||||||
|
// encodeBlock has the same semantics as in encode_other.go.
|
||||||
|
//
|
||||||
|
//go:noescape
|
||||||
|
func encodeBlock(dst, src []byte) (d int)
|
|
@ -0,0 +1,730 @@
|
||||||
|
// Copyright 2016 The Go Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
// +build !appengine
|
||||||
|
// +build gc
|
||||||
|
// +build !noasm
|
||||||
|
|
||||||
|
#include "textflag.h"
|
||||||
|
|
||||||
|
// The XXX lines assemble on Go 1.4, 1.5 and 1.7, but not 1.6, due to a
|
||||||
|
// Go toolchain regression. See https://github.com/golang/go/issues/15426 and
|
||||||
|
// https://github.com/golang/snappy/issues/29
|
||||||
|
//
|
||||||
|
// As a workaround, the package was built with a known good assembler, and
|
||||||
|
// those instructions were disassembled by "objdump -d" to yield the
|
||||||
|
// 4e 0f b7 7c 5c 78 movzwq 0x78(%rsp,%r11,2),%r15
|
||||||
|
// style comments, in AT&T asm syntax. Note that rsp here is a physical
|
||||||
|
// register, not Go/asm's SP pseudo-register (see https://golang.org/doc/asm).
|
||||||
|
// The instructions were then encoded as "BYTE $0x.." sequences, which assemble
|
||||||
|
// fine on Go 1.6.
|
||||||
|
|
||||||
|
// The asm code generally follows the pure Go code in encode_other.go, except
|
||||||
|
// where marked with a "!!!".
|
||||||
|
|
||||||
|
// ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
// func emitLiteral(dst, lit []byte) int
|
||||||
|
//
|
||||||
|
// All local variables fit into registers. The register allocation:
|
||||||
|
// - AX len(lit)
|
||||||
|
// - BX n
|
||||||
|
// - DX return value
|
||||||
|
// - DI &dst[i]
|
||||||
|
// - R10 &lit[0]
|
||||||
|
//
|
||||||
|
// The 24 bytes of stack space is to call runtime·memmove.
|
||||||
|
//
|
||||||
|
// The unusual register allocation of local variables, such as R10 for the
|
||||||
|
// source pointer, matches the allocation used at the call site in encodeBlock,
|
||||||
|
// which makes it easier to manually inline this function.
|
||||||
|
TEXT ·emitLiteral(SB), NOSPLIT, $24-56
|
||||||
|
MOVQ dst_base+0(FP), DI
|
||||||
|
MOVQ lit_base+24(FP), R10
|
||||||
|
MOVQ lit_len+32(FP), AX
|
||||||
|
MOVQ AX, DX
|
||||||
|
MOVL AX, BX
|
||||||
|
SUBL $1, BX
|
||||||
|
|
||||||
|
CMPL BX, $60
|
||||||
|
JLT oneByte
|
||||||
|
CMPL BX, $256
|
||||||
|
JLT twoBytes
|
||||||
|
|
||||||
|
threeBytes:
|
||||||
|
MOVB $0xf4, 0(DI)
|
||||||
|
MOVW BX, 1(DI)
|
||||||
|
ADDQ $3, DI
|
||||||
|
ADDQ $3, DX
|
||||||
|
JMP memmove
|
||||||
|
|
||||||
|
twoBytes:
|
||||||
|
MOVB $0xf0, 0(DI)
|
||||||
|
MOVB BX, 1(DI)
|
||||||
|
ADDQ $2, DI
|
||||||
|
ADDQ $2, DX
|
||||||
|
JMP memmove
|
||||||
|
|
||||||
|
oneByte:
|
||||||
|
SHLB $2, BX
|
||||||
|
MOVB BX, 0(DI)
|
||||||
|
ADDQ $1, DI
|
||||||
|
ADDQ $1, DX
|
||||||
|
|
||||||
|
memmove:
|
||||||
|
MOVQ DX, ret+48(FP)
|
||||||
|
|
||||||
|
// copy(dst[i:], lit)
|
||||||
|
//
|
||||||
|
// This means calling runtime·memmove(&dst[i], &lit[0], len(lit)), so we push
|
||||||
|
// DI, R10 and AX as arguments.
|
||||||
|
MOVQ DI, 0(SP)
|
||||||
|
MOVQ R10, 8(SP)
|
||||||
|
MOVQ AX, 16(SP)
|
||||||
|
CALL runtime·memmove(SB)
|
||||||
|
RET
|
||||||
|
|
||||||
|
// ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
// func emitCopy(dst []byte, offset, length int) int
|
||||||
|
//
|
||||||
|
// All local variables fit into registers. The register allocation:
|
||||||
|
// - AX length
|
||||||
|
// - SI &dst[0]
|
||||||
|
// - DI &dst[i]
|
||||||
|
// - R11 offset
|
||||||
|
//
|
||||||
|
// The unusual register allocation of local variables, such as R11 for the
|
||||||
|
// offset, matches the allocation used at the call site in encodeBlock, which
|
||||||
|
// makes it easier to manually inline this function.
|
||||||
|
TEXT ·emitCopy(SB), NOSPLIT, $0-48
|
||||||
|
MOVQ dst_base+0(FP), DI
|
||||||
|
MOVQ DI, SI
|
||||||
|
MOVQ offset+24(FP), R11
|
||||||
|
MOVQ length+32(FP), AX
|
||||||
|
|
||||||
|
loop0:
|
||||||
|
// for length >= 68 { etc }
|
||||||
|
CMPL AX, $68
|
||||||
|
JLT step1
|
||||||
|
|
||||||
|
// Emit a length 64 copy, encoded as 3 bytes.
|
||||||
|
MOVB $0xfe, 0(DI)
|
||||||
|
MOVW R11, 1(DI)
|
||||||
|
ADDQ $3, DI
|
||||||
|
SUBL $64, AX
|
||||||
|
JMP loop0
|
||||||
|
|
||||||
|
step1:
|
||||||
|
// if length > 64 { etc }
|
||||||
|
CMPL AX, $64
|
||||||
|
JLE step2
|
||||||
|
|
||||||
|
// Emit a length 60 copy, encoded as 3 bytes.
|
||||||
|
MOVB $0xee, 0(DI)
|
||||||
|
MOVW R11, 1(DI)
|
||||||
|
ADDQ $3, DI
|
||||||
|
SUBL $60, AX
|
||||||
|
|
||||||
|
step2:
|
||||||
|
// if length >= 12 || offset >= 2048 { goto step3 }
|
||||||
|
CMPL AX, $12
|
||||||
|
JGE step3
|
||||||
|
CMPL R11, $2048
|
||||||
|
JGE step3
|
||||||
|
|
||||||
|
// Emit the remaining copy, encoded as 2 bytes.
|
||||||
|
MOVB R11, 1(DI)
|
||||||
|
SHRL $8, R11
|
||||||
|
SHLB $5, R11
|
||||||
|
SUBB $4, AX
|
||||||
|
SHLB $2, AX
|
||||||
|
ORB AX, R11
|
||||||
|
ORB $1, R11
|
||||||
|
MOVB R11, 0(DI)
|
||||||
|
ADDQ $2, DI
|
||||||
|
|
||||||
|
// Return the number of bytes written.
|
||||||
|
SUBQ SI, DI
|
||||||
|
MOVQ DI, ret+40(FP)
|
||||||
|
RET
|
||||||
|
|
||||||
|
step3:
|
||||||
|
// Emit the remaining copy, encoded as 3 bytes.
|
||||||
|
SUBL $1, AX
|
||||||
|
SHLB $2, AX
|
||||||
|
ORB $2, AX
|
||||||
|
MOVB AX, 0(DI)
|
||||||
|
MOVW R11, 1(DI)
|
||||||
|
ADDQ $3, DI
|
||||||
|
|
||||||
|
// Return the number of bytes written.
|
||||||
|
SUBQ SI, DI
|
||||||
|
MOVQ DI, ret+40(FP)
|
||||||
|
RET
|
||||||
|
|
||||||
|
// ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
// func extendMatch(src []byte, i, j int) int
|
||||||
|
//
|
||||||
|
// All local variables fit into registers. The register allocation:
|
||||||
|
// - DX &src[0]
|
||||||
|
// - SI &src[j]
|
||||||
|
// - R13 &src[len(src) - 8]
|
||||||
|
// - R14 &src[len(src)]
|
||||||
|
// - R15 &src[i]
|
||||||
|
//
|
||||||
|
// The unusual register allocation of local variables, such as R15 for a source
|
||||||
|
// pointer, matches the allocation used at the call site in encodeBlock, which
|
||||||
|
// makes it easier to manually inline this function.
|
||||||
|
TEXT ·extendMatch(SB), NOSPLIT, $0-48
|
||||||
|
MOVQ src_base+0(FP), DX
|
||||||
|
MOVQ src_len+8(FP), R14
|
||||||
|
MOVQ i+24(FP), R15
|
||||||
|
MOVQ j+32(FP), SI
|
||||||
|
ADDQ DX, R14
|
||||||
|
ADDQ DX, R15
|
||||||
|
ADDQ DX, SI
|
||||||
|
MOVQ R14, R13
|
||||||
|
SUBQ $8, R13
|
||||||
|
|
||||||
|
cmp8:
|
||||||
|
// As long as we are 8 or more bytes before the end of src, we can load and
|
||||||
|
// compare 8 bytes at a time. If those 8 bytes are equal, repeat.
|
||||||
|
CMPQ SI, R13
|
||||||
|
JA cmp1
|
||||||
|
MOVQ (R15), AX
|
||||||
|
MOVQ (SI), BX
|
||||||
|
CMPQ AX, BX
|
||||||
|
JNE bsf
|
||||||
|
ADDQ $8, R15
|
||||||
|
ADDQ $8, SI
|
||||||
|
JMP cmp8
|
||||||
|
|
||||||
|
bsf:
|
||||||
|
// If those 8 bytes were not equal, XOR the two 8 byte values, and return
|
||||||
|
// the index of the first byte that differs. The BSF instruction finds the
|
||||||
|
// least significant 1 bit, the amd64 architecture is little-endian, and
|
||||||
|
// the shift by 3 converts a bit index to a byte index.
|
||||||
|
XORQ AX, BX
|
||||||
|
BSFQ BX, BX
|
||||||
|
SHRQ $3, BX
|
||||||
|
ADDQ BX, SI
|
||||||
|
|
||||||
|
// Convert from &src[ret] to ret.
|
||||||
|
SUBQ DX, SI
|
||||||
|
MOVQ SI, ret+40(FP)
|
||||||
|
RET
|
||||||
|
|
||||||
|
cmp1:
|
||||||
|
// In src's tail, compare 1 byte at a time.
|
||||||
|
CMPQ SI, R14
|
||||||
|
JAE extendMatchEnd
|
||||||
|
MOVB (R15), AX
|
||||||
|
MOVB (SI), BX
|
||||||
|
CMPB AX, BX
|
||||||
|
JNE extendMatchEnd
|
||||||
|
ADDQ $1, R15
|
||||||
|
ADDQ $1, SI
|
||||||
|
JMP cmp1
|
||||||
|
|
||||||
|
extendMatchEnd:
|
||||||
|
// Convert from &src[ret] to ret.
|
||||||
|
SUBQ DX, SI
|
||||||
|
MOVQ SI, ret+40(FP)
|
||||||
|
RET
|
||||||
|
|
||||||
|
// ----------------------------------------------------------------------------
|
||||||
|
|
||||||
|
// func encodeBlock(dst, src []byte) (d int)
|
||||||
|
//
|
||||||
|
// All local variables fit into registers, other than "var table". The register
|
||||||
|
// allocation:
|
||||||
|
// - AX . .
|
||||||
|
// - BX . .
|
||||||
|
// - CX 56 shift (note that amd64 shifts by non-immediates must use CX).
|
||||||
|
// - DX 64 &src[0], tableSize
|
||||||
|
// - SI 72 &src[s]
|
||||||
|
// - DI 80 &dst[d]
|
||||||
|
// - R9 88 sLimit
|
||||||
|
// - R10 . &src[nextEmit]
|
||||||
|
// - R11 96 prevHash, currHash, nextHash, offset
|
||||||
|
// - R12 104 &src[base], skip
|
||||||
|
// - R13 . &src[nextS], &src[len(src) - 8]
|
||||||
|
// - R14 . len(src), bytesBetweenHashLookups, &src[len(src)], x
|
||||||
|
// - R15 112 candidate
|
||||||
|
//
|
||||||
|
// The second column (56, 64, etc) is the stack offset to spill the registers
|
||||||
|
// when calling other functions. We could pack this slightly tighter, but it's
|
||||||
|
// simpler to have a dedicated spill map independent of the function called.
|
||||||
|
//
|
||||||
|
// "var table [maxTableSize]uint16" takes up 32768 bytes of stack space. An
|
||||||
|
// extra 56 bytes, to call other functions, and an extra 64 bytes, to spill
|
||||||
|
// local variables (registers) during calls gives 32768 + 56 + 64 = 32888.
|
||||||
|
TEXT ·encodeBlock(SB), 0, $32888-56
|
||||||
|
MOVQ dst_base+0(FP), DI
|
||||||
|
MOVQ src_base+24(FP), SI
|
||||||
|
MOVQ src_len+32(FP), R14
|
||||||
|
|
||||||
|
// shift, tableSize := uint32(32-8), 1<<8
|
||||||
|
MOVQ $24, CX
|
||||||
|
MOVQ $256, DX
|
||||||
|
|
||||||
|
calcShift:
|
||||||
|
// for ; tableSize < maxTableSize && tableSize < len(src); tableSize *= 2 {
|
||||||
|
// shift--
|
||||||
|
// }
|
||||||
|
CMPQ DX, $16384
|
||||||
|
JGE varTable
|
||||||
|
CMPQ DX, R14
|
||||||
|
JGE varTable
|
||||||
|
SUBQ $1, CX
|
||||||
|
SHLQ $1, DX
|
||||||
|
JMP calcShift
|
||||||
|
|
||||||
|
varTable:
|
||||||
|
// var table [maxTableSize]uint16
|
||||||
|
//
|
||||||
|
// In the asm code, unlike the Go code, we can zero-initialize only the
|
||||||
|
// first tableSize elements. Each uint16 element is 2 bytes and each MOVOU
|
||||||
|
// writes 16 bytes, so we can do only tableSize/8 writes instead of the
|
||||||
|
// 2048 writes that would zero-initialize all of table's 32768 bytes.
|
||||||
|
SHRQ $3, DX
|
||||||
|
LEAQ table-32768(SP), BX
|
||||||
|
PXOR X0, X0
|
||||||
|
|
||||||
|
memclr:
|
||||||
|
MOVOU X0, 0(BX)
|
||||||
|
ADDQ $16, BX
|
||||||
|
SUBQ $1, DX
|
||||||
|
JNZ memclr
|
||||||
|
|
||||||
|
// !!! DX = &src[0]
|
||||||
|
MOVQ SI, DX
|
||||||
|
|
||||||
|
// sLimit := len(src) - inputMargin
|
||||||
|
MOVQ R14, R9
|
||||||
|
SUBQ $15, R9
|
||||||
|
|
||||||
|
// !!! Pre-emptively spill CX, DX and R9 to the stack. Their values don't
|
||||||
|
// change for the rest of the function.
|
||||||
|
MOVQ CX, 56(SP)
|
||||||
|
MOVQ DX, 64(SP)
|
||||||
|
MOVQ R9, 88(SP)
|
||||||
|
|
||||||
|
// nextEmit := 0
|
||||||
|
MOVQ DX, R10
|
||||||
|
|
||||||
|
// s := 1
|
||||||
|
ADDQ $1, SI
|
||||||
|
|
||||||
|
// nextHash := hash(load32(src, s), shift)
|
||||||
|
MOVL 0(SI), R11
|
||||||
|
IMULL $0x1e35a7bd, R11
|
||||||
|
SHRL CX, R11
|
||||||
|
|
||||||
|
outer:
|
||||||
|
// for { etc }
|
||||||
|
|
||||||
|
// skip := 32
|
||||||
|
MOVQ $32, R12
|
||||||
|
|
||||||
|
// nextS := s
|
||||||
|
MOVQ SI, R13
|
||||||
|
|
||||||
|
// candidate := 0
|
||||||
|
MOVQ $0, R15
|
||||||
|
|
||||||
|
inner0:
|
||||||
|
// for { etc }
|
||||||
|
|
||||||
|
// s := nextS
|
||||||
|
MOVQ R13, SI
|
||||||
|
|
||||||
|
// bytesBetweenHashLookups := skip >> 5
|
||||||
|
MOVQ R12, R14
|
||||||
|
SHRQ $5, R14
|
||||||
|
|
||||||
|
// nextS = s + bytesBetweenHashLookups
|
||||||
|
ADDQ R14, R13
|
||||||
|
|
||||||
|
// skip += bytesBetweenHashLookups
|
||||||
|
ADDQ R14, R12
|
||||||
|
|
||||||
|
// if nextS > sLimit { goto emitRemainder }
|
||||||
|
MOVQ R13, AX
|
||||||
|
SUBQ DX, AX
|
||||||
|
CMPQ AX, R9
|
||||||
|
JA emitRemainder
|
||||||
|
|
||||||
|
// candidate = int(table[nextHash])
|
||||||
|
// XXX: MOVWQZX table-32768(SP)(R11*2), R15
|
||||||
|
// XXX: 4e 0f b7 7c 5c 78 movzwq 0x78(%rsp,%r11,2),%r15
|
||||||
|
BYTE $0x4e
|
||||||
|
BYTE $0x0f
|
||||||
|
BYTE $0xb7
|
||||||
|
BYTE $0x7c
|
||||||
|
BYTE $0x5c
|
||||||
|
BYTE $0x78
|
||||||
|
|
||||||
|
// table[nextHash] = uint16(s)
|
||||||
|
MOVQ SI, AX
|
||||||
|
SUBQ DX, AX
|
||||||
|
|
||||||
|
// XXX: MOVW AX, table-32768(SP)(R11*2)
|
||||||
|
// XXX: 66 42 89 44 5c 78 mov %ax,0x78(%rsp,%r11,2)
|
||||||
|
BYTE $0x66
|
||||||
|
BYTE $0x42
|
||||||
|
BYTE $0x89
|
||||||
|
BYTE $0x44
|
||||||
|
BYTE $0x5c
|
||||||
|
BYTE $0x78
|
||||||
|
|
||||||
|
// nextHash = hash(load32(src, nextS), shift)
|
||||||
|
MOVL 0(R13), R11
|
||||||
|
IMULL $0x1e35a7bd, R11
|
||||||
|
SHRL CX, R11
|
||||||
|
|
||||||
|
// if load32(src, s) != load32(src, candidate) { continue } break
|
||||||
|
MOVL 0(SI), AX
|
||||||
|
MOVL (DX)(R15*1), BX
|
||||||
|
CMPL AX, BX
|
||||||
|
JNE inner0
|
||||||
|
|
||||||
|
fourByteMatch:
|
||||||
|
// As per the encode_other.go code:
|
||||||
|
//
|
||||||
|
// A 4-byte match has been found. We'll later see etc.
|
||||||
|
|
||||||
|
// !!! Jump to a fast path for short (<= 16 byte) literals. See the comment
|
||||||
|
// on inputMargin in encode.go.
|
||||||
|
MOVQ SI, AX
|
||||||
|
SUBQ R10, AX
|
||||||
|
CMPQ AX, $16
|
||||||
|
JLE emitLiteralFastPath
|
||||||
|
|
||||||
|
// ----------------------------------------
|
||||||
|
// Begin inline of the emitLiteral call.
|
||||||
|
//
|
||||||
|
// d += emitLiteral(dst[d:], src[nextEmit:s])
|
||||||
|
|
||||||
|
MOVL AX, BX
|
||||||
|
SUBL $1, BX
|
||||||
|
|
||||||
|
CMPL BX, $60
|
||||||
|
JLT inlineEmitLiteralOneByte
|
||||||
|
CMPL BX, $256
|
||||||
|
JLT inlineEmitLiteralTwoBytes
|
||||||
|
|
||||||
|
inlineEmitLiteralThreeBytes:
|
||||||
|
MOVB $0xf4, 0(DI)
|
||||||
|
MOVW BX, 1(DI)
|
||||||
|
ADDQ $3, DI
|
||||||
|
JMP inlineEmitLiteralMemmove
|
||||||
|
|
||||||
|
inlineEmitLiteralTwoBytes:
|
||||||
|
MOVB $0xf0, 0(DI)
|
||||||
|
MOVB BX, 1(DI)
|
||||||
|
ADDQ $2, DI
|
||||||
|
JMP inlineEmitLiteralMemmove
|
||||||
|
|
||||||
|
inlineEmitLiteralOneByte:
|
||||||
|
SHLB $2, BX
|
||||||
|
MOVB BX, 0(DI)
|
||||||
|
ADDQ $1, DI
|
||||||
|
|
||||||
|
inlineEmitLiteralMemmove:
|
||||||
|
// Spill local variables (registers) onto the stack; call; unspill.
|
||||||
|
//
|
||||||
|
// copy(dst[i:], lit)
|
||||||
|
//
|
||||||
|
// This means calling runtime·memmove(&dst[i], &lit[0], len(lit)), so we push
|
||||||
|
// DI, R10 and AX as arguments.
|
||||||
|
MOVQ DI, 0(SP)
|
||||||
|
MOVQ R10, 8(SP)
|
||||||
|
MOVQ AX, 16(SP)
|
||||||
|
ADDQ AX, DI // Finish the "d +=" part of "d += emitLiteral(etc)".
|
||||||
|
MOVQ SI, 72(SP)
|
||||||
|
MOVQ DI, 80(SP)
|
||||||
|
MOVQ R15, 112(SP)
|
||||||
|
CALL runtime·memmove(SB)
|
||||||
|
MOVQ 56(SP), CX
|
||||||
|
MOVQ 64(SP), DX
|
||||||
|
MOVQ 72(SP), SI
|
||||||
|
MOVQ 80(SP), DI
|
||||||
|
MOVQ 88(SP), R9
|
||||||
|
MOVQ 112(SP), R15
|
||||||
|
JMP inner1
|
||||||
|
|
||||||
|
inlineEmitLiteralEnd:
|
||||||
|
// End inline of the emitLiteral call.
|
||||||
|
// ----------------------------------------
|
||||||
|
|
||||||
|
emitLiteralFastPath:
|
||||||
|
// !!! Emit the 1-byte encoding "uint8(len(lit)-1)<<2".
|
||||||
|
MOVB AX, BX
|
||||||
|
SUBB $1, BX
|
||||||
|
SHLB $2, BX
|
||||||
|
MOVB BX, (DI)
|
||||||
|
ADDQ $1, DI
|
||||||
|
|
||||||
|
// !!! Implement the copy from lit to dst as a 16-byte load and store.
|
||||||
|
// (Encode's documentation says that dst and src must not overlap.)
|
||||||
|
//
|
||||||
|
// This always copies 16 bytes, instead of only len(lit) bytes, but that's
|
||||||
|
// OK. Subsequent iterations will fix up the overrun.
|
||||||
|
//
|
||||||
|
// Note that on amd64, it is legal and cheap to issue unaligned 8-byte or
|
||||||
|
// 16-byte loads and stores. This technique probably wouldn't be as
|
||||||
|
// effective on architectures that are fussier about alignment.
|
||||||
|
MOVOU 0(R10), X0
|
||||||
|
MOVOU X0, 0(DI)
|
||||||
|
ADDQ AX, DI
|
||||||
|
|
||||||
|
inner1:
|
||||||
|
// for { etc }
|
||||||
|
|
||||||
|
// base := s
|
||||||
|
MOVQ SI, R12
|
||||||
|
|
||||||
|
// !!! offset := base - candidate
|
||||||
|
MOVQ R12, R11
|
||||||
|
SUBQ R15, R11
|
||||||
|
SUBQ DX, R11
|
||||||
|
|
||||||
|
// ----------------------------------------
|
||||||
|
// Begin inline of the extendMatch call.
|
||||||
|
//
|
||||||
|
// s = extendMatch(src, candidate+4, s+4)
|
||||||
|
|
||||||
|
// !!! R14 = &src[len(src)]
|
||||||
|
MOVQ src_len+32(FP), R14
|
||||||
|
ADDQ DX, R14
|
||||||
|
|
||||||
|
// !!! R13 = &src[len(src) - 8]
|
||||||
|
MOVQ R14, R13
|
||||||
|
SUBQ $8, R13
|
||||||
|
|
||||||
|
// !!! R15 = &src[candidate + 4]
|
||||||
|
ADDQ $4, R15
|
||||||
|
ADDQ DX, R15
|
||||||
|
|
||||||
|
// !!! s += 4
|
||||||
|
ADDQ $4, SI
|
||||||
|
|
||||||
|
inlineExtendMatchCmp8:
|
||||||
|
// As long as we are 8 or more bytes before the end of src, we can load and
|
||||||
|
// compare 8 bytes at a time. If those 8 bytes are equal, repeat.
|
||||||
|
CMPQ SI, R13
|
||||||
|
JA inlineExtendMatchCmp1
|
||||||
|
MOVQ (R15), AX
|
||||||
|
MOVQ (SI), BX
|
||||||
|
CMPQ AX, BX
|
||||||
|
JNE inlineExtendMatchBSF
|
||||||
|
ADDQ $8, R15
|
||||||
|
ADDQ $8, SI
|
||||||
|
JMP inlineExtendMatchCmp8
|
||||||
|
|
||||||
|
inlineExtendMatchBSF:
|
||||||
|
// If those 8 bytes were not equal, XOR the two 8 byte values, and return
|
||||||
|
// the index of the first byte that differs. The BSF instruction finds the
|
||||||
|
// least significant 1 bit, the amd64 architecture is little-endian, and
|
||||||
|
// the shift by 3 converts a bit index to a byte index.
|
||||||
|
XORQ AX, BX
|
||||||
|
BSFQ BX, BX
|
||||||
|
SHRQ $3, BX
|
||||||
|
ADDQ BX, SI
|
||||||
|
JMP inlineExtendMatchEnd
|
||||||
|
|
||||||
|
inlineExtendMatchCmp1:
|
||||||
|
// In src's tail, compare 1 byte at a time.
|
||||||
|
CMPQ SI, R14
|
||||||
|
JAE inlineExtendMatchEnd
|
||||||
|
MOVB (R15), AX
|
||||||
|
MOVB (SI), BX
|
||||||
|
CMPB AX, BX
|
||||||
|
JNE inlineExtendMatchEnd
|
||||||
|
ADDQ $1, R15
|
||||||
|
ADDQ $1, SI
|
||||||
|
JMP inlineExtendMatchCmp1
|
||||||
|
|
||||||
|
inlineExtendMatchEnd:
|
||||||
|
// End inline of the extendMatch call.
|
||||||
|
// ----------------------------------------
|
||||||
|
|
||||||
|
// ----------------------------------------
|
||||||
|
// Begin inline of the emitCopy call.
|
||||||
|
//
|
||||||
|
// d += emitCopy(dst[d:], base-candidate, s-base)
|
||||||
|
|
||||||
|
// !!! length := s - base
|
||||||
|
MOVQ SI, AX
|
||||||
|
SUBQ R12, AX
|
||||||
|
|
||||||
|
inlineEmitCopyLoop0:
|
||||||
|
// for length >= 68 { etc }
|
||||||
|
CMPL AX, $68
|
||||||
|
JLT inlineEmitCopyStep1
|
||||||
|
|
||||||
|
// Emit a length 64 copy, encoded as 3 bytes.
|
||||||
|
MOVB $0xfe, 0(DI)
|
||||||
|
MOVW R11, 1(DI)
|
||||||
|
ADDQ $3, DI
|
||||||
|
SUBL $64, AX
|
||||||
|
JMP inlineEmitCopyLoop0
|
||||||
|
|
||||||
|
inlineEmitCopyStep1:
|
||||||
|
// if length > 64 { etc }
|
||||||
|
CMPL AX, $64
|
||||||
|
JLE inlineEmitCopyStep2
|
||||||
|
|
||||||
|
// Emit a length 60 copy, encoded as 3 bytes.
|
||||||
|
MOVB $0xee, 0(DI)
|
||||||
|
MOVW R11, 1(DI)
|
||||||
|
ADDQ $3, DI
|
||||||
|
SUBL $60, AX
|
||||||
|
|
||||||
|
inlineEmitCopyStep2:
|
||||||
|
// if length >= 12 || offset >= 2048 { goto inlineEmitCopyStep3 }
|
||||||
|
CMPL AX, $12
|
||||||
|
JGE inlineEmitCopyStep3
|
||||||
|
CMPL R11, $2048
|
||||||
|
JGE inlineEmitCopyStep3
|
||||||
|
|
||||||
|
// Emit the remaining copy, encoded as 2 bytes.
|
||||||
|
MOVB R11, 1(DI)
|
||||||
|
SHRL $8, R11
|
||||||
|
SHLB $5, R11
|
||||||
|
SUBB $4, AX
|
||||||
|
SHLB $2, AX
|
||||||
|
ORB AX, R11
|
||||||
|
ORB $1, R11
|
||||||
|
MOVB R11, 0(DI)
|
||||||
|
ADDQ $2, DI
|
||||||
|
JMP inlineEmitCopyEnd
|
||||||
|
|
||||||
|
inlineEmitCopyStep3:
|
||||||
|
// Emit the remaining copy, encoded as 3 bytes.
|
||||||
|
SUBL $1, AX
|
||||||
|
SHLB $2, AX
|
||||||
|
ORB $2, AX
|
||||||
|
MOVB AX, 0(DI)
|
||||||
|
MOVW R11, 1(DI)
|
||||||
|
ADDQ $3, DI
|
||||||
|
|
||||||
|
inlineEmitCopyEnd:
|
||||||
|
// End inline of the emitCopy call.
|
||||||
|
// ----------------------------------------
|
||||||
|
|
||||||
|
// nextEmit = s
|
||||||
|
MOVQ SI, R10
|
||||||
|
|
||||||
|
// if s >= sLimit { goto emitRemainder }
|
||||||
|
MOVQ SI, AX
|
||||||
|
SUBQ DX, AX
|
||||||
|
CMPQ AX, R9
|
||||||
|
JAE emitRemainder
|
||||||
|
|
||||||
|
// As per the encode_other.go code:
|
||||||
|
//
|
||||||
|
// We could immediately etc.
|
||||||
|
|
||||||
|
// x := load64(src, s-1)
|
||||||
|
MOVQ -1(SI), R14
|
||||||
|
|
||||||
|
// prevHash := hash(uint32(x>>0), shift)
|
||||||
|
MOVL R14, R11
|
||||||
|
IMULL $0x1e35a7bd, R11
|
||||||
|
SHRL CX, R11
|
||||||
|
|
||||||
|
// table[prevHash] = uint16(s-1)
|
||||||
|
MOVQ SI, AX
|
||||||
|
SUBQ DX, AX
|
||||||
|
SUBQ $1, AX
|
||||||
|
|
||||||
|
// XXX: MOVW AX, table-32768(SP)(R11*2)
|
||||||
|
// XXX: 66 42 89 44 5c 78 mov %ax,0x78(%rsp,%r11,2)
|
||||||
|
BYTE $0x66
|
||||||
|
BYTE $0x42
|
||||||
|
BYTE $0x89
|
||||||
|
BYTE $0x44
|
||||||
|
BYTE $0x5c
|
||||||
|
BYTE $0x78
|
||||||
|
|
||||||
|
// currHash := hash(uint32(x>>8), shift)
|
||||||
|
SHRQ $8, R14
|
||||||
|
MOVL R14, R11
|
||||||
|
IMULL $0x1e35a7bd, R11
|
||||||
|
SHRL CX, R11
|
||||||
|
|
||||||
|
// candidate = int(table[currHash])
|
||||||
|
// XXX: MOVWQZX table-32768(SP)(R11*2), R15
|
||||||
|
// XXX: 4e 0f b7 7c 5c 78 movzwq 0x78(%rsp,%r11,2),%r15
|
||||||
|
BYTE $0x4e
|
||||||
|
BYTE $0x0f
|
||||||
|
BYTE $0xb7
|
||||||
|
BYTE $0x7c
|
||||||
|
BYTE $0x5c
|
||||||
|
BYTE $0x78
|
||||||
|
|
||||||
|
// table[currHash] = uint16(s)
|
||||||
|
ADDQ $1, AX
|
||||||
|
|
||||||
|
// XXX: MOVW AX, table-32768(SP)(R11*2)
|
||||||
|
// XXX: 66 42 89 44 5c 78 mov %ax,0x78(%rsp,%r11,2)
|
||||||
|
BYTE $0x66
|
||||||
|
BYTE $0x42
|
||||||
|
BYTE $0x89
|
||||||
|
BYTE $0x44
|
||||||
|
BYTE $0x5c
|
||||||
|
BYTE $0x78
|
||||||
|
|
||||||
|
// if uint32(x>>8) == load32(src, candidate) { continue }
|
||||||
|
MOVL (DX)(R15*1), BX
|
||||||
|
CMPL R14, BX
|
||||||
|
JEQ inner1
|
||||||
|
|
||||||
|
// nextHash = hash(uint32(x>>16), shift)
|
||||||
|
SHRQ $8, R14
|
||||||
|
MOVL R14, R11
|
||||||
|
IMULL $0x1e35a7bd, R11
|
||||||
|
SHRL CX, R11
|
||||||
|
|
||||||
|
// s++
|
||||||
|
ADDQ $1, SI
|
||||||
|
|
||||||
|
// break out of the inner1 for loop, i.e. continue the outer loop.
|
||||||
|
JMP outer
|
||||||
|
|
||||||
|
emitRemainder:
|
||||||
|
// if nextEmit < len(src) { etc }
|
||||||
|
MOVQ src_len+32(FP), AX
|
||||||
|
ADDQ DX, AX
|
||||||
|
CMPQ R10, AX
|
||||||
|
JEQ encodeBlockEnd
|
||||||
|
|
||||||
|
// d += emitLiteral(dst[d:], src[nextEmit:])
|
||||||
|
//
|
||||||
|
// Push args.
|
||||||
|
MOVQ DI, 0(SP)
|
||||||
|
MOVQ $0, 8(SP) // Unnecessary, as the callee ignores it, but conservative.
|
||||||
|
MOVQ $0, 16(SP) // Unnecessary, as the callee ignores it, but conservative.
|
||||||
|
MOVQ R10, 24(SP)
|
||||||
|
SUBQ R10, AX
|
||||||
|
MOVQ AX, 32(SP)
|
||||||
|
MOVQ AX, 40(SP) // Unnecessary, as the callee ignores it, but conservative.
|
||||||
|
|
||||||
|
// Spill local variables (registers) onto the stack; call; unspill.
|
||||||
|
MOVQ DI, 80(SP)
|
||||||
|
CALL ·emitLiteral(SB)
|
||||||
|
MOVQ 80(SP), DI
|
||||||
|
|
||||||
|
// Finish the "d +=" part of "d += emitLiteral(etc)".
|
||||||
|
ADDQ 48(SP), DI
|
||||||
|
|
||||||
|
encodeBlockEnd:
|
||||||
|
MOVQ dst_base+0(FP), AX
|
||||||
|
SUBQ AX, DI
|
||||||
|
MOVQ DI, d+48(FP)
|
||||||
|
RET
|
|
@ -0,0 +1,238 @@
|
||||||
|
// Copyright 2016 The Snappy-Go Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
// +build !amd64 appengine !gc noasm
|
||||||
|
|
||||||
|
package snappy
|
||||||
|
|
||||||
|
func load32(b []byte, i int) uint32 {
|
||||||
|
b = b[i : i+4 : len(b)] // Help the compiler eliminate bounds checks on the next line.
|
||||||
|
return uint32(b[0]) | uint32(b[1])<<8 | uint32(b[2])<<16 | uint32(b[3])<<24
|
||||||
|
}
|
||||||
|
|
||||||
|
func load64(b []byte, i int) uint64 {
|
||||||
|
b = b[i : i+8 : len(b)] // Help the compiler eliminate bounds checks on the next line.
|
||||||
|
return uint64(b[0]) | uint64(b[1])<<8 | uint64(b[2])<<16 | uint64(b[3])<<24 |
|
||||||
|
uint64(b[4])<<32 | uint64(b[5])<<40 | uint64(b[6])<<48 | uint64(b[7])<<56
|
||||||
|
}
|
||||||
|
|
||||||
|
// emitLiteral writes a literal chunk and returns the number of bytes written.
|
||||||
|
//
|
||||||
|
// It assumes that:
|
||||||
|
// dst is long enough to hold the encoded bytes
|
||||||
|
// 1 <= len(lit) && len(lit) <= 65536
|
||||||
|
func emitLiteral(dst, lit []byte) int {
|
||||||
|
i, n := 0, uint(len(lit)-1)
|
||||||
|
switch {
|
||||||
|
case n < 60:
|
||||||
|
dst[0] = uint8(n)<<2 | tagLiteral
|
||||||
|
i = 1
|
||||||
|
case n < 1<<8:
|
||||||
|
dst[0] = 60<<2 | tagLiteral
|
||||||
|
dst[1] = uint8(n)
|
||||||
|
i = 2
|
||||||
|
default:
|
||||||
|
dst[0] = 61<<2 | tagLiteral
|
||||||
|
dst[1] = uint8(n)
|
||||||
|
dst[2] = uint8(n >> 8)
|
||||||
|
i = 3
|
||||||
|
}
|
||||||
|
return i + copy(dst[i:], lit)
|
||||||
|
}
|
||||||
|
|
||||||
|
// emitCopy writes a copy chunk and returns the number of bytes written.
|
||||||
|
//
|
||||||
|
// It assumes that:
|
||||||
|
// dst is long enough to hold the encoded bytes
|
||||||
|
// 1 <= offset && offset <= 65535
|
||||||
|
// 4 <= length && length <= 65535
|
||||||
|
func emitCopy(dst []byte, offset, length int) int {
|
||||||
|
i := 0
|
||||||
|
// The maximum length for a single tagCopy1 or tagCopy2 op is 64 bytes. The
|
||||||
|
// threshold for this loop is a little higher (at 68 = 64 + 4), and the
|
||||||
|
// length emitted down below is is a little lower (at 60 = 64 - 4), because
|
||||||
|
// it's shorter to encode a length 67 copy as a length 60 tagCopy2 followed
|
||||||
|
// by a length 7 tagCopy1 (which encodes as 3+2 bytes) than to encode it as
|
||||||
|
// a length 64 tagCopy2 followed by a length 3 tagCopy2 (which encodes as
|
||||||
|
// 3+3 bytes). The magic 4 in the 64±4 is because the minimum length for a
|
||||||
|
// tagCopy1 op is 4 bytes, which is why a length 3 copy has to be an
|
||||||
|
// encodes-as-3-bytes tagCopy2 instead of an encodes-as-2-bytes tagCopy1.
|
||||||
|
for length >= 68 {
|
||||||
|
// Emit a length 64 copy, encoded as 3 bytes.
|
||||||
|
dst[i+0] = 63<<2 | tagCopy2
|
||||||
|
dst[i+1] = uint8(offset)
|
||||||
|
dst[i+2] = uint8(offset >> 8)
|
||||||
|
i += 3
|
||||||
|
length -= 64
|
||||||
|
}
|
||||||
|
if length > 64 {
|
||||||
|
// Emit a length 60 copy, encoded as 3 bytes.
|
||||||
|
dst[i+0] = 59<<2 | tagCopy2
|
||||||
|
dst[i+1] = uint8(offset)
|
||||||
|
dst[i+2] = uint8(offset >> 8)
|
||||||
|
i += 3
|
||||||
|
length -= 60
|
||||||
|
}
|
||||||
|
if length >= 12 || offset >= 2048 {
|
||||||
|
// Emit the remaining copy, encoded as 3 bytes.
|
||||||
|
dst[i+0] = uint8(length-1)<<2 | tagCopy2
|
||||||
|
dst[i+1] = uint8(offset)
|
||||||
|
dst[i+2] = uint8(offset >> 8)
|
||||||
|
return i + 3
|
||||||
|
}
|
||||||
|
// Emit the remaining copy, encoded as 2 bytes.
|
||||||
|
dst[i+0] = uint8(offset>>8)<<5 | uint8(length-4)<<2 | tagCopy1
|
||||||
|
dst[i+1] = uint8(offset)
|
||||||
|
return i + 2
|
||||||
|
}
|
||||||
|
|
||||||
|
// extendMatch returns the largest k such that k <= len(src) and that
|
||||||
|
// src[i:i+k-j] and src[j:k] have the same contents.
|
||||||
|
//
|
||||||
|
// It assumes that:
|
||||||
|
// 0 <= i && i < j && j <= len(src)
|
||||||
|
func extendMatch(src []byte, i, j int) int {
|
||||||
|
for ; j < len(src) && src[i] == src[j]; i, j = i+1, j+1 {
|
||||||
|
}
|
||||||
|
return j
|
||||||
|
}
|
||||||
|
|
||||||
|
func hash(u, shift uint32) uint32 {
|
||||||
|
return (u * 0x1e35a7bd) >> shift
|
||||||
|
}
|
||||||
|
|
||||||
|
// encodeBlock encodes a non-empty src to a guaranteed-large-enough dst. It
|
||||||
|
// assumes that the varint-encoded length of the decompressed bytes has already
|
||||||
|
// been written.
|
||||||
|
//
|
||||||
|
// It also assumes that:
|
||||||
|
// len(dst) >= MaxEncodedLen(len(src)) &&
|
||||||
|
// minNonLiteralBlockSize <= len(src) && len(src) <= maxBlockSize
|
||||||
|
func encodeBlock(dst, src []byte) (d int) {
|
||||||
|
// Initialize the hash table. Its size ranges from 1<<8 to 1<<14 inclusive.
|
||||||
|
// The table element type is uint16, as s < sLimit and sLimit < len(src)
|
||||||
|
// and len(src) <= maxBlockSize and maxBlockSize == 65536.
|
||||||
|
const (
|
||||||
|
maxTableSize = 1 << 14
|
||||||
|
// tableMask is redundant, but helps the compiler eliminate bounds
|
||||||
|
// checks.
|
||||||
|
tableMask = maxTableSize - 1
|
||||||
|
)
|
||||||
|
shift := uint32(32 - 8)
|
||||||
|
for tableSize := 1 << 8; tableSize < maxTableSize && tableSize < len(src); tableSize *= 2 {
|
||||||
|
shift--
|
||||||
|
}
|
||||||
|
// In Go, all array elements are zero-initialized, so there is no advantage
|
||||||
|
// to a smaller tableSize per se. However, it matches the C++ algorithm,
|
||||||
|
// and in the asm versions of this code, we can get away with zeroing only
|
||||||
|
// the first tableSize elements.
|
||||||
|
var table [maxTableSize]uint16
|
||||||
|
|
||||||
|
// sLimit is when to stop looking for offset/length copies. The inputMargin
|
||||||
|
// lets us use a fast path for emitLiteral in the main loop, while we are
|
||||||
|
// looking for copies.
|
||||||
|
sLimit := len(src) - inputMargin
|
||||||
|
|
||||||
|
// nextEmit is where in src the next emitLiteral should start from.
|
||||||
|
nextEmit := 0
|
||||||
|
|
||||||
|
// The encoded form must start with a literal, as there are no previous
|
||||||
|
// bytes to copy, so we start looking for hash matches at s == 1.
|
||||||
|
s := 1
|
||||||
|
nextHash := hash(load32(src, s), shift)
|
||||||
|
|
||||||
|
for {
|
||||||
|
// Copied from the C++ snappy implementation:
|
||||||
|
//
|
||||||
|
// Heuristic match skipping: If 32 bytes are scanned with no matches
|
||||||
|
// found, start looking only at every other byte. If 32 more bytes are
|
||||||
|
// scanned (or skipped), look at every third byte, etc.. When a match
|
||||||
|
// is found, immediately go back to looking at every byte. This is a
|
||||||
|
// small loss (~5% performance, ~0.1% density) for compressible data
|
||||||
|
// due to more bookkeeping, but for non-compressible data (such as
|
||||||
|
// JPEG) it's a huge win since the compressor quickly "realizes" the
|
||||||
|
// data is incompressible and doesn't bother looking for matches
|
||||||
|
// everywhere.
|
||||||
|
//
|
||||||
|
// The "skip" variable keeps track of how many bytes there are since
|
||||||
|
// the last match; dividing it by 32 (ie. right-shifting by five) gives
|
||||||
|
// the number of bytes to move ahead for each iteration.
|
||||||
|
skip := 32
|
||||||
|
|
||||||
|
nextS := s
|
||||||
|
candidate := 0
|
||||||
|
for {
|
||||||
|
s = nextS
|
||||||
|
bytesBetweenHashLookups := skip >> 5
|
||||||
|
nextS = s + bytesBetweenHashLookups
|
||||||
|
skip += bytesBetweenHashLookups
|
||||||
|
if nextS > sLimit {
|
||||||
|
goto emitRemainder
|
||||||
|
}
|
||||||
|
candidate = int(table[nextHash&tableMask])
|
||||||
|
table[nextHash&tableMask] = uint16(s)
|
||||||
|
nextHash = hash(load32(src, nextS), shift)
|
||||||
|
if load32(src, s) == load32(src, candidate) {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// A 4-byte match has been found. We'll later see if more than 4 bytes
|
||||||
|
// match. But, prior to the match, src[nextEmit:s] are unmatched. Emit
|
||||||
|
// them as literal bytes.
|
||||||
|
d += emitLiteral(dst[d:], src[nextEmit:s])
|
||||||
|
|
||||||
|
// Call emitCopy, and then see if another emitCopy could be our next
|
||||||
|
// move. Repeat until we find no match for the input immediately after
|
||||||
|
// what was consumed by the last emitCopy call.
|
||||||
|
//
|
||||||
|
// If we exit this loop normally then we need to call emitLiteral next,
|
||||||
|
// though we don't yet know how big the literal will be. We handle that
|
||||||
|
// by proceeding to the next iteration of the main loop. We also can
|
||||||
|
// exit this loop via goto if we get close to exhausting the input.
|
||||||
|
for {
|
||||||
|
// Invariant: we have a 4-byte match at s, and no need to emit any
|
||||||
|
// literal bytes prior to s.
|
||||||
|
base := s
|
||||||
|
|
||||||
|
// Extend the 4-byte match as long as possible.
|
||||||
|
//
|
||||||
|
// This is an inlined version of:
|
||||||
|
// s = extendMatch(src, candidate+4, s+4)
|
||||||
|
s += 4
|
||||||
|
for i := candidate + 4; s < len(src) && src[i] == src[s]; i, s = i+1, s+1 {
|
||||||
|
}
|
||||||
|
|
||||||
|
d += emitCopy(dst[d:], base-candidate, s-base)
|
||||||
|
nextEmit = s
|
||||||
|
if s >= sLimit {
|
||||||
|
goto emitRemainder
|
||||||
|
}
|
||||||
|
|
||||||
|
// We could immediately start working at s now, but to improve
|
||||||
|
// compression we first update the hash table at s-1 and at s. If
|
||||||
|
// another emitCopy is not our next move, also calculate nextHash
|
||||||
|
// at s+1. At least on GOARCH=amd64, these three hash calculations
|
||||||
|
// are faster as one load64 call (with some shifts) instead of
|
||||||
|
// three load32 calls.
|
||||||
|
x := load64(src, s-1)
|
||||||
|
prevHash := hash(uint32(x>>0), shift)
|
||||||
|
table[prevHash&tableMask] = uint16(s - 1)
|
||||||
|
currHash := hash(uint32(x>>8), shift)
|
||||||
|
candidate = int(table[currHash&tableMask])
|
||||||
|
table[currHash&tableMask] = uint16(s)
|
||||||
|
if uint32(x>>8) != load32(src, candidate) {
|
||||||
|
nextHash = hash(uint32(x>>16), shift)
|
||||||
|
s++
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
emitRemainder:
|
||||||
|
if nextEmit < len(src) {
|
||||||
|
d += emitLiteral(dst[d:], src[nextEmit:])
|
||||||
|
}
|
||||||
|
return d
|
||||||
|
}
|
|
@ -0,0 +1,98 @@
|
||||||
|
// Copyright 2011 The Snappy-Go Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
// Package snappy implements the Snappy compression format. It aims for very
|
||||||
|
// high speeds and reasonable compression.
|
||||||
|
//
|
||||||
|
// There are actually two Snappy formats: block and stream. They are related,
|
||||||
|
// but different: trying to decompress block-compressed data as a Snappy stream
|
||||||
|
// will fail, and vice versa. The block format is the Decode and Encode
|
||||||
|
// functions and the stream format is the Reader and Writer types.
|
||||||
|
//
|
||||||
|
// The block format, the more common case, is used when the complete size (the
|
||||||
|
// number of bytes) of the original data is known upfront, at the time
|
||||||
|
// compression starts. The stream format, also known as the framing format, is
|
||||||
|
// for when that isn't always true.
|
||||||
|
//
|
||||||
|
// The canonical, C++ implementation is at https://github.com/google/snappy and
|
||||||
|
// it only implements the block format.
|
||||||
|
package snappy // import "github.com/golang/snappy"
|
||||||
|
|
||||||
|
import (
|
||||||
|
"hash/crc32"
|
||||||
|
)
|
||||||
|
|
||||||
|
/*
|
||||||
|
Each encoded block begins with the varint-encoded length of the decoded data,
|
||||||
|
followed by a sequence of chunks. Chunks begin and end on byte boundaries. The
|
||||||
|
first byte of each chunk is broken into its 2 least and 6 most significant bits
|
||||||
|
called l and m: l ranges in [0, 4) and m ranges in [0, 64). l is the chunk tag.
|
||||||
|
Zero means a literal tag. All other values mean a copy tag.
|
||||||
|
|
||||||
|
For literal tags:
|
||||||
|
- If m < 60, the next 1 + m bytes are literal bytes.
|
||||||
|
- Otherwise, let n be the little-endian unsigned integer denoted by the next
|
||||||
|
m - 59 bytes. The next 1 + n bytes after that are literal bytes.
|
||||||
|
|
||||||
|
For copy tags, length bytes are copied from offset bytes ago, in the style of
|
||||||
|
Lempel-Ziv compression algorithms. In particular:
|
||||||
|
- For l == 1, the offset ranges in [0, 1<<11) and the length in [4, 12).
|
||||||
|
The length is 4 + the low 3 bits of m. The high 3 bits of m form bits 8-10
|
||||||
|
of the offset. The next byte is bits 0-7 of the offset.
|
||||||
|
- For l == 2, the offset ranges in [0, 1<<16) and the length in [1, 65).
|
||||||
|
The length is 1 + m. The offset is the little-endian unsigned integer
|
||||||
|
denoted by the next 2 bytes.
|
||||||
|
- For l == 3, this tag is a legacy format that is no longer issued by most
|
||||||
|
encoders. Nonetheless, the offset ranges in [0, 1<<32) and the length in
|
||||||
|
[1, 65). The length is 1 + m. The offset is the little-endian unsigned
|
||||||
|
integer denoted by the next 4 bytes.
|
||||||
|
*/
|
||||||
|
const (
|
||||||
|
tagLiteral = 0x00
|
||||||
|
tagCopy1 = 0x01
|
||||||
|
tagCopy2 = 0x02
|
||||||
|
tagCopy4 = 0x03
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
checksumSize = 4
|
||||||
|
chunkHeaderSize = 4
|
||||||
|
magicChunk = "\xff\x06\x00\x00" + magicBody
|
||||||
|
magicBody = "sNaPpY"
|
||||||
|
|
||||||
|
// maxBlockSize is the maximum size of the input to encodeBlock. It is not
|
||||||
|
// part of the wire format per se, but some parts of the encoder assume
|
||||||
|
// that an offset fits into a uint16.
|
||||||
|
//
|
||||||
|
// Also, for the framing format (Writer type instead of Encode function),
|
||||||
|
// https://github.com/google/snappy/blob/master/framing_format.txt says
|
||||||
|
// that "the uncompressed data in a chunk must be no longer than 65536
|
||||||
|
// bytes".
|
||||||
|
maxBlockSize = 65536
|
||||||
|
|
||||||
|
// maxEncodedLenOfMaxBlockSize equals MaxEncodedLen(maxBlockSize), but is
|
||||||
|
// hard coded to be a const instead of a variable, so that obufLen can also
|
||||||
|
// be a const. Their equivalence is confirmed by
|
||||||
|
// TestMaxEncodedLenOfMaxBlockSize.
|
||||||
|
maxEncodedLenOfMaxBlockSize = 76490
|
||||||
|
|
||||||
|
obufHeaderLen = len(magicChunk) + checksumSize + chunkHeaderSize
|
||||||
|
obufLen = obufHeaderLen + maxEncodedLenOfMaxBlockSize
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
chunkTypeCompressedData = 0x00
|
||||||
|
chunkTypeUncompressedData = 0x01
|
||||||
|
chunkTypePadding = 0xfe
|
||||||
|
chunkTypeStreamIdentifier = 0xff
|
||||||
|
)
|
||||||
|
|
||||||
|
var crcTable = crc32.MakeTable(crc32.Castagnoli)
|
||||||
|
|
||||||
|
// crc implements the checksum specified in section 3 of
|
||||||
|
// https://github.com/google/snappy/blob/master/framing_format.txt
|
||||||
|
func crc(b []byte) uint32 {
|
||||||
|
c := crc32.Update(0, crcTable, b)
|
||||||
|
return uint32(c>>15|c<<17) + 0xa282ead8
|
||||||
|
}
|
|
@ -14,13 +14,16 @@ makes `retryablehttp` very easy to drop into existing programs.
|
||||||
|
|
||||||
`retryablehttp` performs automatic retries under certain conditions. Mainly, if
|
`retryablehttp` performs automatic retries under certain conditions. Mainly, if
|
||||||
an error is returned by the client (connection errors, etc.), or if a 500-range
|
an error is returned by the client (connection errors, etc.), or if a 500-range
|
||||||
response code is received, then a retry is invoked after a wait period.
|
response code is received (except 501), then a retry is invoked after a wait
|
||||||
Otherwise, the response is returned and left to the caller to interpret.
|
period. Otherwise, the response is returned and left to the caller to
|
||||||
|
interpret.
|
||||||
|
|
||||||
The main difference from `net/http` is that requests which take a request body
|
The main difference from `net/http` is that requests which take a request body
|
||||||
(POST/PUT et. al) require an `io.ReadSeeker` to be provided. This enables the
|
(POST/PUT et. al) can have the body provided in a number of ways (some more or
|
||||||
request body to be "rewound" if the initial request fails so that the full
|
less efficient) that allow "rewinding" the request body if the initial request
|
||||||
request can be attempted again.
|
fails so that the full request can be attempted again. See the
|
||||||
|
[godoc](http://godoc.org/github.com/hashicorp/go-retryablehttp) for more
|
||||||
|
details.
|
||||||
|
|
||||||
Example Use
|
Example Use
|
||||||
===========
|
===========
|
||||||
|
|
|
@ -8,18 +8,28 @@
|
||||||
// response is received, then a retry is invoked. Otherwise, the response is
|
// response is received, then a retry is invoked. Otherwise, the response is
|
||||||
// returned and left to the caller to interpret.
|
// returned and left to the caller to interpret.
|
||||||
//
|
//
|
||||||
// The main difference from net/http is that requests which take a request body
|
// Requests which take a request body should provide a non-nil function
|
||||||
// (POST/PUT et. al) require an io.ReadSeeker to be provided. This enables the
|
// parameter. The best choice is to provide either a function satisfying
|
||||||
// request body to be "rewound" if the initial request fails so that the full
|
// ReaderFunc which provides multiple io.Readers in an efficient manner, a
|
||||||
// request can be attempted again.
|
// *bytes.Buffer (the underlying raw byte slice will be used) or a raw byte
|
||||||
|
// slice. As it is a reference type, and we will wrap it as needed by readers,
|
||||||
|
// we can efficiently re-use the request body without needing to copy it. If an
|
||||||
|
// io.Reader (such as a *bytes.Reader) is provided, the full body will be read
|
||||||
|
// prior to the first request, and will be efficiently re-used for any retries.
|
||||||
|
// ReadSeeker can be used, but some users have observed occasional data races
|
||||||
|
// between the net/http library and the Seek functionality of some
|
||||||
|
// implementations of ReadSeeker, so should be avoided if possible.
|
||||||
package retryablehttp
|
package retryablehttp
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"bytes"
|
||||||
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"log"
|
"log"
|
||||||
"math"
|
"math"
|
||||||
|
"math/rand"
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/url"
|
"net/url"
|
||||||
"os"
|
"os"
|
||||||
|
@ -44,6 +54,9 @@ var (
|
||||||
respReadLimit = int64(4096)
|
respReadLimit = int64(4096)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// ReaderFunc is the type of function that can be given natively to NewRequest
|
||||||
|
type ReaderFunc func() (io.Reader, error)
|
||||||
|
|
||||||
// LenReader is an interface implemented by many in-memory io.Reader's. Used
|
// LenReader is an interface implemented by many in-memory io.Reader's. Used
|
||||||
// for automatically sending the right Content-Length header when possible.
|
// for automatically sending the right Content-Length header when possible.
|
||||||
type LenReader interface {
|
type LenReader interface {
|
||||||
|
@ -54,32 +67,118 @@ type LenReader interface {
|
||||||
type Request struct {
|
type Request struct {
|
||||||
// body is a seekable reader over the request body payload. This is
|
// body is a seekable reader over the request body payload. This is
|
||||||
// used to rewind the request data in between retries.
|
// used to rewind the request data in between retries.
|
||||||
body io.ReadSeeker
|
body ReaderFunc
|
||||||
|
|
||||||
// Embed an HTTP request directly. This makes a *Request act exactly
|
// Embed an HTTP request directly. This makes a *Request act exactly
|
||||||
// like an *http.Request so that all meta methods are supported.
|
// like an *http.Request so that all meta methods are supported.
|
||||||
*http.Request
|
*http.Request
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// WithContext returns wrapped Request with a shallow copy of underlying *http.Request
|
||||||
|
// with its context changed to ctx. The provided ctx must be non-nil.
|
||||||
|
func (r *Request) WithContext(ctx context.Context) *Request {
|
||||||
|
r.Request = r.Request.WithContext(ctx)
|
||||||
|
return r
|
||||||
|
}
|
||||||
|
|
||||||
// NewRequest creates a new wrapped request.
|
// NewRequest creates a new wrapped request.
|
||||||
func NewRequest(method, url string, body io.ReadSeeker) (*Request, error) {
|
func NewRequest(method, url string, rawBody interface{}) (*Request, error) {
|
||||||
// Wrap the body in a noop ReadCloser if non-nil. This prevents the
|
var err error
|
||||||
// reader from being closed by the HTTP client.
|
var body ReaderFunc
|
||||||
var rcBody io.ReadCloser
|
var contentLength int64
|
||||||
if body != nil {
|
|
||||||
rcBody = ioutil.NopCloser(body)
|
if rawBody != nil {
|
||||||
|
switch rawBody.(type) {
|
||||||
|
// If they gave us a function already, great! Use it.
|
||||||
|
case ReaderFunc:
|
||||||
|
body = rawBody.(ReaderFunc)
|
||||||
|
tmp, err := body()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if lr, ok := tmp.(LenReader); ok {
|
||||||
|
contentLength = int64(lr.Len())
|
||||||
|
}
|
||||||
|
if c, ok := tmp.(io.Closer); ok {
|
||||||
|
c.Close()
|
||||||
|
}
|
||||||
|
|
||||||
|
case func() (io.Reader, error):
|
||||||
|
body = rawBody.(func() (io.Reader, error))
|
||||||
|
tmp, err := body()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if lr, ok := tmp.(LenReader); ok {
|
||||||
|
contentLength = int64(lr.Len())
|
||||||
|
}
|
||||||
|
if c, ok := tmp.(io.Closer); ok {
|
||||||
|
c.Close()
|
||||||
|
}
|
||||||
|
|
||||||
|
// If a regular byte slice, we can read it over and over via new
|
||||||
|
// readers
|
||||||
|
case []byte:
|
||||||
|
buf := rawBody.([]byte)
|
||||||
|
body = func() (io.Reader, error) {
|
||||||
|
return bytes.NewReader(buf), nil
|
||||||
|
}
|
||||||
|
contentLength = int64(len(buf))
|
||||||
|
|
||||||
|
// If a bytes.Buffer we can read the underlying byte slice over and
|
||||||
|
// over
|
||||||
|
case *bytes.Buffer:
|
||||||
|
buf := rawBody.(*bytes.Buffer)
|
||||||
|
body = func() (io.Reader, error) {
|
||||||
|
return bytes.NewReader(buf.Bytes()), nil
|
||||||
|
}
|
||||||
|
contentLength = int64(buf.Len())
|
||||||
|
|
||||||
|
// We prioritize *bytes.Reader here because we don't really want to
|
||||||
|
// deal with it seeking so want it to match here instead of the
|
||||||
|
// io.ReadSeeker case.
|
||||||
|
case *bytes.Reader:
|
||||||
|
buf, err := ioutil.ReadAll(rawBody.(*bytes.Reader))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
body = func() (io.Reader, error) {
|
||||||
|
return bytes.NewReader(buf), nil
|
||||||
|
}
|
||||||
|
contentLength = int64(len(buf))
|
||||||
|
|
||||||
|
// Compat case
|
||||||
|
case io.ReadSeeker:
|
||||||
|
raw := rawBody.(io.ReadSeeker)
|
||||||
|
body = func() (io.Reader, error) {
|
||||||
|
raw.Seek(0, 0)
|
||||||
|
return ioutil.NopCloser(raw), nil
|
||||||
|
}
|
||||||
|
if lr, ok := raw.(LenReader); ok {
|
||||||
|
contentLength = int64(lr.Len())
|
||||||
|
}
|
||||||
|
|
||||||
|
// Read all in so we can reset
|
||||||
|
case io.Reader:
|
||||||
|
buf, err := ioutil.ReadAll(rawBody.(io.Reader))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
body = func() (io.Reader, error) {
|
||||||
|
return bytes.NewReader(buf), nil
|
||||||
|
}
|
||||||
|
contentLength = int64(len(buf))
|
||||||
|
|
||||||
|
default:
|
||||||
|
return nil, fmt.Errorf("cannot handle type %T", rawBody)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Make the request with the noop-closer for the body.
|
httpReq, err := http.NewRequest(method, url, nil)
|
||||||
httpReq, err := http.NewRequest(method, url, rcBody)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
httpReq.ContentLength = contentLength
|
||||||
// Check if we can set the Content-Length automatically.
|
|
||||||
if lr, ok := body.(LenReader); ok {
|
|
||||||
httpReq.ContentLength = int64(lr.Len())
|
|
||||||
}
|
|
||||||
|
|
||||||
return &Request{body, httpReq}, nil
|
return &Request{body, httpReq}, nil
|
||||||
}
|
}
|
||||||
|
@ -105,7 +204,18 @@ type ResponseLogHook func(*log.Logger, *http.Response)
|
||||||
// Client will close any response body when retrying, but if the retry is
|
// Client will close any response body when retrying, but if the retry is
|
||||||
// aborted it is up to the CheckResponse callback to properly close any
|
// aborted it is up to the CheckResponse callback to properly close any
|
||||||
// response body before returning.
|
// response body before returning.
|
||||||
type CheckRetry func(resp *http.Response, err error) (bool, error)
|
type CheckRetry func(ctx context.Context, resp *http.Response, err error) (bool, error)
|
||||||
|
|
||||||
|
// Backoff specifies a policy for how long to wait between retries.
|
||||||
|
// It is called after a failing request to determine the amount of time
|
||||||
|
// that should pass before trying again.
|
||||||
|
type Backoff func(min, max time.Duration, attemptNum int, resp *http.Response) time.Duration
|
||||||
|
|
||||||
|
// ErrorHandler is called if retries are expired, containing the last status
|
||||||
|
// from the http library. If not specified, default behavior for the library is
|
||||||
|
// to close the body and return an error indicating how many tries were
|
||||||
|
// attempted. If overriding this, be sure to close the body if needed.
|
||||||
|
type ErrorHandler func(resp *http.Response, err error, numTries int) (*http.Response, error)
|
||||||
|
|
||||||
// Client is used to make HTTP requests. It adds additional functionality
|
// Client is used to make HTTP requests. It adds additional functionality
|
||||||
// like automatic retries to tolerate minor outages.
|
// like automatic retries to tolerate minor outages.
|
||||||
|
@ -128,6 +238,12 @@ type Client struct {
|
||||||
// CheckRetry specifies the policy for handling retries, and is called
|
// CheckRetry specifies the policy for handling retries, and is called
|
||||||
// after each request. The default policy is DefaultRetryPolicy.
|
// after each request. The default policy is DefaultRetryPolicy.
|
||||||
CheckRetry CheckRetry
|
CheckRetry CheckRetry
|
||||||
|
|
||||||
|
// Backoff specifies the policy for how long to wait between retries
|
||||||
|
Backoff Backoff
|
||||||
|
|
||||||
|
// ErrorHandler specifies the custom error handler to use, if any
|
||||||
|
ErrorHandler ErrorHandler
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewClient creates a new Client with default settings.
|
// NewClient creates a new Client with default settings.
|
||||||
|
@ -139,12 +255,18 @@ func NewClient() *Client {
|
||||||
RetryWaitMax: defaultRetryWaitMax,
|
RetryWaitMax: defaultRetryWaitMax,
|
||||||
RetryMax: defaultRetryMax,
|
RetryMax: defaultRetryMax,
|
||||||
CheckRetry: DefaultRetryPolicy,
|
CheckRetry: DefaultRetryPolicy,
|
||||||
|
Backoff: DefaultBackoff,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// DefaultRetryPolicy provides a default callback for Client.CheckRetry, which
|
// DefaultRetryPolicy provides a default callback for Client.CheckRetry, which
|
||||||
// will retry on connection errors and server errors.
|
// will retry on connection errors and server errors.
|
||||||
func DefaultRetryPolicy(resp *http.Response, err error) (bool, error) {
|
func DefaultRetryPolicy(ctx context.Context, resp *http.Response, err error) (bool, error) {
|
||||||
|
// do not retry on context.Canceled or context.DeadlineExceeded
|
||||||
|
if ctx.Err() != nil {
|
||||||
|
return false, ctx.Err()
|
||||||
|
}
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return true, err
|
return true, err
|
||||||
}
|
}
|
||||||
|
@ -152,24 +274,92 @@ func DefaultRetryPolicy(resp *http.Response, err error) (bool, error) {
|
||||||
// the server time to recover, as 500's are typically not permanent
|
// the server time to recover, as 500's are typically not permanent
|
||||||
// errors and may relate to outages on the server side. This will catch
|
// errors and may relate to outages on the server side. This will catch
|
||||||
// invalid response codes as well, like 0 and 999.
|
// invalid response codes as well, like 0 and 999.
|
||||||
if resp.StatusCode == 0 || resp.StatusCode >= 500 {
|
if resp.StatusCode == 0 || (resp.StatusCode >= 500 && resp.StatusCode != 501) {
|
||||||
return true, nil
|
return true, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
return false, nil
|
return false, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// DefaultBackoff provides a default callback for Client.Backoff which
|
||||||
|
// will perform exponential backoff based on the attempt number and limited
|
||||||
|
// by the provided minimum and maximum durations.
|
||||||
|
func DefaultBackoff(min, max time.Duration, attemptNum int, resp *http.Response) time.Duration {
|
||||||
|
mult := math.Pow(2, float64(attemptNum)) * float64(min)
|
||||||
|
sleep := time.Duration(mult)
|
||||||
|
if float64(sleep) != mult || sleep > max {
|
||||||
|
sleep = max
|
||||||
|
}
|
||||||
|
return sleep
|
||||||
|
}
|
||||||
|
|
||||||
|
// LinearJitterBackoff provides a callback for Client.Backoff which will
|
||||||
|
// perform linear backoff based on the attempt number and with jitter to
|
||||||
|
// prevent a thundering herd.
|
||||||
|
//
|
||||||
|
// min and max here are *not* absolute values. The number to be multipled by
|
||||||
|
// the attempt number will be chosen at random from between them, thus they are
|
||||||
|
// bounding the jitter.
|
||||||
|
//
|
||||||
|
// For instance:
|
||||||
|
// * To get strictly linear backoff of one second increasing each retry, set
|
||||||
|
// both to one second (1s, 2s, 3s, 4s, ...)
|
||||||
|
// * To get a small amount of jitter centered around one second increasing each
|
||||||
|
// retry, set to around one second, such as a min of 800ms and max of 1200ms
|
||||||
|
// (892ms, 2102ms, 2945ms, 4312ms, ...)
|
||||||
|
// * To get extreme jitter, set to a very wide spread, such as a min of 100ms
|
||||||
|
// and a max of 20s (15382ms, 292ms, 51321ms, 35234ms, ...)
|
||||||
|
func LinearJitterBackoff(min, max time.Duration, attemptNum int, resp *http.Response) time.Duration {
|
||||||
|
// attemptNum always starts at zero but we want to start at 1 for multiplication
|
||||||
|
attemptNum++
|
||||||
|
|
||||||
|
if max <= min {
|
||||||
|
// Unclear what to do here, or they are the same, so return min *
|
||||||
|
// attemptNum
|
||||||
|
return min * time.Duration(attemptNum)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Seed rand; doing this every time is fine
|
||||||
|
rand := rand.New(rand.NewSource(int64(time.Now().Nanosecond())))
|
||||||
|
|
||||||
|
// Pick a random number that lies somewhere between the min and max and
|
||||||
|
// multiply by the attemptNum. attemptNum starts at zero so we always
|
||||||
|
// increment here. We first get a random percentage, then apply that to the
|
||||||
|
// difference between min and max, and add to min.
|
||||||
|
jitter := rand.Float64() * float64(max-min)
|
||||||
|
jitterMin := int64(jitter) + int64(min)
|
||||||
|
return time.Duration(jitterMin * int64(attemptNum))
|
||||||
|
}
|
||||||
|
|
||||||
|
// PassthroughErrorHandler is an ErrorHandler that directly passes through the
|
||||||
|
// values from the net/http library for the final request. The body is not
|
||||||
|
// closed.
|
||||||
|
func PassthroughErrorHandler(resp *http.Response, err error, _ int) (*http.Response, error) {
|
||||||
|
return resp, err
|
||||||
|
}
|
||||||
|
|
||||||
// Do wraps calling an HTTP method with retries.
|
// Do wraps calling an HTTP method with retries.
|
||||||
func (c *Client) Do(req *Request) (*http.Response, error) {
|
func (c *Client) Do(req *Request) (*http.Response, error) {
|
||||||
c.Logger.Printf("[DEBUG] %s %s", req.Method, req.URL)
|
if c.Logger != nil {
|
||||||
|
c.Logger.Printf("[DEBUG] %s %s", req.Method, req.URL)
|
||||||
|
}
|
||||||
|
|
||||||
|
var resp *http.Response
|
||||||
|
var err error
|
||||||
|
|
||||||
for i := 0; ; i++ {
|
for i := 0; ; i++ {
|
||||||
var code int // HTTP response code
|
var code int // HTTP response code
|
||||||
|
|
||||||
// Always rewind the request body when non-nil.
|
// Always rewind the request body when non-nil.
|
||||||
if req.body != nil {
|
if req.body != nil {
|
||||||
if _, err := req.body.Seek(0, 0); err != nil {
|
body, err := req.body()
|
||||||
return nil, fmt.Errorf("failed to seek body: %v", err)
|
if err != nil {
|
||||||
|
return resp, err
|
||||||
|
}
|
||||||
|
if c, ok := body.(io.ReadCloser); ok {
|
||||||
|
req.Request.Body = c
|
||||||
|
} else {
|
||||||
|
req.Request.Body = ioutil.NopCloser(body)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -178,13 +368,18 @@ func (c *Client) Do(req *Request) (*http.Response, error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Attempt the request
|
// Attempt the request
|
||||||
resp, err := c.HTTPClient.Do(req.Request)
|
resp, err = c.HTTPClient.Do(req.Request)
|
||||||
|
if resp != nil {
|
||||||
|
code = resp.StatusCode
|
||||||
|
}
|
||||||
|
|
||||||
// Check if we should continue with retries.
|
// Check if we should continue with retries.
|
||||||
checkOK, checkErr := c.CheckRetry(resp, err)
|
checkOK, checkErr := c.CheckRetry(req.Request.Context(), resp, err)
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
c.Logger.Printf("[ERR] %s %s request failed: %v", req.Method, req.URL, err)
|
if c.Logger != nil {
|
||||||
|
c.Logger.Printf("[ERR] %s %s request failed: %v", req.Method, req.URL, err)
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
// Call this here to maintain the behavior of logging all requests,
|
// Call this here to maintain the behavior of logging all requests,
|
||||||
// even if CheckRetry signals to stop.
|
// even if CheckRetry signals to stop.
|
||||||
|
@ -202,25 +397,38 @@ func (c *Client) Do(req *Request) (*http.Response, error) {
|
||||||
return resp, err
|
return resp, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// We do this before drainBody beause there's no need for the I/O if
|
||||||
|
// we're breaking out
|
||||||
|
remain := c.RetryMax - i
|
||||||
|
if remain <= 0 {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
// We're going to retry, consume any response to reuse the connection.
|
// We're going to retry, consume any response to reuse the connection.
|
||||||
if err == nil {
|
if err == nil && resp != nil {
|
||||||
c.drainBody(resp.Body)
|
c.drainBody(resp.Body)
|
||||||
}
|
}
|
||||||
|
|
||||||
remain := c.RetryMax - i
|
wait := c.Backoff(c.RetryWaitMin, c.RetryWaitMax, i, resp)
|
||||||
if remain == 0 {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
wait := backoff(c.RetryWaitMin, c.RetryWaitMax, i)
|
|
||||||
desc := fmt.Sprintf("%s %s", req.Method, req.URL)
|
desc := fmt.Sprintf("%s %s", req.Method, req.URL)
|
||||||
if code > 0 {
|
if code > 0 {
|
||||||
desc = fmt.Sprintf("%s (status: %d)", desc, code)
|
desc = fmt.Sprintf("%s (status: %d)", desc, code)
|
||||||
}
|
}
|
||||||
c.Logger.Printf("[DEBUG] %s: retrying in %s (%d left)", desc, wait, remain)
|
if c.Logger != nil {
|
||||||
|
c.Logger.Printf("[DEBUG] %s: retrying in %s (%d left)", desc, wait, remain)
|
||||||
|
}
|
||||||
time.Sleep(wait)
|
time.Sleep(wait)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Return an error if we fall out of the retry loop
|
if c.ErrorHandler != nil {
|
||||||
|
return c.ErrorHandler(resp, err, c.RetryMax+1)
|
||||||
|
}
|
||||||
|
|
||||||
|
// By default, we close the response body and return an error without
|
||||||
|
// returning the response
|
||||||
|
if resp != nil {
|
||||||
|
resp.Body.Close()
|
||||||
|
}
|
||||||
return nil, fmt.Errorf("%s %s giving up after %d attempts",
|
return nil, fmt.Errorf("%s %s giving up after %d attempts",
|
||||||
req.Method, req.URL, c.RetryMax+1)
|
req.Method, req.URL, c.RetryMax+1)
|
||||||
}
|
}
|
||||||
|
@ -230,7 +438,9 @@ func (c *Client) drainBody(body io.ReadCloser) {
|
||||||
defer body.Close()
|
defer body.Close()
|
||||||
_, err := io.Copy(ioutil.Discard, io.LimitReader(body, respReadLimit))
|
_, err := io.Copy(ioutil.Discard, io.LimitReader(body, respReadLimit))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
c.Logger.Printf("[ERR] error reading response body: %v", err)
|
if c.Logger != nil {
|
||||||
|
c.Logger.Printf("[ERR] error reading response body: %v", err)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -263,12 +473,12 @@ func (c *Client) Head(url string) (*http.Response, error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Post is a shortcut for doing a POST request without making a new client.
|
// Post is a shortcut for doing a POST request without making a new client.
|
||||||
func Post(url, bodyType string, body io.ReadSeeker) (*http.Response, error) {
|
func Post(url, bodyType string, body interface{}) (*http.Response, error) {
|
||||||
return defaultClient.Post(url, bodyType, body)
|
return defaultClient.Post(url, bodyType, body)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Post is a convenience method for doing simple POST requests.
|
// Post is a convenience method for doing simple POST requests.
|
||||||
func (c *Client) Post(url, bodyType string, body io.ReadSeeker) (*http.Response, error) {
|
func (c *Client) Post(url, bodyType string, body interface{}) (*http.Response, error) {
|
||||||
req, err := NewRequest("POST", url, body)
|
req, err := NewRequest("POST", url, body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
|
@ -288,15 +498,3 @@ func PostForm(url string, data url.Values) (*http.Response, error) {
|
||||||
func (c *Client) PostForm(url string, data url.Values) (*http.Response, error) {
|
func (c *Client) PostForm(url string, data url.Values) (*http.Response, error) {
|
||||||
return c.Post(url, "application/x-www-form-urlencoded", strings.NewReader(data.Encode()))
|
return c.Post(url, "application/x-www-form-urlencoded", strings.NewReader(data.Encode()))
|
||||||
}
|
}
|
||||||
|
|
||||||
// backoff is used to calculate how long to sleep before retrying
|
|
||||||
// after observing failures. It takes the minimum/maximum wait time and
|
|
||||||
// iteration, and returns the duration to wait.
|
|
||||||
func backoff(min, max time.Duration, iter int) time.Duration {
|
|
||||||
mult := math.Pow(2, float64(iter)) * float64(min)
|
|
||||||
sleep := time.Duration(mult)
|
|
||||||
if float64(sleep) != mult || sleep > max {
|
|
||||||
sleep = max
|
|
||||||
}
|
|
||||||
return sleep
|
|
||||||
}
|
|
||||||
|
|
|
@ -0,0 +1,65 @@
|
||||||
|
TOOLS= golang.org/x/tools/cover
|
||||||
|
GOCOVER_TMPFILE?= $(GOCOVER_FILE).tmp
|
||||||
|
GOCOVER_FILE?= .cover.out
|
||||||
|
GOCOVERHTML?= coverage.html
|
||||||
|
FIND=`/usr/bin/which 2> /dev/null gfind find | /usr/bin/grep -v ^no | /usr/bin/head -n 1`
|
||||||
|
XARGS=`/usr/bin/which 2> /dev/null gxargs xargs | /usr/bin/grep -v ^no | /usr/bin/head -n 1`
|
||||||
|
|
||||||
|
test:: $(GOCOVER_FILE)
|
||||||
|
@$(MAKE) -C cmd/sockaddr test
|
||||||
|
|
||||||
|
cover:: coverage_report
|
||||||
|
|
||||||
|
$(GOCOVER_FILE)::
|
||||||
|
@${FIND} . -type d ! -path '*cmd*' ! -path '*.git*' -print0 | ${XARGS} -0 -I % sh -ec "cd % && rm -f $(GOCOVER_TMPFILE) && go test -coverprofile=$(GOCOVER_TMPFILE)"
|
||||||
|
|
||||||
|
@echo 'mode: set' > $(GOCOVER_FILE)
|
||||||
|
@${FIND} . -type f ! -path '*cmd*' ! -path '*.git*' -name "$(GOCOVER_TMPFILE)" -print0 | ${XARGS} -0 -n1 cat $(GOCOVER_TMPFILE) | grep -v '^mode: ' >> ${PWD}/$(GOCOVER_FILE)
|
||||||
|
|
||||||
|
$(GOCOVERHTML): $(GOCOVER_FILE)
|
||||||
|
go tool cover -html=$(GOCOVER_FILE) -o $(GOCOVERHTML)
|
||||||
|
|
||||||
|
coverage_report:: $(GOCOVER_FILE)
|
||||||
|
go tool cover -html=$(GOCOVER_FILE)
|
||||||
|
|
||||||
|
audit_tools::
|
||||||
|
@go get -u github.com/golang/lint/golint && echo "Installed golint:"
|
||||||
|
@go get -u github.com/fzipp/gocyclo && echo "Installed gocyclo:"
|
||||||
|
@go get -u github.com/remyoudompheng/go-misc/deadcode && echo "Installed deadcode:"
|
||||||
|
@go get -u github.com/client9/misspell/cmd/misspell && echo "Installed misspell:"
|
||||||
|
@go get -u github.com/gordonklaus/ineffassign && echo "Installed ineffassign:"
|
||||||
|
|
||||||
|
audit::
|
||||||
|
deadcode
|
||||||
|
go tool vet -all *.go
|
||||||
|
go tool vet -shadow=true *.go
|
||||||
|
golint *.go
|
||||||
|
ineffassign .
|
||||||
|
gocyclo -over 65 *.go
|
||||||
|
misspell *.go
|
||||||
|
|
||||||
|
clean::
|
||||||
|
rm -f $(GOCOVER_FILE) $(GOCOVERHTML)
|
||||||
|
|
||||||
|
dev::
|
||||||
|
@go build
|
||||||
|
@$(MAKE) -B -C cmd/sockaddr sockaddr
|
||||||
|
|
||||||
|
install::
|
||||||
|
@go install
|
||||||
|
@$(MAKE) -C cmd/sockaddr install
|
||||||
|
|
||||||
|
doc::
|
||||||
|
@echo Visit: http://127.0.0.1:6161/pkg/github.com/hashicorp/go-sockaddr/
|
||||||
|
godoc -http=:6161 -goroot $GOROOT
|
||||||
|
|
||||||
|
world::
|
||||||
|
@set -e; \
|
||||||
|
for os in solaris darwin freebsd linux windows; do \
|
||||||
|
for arch in amd64; do \
|
||||||
|
printf "Building on %s-%s\n" "$${os}" "$${arch}" ; \
|
||||||
|
env GOOS="$${os}" GOARCH="$${arch}" go build -o /dev/null; \
|
||||||
|
done; \
|
||||||
|
done
|
||||||
|
|
||||||
|
$(MAKE) -C cmd/sockaddr world
|
|
@ -0,0 +1,373 @@
|
||||||
|
Mozilla Public License Version 2.0
|
||||||
|
==================================
|
||||||
|
|
||||||
|
1. Definitions
|
||||||
|
--------------
|
||||||
|
|
||||||
|
1.1. "Contributor"
|
||||||
|
means each individual or legal entity that creates, contributes to
|
||||||
|
the creation of, or owns Covered Software.
|
||||||
|
|
||||||
|
1.2. "Contributor Version"
|
||||||
|
means the combination of the Contributions of others (if any) used
|
||||||
|
by a Contributor and that particular Contributor's Contribution.
|
||||||
|
|
||||||
|
1.3. "Contribution"
|
||||||
|
means Covered Software of a particular Contributor.
|
||||||
|
|
||||||
|
1.4. "Covered Software"
|
||||||
|
means Source Code Form to which the initial Contributor has attached
|
||||||
|
the notice in Exhibit A, the Executable Form of such Source Code
|
||||||
|
Form, and Modifications of such Source Code Form, in each case
|
||||||
|
including portions thereof.
|
||||||
|
|
||||||
|
1.5. "Incompatible With Secondary Licenses"
|
||||||
|
means
|
||||||
|
|
||||||
|
(a) that the initial Contributor has attached the notice described
|
||||||
|
in Exhibit B to the Covered Software; or
|
||||||
|
|
||||||
|
(b) that the Covered Software was made available under the terms of
|
||||||
|
version 1.1 or earlier of the License, but not also under the
|
||||||
|
terms of a Secondary License.
|
||||||
|
|
||||||
|
1.6. "Executable Form"
|
||||||
|
means any form of the work other than Source Code Form.
|
||||||
|
|
||||||
|
1.7. "Larger Work"
|
||||||
|
means a work that combines Covered Software with other material, in
|
||||||
|
a separate file or files, that is not Covered Software.
|
||||||
|
|
||||||
|
1.8. "License"
|
||||||
|
means this document.
|
||||||
|
|
||||||
|
1.9. "Licensable"
|
||||||
|
means having the right to grant, to the maximum extent possible,
|
||||||
|
whether at the time of the initial grant or subsequently, any and
|
||||||
|
all of the rights conveyed by this License.
|
||||||
|
|
||||||
|
1.10. "Modifications"
|
||||||
|
means any of the following:
|
||||||
|
|
||||||
|
(a) any file in Source Code Form that results from an addition to,
|
||||||
|
deletion from, or modification of the contents of Covered
|
||||||
|
Software; or
|
||||||
|
|
||||||
|
(b) any new file in Source Code Form that contains any Covered
|
||||||
|
Software.
|
||||||
|
|
||||||
|
1.11. "Patent Claims" of a Contributor
|
||||||
|
means any patent claim(s), including without limitation, method,
|
||||||
|
process, and apparatus claims, in any patent Licensable by such
|
||||||
|
Contributor that would be infringed, but for the grant of the
|
||||||
|
License, by the making, using, selling, offering for sale, having
|
||||||
|
made, import, or transfer of either its Contributions or its
|
||||||
|
Contributor Version.
|
||||||
|
|
||||||
|
1.12. "Secondary License"
|
||||||
|
means either the GNU General Public License, Version 2.0, the GNU
|
||||||
|
Lesser General Public License, Version 2.1, the GNU Affero General
|
||||||
|
Public License, Version 3.0, or any later versions of those
|
||||||
|
licenses.
|
||||||
|
|
||||||
|
1.13. "Source Code Form"
|
||||||
|
means the form of the work preferred for making modifications.
|
||||||
|
|
||||||
|
1.14. "You" (or "Your")
|
||||||
|
means an individual or a legal entity exercising rights under this
|
||||||
|
License. For legal entities, "You" includes any entity that
|
||||||
|
controls, is controlled by, or is under common control with You. For
|
||||||
|
purposes of this definition, "control" means (a) the power, direct
|
||||||
|
or indirect, to cause the direction or management of such entity,
|
||||||
|
whether by contract or otherwise, or (b) ownership of more than
|
||||||
|
fifty percent (50%) of the outstanding shares or beneficial
|
||||||
|
ownership of such entity.
|
||||||
|
|
||||||
|
2. License Grants and Conditions
|
||||||
|
--------------------------------
|
||||||
|
|
||||||
|
2.1. Grants
|
||||||
|
|
||||||
|
Each Contributor hereby grants You a world-wide, royalty-free,
|
||||||
|
non-exclusive license:
|
||||||
|
|
||||||
|
(a) under intellectual property rights (other than patent or trademark)
|
||||||
|
Licensable by such Contributor to use, reproduce, make available,
|
||||||
|
modify, display, perform, distribute, and otherwise exploit its
|
||||||
|
Contributions, either on an unmodified basis, with Modifications, or
|
||||||
|
as part of a Larger Work; and
|
||||||
|
|
||||||
|
(b) under Patent Claims of such Contributor to make, use, sell, offer
|
||||||
|
for sale, have made, import, and otherwise transfer either its
|
||||||
|
Contributions or its Contributor Version.
|
||||||
|
|
||||||
|
2.2. Effective Date
|
||||||
|
|
||||||
|
The licenses granted in Section 2.1 with respect to any Contribution
|
||||||
|
become effective for each Contribution on the date the Contributor first
|
||||||
|
distributes such Contribution.
|
||||||
|
|
||||||
|
2.3. Limitations on Grant Scope
|
||||||
|
|
||||||
|
The licenses granted in this Section 2 are the only rights granted under
|
||||||
|
this License. No additional rights or licenses will be implied from the
|
||||||
|
distribution or licensing of Covered Software under this License.
|
||||||
|
Notwithstanding Section 2.1(b) above, no patent license is granted by a
|
||||||
|
Contributor:
|
||||||
|
|
||||||
|
(a) for any code that a Contributor has removed from Covered Software;
|
||||||
|
or
|
||||||
|
|
||||||
|
(b) for infringements caused by: (i) Your and any other third party's
|
||||||
|
modifications of Covered Software, or (ii) the combination of its
|
||||||
|
Contributions with other software (except as part of its Contributor
|
||||||
|
Version); or
|
||||||
|
|
||||||
|
(c) under Patent Claims infringed by Covered Software in the absence of
|
||||||
|
its Contributions.
|
||||||
|
|
||||||
|
This License does not grant any rights in the trademarks, service marks,
|
||||||
|
or logos of any Contributor (except as may be necessary to comply with
|
||||||
|
the notice requirements in Section 3.4).
|
||||||
|
|
||||||
|
2.4. Subsequent Licenses
|
||||||
|
|
||||||
|
No Contributor makes additional grants as a result of Your choice to
|
||||||
|
distribute the Covered Software under a subsequent version of this
|
||||||
|
License (see Section 10.2) or under the terms of a Secondary License (if
|
||||||
|
permitted under the terms of Section 3.3).
|
||||||
|
|
||||||
|
2.5. Representation
|
||||||
|
|
||||||
|
Each Contributor represents that the Contributor believes its
|
||||||
|
Contributions are its original creation(s) or it has sufficient rights
|
||||||
|
to grant the rights to its Contributions conveyed by this License.
|
||||||
|
|
||||||
|
2.6. Fair Use
|
||||||
|
|
||||||
|
This License is not intended to limit any rights You have under
|
||||||
|
applicable copyright doctrines of fair use, fair dealing, or other
|
||||||
|
equivalents.
|
||||||
|
|
||||||
|
2.7. Conditions
|
||||||
|
|
||||||
|
Sections 3.1, 3.2, 3.3, and 3.4 are conditions of the licenses granted
|
||||||
|
in Section 2.1.
|
||||||
|
|
||||||
|
3. Responsibilities
|
||||||
|
-------------------
|
||||||
|
|
||||||
|
3.1. Distribution of Source Form
|
||||||
|
|
||||||
|
All distribution of Covered Software in Source Code Form, including any
|
||||||
|
Modifications that You create or to which You contribute, must be under
|
||||||
|
the terms of this License. You must inform recipients that the Source
|
||||||
|
Code Form of the Covered Software is governed by the terms of this
|
||||||
|
License, and how they can obtain a copy of this License. You may not
|
||||||
|
attempt to alter or restrict the recipients' rights in the Source Code
|
||||||
|
Form.
|
||||||
|
|
||||||
|
3.2. Distribution of Executable Form
|
||||||
|
|
||||||
|
If You distribute Covered Software in Executable Form then:
|
||||||
|
|
||||||
|
(a) such Covered Software must also be made available in Source Code
|
||||||
|
Form, as described in Section 3.1, and You must inform recipients of
|
||||||
|
the Executable Form how they can obtain a copy of such Source Code
|
||||||
|
Form by reasonable means in a timely manner, at a charge no more
|
||||||
|
than the cost of distribution to the recipient; and
|
||||||
|
|
||||||
|
(b) You may distribute such Executable Form under the terms of this
|
||||||
|
License, or sublicense it under different terms, provided that the
|
||||||
|
license for the Executable Form does not attempt to limit or alter
|
||||||
|
the recipients' rights in the Source Code Form under this License.
|
||||||
|
|
||||||
|
3.3. Distribution of a Larger Work
|
||||||
|
|
||||||
|
You may create and distribute a Larger Work under terms of Your choice,
|
||||||
|
provided that You also comply with the requirements of this License for
|
||||||
|
the Covered Software. If the Larger Work is a combination of Covered
|
||||||
|
Software with a work governed by one or more Secondary Licenses, and the
|
||||||
|
Covered Software is not Incompatible With Secondary Licenses, this
|
||||||
|
License permits You to additionally distribute such Covered Software
|
||||||
|
under the terms of such Secondary License(s), so that the recipient of
|
||||||
|
the Larger Work may, at their option, further distribute the Covered
|
||||||
|
Software under the terms of either this License or such Secondary
|
||||||
|
License(s).
|
||||||
|
|
||||||
|
3.4. Notices
|
||||||
|
|
||||||
|
You may not remove or alter the substance of any license notices
|
||||||
|
(including copyright notices, patent notices, disclaimers of warranty,
|
||||||
|
or limitations of liability) contained within the Source Code Form of
|
||||||
|
the Covered Software, except that You may alter any license notices to
|
||||||
|
the extent required to remedy known factual inaccuracies.
|
||||||
|
|
||||||
|
3.5. Application of Additional Terms
|
||||||
|
|
||||||
|
You may choose to offer, and to charge a fee for, warranty, support,
|
||||||
|
indemnity or liability obligations to one or more recipients of Covered
|
||||||
|
Software. However, You may do so only on Your own behalf, and not on
|
||||||
|
behalf of any Contributor. You must make it absolutely clear that any
|
||||||
|
such warranty, support, indemnity, or liability obligation is offered by
|
||||||
|
You alone, and You hereby agree to indemnify every Contributor for any
|
||||||
|
liability incurred by such Contributor as a result of warranty, support,
|
||||||
|
indemnity or liability terms You offer. You may include additional
|
||||||
|
disclaimers of warranty and limitations of liability specific to any
|
||||||
|
jurisdiction.
|
||||||
|
|
||||||
|
4. Inability to Comply Due to Statute or Regulation
|
||||||
|
---------------------------------------------------
|
||||||
|
|
||||||
|
If it is impossible for You to comply with any of the terms of this
|
||||||
|
License with respect to some or all of the Covered Software due to
|
||||||
|
statute, judicial order, or regulation then You must: (a) comply with
|
||||||
|
the terms of this License to the maximum extent possible; and (b)
|
||||||
|
describe the limitations and the code they affect. Such description must
|
||||||
|
be placed in a text file included with all distributions of the Covered
|
||||||
|
Software under this License. Except to the extent prohibited by statute
|
||||||
|
or regulation, such description must be sufficiently detailed for a
|
||||||
|
recipient of ordinary skill to be able to understand it.
|
||||||
|
|
||||||
|
5. Termination
|
||||||
|
--------------
|
||||||
|
|
||||||
|
5.1. The rights granted under this License will terminate automatically
|
||||||
|
if You fail to comply with any of its terms. However, if You become
|
||||||
|
compliant, then the rights granted under this License from a particular
|
||||||
|
Contributor are reinstated (a) provisionally, unless and until such
|
||||||
|
Contributor explicitly and finally terminates Your grants, and (b) on an
|
||||||
|
ongoing basis, if such Contributor fails to notify You of the
|
||||||
|
non-compliance by some reasonable means prior to 60 days after You have
|
||||||
|
come back into compliance. Moreover, Your grants from a particular
|
||||||
|
Contributor are reinstated on an ongoing basis if such Contributor
|
||||||
|
notifies You of the non-compliance by some reasonable means, this is the
|
||||||
|
first time You have received notice of non-compliance with this License
|
||||||
|
from such Contributor, and You become compliant prior to 30 days after
|
||||||
|
Your receipt of the notice.
|
||||||
|
|
||||||
|
5.2. If You initiate litigation against any entity by asserting a patent
|
||||||
|
infringement claim (excluding declaratory judgment actions,
|
||||||
|
counter-claims, and cross-claims) alleging that a Contributor Version
|
||||||
|
directly or indirectly infringes any patent, then the rights granted to
|
||||||
|
You by any and all Contributors for the Covered Software under Section
|
||||||
|
2.1 of this License shall terminate.
|
||||||
|
|
||||||
|
5.3. In the event of termination under Sections 5.1 or 5.2 above, all
|
||||||
|
end user license agreements (excluding distributors and resellers) which
|
||||||
|
have been validly granted by You or Your distributors under this License
|
||||||
|
prior to termination shall survive termination.
|
||||||
|
|
||||||
|
************************************************************************
|
||||||
|
* *
|
||||||
|
* 6. Disclaimer of Warranty *
|
||||||
|
* ------------------------- *
|
||||||
|
* *
|
||||||
|
* Covered Software is provided under this License on an "as is" *
|
||||||
|
* basis, without warranty of any kind, either expressed, implied, or *
|
||||||
|
* statutory, including, without limitation, warranties that the *
|
||||||
|
* Covered Software is free of defects, merchantable, fit for a *
|
||||||
|
* particular purpose or non-infringing. The entire risk as to the *
|
||||||
|
* quality and performance of the Covered Software is with You. *
|
||||||
|
* Should any Covered Software prove defective in any respect, You *
|
||||||
|
* (not any Contributor) assume the cost of any necessary servicing, *
|
||||||
|
* repair, or correction. This disclaimer of warranty constitutes an *
|
||||||
|
* essential part of this License. No use of any Covered Software is *
|
||||||
|
* authorized under this License except under this disclaimer. *
|
||||||
|
* *
|
||||||
|
************************************************************************
|
||||||
|
|
||||||
|
************************************************************************
|
||||||
|
* *
|
||||||
|
* 7. Limitation of Liability *
|
||||||
|
* -------------------------- *
|
||||||
|
* *
|
||||||
|
* Under no circumstances and under no legal theory, whether tort *
|
||||||
|
* (including negligence), contract, or otherwise, shall any *
|
||||||
|
* Contributor, or anyone who distributes Covered Software as *
|
||||||
|
* permitted above, be liable to You for any direct, indirect, *
|
||||||
|
* special, incidental, or consequential damages of any character *
|
||||||
|
* including, without limitation, damages for lost profits, loss of *
|
||||||
|
* goodwill, work stoppage, computer failure or malfunction, or any *
|
||||||
|
* and all other commercial damages or losses, even if such party *
|
||||||
|
* shall have been informed of the possibility of such damages. This *
|
||||||
|
* limitation of liability shall not apply to liability for death or *
|
||||||
|
* personal injury resulting from such party's negligence to the *
|
||||||
|
* extent applicable law prohibits such limitation. Some *
|
||||||
|
* jurisdictions do not allow the exclusion or limitation of *
|
||||||
|
* incidental or consequential damages, so this exclusion and *
|
||||||
|
* limitation may not apply to You. *
|
||||||
|
* *
|
||||||
|
************************************************************************
|
||||||
|
|
||||||
|
8. Litigation
|
||||||
|
-------------
|
||||||
|
|
||||||
|
Any litigation relating to this License may be brought only in the
|
||||||
|
courts of a jurisdiction where the defendant maintains its principal
|
||||||
|
place of business and such litigation shall be governed by laws of that
|
||||||
|
jurisdiction, without reference to its conflict-of-law provisions.
|
||||||
|
Nothing in this Section shall prevent a party's ability to bring
|
||||||
|
cross-claims or counter-claims.
|
||||||
|
|
||||||
|
9. Miscellaneous
|
||||||
|
----------------
|
||||||
|
|
||||||
|
This License represents the complete agreement concerning the subject
|
||||||
|
matter hereof. If any provision of this License is held to be
|
||||||
|
unenforceable, such provision shall be reformed only to the extent
|
||||||
|
necessary to make it enforceable. Any law or regulation which provides
|
||||||
|
that the language of a contract shall be construed against the drafter
|
||||||
|
shall not be used to construe this License against a Contributor.
|
||||||
|
|
||||||
|
10. Versions of the License
|
||||||
|
---------------------------
|
||||||
|
|
||||||
|
10.1. New Versions
|
||||||
|
|
||||||
|
Mozilla Foundation is the license steward. Except as provided in Section
|
||||||
|
10.3, no one other than the license steward has the right to modify or
|
||||||
|
publish new versions of this License. Each version will be given a
|
||||||
|
distinguishing version number.
|
||||||
|
|
||||||
|
10.2. Effect of New Versions
|
||||||
|
|
||||||
|
You may distribute the Covered Software under the terms of the version
|
||||||
|
of the License under which You originally received the Covered Software,
|
||||||
|
or under the terms of any subsequent version published by the license
|
||||||
|
steward.
|
||||||
|
|
||||||
|
10.3. Modified Versions
|
||||||
|
|
||||||
|
If you create software not governed by this License, and you want to
|
||||||
|
create a new license for such software, you may create and use a
|
||||||
|
modified version of this License if you rename the license and remove
|
||||||
|
any references to the name of the license steward (except to note that
|
||||||
|
such modified license differs from this License).
|
||||||
|
|
||||||
|
10.4. Distributing Source Code Form that is Incompatible With Secondary
|
||||||
|
Licenses
|
||||||
|
|
||||||
|
If You choose to distribute Source Code Form that is Incompatible With
|
||||||
|
Secondary Licenses under the terms of this version of the License, the
|
||||||
|
notice described in Exhibit B of this License must be attached.
|
||||||
|
|
||||||
|
Exhibit A - Source Code Form License Notice
|
||||||
|
-------------------------------------------
|
||||||
|
|
||||||
|
This Source Code Form is subject to the terms of the Mozilla Public
|
||||||
|
License, v. 2.0. If a copy of the MPL was not distributed with this
|
||||||
|
file, You can obtain one at http://mozilla.org/MPL/2.0/.
|
||||||
|
|
||||||
|
If it is not possible or desirable to put the notice in a particular
|
||||||
|
file, then You may include the notice in a location (such as a LICENSE
|
||||||
|
file in a relevant directory) where a recipient would be likely to look
|
||||||
|
for such a notice.
|
||||||
|
|
||||||
|
You may add additional accurate notices of copyright ownership.
|
||||||
|
|
||||||
|
Exhibit B - "Incompatible With Secondary Licenses" Notice
|
||||||
|
---------------------------------------------------------
|
||||||
|
|
||||||
|
This Source Code Form is "Incompatible With Secondary Licenses", as
|
||||||
|
defined by the Mozilla Public License, v. 2.0.
|
|
@ -0,0 +1,118 @@
|
||||||
|
# go-sockaddr
|
||||||
|
|
||||||
|
## `sockaddr` Library
|
||||||
|
|
||||||
|
Socket address convenience functions for Go. `go-sockaddr` is a convenience
|
||||||
|
library that makes doing the right thing with IP addresses easy. `go-sockaddr`
|
||||||
|
is loosely modeled after the UNIX `sockaddr_t` and creates a union of the family
|
||||||
|
of `sockaddr_t` types (see below for an ascii diagram). Library documentation
|
||||||
|
is available
|
||||||
|
at
|
||||||
|
[https://godoc.org/github.com/hashicorp/go-sockaddr](https://godoc.org/github.com/hashicorp/go-sockaddr).
|
||||||
|
The primary intent of the library was to make it possible to define heuristics
|
||||||
|
for selecting the correct IP addresses when a configuration is evaluated at
|
||||||
|
runtime. See
|
||||||
|
the
|
||||||
|
[docs](https://godoc.org/github.com/hashicorp/go-sockaddr),
|
||||||
|
[`template` package](https://godoc.org/github.com/hashicorp/go-sockaddr/template),
|
||||||
|
tests,
|
||||||
|
and
|
||||||
|
[CLI utility](https://github.com/hashicorp/go-sockaddr/tree/master/cmd/sockaddr)
|
||||||
|
for details and hints as to how to use this library.
|
||||||
|
|
||||||
|
For example, with this library it is possible to find an IP address that:
|
||||||
|
|
||||||
|
* is attached to a default route
|
||||||
|
([`GetDefaultInterfaces()`](https://godoc.org/github.com/hashicorp/go-sockaddr#GetDefaultInterfaces))
|
||||||
|
* is contained within a CIDR block ([`IfByNetwork()`](https://godoc.org/github.com/hashicorp/go-sockaddr#IfByNetwork))
|
||||||
|
* is an RFC1918 address
|
||||||
|
([`IfByRFC("1918")`](https://godoc.org/github.com/hashicorp/go-sockaddr#IfByRFC))
|
||||||
|
* is ordered
|
||||||
|
([`OrderedIfAddrBy(args)`](https://godoc.org/github.com/hashicorp/go-sockaddr#OrderedIfAddrBy) where
|
||||||
|
`args` includes, but is not limited
|
||||||
|
to,
|
||||||
|
[`AscIfType`](https://godoc.org/github.com/hashicorp/go-sockaddr#AscIfType),
|
||||||
|
[`AscNetworkSize`](https://godoc.org/github.com/hashicorp/go-sockaddr#AscNetworkSize))
|
||||||
|
* excludes all IPv6 addresses
|
||||||
|
([`IfByType("^(IPv4)$")`](https://godoc.org/github.com/hashicorp/go-sockaddr#IfByType))
|
||||||
|
* is larger than a `/32`
|
||||||
|
([`IfByMaskSize(32)`](https://godoc.org/github.com/hashicorp/go-sockaddr#IfByMaskSize))
|
||||||
|
* is not on a `down` interface
|
||||||
|
([`ExcludeIfs("flags", "down")`](https://godoc.org/github.com/hashicorp/go-sockaddr#ExcludeIfs))
|
||||||
|
* preferences an IPv6 address over an IPv4 address
|
||||||
|
([`SortIfByType()`](https://godoc.org/github.com/hashicorp/go-sockaddr#SortIfByType) +
|
||||||
|
[`ReverseIfAddrs()`](https://godoc.org/github.com/hashicorp/go-sockaddr#ReverseIfAddrs)); and
|
||||||
|
* excludes any IP in RFC6890 address
|
||||||
|
([`IfByRFC("6890")`](https://godoc.org/github.com/hashicorp/go-sockaddr#IfByRFC))
|
||||||
|
|
||||||
|
Or any combination or variation therein.
|
||||||
|
|
||||||
|
There are also a few simple helper functions such as `GetPublicIP` and
|
||||||
|
`GetPrivateIP` which both return strings and select the first public or private
|
||||||
|
IP address on the default interface, respectively. Similarly, there is also a
|
||||||
|
helper function called `GetInterfaceIP` which returns the first usable IP
|
||||||
|
address on the named interface.
|
||||||
|
|
||||||
|
## `sockaddr` CLI
|
||||||
|
|
||||||
|
Given the possible complexity of the `sockaddr` library, there is a CLI utility
|
||||||
|
that accompanies the library, also
|
||||||
|
called
|
||||||
|
[`sockaddr`](https://github.com/hashicorp/go-sockaddr/tree/master/cmd/sockaddr).
|
||||||
|
The
|
||||||
|
[`sockaddr`](https://github.com/hashicorp/go-sockaddr/tree/master/cmd/sockaddr)
|
||||||
|
utility exposes nearly all of the functionality of the library and can be used
|
||||||
|
either as an administrative tool or testing tool. To install
|
||||||
|
the
|
||||||
|
[`sockaddr`](https://github.com/hashicorp/go-sockaddr/tree/master/cmd/sockaddr),
|
||||||
|
run:
|
||||||
|
|
||||||
|
```text
|
||||||
|
$ go get -u github.com/hashicorp/go-sockaddr/cmd/sockaddr
|
||||||
|
```
|
||||||
|
|
||||||
|
If you're familiar with UNIX's `sockaddr` struct's, the following diagram
|
||||||
|
mapping the C `sockaddr` (top) to `go-sockaddr` structs (bottom) and
|
||||||
|
interfaces will be helpful:
|
||||||
|
|
||||||
|
```
|
||||||
|
+-------------------------------------------------------+
|
||||||
|
| |
|
||||||
|
| sockaddr |
|
||||||
|
| SockAddr |
|
||||||
|
| |
|
||||||
|
| +--------------+ +----------------------------------+ |
|
||||||
|
| | sockaddr_un | | | |
|
||||||
|
| | SockAddrUnix | | sockaddr_in{,6} | |
|
||||||
|
| +--------------+ | IPAddr | |
|
||||||
|
| | | |
|
||||||
|
| | +-------------+ +--------------+ | |
|
||||||
|
| | | sockaddr_in | | sockaddr_in6 | | |
|
||||||
|
| | | IPv4Addr | | IPv6Addr | | |
|
||||||
|
| | +-------------+ +--------------+ | |
|
||||||
|
| | | |
|
||||||
|
| +----------------------------------+ |
|
||||||
|
| |
|
||||||
|
+-------------------------------------------------------+
|
||||||
|
```
|
||||||
|
|
||||||
|
## Inspiration and Design
|
||||||
|
|
||||||
|
There were many subtle inspirations that led to this design, but the most direct
|
||||||
|
inspiration for the filtering syntax was
|
||||||
|
OpenBSD's
|
||||||
|
[`pf.conf(5)`](https://www.freebsd.org/cgi/man.cgi?query=pf.conf&apropos=0&sektion=0&arch=default&format=html#PARAMETERS) firewall
|
||||||
|
syntax that lets you select the first IP address on a given named interface.
|
||||||
|
The original problem stemmed from:
|
||||||
|
|
||||||
|
* needing to create immutable images using [Packer](https://www.packer.io) that
|
||||||
|
ran the [Consul](https://www.consul.io) process (Consul can only use one IP
|
||||||
|
address at a time);
|
||||||
|
* images that may or may not have multiple interfaces or IP addresses at
|
||||||
|
runtime; and
|
||||||
|
* we didn't want to rely on configuration management to render out the correct
|
||||||
|
IP address if the VM image was being used in an auto-scaling group.
|
||||||
|
|
||||||
|
Instead we needed some way to codify a heuristic that would correctly select the
|
||||||
|
right IP address but the input parameters were not known when the image was
|
||||||
|
created.
|
|
@ -0,0 +1,5 @@
|
||||||
|
/*
|
||||||
|
Package sockaddr is a Go implementation of the UNIX socket family data types and
|
||||||
|
related helper functions.
|
||||||
|
*/
|
||||||
|
package sockaddr
|
|
@ -0,0 +1,254 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import "strings"
|
||||||
|
|
||||||
|
// ifAddrAttrMap is a map of the IfAddr type-specific attributes.
|
||||||
|
var ifAddrAttrMap map[AttrName]func(IfAddr) string
|
||||||
|
var ifAddrAttrs []AttrName
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
ifAddrAttrInit()
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetPrivateIP returns a string with a single IP address that is part of RFC
|
||||||
|
// 6890 and has a default route. If the system can't determine its IP address
|
||||||
|
// or find an RFC 6890 IP address, an empty string will be returned instead.
|
||||||
|
// This function is the `eval` equivalent of:
|
||||||
|
//
|
||||||
|
// ```
|
||||||
|
// $ sockaddr eval -r '{{GetPrivateInterfaces | attr "address"}}'
|
||||||
|
/// ```
|
||||||
|
func GetPrivateIP() (string, error) {
|
||||||
|
privateIfs, err := GetPrivateInterfaces()
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
if len(privateIfs) < 1 {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
ifAddr := privateIfs[0]
|
||||||
|
ip := *ToIPAddr(ifAddr.SockAddr)
|
||||||
|
return ip.NetIP().String(), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetPrivateIPs returns a string with all IP addresses that are part of RFC
|
||||||
|
// 6890 (regardless of whether or not there is a default route, unlike
|
||||||
|
// GetPublicIP). If the system can't find any RFC 6890 IP addresses, an empty
|
||||||
|
// string will be returned instead. This function is the `eval` equivalent of:
|
||||||
|
//
|
||||||
|
// ```
|
||||||
|
// $ sockaddr eval -r '{{GetAllInterfaces | include "RFC" "6890" | join "address" " "}}'
|
||||||
|
/// ```
|
||||||
|
func GetPrivateIPs() (string, error) {
|
||||||
|
ifAddrs, err := GetAllInterfaces()
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
} else if len(ifAddrs) < 1 {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
ifAddrs, _ = FilterIfByType(ifAddrs, TypeIP)
|
||||||
|
if len(ifAddrs) == 0 {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
OrderedIfAddrBy(AscIfType, AscIfNetworkSize).Sort(ifAddrs)
|
||||||
|
|
||||||
|
ifAddrs, _, err = IfByRFC("6890", ifAddrs)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
} else if len(ifAddrs) == 0 {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
_, ifAddrs, err = IfByRFC(ForwardingBlacklistRFC, ifAddrs)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
} else if len(ifAddrs) == 0 {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
ips := make([]string, 0, len(ifAddrs))
|
||||||
|
for _, ifAddr := range ifAddrs {
|
||||||
|
ip := *ToIPAddr(ifAddr.SockAddr)
|
||||||
|
s := ip.NetIP().String()
|
||||||
|
ips = append(ips, s)
|
||||||
|
}
|
||||||
|
|
||||||
|
return strings.Join(ips, " "), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetPublicIP returns a string with a single IP address that is NOT part of RFC
|
||||||
|
// 6890 and has a default route. If the system can't determine its IP address
|
||||||
|
// or find a non RFC 6890 IP address, an empty string will be returned instead.
|
||||||
|
// This function is the `eval` equivalent of:
|
||||||
|
//
|
||||||
|
// ```
|
||||||
|
// $ sockaddr eval -r '{{GetPublicInterfaces | attr "address"}}'
|
||||||
|
/// ```
|
||||||
|
func GetPublicIP() (string, error) {
|
||||||
|
publicIfs, err := GetPublicInterfaces()
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
} else if len(publicIfs) < 1 {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
ifAddr := publicIfs[0]
|
||||||
|
ip := *ToIPAddr(ifAddr.SockAddr)
|
||||||
|
return ip.NetIP().String(), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetPublicIPs returns a string with all IP addresses that are NOT part of RFC
|
||||||
|
// 6890 (regardless of whether or not there is a default route, unlike
|
||||||
|
// GetPublicIP). If the system can't find any non RFC 6890 IP addresses, an
|
||||||
|
// empty string will be returned instead. This function is the `eval`
|
||||||
|
// equivalent of:
|
||||||
|
//
|
||||||
|
// ```
|
||||||
|
// $ sockaddr eval -r '{{GetAllInterfaces | exclude "RFC" "6890" | join "address" " "}}'
|
||||||
|
/// ```
|
||||||
|
func GetPublicIPs() (string, error) {
|
||||||
|
ifAddrs, err := GetAllInterfaces()
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
} else if len(ifAddrs) < 1 {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
ifAddrs, _ = FilterIfByType(ifAddrs, TypeIP)
|
||||||
|
if len(ifAddrs) == 0 {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
OrderedIfAddrBy(AscIfType, AscIfNetworkSize).Sort(ifAddrs)
|
||||||
|
|
||||||
|
_, ifAddrs, err = IfByRFC("6890", ifAddrs)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
} else if len(ifAddrs) == 0 {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
ips := make([]string, 0, len(ifAddrs))
|
||||||
|
for _, ifAddr := range ifAddrs {
|
||||||
|
ip := *ToIPAddr(ifAddr.SockAddr)
|
||||||
|
s := ip.NetIP().String()
|
||||||
|
ips = append(ips, s)
|
||||||
|
}
|
||||||
|
|
||||||
|
return strings.Join(ips, " "), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetInterfaceIP returns a string with a single IP address sorted by the size
|
||||||
|
// of the network (i.e. IP addresses with a smaller netmask, larger network
|
||||||
|
// size, are sorted first). This function is the `eval` equivalent of:
|
||||||
|
//
|
||||||
|
// ```
|
||||||
|
// $ sockaddr eval -r '{{GetAllInterfaces | include "name" <<ARG>> | sort "type,size" | include "flag" "forwardable" | attr "address" }}'
|
||||||
|
/// ```
|
||||||
|
func GetInterfaceIP(namedIfRE string) (string, error) {
|
||||||
|
ifAddrs, err := GetAllInterfaces()
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
ifAddrs, _, err = IfByName(namedIfRE, ifAddrs)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
ifAddrs, _, err = IfByFlag("forwardable", ifAddrs)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
ifAddrs, err = SortIfBy("+type,+size", ifAddrs)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(ifAddrs) == 0 {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
ip := ToIPAddr(ifAddrs[0].SockAddr)
|
||||||
|
if ip == nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
return IPAddrAttr(*ip, "address"), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetInterfaceIPs returns a string with all IPs, sorted by the size of the
|
||||||
|
// network (i.e. IP addresses with a smaller netmask, larger network size, are
|
||||||
|
// sorted first), on a named interface. This function is the `eval` equivalent
|
||||||
|
// of:
|
||||||
|
//
|
||||||
|
// ```
|
||||||
|
// $ sockaddr eval -r '{{GetAllInterfaces | include "name" <<ARG>> | sort "type,size" | join "address" " "}}'
|
||||||
|
/// ```
|
||||||
|
func GetInterfaceIPs(namedIfRE string) (string, error) {
|
||||||
|
ifAddrs, err := GetAllInterfaces()
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
ifAddrs, _, err = IfByName(namedIfRE, ifAddrs)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
ifAddrs, err = SortIfBy("+type,+size", ifAddrs)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(ifAddrs) == 0 {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
ips := make([]string, 0, len(ifAddrs))
|
||||||
|
for _, ifAddr := range ifAddrs {
|
||||||
|
ip := *ToIPAddr(ifAddr.SockAddr)
|
||||||
|
s := ip.NetIP().String()
|
||||||
|
ips = append(ips, s)
|
||||||
|
}
|
||||||
|
|
||||||
|
return strings.Join(ips, " "), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// IfAddrAttrs returns a list of attributes supported by the IfAddr type
|
||||||
|
func IfAddrAttrs() []AttrName {
|
||||||
|
return ifAddrAttrs
|
||||||
|
}
|
||||||
|
|
||||||
|
// IfAddrAttr returns a string representation of an attribute for the given
|
||||||
|
// IfAddr.
|
||||||
|
func IfAddrAttr(ifAddr IfAddr, attrName AttrName) string {
|
||||||
|
fn, found := ifAddrAttrMap[attrName]
|
||||||
|
if !found {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
return fn(ifAddr)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ifAddrAttrInit is called once at init()
|
||||||
|
func ifAddrAttrInit() {
|
||||||
|
// Sorted for human readability
|
||||||
|
ifAddrAttrs = []AttrName{
|
||||||
|
"flags",
|
||||||
|
"name",
|
||||||
|
}
|
||||||
|
|
||||||
|
ifAddrAttrMap = map[AttrName]func(ifAddr IfAddr) string{
|
||||||
|
"flags": func(ifAddr IfAddr) string {
|
||||||
|
return ifAddr.Interface.Flags.String()
|
||||||
|
},
|
||||||
|
"name": func(ifAddr IfAddr) string {
|
||||||
|
return ifAddr.Interface.Name
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
File diff suppressed because it is too large
Load Diff
|
@ -0,0 +1,65 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"net"
|
||||||
|
)
|
||||||
|
|
||||||
|
// IfAddr is a union of a SockAddr and a net.Interface.
|
||||||
|
type IfAddr struct {
|
||||||
|
SockAddr
|
||||||
|
net.Interface
|
||||||
|
}
|
||||||
|
|
||||||
|
// Attr returns the named attribute as a string
|
||||||
|
func (ifAddr IfAddr) Attr(attrName AttrName) (string, error) {
|
||||||
|
val := IfAddrAttr(ifAddr, attrName)
|
||||||
|
if val != "" {
|
||||||
|
return val, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
return Attr(ifAddr.SockAddr, attrName)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Attr returns the named attribute as a string
|
||||||
|
func Attr(sa SockAddr, attrName AttrName) (string, error) {
|
||||||
|
switch sockType := sa.Type(); {
|
||||||
|
case sockType&TypeIP != 0:
|
||||||
|
ip := *ToIPAddr(sa)
|
||||||
|
attrVal := IPAddrAttr(ip, attrName)
|
||||||
|
if attrVal != "" {
|
||||||
|
return attrVal, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if sockType == TypeIPv4 {
|
||||||
|
ipv4 := *ToIPv4Addr(sa)
|
||||||
|
attrVal := IPv4AddrAttr(ipv4, attrName)
|
||||||
|
if attrVal != "" {
|
||||||
|
return attrVal, nil
|
||||||
|
}
|
||||||
|
} else if sockType == TypeIPv6 {
|
||||||
|
ipv6 := *ToIPv6Addr(sa)
|
||||||
|
attrVal := IPv6AddrAttr(ipv6, attrName)
|
||||||
|
if attrVal != "" {
|
||||||
|
return attrVal, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
case sockType == TypeUnix:
|
||||||
|
us := *ToUnixSock(sa)
|
||||||
|
attrVal := UnixSockAttr(us, attrName)
|
||||||
|
if attrVal != "" {
|
||||||
|
return attrVal, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Non type-specific attributes
|
||||||
|
switch attrName {
|
||||||
|
case "string":
|
||||||
|
return sa.String(), nil
|
||||||
|
case "type":
|
||||||
|
return sa.Type().String(), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
return "", fmt.Errorf("unsupported attribute name %q", attrName)
|
||||||
|
}
|
|
@ -0,0 +1,169 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"math/big"
|
||||||
|
"net"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Constants for the sizes of IPv3, IPv4, and IPv6 address types.
|
||||||
|
const (
|
||||||
|
IPv3len = 6
|
||||||
|
IPv4len = 4
|
||||||
|
IPv6len = 16
|
||||||
|
)
|
||||||
|
|
||||||
|
// IPAddr is a generic IP address interface for IPv4 and IPv6 addresses,
|
||||||
|
// networks, and socket endpoints.
|
||||||
|
type IPAddr interface {
|
||||||
|
SockAddr
|
||||||
|
AddressBinString() string
|
||||||
|
AddressHexString() string
|
||||||
|
Cmp(SockAddr) int
|
||||||
|
CmpAddress(SockAddr) int
|
||||||
|
CmpPort(SockAddr) int
|
||||||
|
FirstUsable() IPAddr
|
||||||
|
Host() IPAddr
|
||||||
|
IPPort() IPPort
|
||||||
|
LastUsable() IPAddr
|
||||||
|
Maskbits() int
|
||||||
|
NetIP() *net.IP
|
||||||
|
NetIPMask() *net.IPMask
|
||||||
|
NetIPNet() *net.IPNet
|
||||||
|
Network() IPAddr
|
||||||
|
Octets() []int
|
||||||
|
}
|
||||||
|
|
||||||
|
// IPPort is the type for an IP port number for the TCP and UDP IP transports.
|
||||||
|
type IPPort uint16
|
||||||
|
|
||||||
|
// IPPrefixLen is a typed integer representing the prefix length for a given
|
||||||
|
// IPAddr.
|
||||||
|
type IPPrefixLen byte
|
||||||
|
|
||||||
|
// ipAddrAttrMap is a map of the IPAddr type-specific attributes.
|
||||||
|
var ipAddrAttrMap map[AttrName]func(IPAddr) string
|
||||||
|
var ipAddrAttrs []AttrName
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
ipAddrInit()
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewIPAddr creates a new IPAddr from a string. Returns nil if the string is
|
||||||
|
// not an IPv4 or an IPv6 address.
|
||||||
|
func NewIPAddr(addr string) (IPAddr, error) {
|
||||||
|
ipv4Addr, err := NewIPv4Addr(addr)
|
||||||
|
if err == nil {
|
||||||
|
return ipv4Addr, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
ipv6Addr, err := NewIPv6Addr(addr)
|
||||||
|
if err == nil {
|
||||||
|
return ipv6Addr, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil, fmt.Errorf("invalid IPAddr %v", addr)
|
||||||
|
}
|
||||||
|
|
||||||
|
// IPAddrAttr returns a string representation of an attribute for the given
|
||||||
|
// IPAddr.
|
||||||
|
func IPAddrAttr(ip IPAddr, selector AttrName) string {
|
||||||
|
fn, found := ipAddrAttrMap[selector]
|
||||||
|
if !found {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
return fn(ip)
|
||||||
|
}
|
||||||
|
|
||||||
|
// IPAttrs returns a list of attributes supported by the IPAddr type
|
||||||
|
func IPAttrs() []AttrName {
|
||||||
|
return ipAddrAttrs
|
||||||
|
}
|
||||||
|
|
||||||
|
// MustIPAddr is a helper method that must return an IPAddr or panic on invalid
|
||||||
|
// input.
|
||||||
|
func MustIPAddr(addr string) IPAddr {
|
||||||
|
ip, err := NewIPAddr(addr)
|
||||||
|
if err != nil {
|
||||||
|
panic(fmt.Sprintf("Unable to create an IPAddr from %+q: %v", addr, err))
|
||||||
|
}
|
||||||
|
return ip
|
||||||
|
}
|
||||||
|
|
||||||
|
// ipAddrInit is called once at init()
|
||||||
|
func ipAddrInit() {
|
||||||
|
// Sorted for human readability
|
||||||
|
ipAddrAttrs = []AttrName{
|
||||||
|
"host",
|
||||||
|
"address",
|
||||||
|
"port",
|
||||||
|
"netmask",
|
||||||
|
"network",
|
||||||
|
"mask_bits",
|
||||||
|
"binary",
|
||||||
|
"hex",
|
||||||
|
"first_usable",
|
||||||
|
"last_usable",
|
||||||
|
"octets",
|
||||||
|
}
|
||||||
|
|
||||||
|
ipAddrAttrMap = map[AttrName]func(ip IPAddr) string{
|
||||||
|
"address": func(ip IPAddr) string {
|
||||||
|
return ip.NetIP().String()
|
||||||
|
},
|
||||||
|
"binary": func(ip IPAddr) string {
|
||||||
|
return ip.AddressBinString()
|
||||||
|
},
|
||||||
|
"first_usable": func(ip IPAddr) string {
|
||||||
|
return ip.FirstUsable().String()
|
||||||
|
},
|
||||||
|
"hex": func(ip IPAddr) string {
|
||||||
|
return ip.AddressHexString()
|
||||||
|
},
|
||||||
|
"host": func(ip IPAddr) string {
|
||||||
|
return ip.Host().String()
|
||||||
|
},
|
||||||
|
"last_usable": func(ip IPAddr) string {
|
||||||
|
return ip.LastUsable().String()
|
||||||
|
},
|
||||||
|
"mask_bits": func(ip IPAddr) string {
|
||||||
|
return fmt.Sprintf("%d", ip.Maskbits())
|
||||||
|
},
|
||||||
|
"netmask": func(ip IPAddr) string {
|
||||||
|
switch v := ip.(type) {
|
||||||
|
case IPv4Addr:
|
||||||
|
ipv4Mask := IPv4Addr{
|
||||||
|
Address: IPv4Address(v.Mask),
|
||||||
|
Mask: IPv4HostMask,
|
||||||
|
}
|
||||||
|
return ipv4Mask.String()
|
||||||
|
case IPv6Addr:
|
||||||
|
ipv6Mask := new(big.Int)
|
||||||
|
ipv6Mask.Set(v.Mask)
|
||||||
|
ipv6MaskAddr := IPv6Addr{
|
||||||
|
Address: IPv6Address(ipv6Mask),
|
||||||
|
Mask: ipv6HostMask,
|
||||||
|
}
|
||||||
|
return ipv6MaskAddr.String()
|
||||||
|
default:
|
||||||
|
return fmt.Sprintf("<unsupported type: %T>", ip)
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"network": func(ip IPAddr) string {
|
||||||
|
return ip.Network().NetIP().String()
|
||||||
|
},
|
||||||
|
"octets": func(ip IPAddr) string {
|
||||||
|
octets := ip.Octets()
|
||||||
|
octetStrs := make([]string, 0, len(octets))
|
||||||
|
for _, octet := range octets {
|
||||||
|
octetStrs = append(octetStrs, fmt.Sprintf("%d", octet))
|
||||||
|
}
|
||||||
|
return strings.Join(octetStrs, " ")
|
||||||
|
},
|
||||||
|
"port": func(ip IPAddr) string {
|
||||||
|
return fmt.Sprintf("%d", ip.IPPort())
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,98 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import "bytes"
|
||||||
|
|
||||||
|
type IPAddrs []IPAddr
|
||||||
|
|
||||||
|
func (s IPAddrs) Len() int { return len(s) }
|
||||||
|
func (s IPAddrs) Swap(i, j int) { s[i], s[j] = s[j], s[i] }
|
||||||
|
|
||||||
|
// // SortIPAddrsByCmp is a type that satisfies sort.Interface and can be used
|
||||||
|
// // by the routines in this package. The SortIPAddrsByCmp type is used to
|
||||||
|
// // sort IPAddrs by Cmp()
|
||||||
|
// type SortIPAddrsByCmp struct{ IPAddrs }
|
||||||
|
|
||||||
|
// // Less reports whether the element with index i should sort before the
|
||||||
|
// // element with index j.
|
||||||
|
// func (s SortIPAddrsByCmp) Less(i, j int) bool {
|
||||||
|
// // Sort by Type, then address, then port number.
|
||||||
|
// return Less(s.IPAddrs[i], s.IPAddrs[j])
|
||||||
|
// }
|
||||||
|
|
||||||
|
// SortIPAddrsBySpecificMaskLen is a type that satisfies sort.Interface and
|
||||||
|
// can be used by the routines in this package. The
|
||||||
|
// SortIPAddrsBySpecificMaskLen type is used to sort IPAddrs by smallest
|
||||||
|
// network (most specific to largest network).
|
||||||
|
type SortIPAddrsByNetworkSize struct{ IPAddrs }
|
||||||
|
|
||||||
|
// Less reports whether the element with index i should sort before the
|
||||||
|
// element with index j.
|
||||||
|
func (s SortIPAddrsByNetworkSize) Less(i, j int) bool {
|
||||||
|
// Sort masks with a larger binary value (i.e. fewer hosts per network
|
||||||
|
// prefix) after masks with a smaller value (larger number of hosts per
|
||||||
|
// prefix).
|
||||||
|
switch bytes.Compare([]byte(*s.IPAddrs[i].NetIPMask()), []byte(*s.IPAddrs[j].NetIPMask())) {
|
||||||
|
case 0:
|
||||||
|
// Fall through to the second test if the net.IPMasks are the
|
||||||
|
// same.
|
||||||
|
break
|
||||||
|
case 1:
|
||||||
|
return true
|
||||||
|
case -1:
|
||||||
|
return false
|
||||||
|
default:
|
||||||
|
panic("bad, m'kay?")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Sort IPs based on the length (i.e. prefer IPv4 over IPv6).
|
||||||
|
iLen := len(*s.IPAddrs[i].NetIP())
|
||||||
|
jLen := len(*s.IPAddrs[j].NetIP())
|
||||||
|
if iLen != jLen {
|
||||||
|
return iLen > jLen
|
||||||
|
}
|
||||||
|
|
||||||
|
// Sort IPs based on their network address from lowest to highest.
|
||||||
|
switch bytes.Compare(s.IPAddrs[i].NetIPNet().IP, s.IPAddrs[j].NetIPNet().IP) {
|
||||||
|
case 0:
|
||||||
|
break
|
||||||
|
case 1:
|
||||||
|
return false
|
||||||
|
case -1:
|
||||||
|
return true
|
||||||
|
default:
|
||||||
|
panic("lol wut?")
|
||||||
|
}
|
||||||
|
|
||||||
|
// If a host does not have a port set, it always sorts after hosts
|
||||||
|
// that have a port (e.g. a host with a /32 and port number is more
|
||||||
|
// specific and should sort first over a host with a /32 but no port
|
||||||
|
// set).
|
||||||
|
if s.IPAddrs[i].IPPort() == 0 || s.IPAddrs[j].IPPort() == 0 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
return s.IPAddrs[i].IPPort() < s.IPAddrs[j].IPPort()
|
||||||
|
}
|
||||||
|
|
||||||
|
// SortIPAddrsBySpecificMaskLen is a type that satisfies sort.Interface and
|
||||||
|
// can be used by the routines in this package. The
|
||||||
|
// SortIPAddrsBySpecificMaskLen type is used to sort IPAddrs by smallest
|
||||||
|
// network (most specific to largest network).
|
||||||
|
type SortIPAddrsBySpecificMaskLen struct{ IPAddrs }
|
||||||
|
|
||||||
|
// Less reports whether the element with index i should sort before the
|
||||||
|
// element with index j.
|
||||||
|
func (s SortIPAddrsBySpecificMaskLen) Less(i, j int) bool {
|
||||||
|
return s.IPAddrs[i].Maskbits() > s.IPAddrs[j].Maskbits()
|
||||||
|
}
|
||||||
|
|
||||||
|
// SortIPAddrsByBroadMaskLen is a type that satisfies sort.Interface and can
|
||||||
|
// be used by the routines in this package. The SortIPAddrsByBroadMaskLen
|
||||||
|
// type is used to sort IPAddrs by largest network (i.e. largest subnets
|
||||||
|
// first).
|
||||||
|
type SortIPAddrsByBroadMaskLen struct{ IPAddrs }
|
||||||
|
|
||||||
|
// Less reports whether the element with index i should sort before the
|
||||||
|
// element with index j.
|
||||||
|
func (s SortIPAddrsByBroadMaskLen) Less(i, j int) bool {
|
||||||
|
return s.IPAddrs[i].Maskbits() < s.IPAddrs[j].Maskbits()
|
||||||
|
}
|
|
@ -0,0 +1,516 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/binary"
|
||||||
|
"fmt"
|
||||||
|
"net"
|
||||||
|
"regexp"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
type (
|
||||||
|
// IPv4Address is a named type representing an IPv4 address.
|
||||||
|
IPv4Address uint32
|
||||||
|
|
||||||
|
// IPv4Network is a named type representing an IPv4 network.
|
||||||
|
IPv4Network uint32
|
||||||
|
|
||||||
|
// IPv4Mask is a named type representing an IPv4 network mask.
|
||||||
|
IPv4Mask uint32
|
||||||
|
)
|
||||||
|
|
||||||
|
// IPv4HostMask is a constant represents a /32 IPv4 Address
|
||||||
|
// (i.e. 255.255.255.255).
|
||||||
|
const IPv4HostMask = IPv4Mask(0xffffffff)
|
||||||
|
|
||||||
|
// ipv4AddrAttrMap is a map of the IPv4Addr type-specific attributes.
|
||||||
|
var ipv4AddrAttrMap map[AttrName]func(IPv4Addr) string
|
||||||
|
var ipv4AddrAttrs []AttrName
|
||||||
|
var trailingHexNetmaskRE *regexp.Regexp
|
||||||
|
|
||||||
|
// IPv4Addr implements a convenience wrapper around the union of Go's
|
||||||
|
// built-in net.IP and net.IPNet types. In UNIX-speak, IPv4Addr implements
|
||||||
|
// `sockaddr` when the the address family is set to AF_INET
|
||||||
|
// (i.e. `sockaddr_in`).
|
||||||
|
type IPv4Addr struct {
|
||||||
|
IPAddr
|
||||||
|
Address IPv4Address
|
||||||
|
Mask IPv4Mask
|
||||||
|
Port IPPort
|
||||||
|
}
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
ipv4AddrInit()
|
||||||
|
trailingHexNetmaskRE = regexp.MustCompile(`/([0f]{8})$`)
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewIPv4Addr creates an IPv4Addr from a string. String can be in the form
|
||||||
|
// of either an IPv4:port (e.g. `1.2.3.4:80`, in which case the mask is
|
||||||
|
// assumed to be a `/32`), an IPv4 address (e.g. `1.2.3.4`, also with a `/32`
|
||||||
|
// mask), or an IPv4 CIDR (e.g. `1.2.3.4/24`, which has its IP port
|
||||||
|
// initialized to zero). ipv4Str can not be a hostname.
|
||||||
|
//
|
||||||
|
// NOTE: Many net.*() routines will initialize and return an IPv6 address.
|
||||||
|
// To create uint32 values from net.IP, always test to make sure the address
|
||||||
|
// returned can be converted to a 4 byte array using To4().
|
||||||
|
func NewIPv4Addr(ipv4Str string) (IPv4Addr, error) {
|
||||||
|
// Strip off any bogus hex-encoded netmasks that will be mis-parsed by Go. In
|
||||||
|
// particular, clients with the Barracuda VPN client will see something like:
|
||||||
|
// `192.168.3.51/00ffffff` as their IP address.
|
||||||
|
trailingHexNetmaskRe := trailingHexNetmaskRE.Copy()
|
||||||
|
if match := trailingHexNetmaskRe.FindStringIndex(ipv4Str); match != nil {
|
||||||
|
ipv4Str = ipv4Str[:match[0]]
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parse as an IPv4 CIDR
|
||||||
|
ipAddr, network, err := net.ParseCIDR(ipv4Str)
|
||||||
|
if err == nil {
|
||||||
|
ipv4 := ipAddr.To4()
|
||||||
|
if ipv4 == nil {
|
||||||
|
return IPv4Addr{}, fmt.Errorf("Unable to convert %s to an IPv4 address", ipv4Str)
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we see an IPv6 netmask, convert it to an IPv4 mask.
|
||||||
|
netmaskSepPos := strings.LastIndexByte(ipv4Str, '/')
|
||||||
|
if netmaskSepPos != -1 && netmaskSepPos+1 < len(ipv4Str) {
|
||||||
|
netMask, err := strconv.ParseUint(ipv4Str[netmaskSepPos+1:], 10, 8)
|
||||||
|
if err != nil {
|
||||||
|
return IPv4Addr{}, fmt.Errorf("Unable to convert %s to an IPv4 address: unable to parse CIDR netmask: %v", ipv4Str, err)
|
||||||
|
} else if netMask > 128 {
|
||||||
|
return IPv4Addr{}, fmt.Errorf("Unable to convert %s to an IPv4 address: invalid CIDR netmask", ipv4Str)
|
||||||
|
}
|
||||||
|
|
||||||
|
if netMask >= 96 {
|
||||||
|
// Convert the IPv6 netmask to an IPv4 netmask
|
||||||
|
network.Mask = net.CIDRMask(int(netMask-96), IPv4len*8)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
ipv4Addr := IPv4Addr{
|
||||||
|
Address: IPv4Address(binary.BigEndian.Uint32(ipv4)),
|
||||||
|
Mask: IPv4Mask(binary.BigEndian.Uint32(network.Mask)),
|
||||||
|
}
|
||||||
|
return ipv4Addr, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Attempt to parse ipv4Str as a /32 host with a port number.
|
||||||
|
tcpAddr, err := net.ResolveTCPAddr("tcp4", ipv4Str)
|
||||||
|
if err == nil {
|
||||||
|
ipv4 := tcpAddr.IP.To4()
|
||||||
|
if ipv4 == nil {
|
||||||
|
return IPv4Addr{}, fmt.Errorf("Unable to resolve %+q as an IPv4 address", ipv4Str)
|
||||||
|
}
|
||||||
|
|
||||||
|
ipv4Uint32 := binary.BigEndian.Uint32(ipv4)
|
||||||
|
ipv4Addr := IPv4Addr{
|
||||||
|
Address: IPv4Address(ipv4Uint32),
|
||||||
|
Mask: IPv4HostMask,
|
||||||
|
Port: IPPort(tcpAddr.Port),
|
||||||
|
}
|
||||||
|
|
||||||
|
return ipv4Addr, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parse as a naked IPv4 address
|
||||||
|
ip := net.ParseIP(ipv4Str)
|
||||||
|
if ip != nil {
|
||||||
|
ipv4 := ip.To4()
|
||||||
|
if ipv4 == nil {
|
||||||
|
return IPv4Addr{}, fmt.Errorf("Unable to string convert %+q to an IPv4 address", ipv4Str)
|
||||||
|
}
|
||||||
|
|
||||||
|
ipv4Uint32 := binary.BigEndian.Uint32(ipv4)
|
||||||
|
ipv4Addr := IPv4Addr{
|
||||||
|
Address: IPv4Address(ipv4Uint32),
|
||||||
|
Mask: IPv4HostMask,
|
||||||
|
}
|
||||||
|
return ipv4Addr, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
return IPv4Addr{}, fmt.Errorf("Unable to parse %+q to an IPv4 address: %v", ipv4Str, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// AddressBinString returns a string with the IPv4Addr's Address represented
|
||||||
|
// as a sequence of '0' and '1' characters. This method is useful for
|
||||||
|
// debugging or by operators who want to inspect an address.
|
||||||
|
func (ipv4 IPv4Addr) AddressBinString() string {
|
||||||
|
return fmt.Sprintf("%032s", strconv.FormatUint(uint64(ipv4.Address), 2))
|
||||||
|
}
|
||||||
|
|
||||||
|
// AddressHexString returns a string with the IPv4Addr address represented as
|
||||||
|
// a sequence of hex characters. This method is useful for debugging or by
|
||||||
|
// operators who want to inspect an address.
|
||||||
|
func (ipv4 IPv4Addr) AddressHexString() string {
|
||||||
|
return fmt.Sprintf("%08s", strconv.FormatUint(uint64(ipv4.Address), 16))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Broadcast is an IPv4Addr-only method that returns the broadcast address of
|
||||||
|
// the network.
|
||||||
|
//
|
||||||
|
// NOTE: IPv6 only supports multicast, so this method only exists for
|
||||||
|
// IPv4Addr.
|
||||||
|
func (ipv4 IPv4Addr) Broadcast() IPAddr {
|
||||||
|
// Nothing should listen on a broadcast address.
|
||||||
|
return IPv4Addr{
|
||||||
|
Address: IPv4Address(ipv4.BroadcastAddress()),
|
||||||
|
Mask: IPv4HostMask,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// BroadcastAddress returns a IPv4Network of the IPv4Addr's broadcast
|
||||||
|
// address.
|
||||||
|
func (ipv4 IPv4Addr) BroadcastAddress() IPv4Network {
|
||||||
|
return IPv4Network(uint32(ipv4.Address)&uint32(ipv4.Mask) | ^uint32(ipv4.Mask))
|
||||||
|
}
|
||||||
|
|
||||||
|
// CmpAddress follows the Cmp() standard protocol and returns:
|
||||||
|
//
|
||||||
|
// - -1 If the receiver should sort first because its address is lower than arg
|
||||||
|
// - 0 if the SockAddr arg is equal to the receiving IPv4Addr or the argument is
|
||||||
|
// of a different type.
|
||||||
|
// - 1 If the argument should sort first.
|
||||||
|
func (ipv4 IPv4Addr) CmpAddress(sa SockAddr) int {
|
||||||
|
ipv4b, ok := sa.(IPv4Addr)
|
||||||
|
if !ok {
|
||||||
|
return sortDeferDecision
|
||||||
|
}
|
||||||
|
|
||||||
|
switch {
|
||||||
|
case ipv4.Address == ipv4b.Address:
|
||||||
|
return sortDeferDecision
|
||||||
|
case ipv4.Address < ipv4b.Address:
|
||||||
|
return sortReceiverBeforeArg
|
||||||
|
default:
|
||||||
|
return sortArgBeforeReceiver
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// CmpPort follows the Cmp() standard protocol and returns:
|
||||||
|
//
|
||||||
|
// - -1 If the receiver should sort first because its port is lower than arg
|
||||||
|
// - 0 if the SockAddr arg's port number is equal to the receiving IPv4Addr,
|
||||||
|
// regardless of type.
|
||||||
|
// - 1 If the argument should sort first.
|
||||||
|
func (ipv4 IPv4Addr) CmpPort(sa SockAddr) int {
|
||||||
|
var saPort IPPort
|
||||||
|
switch v := sa.(type) {
|
||||||
|
case IPv4Addr:
|
||||||
|
saPort = v.Port
|
||||||
|
case IPv6Addr:
|
||||||
|
saPort = v.Port
|
||||||
|
default:
|
||||||
|
return sortDeferDecision
|
||||||
|
}
|
||||||
|
|
||||||
|
switch {
|
||||||
|
case ipv4.Port == saPort:
|
||||||
|
return sortDeferDecision
|
||||||
|
case ipv4.Port < saPort:
|
||||||
|
return sortReceiverBeforeArg
|
||||||
|
default:
|
||||||
|
return sortArgBeforeReceiver
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// CmpRFC follows the Cmp() standard protocol and returns:
|
||||||
|
//
|
||||||
|
// - -1 If the receiver should sort first because it belongs to the RFC and its
|
||||||
|
// arg does not
|
||||||
|
// - 0 if the receiver and arg both belong to the same RFC or neither do.
|
||||||
|
// - 1 If the arg belongs to the RFC but receiver does not.
|
||||||
|
func (ipv4 IPv4Addr) CmpRFC(rfcNum uint, sa SockAddr) int {
|
||||||
|
recvInRFC := IsRFC(rfcNum, ipv4)
|
||||||
|
ipv4b, ok := sa.(IPv4Addr)
|
||||||
|
if !ok {
|
||||||
|
// If the receiver is part of the desired RFC and the SockAddr
|
||||||
|
// argument is not, return -1 so that the receiver sorts before
|
||||||
|
// the non-IPv4 SockAddr. Conversely, if the receiver is not
|
||||||
|
// part of the RFC, punt on sorting and leave it for the next
|
||||||
|
// sorter.
|
||||||
|
if recvInRFC {
|
||||||
|
return sortReceiverBeforeArg
|
||||||
|
} else {
|
||||||
|
return sortDeferDecision
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
argInRFC := IsRFC(rfcNum, ipv4b)
|
||||||
|
switch {
|
||||||
|
case (recvInRFC && argInRFC), (!recvInRFC && !argInRFC):
|
||||||
|
// If a and b both belong to the RFC, or neither belong to
|
||||||
|
// rfcNum, defer sorting to the next sorter.
|
||||||
|
return sortDeferDecision
|
||||||
|
case recvInRFC && !argInRFC:
|
||||||
|
return sortReceiverBeforeArg
|
||||||
|
default:
|
||||||
|
return sortArgBeforeReceiver
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Contains returns true if the SockAddr is contained within the receiver.
|
||||||
|
func (ipv4 IPv4Addr) Contains(sa SockAddr) bool {
|
||||||
|
ipv4b, ok := sa.(IPv4Addr)
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
return ipv4.ContainsNetwork(ipv4b)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ContainsAddress returns true if the IPv4Address is contained within the
|
||||||
|
// receiver.
|
||||||
|
func (ipv4 IPv4Addr) ContainsAddress(x IPv4Address) bool {
|
||||||
|
return IPv4Address(ipv4.NetworkAddress()) <= x &&
|
||||||
|
IPv4Address(ipv4.BroadcastAddress()) >= x
|
||||||
|
}
|
||||||
|
|
||||||
|
// ContainsNetwork returns true if the network from IPv4Addr is contained
|
||||||
|
// within the receiver.
|
||||||
|
func (ipv4 IPv4Addr) ContainsNetwork(x IPv4Addr) bool {
|
||||||
|
return ipv4.NetworkAddress() <= x.NetworkAddress() &&
|
||||||
|
ipv4.BroadcastAddress() >= x.BroadcastAddress()
|
||||||
|
}
|
||||||
|
|
||||||
|
// DialPacketArgs returns the arguments required to be passed to
|
||||||
|
// net.DialUDP(). If the Mask of ipv4 is not a /32 or the Port is 0,
|
||||||
|
// DialPacketArgs() will fail. See Host() to create an IPv4Addr with its
|
||||||
|
// mask set to /32.
|
||||||
|
func (ipv4 IPv4Addr) DialPacketArgs() (network, dialArgs string) {
|
||||||
|
if ipv4.Mask != IPv4HostMask || ipv4.Port == 0 {
|
||||||
|
return "udp4", ""
|
||||||
|
}
|
||||||
|
return "udp4", fmt.Sprintf("%s:%d", ipv4.NetIP().String(), ipv4.Port)
|
||||||
|
}
|
||||||
|
|
||||||
|
// DialStreamArgs returns the arguments required to be passed to
|
||||||
|
// net.DialTCP(). If the Mask of ipv4 is not a /32 or the Port is 0,
|
||||||
|
// DialStreamArgs() will fail. See Host() to create an IPv4Addr with its
|
||||||
|
// mask set to /32.
|
||||||
|
func (ipv4 IPv4Addr) DialStreamArgs() (network, dialArgs string) {
|
||||||
|
if ipv4.Mask != IPv4HostMask || ipv4.Port == 0 {
|
||||||
|
return "tcp4", ""
|
||||||
|
}
|
||||||
|
return "tcp4", fmt.Sprintf("%s:%d", ipv4.NetIP().String(), ipv4.Port)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Equal returns true if a SockAddr is equal to the receiving IPv4Addr.
|
||||||
|
func (ipv4 IPv4Addr) Equal(sa SockAddr) bool {
|
||||||
|
ipv4b, ok := sa.(IPv4Addr)
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
if ipv4.Port != ipv4b.Port {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
if ipv4.Address != ipv4b.Address {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
if ipv4.NetIPNet().String() != ipv4b.NetIPNet().String() {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// FirstUsable returns an IPv4Addr set to the first address following the
|
||||||
|
// network prefix. The first usable address in a network is normally the
|
||||||
|
// gateway and should not be used except by devices forwarding packets
|
||||||
|
// between two administratively distinct networks (i.e. a router). This
|
||||||
|
// function does not discriminate against first usable vs "first address that
|
||||||
|
// should be used." For example, FirstUsable() on "192.168.1.10/24" would
|
||||||
|
// return the address "192.168.1.1/24".
|
||||||
|
func (ipv4 IPv4Addr) FirstUsable() IPAddr {
|
||||||
|
addr := ipv4.NetworkAddress()
|
||||||
|
|
||||||
|
// If /32, return the address itself. If /31 assume a point-to-point
|
||||||
|
// link and return the lower address.
|
||||||
|
if ipv4.Maskbits() < 31 {
|
||||||
|
addr++
|
||||||
|
}
|
||||||
|
|
||||||
|
return IPv4Addr{
|
||||||
|
Address: IPv4Address(addr),
|
||||||
|
Mask: IPv4HostMask,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Host returns a copy of ipv4 with its mask set to /32 so that it can be
|
||||||
|
// used by DialPacketArgs(), DialStreamArgs(), ListenPacketArgs(), or
|
||||||
|
// ListenStreamArgs().
|
||||||
|
func (ipv4 IPv4Addr) Host() IPAddr {
|
||||||
|
// Nothing should listen on a broadcast address.
|
||||||
|
return IPv4Addr{
|
||||||
|
Address: ipv4.Address,
|
||||||
|
Mask: IPv4HostMask,
|
||||||
|
Port: ipv4.Port,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// IPPort returns the Port number attached to the IPv4Addr
|
||||||
|
func (ipv4 IPv4Addr) IPPort() IPPort {
|
||||||
|
return ipv4.Port
|
||||||
|
}
|
||||||
|
|
||||||
|
// LastUsable returns the last address before the broadcast address in a
|
||||||
|
// given network.
|
||||||
|
func (ipv4 IPv4Addr) LastUsable() IPAddr {
|
||||||
|
addr := ipv4.BroadcastAddress()
|
||||||
|
|
||||||
|
// If /32, return the address itself. If /31 assume a point-to-point
|
||||||
|
// link and return the upper address.
|
||||||
|
if ipv4.Maskbits() < 31 {
|
||||||
|
addr--
|
||||||
|
}
|
||||||
|
|
||||||
|
return IPv4Addr{
|
||||||
|
Address: IPv4Address(addr),
|
||||||
|
Mask: IPv4HostMask,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// ListenPacketArgs returns the arguments required to be passed to
|
||||||
|
// net.ListenUDP(). If the Mask of ipv4 is not a /32, ListenPacketArgs()
|
||||||
|
// will fail. See Host() to create an IPv4Addr with its mask set to /32.
|
||||||
|
func (ipv4 IPv4Addr) ListenPacketArgs() (network, listenArgs string) {
|
||||||
|
if ipv4.Mask != IPv4HostMask {
|
||||||
|
return "udp4", ""
|
||||||
|
}
|
||||||
|
return "udp4", fmt.Sprintf("%s:%d", ipv4.NetIP().String(), ipv4.Port)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ListenStreamArgs returns the arguments required to be passed to
|
||||||
|
// net.ListenTCP(). If the Mask of ipv4 is not a /32, ListenStreamArgs()
|
||||||
|
// will fail. See Host() to create an IPv4Addr with its mask set to /32.
|
||||||
|
func (ipv4 IPv4Addr) ListenStreamArgs() (network, listenArgs string) {
|
||||||
|
if ipv4.Mask != IPv4HostMask {
|
||||||
|
return "tcp4", ""
|
||||||
|
}
|
||||||
|
return "tcp4", fmt.Sprintf("%s:%d", ipv4.NetIP().String(), ipv4.Port)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Maskbits returns the number of network mask bits in a given IPv4Addr. For
|
||||||
|
// example, the Maskbits() of "192.168.1.1/24" would return 24.
|
||||||
|
func (ipv4 IPv4Addr) Maskbits() int {
|
||||||
|
mask := make(net.IPMask, IPv4len)
|
||||||
|
binary.BigEndian.PutUint32(mask, uint32(ipv4.Mask))
|
||||||
|
maskOnes, _ := mask.Size()
|
||||||
|
return maskOnes
|
||||||
|
}
|
||||||
|
|
||||||
|
// MustIPv4Addr is a helper method that must return an IPv4Addr or panic on
|
||||||
|
// invalid input.
|
||||||
|
func MustIPv4Addr(addr string) IPv4Addr {
|
||||||
|
ipv4, err := NewIPv4Addr(addr)
|
||||||
|
if err != nil {
|
||||||
|
panic(fmt.Sprintf("Unable to create an IPv4Addr from %+q: %v", addr, err))
|
||||||
|
}
|
||||||
|
return ipv4
|
||||||
|
}
|
||||||
|
|
||||||
|
// NetIP returns the address as a net.IP (address is always presized to
|
||||||
|
// IPv4).
|
||||||
|
func (ipv4 IPv4Addr) NetIP() *net.IP {
|
||||||
|
x := make(net.IP, IPv4len)
|
||||||
|
binary.BigEndian.PutUint32(x, uint32(ipv4.Address))
|
||||||
|
return &x
|
||||||
|
}
|
||||||
|
|
||||||
|
// NetIPMask create a new net.IPMask from the IPv4Addr.
|
||||||
|
func (ipv4 IPv4Addr) NetIPMask() *net.IPMask {
|
||||||
|
ipv4Mask := net.IPMask{}
|
||||||
|
ipv4Mask = make(net.IPMask, IPv4len)
|
||||||
|
binary.BigEndian.PutUint32(ipv4Mask, uint32(ipv4.Mask))
|
||||||
|
return &ipv4Mask
|
||||||
|
}
|
||||||
|
|
||||||
|
// NetIPNet create a new net.IPNet from the IPv4Addr.
|
||||||
|
func (ipv4 IPv4Addr) NetIPNet() *net.IPNet {
|
||||||
|
ipv4net := &net.IPNet{}
|
||||||
|
ipv4net.IP = make(net.IP, IPv4len)
|
||||||
|
binary.BigEndian.PutUint32(ipv4net.IP, uint32(ipv4.NetworkAddress()))
|
||||||
|
ipv4net.Mask = *ipv4.NetIPMask()
|
||||||
|
return ipv4net
|
||||||
|
}
|
||||||
|
|
||||||
|
// Network returns the network prefix or network address for a given network.
|
||||||
|
func (ipv4 IPv4Addr) Network() IPAddr {
|
||||||
|
return IPv4Addr{
|
||||||
|
Address: IPv4Address(ipv4.NetworkAddress()),
|
||||||
|
Mask: ipv4.Mask,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// NetworkAddress returns an IPv4Network of the IPv4Addr's network address.
|
||||||
|
func (ipv4 IPv4Addr) NetworkAddress() IPv4Network {
|
||||||
|
return IPv4Network(uint32(ipv4.Address) & uint32(ipv4.Mask))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Octets returns a slice of the four octets in an IPv4Addr's Address. The
|
||||||
|
// order of the bytes is big endian.
|
||||||
|
func (ipv4 IPv4Addr) Octets() []int {
|
||||||
|
return []int{
|
||||||
|
int(ipv4.Address >> 24),
|
||||||
|
int((ipv4.Address >> 16) & 0xff),
|
||||||
|
int((ipv4.Address >> 8) & 0xff),
|
||||||
|
int(ipv4.Address & 0xff),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// String returns a string representation of the IPv4Addr
|
||||||
|
func (ipv4 IPv4Addr) String() string {
|
||||||
|
if ipv4.Port != 0 {
|
||||||
|
return fmt.Sprintf("%s:%d", ipv4.NetIP().String(), ipv4.Port)
|
||||||
|
}
|
||||||
|
|
||||||
|
if ipv4.Maskbits() == 32 {
|
||||||
|
return ipv4.NetIP().String()
|
||||||
|
}
|
||||||
|
|
||||||
|
return fmt.Sprintf("%s/%d", ipv4.NetIP().String(), ipv4.Maskbits())
|
||||||
|
}
|
||||||
|
|
||||||
|
// Type is used as a type switch and returns TypeIPv4
|
||||||
|
func (IPv4Addr) Type() SockAddrType {
|
||||||
|
return TypeIPv4
|
||||||
|
}
|
||||||
|
|
||||||
|
// IPv4AddrAttr returns a string representation of an attribute for the given
|
||||||
|
// IPv4Addr.
|
||||||
|
func IPv4AddrAttr(ipv4 IPv4Addr, selector AttrName) string {
|
||||||
|
fn, found := ipv4AddrAttrMap[selector]
|
||||||
|
if !found {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
return fn(ipv4)
|
||||||
|
}
|
||||||
|
|
||||||
|
// IPv4Attrs returns a list of attributes supported by the IPv4Addr type
|
||||||
|
func IPv4Attrs() []AttrName {
|
||||||
|
return ipv4AddrAttrs
|
||||||
|
}
|
||||||
|
|
||||||
|
// ipv4AddrInit is called once at init()
|
||||||
|
func ipv4AddrInit() {
|
||||||
|
// Sorted for human readability
|
||||||
|
ipv4AddrAttrs = []AttrName{
|
||||||
|
"size", // Same position as in IPv6 for output consistency
|
||||||
|
"broadcast",
|
||||||
|
"uint32",
|
||||||
|
}
|
||||||
|
|
||||||
|
ipv4AddrAttrMap = map[AttrName]func(ipv4 IPv4Addr) string{
|
||||||
|
"broadcast": func(ipv4 IPv4Addr) string {
|
||||||
|
return ipv4.Broadcast().String()
|
||||||
|
},
|
||||||
|
"size": func(ipv4 IPv4Addr) string {
|
||||||
|
return fmt.Sprintf("%d", 1<<uint(IPv4len*8-ipv4.Maskbits()))
|
||||||
|
},
|
||||||
|
"uint32": func(ipv4 IPv4Addr) string {
|
||||||
|
return fmt.Sprintf("%d", uint32(ipv4.Address))
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,591 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"encoding/binary"
|
||||||
|
"fmt"
|
||||||
|
"math/big"
|
||||||
|
"net"
|
||||||
|
)
|
||||||
|
|
||||||
|
type (
|
||||||
|
// IPv6Address is a named type representing an IPv6 address.
|
||||||
|
IPv6Address *big.Int
|
||||||
|
|
||||||
|
// IPv6Network is a named type representing an IPv6 network.
|
||||||
|
IPv6Network *big.Int
|
||||||
|
|
||||||
|
// IPv6Mask is a named type representing an IPv6 network mask.
|
||||||
|
IPv6Mask *big.Int
|
||||||
|
)
|
||||||
|
|
||||||
|
// IPv6HostPrefix is a constant represents a /128 IPv6 Prefix.
|
||||||
|
const IPv6HostPrefix = IPPrefixLen(128)
|
||||||
|
|
||||||
|
// ipv6HostMask is an unexported big.Int representing a /128 IPv6 address.
|
||||||
|
// This value must be a constant and always set to all ones.
|
||||||
|
var ipv6HostMask IPv6Mask
|
||||||
|
|
||||||
|
// ipv6AddrAttrMap is a map of the IPv6Addr type-specific attributes.
|
||||||
|
var ipv6AddrAttrMap map[AttrName]func(IPv6Addr) string
|
||||||
|
var ipv6AddrAttrs []AttrName
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
biMask := new(big.Int)
|
||||||
|
biMask.SetBytes([]byte{
|
||||||
|
0xff, 0xff,
|
||||||
|
0xff, 0xff,
|
||||||
|
0xff, 0xff,
|
||||||
|
0xff, 0xff,
|
||||||
|
0xff, 0xff,
|
||||||
|
0xff, 0xff,
|
||||||
|
0xff, 0xff,
|
||||||
|
0xff, 0xff,
|
||||||
|
},
|
||||||
|
)
|
||||||
|
ipv6HostMask = IPv6Mask(biMask)
|
||||||
|
|
||||||
|
ipv6AddrInit()
|
||||||
|
}
|
||||||
|
|
||||||
|
// IPv6Addr implements a convenience wrapper around the union of Go's
|
||||||
|
// built-in net.IP and net.IPNet types. In UNIX-speak, IPv6Addr implements
|
||||||
|
// `sockaddr` when the the address family is set to AF_INET6
|
||||||
|
// (i.e. `sockaddr_in6`).
|
||||||
|
type IPv6Addr struct {
|
||||||
|
IPAddr
|
||||||
|
Address IPv6Address
|
||||||
|
Mask IPv6Mask
|
||||||
|
Port IPPort
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewIPv6Addr creates an IPv6Addr from a string. String can be in the form of
|
||||||
|
// an an IPv6:port (e.g. `[2001:4860:0:2001::68]:80`, in which case the mask is
|
||||||
|
// assumed to be a /128), an IPv6 address (e.g. `2001:4860:0:2001::68`, also
|
||||||
|
// with a `/128` mask), an IPv6 CIDR (e.g. `2001:4860:0:2001::68/64`, which has
|
||||||
|
// its IP port initialized to zero). ipv6Str can not be a hostname.
|
||||||
|
//
|
||||||
|
// NOTE: Many net.*() routines will initialize and return an IPv4 address.
|
||||||
|
// Always test to make sure the address returned cannot be converted to a 4 byte
|
||||||
|
// array using To4().
|
||||||
|
func NewIPv6Addr(ipv6Str string) (IPv6Addr, error) {
|
||||||
|
v6Addr := false
|
||||||
|
LOOP:
|
||||||
|
for i := 0; i < len(ipv6Str); i++ {
|
||||||
|
switch ipv6Str[i] {
|
||||||
|
case '.':
|
||||||
|
break LOOP
|
||||||
|
case ':':
|
||||||
|
v6Addr = true
|
||||||
|
break LOOP
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if !v6Addr {
|
||||||
|
return IPv6Addr{}, fmt.Errorf("Unable to resolve %+q as an IPv6 address, appears to be an IPv4 address", ipv6Str)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Attempt to parse ipv6Str as a /128 host with a port number.
|
||||||
|
tcpAddr, err := net.ResolveTCPAddr("tcp6", ipv6Str)
|
||||||
|
if err == nil {
|
||||||
|
ipv6 := tcpAddr.IP.To16()
|
||||||
|
if ipv6 == nil {
|
||||||
|
return IPv6Addr{}, fmt.Errorf("Unable to resolve %+q as a 16byte IPv6 address", ipv6Str)
|
||||||
|
}
|
||||||
|
|
||||||
|
ipv6BigIntAddr := new(big.Int)
|
||||||
|
ipv6BigIntAddr.SetBytes(ipv6)
|
||||||
|
|
||||||
|
ipv6BigIntMask := new(big.Int)
|
||||||
|
ipv6BigIntMask.Set(ipv6HostMask)
|
||||||
|
|
||||||
|
ipv6Addr := IPv6Addr{
|
||||||
|
Address: IPv6Address(ipv6BigIntAddr),
|
||||||
|
Mask: IPv6Mask(ipv6BigIntMask),
|
||||||
|
Port: IPPort(tcpAddr.Port),
|
||||||
|
}
|
||||||
|
|
||||||
|
return ipv6Addr, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parse as a naked IPv6 address. Trim square brackets if present.
|
||||||
|
if len(ipv6Str) > 2 && ipv6Str[0] == '[' && ipv6Str[len(ipv6Str)-1] == ']' {
|
||||||
|
ipv6Str = ipv6Str[1 : len(ipv6Str)-1]
|
||||||
|
}
|
||||||
|
ip := net.ParseIP(ipv6Str)
|
||||||
|
if ip != nil {
|
||||||
|
ipv6 := ip.To16()
|
||||||
|
if ipv6 == nil {
|
||||||
|
return IPv6Addr{}, fmt.Errorf("Unable to string convert %+q to a 16byte IPv6 address", ipv6Str)
|
||||||
|
}
|
||||||
|
|
||||||
|
ipv6BigIntAddr := new(big.Int)
|
||||||
|
ipv6BigIntAddr.SetBytes(ipv6)
|
||||||
|
|
||||||
|
ipv6BigIntMask := new(big.Int)
|
||||||
|
ipv6BigIntMask.Set(ipv6HostMask)
|
||||||
|
|
||||||
|
return IPv6Addr{
|
||||||
|
Address: IPv6Address(ipv6BigIntAddr),
|
||||||
|
Mask: IPv6Mask(ipv6BigIntMask),
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parse as an IPv6 CIDR
|
||||||
|
ipAddr, network, err := net.ParseCIDR(ipv6Str)
|
||||||
|
if err == nil {
|
||||||
|
ipv6 := ipAddr.To16()
|
||||||
|
if ipv6 == nil {
|
||||||
|
return IPv6Addr{}, fmt.Errorf("Unable to convert %+q to a 16byte IPv6 address", ipv6Str)
|
||||||
|
}
|
||||||
|
|
||||||
|
ipv6BigIntAddr := new(big.Int)
|
||||||
|
ipv6BigIntAddr.SetBytes(ipv6)
|
||||||
|
|
||||||
|
ipv6BigIntMask := new(big.Int)
|
||||||
|
ipv6BigIntMask.SetBytes(network.Mask)
|
||||||
|
|
||||||
|
ipv6Addr := IPv6Addr{
|
||||||
|
Address: IPv6Address(ipv6BigIntAddr),
|
||||||
|
Mask: IPv6Mask(ipv6BigIntMask),
|
||||||
|
}
|
||||||
|
return ipv6Addr, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
return IPv6Addr{}, fmt.Errorf("Unable to parse %+q to an IPv6 address: %v", ipv6Str, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// AddressBinString returns a string with the IPv6Addr's Address represented
|
||||||
|
// as a sequence of '0' and '1' characters. This method is useful for
|
||||||
|
// debugging or by operators who want to inspect an address.
|
||||||
|
func (ipv6 IPv6Addr) AddressBinString() string {
|
||||||
|
bi := big.Int(*ipv6.Address)
|
||||||
|
return fmt.Sprintf("%0128s", bi.Text(2))
|
||||||
|
}
|
||||||
|
|
||||||
|
// AddressHexString returns a string with the IPv6Addr address represented as
|
||||||
|
// a sequence of hex characters. This method is useful for debugging or by
|
||||||
|
// operators who want to inspect an address.
|
||||||
|
func (ipv6 IPv6Addr) AddressHexString() string {
|
||||||
|
bi := big.Int(*ipv6.Address)
|
||||||
|
return fmt.Sprintf("%032s", bi.Text(16))
|
||||||
|
}
|
||||||
|
|
||||||
|
// CmpAddress follows the Cmp() standard protocol and returns:
|
||||||
|
//
|
||||||
|
// - -1 If the receiver should sort first because its address is lower than arg
|
||||||
|
// - 0 if the SockAddr arg equal to the receiving IPv6Addr or the argument is of a
|
||||||
|
// different type.
|
||||||
|
// - 1 If the argument should sort first.
|
||||||
|
func (ipv6 IPv6Addr) CmpAddress(sa SockAddr) int {
|
||||||
|
ipv6b, ok := sa.(IPv6Addr)
|
||||||
|
if !ok {
|
||||||
|
return sortDeferDecision
|
||||||
|
}
|
||||||
|
|
||||||
|
ipv6aBigInt := new(big.Int)
|
||||||
|
ipv6aBigInt.Set(ipv6.Address)
|
||||||
|
ipv6bBigInt := new(big.Int)
|
||||||
|
ipv6bBigInt.Set(ipv6b.Address)
|
||||||
|
|
||||||
|
return ipv6aBigInt.Cmp(ipv6bBigInt)
|
||||||
|
}
|
||||||
|
|
||||||
|
// CmpPort follows the Cmp() standard protocol and returns:
|
||||||
|
//
|
||||||
|
// - -1 If the receiver should sort first because its port is lower than arg
|
||||||
|
// - 0 if the SockAddr arg's port number is equal to the receiving IPv6Addr,
|
||||||
|
// regardless of type.
|
||||||
|
// - 1 If the argument should sort first.
|
||||||
|
func (ipv6 IPv6Addr) CmpPort(sa SockAddr) int {
|
||||||
|
var saPort IPPort
|
||||||
|
switch v := sa.(type) {
|
||||||
|
case IPv4Addr:
|
||||||
|
saPort = v.Port
|
||||||
|
case IPv6Addr:
|
||||||
|
saPort = v.Port
|
||||||
|
default:
|
||||||
|
return sortDeferDecision
|
||||||
|
}
|
||||||
|
|
||||||
|
switch {
|
||||||
|
case ipv6.Port == saPort:
|
||||||
|
return sortDeferDecision
|
||||||
|
case ipv6.Port < saPort:
|
||||||
|
return sortReceiverBeforeArg
|
||||||
|
default:
|
||||||
|
return sortArgBeforeReceiver
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// CmpRFC follows the Cmp() standard protocol and returns:
|
||||||
|
//
|
||||||
|
// - -1 If the receiver should sort first because it belongs to the RFC and its
|
||||||
|
// arg does not
|
||||||
|
// - 0 if the receiver and arg both belong to the same RFC or neither do.
|
||||||
|
// - 1 If the arg belongs to the RFC but receiver does not.
|
||||||
|
func (ipv6 IPv6Addr) CmpRFC(rfcNum uint, sa SockAddr) int {
|
||||||
|
recvInRFC := IsRFC(rfcNum, ipv6)
|
||||||
|
ipv6b, ok := sa.(IPv6Addr)
|
||||||
|
if !ok {
|
||||||
|
// If the receiver is part of the desired RFC and the SockAddr
|
||||||
|
// argument is not, sort receiver before the non-IPv6 SockAddr.
|
||||||
|
// Conversely, if the receiver is not part of the RFC, punt on
|
||||||
|
// sorting and leave it for the next sorter.
|
||||||
|
if recvInRFC {
|
||||||
|
return sortReceiverBeforeArg
|
||||||
|
} else {
|
||||||
|
return sortDeferDecision
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
argInRFC := IsRFC(rfcNum, ipv6b)
|
||||||
|
switch {
|
||||||
|
case (recvInRFC && argInRFC), (!recvInRFC && !argInRFC):
|
||||||
|
// If a and b both belong to the RFC, or neither belong to
|
||||||
|
// rfcNum, defer sorting to the next sorter.
|
||||||
|
return sortDeferDecision
|
||||||
|
case recvInRFC && !argInRFC:
|
||||||
|
return sortReceiverBeforeArg
|
||||||
|
default:
|
||||||
|
return sortArgBeforeReceiver
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Contains returns true if the SockAddr is contained within the receiver.
|
||||||
|
func (ipv6 IPv6Addr) Contains(sa SockAddr) bool {
|
||||||
|
ipv6b, ok := sa.(IPv6Addr)
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
return ipv6.ContainsNetwork(ipv6b)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ContainsAddress returns true if the IPv6Address is contained within the
|
||||||
|
// receiver.
|
||||||
|
func (ipv6 IPv6Addr) ContainsAddress(x IPv6Address) bool {
|
||||||
|
xAddr := IPv6Addr{
|
||||||
|
Address: x,
|
||||||
|
Mask: ipv6HostMask,
|
||||||
|
}
|
||||||
|
|
||||||
|
{
|
||||||
|
xIPv6 := xAddr.FirstUsable().(IPv6Addr)
|
||||||
|
yIPv6 := ipv6.FirstUsable().(IPv6Addr)
|
||||||
|
if xIPv6.CmpAddress(yIPv6) >= 1 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
{
|
||||||
|
xIPv6 := xAddr.LastUsable().(IPv6Addr)
|
||||||
|
yIPv6 := ipv6.LastUsable().(IPv6Addr)
|
||||||
|
if xIPv6.CmpAddress(yIPv6) <= -1 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// ContainsNetwork returns true if the network from IPv6Addr is contained within
|
||||||
|
// the receiver.
|
||||||
|
func (x IPv6Addr) ContainsNetwork(y IPv6Addr) bool {
|
||||||
|
{
|
||||||
|
xIPv6 := x.FirstUsable().(IPv6Addr)
|
||||||
|
yIPv6 := y.FirstUsable().(IPv6Addr)
|
||||||
|
if ret := xIPv6.CmpAddress(yIPv6); ret >= 1 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
{
|
||||||
|
xIPv6 := x.LastUsable().(IPv6Addr)
|
||||||
|
yIPv6 := y.LastUsable().(IPv6Addr)
|
||||||
|
if ret := xIPv6.CmpAddress(yIPv6); ret <= -1 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// DialPacketArgs returns the arguments required to be passed to
|
||||||
|
// net.DialUDP(). If the Mask of ipv6 is not a /128 or the Port is 0,
|
||||||
|
// DialPacketArgs() will fail. See Host() to create an IPv6Addr with its
|
||||||
|
// mask set to /128.
|
||||||
|
func (ipv6 IPv6Addr) DialPacketArgs() (network, dialArgs string) {
|
||||||
|
ipv6Mask := big.Int(*ipv6.Mask)
|
||||||
|
if ipv6Mask.Cmp(ipv6HostMask) != 0 || ipv6.Port == 0 {
|
||||||
|
return "udp6", ""
|
||||||
|
}
|
||||||
|
return "udp6", fmt.Sprintf("[%s]:%d", ipv6.NetIP().String(), ipv6.Port)
|
||||||
|
}
|
||||||
|
|
||||||
|
// DialStreamArgs returns the arguments required to be passed to
|
||||||
|
// net.DialTCP(). If the Mask of ipv6 is not a /128 or the Port is 0,
|
||||||
|
// DialStreamArgs() will fail. See Host() to create an IPv6Addr with its
|
||||||
|
// mask set to /128.
|
||||||
|
func (ipv6 IPv6Addr) DialStreamArgs() (network, dialArgs string) {
|
||||||
|
ipv6Mask := big.Int(*ipv6.Mask)
|
||||||
|
if ipv6Mask.Cmp(ipv6HostMask) != 0 || ipv6.Port == 0 {
|
||||||
|
return "tcp6", ""
|
||||||
|
}
|
||||||
|
return "tcp6", fmt.Sprintf("[%s]:%d", ipv6.NetIP().String(), ipv6.Port)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Equal returns true if a SockAddr is equal to the receiving IPv4Addr.
|
||||||
|
func (ipv6a IPv6Addr) Equal(sa SockAddr) bool {
|
||||||
|
ipv6b, ok := sa.(IPv6Addr)
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
if ipv6a.NetIP().String() != ipv6b.NetIP().String() {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
if ipv6a.NetIPNet().String() != ipv6b.NetIPNet().String() {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
if ipv6a.Port != ipv6b.Port {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// FirstUsable returns an IPv6Addr set to the first address following the
|
||||||
|
// network prefix. The first usable address in a network is normally the
|
||||||
|
// gateway and should not be used except by devices forwarding packets
|
||||||
|
// between two administratively distinct networks (i.e. a router). This
|
||||||
|
// function does not discriminate against first usable vs "first address that
|
||||||
|
// should be used." For example, FirstUsable() on "2001:0db8::0003/64" would
|
||||||
|
// return "2001:0db8::00011".
|
||||||
|
func (ipv6 IPv6Addr) FirstUsable() IPAddr {
|
||||||
|
return IPv6Addr{
|
||||||
|
Address: IPv6Address(ipv6.NetworkAddress()),
|
||||||
|
Mask: ipv6HostMask,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Host returns a copy of ipv6 with its mask set to /128 so that it can be
|
||||||
|
// used by DialPacketArgs(), DialStreamArgs(), ListenPacketArgs(), or
|
||||||
|
// ListenStreamArgs().
|
||||||
|
func (ipv6 IPv6Addr) Host() IPAddr {
|
||||||
|
// Nothing should listen on a broadcast address.
|
||||||
|
return IPv6Addr{
|
||||||
|
Address: ipv6.Address,
|
||||||
|
Mask: ipv6HostMask,
|
||||||
|
Port: ipv6.Port,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// IPPort returns the Port number attached to the IPv6Addr
|
||||||
|
func (ipv6 IPv6Addr) IPPort() IPPort {
|
||||||
|
return ipv6.Port
|
||||||
|
}
|
||||||
|
|
||||||
|
// LastUsable returns the last address in a given network.
|
||||||
|
func (ipv6 IPv6Addr) LastUsable() IPAddr {
|
||||||
|
addr := new(big.Int)
|
||||||
|
addr.Set(ipv6.Address)
|
||||||
|
|
||||||
|
mask := new(big.Int)
|
||||||
|
mask.Set(ipv6.Mask)
|
||||||
|
|
||||||
|
negMask := new(big.Int)
|
||||||
|
negMask.Xor(ipv6HostMask, mask)
|
||||||
|
|
||||||
|
lastAddr := new(big.Int)
|
||||||
|
lastAddr.And(addr, mask)
|
||||||
|
lastAddr.Or(lastAddr, negMask)
|
||||||
|
|
||||||
|
return IPv6Addr{
|
||||||
|
Address: IPv6Address(lastAddr),
|
||||||
|
Mask: ipv6HostMask,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// ListenPacketArgs returns the arguments required to be passed to
|
||||||
|
// net.ListenUDP(). If the Mask of ipv6 is not a /128, ListenPacketArgs()
|
||||||
|
// will fail. See Host() to create an IPv6Addr with its mask set to /128.
|
||||||
|
func (ipv6 IPv6Addr) ListenPacketArgs() (network, listenArgs string) {
|
||||||
|
ipv6Mask := big.Int(*ipv6.Mask)
|
||||||
|
if ipv6Mask.Cmp(ipv6HostMask) != 0 {
|
||||||
|
return "udp6", ""
|
||||||
|
}
|
||||||
|
return "udp6", fmt.Sprintf("[%s]:%d", ipv6.NetIP().String(), ipv6.Port)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ListenStreamArgs returns the arguments required to be passed to
|
||||||
|
// net.ListenTCP(). If the Mask of ipv6 is not a /128, ListenStreamArgs()
|
||||||
|
// will fail. See Host() to create an IPv6Addr with its mask set to /128.
|
||||||
|
func (ipv6 IPv6Addr) ListenStreamArgs() (network, listenArgs string) {
|
||||||
|
ipv6Mask := big.Int(*ipv6.Mask)
|
||||||
|
if ipv6Mask.Cmp(ipv6HostMask) != 0 {
|
||||||
|
return "tcp6", ""
|
||||||
|
}
|
||||||
|
return "tcp6", fmt.Sprintf("[%s]:%d", ipv6.NetIP().String(), ipv6.Port)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Maskbits returns the number of network mask bits in a given IPv6Addr. For
|
||||||
|
// example, the Maskbits() of "2001:0db8::0003/64" would return 64.
|
||||||
|
func (ipv6 IPv6Addr) Maskbits() int {
|
||||||
|
maskOnes, _ := ipv6.NetIPNet().Mask.Size()
|
||||||
|
|
||||||
|
return maskOnes
|
||||||
|
}
|
||||||
|
|
||||||
|
// MustIPv6Addr is a helper method that must return an IPv6Addr or panic on
|
||||||
|
// invalid input.
|
||||||
|
func MustIPv6Addr(addr string) IPv6Addr {
|
||||||
|
ipv6, err := NewIPv6Addr(addr)
|
||||||
|
if err != nil {
|
||||||
|
panic(fmt.Sprintf("Unable to create an IPv6Addr from %+q: %v", addr, err))
|
||||||
|
}
|
||||||
|
return ipv6
|
||||||
|
}
|
||||||
|
|
||||||
|
// NetIP returns the address as a net.IP.
|
||||||
|
func (ipv6 IPv6Addr) NetIP() *net.IP {
|
||||||
|
return bigIntToNetIPv6(ipv6.Address)
|
||||||
|
}
|
||||||
|
|
||||||
|
// NetIPMask create a new net.IPMask from the IPv6Addr.
|
||||||
|
func (ipv6 IPv6Addr) NetIPMask() *net.IPMask {
|
||||||
|
ipv6Mask := make(net.IPMask, IPv6len)
|
||||||
|
m := big.Int(*ipv6.Mask)
|
||||||
|
copy(ipv6Mask, m.Bytes())
|
||||||
|
return &ipv6Mask
|
||||||
|
}
|
||||||
|
|
||||||
|
// Network returns a pointer to the net.IPNet within IPv4Addr receiver.
|
||||||
|
func (ipv6 IPv6Addr) NetIPNet() *net.IPNet {
|
||||||
|
ipv6net := &net.IPNet{}
|
||||||
|
ipv6net.IP = make(net.IP, IPv6len)
|
||||||
|
copy(ipv6net.IP, *ipv6.NetIP())
|
||||||
|
ipv6net.Mask = *ipv6.NetIPMask()
|
||||||
|
return ipv6net
|
||||||
|
}
|
||||||
|
|
||||||
|
// Network returns the network prefix or network address for a given network.
|
||||||
|
func (ipv6 IPv6Addr) Network() IPAddr {
|
||||||
|
return IPv6Addr{
|
||||||
|
Address: IPv6Address(ipv6.NetworkAddress()),
|
||||||
|
Mask: ipv6.Mask,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// NetworkAddress returns an IPv6Network of the IPv6Addr's network address.
|
||||||
|
func (ipv6 IPv6Addr) NetworkAddress() IPv6Network {
|
||||||
|
addr := new(big.Int)
|
||||||
|
addr.SetBytes((*ipv6.Address).Bytes())
|
||||||
|
|
||||||
|
mask := new(big.Int)
|
||||||
|
mask.SetBytes(*ipv6.NetIPMask())
|
||||||
|
|
||||||
|
netAddr := new(big.Int)
|
||||||
|
netAddr.And(addr, mask)
|
||||||
|
|
||||||
|
return IPv6Network(netAddr)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Octets returns a slice of the 16 octets in an IPv6Addr's Address. The
|
||||||
|
// order of the bytes is big endian.
|
||||||
|
func (ipv6 IPv6Addr) Octets() []int {
|
||||||
|
x := make([]int, IPv6len)
|
||||||
|
for i, b := range *bigIntToNetIPv6(ipv6.Address) {
|
||||||
|
x[i] = int(b)
|
||||||
|
}
|
||||||
|
|
||||||
|
return x
|
||||||
|
}
|
||||||
|
|
||||||
|
// String returns a string representation of the IPv6Addr
|
||||||
|
func (ipv6 IPv6Addr) String() string {
|
||||||
|
if ipv6.Port != 0 {
|
||||||
|
return fmt.Sprintf("[%s]:%d", ipv6.NetIP().String(), ipv6.Port)
|
||||||
|
}
|
||||||
|
|
||||||
|
if ipv6.Maskbits() == 128 {
|
||||||
|
return ipv6.NetIP().String()
|
||||||
|
}
|
||||||
|
|
||||||
|
return fmt.Sprintf("%s/%d", ipv6.NetIP().String(), ipv6.Maskbits())
|
||||||
|
}
|
||||||
|
|
||||||
|
// Type is used as a type switch and returns TypeIPv6
|
||||||
|
func (IPv6Addr) Type() SockAddrType {
|
||||||
|
return TypeIPv6
|
||||||
|
}
|
||||||
|
|
||||||
|
// IPv6Attrs returns a list of attributes supported by the IPv6Addr type
|
||||||
|
func IPv6Attrs() []AttrName {
|
||||||
|
return ipv6AddrAttrs
|
||||||
|
}
|
||||||
|
|
||||||
|
// IPv6AddrAttr returns a string representation of an attribute for the given
|
||||||
|
// IPv6Addr.
|
||||||
|
func IPv6AddrAttr(ipv6 IPv6Addr, selector AttrName) string {
|
||||||
|
fn, found := ipv6AddrAttrMap[selector]
|
||||||
|
if !found {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
return fn(ipv6)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ipv6AddrInit is called once at init()
|
||||||
|
func ipv6AddrInit() {
|
||||||
|
// Sorted for human readability
|
||||||
|
ipv6AddrAttrs = []AttrName{
|
||||||
|
"size", // Same position as in IPv6 for output consistency
|
||||||
|
"uint128",
|
||||||
|
}
|
||||||
|
|
||||||
|
ipv6AddrAttrMap = map[AttrName]func(ipv6 IPv6Addr) string{
|
||||||
|
"size": func(ipv6 IPv6Addr) string {
|
||||||
|
netSize := big.NewInt(1)
|
||||||
|
netSize = netSize.Lsh(netSize, uint(IPv6len*8-ipv6.Maskbits()))
|
||||||
|
return netSize.Text(10)
|
||||||
|
},
|
||||||
|
"uint128": func(ipv6 IPv6Addr) string {
|
||||||
|
b := big.Int(*ipv6.Address)
|
||||||
|
return b.Text(10)
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// bigIntToNetIPv6 is a helper function that correctly returns a net.IP with the
|
||||||
|
// correctly padded values.
|
||||||
|
func bigIntToNetIPv6(bi *big.Int) *net.IP {
|
||||||
|
x := make(net.IP, IPv6len)
|
||||||
|
ipv6Bytes := bi.Bytes()
|
||||||
|
|
||||||
|
// It's possibe for ipv6Bytes to be less than IPv6len bytes in size. If
|
||||||
|
// they are different sizes we to pad the size of response.
|
||||||
|
if len(ipv6Bytes) < IPv6len {
|
||||||
|
buf := new(bytes.Buffer)
|
||||||
|
buf.Grow(IPv6len)
|
||||||
|
|
||||||
|
for i := len(ipv6Bytes); i < IPv6len; i++ {
|
||||||
|
if err := binary.Write(buf, binary.BigEndian, byte(0)); err != nil {
|
||||||
|
panic(fmt.Sprintf("Unable to pad byte %d of input %v: %v", i, bi, err))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, b := range ipv6Bytes {
|
||||||
|
if err := binary.Write(buf, binary.BigEndian, b); err != nil {
|
||||||
|
panic(fmt.Sprintf("Unable to preserve endianness of input %v: %v", bi, err))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
ipv6Bytes = buf.Bytes()
|
||||||
|
}
|
||||||
|
i := copy(x, ipv6Bytes)
|
||||||
|
if i != IPv6len {
|
||||||
|
panic("IPv6 wrong size")
|
||||||
|
}
|
||||||
|
return &x
|
||||||
|
}
|
|
@ -0,0 +1,948 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
// ForwardingBlacklist is a faux RFC that includes a list of non-forwardable IP
|
||||||
|
// blocks.
|
||||||
|
const ForwardingBlacklist = 4294967295
|
||||||
|
const ForwardingBlacklistRFC = "4294967295"
|
||||||
|
|
||||||
|
// IsRFC tests to see if an SockAddr matches the specified RFC
|
||||||
|
func IsRFC(rfcNum uint, sa SockAddr) bool {
|
||||||
|
rfcNetMap := KnownRFCs()
|
||||||
|
rfcNets, ok := rfcNetMap[rfcNum]
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
var contained bool
|
||||||
|
for _, rfcNet := range rfcNets {
|
||||||
|
if rfcNet.Contains(sa) {
|
||||||
|
contained = true
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return contained
|
||||||
|
}
|
||||||
|
|
||||||
|
// KnownRFCs returns an initial set of known RFCs.
|
||||||
|
//
|
||||||
|
// NOTE (sean@): As this list evolves over time, please submit patches to keep
|
||||||
|
// this list current. If something isn't right, inquire, as it may just be a
|
||||||
|
// bug on my part. Some of the inclusions were based on my judgement as to what
|
||||||
|
// would be a useful value (e.g. RFC3330).
|
||||||
|
//
|
||||||
|
// Useful resources:
|
||||||
|
//
|
||||||
|
// * https://www.iana.org/assignments/ipv6-address-space/ipv6-address-space.xhtml
|
||||||
|
// * https://www.iana.org/assignments/ipv6-unicast-address-assignments/ipv6-unicast-address-assignments.xhtml
|
||||||
|
// * https://www.iana.org/assignments/ipv6-address-space/ipv6-address-space.xhtml
|
||||||
|
func KnownRFCs() map[uint]SockAddrs {
|
||||||
|
// NOTE(sean@): Multiple SockAddrs per RFC lend themselves well to a
|
||||||
|
// RADIX tree, but `ENOTIME`. Patches welcome.
|
||||||
|
return map[uint]SockAddrs{
|
||||||
|
919: {
|
||||||
|
// [RFC919] Broadcasting Internet Datagrams
|
||||||
|
MustIPv4Addr("255.255.255.255/32"), // [RFC1122], §7 Broadcast IP Addressing - Proposed Standards
|
||||||
|
},
|
||||||
|
1122: {
|
||||||
|
// [RFC1122] Requirements for Internet Hosts -- Communication Layers
|
||||||
|
MustIPv4Addr("0.0.0.0/8"), // [RFC1122], §3.2.1.3
|
||||||
|
MustIPv4Addr("127.0.0.0/8"), // [RFC1122], §3.2.1.3
|
||||||
|
},
|
||||||
|
1112: {
|
||||||
|
// [RFC1112] Host Extensions for IP Multicasting
|
||||||
|
MustIPv4Addr("224.0.0.0/4"), // [RFC1112], §4 Host Group Addresses
|
||||||
|
},
|
||||||
|
1918: {
|
||||||
|
// [RFC1918] Address Allocation for Private Internets
|
||||||
|
MustIPv4Addr("10.0.0.0/8"),
|
||||||
|
MustIPv4Addr("172.16.0.0/12"),
|
||||||
|
MustIPv4Addr("192.168.0.0/16"),
|
||||||
|
},
|
||||||
|
2544: {
|
||||||
|
// [RFC2544] Benchmarking Methodology for Network
|
||||||
|
// Interconnect Devices
|
||||||
|
MustIPv4Addr("198.18.0.0/15"),
|
||||||
|
},
|
||||||
|
2765: {
|
||||||
|
// [RFC2765] Stateless IP/ICMP Translation Algorithm
|
||||||
|
// (SIIT) (obsoleted by RFCs 6145, which itself was
|
||||||
|
// later obsoleted by 7915).
|
||||||
|
|
||||||
|
// [RFC2765], §2.1 Addresses
|
||||||
|
MustIPv6Addr("0:0:0:0:0:ffff:0:0/96"),
|
||||||
|
},
|
||||||
|
2928: {
|
||||||
|
// [RFC2928] Initial IPv6 Sub-TLA ID Assignments
|
||||||
|
MustIPv6Addr("2001::/16"), // Superblock
|
||||||
|
//MustIPv6Addr("2001:0000::/23"), // IANA
|
||||||
|
//MustIPv6Addr("2001:0200::/23"), // APNIC
|
||||||
|
//MustIPv6Addr("2001:0400::/23"), // ARIN
|
||||||
|
//MustIPv6Addr("2001:0600::/23"), // RIPE NCC
|
||||||
|
//MustIPv6Addr("2001:0800::/23"), // (future assignment)
|
||||||
|
// ...
|
||||||
|
//MustIPv6Addr("2001:FE00::/23"), // (future assignment)
|
||||||
|
},
|
||||||
|
3056: { // 6to4 address
|
||||||
|
// [RFC3056] Connection of IPv6 Domains via IPv4 Clouds
|
||||||
|
|
||||||
|
// [RFC3056], §2 IPv6 Prefix Allocation
|
||||||
|
MustIPv6Addr("2002::/16"),
|
||||||
|
},
|
||||||
|
3068: {
|
||||||
|
// [RFC3068] An Anycast Prefix for 6to4 Relay Routers
|
||||||
|
// (obsolete by RFC7526)
|
||||||
|
|
||||||
|
// [RFC3068], § 6to4 Relay anycast address
|
||||||
|
MustIPv4Addr("192.88.99.0/24"),
|
||||||
|
|
||||||
|
// [RFC3068], §2.5 6to4 IPv6 relay anycast address
|
||||||
|
//
|
||||||
|
// NOTE: /120 == 128-(32-24)
|
||||||
|
MustIPv6Addr("2002:c058:6301::/120"),
|
||||||
|
},
|
||||||
|
3171: {
|
||||||
|
// [RFC3171] IANA Guidelines for IPv4 Multicast Address Assignments
|
||||||
|
MustIPv4Addr("224.0.0.0/4"),
|
||||||
|
},
|
||||||
|
3330: {
|
||||||
|
// [RFC3330] Special-Use IPv4 Addresses
|
||||||
|
|
||||||
|
// Addresses in this block refer to source hosts on
|
||||||
|
// "this" network. Address 0.0.0.0/32 may be used as a
|
||||||
|
// source address for this host on this network; other
|
||||||
|
// addresses within 0.0.0.0/8 may be used to refer to
|
||||||
|
// specified hosts on this network [RFC1700, page 4].
|
||||||
|
MustIPv4Addr("0.0.0.0/8"),
|
||||||
|
|
||||||
|
// 10.0.0.0/8 - This block is set aside for use in
|
||||||
|
// private networks. Its intended use is documented in
|
||||||
|
// [RFC1918]. Addresses within this block should not
|
||||||
|
// appear on the public Internet.
|
||||||
|
MustIPv4Addr("10.0.0.0/8"),
|
||||||
|
|
||||||
|
// 14.0.0.0/8 - This block is set aside for assignments
|
||||||
|
// to the international system of Public Data Networks
|
||||||
|
// [RFC1700, page 181]. The registry of assignments
|
||||||
|
// within this block can be accessed from the "Public
|
||||||
|
// Data Network Numbers" link on the web page at
|
||||||
|
// http://www.iana.org/numbers.html. Addresses within
|
||||||
|
// this block are assigned to users and should be
|
||||||
|
// treated as such.
|
||||||
|
|
||||||
|
// 24.0.0.0/8 - This block was allocated in early 1996
|
||||||
|
// for use in provisioning IP service over cable
|
||||||
|
// television systems. Although the IANA initially was
|
||||||
|
// involved in making assignments to cable operators,
|
||||||
|
// this responsibility was transferred to American
|
||||||
|
// Registry for Internet Numbers (ARIN) in May 2001.
|
||||||
|
// Addresses within this block are assigned in the
|
||||||
|
// normal manner and should be treated as such.
|
||||||
|
|
||||||
|
// 39.0.0.0/8 - This block was used in the "Class A
|
||||||
|
// Subnet Experiment" that commenced in May 1995, as
|
||||||
|
// documented in [RFC1797]. The experiment has been
|
||||||
|
// completed and this block has been returned to the
|
||||||
|
// pool of addresses reserved for future allocation or
|
||||||
|
// assignment. This block therefore no longer has a
|
||||||
|
// special use and is subject to allocation to a
|
||||||
|
// Regional Internet Registry for assignment in the
|
||||||
|
// normal manner.
|
||||||
|
|
||||||
|
// 127.0.0.0/8 - This block is assigned for use as the Internet host
|
||||||
|
// loopback address. A datagram sent by a higher level protocol to an
|
||||||
|
// address anywhere within this block should loop back inside the host.
|
||||||
|
// This is ordinarily implemented using only 127.0.0.1/32 for loopback,
|
||||||
|
// but no addresses within this block should ever appear on any network
|
||||||
|
// anywhere [RFC1700, page 5].
|
||||||
|
MustIPv4Addr("127.0.0.0/8"),
|
||||||
|
|
||||||
|
// 128.0.0.0/16 - This block, corresponding to the
|
||||||
|
// numerically lowest of the former Class B addresses,
|
||||||
|
// was initially and is still reserved by the IANA.
|
||||||
|
// Given the present classless nature of the IP address
|
||||||
|
// space, the basis for the reservation no longer
|
||||||
|
// applies and addresses in this block are subject to
|
||||||
|
// future allocation to a Regional Internet Registry for
|
||||||
|
// assignment in the normal manner.
|
||||||
|
|
||||||
|
// 169.254.0.0/16 - This is the "link local" block. It
|
||||||
|
// is allocated for communication between hosts on a
|
||||||
|
// single link. Hosts obtain these addresses by
|
||||||
|
// auto-configuration, such as when a DHCP server may
|
||||||
|
// not be found.
|
||||||
|
MustIPv4Addr("169.254.0.0/16"),
|
||||||
|
|
||||||
|
// 172.16.0.0/12 - This block is set aside for use in
|
||||||
|
// private networks. Its intended use is documented in
|
||||||
|
// [RFC1918]. Addresses within this block should not
|
||||||
|
// appear on the public Internet.
|
||||||
|
MustIPv4Addr("172.16.0.0/12"),
|
||||||
|
|
||||||
|
// 191.255.0.0/16 - This block, corresponding to the numerically highest
|
||||||
|
// to the former Class B addresses, was initially and is still reserved
|
||||||
|
// by the IANA. Given the present classless nature of the IP address
|
||||||
|
// space, the basis for the reservation no longer applies and addresses
|
||||||
|
// in this block are subject to future allocation to a Regional Internet
|
||||||
|
// Registry for assignment in the normal manner.
|
||||||
|
|
||||||
|
// 192.0.0.0/24 - This block, corresponding to the
|
||||||
|
// numerically lowest of the former Class C addresses,
|
||||||
|
// was initially and is still reserved by the IANA.
|
||||||
|
// Given the present classless nature of the IP address
|
||||||
|
// space, the basis for the reservation no longer
|
||||||
|
// applies and addresses in this block are subject to
|
||||||
|
// future allocation to a Regional Internet Registry for
|
||||||
|
// assignment in the normal manner.
|
||||||
|
|
||||||
|
// 192.0.2.0/24 - This block is assigned as "TEST-NET" for use in
|
||||||
|
// documentation and example code. It is often used in conjunction with
|
||||||
|
// domain names example.com or example.net in vendor and protocol
|
||||||
|
// documentation. Addresses within this block should not appear on the
|
||||||
|
// public Internet.
|
||||||
|
MustIPv4Addr("192.0.2.0/24"),
|
||||||
|
|
||||||
|
// 192.88.99.0/24 - This block is allocated for use as 6to4 relay
|
||||||
|
// anycast addresses, according to [RFC3068].
|
||||||
|
MustIPv4Addr("192.88.99.0/24"),
|
||||||
|
|
||||||
|
// 192.168.0.0/16 - This block is set aside for use in private networks.
|
||||||
|
// Its intended use is documented in [RFC1918]. Addresses within this
|
||||||
|
// block should not appear on the public Internet.
|
||||||
|
MustIPv4Addr("192.168.0.0/16"),
|
||||||
|
|
||||||
|
// 198.18.0.0/15 - This block has been allocated for use
|
||||||
|
// in benchmark tests of network interconnect devices.
|
||||||
|
// Its use is documented in [RFC2544].
|
||||||
|
MustIPv4Addr("198.18.0.0/15"),
|
||||||
|
|
||||||
|
// 223.255.255.0/24 - This block, corresponding to the
|
||||||
|
// numerically highest of the former Class C addresses,
|
||||||
|
// was initially and is still reserved by the IANA.
|
||||||
|
// Given the present classless nature of the IP address
|
||||||
|
// space, the basis for the reservation no longer
|
||||||
|
// applies and addresses in this block are subject to
|
||||||
|
// future allocation to a Regional Internet Registry for
|
||||||
|
// assignment in the normal manner.
|
||||||
|
|
||||||
|
// 224.0.0.0/4 - This block, formerly known as the Class
|
||||||
|
// D address space, is allocated for use in IPv4
|
||||||
|
// multicast address assignments. The IANA guidelines
|
||||||
|
// for assignments from this space are described in
|
||||||
|
// [RFC3171].
|
||||||
|
MustIPv4Addr("224.0.0.0/4"),
|
||||||
|
|
||||||
|
// 240.0.0.0/4 - This block, formerly known as the Class E address
|
||||||
|
// space, is reserved. The "limited broadcast" destination address
|
||||||
|
// 255.255.255.255 should never be forwarded outside the (sub-)net of
|
||||||
|
// the source. The remainder of this space is reserved
|
||||||
|
// for future use. [RFC1700, page 4]
|
||||||
|
MustIPv4Addr("240.0.0.0/4"),
|
||||||
|
},
|
||||||
|
3849: {
|
||||||
|
// [RFC3849] IPv6 Address Prefix Reserved for Documentation
|
||||||
|
MustIPv6Addr("2001:db8::/32"), // [RFC3849], §4 IANA Considerations
|
||||||
|
},
|
||||||
|
3927: {
|
||||||
|
// [RFC3927] Dynamic Configuration of IPv4 Link-Local Addresses
|
||||||
|
MustIPv4Addr("169.254.0.0/16"), // [RFC3927], §2.1 Link-Local Address Selection
|
||||||
|
},
|
||||||
|
4038: {
|
||||||
|
// [RFC4038] Application Aspects of IPv6 Transition
|
||||||
|
|
||||||
|
// [RFC4038], §4.2. IPv6 Applications in a Dual-Stack Node
|
||||||
|
MustIPv6Addr("0:0:0:0:0:ffff::/96"),
|
||||||
|
},
|
||||||
|
4193: {
|
||||||
|
// [RFC4193] Unique Local IPv6 Unicast Addresses
|
||||||
|
MustIPv6Addr("fc00::/7"),
|
||||||
|
},
|
||||||
|
4291: {
|
||||||
|
// [RFC4291] IP Version 6 Addressing Architecture
|
||||||
|
|
||||||
|
// [RFC4291], §2.5.2 The Unspecified Address
|
||||||
|
MustIPv6Addr("::/128"),
|
||||||
|
|
||||||
|
// [RFC4291], §2.5.3 The Loopback Address
|
||||||
|
MustIPv6Addr("::1/128"),
|
||||||
|
|
||||||
|
// [RFC4291], §2.5.5.1. IPv4-Compatible IPv6 Address
|
||||||
|
MustIPv6Addr("::/96"),
|
||||||
|
|
||||||
|
// [RFC4291], §2.5.5.2. IPv4-Mapped IPv6 Address
|
||||||
|
MustIPv6Addr("::ffff:0:0/96"),
|
||||||
|
|
||||||
|
// [RFC4291], §2.5.6 Link-Local IPv6 Unicast Addresses
|
||||||
|
MustIPv6Addr("fe80::/10"),
|
||||||
|
|
||||||
|
// [RFC4291], §2.5.7 Site-Local IPv6 Unicast Addresses
|
||||||
|
// (depreciated)
|
||||||
|
MustIPv6Addr("fec0::/10"),
|
||||||
|
|
||||||
|
// [RFC4291], §2.7 Multicast Addresses
|
||||||
|
MustIPv6Addr("ff00::/8"),
|
||||||
|
|
||||||
|
// IPv6 Multicast Information.
|
||||||
|
//
|
||||||
|
// In the following "table" below, `ff0x` is replaced
|
||||||
|
// with the following values depending on the scope of
|
||||||
|
// the query:
|
||||||
|
//
|
||||||
|
// IPv6 Multicast Scopes:
|
||||||
|
// * ff00/9 // reserved
|
||||||
|
// * ff01/9 // interface-local
|
||||||
|
// * ff02/9 // link-local
|
||||||
|
// * ff03/9 // realm-local
|
||||||
|
// * ff04/9 // admin-local
|
||||||
|
// * ff05/9 // site-local
|
||||||
|
// * ff08/9 // organization-local
|
||||||
|
// * ff0e/9 // global
|
||||||
|
// * ff0f/9 // reserved
|
||||||
|
//
|
||||||
|
// IPv6 Multicast Addresses:
|
||||||
|
// * ff0x::2 // All routers
|
||||||
|
// * ff02::5 // OSPFIGP
|
||||||
|
// * ff02::6 // OSPFIGP Designated Routers
|
||||||
|
// * ff02::9 // RIP Routers
|
||||||
|
// * ff02::a // EIGRP Routers
|
||||||
|
// * ff02::d // All PIM Routers
|
||||||
|
// * ff02::1a // All RPL Routers
|
||||||
|
// * ff0x::fb // mDNSv6
|
||||||
|
// * ff0x::101 // All Network Time Protocol (NTP) servers
|
||||||
|
// * ff02::1:1 // Link Name
|
||||||
|
// * ff02::1:2 // All-dhcp-agents
|
||||||
|
// * ff02::1:3 // Link-local Multicast Name Resolution
|
||||||
|
// * ff05::1:3 // All-dhcp-servers
|
||||||
|
// * ff02::1:ff00:0/104 // Solicited-node multicast address.
|
||||||
|
// * ff02::2:ff00:0/104 // Node Information Queries
|
||||||
|
},
|
||||||
|
4380: {
|
||||||
|
// [RFC4380] Teredo: Tunneling IPv6 over UDP through
|
||||||
|
// Network Address Translations (NATs)
|
||||||
|
|
||||||
|
// [RFC4380], §2.6 Global Teredo IPv6 Service Prefix
|
||||||
|
MustIPv6Addr("2001:0000::/32"),
|
||||||
|
},
|
||||||
|
4773: {
|
||||||
|
// [RFC4773] Administration of the IANA Special Purpose IPv6 Address Block
|
||||||
|
MustIPv6Addr("2001:0000::/23"), // IANA
|
||||||
|
},
|
||||||
|
4843: {
|
||||||
|
// [RFC4843] An IPv6 Prefix for Overlay Routable Cryptographic Hash Identifiers (ORCHID)
|
||||||
|
MustIPv6Addr("2001:10::/28"), // [RFC4843], §7 IANA Considerations
|
||||||
|
},
|
||||||
|
5180: {
|
||||||
|
// [RFC5180] IPv6 Benchmarking Methodology for Network Interconnect Devices
|
||||||
|
MustIPv6Addr("2001:0200::/48"), // [RFC5180], §8 IANA Considerations
|
||||||
|
},
|
||||||
|
5735: {
|
||||||
|
// [RFC5735] Special Use IPv4 Addresses
|
||||||
|
MustIPv4Addr("192.0.2.0/24"), // TEST-NET-1
|
||||||
|
MustIPv4Addr("198.51.100.0/24"), // TEST-NET-2
|
||||||
|
MustIPv4Addr("203.0.113.0/24"), // TEST-NET-3
|
||||||
|
MustIPv4Addr("198.18.0.0/15"), // Benchmarks
|
||||||
|
},
|
||||||
|
5737: {
|
||||||
|
// [RFC5737] IPv4 Address Blocks Reserved for Documentation
|
||||||
|
MustIPv4Addr("192.0.2.0/24"), // TEST-NET-1
|
||||||
|
MustIPv4Addr("198.51.100.0/24"), // TEST-NET-2
|
||||||
|
MustIPv4Addr("203.0.113.0/24"), // TEST-NET-3
|
||||||
|
},
|
||||||
|
6052: {
|
||||||
|
// [RFC6052] IPv6 Addressing of IPv4/IPv6 Translators
|
||||||
|
MustIPv6Addr("64:ff9b::/96"), // [RFC6052], §2.1. Well-Known Prefix
|
||||||
|
},
|
||||||
|
6333: {
|
||||||
|
// [RFC6333] Dual-Stack Lite Broadband Deployments Following IPv4 Exhaustion
|
||||||
|
MustIPv4Addr("192.0.0.0/29"), // [RFC6333], §5.7 Well-Known IPv4 Address
|
||||||
|
},
|
||||||
|
6598: {
|
||||||
|
// [RFC6598] IANA-Reserved IPv4 Prefix for Shared Address Space
|
||||||
|
MustIPv4Addr("100.64.0.0/10"),
|
||||||
|
},
|
||||||
|
6666: {
|
||||||
|
// [RFC6666] A Discard Prefix for IPv6
|
||||||
|
MustIPv6Addr("0100::/64"),
|
||||||
|
},
|
||||||
|
6890: {
|
||||||
|
// [RFC6890] Special-Purpose IP Address Registries
|
||||||
|
|
||||||
|
// From "RFC6890 §2.2.1 Information Requirements":
|
||||||
|
/*
|
||||||
|
The IPv4 and IPv6 Special-Purpose Address Registries maintain the
|
||||||
|
following information regarding each entry:
|
||||||
|
|
||||||
|
o Address Block - A block of IPv4 or IPv6 addresses that has been
|
||||||
|
registered for a special purpose.
|
||||||
|
|
||||||
|
o Name - A descriptive name for the special-purpose address block.
|
||||||
|
|
||||||
|
o RFC - The RFC through which the special-purpose address block was
|
||||||
|
requested.
|
||||||
|
|
||||||
|
o Allocation Date - The date upon which the special-purpose address
|
||||||
|
block was allocated.
|
||||||
|
|
||||||
|
o Termination Date - The date upon which the allocation is to be
|
||||||
|
terminated. This field is applicable for limited-use allocations
|
||||||
|
only.
|
||||||
|
|
||||||
|
o Source - A boolean value indicating whether an address from the
|
||||||
|
allocated special-purpose address block is valid when used as the
|
||||||
|
source address of an IP datagram that transits two devices.
|
||||||
|
|
||||||
|
o Destination - A boolean value indicating whether an address from
|
||||||
|
the allocated special-purpose address block is valid when used as
|
||||||
|
the destination address of an IP datagram that transits two
|
||||||
|
devices.
|
||||||
|
|
||||||
|
o Forwardable - A boolean value indicating whether a router may
|
||||||
|
forward an IP datagram whose destination address is drawn from the
|
||||||
|
allocated special-purpose address block between external
|
||||||
|
interfaces.
|
||||||
|
|
||||||
|
o Global - A boolean value indicating whether an IP datagram whose
|
||||||
|
destination address is drawn from the allocated special-purpose
|
||||||
|
address block is forwardable beyond a specified administrative
|
||||||
|
domain.
|
||||||
|
|
||||||
|
o Reserved-by-Protocol - A boolean value indicating whether the
|
||||||
|
special-purpose address block is reserved by IP, itself. This
|
||||||
|
value is "TRUE" if the RFC that created the special-purpose
|
||||||
|
address block requires all compliant IP implementations to behave
|
||||||
|
in a special way when processing packets either to or from
|
||||||
|
addresses contained by the address block.
|
||||||
|
|
||||||
|
If the value of "Destination" is FALSE, the values of "Forwardable"
|
||||||
|
and "Global" must also be false.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/*+----------------------+----------------------------+
|
||||||
|
* | Attribute | Value |
|
||||||
|
* +----------------------+----------------------------+
|
||||||
|
* | Address Block | 0.0.0.0/8 |
|
||||||
|
* | Name | "This host on this network"|
|
||||||
|
* | RFC | [RFC1122], Section 3.2.1.3 |
|
||||||
|
* | Allocation Date | September 1981 |
|
||||||
|
* | Termination Date | N/A |
|
||||||
|
* | Source | True |
|
||||||
|
* | Destination | False |
|
||||||
|
* | Forwardable | False |
|
||||||
|
* | Global | False |
|
||||||
|
* | Reserved-by-Protocol | True |
|
||||||
|
* +----------------------+----------------------------+*/
|
||||||
|
MustIPv4Addr("0.0.0.0/8"),
|
||||||
|
|
||||||
|
/*+----------------------+---------------+
|
||||||
|
* | Attribute | Value |
|
||||||
|
* +----------------------+---------------+
|
||||||
|
* | Address Block | 10.0.0.0/8 |
|
||||||
|
* | Name | Private-Use |
|
||||||
|
* | RFC | [RFC1918] |
|
||||||
|
* | Allocation Date | February 1996 |
|
||||||
|
* | Termination Date | N/A |
|
||||||
|
* | Source | True |
|
||||||
|
* | Destination | True |
|
||||||
|
* | Forwardable | True |
|
||||||
|
* | Global | False |
|
||||||
|
* | Reserved-by-Protocol | False |
|
||||||
|
* +----------------------+---------------+ */
|
||||||
|
MustIPv4Addr("10.0.0.0/8"),
|
||||||
|
|
||||||
|
/*+----------------------+----------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+----------------------+
|
||||||
|
| Address Block | 100.64.0.0/10 |
|
||||||
|
| Name | Shared Address Space |
|
||||||
|
| RFC | [RFC6598] |
|
||||||
|
| Allocation Date | April 2012 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | True |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+----------------------+*/
|
||||||
|
MustIPv4Addr("100.64.0.0/10"),
|
||||||
|
|
||||||
|
/*+----------------------+----------------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+----------------------------+
|
||||||
|
| Address Block | 127.0.0.0/8 |
|
||||||
|
| Name | Loopback |
|
||||||
|
| RFC | [RFC1122], Section 3.2.1.3 |
|
||||||
|
| Allocation Date | September 1981 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | False [1] |
|
||||||
|
| Destination | False [1] |
|
||||||
|
| Forwardable | False [1] |
|
||||||
|
| Global | False [1] |
|
||||||
|
| Reserved-by-Protocol | True |
|
||||||
|
+----------------------+----------------------------+*/
|
||||||
|
// [1] Several protocols have been granted exceptions to
|
||||||
|
// this rule. For examples, see [RFC4379] and
|
||||||
|
// [RFC5884].
|
||||||
|
MustIPv4Addr("127.0.0.0/8"),
|
||||||
|
|
||||||
|
/*+----------------------+----------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+----------------+
|
||||||
|
| Address Block | 169.254.0.0/16 |
|
||||||
|
| Name | Link Local |
|
||||||
|
| RFC | [RFC3927] |
|
||||||
|
| Allocation Date | May 2005 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | False |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | True |
|
||||||
|
+----------------------+----------------+*/
|
||||||
|
MustIPv4Addr("169.254.0.0/16"),
|
||||||
|
|
||||||
|
/*+----------------------+---------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+---------------+
|
||||||
|
| Address Block | 172.16.0.0/12 |
|
||||||
|
| Name | Private-Use |
|
||||||
|
| RFC | [RFC1918] |
|
||||||
|
| Allocation Date | February 1996 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | True |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+---------------+*/
|
||||||
|
MustIPv4Addr("172.16.0.0/12"),
|
||||||
|
|
||||||
|
/*+----------------------+---------------------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+---------------------------------+
|
||||||
|
| Address Block | 192.0.0.0/24 [2] |
|
||||||
|
| Name | IETF Protocol Assignments |
|
||||||
|
| RFC | Section 2.1 of this document |
|
||||||
|
| Allocation Date | January 2010 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | False |
|
||||||
|
| Destination | False |
|
||||||
|
| Forwardable | False |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+---------------------------------+*/
|
||||||
|
// [2] Not usable unless by virtue of a more specific
|
||||||
|
// reservation.
|
||||||
|
MustIPv4Addr("192.0.0.0/24"),
|
||||||
|
|
||||||
|
/*+----------------------+--------------------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+--------------------------------+
|
||||||
|
| Address Block | 192.0.0.0/29 |
|
||||||
|
| Name | IPv4 Service Continuity Prefix |
|
||||||
|
| RFC | [RFC6333], [RFC7335] |
|
||||||
|
| Allocation Date | June 2011 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | True |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+--------------------------------+*/
|
||||||
|
MustIPv4Addr("192.0.0.0/29"),
|
||||||
|
|
||||||
|
/*+----------------------+----------------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+----------------------------+
|
||||||
|
| Address Block | 192.0.2.0/24 |
|
||||||
|
| Name | Documentation (TEST-NET-1) |
|
||||||
|
| RFC | [RFC5737] |
|
||||||
|
| Allocation Date | January 2010 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | False |
|
||||||
|
| Destination | False |
|
||||||
|
| Forwardable | False |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+----------------------------+*/
|
||||||
|
MustIPv4Addr("192.0.2.0/24"),
|
||||||
|
|
||||||
|
/*+----------------------+--------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+--------------------+
|
||||||
|
| Address Block | 192.88.99.0/24 |
|
||||||
|
| Name | 6to4 Relay Anycast |
|
||||||
|
| RFC | [RFC3068] |
|
||||||
|
| Allocation Date | June 2001 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | True |
|
||||||
|
| Global | True |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+--------------------+*/
|
||||||
|
MustIPv4Addr("192.88.99.0/24"),
|
||||||
|
|
||||||
|
/*+----------------------+----------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+----------------+
|
||||||
|
| Address Block | 192.168.0.0/16 |
|
||||||
|
| Name | Private-Use |
|
||||||
|
| RFC | [RFC1918] |
|
||||||
|
| Allocation Date | February 1996 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | True |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+----------------+*/
|
||||||
|
MustIPv4Addr("192.168.0.0/16"),
|
||||||
|
|
||||||
|
/*+----------------------+---------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+---------------+
|
||||||
|
| Address Block | 198.18.0.0/15 |
|
||||||
|
| Name | Benchmarking |
|
||||||
|
| RFC | [RFC2544] |
|
||||||
|
| Allocation Date | March 1999 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | True |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+---------------+*/
|
||||||
|
MustIPv4Addr("198.18.0.0/15"),
|
||||||
|
|
||||||
|
/*+----------------------+----------------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+----------------------------+
|
||||||
|
| Address Block | 198.51.100.0/24 |
|
||||||
|
| Name | Documentation (TEST-NET-2) |
|
||||||
|
| RFC | [RFC5737] |
|
||||||
|
| Allocation Date | January 2010 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | False |
|
||||||
|
| Destination | False |
|
||||||
|
| Forwardable | False |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+----------------------------+*/
|
||||||
|
MustIPv4Addr("198.51.100.0/24"),
|
||||||
|
|
||||||
|
/*+----------------------+----------------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+----------------------------+
|
||||||
|
| Address Block | 203.0.113.0/24 |
|
||||||
|
| Name | Documentation (TEST-NET-3) |
|
||||||
|
| RFC | [RFC5737] |
|
||||||
|
| Allocation Date | January 2010 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | False |
|
||||||
|
| Destination | False |
|
||||||
|
| Forwardable | False |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+----------------------------+*/
|
||||||
|
MustIPv4Addr("203.0.113.0/24"),
|
||||||
|
|
||||||
|
/*+----------------------+----------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+----------------------+
|
||||||
|
| Address Block | 240.0.0.0/4 |
|
||||||
|
| Name | Reserved |
|
||||||
|
| RFC | [RFC1112], Section 4 |
|
||||||
|
| Allocation Date | August 1989 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | False |
|
||||||
|
| Destination | False |
|
||||||
|
| Forwardable | False |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | True |
|
||||||
|
+----------------------+----------------------+*/
|
||||||
|
MustIPv4Addr("240.0.0.0/4"),
|
||||||
|
|
||||||
|
/*+----------------------+----------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+----------------------+
|
||||||
|
| Address Block | 255.255.255.255/32 |
|
||||||
|
| Name | Limited Broadcast |
|
||||||
|
| RFC | [RFC0919], Section 7 |
|
||||||
|
| Allocation Date | October 1984 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | False |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | False |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+----------------------+*/
|
||||||
|
MustIPv4Addr("255.255.255.255/32"),
|
||||||
|
|
||||||
|
/*+----------------------+------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+------------------+
|
||||||
|
| Address Block | ::1/128 |
|
||||||
|
| Name | Loopback Address |
|
||||||
|
| RFC | [RFC4291] |
|
||||||
|
| Allocation Date | February 2006 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | False |
|
||||||
|
| Destination | False |
|
||||||
|
| Forwardable | False |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | True |
|
||||||
|
+----------------------+------------------+*/
|
||||||
|
MustIPv6Addr("::1/128"),
|
||||||
|
|
||||||
|
/*+----------------------+---------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+---------------------+
|
||||||
|
| Address Block | ::/128 |
|
||||||
|
| Name | Unspecified Address |
|
||||||
|
| RFC | [RFC4291] |
|
||||||
|
| Allocation Date | February 2006 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | False |
|
||||||
|
| Forwardable | False |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | True |
|
||||||
|
+----------------------+---------------------+*/
|
||||||
|
MustIPv6Addr("::/128"),
|
||||||
|
|
||||||
|
/*+----------------------+---------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+---------------------+
|
||||||
|
| Address Block | 64:ff9b::/96 |
|
||||||
|
| Name | IPv4-IPv6 Translat. |
|
||||||
|
| RFC | [RFC6052] |
|
||||||
|
| Allocation Date | October 2010 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | True |
|
||||||
|
| Global | True |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+---------------------+*/
|
||||||
|
MustIPv6Addr("64:ff9b::/96"),
|
||||||
|
|
||||||
|
/*+----------------------+---------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+---------------------+
|
||||||
|
| Address Block | ::ffff:0:0/96 |
|
||||||
|
| Name | IPv4-mapped Address |
|
||||||
|
| RFC | [RFC4291] |
|
||||||
|
| Allocation Date | February 2006 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | False |
|
||||||
|
| Destination | False |
|
||||||
|
| Forwardable | False |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | True |
|
||||||
|
+----------------------+---------------------+*/
|
||||||
|
MustIPv6Addr("::ffff:0:0/96"),
|
||||||
|
|
||||||
|
/*+----------------------+----------------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+----------------------------+
|
||||||
|
| Address Block | 100::/64 |
|
||||||
|
| Name | Discard-Only Address Block |
|
||||||
|
| RFC | [RFC6666] |
|
||||||
|
| Allocation Date | June 2012 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | True |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+----------------------------+*/
|
||||||
|
MustIPv6Addr("100::/64"),
|
||||||
|
|
||||||
|
/*+----------------------+---------------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+---------------------------+
|
||||||
|
| Address Block | 2001::/23 |
|
||||||
|
| Name | IETF Protocol Assignments |
|
||||||
|
| RFC | [RFC2928] |
|
||||||
|
| Allocation Date | September 2000 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | False[1] |
|
||||||
|
| Destination | False[1] |
|
||||||
|
| Forwardable | False[1] |
|
||||||
|
| Global | False[1] |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+---------------------------+*/
|
||||||
|
// [1] Unless allowed by a more specific allocation.
|
||||||
|
MustIPv6Addr("2001::/16"),
|
||||||
|
|
||||||
|
/*+----------------------+----------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+----------------+
|
||||||
|
| Address Block | 2001::/32 |
|
||||||
|
| Name | TEREDO |
|
||||||
|
| RFC | [RFC4380] |
|
||||||
|
| Allocation Date | January 2006 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | True |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+----------------+*/
|
||||||
|
// Covered by previous entry, included for completeness.
|
||||||
|
//
|
||||||
|
// MustIPv6Addr("2001::/16"),
|
||||||
|
|
||||||
|
/*+----------------------+----------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+----------------+
|
||||||
|
| Address Block | 2001:2::/48 |
|
||||||
|
| Name | Benchmarking |
|
||||||
|
| RFC | [RFC5180] |
|
||||||
|
| Allocation Date | April 2008 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | True |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+----------------+*/
|
||||||
|
// Covered by previous entry, included for completeness.
|
||||||
|
//
|
||||||
|
// MustIPv6Addr("2001:2::/48"),
|
||||||
|
|
||||||
|
/*+----------------------+---------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+---------------+
|
||||||
|
| Address Block | 2001:db8::/32 |
|
||||||
|
| Name | Documentation |
|
||||||
|
| RFC | [RFC3849] |
|
||||||
|
| Allocation Date | July 2004 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | False |
|
||||||
|
| Destination | False |
|
||||||
|
| Forwardable | False |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+---------------+*/
|
||||||
|
// Covered by previous entry, included for completeness.
|
||||||
|
//
|
||||||
|
// MustIPv6Addr("2001:db8::/32"),
|
||||||
|
|
||||||
|
/*+----------------------+--------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+--------------+
|
||||||
|
| Address Block | 2001:10::/28 |
|
||||||
|
| Name | ORCHID |
|
||||||
|
| RFC | [RFC4843] |
|
||||||
|
| Allocation Date | March 2007 |
|
||||||
|
| Termination Date | March 2014 |
|
||||||
|
| Source | False |
|
||||||
|
| Destination | False |
|
||||||
|
| Forwardable | False |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+--------------+*/
|
||||||
|
// Covered by previous entry, included for completeness.
|
||||||
|
//
|
||||||
|
// MustIPv6Addr("2001:10::/28"),
|
||||||
|
|
||||||
|
/*+----------------------+---------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+---------------+
|
||||||
|
| Address Block | 2002::/16 [2] |
|
||||||
|
| Name | 6to4 |
|
||||||
|
| RFC | [RFC3056] |
|
||||||
|
| Allocation Date | February 2001 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | True |
|
||||||
|
| Global | N/A [2] |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+---------------+*/
|
||||||
|
// [2] See [RFC3056] for details.
|
||||||
|
MustIPv6Addr("2002::/16"),
|
||||||
|
|
||||||
|
/*+----------------------+--------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+--------------+
|
||||||
|
| Address Block | fc00::/7 |
|
||||||
|
| Name | Unique-Local |
|
||||||
|
| RFC | [RFC4193] |
|
||||||
|
| Allocation Date | October 2005 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | True |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | False |
|
||||||
|
+----------------------+--------------+*/
|
||||||
|
MustIPv6Addr("fc00::/7"),
|
||||||
|
|
||||||
|
/*+----------------------+-----------------------+
|
||||||
|
| Attribute | Value |
|
||||||
|
+----------------------+-----------------------+
|
||||||
|
| Address Block | fe80::/10 |
|
||||||
|
| Name | Linked-Scoped Unicast |
|
||||||
|
| RFC | [RFC4291] |
|
||||||
|
| Allocation Date | February 2006 |
|
||||||
|
| Termination Date | N/A |
|
||||||
|
| Source | True |
|
||||||
|
| Destination | True |
|
||||||
|
| Forwardable | False |
|
||||||
|
| Global | False |
|
||||||
|
| Reserved-by-Protocol | True |
|
||||||
|
+----------------------+-----------------------+*/
|
||||||
|
MustIPv6Addr("fe80::/10"),
|
||||||
|
},
|
||||||
|
7335: {
|
||||||
|
// [RFC7335] IPv4 Service Continuity Prefix
|
||||||
|
MustIPv4Addr("192.0.0.0/29"), // [RFC7335], §6 IANA Considerations
|
||||||
|
},
|
||||||
|
ForwardingBlacklist: { // Pseudo-RFC
|
||||||
|
// Blacklist of non-forwardable IP blocks taken from RFC6890
|
||||||
|
//
|
||||||
|
// TODO: the attributes for forwardable should be
|
||||||
|
// searcahble and embedded in the main list of RFCs
|
||||||
|
// above.
|
||||||
|
MustIPv4Addr("0.0.0.0/8"),
|
||||||
|
MustIPv4Addr("127.0.0.0/8"),
|
||||||
|
MustIPv4Addr("169.254.0.0/16"),
|
||||||
|
MustIPv4Addr("192.0.0.0/24"),
|
||||||
|
MustIPv4Addr("192.0.2.0/24"),
|
||||||
|
MustIPv4Addr("198.51.100.0/24"),
|
||||||
|
MustIPv4Addr("203.0.113.0/24"),
|
||||||
|
MustIPv4Addr("240.0.0.0/4"),
|
||||||
|
MustIPv4Addr("255.255.255.255/32"),
|
||||||
|
MustIPv6Addr("::1/128"),
|
||||||
|
MustIPv6Addr("::/128"),
|
||||||
|
MustIPv6Addr("::ffff:0:0/96"),
|
||||||
|
|
||||||
|
// There is no way of expressing a whitelist per RFC2928
|
||||||
|
// atm without creating a negative mask, which I don't
|
||||||
|
// want to do atm.
|
||||||
|
//MustIPv6Addr("2001::/23"),
|
||||||
|
|
||||||
|
MustIPv6Addr("2001:db8::/32"),
|
||||||
|
MustIPv6Addr("2001:10::/28"),
|
||||||
|
MustIPv6Addr("fe80::/10"),
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// VisitAllRFCs iterates over all known RFCs and calls the visitor
|
||||||
|
func VisitAllRFCs(fn func(rfcNum uint, sockaddrs SockAddrs)) {
|
||||||
|
rfcNetMap := KnownRFCs()
|
||||||
|
|
||||||
|
// Blacklist of faux-RFCs. Don't show the world that we're abusing the
|
||||||
|
// RFC system in this library.
|
||||||
|
rfcBlacklist := map[uint]struct{}{
|
||||||
|
ForwardingBlacklist: {},
|
||||||
|
}
|
||||||
|
|
||||||
|
for rfcNum, sas := range rfcNetMap {
|
||||||
|
if _, found := rfcBlacklist[rfcNum]; !found {
|
||||||
|
fn(rfcNum, sas)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,19 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
// RouteInterface specifies an interface for obtaining memoized route table and
|
||||||
|
// network information from a given OS.
|
||||||
|
type RouteInterface interface {
|
||||||
|
// GetDefaultInterfaceName returns the name of the interface that has a
|
||||||
|
// default route or an error and an empty string if a problem was
|
||||||
|
// encountered.
|
||||||
|
GetDefaultInterfaceName() (string, error)
|
||||||
|
}
|
||||||
|
|
||||||
|
// VisitCommands visits each command used by the platform-specific RouteInfo
|
||||||
|
// implementation.
|
||||||
|
func (ri routeInfo) VisitCommands(fn func(name string, cmd []string)) {
|
||||||
|
for k, v := range ri.cmds {
|
||||||
|
cmds := append([]string(nil), v...)
|
||||||
|
fn(k, cmds)
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,36 @@
|
||||||
|
// +build darwin dragonfly freebsd netbsd openbsd
|
||||||
|
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import "os/exec"
|
||||||
|
|
||||||
|
var cmds map[string][]string = map[string][]string{
|
||||||
|
"route": {"/sbin/route", "-n", "get", "default"},
|
||||||
|
}
|
||||||
|
|
||||||
|
type routeInfo struct {
|
||||||
|
cmds map[string][]string
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewRouteInfo returns a BSD-specific implementation of the RouteInfo
|
||||||
|
// interface.
|
||||||
|
func NewRouteInfo() (routeInfo, error) {
|
||||||
|
return routeInfo{
|
||||||
|
cmds: cmds,
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetDefaultInterfaceName returns the interface name attached to the default
|
||||||
|
// route on the default interface.
|
||||||
|
func (ri routeInfo) GetDefaultInterfaceName() (string, error) {
|
||||||
|
out, err := exec.Command(cmds["route"][0], cmds["route"][1:]...).Output()
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
var ifName string
|
||||||
|
if ifName, err = parseDefaultIfNameFromRoute(string(out)); err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
return ifName, nil
|
||||||
|
}
|
|
@ -0,0 +1,10 @@
|
||||||
|
// +build android nacl plan9
|
||||||
|
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import "errors"
|
||||||
|
|
||||||
|
// getDefaultIfName is the default interface function for unsupported platforms.
|
||||||
|
func getDefaultIfName() (string, error) {
|
||||||
|
return "", errors.New("No default interface found (unsupported platform)")
|
||||||
|
}
|
|
@ -0,0 +1,40 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"os/exec"
|
||||||
|
)
|
||||||
|
|
||||||
|
type routeInfo struct {
|
||||||
|
cmds map[string][]string
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewRouteInfo returns a Linux-specific implementation of the RouteInfo
|
||||||
|
// interface.
|
||||||
|
func NewRouteInfo() (routeInfo, error) {
|
||||||
|
// CoreOS Container Linux moved ip to /usr/bin/ip, so look it up on
|
||||||
|
// $PATH and fallback to /sbin/ip on error.
|
||||||
|
path, _ := exec.LookPath("ip")
|
||||||
|
if path == "" {
|
||||||
|
path = "/sbin/ip"
|
||||||
|
}
|
||||||
|
|
||||||
|
return routeInfo{
|
||||||
|
cmds: map[string][]string{"ip": {path, "route"}},
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetDefaultInterfaceName returns the interface name attached to the default
|
||||||
|
// route on the default interface.
|
||||||
|
func (ri routeInfo) GetDefaultInterfaceName() (string, error) {
|
||||||
|
out, err := exec.Command(ri.cmds["ip"][0], ri.cmds["ip"][1:]...).Output()
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
var ifName string
|
||||||
|
if ifName, err = parseDefaultIfNameFromIPCmd(string(out)); err != nil {
|
||||||
|
return "", errors.New("No default interface found")
|
||||||
|
}
|
||||||
|
return ifName, nil
|
||||||
|
}
|
|
@ -0,0 +1,37 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"os/exec"
|
||||||
|
)
|
||||||
|
|
||||||
|
var cmds map[string][]string = map[string][]string{
|
||||||
|
"route": {"/usr/sbin/route", "-n", "get", "default"},
|
||||||
|
}
|
||||||
|
|
||||||
|
type routeInfo struct {
|
||||||
|
cmds map[string][]string
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewRouteInfo returns a BSD-specific implementation of the RouteInfo
|
||||||
|
// interface.
|
||||||
|
func NewRouteInfo() (routeInfo, error) {
|
||||||
|
return routeInfo{
|
||||||
|
cmds: cmds,
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetDefaultInterfaceName returns the interface name attached to the default
|
||||||
|
// route on the default interface.
|
||||||
|
func (ri routeInfo) GetDefaultInterfaceName() (string, error) {
|
||||||
|
out, err := exec.Command(cmds["route"][0], cmds["route"][1:]...).Output()
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
var ifName string
|
||||||
|
if ifName, err = parseDefaultIfNameFromRoute(string(out)); err != nil {
|
||||||
|
return "", errors.New("No default interface found")
|
||||||
|
}
|
||||||
|
return ifName, nil
|
||||||
|
}
|
|
@ -0,0 +1,41 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import "os/exec"
|
||||||
|
|
||||||
|
var cmds map[string][]string = map[string][]string{
|
||||||
|
"netstat": {"netstat", "-rn"},
|
||||||
|
"ipconfig": {"ipconfig"},
|
||||||
|
}
|
||||||
|
|
||||||
|
type routeInfo struct {
|
||||||
|
cmds map[string][]string
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewRouteInfo returns a BSD-specific implementation of the RouteInfo
|
||||||
|
// interface.
|
||||||
|
func NewRouteInfo() (routeInfo, error) {
|
||||||
|
return routeInfo{
|
||||||
|
cmds: cmds,
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetDefaultInterfaceName returns the interface name attached to the default
|
||||||
|
// route on the default interface.
|
||||||
|
func (ri routeInfo) GetDefaultInterfaceName() (string, error) {
|
||||||
|
ifNameOut, err := exec.Command(cmds["netstat"][0], cmds["netstat"][1:]...).Output()
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
ipconfigOut, err := exec.Command(cmds["ipconfig"][0], cmds["ipconfig"][1:]...).Output()
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
ifName, err := parseDefaultIfNameWindows(string(ifNameOut), string(ipconfigOut))
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
return ifName, nil
|
||||||
|
}
|
|
@ -0,0 +1,206 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
type SockAddrType int
|
||||||
|
type AttrName string
|
||||||
|
|
||||||
|
const (
|
||||||
|
TypeUnknown SockAddrType = 0x0
|
||||||
|
TypeUnix = 0x1
|
||||||
|
TypeIPv4 = 0x2
|
||||||
|
TypeIPv6 = 0x4
|
||||||
|
|
||||||
|
// TypeIP is the union of TypeIPv4 and TypeIPv6
|
||||||
|
TypeIP = 0x6
|
||||||
|
)
|
||||||
|
|
||||||
|
type SockAddr interface {
|
||||||
|
// CmpRFC returns 0 if SockAddr exactly matches one of the matched RFC
|
||||||
|
// networks, -1 if the receiver is contained within the RFC network, or
|
||||||
|
// 1 if the address is not contained within the RFC.
|
||||||
|
CmpRFC(rfcNum uint, sa SockAddr) int
|
||||||
|
|
||||||
|
// Contains returns true if the SockAddr arg is contained within the
|
||||||
|
// receiver
|
||||||
|
Contains(SockAddr) bool
|
||||||
|
|
||||||
|
// Equal allows for the comparison of two SockAddrs
|
||||||
|
Equal(SockAddr) bool
|
||||||
|
|
||||||
|
DialPacketArgs() (string, string)
|
||||||
|
DialStreamArgs() (string, string)
|
||||||
|
ListenPacketArgs() (string, string)
|
||||||
|
ListenStreamArgs() (string, string)
|
||||||
|
|
||||||
|
// String returns the string representation of SockAddr
|
||||||
|
String() string
|
||||||
|
|
||||||
|
// Type returns the SockAddrType
|
||||||
|
Type() SockAddrType
|
||||||
|
}
|
||||||
|
|
||||||
|
// sockAddrAttrMap is a map of the SockAddr type-specific attributes.
|
||||||
|
var sockAddrAttrMap map[AttrName]func(SockAddr) string
|
||||||
|
var sockAddrAttrs []AttrName
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
sockAddrInit()
|
||||||
|
}
|
||||||
|
|
||||||
|
// New creates a new SockAddr from the string. The order in which New()
|
||||||
|
// attempts to construct a SockAddr is: IPv4Addr, IPv6Addr, SockAddrUnix.
|
||||||
|
//
|
||||||
|
// NOTE: New() relies on the heuristic wherein if the path begins with either a
|
||||||
|
// '.' or '/' character before creating a new UnixSock. For UNIX sockets that
|
||||||
|
// are absolute paths or are nested within a sub-directory, this works as
|
||||||
|
// expected, however if the UNIX socket is contained in the current working
|
||||||
|
// directory, this will fail unless the path begins with "./"
|
||||||
|
// (e.g. "./my-local-socket"). Calls directly to NewUnixSock() do not suffer
|
||||||
|
// this limitation. Invalid IP addresses such as "256.0.0.0/-1" will run afoul
|
||||||
|
// of this heuristic and be assumed to be a valid UNIX socket path (which they
|
||||||
|
// are, but it is probably not what you want and you won't realize it until you
|
||||||
|
// stat(2) the file system to discover it doesn't exist).
|
||||||
|
func NewSockAddr(s string) (SockAddr, error) {
|
||||||
|
ipv4Addr, err := NewIPv4Addr(s)
|
||||||
|
if err == nil {
|
||||||
|
return ipv4Addr, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
ipv6Addr, err := NewIPv6Addr(s)
|
||||||
|
if err == nil {
|
||||||
|
return ipv6Addr, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check to make sure the string begins with either a '.' or '/', or
|
||||||
|
// contains a '/'.
|
||||||
|
if len(s) > 1 && (strings.IndexAny(s[0:1], "./") != -1 || strings.IndexByte(s, '/') != -1) {
|
||||||
|
unixSock, err := NewUnixSock(s)
|
||||||
|
if err == nil {
|
||||||
|
return unixSock, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil, fmt.Errorf("Unable to convert %q to an IPv4 or IPv6 address, or a UNIX Socket", s)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ToIPAddr returns an IPAddr type or nil if the type conversion fails.
|
||||||
|
func ToIPAddr(sa SockAddr) *IPAddr {
|
||||||
|
ipa, ok := sa.(IPAddr)
|
||||||
|
if !ok {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return &ipa
|
||||||
|
}
|
||||||
|
|
||||||
|
// ToIPv4Addr returns an IPv4Addr type or nil if the type conversion fails.
|
||||||
|
func ToIPv4Addr(sa SockAddr) *IPv4Addr {
|
||||||
|
switch v := sa.(type) {
|
||||||
|
case IPv4Addr:
|
||||||
|
return &v
|
||||||
|
default:
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// ToIPv6Addr returns an IPv6Addr type or nil if the type conversion fails.
|
||||||
|
func ToIPv6Addr(sa SockAddr) *IPv6Addr {
|
||||||
|
switch v := sa.(type) {
|
||||||
|
case IPv6Addr:
|
||||||
|
return &v
|
||||||
|
default:
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// ToUnixSock returns a UnixSock type or nil if the type conversion fails.
|
||||||
|
func ToUnixSock(sa SockAddr) *UnixSock {
|
||||||
|
switch v := sa.(type) {
|
||||||
|
case UnixSock:
|
||||||
|
return &v
|
||||||
|
default:
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// SockAddrAttr returns a string representation of an attribute for the given
|
||||||
|
// SockAddr.
|
||||||
|
func SockAddrAttr(sa SockAddr, selector AttrName) string {
|
||||||
|
fn, found := sockAddrAttrMap[selector]
|
||||||
|
if !found {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
return fn(sa)
|
||||||
|
}
|
||||||
|
|
||||||
|
// String() for SockAddrType returns a string representation of the
|
||||||
|
// SockAddrType (e.g. "IPv4", "IPv6", "UNIX", "IP", or "unknown").
|
||||||
|
func (sat SockAddrType) String() string {
|
||||||
|
switch sat {
|
||||||
|
case TypeIPv4:
|
||||||
|
return "IPv4"
|
||||||
|
case TypeIPv6:
|
||||||
|
return "IPv6"
|
||||||
|
// There is no concrete "IP" type. Leaving here as a reminder.
|
||||||
|
// case TypeIP:
|
||||||
|
// return "IP"
|
||||||
|
case TypeUnix:
|
||||||
|
return "UNIX"
|
||||||
|
default:
|
||||||
|
panic("unsupported type")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// sockAddrInit is called once at init()
|
||||||
|
func sockAddrInit() {
|
||||||
|
sockAddrAttrs = []AttrName{
|
||||||
|
"type", // type should be first
|
||||||
|
"string",
|
||||||
|
}
|
||||||
|
|
||||||
|
sockAddrAttrMap = map[AttrName]func(sa SockAddr) string{
|
||||||
|
"string": func(sa SockAddr) string {
|
||||||
|
return sa.String()
|
||||||
|
},
|
||||||
|
"type": func(sa SockAddr) string {
|
||||||
|
return sa.Type().String()
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// UnixSockAttrs returns a list of attributes supported by the UnixSock type
|
||||||
|
func SockAddrAttrs() []AttrName {
|
||||||
|
return sockAddrAttrs
|
||||||
|
}
|
||||||
|
|
||||||
|
// Although this is pretty trivial to do in a program, having the logic here is
|
||||||
|
// useful all around. Note that this marshals into a *string* -- the underlying
|
||||||
|
// string representation of the sockaddr. If you then unmarshal into this type
|
||||||
|
// in Go, all will work as expected, but externally you can take what comes out
|
||||||
|
// and use the string value directly.
|
||||||
|
type SockAddrMarshaler struct {
|
||||||
|
SockAddr
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *SockAddrMarshaler) MarshalJSON() ([]byte, error) {
|
||||||
|
return json.Marshal(s.SockAddr.String())
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *SockAddrMarshaler) UnmarshalJSON(in []byte) error {
|
||||||
|
var str string
|
||||||
|
err := json.Unmarshal(in, &str)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
sa, err := NewSockAddr(str)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
s.SockAddr = sa
|
||||||
|
return nil
|
||||||
|
}
|
|
@ -0,0 +1,193 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"sort"
|
||||||
|
)
|
||||||
|
|
||||||
|
// SockAddrs is a slice of SockAddrs
|
||||||
|
type SockAddrs []SockAddr
|
||||||
|
|
||||||
|
func (s SockAddrs) Len() int { return len(s) }
|
||||||
|
func (s SockAddrs) Swap(i, j int) { s[i], s[j] = s[j], s[i] }
|
||||||
|
|
||||||
|
// CmpAddrFunc is the function signature that must be met to be used in the
|
||||||
|
// OrderedAddrBy multiAddrSorter
|
||||||
|
type CmpAddrFunc func(p1, p2 *SockAddr) int
|
||||||
|
|
||||||
|
// multiAddrSorter implements the Sort interface, sorting the SockAddrs within.
|
||||||
|
type multiAddrSorter struct {
|
||||||
|
addrs SockAddrs
|
||||||
|
cmp []CmpAddrFunc
|
||||||
|
}
|
||||||
|
|
||||||
|
// Sort sorts the argument slice according to the Cmp functions passed to
|
||||||
|
// OrderedAddrBy.
|
||||||
|
func (ms *multiAddrSorter) Sort(sockAddrs SockAddrs) {
|
||||||
|
ms.addrs = sockAddrs
|
||||||
|
sort.Sort(ms)
|
||||||
|
}
|
||||||
|
|
||||||
|
// OrderedAddrBy sorts SockAddr by the list of sort function pointers.
|
||||||
|
func OrderedAddrBy(cmpFuncs ...CmpAddrFunc) *multiAddrSorter {
|
||||||
|
return &multiAddrSorter{
|
||||||
|
cmp: cmpFuncs,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Len is part of sort.Interface.
|
||||||
|
func (ms *multiAddrSorter) Len() int {
|
||||||
|
return len(ms.addrs)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Less is part of sort.Interface. It is implemented by looping along the
|
||||||
|
// Cmp() functions until it finds a comparison that is either less than,
|
||||||
|
// equal to, or greater than.
|
||||||
|
func (ms *multiAddrSorter) Less(i, j int) bool {
|
||||||
|
p, q := &ms.addrs[i], &ms.addrs[j]
|
||||||
|
// Try all but the last comparison.
|
||||||
|
var k int
|
||||||
|
for k = 0; k < len(ms.cmp)-1; k++ {
|
||||||
|
cmp := ms.cmp[k]
|
||||||
|
x := cmp(p, q)
|
||||||
|
switch x {
|
||||||
|
case -1:
|
||||||
|
// p < q, so we have a decision.
|
||||||
|
return true
|
||||||
|
case 1:
|
||||||
|
// p > q, so we have a decision.
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
// p == q; try the next comparison.
|
||||||
|
}
|
||||||
|
// All comparisons to here said "equal", so just return whatever the
|
||||||
|
// final comparison reports.
|
||||||
|
switch ms.cmp[k](p, q) {
|
||||||
|
case -1:
|
||||||
|
return true
|
||||||
|
case 1:
|
||||||
|
return false
|
||||||
|
default:
|
||||||
|
// Still a tie! Now what?
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Swap is part of sort.Interface.
|
||||||
|
func (ms *multiAddrSorter) Swap(i, j int) {
|
||||||
|
ms.addrs[i], ms.addrs[j] = ms.addrs[j], ms.addrs[i]
|
||||||
|
}
|
||||||
|
|
||||||
|
const (
|
||||||
|
// NOTE (sean@): These constants are here for code readability only and
|
||||||
|
// are sprucing up the code for readability purposes. Some of the
|
||||||
|
// Cmp*() variants have confusing logic (especially when dealing with
|
||||||
|
// mixed-type comparisons) and this, I think, has made it easier to grok
|
||||||
|
// the code faster.
|
||||||
|
sortReceiverBeforeArg = -1
|
||||||
|
sortDeferDecision = 0
|
||||||
|
sortArgBeforeReceiver = 1
|
||||||
|
)
|
||||||
|
|
||||||
|
// AscAddress is a sorting function to sort SockAddrs by their respective
|
||||||
|
// address type. Non-equal types are deferred in the sort.
|
||||||
|
func AscAddress(p1Ptr, p2Ptr *SockAddr) int {
|
||||||
|
p1 := *p1Ptr
|
||||||
|
p2 := *p2Ptr
|
||||||
|
|
||||||
|
switch v := p1.(type) {
|
||||||
|
case IPv4Addr:
|
||||||
|
return v.CmpAddress(p2)
|
||||||
|
case IPv6Addr:
|
||||||
|
return v.CmpAddress(p2)
|
||||||
|
case UnixSock:
|
||||||
|
return v.CmpAddress(p2)
|
||||||
|
default:
|
||||||
|
return sortDeferDecision
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// AscPort is a sorting function to sort SockAddrs by their respective address
|
||||||
|
// type. Non-equal types are deferred in the sort.
|
||||||
|
func AscPort(p1Ptr, p2Ptr *SockAddr) int {
|
||||||
|
p1 := *p1Ptr
|
||||||
|
p2 := *p2Ptr
|
||||||
|
|
||||||
|
switch v := p1.(type) {
|
||||||
|
case IPv4Addr:
|
||||||
|
return v.CmpPort(p2)
|
||||||
|
case IPv6Addr:
|
||||||
|
return v.CmpPort(p2)
|
||||||
|
default:
|
||||||
|
return sortDeferDecision
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// AscPrivate is a sorting function to sort "more secure" private values before
|
||||||
|
// "more public" values. Both IPv4 and IPv6 are compared against RFC6890
|
||||||
|
// (RFC6890 includes, and is not limited to, RFC1918 and RFC6598 for IPv4, and
|
||||||
|
// IPv6 includes RFC4193).
|
||||||
|
func AscPrivate(p1Ptr, p2Ptr *SockAddr) int {
|
||||||
|
p1 := *p1Ptr
|
||||||
|
p2 := *p2Ptr
|
||||||
|
|
||||||
|
switch v := p1.(type) {
|
||||||
|
case IPv4Addr, IPv6Addr:
|
||||||
|
return v.CmpRFC(6890, p2)
|
||||||
|
default:
|
||||||
|
return sortDeferDecision
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// AscNetworkSize is a sorting function to sort SockAddrs based on their network
|
||||||
|
// size. Non-equal types are deferred in the sort.
|
||||||
|
func AscNetworkSize(p1Ptr, p2Ptr *SockAddr) int {
|
||||||
|
p1 := *p1Ptr
|
||||||
|
p2 := *p2Ptr
|
||||||
|
p1Type := p1.Type()
|
||||||
|
p2Type := p2.Type()
|
||||||
|
|
||||||
|
// Network size operations on non-IP types make no sense
|
||||||
|
if p1Type != p2Type && p1Type != TypeIP {
|
||||||
|
return sortDeferDecision
|
||||||
|
}
|
||||||
|
|
||||||
|
ipA := p1.(IPAddr)
|
||||||
|
ipB := p2.(IPAddr)
|
||||||
|
|
||||||
|
return bytes.Compare([]byte(*ipA.NetIPMask()), []byte(*ipB.NetIPMask()))
|
||||||
|
}
|
||||||
|
|
||||||
|
// AscType is a sorting function to sort "more secure" types before
|
||||||
|
// "less-secure" types.
|
||||||
|
func AscType(p1Ptr, p2Ptr *SockAddr) int {
|
||||||
|
p1 := *p1Ptr
|
||||||
|
p2 := *p2Ptr
|
||||||
|
p1Type := p1.Type()
|
||||||
|
p2Type := p2.Type()
|
||||||
|
switch {
|
||||||
|
case p1Type < p2Type:
|
||||||
|
return sortReceiverBeforeArg
|
||||||
|
case p1Type == p2Type:
|
||||||
|
return sortDeferDecision
|
||||||
|
case p1Type > p2Type:
|
||||||
|
return sortArgBeforeReceiver
|
||||||
|
default:
|
||||||
|
return sortDeferDecision
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// FilterByType returns two lists: a list of matched and unmatched SockAddrs
|
||||||
|
func (sas SockAddrs) FilterByType(type_ SockAddrType) (matched, excluded SockAddrs) {
|
||||||
|
matched = make(SockAddrs, 0, len(sas))
|
||||||
|
excluded = make(SockAddrs, 0, len(sas))
|
||||||
|
|
||||||
|
for _, sa := range sas {
|
||||||
|
if sa.Type()&type_ != 0 {
|
||||||
|
matched = append(matched, sa)
|
||||||
|
} else {
|
||||||
|
excluded = append(excluded, sa)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return matched, excluded
|
||||||
|
}
|
|
@ -0,0 +1,135 @@
|
||||||
|
package sockaddr
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
type UnixSock struct {
|
||||||
|
SockAddr
|
||||||
|
path string
|
||||||
|
}
|
||||||
|
type UnixSocks []*UnixSock
|
||||||
|
|
||||||
|
// unixAttrMap is a map of the UnixSockAddr type-specific attributes.
|
||||||
|
var unixAttrMap map[AttrName]func(UnixSock) string
|
||||||
|
var unixAttrs []AttrName
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
unixAttrInit()
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewUnixSock creates an UnixSock from a string path. String can be in the
|
||||||
|
// form of either URI-based string (e.g. `file:///etc/passwd`), an absolute
|
||||||
|
// path (e.g. `/etc/passwd`), or a relative path (e.g. `./foo`).
|
||||||
|
func NewUnixSock(s string) (ret UnixSock, err error) {
|
||||||
|
ret.path = s
|
||||||
|
return ret, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// CmpAddress follows the Cmp() standard protocol and returns:
|
||||||
|
//
|
||||||
|
// - -1 If the receiver should sort first because its name lexically sorts before arg
|
||||||
|
// - 0 if the SockAddr arg is not a UnixSock, or is a UnixSock with the same path.
|
||||||
|
// - 1 If the argument should sort first.
|
||||||
|
func (us UnixSock) CmpAddress(sa SockAddr) int {
|
||||||
|
usb, ok := sa.(UnixSock)
|
||||||
|
if !ok {
|
||||||
|
return sortDeferDecision
|
||||||
|
}
|
||||||
|
|
||||||
|
return strings.Compare(us.Path(), usb.Path())
|
||||||
|
}
|
||||||
|
|
||||||
|
// DialPacketArgs returns the arguments required to be passed to net.DialUnix()
|
||||||
|
// with the `unixgram` network type.
|
||||||
|
func (us UnixSock) DialPacketArgs() (network, dialArgs string) {
|
||||||
|
return "unixgram", us.path
|
||||||
|
}
|
||||||
|
|
||||||
|
// DialStreamArgs returns the arguments required to be passed to net.DialUnix()
|
||||||
|
// with the `unix` network type.
|
||||||
|
func (us UnixSock) DialStreamArgs() (network, dialArgs string) {
|
||||||
|
return "unix", us.path
|
||||||
|
}
|
||||||
|
|
||||||
|
// Equal returns true if a SockAddr is equal to the receiving UnixSock.
|
||||||
|
func (us UnixSock) Equal(sa SockAddr) bool {
|
||||||
|
usb, ok := sa.(UnixSock)
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
if us.Path() != usb.Path() {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// ListenPacketArgs returns the arguments required to be passed to
|
||||||
|
// net.ListenUnixgram() with the `unixgram` network type.
|
||||||
|
func (us UnixSock) ListenPacketArgs() (network, dialArgs string) {
|
||||||
|
return "unixgram", us.path
|
||||||
|
}
|
||||||
|
|
||||||
|
// ListenStreamArgs returns the arguments required to be passed to
|
||||||
|
// net.ListenUnix() with the `unix` network type.
|
||||||
|
func (us UnixSock) ListenStreamArgs() (network, dialArgs string) {
|
||||||
|
return "unix", us.path
|
||||||
|
}
|
||||||
|
|
||||||
|
// MustUnixSock is a helper method that must return an UnixSock or panic on
|
||||||
|
// invalid input.
|
||||||
|
func MustUnixSock(addr string) UnixSock {
|
||||||
|
us, err := NewUnixSock(addr)
|
||||||
|
if err != nil {
|
||||||
|
panic(fmt.Sprintf("Unable to create a UnixSock from %+q: %v", addr, err))
|
||||||
|
}
|
||||||
|
return us
|
||||||
|
}
|
||||||
|
|
||||||
|
// Path returns the given path of the UnixSock
|
||||||
|
func (us UnixSock) Path() string {
|
||||||
|
return us.path
|
||||||
|
}
|
||||||
|
|
||||||
|
// String returns the path of the UnixSock
|
||||||
|
func (us UnixSock) String() string {
|
||||||
|
return fmt.Sprintf("%+q", us.path)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Type is used as a type switch and returns TypeUnix
|
||||||
|
func (UnixSock) Type() SockAddrType {
|
||||||
|
return TypeUnix
|
||||||
|
}
|
||||||
|
|
||||||
|
// UnixSockAttrs returns a list of attributes supported by the UnixSockAddr type
|
||||||
|
func UnixSockAttrs() []AttrName {
|
||||||
|
return unixAttrs
|
||||||
|
}
|
||||||
|
|
||||||
|
// UnixSockAttr returns a string representation of an attribute for the given
|
||||||
|
// UnixSock.
|
||||||
|
func UnixSockAttr(us UnixSock, attrName AttrName) string {
|
||||||
|
fn, found := unixAttrMap[attrName]
|
||||||
|
if !found {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
return fn(us)
|
||||||
|
}
|
||||||
|
|
||||||
|
// unixAttrInit is called once at init()
|
||||||
|
func unixAttrInit() {
|
||||||
|
// Sorted for human readability
|
||||||
|
unixAttrs = []AttrName{
|
||||||
|
"path",
|
||||||
|
}
|
||||||
|
|
||||||
|
unixAttrMap = map[AttrName]func(us UnixSock) string{
|
||||||
|
"path": func(us UnixSock) string {
|
||||||
|
return us.Path()
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,354 @@
|
||||||
|
Mozilla Public License, version 2.0
|
||||||
|
|
||||||
|
1. Definitions
|
||||||
|
|
||||||
|
1.1. “Contributor”
|
||||||
|
|
||||||
|
means each individual or legal entity that creates, contributes to the
|
||||||
|
creation of, or owns Covered Software.
|
||||||
|
|
||||||
|
1.2. “Contributor Version”
|
||||||
|
|
||||||
|
means the combination of the Contributions of others (if any) used by a
|
||||||
|
Contributor and that particular Contributor’s Contribution.
|
||||||
|
|
||||||
|
1.3. “Contribution”
|
||||||
|
|
||||||
|
means Covered Software of a particular Contributor.
|
||||||
|
|
||||||
|
1.4. “Covered Software”
|
||||||
|
|
||||||
|
means Source Code Form to which the initial Contributor has attached the
|
||||||
|
notice in Exhibit A, the Executable Form of such Source Code Form, and
|
||||||
|
Modifications of such Source Code Form, in each case including portions
|
||||||
|
thereof.
|
||||||
|
|
||||||
|
1.5. “Incompatible With Secondary Licenses”
|
||||||
|
means
|
||||||
|
|
||||||
|
a. that the initial Contributor has attached the notice described in
|
||||||
|
Exhibit B to the Covered Software; or
|
||||||
|
|
||||||
|
b. that the Covered Software was made available under the terms of version
|
||||||
|
1.1 or earlier of the License, but not also under the terms of a
|
||||||
|
Secondary License.
|
||||||
|
|
||||||
|
1.6. “Executable Form”
|
||||||
|
|
||||||
|
means any form of the work other than Source Code Form.
|
||||||
|
|
||||||
|
1.7. “Larger Work”
|
||||||
|
|
||||||
|
means a work that combines Covered Software with other material, in a separate
|
||||||
|
file or files, that is not Covered Software.
|
||||||
|
|
||||||
|
1.8. “License”
|
||||||
|
|
||||||
|
means this document.
|
||||||
|
|
||||||
|
1.9. “Licensable”
|
||||||
|
|
||||||
|
means having the right to grant, to the maximum extent possible, whether at the
|
||||||
|
time of the initial grant or subsequently, any and all of the rights conveyed by
|
||||||
|
this License.
|
||||||
|
|
||||||
|
1.10. “Modifications”
|
||||||
|
|
||||||
|
means any of the following:
|
||||||
|
|
||||||
|
a. any file in Source Code Form that results from an addition to, deletion
|
||||||
|
from, or modification of the contents of Covered Software; or
|
||||||
|
|
||||||
|
b. any new file in Source Code Form that contains any Covered Software.
|
||||||
|
|
||||||
|
1.11. “Patent Claims” of a Contributor
|
||||||
|
|
||||||
|
means any patent claim(s), including without limitation, method, process,
|
||||||
|
and apparatus claims, in any patent Licensable by such Contributor that
|
||||||
|
would be infringed, but for the grant of the License, by the making,
|
||||||
|
using, selling, offering for sale, having made, import, or transfer of
|
||||||
|
either its Contributions or its Contributor Version.
|
||||||
|
|
||||||
|
1.12. “Secondary License”
|
||||||
|
|
||||||
|
means either the GNU General Public License, Version 2.0, the GNU Lesser
|
||||||
|
General Public License, Version 2.1, the GNU Affero General Public
|
||||||
|
License, Version 3.0, or any later versions of those licenses.
|
||||||
|
|
||||||
|
1.13. “Source Code Form”
|
||||||
|
|
||||||
|
means the form of the work preferred for making modifications.
|
||||||
|
|
||||||
|
1.14. “You” (or “Your”)
|
||||||
|
|
||||||
|
means an individual or a legal entity exercising rights under this
|
||||||
|
License. For legal entities, “You” includes any entity that controls, is
|
||||||
|
controlled by, or is under common control with You. For purposes of this
|
||||||
|
definition, “control” means (a) the power, direct or indirect, to cause
|
||||||
|
the direction or management of such entity, whether by contract or
|
||||||
|
otherwise, or (b) ownership of more than fifty percent (50%) of the
|
||||||
|
outstanding shares or beneficial ownership of such entity.
|
||||||
|
|
||||||
|
|
||||||
|
2. License Grants and Conditions
|
||||||
|
|
||||||
|
2.1. Grants
|
||||||
|
|
||||||
|
Each Contributor hereby grants You a world-wide, royalty-free,
|
||||||
|
non-exclusive license:
|
||||||
|
|
||||||
|
a. under intellectual property rights (other than patent or trademark)
|
||||||
|
Licensable by such Contributor to use, reproduce, make available,
|
||||||
|
modify, display, perform, distribute, and otherwise exploit its
|
||||||
|
Contributions, either on an unmodified basis, with Modifications, or as
|
||||||
|
part of a Larger Work; and
|
||||||
|
|
||||||
|
b. under Patent Claims of such Contributor to make, use, sell, offer for
|
||||||
|
sale, have made, import, and otherwise transfer either its Contributions
|
||||||
|
or its Contributor Version.
|
||||||
|
|
||||||
|
2.2. Effective Date
|
||||||
|
|
||||||
|
The licenses granted in Section 2.1 with respect to any Contribution become
|
||||||
|
effective for each Contribution on the date the Contributor first distributes
|
||||||
|
such Contribution.
|
||||||
|
|
||||||
|
2.3. Limitations on Grant Scope
|
||||||
|
|
||||||
|
The licenses granted in this Section 2 are the only rights granted under this
|
||||||
|
License. No additional rights or licenses will be implied from the distribution
|
||||||
|
or licensing of Covered Software under this License. Notwithstanding Section
|
||||||
|
2.1(b) above, no patent license is granted by a Contributor:
|
||||||
|
|
||||||
|
a. for any code that a Contributor has removed from Covered Software; or
|
||||||
|
|
||||||
|
b. for infringements caused by: (i) Your and any other third party’s
|
||||||
|
modifications of Covered Software, or (ii) the combination of its
|
||||||
|
Contributions with other software (except as part of its Contributor
|
||||||
|
Version); or
|
||||||
|
|
||||||
|
c. under Patent Claims infringed by Covered Software in the absence of its
|
||||||
|
Contributions.
|
||||||
|
|
||||||
|
This License does not grant any rights in the trademarks, service marks, or
|
||||||
|
logos of any Contributor (except as may be necessary to comply with the
|
||||||
|
notice requirements in Section 3.4).
|
||||||
|
|
||||||
|
2.4. Subsequent Licenses
|
||||||
|
|
||||||
|
No Contributor makes additional grants as a result of Your choice to
|
||||||
|
distribute the Covered Software under a subsequent version of this License
|
||||||
|
(see Section 10.2) or under the terms of a Secondary License (if permitted
|
||||||
|
under the terms of Section 3.3).
|
||||||
|
|
||||||
|
2.5. Representation
|
||||||
|
|
||||||
|
Each Contributor represents that the Contributor believes its Contributions
|
||||||
|
are its original creation(s) or it has sufficient rights to grant the
|
||||||
|
rights to its Contributions conveyed by this License.
|
||||||
|
|
||||||
|
2.6. Fair Use
|
||||||
|
|
||||||
|
This License is not intended to limit any rights You have under applicable
|
||||||
|
copyright doctrines of fair use, fair dealing, or other equivalents.
|
||||||
|
|
||||||
|
2.7. Conditions
|
||||||
|
|
||||||
|
Sections 3.1, 3.2, 3.3, and 3.4 are conditions of the licenses granted in
|
||||||
|
Section 2.1.
|
||||||
|
|
||||||
|
|
||||||
|
3. Responsibilities
|
||||||
|
|
||||||
|
3.1. Distribution of Source Form
|
||||||
|
|
||||||
|
All distribution of Covered Software in Source Code Form, including any
|
||||||
|
Modifications that You create or to which You contribute, must be under the
|
||||||
|
terms of this License. You must inform recipients that the Source Code Form
|
||||||
|
of the Covered Software is governed by the terms of this License, and how
|
||||||
|
they can obtain a copy of this License. You may not attempt to alter or
|
||||||
|
restrict the recipients’ rights in the Source Code Form.
|
||||||
|
|
||||||
|
3.2. Distribution of Executable Form
|
||||||
|
|
||||||
|
If You distribute Covered Software in Executable Form then:
|
||||||
|
|
||||||
|
a. such Covered Software must also be made available in Source Code Form,
|
||||||
|
as described in Section 3.1, and You must inform recipients of the
|
||||||
|
Executable Form how they can obtain a copy of such Source Code Form by
|
||||||
|
reasonable means in a timely manner, at a charge no more than the cost
|
||||||
|
of distribution to the recipient; and
|
||||||
|
|
||||||
|
b. You may distribute such Executable Form under the terms of this License,
|
||||||
|
or sublicense it under different terms, provided that the license for
|
||||||
|
the Executable Form does not attempt to limit or alter the recipients’
|
||||||
|
rights in the Source Code Form under this License.
|
||||||
|
|
||||||
|
3.3. Distribution of a Larger Work
|
||||||
|
|
||||||
|
You may create and distribute a Larger Work under terms of Your choice,
|
||||||
|
provided that You also comply with the requirements of this License for the
|
||||||
|
Covered Software. If the Larger Work is a combination of Covered Software
|
||||||
|
with a work governed by one or more Secondary Licenses, and the Covered
|
||||||
|
Software is not Incompatible With Secondary Licenses, this License permits
|
||||||
|
You to additionally distribute such Covered Software under the terms of
|
||||||
|
such Secondary License(s), so that the recipient of the Larger Work may, at
|
||||||
|
their option, further distribute the Covered Software under the terms of
|
||||||
|
either this License or such Secondary License(s).
|
||||||
|
|
||||||
|
3.4. Notices
|
||||||
|
|
||||||
|
You may not remove or alter the substance of any license notices (including
|
||||||
|
copyright notices, patent notices, disclaimers of warranty, or limitations
|
||||||
|
of liability) contained within the Source Code Form of the Covered
|
||||||
|
Software, except that You may alter any license notices to the extent
|
||||||
|
required to remedy known factual inaccuracies.
|
||||||
|
|
||||||
|
3.5. Application of Additional Terms
|
||||||
|
|
||||||
|
You may choose to offer, and to charge a fee for, warranty, support,
|
||||||
|
indemnity or liability obligations to one or more recipients of Covered
|
||||||
|
Software. However, You may do so only on Your own behalf, and not on behalf
|
||||||
|
of any Contributor. You must make it absolutely clear that any such
|
||||||
|
warranty, support, indemnity, or liability obligation is offered by You
|
||||||
|
alone, and You hereby agree to indemnify every Contributor for any
|
||||||
|
liability incurred by such Contributor as a result of warranty, support,
|
||||||
|
indemnity or liability terms You offer. You may include additional
|
||||||
|
disclaimers of warranty and limitations of liability specific to any
|
||||||
|
jurisdiction.
|
||||||
|
|
||||||
|
4. Inability to Comply Due to Statute or Regulation
|
||||||
|
|
||||||
|
If it is impossible for You to comply with any of the terms of this License
|
||||||
|
with respect to some or all of the Covered Software due to statute, judicial
|
||||||
|
order, or regulation then You must: (a) comply with the terms of this License
|
||||||
|
to the maximum extent possible; and (b) describe the limitations and the code
|
||||||
|
they affect. Such description must be placed in a text file included with all
|
||||||
|
distributions of the Covered Software under this License. Except to the
|
||||||
|
extent prohibited by statute or regulation, such description must be
|
||||||
|
sufficiently detailed for a recipient of ordinary skill to be able to
|
||||||
|
understand it.
|
||||||
|
|
||||||
|
5. Termination
|
||||||
|
|
||||||
|
5.1. The rights granted under this License will terminate automatically if You
|
||||||
|
fail to comply with any of its terms. However, if You become compliant,
|
||||||
|
then the rights granted under this License from a particular Contributor
|
||||||
|
are reinstated (a) provisionally, unless and until such Contributor
|
||||||
|
explicitly and finally terminates Your grants, and (b) on an ongoing basis,
|
||||||
|
if such Contributor fails to notify You of the non-compliance by some
|
||||||
|
reasonable means prior to 60 days after You have come back into compliance.
|
||||||
|
Moreover, Your grants from a particular Contributor are reinstated on an
|
||||||
|
ongoing basis if such Contributor notifies You of the non-compliance by
|
||||||
|
some reasonable means, this is the first time You have received notice of
|
||||||
|
non-compliance with this License from such Contributor, and You become
|
||||||
|
compliant prior to 30 days after Your receipt of the notice.
|
||||||
|
|
||||||
|
5.2. If You initiate litigation against any entity by asserting a patent
|
||||||
|
infringement claim (excluding declaratory judgment actions, counter-claims,
|
||||||
|
and cross-claims) alleging that a Contributor Version directly or
|
||||||
|
indirectly infringes any patent, then the rights granted to You by any and
|
||||||
|
all Contributors for the Covered Software under Section 2.1 of this License
|
||||||
|
shall terminate.
|
||||||
|
|
||||||
|
5.3. In the event of termination under Sections 5.1 or 5.2 above, all end user
|
||||||
|
license agreements (excluding distributors and resellers) which have been
|
||||||
|
validly granted by You or Your distributors under this License prior to
|
||||||
|
termination shall survive termination.
|
||||||
|
|
||||||
|
6. Disclaimer of Warranty
|
||||||
|
|
||||||
|
Covered Software is provided under this License on an “as is” basis, without
|
||||||
|
warranty of any kind, either expressed, implied, or statutory, including,
|
||||||
|
without limitation, warranties that the Covered Software is free of defects,
|
||||||
|
merchantable, fit for a particular purpose or non-infringing. The entire
|
||||||
|
risk as to the quality and performance of the Covered Software is with You.
|
||||||
|
Should any Covered Software prove defective in any respect, You (not any
|
||||||
|
Contributor) assume the cost of any necessary servicing, repair, or
|
||||||
|
correction. This disclaimer of warranty constitutes an essential part of this
|
||||||
|
License. No use of any Covered Software is authorized under this License
|
||||||
|
except under this disclaimer.
|
||||||
|
|
||||||
|
7. Limitation of Liability
|
||||||
|
|
||||||
|
Under no circumstances and under no legal theory, whether tort (including
|
||||||
|
negligence), contract, or otherwise, shall any Contributor, or anyone who
|
||||||
|
distributes Covered Software as permitted above, be liable to You for any
|
||||||
|
direct, indirect, special, incidental, or consequential damages of any
|
||||||
|
character including, without limitation, damages for lost profits, loss of
|
||||||
|
goodwill, work stoppage, computer failure or malfunction, or any and all
|
||||||
|
other commercial damages or losses, even if such party shall have been
|
||||||
|
informed of the possibility of such damages. This limitation of liability
|
||||||
|
shall not apply to liability for death or personal injury resulting from such
|
||||||
|
party’s negligence to the extent applicable law prohibits such limitation.
|
||||||
|
Some jurisdictions do not allow the exclusion or limitation of incidental or
|
||||||
|
consequential damages, so this exclusion and limitation may not apply to You.
|
||||||
|
|
||||||
|
8. Litigation
|
||||||
|
|
||||||
|
Any litigation relating to this License may be brought only in the courts of
|
||||||
|
a jurisdiction where the defendant maintains its principal place of business
|
||||||
|
and such litigation shall be governed by laws of that jurisdiction, without
|
||||||
|
reference to its conflict-of-law provisions. Nothing in this Section shall
|
||||||
|
prevent a party’s ability to bring cross-claims or counter-claims.
|
||||||
|
|
||||||
|
9. Miscellaneous
|
||||||
|
|
||||||
|
This License represents the complete agreement concerning the subject matter
|
||||||
|
hereof. If any provision of this License is held to be unenforceable, such
|
||||||
|
provision shall be reformed only to the extent necessary to make it
|
||||||
|
enforceable. Any law or regulation which provides that the language of a
|
||||||
|
contract shall be construed against the drafter shall not be used to construe
|
||||||
|
this License against a Contributor.
|
||||||
|
|
||||||
|
|
||||||
|
10. Versions of the License
|
||||||
|
|
||||||
|
10.1. New Versions
|
||||||
|
|
||||||
|
Mozilla Foundation is the license steward. Except as provided in Section
|
||||||
|
10.3, no one other than the license steward has the right to modify or
|
||||||
|
publish new versions of this License. Each version will be given a
|
||||||
|
distinguishing version number.
|
||||||
|
|
||||||
|
10.2. Effect of New Versions
|
||||||
|
|
||||||
|
You may distribute the Covered Software under the terms of the version of
|
||||||
|
the License under which You originally received the Covered Software, or
|
||||||
|
under the terms of any subsequent version published by the license
|
||||||
|
steward.
|
||||||
|
|
||||||
|
10.3. Modified Versions
|
||||||
|
|
||||||
|
If you create software not governed by this License, and you want to
|
||||||
|
create a new license for such software, you may create and use a modified
|
||||||
|
version of this License if you rename the license and remove any
|
||||||
|
references to the name of the license steward (except to note that such
|
||||||
|
modified license differs from this License).
|
||||||
|
|
||||||
|
10.4. Distributing Source Code Form that is Incompatible With Secondary Licenses
|
||||||
|
If You choose to distribute Source Code Form that is Incompatible With
|
||||||
|
Secondary Licenses under the terms of this version of the License, the
|
||||||
|
notice described in Exhibit B of this License must be attached.
|
||||||
|
|
||||||
|
Exhibit A - Source Code Form License Notice
|
||||||
|
|
||||||
|
This Source Code Form is subject to the
|
||||||
|
terms of the Mozilla Public License, v.
|
||||||
|
2.0. If a copy of the MPL was not
|
||||||
|
distributed with this file, You can
|
||||||
|
obtain one at
|
||||||
|
http://mozilla.org/MPL/2.0/.
|
||||||
|
|
||||||
|
If it is not possible or desirable to put the notice in a particular file, then
|
||||||
|
You may include the notice in a location (such as a LICENSE file in a relevant
|
||||||
|
directory) where a recipient would be likely to look for such a notice.
|
||||||
|
|
||||||
|
You may add additional accurate notices of copyright ownership.
|
||||||
|
|
||||||
|
Exhibit B - “Incompatible With Secondary Licenses” Notice
|
||||||
|
|
||||||
|
This Source Code Form is “Incompatible
|
||||||
|
With Secondary Licenses”, as defined by
|
||||||
|
the Mozilla Public License, v. 2.0.
|
||||||
|
|
|
@ -0,0 +1,18 @@
|
||||||
|
TEST?=./...
|
||||||
|
|
||||||
|
default: test
|
||||||
|
|
||||||
|
fmt: generate
|
||||||
|
go fmt ./...
|
||||||
|
|
||||||
|
test: generate
|
||||||
|
go get -t ./...
|
||||||
|
go test $(TEST) $(TESTARGS)
|
||||||
|
|
||||||
|
generate:
|
||||||
|
go generate ./...
|
||||||
|
|
||||||
|
updatedeps:
|
||||||
|
go get -u golang.org/x/tools/cmd/stringer
|
||||||
|
|
||||||
|
.PHONY: default generate test updatedeps
|
|
@ -0,0 +1,125 @@
|
||||||
|
# HCL
|
||||||
|
|
||||||
|
[![GoDoc](https://godoc.org/github.com/hashicorp/hcl?status.png)](https://godoc.org/github.com/hashicorp/hcl) [![Build Status](https://travis-ci.org/hashicorp/hcl.svg?branch=master)](https://travis-ci.org/hashicorp/hcl)
|
||||||
|
|
||||||
|
HCL (HashiCorp Configuration Language) is a configuration language built
|
||||||
|
by HashiCorp. The goal of HCL is to build a structured configuration language
|
||||||
|
that is both human and machine friendly for use with command-line tools, but
|
||||||
|
specifically targeted towards DevOps tools, servers, etc.
|
||||||
|
|
||||||
|
HCL is also fully JSON compatible. That is, JSON can be used as completely
|
||||||
|
valid input to a system expecting HCL. This helps makes systems
|
||||||
|
interoperable with other systems.
|
||||||
|
|
||||||
|
HCL is heavily inspired by
|
||||||
|
[libucl](https://github.com/vstakhov/libucl),
|
||||||
|
nginx configuration, and others similar.
|
||||||
|
|
||||||
|
## Why?
|
||||||
|
|
||||||
|
A common question when viewing HCL is to ask the question: why not
|
||||||
|
JSON, YAML, etc.?
|
||||||
|
|
||||||
|
Prior to HCL, the tools we built at [HashiCorp](http://www.hashicorp.com)
|
||||||
|
used a variety of configuration languages from full programming languages
|
||||||
|
such as Ruby to complete data structure languages such as JSON. What we
|
||||||
|
learned is that some people wanted human-friendly configuration languages
|
||||||
|
and some people wanted machine-friendly languages.
|
||||||
|
|
||||||
|
JSON fits a nice balance in this, but is fairly verbose and most
|
||||||
|
importantly doesn't support comments. With YAML, we found that beginners
|
||||||
|
had a really hard time determining what the actual structure was, and
|
||||||
|
ended up guessing more often than not whether to use a hyphen, colon, etc.
|
||||||
|
in order to represent some configuration key.
|
||||||
|
|
||||||
|
Full programming languages such as Ruby enable complex behavior
|
||||||
|
a configuration language shouldn't usually allow, and also forces
|
||||||
|
people to learn some set of Ruby.
|
||||||
|
|
||||||
|
Because of this, we decided to create our own configuration language
|
||||||
|
that is JSON-compatible. Our configuration language (HCL) is designed
|
||||||
|
to be written and modified by humans. The API for HCL allows JSON
|
||||||
|
as an input so that it is also machine-friendly (machines can generate
|
||||||
|
JSON instead of trying to generate HCL).
|
||||||
|
|
||||||
|
Our goal with HCL is not to alienate other configuration languages.
|
||||||
|
It is instead to provide HCL as a specialized language for our tools,
|
||||||
|
and JSON as the interoperability layer.
|
||||||
|
|
||||||
|
## Syntax
|
||||||
|
|
||||||
|
For a complete grammar, please see the parser itself. A high-level overview
|
||||||
|
of the syntax and grammar is listed here.
|
||||||
|
|
||||||
|
* Single line comments start with `#` or `//`
|
||||||
|
|
||||||
|
* Multi-line comments are wrapped in `/*` and `*/`. Nested block comments
|
||||||
|
are not allowed. A multi-line comment (also known as a block comment)
|
||||||
|
terminates at the first `*/` found.
|
||||||
|
|
||||||
|
* Values are assigned with the syntax `key = value` (whitespace doesn't
|
||||||
|
matter). The value can be any primitive: a string, number, boolean,
|
||||||
|
object, or list.
|
||||||
|
|
||||||
|
* Strings are double-quoted and can contain any UTF-8 characters.
|
||||||
|
Example: `"Hello, World"`
|
||||||
|
|
||||||
|
* Multi-line strings start with `<<EOF` at the end of a line, and end
|
||||||
|
with `EOF` on its own line ([here documents](https://en.wikipedia.org/wiki/Here_document)).
|
||||||
|
Any text may be used in place of `EOF`. Example:
|
||||||
|
```
|
||||||
|
<<FOO
|
||||||
|
hello
|
||||||
|
world
|
||||||
|
FOO
|
||||||
|
```
|
||||||
|
|
||||||
|
* Numbers are assumed to be base 10. If you prefix a number with 0x,
|
||||||
|
it is treated as a hexadecimal. If it is prefixed with 0, it is
|
||||||
|
treated as an octal. Numbers can be in scientific notation: "1e10".
|
||||||
|
|
||||||
|
* Boolean values: `true`, `false`
|
||||||
|
|
||||||
|
* Arrays can be made by wrapping it in `[]`. Example:
|
||||||
|
`["foo", "bar", 42]`. Arrays can contain primitives,
|
||||||
|
other arrays, and objects. As an alternative, lists
|
||||||
|
of objects can be created with repeated blocks, using
|
||||||
|
this structure:
|
||||||
|
|
||||||
|
```hcl
|
||||||
|
service {
|
||||||
|
key = "value"
|
||||||
|
}
|
||||||
|
|
||||||
|
service {
|
||||||
|
key = "value"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
Objects and nested objects are created using the structure shown below:
|
||||||
|
|
||||||
|
```
|
||||||
|
variable "ami" {
|
||||||
|
description = "the AMI to use"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
This would be equivalent to the following json:
|
||||||
|
``` json
|
||||||
|
{
|
||||||
|
"variable": {
|
||||||
|
"ami": {
|
||||||
|
"description": "the AMI to use"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
## Thanks
|
||||||
|
|
||||||
|
Thanks to:
|
||||||
|
|
||||||
|
* [@vstakhov](https://github.com/vstakhov) - The original libucl parser
|
||||||
|
and syntax that HCL was based off of.
|
||||||
|
|
||||||
|
* [@fatih](https://github.com/fatih) - The rewritten HCL parser
|
||||||
|
in pure Go (no goyacc) and support for a printer.
|
|
@ -0,0 +1,19 @@
|
||||||
|
version: "build-{branch}-{build}"
|
||||||
|
image: Visual Studio 2015
|
||||||
|
clone_folder: c:\gopath\src\github.com\hashicorp\hcl
|
||||||
|
environment:
|
||||||
|
GOPATH: c:\gopath
|
||||||
|
init:
|
||||||
|
- git config --global core.autocrlf false
|
||||||
|
install:
|
||||||
|
- cmd: >-
|
||||||
|
echo %Path%
|
||||||
|
|
||||||
|
go version
|
||||||
|
|
||||||
|
go env
|
||||||
|
|
||||||
|
go get -t ./...
|
||||||
|
|
||||||
|
build_script:
|
||||||
|
- cmd: go test -v ./...
|
|
@ -0,0 +1,729 @@
|
||||||
|
package hcl
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"reflect"
|
||||||
|
"sort"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/hashicorp/hcl/hcl/ast"
|
||||||
|
"github.com/hashicorp/hcl/hcl/parser"
|
||||||
|
"github.com/hashicorp/hcl/hcl/token"
|
||||||
|
)
|
||||||
|
|
||||||
|
// This is the tag to use with structures to have settings for HCL
|
||||||
|
const tagName = "hcl"
|
||||||
|
|
||||||
|
var (
|
||||||
|
// nodeType holds a reference to the type of ast.Node
|
||||||
|
nodeType reflect.Type = findNodeType()
|
||||||
|
)
|
||||||
|
|
||||||
|
// Unmarshal accepts a byte slice as input and writes the
|
||||||
|
// data to the value pointed to by v.
|
||||||
|
func Unmarshal(bs []byte, v interface{}) error {
|
||||||
|
root, err := parse(bs)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
return DecodeObject(v, root)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Decode reads the given input and decodes it into the structure
|
||||||
|
// given by `out`.
|
||||||
|
func Decode(out interface{}, in string) error {
|
||||||
|
obj, err := Parse(in)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
return DecodeObject(out, obj)
|
||||||
|
}
|
||||||
|
|
||||||
|
// DecodeObject is a lower-level version of Decode. It decodes a
|
||||||
|
// raw Object into the given output.
|
||||||
|
func DecodeObject(out interface{}, n ast.Node) error {
|
||||||
|
val := reflect.ValueOf(out)
|
||||||
|
if val.Kind() != reflect.Ptr {
|
||||||
|
return errors.New("result must be a pointer")
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we have the file, we really decode the root node
|
||||||
|
if f, ok := n.(*ast.File); ok {
|
||||||
|
n = f.Node
|
||||||
|
}
|
||||||
|
|
||||||
|
var d decoder
|
||||||
|
return d.decode("root", n, val.Elem())
|
||||||
|
}
|
||||||
|
|
||||||
|
type decoder struct {
|
||||||
|
stack []reflect.Kind
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *decoder) decode(name string, node ast.Node, result reflect.Value) error {
|
||||||
|
k := result
|
||||||
|
|
||||||
|
// If we have an interface with a valid value, we use that
|
||||||
|
// for the check.
|
||||||
|
if result.Kind() == reflect.Interface {
|
||||||
|
elem := result.Elem()
|
||||||
|
if elem.IsValid() {
|
||||||
|
k = elem
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Push current onto stack unless it is an interface.
|
||||||
|
if k.Kind() != reflect.Interface {
|
||||||
|
d.stack = append(d.stack, k.Kind())
|
||||||
|
|
||||||
|
// Schedule a pop
|
||||||
|
defer func() {
|
||||||
|
d.stack = d.stack[:len(d.stack)-1]
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
|
switch k.Kind() {
|
||||||
|
case reflect.Bool:
|
||||||
|
return d.decodeBool(name, node, result)
|
||||||
|
case reflect.Float32, reflect.Float64:
|
||||||
|
return d.decodeFloat(name, node, result)
|
||||||
|
case reflect.Int, reflect.Int32, reflect.Int64:
|
||||||
|
return d.decodeInt(name, node, result)
|
||||||
|
case reflect.Interface:
|
||||||
|
// When we see an interface, we make our own thing
|
||||||
|
return d.decodeInterface(name, node, result)
|
||||||
|
case reflect.Map:
|
||||||
|
return d.decodeMap(name, node, result)
|
||||||
|
case reflect.Ptr:
|
||||||
|
return d.decodePtr(name, node, result)
|
||||||
|
case reflect.Slice:
|
||||||
|
return d.decodeSlice(name, node, result)
|
||||||
|
case reflect.String:
|
||||||
|
return d.decodeString(name, node, result)
|
||||||
|
case reflect.Struct:
|
||||||
|
return d.decodeStruct(name, node, result)
|
||||||
|
default:
|
||||||
|
return &parser.PosError{
|
||||||
|
Pos: node.Pos(),
|
||||||
|
Err: fmt.Errorf("%s: unknown kind to decode into: %s", name, k.Kind()),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *decoder) decodeBool(name string, node ast.Node, result reflect.Value) error {
|
||||||
|
switch n := node.(type) {
|
||||||
|
case *ast.LiteralType:
|
||||||
|
if n.Token.Type == token.BOOL {
|
||||||
|
v, err := strconv.ParseBool(n.Token.Text)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
result.Set(reflect.ValueOf(v))
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return &parser.PosError{
|
||||||
|
Pos: node.Pos(),
|
||||||
|
Err: fmt.Errorf("%s: unknown type %T", name, node),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *decoder) decodeFloat(name string, node ast.Node, result reflect.Value) error {
|
||||||
|
switch n := node.(type) {
|
||||||
|
case *ast.LiteralType:
|
||||||
|
if n.Token.Type == token.FLOAT || n.Token.Type == token.NUMBER {
|
||||||
|
v, err := strconv.ParseFloat(n.Token.Text, 64)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
result.Set(reflect.ValueOf(v).Convert(result.Type()))
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return &parser.PosError{
|
||||||
|
Pos: node.Pos(),
|
||||||
|
Err: fmt.Errorf("%s: unknown type %T", name, node),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *decoder) decodeInt(name string, node ast.Node, result reflect.Value) error {
|
||||||
|
switch n := node.(type) {
|
||||||
|
case *ast.LiteralType:
|
||||||
|
switch n.Token.Type {
|
||||||
|
case token.NUMBER:
|
||||||
|
v, err := strconv.ParseInt(n.Token.Text, 0, 0)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
if result.Kind() == reflect.Interface {
|
||||||
|
result.Set(reflect.ValueOf(int(v)))
|
||||||
|
} else {
|
||||||
|
result.SetInt(v)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
case token.STRING:
|
||||||
|
v, err := strconv.ParseInt(n.Token.Value().(string), 0, 0)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
if result.Kind() == reflect.Interface {
|
||||||
|
result.Set(reflect.ValueOf(int(v)))
|
||||||
|
} else {
|
||||||
|
result.SetInt(v)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return &parser.PosError{
|
||||||
|
Pos: node.Pos(),
|
||||||
|
Err: fmt.Errorf("%s: unknown type %T", name, node),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *decoder) decodeInterface(name string, node ast.Node, result reflect.Value) error {
|
||||||
|
// When we see an ast.Node, we retain the value to enable deferred decoding.
|
||||||
|
// Very useful in situations where we want to preserve ast.Node information
|
||||||
|
// like Pos
|
||||||
|
if result.Type() == nodeType && result.CanSet() {
|
||||||
|
result.Set(reflect.ValueOf(node))
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
var set reflect.Value
|
||||||
|
redecode := true
|
||||||
|
|
||||||
|
// For testing types, ObjectType should just be treated as a list. We
|
||||||
|
// set this to a temporary var because we want to pass in the real node.
|
||||||
|
testNode := node
|
||||||
|
if ot, ok := node.(*ast.ObjectType); ok {
|
||||||
|
testNode = ot.List
|
||||||
|
}
|
||||||
|
|
||||||
|
switch n := testNode.(type) {
|
||||||
|
case *ast.ObjectList:
|
||||||
|
// If we're at the root or we're directly within a slice, then we
|
||||||
|
// decode objects into map[string]interface{}, otherwise we decode
|
||||||
|
// them into lists.
|
||||||
|
if len(d.stack) == 0 || d.stack[len(d.stack)-1] == reflect.Slice {
|
||||||
|
var temp map[string]interface{}
|
||||||
|
tempVal := reflect.ValueOf(temp)
|
||||||
|
result := reflect.MakeMap(
|
||||||
|
reflect.MapOf(
|
||||||
|
reflect.TypeOf(""),
|
||||||
|
tempVal.Type().Elem()))
|
||||||
|
|
||||||
|
set = result
|
||||||
|
} else {
|
||||||
|
var temp []map[string]interface{}
|
||||||
|
tempVal := reflect.ValueOf(temp)
|
||||||
|
result := reflect.MakeSlice(
|
||||||
|
reflect.SliceOf(tempVal.Type().Elem()), 0, len(n.Items))
|
||||||
|
set = result
|
||||||
|
}
|
||||||
|
case *ast.ObjectType:
|
||||||
|
// If we're at the root or we're directly within a slice, then we
|
||||||
|
// decode objects into map[string]interface{}, otherwise we decode
|
||||||
|
// them into lists.
|
||||||
|
if len(d.stack) == 0 || d.stack[len(d.stack)-1] == reflect.Slice {
|
||||||
|
var temp map[string]interface{}
|
||||||
|
tempVal := reflect.ValueOf(temp)
|
||||||
|
result := reflect.MakeMap(
|
||||||
|
reflect.MapOf(
|
||||||
|
reflect.TypeOf(""),
|
||||||
|
tempVal.Type().Elem()))
|
||||||
|
|
||||||
|
set = result
|
||||||
|
} else {
|
||||||
|
var temp []map[string]interface{}
|
||||||
|
tempVal := reflect.ValueOf(temp)
|
||||||
|
result := reflect.MakeSlice(
|
||||||
|
reflect.SliceOf(tempVal.Type().Elem()), 0, 1)
|
||||||
|
set = result
|
||||||
|
}
|
||||||
|
case *ast.ListType:
|
||||||
|
var temp []interface{}
|
||||||
|
tempVal := reflect.ValueOf(temp)
|
||||||
|
result := reflect.MakeSlice(
|
||||||
|
reflect.SliceOf(tempVal.Type().Elem()), 0, 0)
|
||||||
|
set = result
|
||||||
|
case *ast.LiteralType:
|
||||||
|
switch n.Token.Type {
|
||||||
|
case token.BOOL:
|
||||||
|
var result bool
|
||||||
|
set = reflect.Indirect(reflect.New(reflect.TypeOf(result)))
|
||||||
|
case token.FLOAT:
|
||||||
|
var result float64
|
||||||
|
set = reflect.Indirect(reflect.New(reflect.TypeOf(result)))
|
||||||
|
case token.NUMBER:
|
||||||
|
var result int
|
||||||
|
set = reflect.Indirect(reflect.New(reflect.TypeOf(result)))
|
||||||
|
case token.STRING, token.HEREDOC:
|
||||||
|
set = reflect.Indirect(reflect.New(reflect.TypeOf("")))
|
||||||
|
default:
|
||||||
|
return &parser.PosError{
|
||||||
|
Pos: node.Pos(),
|
||||||
|
Err: fmt.Errorf("%s: cannot decode into interface: %T", name, node),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
return fmt.Errorf(
|
||||||
|
"%s: cannot decode into interface: %T",
|
||||||
|
name, node)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Set the result to what its supposed to be, then reset
|
||||||
|
// result so we don't reflect into this method anymore.
|
||||||
|
result.Set(set)
|
||||||
|
|
||||||
|
if redecode {
|
||||||
|
// Revisit the node so that we can use the newly instantiated
|
||||||
|
// thing and populate it.
|
||||||
|
if err := d.decode(name, node, result); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *decoder) decodeMap(name string, node ast.Node, result reflect.Value) error {
|
||||||
|
if item, ok := node.(*ast.ObjectItem); ok {
|
||||||
|
node = &ast.ObjectList{Items: []*ast.ObjectItem{item}}
|
||||||
|
}
|
||||||
|
|
||||||
|
if ot, ok := node.(*ast.ObjectType); ok {
|
||||||
|
node = ot.List
|
||||||
|
}
|
||||||
|
|
||||||
|
n, ok := node.(*ast.ObjectList)
|
||||||
|
if !ok {
|
||||||
|
return &parser.PosError{
|
||||||
|
Pos: node.Pos(),
|
||||||
|
Err: fmt.Errorf("%s: not an object type for map (%T)", name, node),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we have an interface, then we can address the interface,
|
||||||
|
// but not the slice itself, so get the element but set the interface
|
||||||
|
set := result
|
||||||
|
if result.Kind() == reflect.Interface {
|
||||||
|
result = result.Elem()
|
||||||
|
}
|
||||||
|
|
||||||
|
resultType := result.Type()
|
||||||
|
resultElemType := resultType.Elem()
|
||||||
|
resultKeyType := resultType.Key()
|
||||||
|
if resultKeyType.Kind() != reflect.String {
|
||||||
|
return &parser.PosError{
|
||||||
|
Pos: node.Pos(),
|
||||||
|
Err: fmt.Errorf("%s: map must have string keys", name),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Make a map if it is nil
|
||||||
|
resultMap := result
|
||||||
|
if result.IsNil() {
|
||||||
|
resultMap = reflect.MakeMap(
|
||||||
|
reflect.MapOf(resultKeyType, resultElemType))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Go through each element and decode it.
|
||||||
|
done := make(map[string]struct{})
|
||||||
|
for _, item := range n.Items {
|
||||||
|
if item.Val == nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
// github.com/hashicorp/terraform/issue/5740
|
||||||
|
if len(item.Keys) == 0 {
|
||||||
|
return &parser.PosError{
|
||||||
|
Pos: node.Pos(),
|
||||||
|
Err: fmt.Errorf("%s: map must have string keys", name),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Get the key we're dealing with, which is the first item
|
||||||
|
keyStr := item.Keys[0].Token.Value().(string)
|
||||||
|
|
||||||
|
// If we've already processed this key, then ignore it
|
||||||
|
if _, ok := done[keyStr]; ok {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
// Determine the value. If we have more than one key, then we
|
||||||
|
// get the objectlist of only these keys.
|
||||||
|
itemVal := item.Val
|
||||||
|
if len(item.Keys) > 1 {
|
||||||
|
itemVal = n.Filter(keyStr)
|
||||||
|
done[keyStr] = struct{}{}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Make the field name
|
||||||
|
fieldName := fmt.Sprintf("%s.%s", name, keyStr)
|
||||||
|
|
||||||
|
// Get the key/value as reflection values
|
||||||
|
key := reflect.ValueOf(keyStr)
|
||||||
|
val := reflect.Indirect(reflect.New(resultElemType))
|
||||||
|
|
||||||
|
// If we have a pre-existing value in the map, use that
|
||||||
|
oldVal := resultMap.MapIndex(key)
|
||||||
|
if oldVal.IsValid() {
|
||||||
|
val.Set(oldVal)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Decode!
|
||||||
|
if err := d.decode(fieldName, itemVal, val); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Set the value on the map
|
||||||
|
resultMap.SetMapIndex(key, val)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Set the final map if we can
|
||||||
|
set.Set(resultMap)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *decoder) decodePtr(name string, node ast.Node, result reflect.Value) error {
|
||||||
|
// Create an element of the concrete (non pointer) type and decode
|
||||||
|
// into that. Then set the value of the pointer to this type.
|
||||||
|
resultType := result.Type()
|
||||||
|
resultElemType := resultType.Elem()
|
||||||
|
val := reflect.New(resultElemType)
|
||||||
|
if err := d.decode(name, node, reflect.Indirect(val)); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
result.Set(val)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *decoder) decodeSlice(name string, node ast.Node, result reflect.Value) error {
|
||||||
|
// If we have an interface, then we can address the interface,
|
||||||
|
// but not the slice itself, so get the element but set the interface
|
||||||
|
set := result
|
||||||
|
if result.Kind() == reflect.Interface {
|
||||||
|
result = result.Elem()
|
||||||
|
}
|
||||||
|
// Create the slice if it isn't nil
|
||||||
|
resultType := result.Type()
|
||||||
|
resultElemType := resultType.Elem()
|
||||||
|
if result.IsNil() {
|
||||||
|
resultSliceType := reflect.SliceOf(resultElemType)
|
||||||
|
result = reflect.MakeSlice(
|
||||||
|
resultSliceType, 0, 0)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Figure out the items we'll be copying into the slice
|
||||||
|
var items []ast.Node
|
||||||
|
switch n := node.(type) {
|
||||||
|
case *ast.ObjectList:
|
||||||
|
items = make([]ast.Node, len(n.Items))
|
||||||
|
for i, item := range n.Items {
|
||||||
|
items[i] = item
|
||||||
|
}
|
||||||
|
case *ast.ObjectType:
|
||||||
|
items = []ast.Node{n}
|
||||||
|
case *ast.ListType:
|
||||||
|
items = n.List
|
||||||
|
default:
|
||||||
|
return &parser.PosError{
|
||||||
|
Pos: node.Pos(),
|
||||||
|
Err: fmt.Errorf("unknown slice type: %T", node),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for i, item := range items {
|
||||||
|
fieldName := fmt.Sprintf("%s[%d]", name, i)
|
||||||
|
|
||||||
|
// Decode
|
||||||
|
val := reflect.Indirect(reflect.New(resultElemType))
|
||||||
|
|
||||||
|
// if item is an object that was decoded from ambiguous JSON and
|
||||||
|
// flattened, make sure it's expanded if it needs to decode into a
|
||||||
|
// defined structure.
|
||||||
|
item := expandObject(item, val)
|
||||||
|
|
||||||
|
if err := d.decode(fieldName, item, val); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Append it onto the slice
|
||||||
|
result = reflect.Append(result, val)
|
||||||
|
}
|
||||||
|
|
||||||
|
set.Set(result)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// expandObject detects if an ambiguous JSON object was flattened to a List which
|
||||||
|
// should be decoded into a struct, and expands the ast to properly deocode.
|
||||||
|
func expandObject(node ast.Node, result reflect.Value) ast.Node {
|
||||||
|
item, ok := node.(*ast.ObjectItem)
|
||||||
|
if !ok {
|
||||||
|
return node
|
||||||
|
}
|
||||||
|
|
||||||
|
elemType := result.Type()
|
||||||
|
|
||||||
|
// our target type must be a struct
|
||||||
|
switch elemType.Kind() {
|
||||||
|
case reflect.Ptr:
|
||||||
|
switch elemType.Elem().Kind() {
|
||||||
|
case reflect.Struct:
|
||||||
|
//OK
|
||||||
|
default:
|
||||||
|
return node
|
||||||
|
}
|
||||||
|
case reflect.Struct:
|
||||||
|
//OK
|
||||||
|
default:
|
||||||
|
return node
|
||||||
|
}
|
||||||
|
|
||||||
|
// A list value will have a key and field name. If it had more fields,
|
||||||
|
// it wouldn't have been flattened.
|
||||||
|
if len(item.Keys) != 2 {
|
||||||
|
return node
|
||||||
|
}
|
||||||
|
|
||||||
|
keyToken := item.Keys[0].Token
|
||||||
|
item.Keys = item.Keys[1:]
|
||||||
|
|
||||||
|
// we need to un-flatten the ast enough to decode
|
||||||
|
newNode := &ast.ObjectItem{
|
||||||
|
Keys: []*ast.ObjectKey{
|
||||||
|
&ast.ObjectKey{
|
||||||
|
Token: keyToken,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
Val: &ast.ObjectType{
|
||||||
|
List: &ast.ObjectList{
|
||||||
|
Items: []*ast.ObjectItem{item},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
return newNode
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *decoder) decodeString(name string, node ast.Node, result reflect.Value) error {
|
||||||
|
switch n := node.(type) {
|
||||||
|
case *ast.LiteralType:
|
||||||
|
switch n.Token.Type {
|
||||||
|
case token.NUMBER:
|
||||||
|
result.Set(reflect.ValueOf(n.Token.Text).Convert(result.Type()))
|
||||||
|
return nil
|
||||||
|
case token.STRING, token.HEREDOC:
|
||||||
|
result.Set(reflect.ValueOf(n.Token.Value()).Convert(result.Type()))
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return &parser.PosError{
|
||||||
|
Pos: node.Pos(),
|
||||||
|
Err: fmt.Errorf("%s: unknown type for string %T", name, node),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *decoder) decodeStruct(name string, node ast.Node, result reflect.Value) error {
|
||||||
|
var item *ast.ObjectItem
|
||||||
|
if it, ok := node.(*ast.ObjectItem); ok {
|
||||||
|
item = it
|
||||||
|
node = it.Val
|
||||||
|
}
|
||||||
|
|
||||||
|
if ot, ok := node.(*ast.ObjectType); ok {
|
||||||
|
node = ot.List
|
||||||
|
}
|
||||||
|
|
||||||
|
// Handle the special case where the object itself is a literal. Previously
|
||||||
|
// the yacc parser would always ensure top-level elements were arrays. The new
|
||||||
|
// parser does not make the same guarantees, thus we need to convert any
|
||||||
|
// top-level literal elements into a list.
|
||||||
|
if _, ok := node.(*ast.LiteralType); ok && item != nil {
|
||||||
|
node = &ast.ObjectList{Items: []*ast.ObjectItem{item}}
|
||||||
|
}
|
||||||
|
|
||||||
|
list, ok := node.(*ast.ObjectList)
|
||||||
|
if !ok {
|
||||||
|
return &parser.PosError{
|
||||||
|
Pos: node.Pos(),
|
||||||
|
Err: fmt.Errorf("%s: not an object type for struct (%T)", name, node),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// This slice will keep track of all the structs we'll be decoding.
|
||||||
|
// There can be more than one struct if there are embedded structs
|
||||||
|
// that are squashed.
|
||||||
|
structs := make([]reflect.Value, 1, 5)
|
||||||
|
structs[0] = result
|
||||||
|
|
||||||
|
// Compile the list of all the fields that we're going to be decoding
|
||||||
|
// from all the structs.
|
||||||
|
type field struct {
|
||||||
|
field reflect.StructField
|
||||||
|
val reflect.Value
|
||||||
|
}
|
||||||
|
fields := []field{}
|
||||||
|
for len(structs) > 0 {
|
||||||
|
structVal := structs[0]
|
||||||
|
structs = structs[1:]
|
||||||
|
|
||||||
|
structType := structVal.Type()
|
||||||
|
for i := 0; i < structType.NumField(); i++ {
|
||||||
|
fieldType := structType.Field(i)
|
||||||
|
tagParts := strings.Split(fieldType.Tag.Get(tagName), ",")
|
||||||
|
|
||||||
|
// Ignore fields with tag name "-"
|
||||||
|
if tagParts[0] == "-" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
if fieldType.Anonymous {
|
||||||
|
fieldKind := fieldType.Type.Kind()
|
||||||
|
if fieldKind != reflect.Struct {
|
||||||
|
return &parser.PosError{
|
||||||
|
Pos: node.Pos(),
|
||||||
|
Err: fmt.Errorf("%s: unsupported type to struct: %s",
|
||||||
|
fieldType.Name, fieldKind),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// We have an embedded field. We "squash" the fields down
|
||||||
|
// if specified in the tag.
|
||||||
|
squash := false
|
||||||
|
for _, tag := range tagParts[1:] {
|
||||||
|
if tag == "squash" {
|
||||||
|
squash = true
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if squash {
|
||||||
|
structs = append(
|
||||||
|
structs, result.FieldByName(fieldType.Name))
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Normal struct field, store it away
|
||||||
|
fields = append(fields, field{fieldType, structVal.Field(i)})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
usedKeys := make(map[string]struct{})
|
||||||
|
decodedFields := make([]string, 0, len(fields))
|
||||||
|
decodedFieldsVal := make([]reflect.Value, 0)
|
||||||
|
unusedKeysVal := make([]reflect.Value, 0)
|
||||||
|
for _, f := range fields {
|
||||||
|
field, fieldValue := f.field, f.val
|
||||||
|
if !fieldValue.IsValid() {
|
||||||
|
// This should never happen
|
||||||
|
panic("field is not valid")
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we can't set the field, then it is unexported or something,
|
||||||
|
// and we just continue onwards.
|
||||||
|
if !fieldValue.CanSet() {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
fieldName := field.Name
|
||||||
|
|
||||||
|
tagValue := field.Tag.Get(tagName)
|
||||||
|
tagParts := strings.SplitN(tagValue, ",", 2)
|
||||||
|
if len(tagParts) >= 2 {
|
||||||
|
switch tagParts[1] {
|
||||||
|
case "decodedFields":
|
||||||
|
decodedFieldsVal = append(decodedFieldsVal, fieldValue)
|
||||||
|
continue
|
||||||
|
case "key":
|
||||||
|
if item == nil {
|
||||||
|
return &parser.PosError{
|
||||||
|
Pos: node.Pos(),
|
||||||
|
Err: fmt.Errorf("%s: %s asked for 'key', impossible",
|
||||||
|
name, fieldName),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fieldValue.SetString(item.Keys[0].Token.Value().(string))
|
||||||
|
continue
|
||||||
|
case "unusedKeys":
|
||||||
|
unusedKeysVal = append(unusedKeysVal, fieldValue)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if tagParts[0] != "" {
|
||||||
|
fieldName = tagParts[0]
|
||||||
|
}
|
||||||
|
|
||||||
|
// Determine the element we'll use to decode. If it is a single
|
||||||
|
// match (only object with the field), then we decode it exactly.
|
||||||
|
// If it is a prefix match, then we decode the matches.
|
||||||
|
filter := list.Filter(fieldName)
|
||||||
|
|
||||||
|
prefixMatches := filter.Children()
|
||||||
|
matches := filter.Elem()
|
||||||
|
if len(matches.Items) == 0 && len(prefixMatches.Items) == 0 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
// Track the used key
|
||||||
|
usedKeys[fieldName] = struct{}{}
|
||||||
|
|
||||||
|
// Create the field name and decode. We range over the elements
|
||||||
|
// because we actually want the value.
|
||||||
|
fieldName = fmt.Sprintf("%s.%s", name, fieldName)
|
||||||
|
if len(prefixMatches.Items) > 0 {
|
||||||
|
if err := d.decode(fieldName, prefixMatches, fieldValue); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for _, match := range matches.Items {
|
||||||
|
var decodeNode ast.Node = match.Val
|
||||||
|
if ot, ok := decodeNode.(*ast.ObjectType); ok {
|
||||||
|
decodeNode = &ast.ObjectList{Items: ot.List.Items}
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := d.decode(fieldName, decodeNode, fieldValue); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
decodedFields = append(decodedFields, field.Name)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(decodedFieldsVal) > 0 {
|
||||||
|
// Sort it so that it is deterministic
|
||||||
|
sort.Strings(decodedFields)
|
||||||
|
|
||||||
|
for _, v := range decodedFieldsVal {
|
||||||
|
v.Set(reflect.ValueOf(decodedFields))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// findNodeType returns the type of ast.Node
|
||||||
|
func findNodeType() reflect.Type {
|
||||||
|
var nodeContainer struct {
|
||||||
|
Node ast.Node
|
||||||
|
}
|
||||||
|
value := reflect.ValueOf(nodeContainer).FieldByName("Node")
|
||||||
|
return value.Type()
|
||||||
|
}
|
|
@ -0,0 +1,11 @@
|
||||||
|
// Package hcl decodes HCL into usable Go structures.
|
||||||
|
//
|
||||||
|
// hcl input can come in either pure HCL format or JSON format.
|
||||||
|
// It can be parsed into an AST, and then decoded into a structure,
|
||||||
|
// or it can be decoded directly from a string into a structure.
|
||||||
|
//
|
||||||
|
// If you choose to parse HCL into a raw AST, the benefit is that you
|
||||||
|
// can write custom visitor implementations to implement custom
|
||||||
|
// semantic checks. By default, HCL does not perform any semantic
|
||||||
|
// checks.
|
||||||
|
package hcl
|
|
@ -0,0 +1,219 @@
|
||||||
|
// Package ast declares the types used to represent syntax trees for HCL
|
||||||
|
// (HashiCorp Configuration Language)
|
||||||
|
package ast
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/hashicorp/hcl/hcl/token"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Node is an element in the abstract syntax tree.
|
||||||
|
type Node interface {
|
||||||
|
node()
|
||||||
|
Pos() token.Pos
|
||||||
|
}
|
||||||
|
|
||||||
|
func (File) node() {}
|
||||||
|
func (ObjectList) node() {}
|
||||||
|
func (ObjectKey) node() {}
|
||||||
|
func (ObjectItem) node() {}
|
||||||
|
func (Comment) node() {}
|
||||||
|
func (CommentGroup) node() {}
|
||||||
|
func (ObjectType) node() {}
|
||||||
|
func (LiteralType) node() {}
|
||||||
|
func (ListType) node() {}
|
||||||
|
|
||||||
|
// File represents a single HCL file
|
||||||
|
type File struct {
|
||||||
|
Node Node // usually a *ObjectList
|
||||||
|
Comments []*CommentGroup // list of all comments in the source
|
||||||
|
}
|
||||||
|
|
||||||
|
func (f *File) Pos() token.Pos {
|
||||||
|
return f.Node.Pos()
|
||||||
|
}
|
||||||
|
|
||||||
|
// ObjectList represents a list of ObjectItems. An HCL file itself is an
|
||||||
|
// ObjectList.
|
||||||
|
type ObjectList struct {
|
||||||
|
Items []*ObjectItem
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *ObjectList) Add(item *ObjectItem) {
|
||||||
|
o.Items = append(o.Items, item)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Filter filters out the objects with the given key list as a prefix.
|
||||||
|
//
|
||||||
|
// The returned list of objects contain ObjectItems where the keys have
|
||||||
|
// this prefix already stripped off. This might result in objects with
|
||||||
|
// zero-length key lists if they have no children.
|
||||||
|
//
|
||||||
|
// If no matches are found, an empty ObjectList (non-nil) is returned.
|
||||||
|
func (o *ObjectList) Filter(keys ...string) *ObjectList {
|
||||||
|
var result ObjectList
|
||||||
|
for _, item := range o.Items {
|
||||||
|
// If there aren't enough keys, then ignore this
|
||||||
|
if len(item.Keys) < len(keys) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
match := true
|
||||||
|
for i, key := range item.Keys[:len(keys)] {
|
||||||
|
key := key.Token.Value().(string)
|
||||||
|
if key != keys[i] && !strings.EqualFold(key, keys[i]) {
|
||||||
|
match = false
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if !match {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
// Strip off the prefix from the children
|
||||||
|
newItem := *item
|
||||||
|
newItem.Keys = newItem.Keys[len(keys):]
|
||||||
|
result.Add(&newItem)
|
||||||
|
}
|
||||||
|
|
||||||
|
return &result
|
||||||
|
}
|
||||||
|
|
||||||
|
// Children returns further nested objects (key length > 0) within this
|
||||||
|
// ObjectList. This should be used with Filter to get at child items.
|
||||||
|
func (o *ObjectList) Children() *ObjectList {
|
||||||
|
var result ObjectList
|
||||||
|
for _, item := range o.Items {
|
||||||
|
if len(item.Keys) > 0 {
|
||||||
|
result.Add(item)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return &result
|
||||||
|
}
|
||||||
|
|
||||||
|
// Elem returns items in the list that are direct element assignments
|
||||||
|
// (key length == 0). This should be used with Filter to get at elements.
|
||||||
|
func (o *ObjectList) Elem() *ObjectList {
|
||||||
|
var result ObjectList
|
||||||
|
for _, item := range o.Items {
|
||||||
|
if len(item.Keys) == 0 {
|
||||||
|
result.Add(item)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return &result
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *ObjectList) Pos() token.Pos {
|
||||||
|
// always returns the uninitiliazed position
|
||||||
|
return o.Items[0].Pos()
|
||||||
|
}
|
||||||
|
|
||||||
|
// ObjectItem represents a HCL Object Item. An item is represented with a key
|
||||||
|
// (or keys). It can be an assignment or an object (both normal and nested)
|
||||||
|
type ObjectItem struct {
|
||||||
|
// keys is only one length long if it's of type assignment. If it's a
|
||||||
|
// nested object it can be larger than one. In that case "assign" is
|
||||||
|
// invalid as there is no assignments for a nested object.
|
||||||
|
Keys []*ObjectKey
|
||||||
|
|
||||||
|
// assign contains the position of "=", if any
|
||||||
|
Assign token.Pos
|
||||||
|
|
||||||
|
// val is the item itself. It can be an object,list, number, bool or a
|
||||||
|
// string. If key length is larger than one, val can be only of type
|
||||||
|
// Object.
|
||||||
|
Val Node
|
||||||
|
|
||||||
|
LeadComment *CommentGroup // associated lead comment
|
||||||
|
LineComment *CommentGroup // associated line comment
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *ObjectItem) Pos() token.Pos {
|
||||||
|
// I'm not entirely sure what causes this, but removing this causes
|
||||||
|
// a test failure. We should investigate at some point.
|
||||||
|
if len(o.Keys) == 0 {
|
||||||
|
return token.Pos{}
|
||||||
|
}
|
||||||
|
|
||||||
|
return o.Keys[0].Pos()
|
||||||
|
}
|
||||||
|
|
||||||
|
// ObjectKeys are either an identifier or of type string.
|
||||||
|
type ObjectKey struct {
|
||||||
|
Token token.Token
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *ObjectKey) Pos() token.Pos {
|
||||||
|
return o.Token.Pos
|
||||||
|
}
|
||||||
|
|
||||||
|
// LiteralType represents a literal of basic type. Valid types are:
|
||||||
|
// token.NUMBER, token.FLOAT, token.BOOL and token.STRING
|
||||||
|
type LiteralType struct {
|
||||||
|
Token token.Token
|
||||||
|
|
||||||
|
// comment types, only used when in a list
|
||||||
|
LeadComment *CommentGroup
|
||||||
|
LineComment *CommentGroup
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l *LiteralType) Pos() token.Pos {
|
||||||
|
return l.Token.Pos
|
||||||
|
}
|
||||||
|
|
||||||
|
// ListStatement represents a HCL List type
|
||||||
|
type ListType struct {
|
||||||
|
Lbrack token.Pos // position of "["
|
||||||
|
Rbrack token.Pos // position of "]"
|
||||||
|
List []Node // the elements in lexical order
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l *ListType) Pos() token.Pos {
|
||||||
|
return l.Lbrack
|
||||||
|
}
|
||||||
|
|
||||||
|
func (l *ListType) Add(node Node) {
|
||||||
|
l.List = append(l.List, node)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ObjectType represents a HCL Object Type
|
||||||
|
type ObjectType struct {
|
||||||
|
Lbrace token.Pos // position of "{"
|
||||||
|
Rbrace token.Pos // position of "}"
|
||||||
|
List *ObjectList // the nodes in lexical order
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *ObjectType) Pos() token.Pos {
|
||||||
|
return o.Lbrace
|
||||||
|
}
|
||||||
|
|
||||||
|
// Comment node represents a single //, # style or /*- style commment
|
||||||
|
type Comment struct {
|
||||||
|
Start token.Pos // position of / or #
|
||||||
|
Text string
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Comment) Pos() token.Pos {
|
||||||
|
return c.Start
|
||||||
|
}
|
||||||
|
|
||||||
|
// CommentGroup node represents a sequence of comments with no other tokens and
|
||||||
|
// no empty lines between.
|
||||||
|
type CommentGroup struct {
|
||||||
|
List []*Comment // len(List) > 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *CommentGroup) Pos() token.Pos {
|
||||||
|
return c.List[0].Pos()
|
||||||
|
}
|
||||||
|
|
||||||
|
//-------------------------------------------------------------------
|
||||||
|
// GoStringer
|
||||||
|
//-------------------------------------------------------------------
|
||||||
|
|
||||||
|
func (o *ObjectKey) GoString() string { return fmt.Sprintf("*%#v", *o) }
|
||||||
|
func (o *ObjectList) GoString() string { return fmt.Sprintf("*%#v", *o) }
|
|
@ -0,0 +1,52 @@
|
||||||
|
package ast
|
||||||
|
|
||||||
|
import "fmt"
|
||||||
|
|
||||||
|
// WalkFunc describes a function to be called for each node during a Walk. The
|
||||||
|
// returned node can be used to rewrite the AST. Walking stops the returned
|
||||||
|
// bool is false.
|
||||||
|
type WalkFunc func(Node) (Node, bool)
|
||||||
|
|
||||||
|
// Walk traverses an AST in depth-first order: It starts by calling fn(node);
|
||||||
|
// node must not be nil. If fn returns true, Walk invokes fn recursively for
|
||||||
|
// each of the non-nil children of node, followed by a call of fn(nil). The
|
||||||
|
// returned node of fn can be used to rewrite the passed node to fn.
|
||||||
|
func Walk(node Node, fn WalkFunc) Node {
|
||||||
|
rewritten, ok := fn(node)
|
||||||
|
if !ok {
|
||||||
|
return rewritten
|
||||||
|
}
|
||||||
|
|
||||||
|
switch n := node.(type) {
|
||||||
|
case *File:
|
||||||
|
n.Node = Walk(n.Node, fn)
|
||||||
|
case *ObjectList:
|
||||||
|
for i, item := range n.Items {
|
||||||
|
n.Items[i] = Walk(item, fn).(*ObjectItem)
|
||||||
|
}
|
||||||
|
case *ObjectKey:
|
||||||
|
// nothing to do
|
||||||
|
case *ObjectItem:
|
||||||
|
for i, k := range n.Keys {
|
||||||
|
n.Keys[i] = Walk(k, fn).(*ObjectKey)
|
||||||
|
}
|
||||||
|
|
||||||
|
if n.Val != nil {
|
||||||
|
n.Val = Walk(n.Val, fn)
|
||||||
|
}
|
||||||
|
case *LiteralType:
|
||||||
|
// nothing to do
|
||||||
|
case *ListType:
|
||||||
|
for i, l := range n.List {
|
||||||
|
n.List[i] = Walk(l, fn)
|
||||||
|
}
|
||||||
|
case *ObjectType:
|
||||||
|
n.List = Walk(n.List, fn).(*ObjectList)
|
||||||
|
default:
|
||||||
|
// should we panic here?
|
||||||
|
fmt.Printf("unknown type: %T\n", n)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn(nil)
|
||||||
|
return rewritten
|
||||||
|
}
|
|
@ -0,0 +1,17 @@
|
||||||
|
package parser
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/hashicorp/hcl/hcl/token"
|
||||||
|
)
|
||||||
|
|
||||||
|
// PosError is a parse error that contains a position.
|
||||||
|
type PosError struct {
|
||||||
|
Pos token.Pos
|
||||||
|
Err error
|
||||||
|
}
|
||||||
|
|
||||||
|
func (e *PosError) Error() string {
|
||||||
|
return fmt.Sprintf("At %s: %s", e.Pos, e.Err)
|
||||||
|
}
|
|
@ -0,0 +1,532 @@
|
||||||
|
// Package parser implements a parser for HCL (HashiCorp Configuration
|
||||||
|
// Language)
|
||||||
|
package parser
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/hashicorp/hcl/hcl/ast"
|
||||||
|
"github.com/hashicorp/hcl/hcl/scanner"
|
||||||
|
"github.com/hashicorp/hcl/hcl/token"
|
||||||
|
)
|
||||||
|
|
||||||
|
type Parser struct {
|
||||||
|
sc *scanner.Scanner
|
||||||
|
|
||||||
|
// Last read token
|
||||||
|
tok token.Token
|
||||||
|
commaPrev token.Token
|
||||||
|
|
||||||
|
comments []*ast.CommentGroup
|
||||||
|
leadComment *ast.CommentGroup // last lead comment
|
||||||
|
lineComment *ast.CommentGroup // last line comment
|
||||||
|
|
||||||
|
enableTrace bool
|
||||||
|
indent int
|
||||||
|
n int // buffer size (max = 1)
|
||||||
|
}
|
||||||
|
|
||||||
|
func newParser(src []byte) *Parser {
|
||||||
|
return &Parser{
|
||||||
|
sc: scanner.New(src),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parse returns the fully parsed source and returns the abstract syntax tree.
|
||||||
|
func Parse(src []byte) (*ast.File, error) {
|
||||||
|
// normalize all line endings
|
||||||
|
// since the scanner and output only work with "\n" line endings, we may
|
||||||
|
// end up with dangling "\r" characters in the parsed data.
|
||||||
|
src = bytes.Replace(src, []byte("\r\n"), []byte("\n"), -1)
|
||||||
|
|
||||||
|
p := newParser(src)
|
||||||
|
return p.Parse()
|
||||||
|
}
|
||||||
|
|
||||||
|
var errEofToken = errors.New("EOF token found")
|
||||||
|
|
||||||
|
// Parse returns the fully parsed source and returns the abstract syntax tree.
|
||||||
|
func (p *Parser) Parse() (*ast.File, error) {
|
||||||
|
f := &ast.File{}
|
||||||
|
var err, scerr error
|
||||||
|
p.sc.Error = func(pos token.Pos, msg string) {
|
||||||
|
scerr = &PosError{Pos: pos, Err: errors.New(msg)}
|
||||||
|
}
|
||||||
|
|
||||||
|
f.Node, err = p.objectList(false)
|
||||||
|
if scerr != nil {
|
||||||
|
return nil, scerr
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
f.Comments = p.comments
|
||||||
|
return f, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// objectList parses a list of items within an object (generally k/v pairs).
|
||||||
|
// The parameter" obj" tells this whether to we are within an object (braces:
|
||||||
|
// '{', '}') or just at the top level. If we're within an object, we end
|
||||||
|
// at an RBRACE.
|
||||||
|
func (p *Parser) objectList(obj bool) (*ast.ObjectList, error) {
|
||||||
|
defer un(trace(p, "ParseObjectList"))
|
||||||
|
node := &ast.ObjectList{}
|
||||||
|
|
||||||
|
for {
|
||||||
|
if obj {
|
||||||
|
tok := p.scan()
|
||||||
|
p.unscan()
|
||||||
|
if tok.Type == token.RBRACE {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
n, err := p.objectItem()
|
||||||
|
if err == errEofToken {
|
||||||
|
break // we are finished
|
||||||
|
}
|
||||||
|
|
||||||
|
// we don't return a nil node, because might want to use already
|
||||||
|
// collected items.
|
||||||
|
if err != nil {
|
||||||
|
return node, err
|
||||||
|
}
|
||||||
|
|
||||||
|
node.Add(n)
|
||||||
|
|
||||||
|
// object lists can be optionally comma-delimited e.g. when a list of maps
|
||||||
|
// is being expressed, so a comma is allowed here - it's simply consumed
|
||||||
|
tok := p.scan()
|
||||||
|
if tok.Type != token.COMMA {
|
||||||
|
p.unscan()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return node, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *Parser) consumeComment() (comment *ast.Comment, endline int) {
|
||||||
|
endline = p.tok.Pos.Line
|
||||||
|
|
||||||
|
// count the endline if it's multiline comment, ie starting with /*
|
||||||
|
if len(p.tok.Text) > 1 && p.tok.Text[1] == '*' {
|
||||||
|
// don't use range here - no need to decode Unicode code points
|
||||||
|
for i := 0; i < len(p.tok.Text); i++ {
|
||||||
|
if p.tok.Text[i] == '\n' {
|
||||||
|
endline++
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
comment = &ast.Comment{Start: p.tok.Pos, Text: p.tok.Text}
|
||||||
|
p.tok = p.sc.Scan()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *Parser) consumeCommentGroup(n int) (comments *ast.CommentGroup, endline int) {
|
||||||
|
var list []*ast.Comment
|
||||||
|
endline = p.tok.Pos.Line
|
||||||
|
|
||||||
|
for p.tok.Type == token.COMMENT && p.tok.Pos.Line <= endline+n {
|
||||||
|
var comment *ast.Comment
|
||||||
|
comment, endline = p.consumeComment()
|
||||||
|
list = append(list, comment)
|
||||||
|
}
|
||||||
|
|
||||||
|
// add comment group to the comments list
|
||||||
|
comments = &ast.CommentGroup{List: list}
|
||||||
|
p.comments = append(p.comments, comments)
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// objectItem parses a single object item
|
||||||
|
func (p *Parser) objectItem() (*ast.ObjectItem, error) {
|
||||||
|
defer un(trace(p, "ParseObjectItem"))
|
||||||
|
|
||||||
|
keys, err := p.objectKey()
|
||||||
|
if len(keys) > 0 && err == errEofToken {
|
||||||
|
// We ignore eof token here since it is an error if we didn't
|
||||||
|
// receive a value (but we did receive a key) for the item.
|
||||||
|
err = nil
|
||||||
|
}
|
||||||
|
if len(keys) > 0 && err != nil && p.tok.Type == token.RBRACE {
|
||||||
|
// This is a strange boolean statement, but what it means is:
|
||||||
|
// We have keys with no value, and we're likely in an object
|
||||||
|
// (since RBrace ends an object). For this, we set err to nil so
|
||||||
|
// we continue and get the error below of having the wrong value
|
||||||
|
// type.
|
||||||
|
err = nil
|
||||||
|
|
||||||
|
// Reset the token type so we don't think it completed fine. See
|
||||||
|
// objectType which uses p.tok.Type to check if we're done with
|
||||||
|
// the object.
|
||||||
|
p.tok.Type = token.EOF
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
o := &ast.ObjectItem{
|
||||||
|
Keys: keys,
|
||||||
|
}
|
||||||
|
|
||||||
|
if p.leadComment != nil {
|
||||||
|
o.LeadComment = p.leadComment
|
||||||
|
p.leadComment = nil
|
||||||
|
}
|
||||||
|
|
||||||
|
switch p.tok.Type {
|
||||||
|
case token.ASSIGN:
|
||||||
|
o.Assign = p.tok.Pos
|
||||||
|
o.Val, err = p.object()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
case token.LBRACE:
|
||||||
|
o.Val, err = p.objectType()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
keyStr := make([]string, 0, len(keys))
|
||||||
|
for _, k := range keys {
|
||||||
|
keyStr = append(keyStr, k.Token.Text)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil, &PosError{
|
||||||
|
Pos: p.tok.Pos,
|
||||||
|
Err: fmt.Errorf(
|
||||||
|
"key '%s' expected start of object ('{') or assignment ('=')",
|
||||||
|
strings.Join(keyStr, " ")),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// key=#comment
|
||||||
|
// val
|
||||||
|
if p.lineComment != nil {
|
||||||
|
o.LineComment, p.lineComment = p.lineComment, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// do a look-ahead for line comment
|
||||||
|
p.scan()
|
||||||
|
if len(keys) > 0 && o.Val.Pos().Line == keys[0].Pos().Line && p.lineComment != nil {
|
||||||
|
o.LineComment = p.lineComment
|
||||||
|
p.lineComment = nil
|
||||||
|
}
|
||||||
|
p.unscan()
|
||||||
|
return o, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// objectKey parses an object key and returns a ObjectKey AST
|
||||||
|
func (p *Parser) objectKey() ([]*ast.ObjectKey, error) {
|
||||||
|
keyCount := 0
|
||||||
|
keys := make([]*ast.ObjectKey, 0)
|
||||||
|
|
||||||
|
for {
|
||||||
|
tok := p.scan()
|
||||||
|
switch tok.Type {
|
||||||
|
case token.EOF:
|
||||||
|
// It is very important to also return the keys here as well as
|
||||||
|
// the error. This is because we need to be able to tell if we
|
||||||
|
// did parse keys prior to finding the EOF, or if we just found
|
||||||
|
// a bare EOF.
|
||||||
|
return keys, errEofToken
|
||||||
|
case token.ASSIGN:
|
||||||
|
// assignment or object only, but not nested objects. this is not
|
||||||
|
// allowed: `foo bar = {}`
|
||||||
|
if keyCount > 1 {
|
||||||
|
return nil, &PosError{
|
||||||
|
Pos: p.tok.Pos,
|
||||||
|
Err: fmt.Errorf("nested object expected: LBRACE got: %s", p.tok.Type),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if keyCount == 0 {
|
||||||
|
return nil, &PosError{
|
||||||
|
Pos: p.tok.Pos,
|
||||||
|
Err: errors.New("no object keys found!"),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return keys, nil
|
||||||
|
case token.LBRACE:
|
||||||
|
var err error
|
||||||
|
|
||||||
|
// If we have no keys, then it is a syntax error. i.e. {{}} is not
|
||||||
|
// allowed.
|
||||||
|
if len(keys) == 0 {
|
||||||
|
err = &PosError{
|
||||||
|
Pos: p.tok.Pos,
|
||||||
|
Err: fmt.Errorf("expected: IDENT | STRING got: %s", p.tok.Type),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// object
|
||||||
|
return keys, err
|
||||||
|
case token.IDENT, token.STRING:
|
||||||
|
keyCount++
|
||||||
|
keys = append(keys, &ast.ObjectKey{Token: p.tok})
|
||||||
|
case token.ILLEGAL:
|
||||||
|
return keys, &PosError{
|
||||||
|
Pos: p.tok.Pos,
|
||||||
|
Err: fmt.Errorf("illegal character"),
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
return keys, &PosError{
|
||||||
|
Pos: p.tok.Pos,
|
||||||
|
Err: fmt.Errorf("expected: IDENT | STRING | ASSIGN | LBRACE got: %s", p.tok.Type),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// object parses any type of object, such as number, bool, string, object or
|
||||||
|
// list.
|
||||||
|
func (p *Parser) object() (ast.Node, error) {
|
||||||
|
defer un(trace(p, "ParseType"))
|
||||||
|
tok := p.scan()
|
||||||
|
|
||||||
|
switch tok.Type {
|
||||||
|
case token.NUMBER, token.FLOAT, token.BOOL, token.STRING, token.HEREDOC:
|
||||||
|
return p.literalType()
|
||||||
|
case token.LBRACE:
|
||||||
|
return p.objectType()
|
||||||
|
case token.LBRACK:
|
||||||
|
return p.listType()
|
||||||
|
case token.COMMENT:
|
||||||
|
// implement comment
|
||||||
|
case token.EOF:
|
||||||
|
return nil, errEofToken
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil, &PosError{
|
||||||
|
Pos: tok.Pos,
|
||||||
|
Err: fmt.Errorf("Unknown token: %+v", tok),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// objectType parses an object type and returns a ObjectType AST
|
||||||
|
func (p *Parser) objectType() (*ast.ObjectType, error) {
|
||||||
|
defer un(trace(p, "ParseObjectType"))
|
||||||
|
|
||||||
|
// we assume that the currently scanned token is a LBRACE
|
||||||
|
o := &ast.ObjectType{
|
||||||
|
Lbrace: p.tok.Pos,
|
||||||
|
}
|
||||||
|
|
||||||
|
l, err := p.objectList(true)
|
||||||
|
|
||||||
|
// if we hit RBRACE, we are good to go (means we parsed all Items), if it's
|
||||||
|
// not a RBRACE, it's an syntax error and we just return it.
|
||||||
|
if err != nil && p.tok.Type != token.RBRACE {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// No error, scan and expect the ending to be a brace
|
||||||
|
if tok := p.scan(); tok.Type != token.RBRACE {
|
||||||
|
return nil, &PosError{
|
||||||
|
Pos: tok.Pos,
|
||||||
|
Err: fmt.Errorf("object expected closing RBRACE got: %s", tok.Type),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
o.List = l
|
||||||
|
o.Rbrace = p.tok.Pos // advanced via parseObjectList
|
||||||
|
return o, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// listType parses a list type and returns a ListType AST
|
||||||
|
func (p *Parser) listType() (*ast.ListType, error) {
|
||||||
|
defer un(trace(p, "ParseListType"))
|
||||||
|
|
||||||
|
// we assume that the currently scanned token is a LBRACK
|
||||||
|
l := &ast.ListType{
|
||||||
|
Lbrack: p.tok.Pos,
|
||||||
|
}
|
||||||
|
|
||||||
|
needComma := false
|
||||||
|
for {
|
||||||
|
tok := p.scan()
|
||||||
|
if needComma {
|
||||||
|
switch tok.Type {
|
||||||
|
case token.COMMA, token.RBRACK:
|
||||||
|
default:
|
||||||
|
return nil, &PosError{
|
||||||
|
Pos: tok.Pos,
|
||||||
|
Err: fmt.Errorf(
|
||||||
|
"error parsing list, expected comma or list end, got: %s",
|
||||||
|
tok.Type),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
switch tok.Type {
|
||||||
|
case token.BOOL, token.NUMBER, token.FLOAT, token.STRING, token.HEREDOC:
|
||||||
|
node, err := p.literalType()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// If there is a lead comment, apply it
|
||||||
|
if p.leadComment != nil {
|
||||||
|
node.LeadComment = p.leadComment
|
||||||
|
p.leadComment = nil
|
||||||
|
}
|
||||||
|
|
||||||
|
l.Add(node)
|
||||||
|
needComma = true
|
||||||
|
case token.COMMA:
|
||||||
|
// get next list item or we are at the end
|
||||||
|
// do a look-ahead for line comment
|
||||||
|
p.scan()
|
||||||
|
if p.lineComment != nil && len(l.List) > 0 {
|
||||||
|
lit, ok := l.List[len(l.List)-1].(*ast.LiteralType)
|
||||||
|
if ok {
|
||||||
|
lit.LineComment = p.lineComment
|
||||||
|
l.List[len(l.List)-1] = lit
|
||||||
|
p.lineComment = nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
p.unscan()
|
||||||
|
|
||||||
|
needComma = false
|
||||||
|
continue
|
||||||
|
case token.LBRACE:
|
||||||
|
// Looks like a nested object, so parse it out
|
||||||
|
node, err := p.objectType()
|
||||||
|
if err != nil {
|
||||||
|
return nil, &PosError{
|
||||||
|
Pos: tok.Pos,
|
||||||
|
Err: fmt.Errorf(
|
||||||
|
"error while trying to parse object within list: %s", err),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
l.Add(node)
|
||||||
|
needComma = true
|
||||||
|
case token.LBRACK:
|
||||||
|
node, err := p.listType()
|
||||||
|
if err != nil {
|
||||||
|
return nil, &PosError{
|
||||||
|
Pos: tok.Pos,
|
||||||
|
Err: fmt.Errorf(
|
||||||
|
"error while trying to parse list within list: %s", err),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
l.Add(node)
|
||||||
|
case token.RBRACK:
|
||||||
|
// finished
|
||||||
|
l.Rbrack = p.tok.Pos
|
||||||
|
return l, nil
|
||||||
|
default:
|
||||||
|
return nil, &PosError{
|
||||||
|
Pos: tok.Pos,
|
||||||
|
Err: fmt.Errorf("unexpected token while parsing list: %s", tok.Type),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// literalType parses a literal type and returns a LiteralType AST
|
||||||
|
func (p *Parser) literalType() (*ast.LiteralType, error) {
|
||||||
|
defer un(trace(p, "ParseLiteral"))
|
||||||
|
|
||||||
|
return &ast.LiteralType{
|
||||||
|
Token: p.tok,
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// scan returns the next token from the underlying scanner. If a token has
|
||||||
|
// been unscanned then read that instead. In the process, it collects any
|
||||||
|
// comment groups encountered, and remembers the last lead and line comments.
|
||||||
|
func (p *Parser) scan() token.Token {
|
||||||
|
// If we have a token on the buffer, then return it.
|
||||||
|
if p.n != 0 {
|
||||||
|
p.n = 0
|
||||||
|
return p.tok
|
||||||
|
}
|
||||||
|
|
||||||
|
// Otherwise read the next token from the scanner and Save it to the buffer
|
||||||
|
// in case we unscan later.
|
||||||
|
prev := p.tok
|
||||||
|
p.tok = p.sc.Scan()
|
||||||
|
|
||||||
|
if p.tok.Type == token.COMMENT {
|
||||||
|
var comment *ast.CommentGroup
|
||||||
|
var endline int
|
||||||
|
|
||||||
|
// fmt.Printf("p.tok.Pos.Line = %+v prev: %d endline %d \n",
|
||||||
|
// p.tok.Pos.Line, prev.Pos.Line, endline)
|
||||||
|
if p.tok.Pos.Line == prev.Pos.Line {
|
||||||
|
// The comment is on same line as the previous token; it
|
||||||
|
// cannot be a lead comment but may be a line comment.
|
||||||
|
comment, endline = p.consumeCommentGroup(0)
|
||||||
|
if p.tok.Pos.Line != endline {
|
||||||
|
// The next token is on a different line, thus
|
||||||
|
// the last comment group is a line comment.
|
||||||
|
p.lineComment = comment
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// consume successor comments, if any
|
||||||
|
endline = -1
|
||||||
|
for p.tok.Type == token.COMMENT {
|
||||||
|
comment, endline = p.consumeCommentGroup(1)
|
||||||
|
}
|
||||||
|
|
||||||
|
if endline+1 == p.tok.Pos.Line && p.tok.Type != token.RBRACE {
|
||||||
|
switch p.tok.Type {
|
||||||
|
case token.RBRACE, token.RBRACK:
|
||||||
|
// Do not count for these cases
|
||||||
|
default:
|
||||||
|
// The next token is following on the line immediately after the
|
||||||
|
// comment group, thus the last comment group is a lead comment.
|
||||||
|
p.leadComment = comment
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
return p.tok
|
||||||
|
}
|
||||||
|
|
||||||
|
// unscan pushes the previously read token back onto the buffer.
|
||||||
|
func (p *Parser) unscan() {
|
||||||
|
p.n = 1
|
||||||
|
}
|
||||||
|
|
||||||
|
// ----------------------------------------------------------------------------
|
||||||
|
// Parsing support
|
||||||
|
|
||||||
|
func (p *Parser) printTrace(a ...interface{}) {
|
||||||
|
if !p.enableTrace {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
const dots = ". . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . "
|
||||||
|
const n = len(dots)
|
||||||
|
fmt.Printf("%5d:%3d: ", p.tok.Pos.Line, p.tok.Pos.Column)
|
||||||
|
|
||||||
|
i := 2 * p.indent
|
||||||
|
for i > n {
|
||||||
|
fmt.Print(dots)
|
||||||
|
i -= n
|
||||||
|
}
|
||||||
|
// i <= n
|
||||||
|
fmt.Print(dots[0:i])
|
||||||
|
fmt.Println(a...)
|
||||||
|
}
|
||||||
|
|
||||||
|
func trace(p *Parser, msg string) *Parser {
|
||||||
|
p.printTrace(msg, "(")
|
||||||
|
p.indent++
|
||||||
|
return p
|
||||||
|
}
|
||||||
|
|
||||||
|
// Usage pattern: defer un(trace(p, "..."))
|
||||||
|
func un(p *Parser) {
|
||||||
|
p.indent--
|
||||||
|
p.printTrace(")")
|
||||||
|
}
|
|
@ -0,0 +1,652 @@
|
||||||
|
// Package scanner implements a scanner for HCL (HashiCorp Configuration
|
||||||
|
// Language) source text.
|
||||||
|
package scanner
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"os"
|
||||||
|
"regexp"
|
||||||
|
"unicode"
|
||||||
|
"unicode/utf8"
|
||||||
|
|
||||||
|
"github.com/hashicorp/hcl/hcl/token"
|
||||||
|
)
|
||||||
|
|
||||||
|
// eof represents a marker rune for the end of the reader.
|
||||||
|
const eof = rune(0)
|
||||||
|
|
||||||
|
// Scanner defines a lexical scanner
|
||||||
|
type Scanner struct {
|
||||||
|
buf *bytes.Buffer // Source buffer for advancing and scanning
|
||||||
|
src []byte // Source buffer for immutable access
|
||||||
|
|
||||||
|
// Source Position
|
||||||
|
srcPos token.Pos // current position
|
||||||
|
prevPos token.Pos // previous position, used for peek() method
|
||||||
|
|
||||||
|
lastCharLen int // length of last character in bytes
|
||||||
|
lastLineLen int // length of last line in characters (for correct column reporting)
|
||||||
|
|
||||||
|
tokStart int // token text start position
|
||||||
|
tokEnd int // token text end position
|
||||||
|
|
||||||
|
// Error is called for each error encountered. If no Error
|
||||||
|
// function is set, the error is reported to os.Stderr.
|
||||||
|
Error func(pos token.Pos, msg string)
|
||||||
|
|
||||||
|
// ErrorCount is incremented by one for each error encountered.
|
||||||
|
ErrorCount int
|
||||||
|
|
||||||
|
// tokPos is the start position of most recently scanned token; set by
|
||||||
|
// Scan. The Filename field is always left untouched by the Scanner. If
|
||||||
|
// an error is reported (via Error) and Position is invalid, the scanner is
|
||||||
|
// not inside a token.
|
||||||
|
tokPos token.Pos
|
||||||
|
}
|
||||||
|
|
||||||
|
// New creates and initializes a new instance of Scanner using src as
|
||||||
|
// its source content.
|
||||||
|
func New(src []byte) *Scanner {
|
||||||
|
// even though we accept a src, we read from a io.Reader compatible type
|
||||||
|
// (*bytes.Buffer). So in the future we might easily change it to streaming
|
||||||
|
// read.
|
||||||
|
b := bytes.NewBuffer(src)
|
||||||
|
s := &Scanner{
|
||||||
|
buf: b,
|
||||||
|
src: src,
|
||||||
|
}
|
||||||
|
|
||||||
|
// srcPosition always starts with 1
|
||||||
|
s.srcPos.Line = 1
|
||||||
|
return s
|
||||||
|
}
|
||||||
|
|
||||||
|
// next reads the next rune from the bufferred reader. Returns the rune(0) if
|
||||||
|
// an error occurs (or io.EOF is returned).
|
||||||
|
func (s *Scanner) next() rune {
|
||||||
|
ch, size, err := s.buf.ReadRune()
|
||||||
|
if err != nil {
|
||||||
|
// advance for error reporting
|
||||||
|
s.srcPos.Column++
|
||||||
|
s.srcPos.Offset += size
|
||||||
|
s.lastCharLen = size
|
||||||
|
return eof
|
||||||
|
}
|
||||||
|
|
||||||
|
// remember last position
|
||||||
|
s.prevPos = s.srcPos
|
||||||
|
|
||||||
|
s.srcPos.Column++
|
||||||
|
s.lastCharLen = size
|
||||||
|
s.srcPos.Offset += size
|
||||||
|
|
||||||
|
if ch == utf8.RuneError && size == 1 {
|
||||||
|
s.err("illegal UTF-8 encoding")
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == '\n' {
|
||||||
|
s.srcPos.Line++
|
||||||
|
s.lastLineLen = s.srcPos.Column
|
||||||
|
s.srcPos.Column = 0
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == '\x00' {
|
||||||
|
s.err("unexpected null character (0x00)")
|
||||||
|
return eof
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == '\uE123' {
|
||||||
|
s.err("unicode code point U+E123 reserved for internal use")
|
||||||
|
return utf8.RuneError
|
||||||
|
}
|
||||||
|
|
||||||
|
// debug
|
||||||
|
// fmt.Printf("ch: %q, offset:column: %d:%d\n", ch, s.srcPos.Offset, s.srcPos.Column)
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
// unread unreads the previous read Rune and updates the source position
|
||||||
|
func (s *Scanner) unread() {
|
||||||
|
if err := s.buf.UnreadRune(); err != nil {
|
||||||
|
panic(err) // this is user fault, we should catch it
|
||||||
|
}
|
||||||
|
s.srcPos = s.prevPos // put back last position
|
||||||
|
}
|
||||||
|
|
||||||
|
// peek returns the next rune without advancing the reader.
|
||||||
|
func (s *Scanner) peek() rune {
|
||||||
|
peek, _, err := s.buf.ReadRune()
|
||||||
|
if err != nil {
|
||||||
|
return eof
|
||||||
|
}
|
||||||
|
|
||||||
|
s.buf.UnreadRune()
|
||||||
|
return peek
|
||||||
|
}
|
||||||
|
|
||||||
|
// Scan scans the next token and returns the token.
|
||||||
|
func (s *Scanner) Scan() token.Token {
|
||||||
|
ch := s.next()
|
||||||
|
|
||||||
|
// skip white space
|
||||||
|
for isWhitespace(ch) {
|
||||||
|
ch = s.next()
|
||||||
|
}
|
||||||
|
|
||||||
|
var tok token.Type
|
||||||
|
|
||||||
|
// token text markings
|
||||||
|
s.tokStart = s.srcPos.Offset - s.lastCharLen
|
||||||
|
|
||||||
|
// token position, initial next() is moving the offset by one(size of rune
|
||||||
|
// actually), though we are interested with the starting point
|
||||||
|
s.tokPos.Offset = s.srcPos.Offset - s.lastCharLen
|
||||||
|
if s.srcPos.Column > 0 {
|
||||||
|
// common case: last character was not a '\n'
|
||||||
|
s.tokPos.Line = s.srcPos.Line
|
||||||
|
s.tokPos.Column = s.srcPos.Column
|
||||||
|
} else {
|
||||||
|
// last character was a '\n'
|
||||||
|
// (we cannot be at the beginning of the source
|
||||||
|
// since we have called next() at least once)
|
||||||
|
s.tokPos.Line = s.srcPos.Line - 1
|
||||||
|
s.tokPos.Column = s.lastLineLen
|
||||||
|
}
|
||||||
|
|
||||||
|
switch {
|
||||||
|
case isLetter(ch):
|
||||||
|
tok = token.IDENT
|
||||||
|
lit := s.scanIdentifier()
|
||||||
|
if lit == "true" || lit == "false" {
|
||||||
|
tok = token.BOOL
|
||||||
|
}
|
||||||
|
case isDecimal(ch):
|
||||||
|
tok = s.scanNumber(ch)
|
||||||
|
default:
|
||||||
|
switch ch {
|
||||||
|
case eof:
|
||||||
|
tok = token.EOF
|
||||||
|
case '"':
|
||||||
|
tok = token.STRING
|
||||||
|
s.scanString()
|
||||||
|
case '#', '/':
|
||||||
|
tok = token.COMMENT
|
||||||
|
s.scanComment(ch)
|
||||||
|
case '.':
|
||||||
|
tok = token.PERIOD
|
||||||
|
ch = s.peek()
|
||||||
|
if isDecimal(ch) {
|
||||||
|
tok = token.FLOAT
|
||||||
|
ch = s.scanMantissa(ch)
|
||||||
|
ch = s.scanExponent(ch)
|
||||||
|
}
|
||||||
|
case '<':
|
||||||
|
tok = token.HEREDOC
|
||||||
|
s.scanHeredoc()
|
||||||
|
case '[':
|
||||||
|
tok = token.LBRACK
|
||||||
|
case ']':
|
||||||
|
tok = token.RBRACK
|
||||||
|
case '{':
|
||||||
|
tok = token.LBRACE
|
||||||
|
case '}':
|
||||||
|
tok = token.RBRACE
|
||||||
|
case ',':
|
||||||
|
tok = token.COMMA
|
||||||
|
case '=':
|
||||||
|
tok = token.ASSIGN
|
||||||
|
case '+':
|
||||||
|
tok = token.ADD
|
||||||
|
case '-':
|
||||||
|
if isDecimal(s.peek()) {
|
||||||
|
ch := s.next()
|
||||||
|
tok = s.scanNumber(ch)
|
||||||
|
} else {
|
||||||
|
tok = token.SUB
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
s.err("illegal char")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// finish token ending
|
||||||
|
s.tokEnd = s.srcPos.Offset
|
||||||
|
|
||||||
|
// create token literal
|
||||||
|
var tokenText string
|
||||||
|
if s.tokStart >= 0 {
|
||||||
|
tokenText = string(s.src[s.tokStart:s.tokEnd])
|
||||||
|
}
|
||||||
|
s.tokStart = s.tokEnd // ensure idempotency of tokenText() call
|
||||||
|
|
||||||
|
return token.Token{
|
||||||
|
Type: tok,
|
||||||
|
Pos: s.tokPos,
|
||||||
|
Text: tokenText,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Scanner) scanComment(ch rune) {
|
||||||
|
// single line comments
|
||||||
|
if ch == '#' || (ch == '/' && s.peek() != '*') {
|
||||||
|
if ch == '/' && s.peek() != '/' {
|
||||||
|
s.err("expected '/' for comment")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
ch = s.next()
|
||||||
|
for ch != '\n' && ch >= 0 && ch != eof {
|
||||||
|
ch = s.next()
|
||||||
|
}
|
||||||
|
if ch != eof && ch >= 0 {
|
||||||
|
s.unread()
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// be sure we get the character after /* This allows us to find comment's
|
||||||
|
// that are not erminated
|
||||||
|
if ch == '/' {
|
||||||
|
s.next()
|
||||||
|
ch = s.next() // read character after "/*"
|
||||||
|
}
|
||||||
|
|
||||||
|
// look for /* - style comments
|
||||||
|
for {
|
||||||
|
if ch < 0 || ch == eof {
|
||||||
|
s.err("comment not terminated")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
ch0 := ch
|
||||||
|
ch = s.next()
|
||||||
|
if ch0 == '*' && ch == '/' {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanNumber scans a HCL number definition starting with the given rune
|
||||||
|
func (s *Scanner) scanNumber(ch rune) token.Type {
|
||||||
|
if ch == '0' {
|
||||||
|
// check for hexadecimal, octal or float
|
||||||
|
ch = s.next()
|
||||||
|
if ch == 'x' || ch == 'X' {
|
||||||
|
// hexadecimal
|
||||||
|
ch = s.next()
|
||||||
|
found := false
|
||||||
|
for isHexadecimal(ch) {
|
||||||
|
ch = s.next()
|
||||||
|
found = true
|
||||||
|
}
|
||||||
|
|
||||||
|
if !found {
|
||||||
|
s.err("illegal hexadecimal number")
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch != eof {
|
||||||
|
s.unread()
|
||||||
|
}
|
||||||
|
|
||||||
|
return token.NUMBER
|
||||||
|
}
|
||||||
|
|
||||||
|
// now it's either something like: 0421(octal) or 0.1231(float)
|
||||||
|
illegalOctal := false
|
||||||
|
for isDecimal(ch) {
|
||||||
|
ch = s.next()
|
||||||
|
if ch == '8' || ch == '9' {
|
||||||
|
// this is just a possibility. For example 0159 is illegal, but
|
||||||
|
// 0159.23 is valid. So we mark a possible illegal octal. If
|
||||||
|
// the next character is not a period, we'll print the error.
|
||||||
|
illegalOctal = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == 'e' || ch == 'E' {
|
||||||
|
ch = s.scanExponent(ch)
|
||||||
|
return token.FLOAT
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == '.' {
|
||||||
|
ch = s.scanFraction(ch)
|
||||||
|
|
||||||
|
if ch == 'e' || ch == 'E' {
|
||||||
|
ch = s.next()
|
||||||
|
ch = s.scanExponent(ch)
|
||||||
|
}
|
||||||
|
return token.FLOAT
|
||||||
|
}
|
||||||
|
|
||||||
|
if illegalOctal {
|
||||||
|
s.err("illegal octal number")
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch != eof {
|
||||||
|
s.unread()
|
||||||
|
}
|
||||||
|
return token.NUMBER
|
||||||
|
}
|
||||||
|
|
||||||
|
s.scanMantissa(ch)
|
||||||
|
ch = s.next() // seek forward
|
||||||
|
if ch == 'e' || ch == 'E' {
|
||||||
|
ch = s.scanExponent(ch)
|
||||||
|
return token.FLOAT
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == '.' {
|
||||||
|
ch = s.scanFraction(ch)
|
||||||
|
if ch == 'e' || ch == 'E' {
|
||||||
|
ch = s.next()
|
||||||
|
ch = s.scanExponent(ch)
|
||||||
|
}
|
||||||
|
return token.FLOAT
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch != eof {
|
||||||
|
s.unread()
|
||||||
|
}
|
||||||
|
return token.NUMBER
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanMantissa scans the mantissa beginning from the rune. It returns the next
|
||||||
|
// non decimal rune. It's used to determine wheter it's a fraction or exponent.
|
||||||
|
func (s *Scanner) scanMantissa(ch rune) rune {
|
||||||
|
scanned := false
|
||||||
|
for isDecimal(ch) {
|
||||||
|
ch = s.next()
|
||||||
|
scanned = true
|
||||||
|
}
|
||||||
|
|
||||||
|
if scanned && ch != eof {
|
||||||
|
s.unread()
|
||||||
|
}
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanFraction scans the fraction after the '.' rune
|
||||||
|
func (s *Scanner) scanFraction(ch rune) rune {
|
||||||
|
if ch == '.' {
|
||||||
|
ch = s.peek() // we peek just to see if we can move forward
|
||||||
|
ch = s.scanMantissa(ch)
|
||||||
|
}
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanExponent scans the remaining parts of an exponent after the 'e' or 'E'
|
||||||
|
// rune.
|
||||||
|
func (s *Scanner) scanExponent(ch rune) rune {
|
||||||
|
if ch == 'e' || ch == 'E' {
|
||||||
|
ch = s.next()
|
||||||
|
if ch == '-' || ch == '+' {
|
||||||
|
ch = s.next()
|
||||||
|
}
|
||||||
|
ch = s.scanMantissa(ch)
|
||||||
|
}
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanHeredoc scans a heredoc string
|
||||||
|
func (s *Scanner) scanHeredoc() {
|
||||||
|
// Scan the second '<' in example: '<<EOF'
|
||||||
|
if s.next() != '<' {
|
||||||
|
s.err("heredoc expected second '<', didn't see it")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Get the original offset so we can read just the heredoc ident
|
||||||
|
offs := s.srcPos.Offset
|
||||||
|
|
||||||
|
// Scan the identifier
|
||||||
|
ch := s.next()
|
||||||
|
|
||||||
|
// Indented heredoc syntax
|
||||||
|
if ch == '-' {
|
||||||
|
ch = s.next()
|
||||||
|
}
|
||||||
|
|
||||||
|
for isLetter(ch) || isDigit(ch) {
|
||||||
|
ch = s.next()
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we reached an EOF then that is not good
|
||||||
|
if ch == eof {
|
||||||
|
s.err("heredoc not terminated")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Ignore the '\r' in Windows line endings
|
||||||
|
if ch == '\r' {
|
||||||
|
if s.peek() == '\n' {
|
||||||
|
ch = s.next()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we didn't reach a newline then that is also not good
|
||||||
|
if ch != '\n' {
|
||||||
|
s.err("invalid characters in heredoc anchor")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Read the identifier
|
||||||
|
identBytes := s.src[offs : s.srcPos.Offset-s.lastCharLen]
|
||||||
|
if len(identBytes) == 0 || (len(identBytes) == 1 && identBytes[0] == '-') {
|
||||||
|
s.err("zero-length heredoc anchor")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
var identRegexp *regexp.Regexp
|
||||||
|
if identBytes[0] == '-' {
|
||||||
|
identRegexp = regexp.MustCompile(fmt.Sprintf(`^[[:space:]]*%s\r*\z`, identBytes[1:]))
|
||||||
|
} else {
|
||||||
|
identRegexp = regexp.MustCompile(fmt.Sprintf(`^[[:space:]]*%s\r*\z`, identBytes))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Read the actual string value
|
||||||
|
lineStart := s.srcPos.Offset
|
||||||
|
for {
|
||||||
|
ch := s.next()
|
||||||
|
|
||||||
|
// Special newline handling.
|
||||||
|
if ch == '\n' {
|
||||||
|
// Math is fast, so we first compare the byte counts to see if we have a chance
|
||||||
|
// of seeing the same identifier - if the length is less than the number of bytes
|
||||||
|
// in the identifier, this cannot be a valid terminator.
|
||||||
|
lineBytesLen := s.srcPos.Offset - s.lastCharLen - lineStart
|
||||||
|
if lineBytesLen >= len(identBytes) && identRegexp.Match(s.src[lineStart:s.srcPos.Offset-s.lastCharLen]) {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
// Not an anchor match, record the start of a new line
|
||||||
|
lineStart = s.srcPos.Offset
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == eof {
|
||||||
|
s.err("heredoc not terminated")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanString scans a quoted string
|
||||||
|
func (s *Scanner) scanString() {
|
||||||
|
braces := 0
|
||||||
|
for {
|
||||||
|
// '"' opening already consumed
|
||||||
|
// read character after quote
|
||||||
|
ch := s.next()
|
||||||
|
|
||||||
|
if (ch == '\n' && braces == 0) || ch < 0 || ch == eof {
|
||||||
|
s.err("literal not terminated")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == '"' && braces == 0 {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we're going into a ${} then we can ignore quotes for awhile
|
||||||
|
if braces == 0 && ch == '$' && s.peek() == '{' {
|
||||||
|
braces++
|
||||||
|
s.next()
|
||||||
|
} else if braces > 0 && ch == '{' {
|
||||||
|
braces++
|
||||||
|
}
|
||||||
|
if braces > 0 && ch == '}' {
|
||||||
|
braces--
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == '\\' {
|
||||||
|
s.scanEscape()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanEscape scans an escape sequence
|
||||||
|
func (s *Scanner) scanEscape() rune {
|
||||||
|
// http://en.cppreference.com/w/cpp/language/escape
|
||||||
|
ch := s.next() // read character after '/'
|
||||||
|
switch ch {
|
||||||
|
case 'a', 'b', 'f', 'n', 'r', 't', 'v', '\\', '"':
|
||||||
|
// nothing to do
|
||||||
|
case '0', '1', '2', '3', '4', '5', '6', '7':
|
||||||
|
// octal notation
|
||||||
|
ch = s.scanDigits(ch, 8, 3)
|
||||||
|
case 'x':
|
||||||
|
// hexademical notation
|
||||||
|
ch = s.scanDigits(s.next(), 16, 2)
|
||||||
|
case 'u':
|
||||||
|
// universal character name
|
||||||
|
ch = s.scanDigits(s.next(), 16, 4)
|
||||||
|
case 'U':
|
||||||
|
// universal character name
|
||||||
|
ch = s.scanDigits(s.next(), 16, 8)
|
||||||
|
default:
|
||||||
|
s.err("illegal char escape")
|
||||||
|
}
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanDigits scans a rune with the given base for n times. For example an
|
||||||
|
// octal notation \184 would yield in scanDigits(ch, 8, 3)
|
||||||
|
func (s *Scanner) scanDigits(ch rune, base, n int) rune {
|
||||||
|
start := n
|
||||||
|
for n > 0 && digitVal(ch) < base {
|
||||||
|
ch = s.next()
|
||||||
|
if ch == eof {
|
||||||
|
// If we see an EOF, we halt any more scanning of digits
|
||||||
|
// immediately.
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
n--
|
||||||
|
}
|
||||||
|
if n > 0 {
|
||||||
|
s.err("illegal char escape")
|
||||||
|
}
|
||||||
|
|
||||||
|
if n != start && ch != eof {
|
||||||
|
// we scanned all digits, put the last non digit char back,
|
||||||
|
// only if we read anything at all
|
||||||
|
s.unread()
|
||||||
|
}
|
||||||
|
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanIdentifier scans an identifier and returns the literal string
|
||||||
|
func (s *Scanner) scanIdentifier() string {
|
||||||
|
offs := s.srcPos.Offset - s.lastCharLen
|
||||||
|
ch := s.next()
|
||||||
|
for isLetter(ch) || isDigit(ch) || ch == '-' || ch == '.' {
|
||||||
|
ch = s.next()
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch != eof {
|
||||||
|
s.unread() // we got identifier, put back latest char
|
||||||
|
}
|
||||||
|
|
||||||
|
return string(s.src[offs:s.srcPos.Offset])
|
||||||
|
}
|
||||||
|
|
||||||
|
// recentPosition returns the position of the character immediately after the
|
||||||
|
// character or token returned by the last call to Scan.
|
||||||
|
func (s *Scanner) recentPosition() (pos token.Pos) {
|
||||||
|
pos.Offset = s.srcPos.Offset - s.lastCharLen
|
||||||
|
switch {
|
||||||
|
case s.srcPos.Column > 0:
|
||||||
|
// common case: last character was not a '\n'
|
||||||
|
pos.Line = s.srcPos.Line
|
||||||
|
pos.Column = s.srcPos.Column
|
||||||
|
case s.lastLineLen > 0:
|
||||||
|
// last character was a '\n'
|
||||||
|
// (we cannot be at the beginning of the source
|
||||||
|
// since we have called next() at least once)
|
||||||
|
pos.Line = s.srcPos.Line - 1
|
||||||
|
pos.Column = s.lastLineLen
|
||||||
|
default:
|
||||||
|
// at the beginning of the source
|
||||||
|
pos.Line = 1
|
||||||
|
pos.Column = 1
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// err prints the error of any scanning to s.Error function. If the function is
|
||||||
|
// not defined, by default it prints them to os.Stderr
|
||||||
|
func (s *Scanner) err(msg string) {
|
||||||
|
s.ErrorCount++
|
||||||
|
pos := s.recentPosition()
|
||||||
|
|
||||||
|
if s.Error != nil {
|
||||||
|
s.Error(pos, msg)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
fmt.Fprintf(os.Stderr, "%s: %s\n", pos, msg)
|
||||||
|
}
|
||||||
|
|
||||||
|
// isHexadecimal returns true if the given rune is a letter
|
||||||
|
func isLetter(ch rune) bool {
|
||||||
|
return 'a' <= ch && ch <= 'z' || 'A' <= ch && ch <= 'Z' || ch == '_' || ch >= 0x80 && unicode.IsLetter(ch)
|
||||||
|
}
|
||||||
|
|
||||||
|
// isDigit returns true if the given rune is a decimal digit
|
||||||
|
func isDigit(ch rune) bool {
|
||||||
|
return '0' <= ch && ch <= '9' || ch >= 0x80 && unicode.IsDigit(ch)
|
||||||
|
}
|
||||||
|
|
||||||
|
// isDecimal returns true if the given rune is a decimal number
|
||||||
|
func isDecimal(ch rune) bool {
|
||||||
|
return '0' <= ch && ch <= '9'
|
||||||
|
}
|
||||||
|
|
||||||
|
// isHexadecimal returns true if the given rune is an hexadecimal number
|
||||||
|
func isHexadecimal(ch rune) bool {
|
||||||
|
return '0' <= ch && ch <= '9' || 'a' <= ch && ch <= 'f' || 'A' <= ch && ch <= 'F'
|
||||||
|
}
|
||||||
|
|
||||||
|
// isWhitespace returns true if the rune is a space, tab, newline or carriage return
|
||||||
|
func isWhitespace(ch rune) bool {
|
||||||
|
return ch == ' ' || ch == '\t' || ch == '\n' || ch == '\r'
|
||||||
|
}
|
||||||
|
|
||||||
|
// digitVal returns the integer value of a given octal,decimal or hexadecimal rune
|
||||||
|
func digitVal(ch rune) int {
|
||||||
|
switch {
|
||||||
|
case '0' <= ch && ch <= '9':
|
||||||
|
return int(ch - '0')
|
||||||
|
case 'a' <= ch && ch <= 'f':
|
||||||
|
return int(ch - 'a' + 10)
|
||||||
|
case 'A' <= ch && ch <= 'F':
|
||||||
|
return int(ch - 'A' + 10)
|
||||||
|
}
|
||||||
|
return 16 // larger than any legal digit val
|
||||||
|
}
|
|
@ -0,0 +1,241 @@
|
||||||
|
package strconv
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"unicode/utf8"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ErrSyntax indicates that a value does not have the right syntax for the target type.
|
||||||
|
var ErrSyntax = errors.New("invalid syntax")
|
||||||
|
|
||||||
|
// Unquote interprets s as a single-quoted, double-quoted,
|
||||||
|
// or backquoted Go string literal, returning the string value
|
||||||
|
// that s quotes. (If s is single-quoted, it would be a Go
|
||||||
|
// character literal; Unquote returns the corresponding
|
||||||
|
// one-character string.)
|
||||||
|
func Unquote(s string) (t string, err error) {
|
||||||
|
n := len(s)
|
||||||
|
if n < 2 {
|
||||||
|
return "", ErrSyntax
|
||||||
|
}
|
||||||
|
quote := s[0]
|
||||||
|
if quote != s[n-1] {
|
||||||
|
return "", ErrSyntax
|
||||||
|
}
|
||||||
|
s = s[1 : n-1]
|
||||||
|
|
||||||
|
if quote != '"' {
|
||||||
|
return "", ErrSyntax
|
||||||
|
}
|
||||||
|
if !contains(s, '$') && !contains(s, '{') && contains(s, '\n') {
|
||||||
|
return "", ErrSyntax
|
||||||
|
}
|
||||||
|
|
||||||
|
// Is it trivial? Avoid allocation.
|
||||||
|
if !contains(s, '\\') && !contains(s, quote) && !contains(s, '$') {
|
||||||
|
switch quote {
|
||||||
|
case '"':
|
||||||
|
return s, nil
|
||||||
|
case '\'':
|
||||||
|
r, size := utf8.DecodeRuneInString(s)
|
||||||
|
if size == len(s) && (r != utf8.RuneError || size != 1) {
|
||||||
|
return s, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
var runeTmp [utf8.UTFMax]byte
|
||||||
|
buf := make([]byte, 0, 3*len(s)/2) // Try to avoid more allocations.
|
||||||
|
for len(s) > 0 {
|
||||||
|
// If we're starting a '${}' then let it through un-unquoted.
|
||||||
|
// Specifically: we don't unquote any characters within the `${}`
|
||||||
|
// section.
|
||||||
|
if s[0] == '$' && len(s) > 1 && s[1] == '{' {
|
||||||
|
buf = append(buf, '$', '{')
|
||||||
|
s = s[2:]
|
||||||
|
|
||||||
|
// Continue reading until we find the closing brace, copying as-is
|
||||||
|
braces := 1
|
||||||
|
for len(s) > 0 && braces > 0 {
|
||||||
|
r, size := utf8.DecodeRuneInString(s)
|
||||||
|
if r == utf8.RuneError {
|
||||||
|
return "", ErrSyntax
|
||||||
|
}
|
||||||
|
|
||||||
|
s = s[size:]
|
||||||
|
|
||||||
|
n := utf8.EncodeRune(runeTmp[:], r)
|
||||||
|
buf = append(buf, runeTmp[:n]...)
|
||||||
|
|
||||||
|
switch r {
|
||||||
|
case '{':
|
||||||
|
braces++
|
||||||
|
case '}':
|
||||||
|
braces--
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if braces != 0 {
|
||||||
|
return "", ErrSyntax
|
||||||
|
}
|
||||||
|
if len(s) == 0 {
|
||||||
|
// If there's no string left, we're done!
|
||||||
|
break
|
||||||
|
} else {
|
||||||
|
// If there's more left, we need to pop back up to the top of the loop
|
||||||
|
// in case there's another interpolation in this string.
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if s[0] == '\n' {
|
||||||
|
return "", ErrSyntax
|
||||||
|
}
|
||||||
|
|
||||||
|
c, multibyte, ss, err := unquoteChar(s, quote)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
s = ss
|
||||||
|
if c < utf8.RuneSelf || !multibyte {
|
||||||
|
buf = append(buf, byte(c))
|
||||||
|
} else {
|
||||||
|
n := utf8.EncodeRune(runeTmp[:], c)
|
||||||
|
buf = append(buf, runeTmp[:n]...)
|
||||||
|
}
|
||||||
|
if quote == '\'' && len(s) != 0 {
|
||||||
|
// single-quoted must be single character
|
||||||
|
return "", ErrSyntax
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return string(buf), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// contains reports whether the string contains the byte c.
|
||||||
|
func contains(s string, c byte) bool {
|
||||||
|
for i := 0; i < len(s); i++ {
|
||||||
|
if s[i] == c {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
func unhex(b byte) (v rune, ok bool) {
|
||||||
|
c := rune(b)
|
||||||
|
switch {
|
||||||
|
case '0' <= c && c <= '9':
|
||||||
|
return c - '0', true
|
||||||
|
case 'a' <= c && c <= 'f':
|
||||||
|
return c - 'a' + 10, true
|
||||||
|
case 'A' <= c && c <= 'F':
|
||||||
|
return c - 'A' + 10, true
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func unquoteChar(s string, quote byte) (value rune, multibyte bool, tail string, err error) {
|
||||||
|
// easy cases
|
||||||
|
switch c := s[0]; {
|
||||||
|
case c == quote && (quote == '\'' || quote == '"'):
|
||||||
|
err = ErrSyntax
|
||||||
|
return
|
||||||
|
case c >= utf8.RuneSelf:
|
||||||
|
r, size := utf8.DecodeRuneInString(s)
|
||||||
|
return r, true, s[size:], nil
|
||||||
|
case c != '\\':
|
||||||
|
return rune(s[0]), false, s[1:], nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// hard case: c is backslash
|
||||||
|
if len(s) <= 1 {
|
||||||
|
err = ErrSyntax
|
||||||
|
return
|
||||||
|
}
|
||||||
|
c := s[1]
|
||||||
|
s = s[2:]
|
||||||
|
|
||||||
|
switch c {
|
||||||
|
case 'a':
|
||||||
|
value = '\a'
|
||||||
|
case 'b':
|
||||||
|
value = '\b'
|
||||||
|
case 'f':
|
||||||
|
value = '\f'
|
||||||
|
case 'n':
|
||||||
|
value = '\n'
|
||||||
|
case 'r':
|
||||||
|
value = '\r'
|
||||||
|
case 't':
|
||||||
|
value = '\t'
|
||||||
|
case 'v':
|
||||||
|
value = '\v'
|
||||||
|
case 'x', 'u', 'U':
|
||||||
|
n := 0
|
||||||
|
switch c {
|
||||||
|
case 'x':
|
||||||
|
n = 2
|
||||||
|
case 'u':
|
||||||
|
n = 4
|
||||||
|
case 'U':
|
||||||
|
n = 8
|
||||||
|
}
|
||||||
|
var v rune
|
||||||
|
if len(s) < n {
|
||||||
|
err = ErrSyntax
|
||||||
|
return
|
||||||
|
}
|
||||||
|
for j := 0; j < n; j++ {
|
||||||
|
x, ok := unhex(s[j])
|
||||||
|
if !ok {
|
||||||
|
err = ErrSyntax
|
||||||
|
return
|
||||||
|
}
|
||||||
|
v = v<<4 | x
|
||||||
|
}
|
||||||
|
s = s[n:]
|
||||||
|
if c == 'x' {
|
||||||
|
// single-byte string, possibly not UTF-8
|
||||||
|
value = v
|
||||||
|
break
|
||||||
|
}
|
||||||
|
if v > utf8.MaxRune {
|
||||||
|
err = ErrSyntax
|
||||||
|
return
|
||||||
|
}
|
||||||
|
value = v
|
||||||
|
multibyte = true
|
||||||
|
case '0', '1', '2', '3', '4', '5', '6', '7':
|
||||||
|
v := rune(c) - '0'
|
||||||
|
if len(s) < 2 {
|
||||||
|
err = ErrSyntax
|
||||||
|
return
|
||||||
|
}
|
||||||
|
for j := 0; j < 2; j++ { // one digit already; two more
|
||||||
|
x := rune(s[j]) - '0'
|
||||||
|
if x < 0 || x > 7 {
|
||||||
|
err = ErrSyntax
|
||||||
|
return
|
||||||
|
}
|
||||||
|
v = (v << 3) | x
|
||||||
|
}
|
||||||
|
s = s[2:]
|
||||||
|
if v > 255 {
|
||||||
|
err = ErrSyntax
|
||||||
|
return
|
||||||
|
}
|
||||||
|
value = v
|
||||||
|
case '\\':
|
||||||
|
value = '\\'
|
||||||
|
case '\'', '"':
|
||||||
|
if c != quote {
|
||||||
|
err = ErrSyntax
|
||||||
|
return
|
||||||
|
}
|
||||||
|
value = rune(c)
|
||||||
|
default:
|
||||||
|
err = ErrSyntax
|
||||||
|
return
|
||||||
|
}
|
||||||
|
tail = s
|
||||||
|
return
|
||||||
|
}
|
|
@ -0,0 +1,46 @@
|
||||||
|
package token
|
||||||
|
|
||||||
|
import "fmt"
|
||||||
|
|
||||||
|
// Pos describes an arbitrary source position
|
||||||
|
// including the file, line, and column location.
|
||||||
|
// A Position is valid if the line number is > 0.
|
||||||
|
type Pos struct {
|
||||||
|
Filename string // filename, if any
|
||||||
|
Offset int // offset, starting at 0
|
||||||
|
Line int // line number, starting at 1
|
||||||
|
Column int // column number, starting at 1 (character count)
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsValid returns true if the position is valid.
|
||||||
|
func (p *Pos) IsValid() bool { return p.Line > 0 }
|
||||||
|
|
||||||
|
// String returns a string in one of several forms:
|
||||||
|
//
|
||||||
|
// file:line:column valid position with file name
|
||||||
|
// line:column valid position without file name
|
||||||
|
// file invalid position with file name
|
||||||
|
// - invalid position without file name
|
||||||
|
func (p Pos) String() string {
|
||||||
|
s := p.Filename
|
||||||
|
if p.IsValid() {
|
||||||
|
if s != "" {
|
||||||
|
s += ":"
|
||||||
|
}
|
||||||
|
s += fmt.Sprintf("%d:%d", p.Line, p.Column)
|
||||||
|
}
|
||||||
|
if s == "" {
|
||||||
|
s = "-"
|
||||||
|
}
|
||||||
|
return s
|
||||||
|
}
|
||||||
|
|
||||||
|
// Before reports whether the position p is before u.
|
||||||
|
func (p Pos) Before(u Pos) bool {
|
||||||
|
return u.Offset > p.Offset || u.Line > p.Line
|
||||||
|
}
|
||||||
|
|
||||||
|
// After reports whether the position p is after u.
|
||||||
|
func (p Pos) After(u Pos) bool {
|
||||||
|
return u.Offset < p.Offset || u.Line < p.Line
|
||||||
|
}
|
|
@ -0,0 +1,219 @@
|
||||||
|
// Package token defines constants representing the lexical tokens for HCL
|
||||||
|
// (HashiCorp Configuration Language)
|
||||||
|
package token
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
hclstrconv "github.com/hashicorp/hcl/hcl/strconv"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Token defines a single HCL token which can be obtained via the Scanner
|
||||||
|
type Token struct {
|
||||||
|
Type Type
|
||||||
|
Pos Pos
|
||||||
|
Text string
|
||||||
|
JSON bool
|
||||||
|
}
|
||||||
|
|
||||||
|
// Type is the set of lexical tokens of the HCL (HashiCorp Configuration Language)
|
||||||
|
type Type int
|
||||||
|
|
||||||
|
const (
|
||||||
|
// Special tokens
|
||||||
|
ILLEGAL Type = iota
|
||||||
|
EOF
|
||||||
|
COMMENT
|
||||||
|
|
||||||
|
identifier_beg
|
||||||
|
IDENT // literals
|
||||||
|
literal_beg
|
||||||
|
NUMBER // 12345
|
||||||
|
FLOAT // 123.45
|
||||||
|
BOOL // true,false
|
||||||
|
STRING // "abc"
|
||||||
|
HEREDOC // <<FOO\nbar\nFOO
|
||||||
|
literal_end
|
||||||
|
identifier_end
|
||||||
|
|
||||||
|
operator_beg
|
||||||
|
LBRACK // [
|
||||||
|
LBRACE // {
|
||||||
|
COMMA // ,
|
||||||
|
PERIOD // .
|
||||||
|
|
||||||
|
RBRACK // ]
|
||||||
|
RBRACE // }
|
||||||
|
|
||||||
|
ASSIGN // =
|
||||||
|
ADD // +
|
||||||
|
SUB // -
|
||||||
|
operator_end
|
||||||
|
)
|
||||||
|
|
||||||
|
var tokens = [...]string{
|
||||||
|
ILLEGAL: "ILLEGAL",
|
||||||
|
|
||||||
|
EOF: "EOF",
|
||||||
|
COMMENT: "COMMENT",
|
||||||
|
|
||||||
|
IDENT: "IDENT",
|
||||||
|
NUMBER: "NUMBER",
|
||||||
|
FLOAT: "FLOAT",
|
||||||
|
BOOL: "BOOL",
|
||||||
|
STRING: "STRING",
|
||||||
|
|
||||||
|
LBRACK: "LBRACK",
|
||||||
|
LBRACE: "LBRACE",
|
||||||
|
COMMA: "COMMA",
|
||||||
|
PERIOD: "PERIOD",
|
||||||
|
HEREDOC: "HEREDOC",
|
||||||
|
|
||||||
|
RBRACK: "RBRACK",
|
||||||
|
RBRACE: "RBRACE",
|
||||||
|
|
||||||
|
ASSIGN: "ASSIGN",
|
||||||
|
ADD: "ADD",
|
||||||
|
SUB: "SUB",
|
||||||
|
}
|
||||||
|
|
||||||
|
// String returns the string corresponding to the token tok.
|
||||||
|
func (t Type) String() string {
|
||||||
|
s := ""
|
||||||
|
if 0 <= t && t < Type(len(tokens)) {
|
||||||
|
s = tokens[t]
|
||||||
|
}
|
||||||
|
if s == "" {
|
||||||
|
s = "token(" + strconv.Itoa(int(t)) + ")"
|
||||||
|
}
|
||||||
|
return s
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsIdentifier returns true for tokens corresponding to identifiers and basic
|
||||||
|
// type literals; it returns false otherwise.
|
||||||
|
func (t Type) IsIdentifier() bool { return identifier_beg < t && t < identifier_end }
|
||||||
|
|
||||||
|
// IsLiteral returns true for tokens corresponding to basic type literals; it
|
||||||
|
// returns false otherwise.
|
||||||
|
func (t Type) IsLiteral() bool { return literal_beg < t && t < literal_end }
|
||||||
|
|
||||||
|
// IsOperator returns true for tokens corresponding to operators and
|
||||||
|
// delimiters; it returns false otherwise.
|
||||||
|
func (t Type) IsOperator() bool { return operator_beg < t && t < operator_end }
|
||||||
|
|
||||||
|
// String returns the token's literal text. Note that this is only
|
||||||
|
// applicable for certain token types, such as token.IDENT,
|
||||||
|
// token.STRING, etc..
|
||||||
|
func (t Token) String() string {
|
||||||
|
return fmt.Sprintf("%s %s %s", t.Pos.String(), t.Type.String(), t.Text)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Value returns the properly typed value for this token. The type of
|
||||||
|
// the returned interface{} is guaranteed based on the Type field.
|
||||||
|
//
|
||||||
|
// This can only be called for literal types. If it is called for any other
|
||||||
|
// type, this will panic.
|
||||||
|
func (t Token) Value() interface{} {
|
||||||
|
switch t.Type {
|
||||||
|
case BOOL:
|
||||||
|
if t.Text == "true" {
|
||||||
|
return true
|
||||||
|
} else if t.Text == "false" {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
panic("unknown bool value: " + t.Text)
|
||||||
|
case FLOAT:
|
||||||
|
v, err := strconv.ParseFloat(t.Text, 64)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return float64(v)
|
||||||
|
case NUMBER:
|
||||||
|
v, err := strconv.ParseInt(t.Text, 0, 64)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return int64(v)
|
||||||
|
case IDENT:
|
||||||
|
return t.Text
|
||||||
|
case HEREDOC:
|
||||||
|
return unindentHeredoc(t.Text)
|
||||||
|
case STRING:
|
||||||
|
// Determine the Unquote method to use. If it came from JSON,
|
||||||
|
// then we need to use the built-in unquote since we have to
|
||||||
|
// escape interpolations there.
|
||||||
|
f := hclstrconv.Unquote
|
||||||
|
if t.JSON {
|
||||||
|
f = strconv.Unquote
|
||||||
|
}
|
||||||
|
|
||||||
|
// This case occurs if json null is used
|
||||||
|
if t.Text == "" {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
v, err := f(t.Text)
|
||||||
|
if err != nil {
|
||||||
|
panic(fmt.Sprintf("unquote %s err: %s", t.Text, err))
|
||||||
|
}
|
||||||
|
|
||||||
|
return v
|
||||||
|
default:
|
||||||
|
panic(fmt.Sprintf("unimplemented Value for type: %s", t.Type))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// unindentHeredoc returns the string content of a HEREDOC if it is started with <<
|
||||||
|
// and the content of a HEREDOC with the hanging indent removed if it is started with
|
||||||
|
// a <<-, and the terminating line is at least as indented as the least indented line.
|
||||||
|
func unindentHeredoc(heredoc string) string {
|
||||||
|
// We need to find the end of the marker
|
||||||
|
idx := strings.IndexByte(heredoc, '\n')
|
||||||
|
if idx == -1 {
|
||||||
|
panic("heredoc doesn't contain newline")
|
||||||
|
}
|
||||||
|
|
||||||
|
unindent := heredoc[2] == '-'
|
||||||
|
|
||||||
|
// We can optimize if the heredoc isn't marked for indentation
|
||||||
|
if !unindent {
|
||||||
|
return string(heredoc[idx+1 : len(heredoc)-idx+1])
|
||||||
|
}
|
||||||
|
|
||||||
|
// We need to unindent each line based on the indentation level of the marker
|
||||||
|
lines := strings.Split(string(heredoc[idx+1:len(heredoc)-idx+2]), "\n")
|
||||||
|
whitespacePrefix := lines[len(lines)-1]
|
||||||
|
|
||||||
|
isIndented := true
|
||||||
|
for _, v := range lines {
|
||||||
|
if strings.HasPrefix(v, whitespacePrefix) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
isIndented = false
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
// If all lines are not at least as indented as the terminating mark, return the
|
||||||
|
// heredoc as is, but trim the leading space from the marker on the final line.
|
||||||
|
if !isIndented {
|
||||||
|
return strings.TrimRight(string(heredoc[idx+1:len(heredoc)-idx+1]), " \t")
|
||||||
|
}
|
||||||
|
|
||||||
|
unindentedLines := make([]string, len(lines))
|
||||||
|
for k, v := range lines {
|
||||||
|
if k == len(lines)-1 {
|
||||||
|
unindentedLines[k] = ""
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
unindentedLines[k] = strings.TrimPrefix(v, whitespacePrefix)
|
||||||
|
}
|
||||||
|
|
||||||
|
return strings.Join(unindentedLines, "\n")
|
||||||
|
}
|
|
@ -0,0 +1,117 @@
|
||||||
|
package parser
|
||||||
|
|
||||||
|
import "github.com/hashicorp/hcl/hcl/ast"
|
||||||
|
|
||||||
|
// flattenObjects takes an AST node, walks it, and flattens
|
||||||
|
func flattenObjects(node ast.Node) {
|
||||||
|
ast.Walk(node, func(n ast.Node) (ast.Node, bool) {
|
||||||
|
// We only care about lists, because this is what we modify
|
||||||
|
list, ok := n.(*ast.ObjectList)
|
||||||
|
if !ok {
|
||||||
|
return n, true
|
||||||
|
}
|
||||||
|
|
||||||
|
// Rebuild the item list
|
||||||
|
items := make([]*ast.ObjectItem, 0, len(list.Items))
|
||||||
|
frontier := make([]*ast.ObjectItem, len(list.Items))
|
||||||
|
copy(frontier, list.Items)
|
||||||
|
for len(frontier) > 0 {
|
||||||
|
// Pop the current item
|
||||||
|
n := len(frontier)
|
||||||
|
item := frontier[n-1]
|
||||||
|
frontier = frontier[:n-1]
|
||||||
|
|
||||||
|
switch v := item.Val.(type) {
|
||||||
|
case *ast.ObjectType:
|
||||||
|
items, frontier = flattenObjectType(v, item, items, frontier)
|
||||||
|
case *ast.ListType:
|
||||||
|
items, frontier = flattenListType(v, item, items, frontier)
|
||||||
|
default:
|
||||||
|
items = append(items, item)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Reverse the list since the frontier model runs things backwards
|
||||||
|
for i := len(items)/2 - 1; i >= 0; i-- {
|
||||||
|
opp := len(items) - 1 - i
|
||||||
|
items[i], items[opp] = items[opp], items[i]
|
||||||
|
}
|
||||||
|
|
||||||
|
// Done! Set the original items
|
||||||
|
list.Items = items
|
||||||
|
return n, true
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func flattenListType(
|
||||||
|
ot *ast.ListType,
|
||||||
|
item *ast.ObjectItem,
|
||||||
|
items []*ast.ObjectItem,
|
||||||
|
frontier []*ast.ObjectItem) ([]*ast.ObjectItem, []*ast.ObjectItem) {
|
||||||
|
// If the list is empty, keep the original list
|
||||||
|
if len(ot.List) == 0 {
|
||||||
|
items = append(items, item)
|
||||||
|
return items, frontier
|
||||||
|
}
|
||||||
|
|
||||||
|
// All the elements of this object must also be objects!
|
||||||
|
for _, subitem := range ot.List {
|
||||||
|
if _, ok := subitem.(*ast.ObjectType); !ok {
|
||||||
|
items = append(items, item)
|
||||||
|
return items, frontier
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Great! We have a match go through all the items and flatten
|
||||||
|
for _, elem := range ot.List {
|
||||||
|
// Add it to the frontier so that we can recurse
|
||||||
|
frontier = append(frontier, &ast.ObjectItem{
|
||||||
|
Keys: item.Keys,
|
||||||
|
Assign: item.Assign,
|
||||||
|
Val: elem,
|
||||||
|
LeadComment: item.LeadComment,
|
||||||
|
LineComment: item.LineComment,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
return items, frontier
|
||||||
|
}
|
||||||
|
|
||||||
|
func flattenObjectType(
|
||||||
|
ot *ast.ObjectType,
|
||||||
|
item *ast.ObjectItem,
|
||||||
|
items []*ast.ObjectItem,
|
||||||
|
frontier []*ast.ObjectItem) ([]*ast.ObjectItem, []*ast.ObjectItem) {
|
||||||
|
// If the list has no items we do not have to flatten anything
|
||||||
|
if ot.List.Items == nil {
|
||||||
|
items = append(items, item)
|
||||||
|
return items, frontier
|
||||||
|
}
|
||||||
|
|
||||||
|
// All the elements of this object must also be objects!
|
||||||
|
for _, subitem := range ot.List.Items {
|
||||||
|
if _, ok := subitem.Val.(*ast.ObjectType); !ok {
|
||||||
|
items = append(items, item)
|
||||||
|
return items, frontier
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Great! We have a match go through all the items and flatten
|
||||||
|
for _, subitem := range ot.List.Items {
|
||||||
|
// Copy the new key
|
||||||
|
keys := make([]*ast.ObjectKey, len(item.Keys)+len(subitem.Keys))
|
||||||
|
copy(keys, item.Keys)
|
||||||
|
copy(keys[len(item.Keys):], subitem.Keys)
|
||||||
|
|
||||||
|
// Add it to the frontier so that we can recurse
|
||||||
|
frontier = append(frontier, &ast.ObjectItem{
|
||||||
|
Keys: keys,
|
||||||
|
Assign: item.Assign,
|
||||||
|
Val: subitem.Val,
|
||||||
|
LeadComment: item.LeadComment,
|
||||||
|
LineComment: item.LineComment,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
return items, frontier
|
||||||
|
}
|
|
@ -0,0 +1,313 @@
|
||||||
|
package parser
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/hashicorp/hcl/hcl/ast"
|
||||||
|
hcltoken "github.com/hashicorp/hcl/hcl/token"
|
||||||
|
"github.com/hashicorp/hcl/json/scanner"
|
||||||
|
"github.com/hashicorp/hcl/json/token"
|
||||||
|
)
|
||||||
|
|
||||||
|
type Parser struct {
|
||||||
|
sc *scanner.Scanner
|
||||||
|
|
||||||
|
// Last read token
|
||||||
|
tok token.Token
|
||||||
|
commaPrev token.Token
|
||||||
|
|
||||||
|
enableTrace bool
|
||||||
|
indent int
|
||||||
|
n int // buffer size (max = 1)
|
||||||
|
}
|
||||||
|
|
||||||
|
func newParser(src []byte) *Parser {
|
||||||
|
return &Parser{
|
||||||
|
sc: scanner.New(src),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parse returns the fully parsed source and returns the abstract syntax tree.
|
||||||
|
func Parse(src []byte) (*ast.File, error) {
|
||||||
|
p := newParser(src)
|
||||||
|
return p.Parse()
|
||||||
|
}
|
||||||
|
|
||||||
|
var errEofToken = errors.New("EOF token found")
|
||||||
|
|
||||||
|
// Parse returns the fully parsed source and returns the abstract syntax tree.
|
||||||
|
func (p *Parser) Parse() (*ast.File, error) {
|
||||||
|
f := &ast.File{}
|
||||||
|
var err, scerr error
|
||||||
|
p.sc.Error = func(pos token.Pos, msg string) {
|
||||||
|
scerr = fmt.Errorf("%s: %s", pos, msg)
|
||||||
|
}
|
||||||
|
|
||||||
|
// The root must be an object in JSON
|
||||||
|
object, err := p.object()
|
||||||
|
if scerr != nil {
|
||||||
|
return nil, scerr
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// We make our final node an object list so it is more HCL compatible
|
||||||
|
f.Node = object.List
|
||||||
|
|
||||||
|
// Flatten it, which finds patterns and turns them into more HCL-like
|
||||||
|
// AST trees.
|
||||||
|
flattenObjects(f.Node)
|
||||||
|
|
||||||
|
return f, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *Parser) objectList() (*ast.ObjectList, error) {
|
||||||
|
defer un(trace(p, "ParseObjectList"))
|
||||||
|
node := &ast.ObjectList{}
|
||||||
|
|
||||||
|
for {
|
||||||
|
n, err := p.objectItem()
|
||||||
|
if err == errEofToken {
|
||||||
|
break // we are finished
|
||||||
|
}
|
||||||
|
|
||||||
|
// we don't return a nil node, because might want to use already
|
||||||
|
// collected items.
|
||||||
|
if err != nil {
|
||||||
|
return node, err
|
||||||
|
}
|
||||||
|
|
||||||
|
node.Add(n)
|
||||||
|
|
||||||
|
// Check for a followup comma. If it isn't a comma, then we're done
|
||||||
|
if tok := p.scan(); tok.Type != token.COMMA {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return node, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// objectItem parses a single object item
|
||||||
|
func (p *Parser) objectItem() (*ast.ObjectItem, error) {
|
||||||
|
defer un(trace(p, "ParseObjectItem"))
|
||||||
|
|
||||||
|
keys, err := p.objectKey()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
o := &ast.ObjectItem{
|
||||||
|
Keys: keys,
|
||||||
|
}
|
||||||
|
|
||||||
|
switch p.tok.Type {
|
||||||
|
case token.COLON:
|
||||||
|
pos := p.tok.Pos
|
||||||
|
o.Assign = hcltoken.Pos{
|
||||||
|
Filename: pos.Filename,
|
||||||
|
Offset: pos.Offset,
|
||||||
|
Line: pos.Line,
|
||||||
|
Column: pos.Column,
|
||||||
|
}
|
||||||
|
|
||||||
|
o.Val, err = p.objectValue()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return o, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// objectKey parses an object key and returns a ObjectKey AST
|
||||||
|
func (p *Parser) objectKey() ([]*ast.ObjectKey, error) {
|
||||||
|
keyCount := 0
|
||||||
|
keys := make([]*ast.ObjectKey, 0)
|
||||||
|
|
||||||
|
for {
|
||||||
|
tok := p.scan()
|
||||||
|
switch tok.Type {
|
||||||
|
case token.EOF:
|
||||||
|
return nil, errEofToken
|
||||||
|
case token.STRING:
|
||||||
|
keyCount++
|
||||||
|
keys = append(keys, &ast.ObjectKey{
|
||||||
|
Token: p.tok.HCLToken(),
|
||||||
|
})
|
||||||
|
case token.COLON:
|
||||||
|
// If we have a zero keycount it means that we never got
|
||||||
|
// an object key, i.e. `{ :`. This is a syntax error.
|
||||||
|
if keyCount == 0 {
|
||||||
|
return nil, fmt.Errorf("expected: STRING got: %s", p.tok.Type)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Done
|
||||||
|
return keys, nil
|
||||||
|
case token.ILLEGAL:
|
||||||
|
return nil, errors.New("illegal")
|
||||||
|
default:
|
||||||
|
return nil, fmt.Errorf("expected: STRING got: %s", p.tok.Type)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// object parses any type of object, such as number, bool, string, object or
|
||||||
|
// list.
|
||||||
|
func (p *Parser) objectValue() (ast.Node, error) {
|
||||||
|
defer un(trace(p, "ParseObjectValue"))
|
||||||
|
tok := p.scan()
|
||||||
|
|
||||||
|
switch tok.Type {
|
||||||
|
case token.NUMBER, token.FLOAT, token.BOOL, token.NULL, token.STRING:
|
||||||
|
return p.literalType()
|
||||||
|
case token.LBRACE:
|
||||||
|
return p.objectType()
|
||||||
|
case token.LBRACK:
|
||||||
|
return p.listType()
|
||||||
|
case token.EOF:
|
||||||
|
return nil, errEofToken
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil, fmt.Errorf("Expected object value, got unknown token: %+v", tok)
|
||||||
|
}
|
||||||
|
|
||||||
|
// object parses any type of object, such as number, bool, string, object or
|
||||||
|
// list.
|
||||||
|
func (p *Parser) object() (*ast.ObjectType, error) {
|
||||||
|
defer un(trace(p, "ParseType"))
|
||||||
|
tok := p.scan()
|
||||||
|
|
||||||
|
switch tok.Type {
|
||||||
|
case token.LBRACE:
|
||||||
|
return p.objectType()
|
||||||
|
case token.EOF:
|
||||||
|
return nil, errEofToken
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil, fmt.Errorf("Expected object, got unknown token: %+v", tok)
|
||||||
|
}
|
||||||
|
|
||||||
|
// objectType parses an object type and returns a ObjectType AST
|
||||||
|
func (p *Parser) objectType() (*ast.ObjectType, error) {
|
||||||
|
defer un(trace(p, "ParseObjectType"))
|
||||||
|
|
||||||
|
// we assume that the currently scanned token is a LBRACE
|
||||||
|
o := &ast.ObjectType{}
|
||||||
|
|
||||||
|
l, err := p.objectList()
|
||||||
|
|
||||||
|
// if we hit RBRACE, we are good to go (means we parsed all Items), if it's
|
||||||
|
// not a RBRACE, it's an syntax error and we just return it.
|
||||||
|
if err != nil && p.tok.Type != token.RBRACE {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
o.List = l
|
||||||
|
return o, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// listType parses a list type and returns a ListType AST
|
||||||
|
func (p *Parser) listType() (*ast.ListType, error) {
|
||||||
|
defer un(trace(p, "ParseListType"))
|
||||||
|
|
||||||
|
// we assume that the currently scanned token is a LBRACK
|
||||||
|
l := &ast.ListType{}
|
||||||
|
|
||||||
|
for {
|
||||||
|
tok := p.scan()
|
||||||
|
switch tok.Type {
|
||||||
|
case token.NUMBER, token.FLOAT, token.STRING:
|
||||||
|
node, err := p.literalType()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
l.Add(node)
|
||||||
|
case token.COMMA:
|
||||||
|
continue
|
||||||
|
case token.LBRACE:
|
||||||
|
node, err := p.objectType()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
l.Add(node)
|
||||||
|
case token.BOOL:
|
||||||
|
// TODO(arslan) should we support? not supported by HCL yet
|
||||||
|
case token.LBRACK:
|
||||||
|
// TODO(arslan) should we support nested lists? Even though it's
|
||||||
|
// written in README of HCL, it's not a part of the grammar
|
||||||
|
// (not defined in parse.y)
|
||||||
|
case token.RBRACK:
|
||||||
|
// finished
|
||||||
|
return l, nil
|
||||||
|
default:
|
||||||
|
return nil, fmt.Errorf("unexpected token while parsing list: %s", tok.Type)
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// literalType parses a literal type and returns a LiteralType AST
|
||||||
|
func (p *Parser) literalType() (*ast.LiteralType, error) {
|
||||||
|
defer un(trace(p, "ParseLiteral"))
|
||||||
|
|
||||||
|
return &ast.LiteralType{
|
||||||
|
Token: p.tok.HCLToken(),
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// scan returns the next token from the underlying scanner. If a token has
|
||||||
|
// been unscanned then read that instead.
|
||||||
|
func (p *Parser) scan() token.Token {
|
||||||
|
// If we have a token on the buffer, then return it.
|
||||||
|
if p.n != 0 {
|
||||||
|
p.n = 0
|
||||||
|
return p.tok
|
||||||
|
}
|
||||||
|
|
||||||
|
p.tok = p.sc.Scan()
|
||||||
|
return p.tok
|
||||||
|
}
|
||||||
|
|
||||||
|
// unscan pushes the previously read token back onto the buffer.
|
||||||
|
func (p *Parser) unscan() {
|
||||||
|
p.n = 1
|
||||||
|
}
|
||||||
|
|
||||||
|
// ----------------------------------------------------------------------------
|
||||||
|
// Parsing support
|
||||||
|
|
||||||
|
func (p *Parser) printTrace(a ...interface{}) {
|
||||||
|
if !p.enableTrace {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
const dots = ". . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . "
|
||||||
|
const n = len(dots)
|
||||||
|
fmt.Printf("%5d:%3d: ", p.tok.Pos.Line, p.tok.Pos.Column)
|
||||||
|
|
||||||
|
i := 2 * p.indent
|
||||||
|
for i > n {
|
||||||
|
fmt.Print(dots)
|
||||||
|
i -= n
|
||||||
|
}
|
||||||
|
// i <= n
|
||||||
|
fmt.Print(dots[0:i])
|
||||||
|
fmt.Println(a...)
|
||||||
|
}
|
||||||
|
|
||||||
|
func trace(p *Parser, msg string) *Parser {
|
||||||
|
p.printTrace(msg, "(")
|
||||||
|
p.indent++
|
||||||
|
return p
|
||||||
|
}
|
||||||
|
|
||||||
|
// Usage pattern: defer un(trace(p, "..."))
|
||||||
|
func un(p *Parser) {
|
||||||
|
p.indent--
|
||||||
|
p.printTrace(")")
|
||||||
|
}
|
|
@ -0,0 +1,451 @@
|
||||||
|
package scanner
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"os"
|
||||||
|
"unicode"
|
||||||
|
"unicode/utf8"
|
||||||
|
|
||||||
|
"github.com/hashicorp/hcl/json/token"
|
||||||
|
)
|
||||||
|
|
||||||
|
// eof represents a marker rune for the end of the reader.
|
||||||
|
const eof = rune(0)
|
||||||
|
|
||||||
|
// Scanner defines a lexical scanner
|
||||||
|
type Scanner struct {
|
||||||
|
buf *bytes.Buffer // Source buffer for advancing and scanning
|
||||||
|
src []byte // Source buffer for immutable access
|
||||||
|
|
||||||
|
// Source Position
|
||||||
|
srcPos token.Pos // current position
|
||||||
|
prevPos token.Pos // previous position, used for peek() method
|
||||||
|
|
||||||
|
lastCharLen int // length of last character in bytes
|
||||||
|
lastLineLen int // length of last line in characters (for correct column reporting)
|
||||||
|
|
||||||
|
tokStart int // token text start position
|
||||||
|
tokEnd int // token text end position
|
||||||
|
|
||||||
|
// Error is called for each error encountered. If no Error
|
||||||
|
// function is set, the error is reported to os.Stderr.
|
||||||
|
Error func(pos token.Pos, msg string)
|
||||||
|
|
||||||
|
// ErrorCount is incremented by one for each error encountered.
|
||||||
|
ErrorCount int
|
||||||
|
|
||||||
|
// tokPos is the start position of most recently scanned token; set by
|
||||||
|
// Scan. The Filename field is always left untouched by the Scanner. If
|
||||||
|
// an error is reported (via Error) and Position is invalid, the scanner is
|
||||||
|
// not inside a token.
|
||||||
|
tokPos token.Pos
|
||||||
|
}
|
||||||
|
|
||||||
|
// New creates and initializes a new instance of Scanner using src as
|
||||||
|
// its source content.
|
||||||
|
func New(src []byte) *Scanner {
|
||||||
|
// even though we accept a src, we read from a io.Reader compatible type
|
||||||
|
// (*bytes.Buffer). So in the future we might easily change it to streaming
|
||||||
|
// read.
|
||||||
|
b := bytes.NewBuffer(src)
|
||||||
|
s := &Scanner{
|
||||||
|
buf: b,
|
||||||
|
src: src,
|
||||||
|
}
|
||||||
|
|
||||||
|
// srcPosition always starts with 1
|
||||||
|
s.srcPos.Line = 1
|
||||||
|
return s
|
||||||
|
}
|
||||||
|
|
||||||
|
// next reads the next rune from the bufferred reader. Returns the rune(0) if
|
||||||
|
// an error occurs (or io.EOF is returned).
|
||||||
|
func (s *Scanner) next() rune {
|
||||||
|
ch, size, err := s.buf.ReadRune()
|
||||||
|
if err != nil {
|
||||||
|
// advance for error reporting
|
||||||
|
s.srcPos.Column++
|
||||||
|
s.srcPos.Offset += size
|
||||||
|
s.lastCharLen = size
|
||||||
|
return eof
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == utf8.RuneError && size == 1 {
|
||||||
|
s.srcPos.Column++
|
||||||
|
s.srcPos.Offset += size
|
||||||
|
s.lastCharLen = size
|
||||||
|
s.err("illegal UTF-8 encoding")
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
// remember last position
|
||||||
|
s.prevPos = s.srcPos
|
||||||
|
|
||||||
|
s.srcPos.Column++
|
||||||
|
s.lastCharLen = size
|
||||||
|
s.srcPos.Offset += size
|
||||||
|
|
||||||
|
if ch == '\n' {
|
||||||
|
s.srcPos.Line++
|
||||||
|
s.lastLineLen = s.srcPos.Column
|
||||||
|
s.srcPos.Column = 0
|
||||||
|
}
|
||||||
|
|
||||||
|
// debug
|
||||||
|
// fmt.Printf("ch: %q, offset:column: %d:%d\n", ch, s.srcPos.Offset, s.srcPos.Column)
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
// unread unreads the previous read Rune and updates the source position
|
||||||
|
func (s *Scanner) unread() {
|
||||||
|
if err := s.buf.UnreadRune(); err != nil {
|
||||||
|
panic(err) // this is user fault, we should catch it
|
||||||
|
}
|
||||||
|
s.srcPos = s.prevPos // put back last position
|
||||||
|
}
|
||||||
|
|
||||||
|
// peek returns the next rune without advancing the reader.
|
||||||
|
func (s *Scanner) peek() rune {
|
||||||
|
peek, _, err := s.buf.ReadRune()
|
||||||
|
if err != nil {
|
||||||
|
return eof
|
||||||
|
}
|
||||||
|
|
||||||
|
s.buf.UnreadRune()
|
||||||
|
return peek
|
||||||
|
}
|
||||||
|
|
||||||
|
// Scan scans the next token and returns the token.
|
||||||
|
func (s *Scanner) Scan() token.Token {
|
||||||
|
ch := s.next()
|
||||||
|
|
||||||
|
// skip white space
|
||||||
|
for isWhitespace(ch) {
|
||||||
|
ch = s.next()
|
||||||
|
}
|
||||||
|
|
||||||
|
var tok token.Type
|
||||||
|
|
||||||
|
// token text markings
|
||||||
|
s.tokStart = s.srcPos.Offset - s.lastCharLen
|
||||||
|
|
||||||
|
// token position, initial next() is moving the offset by one(size of rune
|
||||||
|
// actually), though we are interested with the starting point
|
||||||
|
s.tokPos.Offset = s.srcPos.Offset - s.lastCharLen
|
||||||
|
if s.srcPos.Column > 0 {
|
||||||
|
// common case: last character was not a '\n'
|
||||||
|
s.tokPos.Line = s.srcPos.Line
|
||||||
|
s.tokPos.Column = s.srcPos.Column
|
||||||
|
} else {
|
||||||
|
// last character was a '\n'
|
||||||
|
// (we cannot be at the beginning of the source
|
||||||
|
// since we have called next() at least once)
|
||||||
|
s.tokPos.Line = s.srcPos.Line - 1
|
||||||
|
s.tokPos.Column = s.lastLineLen
|
||||||
|
}
|
||||||
|
|
||||||
|
switch {
|
||||||
|
case isLetter(ch):
|
||||||
|
lit := s.scanIdentifier()
|
||||||
|
if lit == "true" || lit == "false" {
|
||||||
|
tok = token.BOOL
|
||||||
|
} else if lit == "null" {
|
||||||
|
tok = token.NULL
|
||||||
|
} else {
|
||||||
|
s.err("illegal char")
|
||||||
|
}
|
||||||
|
case isDecimal(ch):
|
||||||
|
tok = s.scanNumber(ch)
|
||||||
|
default:
|
||||||
|
switch ch {
|
||||||
|
case eof:
|
||||||
|
tok = token.EOF
|
||||||
|
case '"':
|
||||||
|
tok = token.STRING
|
||||||
|
s.scanString()
|
||||||
|
case '.':
|
||||||
|
tok = token.PERIOD
|
||||||
|
ch = s.peek()
|
||||||
|
if isDecimal(ch) {
|
||||||
|
tok = token.FLOAT
|
||||||
|
ch = s.scanMantissa(ch)
|
||||||
|
ch = s.scanExponent(ch)
|
||||||
|
}
|
||||||
|
case '[':
|
||||||
|
tok = token.LBRACK
|
||||||
|
case ']':
|
||||||
|
tok = token.RBRACK
|
||||||
|
case '{':
|
||||||
|
tok = token.LBRACE
|
||||||
|
case '}':
|
||||||
|
tok = token.RBRACE
|
||||||
|
case ',':
|
||||||
|
tok = token.COMMA
|
||||||
|
case ':':
|
||||||
|
tok = token.COLON
|
||||||
|
case '-':
|
||||||
|
if isDecimal(s.peek()) {
|
||||||
|
ch := s.next()
|
||||||
|
tok = s.scanNumber(ch)
|
||||||
|
} else {
|
||||||
|
s.err("illegal char")
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
s.err("illegal char: " + string(ch))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// finish token ending
|
||||||
|
s.tokEnd = s.srcPos.Offset
|
||||||
|
|
||||||
|
// create token literal
|
||||||
|
var tokenText string
|
||||||
|
if s.tokStart >= 0 {
|
||||||
|
tokenText = string(s.src[s.tokStart:s.tokEnd])
|
||||||
|
}
|
||||||
|
s.tokStart = s.tokEnd // ensure idempotency of tokenText() call
|
||||||
|
|
||||||
|
return token.Token{
|
||||||
|
Type: tok,
|
||||||
|
Pos: s.tokPos,
|
||||||
|
Text: tokenText,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanNumber scans a HCL number definition starting with the given rune
|
||||||
|
func (s *Scanner) scanNumber(ch rune) token.Type {
|
||||||
|
zero := ch == '0'
|
||||||
|
pos := s.srcPos
|
||||||
|
|
||||||
|
s.scanMantissa(ch)
|
||||||
|
ch = s.next() // seek forward
|
||||||
|
if ch == 'e' || ch == 'E' {
|
||||||
|
ch = s.scanExponent(ch)
|
||||||
|
return token.FLOAT
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == '.' {
|
||||||
|
ch = s.scanFraction(ch)
|
||||||
|
if ch == 'e' || ch == 'E' {
|
||||||
|
ch = s.next()
|
||||||
|
ch = s.scanExponent(ch)
|
||||||
|
}
|
||||||
|
return token.FLOAT
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch != eof {
|
||||||
|
s.unread()
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we have a larger number and this is zero, error
|
||||||
|
if zero && pos != s.srcPos {
|
||||||
|
s.err("numbers cannot start with 0")
|
||||||
|
}
|
||||||
|
|
||||||
|
return token.NUMBER
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanMantissa scans the mantissa beginning from the rune. It returns the next
|
||||||
|
// non decimal rune. It's used to determine wheter it's a fraction or exponent.
|
||||||
|
func (s *Scanner) scanMantissa(ch rune) rune {
|
||||||
|
scanned := false
|
||||||
|
for isDecimal(ch) {
|
||||||
|
ch = s.next()
|
||||||
|
scanned = true
|
||||||
|
}
|
||||||
|
|
||||||
|
if scanned && ch != eof {
|
||||||
|
s.unread()
|
||||||
|
}
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanFraction scans the fraction after the '.' rune
|
||||||
|
func (s *Scanner) scanFraction(ch rune) rune {
|
||||||
|
if ch == '.' {
|
||||||
|
ch = s.peek() // we peek just to see if we can move forward
|
||||||
|
ch = s.scanMantissa(ch)
|
||||||
|
}
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanExponent scans the remaining parts of an exponent after the 'e' or 'E'
|
||||||
|
// rune.
|
||||||
|
func (s *Scanner) scanExponent(ch rune) rune {
|
||||||
|
if ch == 'e' || ch == 'E' {
|
||||||
|
ch = s.next()
|
||||||
|
if ch == '-' || ch == '+' {
|
||||||
|
ch = s.next()
|
||||||
|
}
|
||||||
|
ch = s.scanMantissa(ch)
|
||||||
|
}
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanString scans a quoted string
|
||||||
|
func (s *Scanner) scanString() {
|
||||||
|
braces := 0
|
||||||
|
for {
|
||||||
|
// '"' opening already consumed
|
||||||
|
// read character after quote
|
||||||
|
ch := s.next()
|
||||||
|
|
||||||
|
if ch == '\n' || ch < 0 || ch == eof {
|
||||||
|
s.err("literal not terminated")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == '"' {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we're going into a ${} then we can ignore quotes for awhile
|
||||||
|
if braces == 0 && ch == '$' && s.peek() == '{' {
|
||||||
|
braces++
|
||||||
|
s.next()
|
||||||
|
} else if braces > 0 && ch == '{' {
|
||||||
|
braces++
|
||||||
|
}
|
||||||
|
if braces > 0 && ch == '}' {
|
||||||
|
braces--
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == '\\' {
|
||||||
|
s.scanEscape()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanEscape scans an escape sequence
|
||||||
|
func (s *Scanner) scanEscape() rune {
|
||||||
|
// http://en.cppreference.com/w/cpp/language/escape
|
||||||
|
ch := s.next() // read character after '/'
|
||||||
|
switch ch {
|
||||||
|
case 'a', 'b', 'f', 'n', 'r', 't', 'v', '\\', '"':
|
||||||
|
// nothing to do
|
||||||
|
case '0', '1', '2', '3', '4', '5', '6', '7':
|
||||||
|
// octal notation
|
||||||
|
ch = s.scanDigits(ch, 8, 3)
|
||||||
|
case 'x':
|
||||||
|
// hexademical notation
|
||||||
|
ch = s.scanDigits(s.next(), 16, 2)
|
||||||
|
case 'u':
|
||||||
|
// universal character name
|
||||||
|
ch = s.scanDigits(s.next(), 16, 4)
|
||||||
|
case 'U':
|
||||||
|
// universal character name
|
||||||
|
ch = s.scanDigits(s.next(), 16, 8)
|
||||||
|
default:
|
||||||
|
s.err("illegal char escape")
|
||||||
|
}
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanDigits scans a rune with the given base for n times. For example an
|
||||||
|
// octal notation \184 would yield in scanDigits(ch, 8, 3)
|
||||||
|
func (s *Scanner) scanDigits(ch rune, base, n int) rune {
|
||||||
|
for n > 0 && digitVal(ch) < base {
|
||||||
|
ch = s.next()
|
||||||
|
n--
|
||||||
|
}
|
||||||
|
if n > 0 {
|
||||||
|
s.err("illegal char escape")
|
||||||
|
}
|
||||||
|
|
||||||
|
// we scanned all digits, put the last non digit char back
|
||||||
|
s.unread()
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
|
||||||
|
// scanIdentifier scans an identifier and returns the literal string
|
||||||
|
func (s *Scanner) scanIdentifier() string {
|
||||||
|
offs := s.srcPos.Offset - s.lastCharLen
|
||||||
|
ch := s.next()
|
||||||
|
for isLetter(ch) || isDigit(ch) || ch == '-' {
|
||||||
|
ch = s.next()
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch != eof {
|
||||||
|
s.unread() // we got identifier, put back latest char
|
||||||
|
}
|
||||||
|
|
||||||
|
return string(s.src[offs:s.srcPos.Offset])
|
||||||
|
}
|
||||||
|
|
||||||
|
// recentPosition returns the position of the character immediately after the
|
||||||
|
// character or token returned by the last call to Scan.
|
||||||
|
func (s *Scanner) recentPosition() (pos token.Pos) {
|
||||||
|
pos.Offset = s.srcPos.Offset - s.lastCharLen
|
||||||
|
switch {
|
||||||
|
case s.srcPos.Column > 0:
|
||||||
|
// common case: last character was not a '\n'
|
||||||
|
pos.Line = s.srcPos.Line
|
||||||
|
pos.Column = s.srcPos.Column
|
||||||
|
case s.lastLineLen > 0:
|
||||||
|
// last character was a '\n'
|
||||||
|
// (we cannot be at the beginning of the source
|
||||||
|
// since we have called next() at least once)
|
||||||
|
pos.Line = s.srcPos.Line - 1
|
||||||
|
pos.Column = s.lastLineLen
|
||||||
|
default:
|
||||||
|
// at the beginning of the source
|
||||||
|
pos.Line = 1
|
||||||
|
pos.Column = 1
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// err prints the error of any scanning to s.Error function. If the function is
|
||||||
|
// not defined, by default it prints them to os.Stderr
|
||||||
|
func (s *Scanner) err(msg string) {
|
||||||
|
s.ErrorCount++
|
||||||
|
pos := s.recentPosition()
|
||||||
|
|
||||||
|
if s.Error != nil {
|
||||||
|
s.Error(pos, msg)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
fmt.Fprintf(os.Stderr, "%s: %s\n", pos, msg)
|
||||||
|
}
|
||||||
|
|
||||||
|
// isHexadecimal returns true if the given rune is a letter
|
||||||
|
func isLetter(ch rune) bool {
|
||||||
|
return 'a' <= ch && ch <= 'z' || 'A' <= ch && ch <= 'Z' || ch == '_' || ch >= 0x80 && unicode.IsLetter(ch)
|
||||||
|
}
|
||||||
|
|
||||||
|
// isHexadecimal returns true if the given rune is a decimal digit
|
||||||
|
func isDigit(ch rune) bool {
|
||||||
|
return '0' <= ch && ch <= '9' || ch >= 0x80 && unicode.IsDigit(ch)
|
||||||
|
}
|
||||||
|
|
||||||
|
// isHexadecimal returns true if the given rune is a decimal number
|
||||||
|
func isDecimal(ch rune) bool {
|
||||||
|
return '0' <= ch && ch <= '9'
|
||||||
|
}
|
||||||
|
|
||||||
|
// isHexadecimal returns true if the given rune is an hexadecimal number
|
||||||
|
func isHexadecimal(ch rune) bool {
|
||||||
|
return '0' <= ch && ch <= '9' || 'a' <= ch && ch <= 'f' || 'A' <= ch && ch <= 'F'
|
||||||
|
}
|
||||||
|
|
||||||
|
// isWhitespace returns true if the rune is a space, tab, newline or carriage return
|
||||||
|
func isWhitespace(ch rune) bool {
|
||||||
|
return ch == ' ' || ch == '\t' || ch == '\n' || ch == '\r'
|
||||||
|
}
|
||||||
|
|
||||||
|
// digitVal returns the integer value of a given octal,decimal or hexadecimal rune
|
||||||
|
func digitVal(ch rune) int {
|
||||||
|
switch {
|
||||||
|
case '0' <= ch && ch <= '9':
|
||||||
|
return int(ch - '0')
|
||||||
|
case 'a' <= ch && ch <= 'f':
|
||||||
|
return int(ch - 'a' + 10)
|
||||||
|
case 'A' <= ch && ch <= 'F':
|
||||||
|
return int(ch - 'A' + 10)
|
||||||
|
}
|
||||||
|
return 16 // larger than any legal digit val
|
||||||
|
}
|
|
@ -0,0 +1,46 @@
|
||||||
|
package token
|
||||||
|
|
||||||
|
import "fmt"
|
||||||
|
|
||||||
|
// Pos describes an arbitrary source position
|
||||||
|
// including the file, line, and column location.
|
||||||
|
// A Position is valid if the line number is > 0.
|
||||||
|
type Pos struct {
|
||||||
|
Filename string // filename, if any
|
||||||
|
Offset int // offset, starting at 0
|
||||||
|
Line int // line number, starting at 1
|
||||||
|
Column int // column number, starting at 1 (character count)
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsValid returns true if the position is valid.
|
||||||
|
func (p *Pos) IsValid() bool { return p.Line > 0 }
|
||||||
|
|
||||||
|
// String returns a string in one of several forms:
|
||||||
|
//
|
||||||
|
// file:line:column valid position with file name
|
||||||
|
// line:column valid position without file name
|
||||||
|
// file invalid position with file name
|
||||||
|
// - invalid position without file name
|
||||||
|
func (p Pos) String() string {
|
||||||
|
s := p.Filename
|
||||||
|
if p.IsValid() {
|
||||||
|
if s != "" {
|
||||||
|
s += ":"
|
||||||
|
}
|
||||||
|
s += fmt.Sprintf("%d:%d", p.Line, p.Column)
|
||||||
|
}
|
||||||
|
if s == "" {
|
||||||
|
s = "-"
|
||||||
|
}
|
||||||
|
return s
|
||||||
|
}
|
||||||
|
|
||||||
|
// Before reports whether the position p is before u.
|
||||||
|
func (p Pos) Before(u Pos) bool {
|
||||||
|
return u.Offset > p.Offset || u.Line > p.Line
|
||||||
|
}
|
||||||
|
|
||||||
|
// After reports whether the position p is after u.
|
||||||
|
func (p Pos) After(u Pos) bool {
|
||||||
|
return u.Offset < p.Offset || u.Line < p.Line
|
||||||
|
}
|
|
@ -0,0 +1,118 @@
|
||||||
|
package token
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"strconv"
|
||||||
|
|
||||||
|
hcltoken "github.com/hashicorp/hcl/hcl/token"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Token defines a single HCL token which can be obtained via the Scanner
|
||||||
|
type Token struct {
|
||||||
|
Type Type
|
||||||
|
Pos Pos
|
||||||
|
Text string
|
||||||
|
}
|
||||||
|
|
||||||
|
// Type is the set of lexical tokens of the HCL (HashiCorp Configuration Language)
|
||||||
|
type Type int
|
||||||
|
|
||||||
|
const (
|
||||||
|
// Special tokens
|
||||||
|
ILLEGAL Type = iota
|
||||||
|
EOF
|
||||||
|
|
||||||
|
identifier_beg
|
||||||
|
literal_beg
|
||||||
|
NUMBER // 12345
|
||||||
|
FLOAT // 123.45
|
||||||
|
BOOL // true,false
|
||||||
|
STRING // "abc"
|
||||||
|
NULL // null
|
||||||
|
literal_end
|
||||||
|
identifier_end
|
||||||
|
|
||||||
|
operator_beg
|
||||||
|
LBRACK // [
|
||||||
|
LBRACE // {
|
||||||
|
COMMA // ,
|
||||||
|
PERIOD // .
|
||||||
|
COLON // :
|
||||||
|
|
||||||
|
RBRACK // ]
|
||||||
|
RBRACE // }
|
||||||
|
|
||||||
|
operator_end
|
||||||
|
)
|
||||||
|
|
||||||
|
var tokens = [...]string{
|
||||||
|
ILLEGAL: "ILLEGAL",
|
||||||
|
|
||||||
|
EOF: "EOF",
|
||||||
|
|
||||||
|
NUMBER: "NUMBER",
|
||||||
|
FLOAT: "FLOAT",
|
||||||
|
BOOL: "BOOL",
|
||||||
|
STRING: "STRING",
|
||||||
|
NULL: "NULL",
|
||||||
|
|
||||||
|
LBRACK: "LBRACK",
|
||||||
|
LBRACE: "LBRACE",
|
||||||
|
COMMA: "COMMA",
|
||||||
|
PERIOD: "PERIOD",
|
||||||
|
COLON: "COLON",
|
||||||
|
|
||||||
|
RBRACK: "RBRACK",
|
||||||
|
RBRACE: "RBRACE",
|
||||||
|
}
|
||||||
|
|
||||||
|
// String returns the string corresponding to the token tok.
|
||||||
|
func (t Type) String() string {
|
||||||
|
s := ""
|
||||||
|
if 0 <= t && t < Type(len(tokens)) {
|
||||||
|
s = tokens[t]
|
||||||
|
}
|
||||||
|
if s == "" {
|
||||||
|
s = "token(" + strconv.Itoa(int(t)) + ")"
|
||||||
|
}
|
||||||
|
return s
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsIdentifier returns true for tokens corresponding to identifiers and basic
|
||||||
|
// type literals; it returns false otherwise.
|
||||||
|
func (t Type) IsIdentifier() bool { return identifier_beg < t && t < identifier_end }
|
||||||
|
|
||||||
|
// IsLiteral returns true for tokens corresponding to basic type literals; it
|
||||||
|
// returns false otherwise.
|
||||||
|
func (t Type) IsLiteral() bool { return literal_beg < t && t < literal_end }
|
||||||
|
|
||||||
|
// IsOperator returns true for tokens corresponding to operators and
|
||||||
|
// delimiters; it returns false otherwise.
|
||||||
|
func (t Type) IsOperator() bool { return operator_beg < t && t < operator_end }
|
||||||
|
|
||||||
|
// String returns the token's literal text. Note that this is only
|
||||||
|
// applicable for certain token types, such as token.IDENT,
|
||||||
|
// token.STRING, etc..
|
||||||
|
func (t Token) String() string {
|
||||||
|
return fmt.Sprintf("%s %s %s", t.Pos.String(), t.Type.String(), t.Text)
|
||||||
|
}
|
||||||
|
|
||||||
|
// HCLToken converts this token to an HCL token.
|
||||||
|
//
|
||||||
|
// The token type must be a literal type or this will panic.
|
||||||
|
func (t Token) HCLToken() hcltoken.Token {
|
||||||
|
switch t.Type {
|
||||||
|
case BOOL:
|
||||||
|
return hcltoken.Token{Type: hcltoken.BOOL, Text: t.Text}
|
||||||
|
case FLOAT:
|
||||||
|
return hcltoken.Token{Type: hcltoken.FLOAT, Text: t.Text}
|
||||||
|
case NULL:
|
||||||
|
return hcltoken.Token{Type: hcltoken.STRING, Text: ""}
|
||||||
|
case NUMBER:
|
||||||
|
return hcltoken.Token{Type: hcltoken.NUMBER, Text: t.Text}
|
||||||
|
case STRING:
|
||||||
|
return hcltoken.Token{Type: hcltoken.STRING, Text: t.Text, JSON: true}
|
||||||
|
default:
|
||||||
|
panic(fmt.Sprintf("unimplemented HCLToken for type: %s", t.Type))
|
||||||
|
}
|
||||||
|
}
|
|
@ -0,0 +1,38 @@
|
||||||
|
package hcl
|
||||||
|
|
||||||
|
import (
|
||||||
|
"unicode"
|
||||||
|
"unicode/utf8"
|
||||||
|
)
|
||||||
|
|
||||||
|
type lexModeValue byte
|
||||||
|
|
||||||
|
const (
|
||||||
|
lexModeUnknown lexModeValue = iota
|
||||||
|
lexModeHcl
|
||||||
|
lexModeJson
|
||||||
|
)
|
||||||
|
|
||||||
|
// lexMode returns whether we're going to be parsing in JSON
|
||||||
|
// mode or HCL mode.
|
||||||
|
func lexMode(v []byte) lexModeValue {
|
||||||
|
var (
|
||||||
|
r rune
|
||||||
|
w int
|
||||||
|
offset int
|
||||||
|
)
|
||||||
|
|
||||||
|
for {
|
||||||
|
r, w = utf8.DecodeRune(v[offset:])
|
||||||
|
offset += w
|
||||||
|
if unicode.IsSpace(r) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if r == '{' {
|
||||||
|
return lexModeJson
|
||||||
|
}
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
return lexModeHcl
|
||||||
|
}
|
|
@ -0,0 +1,39 @@
|
||||||
|
package hcl
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/hashicorp/hcl/hcl/ast"
|
||||||
|
hclParser "github.com/hashicorp/hcl/hcl/parser"
|
||||||
|
jsonParser "github.com/hashicorp/hcl/json/parser"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ParseBytes accepts as input byte slice and returns ast tree.
|
||||||
|
//
|
||||||
|
// Input can be either JSON or HCL
|
||||||
|
func ParseBytes(in []byte) (*ast.File, error) {
|
||||||
|
return parse(in)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ParseString accepts input as a string and returns ast tree.
|
||||||
|
func ParseString(input string) (*ast.File, error) {
|
||||||
|
return parse([]byte(input))
|
||||||
|
}
|
||||||
|
|
||||||
|
func parse(in []byte) (*ast.File, error) {
|
||||||
|
switch lexMode(in) {
|
||||||
|
case lexModeHcl:
|
||||||
|
return hclParser.Parse(in)
|
||||||
|
case lexModeJson:
|
||||||
|
return jsonParser.Parse(in)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil, fmt.Errorf("unknown config format")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parse parses the given input and returns the root object.
|
||||||
|
//
|
||||||
|
// The input format can be either HCL or JSON.
|
||||||
|
func Parse(input string) (*ast.File, error) {
|
||||||
|
return parse([]byte(input))
|
||||||
|
}
|
|
@ -0,0 +1,363 @@
|
||||||
|
Mozilla Public License, version 2.0
|
||||||
|
|
||||||
|
1. Definitions
|
||||||
|
|
||||||
|
1.1. "Contributor"
|
||||||
|
|
||||||
|
means each individual or legal entity that creates, contributes to the
|
||||||
|
creation of, or owns Covered Software.
|
||||||
|
|
||||||
|
1.2. "Contributor Version"
|
||||||
|
|
||||||
|
means the combination of the Contributions of others (if any) used by a
|
||||||
|
Contributor and that particular Contributor's Contribution.
|
||||||
|
|
||||||
|
1.3. "Contribution"
|
||||||
|
|
||||||
|
means Covered Software of a particular Contributor.
|
||||||
|
|
||||||
|
1.4. "Covered Software"
|
||||||
|
|
||||||
|
means Source Code Form to which the initial Contributor has attached the
|
||||||
|
notice in Exhibit A, the Executable Form of such Source Code Form, and
|
||||||
|
Modifications of such Source Code Form, in each case including portions
|
||||||
|
thereof.
|
||||||
|
|
||||||
|
1.5. "Incompatible With Secondary Licenses"
|
||||||
|
means
|
||||||
|
|
||||||
|
a. that the initial Contributor has attached the notice described in
|
||||||
|
Exhibit B to the Covered Software; or
|
||||||
|
|
||||||
|
b. that the Covered Software was made available under the terms of
|
||||||
|
version 1.1 or earlier of the License, but not also under the terms of
|
||||||
|
a Secondary License.
|
||||||
|
|
||||||
|
1.6. "Executable Form"
|
||||||
|
|
||||||
|
means any form of the work other than Source Code Form.
|
||||||
|
|
||||||
|
1.7. "Larger Work"
|
||||||
|
|
||||||
|
means a work that combines Covered Software with other material, in a
|
||||||
|
separate file or files, that is not Covered Software.
|
||||||
|
|
||||||
|
1.8. "License"
|
||||||
|
|
||||||
|
means this document.
|
||||||
|
|
||||||
|
1.9. "Licensable"
|
||||||
|
|
||||||
|
means having the right to grant, to the maximum extent possible, whether
|
||||||
|
at the time of the initial grant or subsequently, any and all of the
|
||||||
|
rights conveyed by this License.
|
||||||
|
|
||||||
|
1.10. "Modifications"
|
||||||
|
|
||||||
|
means any of the following:
|
||||||
|
|
||||||
|
a. any file in Source Code Form that results from an addition to,
|
||||||
|
deletion from, or modification of the contents of Covered Software; or
|
||||||
|
|
||||||
|
b. any new file in Source Code Form that contains any Covered Software.
|
||||||
|
|
||||||
|
1.11. "Patent Claims" of a Contributor
|
||||||
|
|
||||||
|
means any patent claim(s), including without limitation, method,
|
||||||
|
process, and apparatus claims, in any patent Licensable by such
|
||||||
|
Contributor that would be infringed, but for the grant of the License,
|
||||||
|
by the making, using, selling, offering for sale, having made, import,
|
||||||
|
or transfer of either its Contributions or its Contributor Version.
|
||||||
|
|
||||||
|
1.12. "Secondary License"
|
||||||
|
|
||||||
|
means either the GNU General Public License, Version 2.0, the GNU Lesser
|
||||||
|
General Public License, Version 2.1, the GNU Affero General Public
|
||||||
|
License, Version 3.0, or any later versions of those licenses.
|
||||||
|
|
||||||
|
1.13. "Source Code Form"
|
||||||
|
|
||||||
|
means the form of the work preferred for making modifications.
|
||||||
|
|
||||||
|
1.14. "You" (or "Your")
|
||||||
|
|
||||||
|
means an individual or a legal entity exercising rights under this
|
||||||
|
License. For legal entities, "You" includes any entity that controls, is
|
||||||
|
controlled by, or is under common control with You. For purposes of this
|
||||||
|
definition, "control" means (a) the power, direct or indirect, to cause
|
||||||
|
the direction or management of such entity, whether by contract or
|
||||||
|
otherwise, or (b) ownership of more than fifty percent (50%) of the
|
||||||
|
outstanding shares or beneficial ownership of such entity.
|
||||||
|
|
||||||
|
|
||||||
|
2. License Grants and Conditions
|
||||||
|
|
||||||
|
2.1. Grants
|
||||||
|
|
||||||
|
Each Contributor hereby grants You a world-wide, royalty-free,
|
||||||
|
non-exclusive license:
|
||||||
|
|
||||||
|
a. under intellectual property rights (other than patent or trademark)
|
||||||
|
Licensable by such Contributor to use, reproduce, make available,
|
||||||
|
modify, display, perform, distribute, and otherwise exploit its
|
||||||
|
Contributions, either on an unmodified basis, with Modifications, or
|
||||||
|
as part of a Larger Work; and
|
||||||
|
|
||||||
|
b. under Patent Claims of such Contributor to make, use, sell, offer for
|
||||||
|
sale, have made, import, and otherwise transfer either its
|
||||||
|
Contributions or its Contributor Version.
|
||||||
|
|
||||||
|
2.2. Effective Date
|
||||||
|
|
||||||
|
The licenses granted in Section 2.1 with respect to any Contribution
|
||||||
|
become effective for each Contribution on the date the Contributor first
|
||||||
|
distributes such Contribution.
|
||||||
|
|
||||||
|
2.3. Limitations on Grant Scope
|
||||||
|
|
||||||
|
The licenses granted in this Section 2 are the only rights granted under
|
||||||
|
this License. No additional rights or licenses will be implied from the
|
||||||
|
distribution or licensing of Covered Software under this License.
|
||||||
|
Notwithstanding Section 2.1(b) above, no patent license is granted by a
|
||||||
|
Contributor:
|
||||||
|
|
||||||
|
a. for any code that a Contributor has removed from Covered Software; or
|
||||||
|
|
||||||
|
b. for infringements caused by: (i) Your and any other third party's
|
||||||
|
modifications of Covered Software, or (ii) the combination of its
|
||||||
|
Contributions with other software (except as part of its Contributor
|
||||||
|
Version); or
|
||||||
|
|
||||||
|
c. under Patent Claims infringed by Covered Software in the absence of
|
||||||
|
its Contributions.
|
||||||
|
|
||||||
|
This License does not grant any rights in the trademarks, service marks,
|
||||||
|
or logos of any Contributor (except as may be necessary to comply with
|
||||||
|
the notice requirements in Section 3.4).
|
||||||
|
|
||||||
|
2.4. Subsequent Licenses
|
||||||
|
|
||||||
|
No Contributor makes additional grants as a result of Your choice to
|
||||||
|
distribute the Covered Software under a subsequent version of this
|
||||||
|
License (see Section 10.2) or under the terms of a Secondary License (if
|
||||||
|
permitted under the terms of Section 3.3).
|
||||||
|
|
||||||
|
2.5. Representation
|
||||||
|
|
||||||
|
Each Contributor represents that the Contributor believes its
|
||||||
|
Contributions are its original creation(s) or it has sufficient rights to
|
||||||
|
grant the rights to its Contributions conveyed by this License.
|
||||||
|
|
||||||
|
2.6. Fair Use
|
||||||
|
|
||||||
|
This License is not intended to limit any rights You have under
|
||||||
|
applicable copyright doctrines of fair use, fair dealing, or other
|
||||||
|
equivalents.
|
||||||
|
|
||||||
|
2.7. Conditions
|
||||||
|
|
||||||
|
Sections 3.1, 3.2, 3.3, and 3.4 are conditions of the licenses granted in
|
||||||
|
Section 2.1.
|
||||||
|
|
||||||
|
|
||||||
|
3. Responsibilities
|
||||||
|
|
||||||
|
3.1. Distribution of Source Form
|
||||||
|
|
||||||
|
All distribution of Covered Software in Source Code Form, including any
|
||||||
|
Modifications that You create or to which You contribute, must be under
|
||||||
|
the terms of this License. You must inform recipients that the Source
|
||||||
|
Code Form of the Covered Software is governed by the terms of this
|
||||||
|
License, and how they can obtain a copy of this License. You may not
|
||||||
|
attempt to alter or restrict the recipients' rights in the Source Code
|
||||||
|
Form.
|
||||||
|
|
||||||
|
3.2. Distribution of Executable Form
|
||||||
|
|
||||||
|
If You distribute Covered Software in Executable Form then:
|
||||||
|
|
||||||
|
a. such Covered Software must also be made available in Source Code Form,
|
||||||
|
as described in Section 3.1, and You must inform recipients of the
|
||||||
|
Executable Form how they can obtain a copy of such Source Code Form by
|
||||||
|
reasonable means in a timely manner, at a charge no more than the cost
|
||||||
|
of distribution to the recipient; and
|
||||||
|
|
||||||
|
b. You may distribute such Executable Form under the terms of this
|
||||||
|
License, or sublicense it under different terms, provided that the
|
||||||
|
license for the Executable Form does not attempt to limit or alter the
|
||||||
|
recipients' rights in the Source Code Form under this License.
|
||||||
|
|
||||||
|
3.3. Distribution of a Larger Work
|
||||||
|
|
||||||
|
You may create and distribute a Larger Work under terms of Your choice,
|
||||||
|
provided that You also comply with the requirements of this License for
|
||||||
|
the Covered Software. If the Larger Work is a combination of Covered
|
||||||
|
Software with a work governed by one or more Secondary Licenses, and the
|
||||||
|
Covered Software is not Incompatible With Secondary Licenses, this
|
||||||
|
License permits You to additionally distribute such Covered Software
|
||||||
|
under the terms of such Secondary License(s), so that the recipient of
|
||||||
|
the Larger Work may, at their option, further distribute the Covered
|
||||||
|
Software under the terms of either this License or such Secondary
|
||||||
|
License(s).
|
||||||
|
|
||||||
|
3.4. Notices
|
||||||
|
|
||||||
|
You may not remove or alter the substance of any license notices
|
||||||
|
(including copyright notices, patent notices, disclaimers of warranty, or
|
||||||
|
limitations of liability) contained within the Source Code Form of the
|
||||||
|
Covered Software, except that You may alter any license notices to the
|
||||||
|
extent required to remedy known factual inaccuracies.
|
||||||
|
|
||||||
|
3.5. Application of Additional Terms
|
||||||
|
|
||||||
|
You may choose to offer, and to charge a fee for, warranty, support,
|
||||||
|
indemnity or liability obligations to one or more recipients of Covered
|
||||||
|
Software. However, You may do so only on Your own behalf, and not on
|
||||||
|
behalf of any Contributor. You must make it absolutely clear that any
|
||||||
|
such warranty, support, indemnity, or liability obligation is offered by
|
||||||
|
You alone, and You hereby agree to indemnify every Contributor for any
|
||||||
|
liability incurred by such Contributor as a result of warranty, support,
|
||||||
|
indemnity or liability terms You offer. You may include additional
|
||||||
|
disclaimers of warranty and limitations of liability specific to any
|
||||||
|
jurisdiction.
|
||||||
|
|
||||||
|
4. Inability to Comply Due to Statute or Regulation
|
||||||
|
|
||||||
|
If it is impossible for You to comply with any of the terms of this License
|
||||||
|
with respect to some or all of the Covered Software due to statute,
|
||||||
|
judicial order, or regulation then You must: (a) comply with the terms of
|
||||||
|
this License to the maximum extent possible; and (b) describe the
|
||||||
|
limitations and the code they affect. Such description must be placed in a
|
||||||
|
text file included with all distributions of the Covered Software under
|
||||||
|
this License. Except to the extent prohibited by statute or regulation,
|
||||||
|
such description must be sufficiently detailed for a recipient of ordinary
|
||||||
|
skill to be able to understand it.
|
||||||
|
|
||||||
|
5. Termination
|
||||||
|
|
||||||
|
5.1. The rights granted under this License will terminate automatically if You
|
||||||
|
fail to comply with any of its terms. However, if You become compliant,
|
||||||
|
then the rights granted under this License from a particular Contributor
|
||||||
|
are reinstated (a) provisionally, unless and until such Contributor
|
||||||
|
explicitly and finally terminates Your grants, and (b) on an ongoing
|
||||||
|
basis, if such Contributor fails to notify You of the non-compliance by
|
||||||
|
some reasonable means prior to 60 days after You have come back into
|
||||||
|
compliance. Moreover, Your grants from a particular Contributor are
|
||||||
|
reinstated on an ongoing basis if such Contributor notifies You of the
|
||||||
|
non-compliance by some reasonable means, this is the first time You have
|
||||||
|
received notice of non-compliance with this License from such
|
||||||
|
Contributor, and You become compliant prior to 30 days after Your receipt
|
||||||
|
of the notice.
|
||||||
|
|
||||||
|
5.2. If You initiate litigation against any entity by asserting a patent
|
||||||
|
infringement claim (excluding declaratory judgment actions,
|
||||||
|
counter-claims, and cross-claims) alleging that a Contributor Version
|
||||||
|
directly or indirectly infringes any patent, then the rights granted to
|
||||||
|
You by any and all Contributors for the Covered Software under Section
|
||||||
|
2.1 of this License shall terminate.
|
||||||
|
|
||||||
|
5.3. In the event of termination under Sections 5.1 or 5.2 above, all end user
|
||||||
|
license agreements (excluding distributors and resellers) which have been
|
||||||
|
validly granted by You or Your distributors under this License prior to
|
||||||
|
termination shall survive termination.
|
||||||
|
|
||||||
|
6. Disclaimer of Warranty
|
||||||
|
|
||||||
|
Covered Software is provided under this License on an "as is" basis,
|
||||||
|
without warranty of any kind, either expressed, implied, or statutory,
|
||||||
|
including, without limitation, warranties that the Covered Software is free
|
||||||
|
of defects, merchantable, fit for a particular purpose or non-infringing.
|
||||||
|
The entire risk as to the quality and performance of the Covered Software
|
||||||
|
is with You. Should any Covered Software prove defective in any respect,
|
||||||
|
You (not any Contributor) assume the cost of any necessary servicing,
|
||||||
|
repair, or correction. This disclaimer of warranty constitutes an essential
|
||||||
|
part of this License. No use of any Covered Software is authorized under
|
||||||
|
this License except under this disclaimer.
|
||||||
|
|
||||||
|
7. Limitation of Liability
|
||||||
|
|
||||||
|
Under no circumstances and under no legal theory, whether tort (including
|
||||||
|
negligence), contract, or otherwise, shall any Contributor, or anyone who
|
||||||
|
distributes Covered Software as permitted above, be liable to You for any
|
||||||
|
direct, indirect, special, incidental, or consequential damages of any
|
||||||
|
character including, without limitation, damages for lost profits, loss of
|
||||||
|
goodwill, work stoppage, computer failure or malfunction, or any and all
|
||||||
|
other commercial damages or losses, even if such party shall have been
|
||||||
|
informed of the possibility of such damages. This limitation of liability
|
||||||
|
shall not apply to liability for death or personal injury resulting from
|
||||||
|
such party's negligence to the extent applicable law prohibits such
|
||||||
|
limitation. Some jurisdictions do not allow the exclusion or limitation of
|
||||||
|
incidental or consequential damages, so this exclusion and limitation may
|
||||||
|
not apply to You.
|
||||||
|
|
||||||
|
8. Litigation
|
||||||
|
|
||||||
|
Any litigation relating to this License may be brought only in the courts
|
||||||
|
of a jurisdiction where the defendant maintains its principal place of
|
||||||
|
business and such litigation shall be governed by laws of that
|
||||||
|
jurisdiction, without reference to its conflict-of-law provisions. Nothing
|
||||||
|
in this Section shall prevent a party's ability to bring cross-claims or
|
||||||
|
counter-claims.
|
||||||
|
|
||||||
|
9. Miscellaneous
|
||||||
|
|
||||||
|
This License represents the complete agreement concerning the subject
|
||||||
|
matter hereof. If any provision of this License is held to be
|
||||||
|
unenforceable, such provision shall be reformed only to the extent
|
||||||
|
necessary to make it enforceable. Any law or regulation which provides that
|
||||||
|
the language of a contract shall be construed against the drafter shall not
|
||||||
|
be used to construe this License against a Contributor.
|
||||||
|
|
||||||
|
|
||||||
|
10. Versions of the License
|
||||||
|
|
||||||
|
10.1. New Versions
|
||||||
|
|
||||||
|
Mozilla Foundation is the license steward. Except as provided in Section
|
||||||
|
10.3, no one other than the license steward has the right to modify or
|
||||||
|
publish new versions of this License. Each version will be given a
|
||||||
|
distinguishing version number.
|
||||||
|
|
||||||
|
10.2. Effect of New Versions
|
||||||
|
|
||||||
|
You may distribute the Covered Software under the terms of the version
|
||||||
|
of the License under which You originally received the Covered Software,
|
||||||
|
or under the terms of any subsequent version published by the license
|
||||||
|
steward.
|
||||||
|
|
||||||
|
10.3. Modified Versions
|
||||||
|
|
||||||
|
If you create software not governed by this License, and you want to
|
||||||
|
create a new license for such software, you may create and use a
|
||||||
|
modified version of this License if you rename the license and remove
|
||||||
|
any references to the name of the license steward (except to note that
|
||||||
|
such modified license differs from this License).
|
||||||
|
|
||||||
|
10.4. Distributing Source Code Form that is Incompatible With Secondary
|
||||||
|
Licenses If You choose to distribute Source Code Form that is
|
||||||
|
Incompatible With Secondary Licenses under the terms of this version of
|
||||||
|
the License, the notice described in Exhibit B of this License must be
|
||||||
|
attached.
|
||||||
|
|
||||||
|
Exhibit A - Source Code Form License Notice
|
||||||
|
|
||||||
|
This Source Code Form is subject to the
|
||||||
|
terms of the Mozilla Public License, v.
|
||||||
|
2.0. If a copy of the MPL was not
|
||||||
|
distributed with this file, You can
|
||||||
|
obtain one at
|
||||||
|
http://mozilla.org/MPL/2.0/.
|
||||||
|
|
||||||
|
If it is not possible or desirable to put the notice in a particular file,
|
||||||
|
then You may include the notice in a location (such as a LICENSE file in a
|
||||||
|
relevant directory) where a recipient would be likely to look for such a
|
||||||
|
notice.
|
||||||
|
|
||||||
|
You may add additional accurate notices of copyright ownership.
|
||||||
|
|
||||||
|
Exhibit B - "Incompatible With Secondary Licenses" Notice
|
||||||
|
|
||||||
|
This Source Code Form is "Incompatible
|
||||||
|
With Secondary Licenses", as defined by
|
||||||
|
the Mozilla Public License, v. 2.0.
|
||||||
|
|
|
@ -0,0 +1,11 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
// Auth is used to perform credential backend related operations.
|
||||||
|
type Auth struct {
|
||||||
|
c *Client
|
||||||
|
}
|
||||||
|
|
||||||
|
// Auth is used to return the client for credential-backend API calls.
|
||||||
|
func (c *Client) Auth() *Auth {
|
||||||
|
return &Auth{c: c}
|
||||||
|
}
|
|
@ -0,0 +1,243 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
// TokenAuth is used to perform token backend operations on Vault
|
||||||
|
type TokenAuth struct {
|
||||||
|
c *Client
|
||||||
|
}
|
||||||
|
|
||||||
|
// Token is used to return the client for token-backend API calls
|
||||||
|
func (a *Auth) Token() *TokenAuth {
|
||||||
|
return &TokenAuth{c: a.c}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *TokenAuth) Create(opts *TokenCreateRequest) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("POST", "/v1/auth/token/create")
|
||||||
|
if err := r.SetJSONBody(opts); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *TokenAuth) CreateOrphan(opts *TokenCreateRequest) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("POST", "/v1/auth/token/create-orphan")
|
||||||
|
if err := r.SetJSONBody(opts); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *TokenAuth) CreateWithRole(opts *TokenCreateRequest, roleName string) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("POST", "/v1/auth/token/create/"+roleName)
|
||||||
|
if err := r.SetJSONBody(opts); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *TokenAuth) Lookup(token string) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("POST", "/v1/auth/token/lookup")
|
||||||
|
if err := r.SetJSONBody(map[string]interface{}{
|
||||||
|
"token": token,
|
||||||
|
}); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *TokenAuth) LookupAccessor(accessor string) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("POST", "/v1/auth/token/lookup-accessor")
|
||||||
|
if err := r.SetJSONBody(map[string]interface{}{
|
||||||
|
"accessor": accessor,
|
||||||
|
}); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *TokenAuth) LookupSelf() (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/auth/token/lookup-self")
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *TokenAuth) Renew(token string, increment int) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/auth/token/renew")
|
||||||
|
if err := r.SetJSONBody(map[string]interface{}{
|
||||||
|
"token": token,
|
||||||
|
"increment": increment,
|
||||||
|
}); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *TokenAuth) RenewSelf(increment int) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/auth/token/renew-self")
|
||||||
|
|
||||||
|
body := map[string]interface{}{"increment": increment}
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
// RenewTokenAsSelf behaves like renew-self, but authenticates using a provided
|
||||||
|
// token instead of the token attached to the client.
|
||||||
|
func (c *TokenAuth) RenewTokenAsSelf(token string, increment int) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/auth/token/renew-self")
|
||||||
|
r.ClientToken = token
|
||||||
|
|
||||||
|
body := map[string]interface{}{"increment": increment}
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
// RevokeAccessor revokes a token associated with the given accessor
|
||||||
|
// along with all the child tokens.
|
||||||
|
func (c *TokenAuth) RevokeAccessor(accessor string) error {
|
||||||
|
r := c.c.NewRequest("POST", "/v1/auth/token/revoke-accessor")
|
||||||
|
if err := r.SetJSONBody(map[string]interface{}{
|
||||||
|
"accessor": accessor,
|
||||||
|
}); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// RevokeOrphan revokes a token without revoking the tree underneath it (so
|
||||||
|
// child tokens are orphaned rather than revoked)
|
||||||
|
func (c *TokenAuth) RevokeOrphan(token string) error {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/auth/token/revoke-orphan")
|
||||||
|
if err := r.SetJSONBody(map[string]interface{}{
|
||||||
|
"token": token,
|
||||||
|
}); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// RevokeSelf revokes the token making the call. The `token` parameter is kept
|
||||||
|
// for backwards compatibility but is ignored; only the client's set token has
|
||||||
|
// an effect.
|
||||||
|
func (c *TokenAuth) RevokeSelf(token string) error {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/auth/token/revoke-self")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// RevokeTree is the "normal" revoke operation that revokes the given token and
|
||||||
|
// the entire tree underneath -- all of its child tokens, their child tokens,
|
||||||
|
// etc.
|
||||||
|
func (c *TokenAuth) RevokeTree(token string) error {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/auth/token/revoke")
|
||||||
|
if err := r.SetJSONBody(map[string]interface{}{
|
||||||
|
"token": token,
|
||||||
|
}); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// TokenCreateRequest is the options structure for creating a token.
|
||||||
|
type TokenCreateRequest struct {
|
||||||
|
ID string `json:"id,omitempty"`
|
||||||
|
Policies []string `json:"policies,omitempty"`
|
||||||
|
Metadata map[string]string `json:"meta,omitempty"`
|
||||||
|
Lease string `json:"lease,omitempty"`
|
||||||
|
TTL string `json:"ttl,omitempty"`
|
||||||
|
ExplicitMaxTTL string `json:"explicit_max_ttl,omitempty"`
|
||||||
|
Period string `json:"period,omitempty"`
|
||||||
|
NoParent bool `json:"no_parent,omitempty"`
|
||||||
|
NoDefaultPolicy bool `json:"no_default_policy,omitempty"`
|
||||||
|
DisplayName string `json:"display_name"`
|
||||||
|
NumUses int `json:"num_uses"`
|
||||||
|
Renewable *bool `json:"renewable,omitempty"`
|
||||||
|
}
|
|
@ -0,0 +1,724 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"crypto/tls"
|
||||||
|
"fmt"
|
||||||
|
"net"
|
||||||
|
"net/http"
|
||||||
|
"net/url"
|
||||||
|
"os"
|
||||||
|
"path"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
"sync"
|
||||||
|
"time"
|
||||||
|
"unicode"
|
||||||
|
|
||||||
|
"github.com/hashicorp/errwrap"
|
||||||
|
"github.com/hashicorp/go-cleanhttp"
|
||||||
|
retryablehttp "github.com/hashicorp/go-retryablehttp"
|
||||||
|
"github.com/hashicorp/go-rootcerts"
|
||||||
|
"github.com/hashicorp/vault/helper/parseutil"
|
||||||
|
"golang.org/x/net/http2"
|
||||||
|
"golang.org/x/time/rate"
|
||||||
|
)
|
||||||
|
|
||||||
|
const EnvVaultAddress = "VAULT_ADDR"
|
||||||
|
const EnvVaultCACert = "VAULT_CACERT"
|
||||||
|
const EnvVaultCAPath = "VAULT_CAPATH"
|
||||||
|
const EnvVaultClientCert = "VAULT_CLIENT_CERT"
|
||||||
|
const EnvVaultClientKey = "VAULT_CLIENT_KEY"
|
||||||
|
const EnvVaultClientTimeout = "VAULT_CLIENT_TIMEOUT"
|
||||||
|
const EnvVaultInsecure = "VAULT_SKIP_VERIFY"
|
||||||
|
const EnvVaultTLSServerName = "VAULT_TLS_SERVER_NAME"
|
||||||
|
const EnvVaultWrapTTL = "VAULT_WRAP_TTL"
|
||||||
|
const EnvVaultMaxRetries = "VAULT_MAX_RETRIES"
|
||||||
|
const EnvVaultToken = "VAULT_TOKEN"
|
||||||
|
const EnvVaultMFA = "VAULT_MFA"
|
||||||
|
const EnvRateLimit = "VAULT_RATE_LIMIT"
|
||||||
|
|
||||||
|
// WrappingLookupFunc is a function that, given an HTTP verb and a path,
|
||||||
|
// returns an optional string duration to be used for response wrapping (e.g.
|
||||||
|
// "15s", or simply "15"). The path will not begin with "/v1/" or "v1/" or "/",
|
||||||
|
// however, end-of-path forward slashes are not trimmed, so must match your
|
||||||
|
// called path precisely.
|
||||||
|
type WrappingLookupFunc func(operation, path string) string
|
||||||
|
|
||||||
|
// Config is used to configure the creation of the client.
|
||||||
|
type Config struct {
|
||||||
|
modifyLock sync.RWMutex
|
||||||
|
|
||||||
|
// Address is the address of the Vault server. This should be a complete
|
||||||
|
// URL such as "http://vault.example.com". If you need a custom SSL
|
||||||
|
// cert or want to enable insecure mode, you need to specify a custom
|
||||||
|
// HttpClient.
|
||||||
|
Address string
|
||||||
|
|
||||||
|
// HttpClient is the HTTP client to use. Vault sets sane defaults for the
|
||||||
|
// http.Client and its associated http.Transport created in DefaultConfig.
|
||||||
|
// If you must modify Vault's defaults, it is suggested that you start with
|
||||||
|
// that client and modify as needed rather than start with an empty client
|
||||||
|
// (or http.DefaultClient).
|
||||||
|
HttpClient *http.Client
|
||||||
|
|
||||||
|
// MaxRetries controls the maximum number of times to retry when a 5xx
|
||||||
|
// error occurs. Set to 0 to disable retrying. Defaults to 2 (for a total
|
||||||
|
// of three tries).
|
||||||
|
MaxRetries int
|
||||||
|
|
||||||
|
// Timeout is for setting custom timeout parameter in the HttpClient
|
||||||
|
Timeout time.Duration
|
||||||
|
|
||||||
|
// If there is an error when creating the configuration, this will be the
|
||||||
|
// error
|
||||||
|
Error error
|
||||||
|
|
||||||
|
// The Backoff function to use; a default is used if not provided
|
||||||
|
Backoff retryablehttp.Backoff
|
||||||
|
|
||||||
|
// Limiter is the rate limiter used by the client.
|
||||||
|
// If this pointer is nil, then there will be no limit set.
|
||||||
|
// In contrast, if this pointer is set, even to an empty struct,
|
||||||
|
// then that limiter will be used. Note that an empty Limiter
|
||||||
|
// is equivalent blocking all events.
|
||||||
|
Limiter *rate.Limiter
|
||||||
|
}
|
||||||
|
|
||||||
|
// TLSConfig contains the parameters needed to configure TLS on the HTTP client
|
||||||
|
// used to communicate with Vault.
|
||||||
|
type TLSConfig struct {
|
||||||
|
// CACert is the path to a PEM-encoded CA cert file to use to verify the
|
||||||
|
// Vault server SSL certificate.
|
||||||
|
CACert string
|
||||||
|
|
||||||
|
// CAPath is the path to a directory of PEM-encoded CA cert files to verify
|
||||||
|
// the Vault server SSL certificate.
|
||||||
|
CAPath string
|
||||||
|
|
||||||
|
// ClientCert is the path to the certificate for Vault communication
|
||||||
|
ClientCert string
|
||||||
|
|
||||||
|
// ClientKey is the path to the private key for Vault communication
|
||||||
|
ClientKey string
|
||||||
|
|
||||||
|
// TLSServerName, if set, is used to set the SNI host when connecting via
|
||||||
|
// TLS.
|
||||||
|
TLSServerName string
|
||||||
|
|
||||||
|
// Insecure enables or disables SSL verification
|
||||||
|
Insecure bool
|
||||||
|
}
|
||||||
|
|
||||||
|
// DefaultConfig returns a default configuration for the client. It is
|
||||||
|
// safe to modify the return value of this function.
|
||||||
|
//
|
||||||
|
// The default Address is https://127.0.0.1:8200, but this can be overridden by
|
||||||
|
// setting the `VAULT_ADDR` environment variable.
|
||||||
|
//
|
||||||
|
// If an error is encountered, this will return nil.
|
||||||
|
func DefaultConfig() *Config {
|
||||||
|
config := &Config{
|
||||||
|
Address: "https://127.0.0.1:8200",
|
||||||
|
HttpClient: cleanhttp.DefaultClient(),
|
||||||
|
}
|
||||||
|
config.HttpClient.Timeout = time.Second * 60
|
||||||
|
|
||||||
|
transport := config.HttpClient.Transport.(*http.Transport)
|
||||||
|
transport.TLSHandshakeTimeout = 10 * time.Second
|
||||||
|
transport.TLSClientConfig = &tls.Config{
|
||||||
|
MinVersion: tls.VersionTLS12,
|
||||||
|
}
|
||||||
|
if err := http2.ConfigureTransport(transport); err != nil {
|
||||||
|
config.Error = err
|
||||||
|
return config
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := config.ReadEnvironment(); err != nil {
|
||||||
|
config.Error = err
|
||||||
|
return config
|
||||||
|
}
|
||||||
|
|
||||||
|
// Ensure redirects are not automatically followed
|
||||||
|
// Note that this is sane for the API client as it has its own
|
||||||
|
// redirect handling logic (and thus also for command/meta),
|
||||||
|
// but in e.g. http_test actual redirect handling is necessary
|
||||||
|
config.HttpClient.CheckRedirect = func(req *http.Request, via []*http.Request) error {
|
||||||
|
// Returning this value causes the Go net library to not close the
|
||||||
|
// response body and to nil out the error. Otherwise retry clients may
|
||||||
|
// try three times on every redirect because it sees an error from this
|
||||||
|
// function (to prevent redirects) passing through to it.
|
||||||
|
return http.ErrUseLastResponse
|
||||||
|
}
|
||||||
|
|
||||||
|
config.Backoff = retryablehttp.LinearJitterBackoff
|
||||||
|
config.MaxRetries = 2
|
||||||
|
|
||||||
|
return config
|
||||||
|
}
|
||||||
|
|
||||||
|
// ConfigureTLS takes a set of TLS configurations and applies those to the the
|
||||||
|
// HTTP client.
|
||||||
|
func (c *Config) ConfigureTLS(t *TLSConfig) error {
|
||||||
|
if c.HttpClient == nil {
|
||||||
|
c.HttpClient = DefaultConfig().HttpClient
|
||||||
|
}
|
||||||
|
clientTLSConfig := c.HttpClient.Transport.(*http.Transport).TLSClientConfig
|
||||||
|
|
||||||
|
var clientCert tls.Certificate
|
||||||
|
foundClientCert := false
|
||||||
|
|
||||||
|
switch {
|
||||||
|
case t.ClientCert != "" && t.ClientKey != "":
|
||||||
|
var err error
|
||||||
|
clientCert, err = tls.LoadX509KeyPair(t.ClientCert, t.ClientKey)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
foundClientCert = true
|
||||||
|
case t.ClientCert != "" || t.ClientKey != "":
|
||||||
|
return fmt.Errorf("both client cert and client key must be provided")
|
||||||
|
}
|
||||||
|
|
||||||
|
if t.CACert != "" || t.CAPath != "" {
|
||||||
|
rootConfig := &rootcerts.Config{
|
||||||
|
CAFile: t.CACert,
|
||||||
|
CAPath: t.CAPath,
|
||||||
|
}
|
||||||
|
if err := rootcerts.ConfigureTLS(clientTLSConfig, rootConfig); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if t.Insecure {
|
||||||
|
clientTLSConfig.InsecureSkipVerify = true
|
||||||
|
}
|
||||||
|
|
||||||
|
if foundClientCert {
|
||||||
|
// We use this function to ignore the server's preferential list of
|
||||||
|
// CAs, otherwise any CA used for the cert auth backend must be in the
|
||||||
|
// server's CA pool
|
||||||
|
clientTLSConfig.GetClientCertificate = func(*tls.CertificateRequestInfo) (*tls.Certificate, error) {
|
||||||
|
return &clientCert, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if t.TLSServerName != "" {
|
||||||
|
clientTLSConfig.ServerName = t.TLSServerName
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// ReadEnvironment reads configuration information from the environment. If
|
||||||
|
// there is an error, no configuration value is updated.
|
||||||
|
func (c *Config) ReadEnvironment() error {
|
||||||
|
var envAddress string
|
||||||
|
var envCACert string
|
||||||
|
var envCAPath string
|
||||||
|
var envClientCert string
|
||||||
|
var envClientKey string
|
||||||
|
var envClientTimeout time.Duration
|
||||||
|
var envInsecure bool
|
||||||
|
var envTLSServerName string
|
||||||
|
var envMaxRetries *uint64
|
||||||
|
var limit *rate.Limiter
|
||||||
|
|
||||||
|
// Parse the environment variables
|
||||||
|
if v := os.Getenv(EnvVaultAddress); v != "" {
|
||||||
|
envAddress = v
|
||||||
|
}
|
||||||
|
if v := os.Getenv(EnvVaultMaxRetries); v != "" {
|
||||||
|
maxRetries, err := strconv.ParseUint(v, 10, 32)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
envMaxRetries = &maxRetries
|
||||||
|
}
|
||||||
|
if v := os.Getenv(EnvVaultCACert); v != "" {
|
||||||
|
envCACert = v
|
||||||
|
}
|
||||||
|
if v := os.Getenv(EnvVaultCAPath); v != "" {
|
||||||
|
envCAPath = v
|
||||||
|
}
|
||||||
|
if v := os.Getenv(EnvVaultClientCert); v != "" {
|
||||||
|
envClientCert = v
|
||||||
|
}
|
||||||
|
if v := os.Getenv(EnvVaultClientKey); v != "" {
|
||||||
|
envClientKey = v
|
||||||
|
}
|
||||||
|
if v := os.Getenv(EnvRateLimit); v != "" {
|
||||||
|
rateLimit, burstLimit, err := parseRateLimit(v)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
limit = rate.NewLimiter(rate.Limit(rateLimit), burstLimit)
|
||||||
|
}
|
||||||
|
if t := os.Getenv(EnvVaultClientTimeout); t != "" {
|
||||||
|
clientTimeout, err := parseutil.ParseDurationSecond(t)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("could not parse %q", EnvVaultClientTimeout)
|
||||||
|
}
|
||||||
|
envClientTimeout = clientTimeout
|
||||||
|
}
|
||||||
|
if v := os.Getenv(EnvVaultInsecure); v != "" {
|
||||||
|
var err error
|
||||||
|
envInsecure, err = strconv.ParseBool(v)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("could not parse VAULT_SKIP_VERIFY")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if v := os.Getenv(EnvVaultTLSServerName); v != "" {
|
||||||
|
envTLSServerName = v
|
||||||
|
}
|
||||||
|
|
||||||
|
// Configure the HTTP clients TLS configuration.
|
||||||
|
t := &TLSConfig{
|
||||||
|
CACert: envCACert,
|
||||||
|
CAPath: envCAPath,
|
||||||
|
ClientCert: envClientCert,
|
||||||
|
ClientKey: envClientKey,
|
||||||
|
TLSServerName: envTLSServerName,
|
||||||
|
Insecure: envInsecure,
|
||||||
|
}
|
||||||
|
|
||||||
|
c.modifyLock.Lock()
|
||||||
|
defer c.modifyLock.Unlock()
|
||||||
|
|
||||||
|
c.Limiter = limit
|
||||||
|
|
||||||
|
if err := c.ConfigureTLS(t); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
if envAddress != "" {
|
||||||
|
c.Address = envAddress
|
||||||
|
}
|
||||||
|
|
||||||
|
if envMaxRetries != nil {
|
||||||
|
c.MaxRetries = int(*envMaxRetries)
|
||||||
|
}
|
||||||
|
|
||||||
|
if envClientTimeout != 0 {
|
||||||
|
c.Timeout = envClientTimeout
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func parseRateLimit(val string) (rate float64, burst int, err error) {
|
||||||
|
|
||||||
|
_, err = fmt.Sscanf(val, "%f:%d", &rate, &burst)
|
||||||
|
if err != nil {
|
||||||
|
rate, err = strconv.ParseFloat(val, 64)
|
||||||
|
if err != nil {
|
||||||
|
err = fmt.Errorf("%v was provided but incorrectly formatted", EnvRateLimit)
|
||||||
|
}
|
||||||
|
burst = int(rate)
|
||||||
|
}
|
||||||
|
|
||||||
|
return rate, burst, err
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
// Client is the client to the Vault API. Create a client with NewClient.
|
||||||
|
type Client struct {
|
||||||
|
modifyLock sync.RWMutex
|
||||||
|
addr *url.URL
|
||||||
|
config *Config
|
||||||
|
token string
|
||||||
|
headers http.Header
|
||||||
|
wrappingLookupFunc WrappingLookupFunc
|
||||||
|
mfaCreds []string
|
||||||
|
policyOverride bool
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewClient returns a new client for the given configuration.
|
||||||
|
//
|
||||||
|
// If the configuration is nil, Vault will use configuration from
|
||||||
|
// DefaultConfig(), which is the recommended starting configuration.
|
||||||
|
//
|
||||||
|
// If the environment variable `VAULT_TOKEN` is present, the token will be
|
||||||
|
// automatically added to the client. Otherwise, you must manually call
|
||||||
|
// `SetToken()`.
|
||||||
|
func NewClient(c *Config) (*Client, error) {
|
||||||
|
def := DefaultConfig()
|
||||||
|
if def == nil {
|
||||||
|
return nil, fmt.Errorf("could not create/read default configuration")
|
||||||
|
}
|
||||||
|
if def.Error != nil {
|
||||||
|
return nil, errwrap.Wrapf("error encountered setting up default configuration: {{err}}", def.Error)
|
||||||
|
}
|
||||||
|
|
||||||
|
if c == nil {
|
||||||
|
c = def
|
||||||
|
}
|
||||||
|
|
||||||
|
c.modifyLock.Lock()
|
||||||
|
defer c.modifyLock.Unlock()
|
||||||
|
|
||||||
|
u, err := url.Parse(c.Address)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if c.HttpClient == nil {
|
||||||
|
c.HttpClient = def.HttpClient
|
||||||
|
}
|
||||||
|
if c.HttpClient.Transport == nil {
|
||||||
|
c.HttpClient.Transport = def.HttpClient.Transport
|
||||||
|
}
|
||||||
|
|
||||||
|
client := &Client{
|
||||||
|
addr: u,
|
||||||
|
config: c,
|
||||||
|
}
|
||||||
|
|
||||||
|
if token := os.Getenv(EnvVaultToken); token != "" {
|
||||||
|
client.token = token
|
||||||
|
}
|
||||||
|
|
||||||
|
return client, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Sets the address of Vault in the client. The format of address should be
|
||||||
|
// "<Scheme>://<Host>:<Port>". Setting this on a client will override the
|
||||||
|
// value of VAULT_ADDR environment variable.
|
||||||
|
func (c *Client) SetAddress(addr string) error {
|
||||||
|
c.modifyLock.Lock()
|
||||||
|
defer c.modifyLock.Unlock()
|
||||||
|
|
||||||
|
parsedAddr, err := url.Parse(addr)
|
||||||
|
if err != nil {
|
||||||
|
return errwrap.Wrapf("failed to set address: {{err}}", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
c.addr = parsedAddr
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Address returns the Vault URL the client is configured to connect to
|
||||||
|
func (c *Client) Address() string {
|
||||||
|
c.modifyLock.RLock()
|
||||||
|
defer c.modifyLock.RUnlock()
|
||||||
|
|
||||||
|
return c.addr.String()
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetLimiter will set the rate limiter for this client.
|
||||||
|
// This method is thread-safe.
|
||||||
|
// rateLimit and burst are specified according to https://godoc.org/golang.org/x/time/rate#NewLimiter
|
||||||
|
func (c *Client) SetLimiter(rateLimit float64, burst int) {
|
||||||
|
c.modifyLock.RLock()
|
||||||
|
c.config.modifyLock.Lock()
|
||||||
|
defer c.config.modifyLock.Unlock()
|
||||||
|
c.modifyLock.RUnlock()
|
||||||
|
|
||||||
|
c.config.Limiter = rate.NewLimiter(rate.Limit(rateLimit), burst)
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetMaxRetries sets the number of retries that will be used in the case of certain errors
|
||||||
|
func (c *Client) SetMaxRetries(retries int) {
|
||||||
|
c.modifyLock.RLock()
|
||||||
|
c.config.modifyLock.Lock()
|
||||||
|
defer c.config.modifyLock.Unlock()
|
||||||
|
c.modifyLock.RUnlock()
|
||||||
|
|
||||||
|
c.config.MaxRetries = retries
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetClientTimeout sets the client request timeout
|
||||||
|
func (c *Client) SetClientTimeout(timeout time.Duration) {
|
||||||
|
c.modifyLock.RLock()
|
||||||
|
c.config.modifyLock.Lock()
|
||||||
|
defer c.config.modifyLock.Unlock()
|
||||||
|
c.modifyLock.RUnlock()
|
||||||
|
|
||||||
|
c.config.Timeout = timeout
|
||||||
|
}
|
||||||
|
|
||||||
|
// CurrentWrappingLookupFunc sets a lookup function that returns desired wrap TTLs
|
||||||
|
// for a given operation and path
|
||||||
|
func (c *Client) CurrentWrappingLookupFunc() WrappingLookupFunc {
|
||||||
|
c.modifyLock.RLock()
|
||||||
|
defer c.modifyLock.RUnlock()
|
||||||
|
|
||||||
|
return c.wrappingLookupFunc
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetWrappingLookupFunc sets a lookup function that returns desired wrap TTLs
|
||||||
|
// for a given operation and path
|
||||||
|
func (c *Client) SetWrappingLookupFunc(lookupFunc WrappingLookupFunc) {
|
||||||
|
c.modifyLock.Lock()
|
||||||
|
defer c.modifyLock.Unlock()
|
||||||
|
|
||||||
|
c.wrappingLookupFunc = lookupFunc
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetMFACreds sets the MFA credentials supplied either via the environment
|
||||||
|
// variable or via the command line.
|
||||||
|
func (c *Client) SetMFACreds(creds []string) {
|
||||||
|
c.modifyLock.Lock()
|
||||||
|
defer c.modifyLock.Unlock()
|
||||||
|
|
||||||
|
c.mfaCreds = creds
|
||||||
|
}
|
||||||
|
|
||||||
|
// Token returns the access token being used by this client. It will
|
||||||
|
// return the empty string if there is no token set.
|
||||||
|
func (c *Client) Token() string {
|
||||||
|
c.modifyLock.RLock()
|
||||||
|
defer c.modifyLock.RUnlock()
|
||||||
|
|
||||||
|
return c.token
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetToken sets the token directly. This won't perform any auth
|
||||||
|
// verification, it simply sets the token properly for future requests.
|
||||||
|
func (c *Client) SetToken(v string) {
|
||||||
|
c.modifyLock.Lock()
|
||||||
|
defer c.modifyLock.Unlock()
|
||||||
|
|
||||||
|
c.token = v
|
||||||
|
}
|
||||||
|
|
||||||
|
// ClearToken deletes the token if it is set or does nothing otherwise.
|
||||||
|
func (c *Client) ClearToken() {
|
||||||
|
c.modifyLock.Lock()
|
||||||
|
defer c.modifyLock.Unlock()
|
||||||
|
|
||||||
|
c.token = ""
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetHeaders sets the headers to be used for future requests.
|
||||||
|
func (c *Client) SetHeaders(headers http.Header) {
|
||||||
|
c.modifyLock.Lock()
|
||||||
|
defer c.modifyLock.Unlock()
|
||||||
|
|
||||||
|
c.headers = headers
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetBackoff sets the backoff function to be used for future requests.
|
||||||
|
func (c *Client) SetBackoff(backoff retryablehttp.Backoff) {
|
||||||
|
c.modifyLock.RLock()
|
||||||
|
c.config.modifyLock.Lock()
|
||||||
|
defer c.config.modifyLock.Unlock()
|
||||||
|
c.modifyLock.RUnlock()
|
||||||
|
|
||||||
|
c.config.Backoff = backoff
|
||||||
|
}
|
||||||
|
|
||||||
|
// Clone creates a new client with the same configuration. Note that the same
|
||||||
|
// underlying http.Client is used; modifying the client from more than one
|
||||||
|
// goroutine at once may not be safe, so modify the client as needed and then
|
||||||
|
// clone.
|
||||||
|
func (c *Client) Clone() (*Client, error) {
|
||||||
|
c.modifyLock.RLock()
|
||||||
|
c.config.modifyLock.RLock()
|
||||||
|
config := c.config
|
||||||
|
c.modifyLock.RUnlock()
|
||||||
|
|
||||||
|
newConfig := &Config{
|
||||||
|
Address: config.Address,
|
||||||
|
HttpClient: config.HttpClient,
|
||||||
|
MaxRetries: config.MaxRetries,
|
||||||
|
Timeout: config.Timeout,
|
||||||
|
Backoff: config.Backoff,
|
||||||
|
Limiter: config.Limiter,
|
||||||
|
}
|
||||||
|
config.modifyLock.RUnlock()
|
||||||
|
|
||||||
|
return NewClient(newConfig)
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetPolicyOverride sets whether requests should be sent with the policy
|
||||||
|
// override flag to request overriding soft-mandatory Sentinel policies (both
|
||||||
|
// RGPs and EGPs)
|
||||||
|
func (c *Client) SetPolicyOverride(override bool) {
|
||||||
|
c.modifyLock.Lock()
|
||||||
|
defer c.modifyLock.Unlock()
|
||||||
|
|
||||||
|
c.policyOverride = override
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewRequest creates a new raw request object to query the Vault server
|
||||||
|
// configured for this client. This is an advanced method and generally
|
||||||
|
// doesn't need to be called externally.
|
||||||
|
func (c *Client) NewRequest(method, requestPath string) *Request {
|
||||||
|
c.modifyLock.RLock()
|
||||||
|
addr := c.addr
|
||||||
|
token := c.token
|
||||||
|
mfaCreds := c.mfaCreds
|
||||||
|
wrappingLookupFunc := c.wrappingLookupFunc
|
||||||
|
headers := c.headers
|
||||||
|
policyOverride := c.policyOverride
|
||||||
|
c.modifyLock.RUnlock()
|
||||||
|
|
||||||
|
// if SRV records exist (see https://tools.ietf.org/html/draft-andrews-http-srv-02), lookup the SRV
|
||||||
|
// record and take the highest match; this is not designed for high-availability, just discovery
|
||||||
|
var host string = addr.Host
|
||||||
|
if addr.Port() == "" {
|
||||||
|
// Internet Draft specifies that the SRV record is ignored if a port is given
|
||||||
|
_, addrs, err := net.LookupSRV("http", "tcp", addr.Hostname())
|
||||||
|
if err == nil && len(addrs) > 0 {
|
||||||
|
host = fmt.Sprintf("%s:%d", addrs[0].Target, addrs[0].Port)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
req := &Request{
|
||||||
|
Method: method,
|
||||||
|
URL: &url.URL{
|
||||||
|
User: addr.User,
|
||||||
|
Scheme: addr.Scheme,
|
||||||
|
Host: host,
|
||||||
|
Path: path.Join(addr.Path, requestPath),
|
||||||
|
},
|
||||||
|
ClientToken: token,
|
||||||
|
Params: make(map[string][]string),
|
||||||
|
}
|
||||||
|
|
||||||
|
var lookupPath string
|
||||||
|
switch {
|
||||||
|
case strings.HasPrefix(requestPath, "/v1/"):
|
||||||
|
lookupPath = strings.TrimPrefix(requestPath, "/v1/")
|
||||||
|
case strings.HasPrefix(requestPath, "v1/"):
|
||||||
|
lookupPath = strings.TrimPrefix(requestPath, "v1/")
|
||||||
|
default:
|
||||||
|
lookupPath = requestPath
|
||||||
|
}
|
||||||
|
|
||||||
|
req.MFAHeaderVals = mfaCreds
|
||||||
|
|
||||||
|
if wrappingLookupFunc != nil {
|
||||||
|
req.WrapTTL = wrappingLookupFunc(method, lookupPath)
|
||||||
|
} else {
|
||||||
|
req.WrapTTL = DefaultWrappingLookupFunc(method, lookupPath)
|
||||||
|
}
|
||||||
|
|
||||||
|
if headers != nil {
|
||||||
|
req.Headers = headers
|
||||||
|
}
|
||||||
|
|
||||||
|
req.PolicyOverride = policyOverride
|
||||||
|
|
||||||
|
return req
|
||||||
|
}
|
||||||
|
|
||||||
|
// RawRequest performs the raw request given. This request may be against
|
||||||
|
// a Vault server not configured with this client. This is an advanced operation
|
||||||
|
// that generally won't need to be called externally.
|
||||||
|
func (c *Client) RawRequest(r *Request) (*Response, error) {
|
||||||
|
c.modifyLock.RLock()
|
||||||
|
token := c.token
|
||||||
|
|
||||||
|
c.config.modifyLock.RLock()
|
||||||
|
limiter := c.config.Limiter
|
||||||
|
maxRetries := c.config.MaxRetries
|
||||||
|
backoff := c.config.Backoff
|
||||||
|
httpClient := c.config.HttpClient
|
||||||
|
timeout := c.config.Timeout
|
||||||
|
c.config.modifyLock.RUnlock()
|
||||||
|
|
||||||
|
c.modifyLock.RUnlock()
|
||||||
|
|
||||||
|
if limiter != nil {
|
||||||
|
limiter.Wait(context.Background())
|
||||||
|
}
|
||||||
|
|
||||||
|
// Sanity check the token before potentially erroring from the API
|
||||||
|
idx := strings.IndexFunc(token, func(c rune) bool {
|
||||||
|
return !unicode.IsPrint(c)
|
||||||
|
})
|
||||||
|
if idx != -1 {
|
||||||
|
return nil, fmt.Errorf("configured Vault token contains non-printable characters and cannot be used")
|
||||||
|
}
|
||||||
|
|
||||||
|
redirectCount := 0
|
||||||
|
START:
|
||||||
|
req, err := r.toRetryableHTTP()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if req == nil {
|
||||||
|
return nil, fmt.Errorf("nil request created")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Set the timeout, if any
|
||||||
|
var cancelFunc context.CancelFunc
|
||||||
|
if timeout != 0 {
|
||||||
|
var ctx context.Context
|
||||||
|
ctx, cancelFunc = context.WithTimeout(context.Background(), timeout)
|
||||||
|
req.Request = req.Request.WithContext(ctx)
|
||||||
|
}
|
||||||
|
|
||||||
|
if backoff == nil {
|
||||||
|
backoff = retryablehttp.LinearJitterBackoff
|
||||||
|
}
|
||||||
|
|
||||||
|
client := &retryablehttp.Client{
|
||||||
|
HTTPClient: httpClient,
|
||||||
|
RetryWaitMin: 1000 * time.Millisecond,
|
||||||
|
RetryWaitMax: 1500 * time.Millisecond,
|
||||||
|
RetryMax: maxRetries,
|
||||||
|
CheckRetry: retryablehttp.DefaultRetryPolicy,
|
||||||
|
Backoff: backoff,
|
||||||
|
ErrorHandler: retryablehttp.PassthroughErrorHandler,
|
||||||
|
}
|
||||||
|
|
||||||
|
var result *Response
|
||||||
|
resp, err := client.Do(req)
|
||||||
|
if cancelFunc != nil {
|
||||||
|
cancelFunc()
|
||||||
|
}
|
||||||
|
if resp != nil {
|
||||||
|
result = &Response{Response: resp}
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
if strings.Contains(err.Error(), "tls: oversized") {
|
||||||
|
err = errwrap.Wrapf(
|
||||||
|
"{{err}}\n\n"+
|
||||||
|
"This error usually means that the server is running with TLS disabled\n"+
|
||||||
|
"but the client is configured to use TLS. Please either enable TLS\n"+
|
||||||
|
"on the server or run the client with -address set to an address\n"+
|
||||||
|
"that uses the http protocol:\n\n"+
|
||||||
|
" vault <command> -address http://<address>\n\n"+
|
||||||
|
"You can also set the VAULT_ADDR environment variable:\n\n\n"+
|
||||||
|
" VAULT_ADDR=http://<address> vault <command>\n\n"+
|
||||||
|
"where <address> is replaced by the actual address to the server.",
|
||||||
|
err)
|
||||||
|
}
|
||||||
|
return result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check for a redirect, only allowing for a single redirect
|
||||||
|
if (resp.StatusCode == 301 || resp.StatusCode == 302 || resp.StatusCode == 307) && redirectCount == 0 {
|
||||||
|
// Parse the updated location
|
||||||
|
respLoc, err := resp.Location()
|
||||||
|
if err != nil {
|
||||||
|
return result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Ensure a protocol downgrade doesn't happen
|
||||||
|
if req.URL.Scheme == "https" && respLoc.Scheme != "https" {
|
||||||
|
return result, fmt.Errorf("redirect would cause protocol downgrade")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update the request
|
||||||
|
r.URL = respLoc
|
||||||
|
|
||||||
|
// Reset the request body if any
|
||||||
|
if err := r.ResetJSONBody(); err != nil {
|
||||||
|
return result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Retry the request
|
||||||
|
redirectCount++
|
||||||
|
goto START
|
||||||
|
}
|
||||||
|
|
||||||
|
if err := result.Error(); err != nil {
|
||||||
|
return result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return result, nil
|
||||||
|
}
|
|
@ -0,0 +1,25 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Help reads the help information for the given path.
|
||||||
|
func (c *Client) Help(path string) (*Help, error) {
|
||||||
|
r := c.NewRequest("GET", fmt.Sprintf("/v1/%s", path))
|
||||||
|
r.Params.Add("help", "1")
|
||||||
|
resp, err := c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result Help
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
type Help struct {
|
||||||
|
Help string `json:"help"`
|
||||||
|
SeeAlso []string `json:"see_also"`
|
||||||
|
}
|
|
@ -0,0 +1,234 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"os"
|
||||||
|
|
||||||
|
"github.com/hashicorp/errwrap"
|
||||||
|
"github.com/hashicorp/vault/helper/jsonutil"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
wrappedResponseLocation = "cubbyhole/response"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
// The default TTL that will be used with `sys/wrapping/wrap`, can be
|
||||||
|
// changed
|
||||||
|
DefaultWrappingTTL = "5m"
|
||||||
|
|
||||||
|
// The default function used if no other function is set, which honors the
|
||||||
|
// env var and wraps `sys/wrapping/wrap`
|
||||||
|
DefaultWrappingLookupFunc = func(operation, path string) string {
|
||||||
|
if os.Getenv(EnvVaultWrapTTL) != "" {
|
||||||
|
return os.Getenv(EnvVaultWrapTTL)
|
||||||
|
}
|
||||||
|
|
||||||
|
if (operation == "PUT" || operation == "POST") && path == "sys/wrapping/wrap" {
|
||||||
|
return DefaultWrappingTTL
|
||||||
|
}
|
||||||
|
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
// Logical is used to perform logical backend operations on Vault.
|
||||||
|
type Logical struct {
|
||||||
|
c *Client
|
||||||
|
}
|
||||||
|
|
||||||
|
// Logical is used to return the client for logical-backend API calls.
|
||||||
|
func (c *Client) Logical() *Logical {
|
||||||
|
return &Logical{c: c}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Logical) Read(path string) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/"+path)
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if resp != nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
if resp != nil && resp.StatusCode == 404 {
|
||||||
|
secret, parseErr := ParseSecret(resp.Body)
|
||||||
|
switch parseErr {
|
||||||
|
case nil:
|
||||||
|
case io.EOF:
|
||||||
|
return nil, nil
|
||||||
|
default:
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if secret != nil && (len(secret.Warnings) > 0 || len(secret.Data) > 0) {
|
||||||
|
return secret, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Logical) List(path string) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("LIST", "/v1/"+path)
|
||||||
|
// Set this for broader compatibility, but we use LIST above to be able to
|
||||||
|
// handle the wrapping lookup function
|
||||||
|
r.Method = "GET"
|
||||||
|
r.Params.Set("list", "true")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if resp != nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
if resp != nil && resp.StatusCode == 404 {
|
||||||
|
secret, parseErr := ParseSecret(resp.Body)
|
||||||
|
switch parseErr {
|
||||||
|
case nil:
|
||||||
|
case io.EOF:
|
||||||
|
return nil, nil
|
||||||
|
default:
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if secret != nil && (len(secret.Warnings) > 0 || len(secret.Data) > 0) {
|
||||||
|
return secret, nil
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Logical) Write(path string, data map[string]interface{}) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/"+path)
|
||||||
|
if err := r.SetJSONBody(data); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if resp != nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
if resp != nil && resp.StatusCode == 404 {
|
||||||
|
secret, parseErr := ParseSecret(resp.Body)
|
||||||
|
switch parseErr {
|
||||||
|
case nil:
|
||||||
|
case io.EOF:
|
||||||
|
return nil, nil
|
||||||
|
default:
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if secret != nil && (len(secret.Warnings) > 0 || len(secret.Data) > 0) {
|
||||||
|
return secret, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Logical) Delete(path string) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("DELETE", "/v1/"+path)
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if resp != nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
if resp != nil && resp.StatusCode == 404 {
|
||||||
|
secret, parseErr := ParseSecret(resp.Body)
|
||||||
|
switch parseErr {
|
||||||
|
case nil:
|
||||||
|
case io.EOF:
|
||||||
|
return nil, nil
|
||||||
|
default:
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if secret != nil && (len(secret.Warnings) > 0 || len(secret.Data) > 0) {
|
||||||
|
return secret, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Logical) Unwrap(wrappingToken string) (*Secret, error) {
|
||||||
|
var data map[string]interface{}
|
||||||
|
if wrappingToken != "" {
|
||||||
|
if c.c.Token() == "" {
|
||||||
|
c.c.SetToken(wrappingToken)
|
||||||
|
} else if wrappingToken != c.c.Token() {
|
||||||
|
data = map[string]interface{}{
|
||||||
|
"token": wrappingToken,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/wrapping/unwrap")
|
||||||
|
if err := r.SetJSONBody(data); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if resp != nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
if resp == nil || resp.StatusCode != 404 {
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if resp == nil {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
// In the 404 case this may actually be a wrapped 404 error
|
||||||
|
secret, parseErr := ParseSecret(resp.Body)
|
||||||
|
switch parseErr {
|
||||||
|
case nil:
|
||||||
|
case io.EOF:
|
||||||
|
return nil, nil
|
||||||
|
default:
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if secret != nil && (len(secret.Warnings) > 0 || len(secret.Data) > 0) {
|
||||||
|
return secret, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Otherwise this might be an old-style wrapping token so attempt the old
|
||||||
|
// method
|
||||||
|
if wrappingToken != "" {
|
||||||
|
origToken := c.c.Token()
|
||||||
|
defer c.c.SetToken(origToken)
|
||||||
|
c.c.SetToken(wrappingToken)
|
||||||
|
}
|
||||||
|
|
||||||
|
secret, err = c.Read(wrappedResponseLocation)
|
||||||
|
if err != nil {
|
||||||
|
return nil, errwrap.Wrapf(fmt.Sprintf("error reading %q: {{err}}", wrappedResponseLocation), err)
|
||||||
|
}
|
||||||
|
if secret == nil {
|
||||||
|
return nil, fmt.Errorf("no value found at %q", wrappedResponseLocation)
|
||||||
|
}
|
||||||
|
if secret.Data == nil {
|
||||||
|
return nil, fmt.Errorf("\"data\" not found in wrapping response")
|
||||||
|
}
|
||||||
|
if _, ok := secret.Data["response"]; !ok {
|
||||||
|
return nil, fmt.Errorf("\"response\" not found in wrapping response \"data\" map")
|
||||||
|
}
|
||||||
|
|
||||||
|
wrappedSecret := new(Secret)
|
||||||
|
buf := bytes.NewBufferString(secret.Data["response"].(string))
|
||||||
|
if err := jsonutil.DecodeJSONFromReader(buf, wrappedSecret); err != nil {
|
||||||
|
return nil, errwrap.Wrapf("error unmarshalling wrapped secret: {{err}}", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return wrappedSecret, nil
|
||||||
|
}
|
|
@ -0,0 +1,349 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"math/rand"
|
||||||
|
"sync"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
ErrRenewerMissingInput = errors.New("missing input to renewer")
|
||||||
|
ErrRenewerMissingSecret = errors.New("missing secret to renew")
|
||||||
|
ErrRenewerNotRenewable = errors.New("secret is not renewable")
|
||||||
|
ErrRenewerNoSecretData = errors.New("returned empty secret data")
|
||||||
|
|
||||||
|
// DefaultRenewerRenewBuffer is the default size of the buffer for renew
|
||||||
|
// messages on the channel.
|
||||||
|
DefaultRenewerRenewBuffer = 5
|
||||||
|
)
|
||||||
|
|
||||||
|
// Renewer is a process for renewing a secret.
|
||||||
|
//
|
||||||
|
// renewer, err := client.NewRenewer(&RenewerInput{
|
||||||
|
// Secret: mySecret,
|
||||||
|
// })
|
||||||
|
// go renewer.Renew()
|
||||||
|
// defer renewer.Stop()
|
||||||
|
//
|
||||||
|
// for {
|
||||||
|
// select {
|
||||||
|
// case err := <-renewer.DoneCh():
|
||||||
|
// if err != nil {
|
||||||
|
// log.Fatal(err)
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// // Renewal is now over
|
||||||
|
// case renewal := <-renewer.RenewCh():
|
||||||
|
// log.Printf("Successfully renewed: %#v", renewal)
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
//
|
||||||
|
// The `DoneCh` will return if renewal fails or if the remaining lease duration
|
||||||
|
// after a renewal is less than or equal to the grace (in number of seconds). In
|
||||||
|
// both cases, the caller should attempt a re-read of the secret. Clients should
|
||||||
|
// check the return value of the channel to see if renewal was successful.
|
||||||
|
type Renewer struct {
|
||||||
|
l sync.Mutex
|
||||||
|
|
||||||
|
client *Client
|
||||||
|
secret *Secret
|
||||||
|
grace time.Duration
|
||||||
|
random *rand.Rand
|
||||||
|
increment int
|
||||||
|
doneCh chan error
|
||||||
|
renewCh chan *RenewOutput
|
||||||
|
|
||||||
|
stopped bool
|
||||||
|
stopCh chan struct{}
|
||||||
|
}
|
||||||
|
|
||||||
|
// RenewerInput is used as input to the renew function.
|
||||||
|
type RenewerInput struct {
|
||||||
|
// Secret is the secret to renew
|
||||||
|
Secret *Secret
|
||||||
|
|
||||||
|
// DEPRECATED: this does not do anything.
|
||||||
|
Grace time.Duration
|
||||||
|
|
||||||
|
// Rand is the randomizer to use for underlying randomization. If not
|
||||||
|
// provided, one will be generated and seeded automatically. If provided, it
|
||||||
|
// is assumed to have already been seeded.
|
||||||
|
Rand *rand.Rand
|
||||||
|
|
||||||
|
// RenewBuffer is the size of the buffered channel where renew messages are
|
||||||
|
// dispatched.
|
||||||
|
RenewBuffer int
|
||||||
|
|
||||||
|
// The new TTL, in seconds, that should be set on the lease. The TTL set
|
||||||
|
// here may or may not be honored by the vault server, based on Vault
|
||||||
|
// configuration or any associated max TTL values.
|
||||||
|
Increment int
|
||||||
|
}
|
||||||
|
|
||||||
|
// RenewOutput is the metadata returned to the client (if it's listening) to
|
||||||
|
// renew messages.
|
||||||
|
type RenewOutput struct {
|
||||||
|
// RenewedAt is the timestamp when the renewal took place (UTC).
|
||||||
|
RenewedAt time.Time
|
||||||
|
|
||||||
|
// Secret is the underlying renewal data. It's the same struct as all data
|
||||||
|
// that is returned from Vault, but since this is renewal data, it will not
|
||||||
|
// usually include the secret itself.
|
||||||
|
Secret *Secret
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewRenewer creates a new renewer from the given input.
|
||||||
|
func (c *Client) NewRenewer(i *RenewerInput) (*Renewer, error) {
|
||||||
|
if i == nil {
|
||||||
|
return nil, ErrRenewerMissingInput
|
||||||
|
}
|
||||||
|
|
||||||
|
secret := i.Secret
|
||||||
|
if secret == nil {
|
||||||
|
return nil, ErrRenewerMissingSecret
|
||||||
|
}
|
||||||
|
|
||||||
|
random := i.Rand
|
||||||
|
if random == nil {
|
||||||
|
random = rand.New(rand.NewSource(int64(time.Now().Nanosecond())))
|
||||||
|
}
|
||||||
|
|
||||||
|
renewBuffer := i.RenewBuffer
|
||||||
|
if renewBuffer == 0 {
|
||||||
|
renewBuffer = DefaultRenewerRenewBuffer
|
||||||
|
}
|
||||||
|
|
||||||
|
return &Renewer{
|
||||||
|
client: c,
|
||||||
|
secret: secret,
|
||||||
|
increment: i.Increment,
|
||||||
|
random: random,
|
||||||
|
doneCh: make(chan error, 1),
|
||||||
|
renewCh: make(chan *RenewOutput, renewBuffer),
|
||||||
|
|
||||||
|
stopped: false,
|
||||||
|
stopCh: make(chan struct{}),
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// DoneCh returns the channel where the renewer will publish when renewal stops.
|
||||||
|
// If there is an error, this will be an error.
|
||||||
|
func (r *Renewer) DoneCh() <-chan error {
|
||||||
|
return r.doneCh
|
||||||
|
}
|
||||||
|
|
||||||
|
// RenewCh is a channel that receives a message when a successful renewal takes
|
||||||
|
// place and includes metadata about the renewal.
|
||||||
|
func (r *Renewer) RenewCh() <-chan *RenewOutput {
|
||||||
|
return r.renewCh
|
||||||
|
}
|
||||||
|
|
||||||
|
// Stop stops the renewer.
|
||||||
|
func (r *Renewer) Stop() {
|
||||||
|
r.l.Lock()
|
||||||
|
if !r.stopped {
|
||||||
|
close(r.stopCh)
|
||||||
|
r.stopped = true
|
||||||
|
}
|
||||||
|
r.l.Unlock()
|
||||||
|
}
|
||||||
|
|
||||||
|
// Renew starts a background process for renewing this secret. When the secret
|
||||||
|
// has auth data, this attempts to renew the auth (token). When the secret has
|
||||||
|
// a lease, this attempts to renew the lease.
|
||||||
|
func (r *Renewer) Renew() {
|
||||||
|
var result error
|
||||||
|
if r.secret.Auth != nil {
|
||||||
|
result = r.renewAuth()
|
||||||
|
} else {
|
||||||
|
result = r.renewLease()
|
||||||
|
}
|
||||||
|
|
||||||
|
r.doneCh <- result
|
||||||
|
}
|
||||||
|
|
||||||
|
// renewAuth is a helper for renewing authentication.
|
||||||
|
func (r *Renewer) renewAuth() error {
|
||||||
|
if !r.secret.Auth.Renewable || r.secret.Auth.ClientToken == "" {
|
||||||
|
return ErrRenewerNotRenewable
|
||||||
|
}
|
||||||
|
|
||||||
|
priorDuration := time.Duration(r.secret.Auth.LeaseDuration) * time.Second
|
||||||
|
r.calculateGrace(priorDuration)
|
||||||
|
|
||||||
|
client, token := r.client, r.secret.Auth.ClientToken
|
||||||
|
|
||||||
|
for {
|
||||||
|
// Check if we are stopped.
|
||||||
|
select {
|
||||||
|
case <-r.stopCh:
|
||||||
|
return nil
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
|
||||||
|
// Renew the auth.
|
||||||
|
renewal, err := client.Auth().Token().RenewTokenAsSelf(token, r.increment)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Push a message that a renewal took place.
|
||||||
|
select {
|
||||||
|
case r.renewCh <- &RenewOutput{time.Now().UTC(), renewal}:
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
|
||||||
|
// Somehow, sometimes, this happens.
|
||||||
|
if renewal == nil || renewal.Auth == nil {
|
||||||
|
return ErrRenewerNoSecretData
|
||||||
|
}
|
||||||
|
|
||||||
|
// Do nothing if we are not renewable
|
||||||
|
if !renewal.Auth.Renewable {
|
||||||
|
return ErrRenewerNotRenewable
|
||||||
|
}
|
||||||
|
|
||||||
|
// Grab the lease duration
|
||||||
|
leaseDuration := time.Duration(renewal.Auth.LeaseDuration) * time.Second
|
||||||
|
|
||||||
|
// We keep evaluating a new grace period so long as the lease is
|
||||||
|
// extending. Once it stops extending, we've hit the max and need to
|
||||||
|
// rely on the grace duration.
|
||||||
|
if leaseDuration > priorDuration {
|
||||||
|
r.calculateGrace(leaseDuration)
|
||||||
|
}
|
||||||
|
priorDuration = leaseDuration
|
||||||
|
|
||||||
|
// The sleep duration is set to 2/3 of the current lease duration plus
|
||||||
|
// 1/3 of the current grace period, which adds jitter.
|
||||||
|
sleepDuration := time.Duration(float64(leaseDuration.Nanoseconds())*2/3 + float64(r.grace.Nanoseconds())/3)
|
||||||
|
|
||||||
|
// If we are within grace, return now; or, if the amount of time we
|
||||||
|
// would sleep would land us in the grace period. This helps with short
|
||||||
|
// tokens; for example, you don't want a current lease duration of 4
|
||||||
|
// seconds, a grace period of 3 seconds, and end up sleeping for more
|
||||||
|
// than three of those seconds and having a very small budget of time
|
||||||
|
// to renew.
|
||||||
|
if leaseDuration <= r.grace || leaseDuration-sleepDuration <= r.grace {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-r.stopCh:
|
||||||
|
return nil
|
||||||
|
case <-time.After(sleepDuration):
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// renewLease is a helper for renewing a lease.
|
||||||
|
func (r *Renewer) renewLease() error {
|
||||||
|
if !r.secret.Renewable || r.secret.LeaseID == "" {
|
||||||
|
return ErrRenewerNotRenewable
|
||||||
|
}
|
||||||
|
|
||||||
|
priorDuration := time.Duration(r.secret.LeaseDuration) * time.Second
|
||||||
|
r.calculateGrace(priorDuration)
|
||||||
|
|
||||||
|
client, leaseID := r.client, r.secret.LeaseID
|
||||||
|
|
||||||
|
for {
|
||||||
|
// Check if we are stopped.
|
||||||
|
select {
|
||||||
|
case <-r.stopCh:
|
||||||
|
return nil
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
|
||||||
|
// Renew the lease.
|
||||||
|
renewal, err := client.Sys().Renew(leaseID, r.increment)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Push a message that a renewal took place.
|
||||||
|
select {
|
||||||
|
case r.renewCh <- &RenewOutput{time.Now().UTC(), renewal}:
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
|
||||||
|
// Somehow, sometimes, this happens.
|
||||||
|
if renewal == nil {
|
||||||
|
return ErrRenewerNoSecretData
|
||||||
|
}
|
||||||
|
|
||||||
|
// Do nothing if we are not renewable
|
||||||
|
if !renewal.Renewable {
|
||||||
|
return ErrRenewerNotRenewable
|
||||||
|
}
|
||||||
|
|
||||||
|
// Grab the lease duration
|
||||||
|
leaseDuration := time.Duration(renewal.LeaseDuration) * time.Second
|
||||||
|
|
||||||
|
// We keep evaluating a new grace period so long as the lease is
|
||||||
|
// extending. Once it stops extending, we've hit the max and need to
|
||||||
|
// rely on the grace duration.
|
||||||
|
if leaseDuration > priorDuration {
|
||||||
|
r.calculateGrace(leaseDuration)
|
||||||
|
}
|
||||||
|
priorDuration = leaseDuration
|
||||||
|
|
||||||
|
// The sleep duration is set to 2/3 of the current lease duration plus
|
||||||
|
// 1/3 of the current grace period, which adds jitter.
|
||||||
|
sleepDuration := time.Duration(float64(leaseDuration.Nanoseconds())*2/3 + float64(r.grace.Nanoseconds())/3)
|
||||||
|
|
||||||
|
// If we are within grace, return now; or, if the amount of time we
|
||||||
|
// would sleep would land us in the grace period. This helps with short
|
||||||
|
// tokens; for example, you don't want a current lease duration of 4
|
||||||
|
// seconds, a grace period of 3 seconds, and end up sleeping for more
|
||||||
|
// than three of those seconds and having a very small budget of time
|
||||||
|
// to renew.
|
||||||
|
if leaseDuration <= r.grace || leaseDuration-sleepDuration <= r.grace {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-r.stopCh:
|
||||||
|
return nil
|
||||||
|
case <-time.After(sleepDuration):
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// sleepDuration calculates the time to sleep given the base lease duration. The
|
||||||
|
// base is the resulting lease duration. It will be reduced to 1/3 and
|
||||||
|
// multiplied by a random float between 0.0 and 1.0. This extra randomness
|
||||||
|
// prevents multiple clients from all trying to renew simultaneously.
|
||||||
|
func (r *Renewer) sleepDuration(base time.Duration) time.Duration {
|
||||||
|
sleep := float64(base)
|
||||||
|
|
||||||
|
// Renew at 1/3 the remaining lease. This will give us an opportunity to retry
|
||||||
|
// at least one more time should the first renewal fail.
|
||||||
|
sleep = sleep / 3.0
|
||||||
|
|
||||||
|
// Use a randomness so many clients do not hit Vault simultaneously.
|
||||||
|
sleep = sleep * (r.random.Float64() + 1) / 2.0
|
||||||
|
|
||||||
|
return time.Duration(sleep)
|
||||||
|
}
|
||||||
|
|
||||||
|
// calculateGrace calculates the grace period based on a reasonable set of
|
||||||
|
// assumptions given the total lease time; it also adds some jitter to not have
|
||||||
|
// clients be in sync.
|
||||||
|
func (r *Renewer) calculateGrace(leaseDuration time.Duration) {
|
||||||
|
if leaseDuration == 0 {
|
||||||
|
r.grace = 0
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
leaseNanos := float64(leaseDuration.Nanoseconds())
|
||||||
|
jitterMax := 0.1 * leaseNanos
|
||||||
|
|
||||||
|
// For a given lease duration, we want to allow 80-90% of that to elapse,
|
||||||
|
// so the remaining amount is the grace period
|
||||||
|
r.grace = time.Duration(jitterMax) + time.Duration(uint64(r.random.Int63())%uint64(jitterMax))
|
||||||
|
}
|
|
@ -0,0 +1,145 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"encoding/json"
|
||||||
|
"io"
|
||||||
|
"io/ioutil"
|
||||||
|
"net/http"
|
||||||
|
"net/url"
|
||||||
|
|
||||||
|
retryablehttp "github.com/hashicorp/go-retryablehttp"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Request is a raw request configuration structure used to initiate
|
||||||
|
// API requests to the Vault server.
|
||||||
|
type Request struct {
|
||||||
|
Method string
|
||||||
|
URL *url.URL
|
||||||
|
Params url.Values
|
||||||
|
Headers http.Header
|
||||||
|
ClientToken string
|
||||||
|
MFAHeaderVals []string
|
||||||
|
WrapTTL string
|
||||||
|
Obj interface{}
|
||||||
|
|
||||||
|
// When possible, use BodyBytes as it is more efficient due to how the
|
||||||
|
// retry logic works
|
||||||
|
BodyBytes []byte
|
||||||
|
|
||||||
|
// Fallback
|
||||||
|
Body io.Reader
|
||||||
|
BodySize int64
|
||||||
|
|
||||||
|
// Whether to request overriding soft-mandatory Sentinel policies (RGPs and
|
||||||
|
// EGPs). If set, the override flag will take effect for all policies
|
||||||
|
// evaluated during the request.
|
||||||
|
PolicyOverride bool
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetJSONBody is used to set a request body that is a JSON-encoded value.
|
||||||
|
func (r *Request) SetJSONBody(val interface{}) error {
|
||||||
|
buf, err := json.Marshal(val)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
r.Obj = val
|
||||||
|
r.BodyBytes = buf
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// ResetJSONBody is used to reset the body for a redirect
|
||||||
|
func (r *Request) ResetJSONBody() error {
|
||||||
|
if r.BodyBytes == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return r.SetJSONBody(r.Obj)
|
||||||
|
}
|
||||||
|
|
||||||
|
// DEPRECATED: ToHTTP turns this request into a valid *http.Request for use
|
||||||
|
// with the net/http package.
|
||||||
|
func (r *Request) ToHTTP() (*http.Request, error) {
|
||||||
|
req, err := r.toRetryableHTTP()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
switch {
|
||||||
|
case r.BodyBytes == nil && r.Body == nil:
|
||||||
|
// No body
|
||||||
|
|
||||||
|
case r.BodyBytes != nil:
|
||||||
|
req.Request.Body = ioutil.NopCloser(bytes.NewReader(r.BodyBytes))
|
||||||
|
|
||||||
|
default:
|
||||||
|
if c, ok := r.Body.(io.ReadCloser); ok {
|
||||||
|
req.Request.Body = c
|
||||||
|
} else {
|
||||||
|
req.Request.Body = ioutil.NopCloser(r.Body)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return req.Request, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *Request) toRetryableHTTP() (*retryablehttp.Request, error) {
|
||||||
|
// Encode the query parameters
|
||||||
|
r.URL.RawQuery = r.Params.Encode()
|
||||||
|
|
||||||
|
// Create the HTTP request, defaulting to retryable
|
||||||
|
var req *retryablehttp.Request
|
||||||
|
|
||||||
|
var err error
|
||||||
|
var body interface{}
|
||||||
|
|
||||||
|
switch {
|
||||||
|
case r.BodyBytes == nil && r.Body == nil:
|
||||||
|
// No body
|
||||||
|
|
||||||
|
case r.BodyBytes != nil:
|
||||||
|
// Use bytes, it's more efficient
|
||||||
|
body = r.BodyBytes
|
||||||
|
|
||||||
|
default:
|
||||||
|
body = r.Body
|
||||||
|
}
|
||||||
|
|
||||||
|
req, err = retryablehttp.NewRequest(r.Method, r.URL.RequestURI(), body)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
req.URL.User = r.URL.User
|
||||||
|
req.URL.Scheme = r.URL.Scheme
|
||||||
|
req.URL.Host = r.URL.Host
|
||||||
|
req.Host = r.URL.Host
|
||||||
|
|
||||||
|
if r.Headers != nil {
|
||||||
|
for header, vals := range r.Headers {
|
||||||
|
for _, val := range vals {
|
||||||
|
req.Header.Add(header, val)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(r.ClientToken) != 0 {
|
||||||
|
req.Header.Set("X-Vault-Token", r.ClientToken)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(r.WrapTTL) != 0 {
|
||||||
|
req.Header.Set("X-Vault-Wrap-TTL", r.WrapTTL)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(r.MFAHeaderVals) != 0 {
|
||||||
|
for _, mfaHeaderVal := range r.MFAHeaderVals {
|
||||||
|
req.Header.Add("X-Vault-MFA", mfaHeaderVal)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if r.PolicyOverride {
|
||||||
|
req.Header.Set("X-Vault-Policy-Override", "true")
|
||||||
|
}
|
||||||
|
|
||||||
|
return req, nil
|
||||||
|
}
|
|
@ -0,0 +1,77 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"io/ioutil"
|
||||||
|
"net/http"
|
||||||
|
|
||||||
|
"github.com/hashicorp/vault/helper/jsonutil"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Response is a raw response that wraps an HTTP response.
|
||||||
|
type Response struct {
|
||||||
|
*http.Response
|
||||||
|
}
|
||||||
|
|
||||||
|
// DecodeJSON will decode the response body to a JSON structure. This
|
||||||
|
// will consume the response body, but will not close it. Close must
|
||||||
|
// still be called.
|
||||||
|
func (r *Response) DecodeJSON(out interface{}) error {
|
||||||
|
return jsonutil.DecodeJSONFromReader(r.Body, out)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Error returns an error response if there is one. If there is an error,
|
||||||
|
// this will fully consume the response body, but will not close it. The
|
||||||
|
// body must still be closed manually.
|
||||||
|
func (r *Response) Error() error {
|
||||||
|
// 200 to 399 are okay status codes. 429 is the code for health status of
|
||||||
|
// standby nodes.
|
||||||
|
if (r.StatusCode >= 200 && r.StatusCode < 400) || r.StatusCode == 429 {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// We have an error. Let's copy the body into our own buffer first,
|
||||||
|
// so that if we can't decode JSON, we can at least copy it raw.
|
||||||
|
bodyBuf := &bytes.Buffer{}
|
||||||
|
if _, err := io.Copy(bodyBuf, r.Body); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
r.Body.Close()
|
||||||
|
r.Body = ioutil.NopCloser(bodyBuf)
|
||||||
|
|
||||||
|
// Decode the error response if we can. Note that we wrap the bodyBuf
|
||||||
|
// in a bytes.Reader here so that the JSON decoder doesn't move the
|
||||||
|
// read pointer for the original buffer.
|
||||||
|
var resp ErrorResponse
|
||||||
|
if err := jsonutil.DecodeJSON(bodyBuf.Bytes(), &resp); err != nil {
|
||||||
|
// Ignore the decoding error and just drop the raw response
|
||||||
|
return fmt.Errorf(
|
||||||
|
"Error making API request.\n\n"+
|
||||||
|
"URL: %s %s\n"+
|
||||||
|
"Code: %d. Raw Message:\n\n%s",
|
||||||
|
r.Request.Method, r.Request.URL.String(),
|
||||||
|
r.StatusCode, bodyBuf.String())
|
||||||
|
}
|
||||||
|
|
||||||
|
var errBody bytes.Buffer
|
||||||
|
errBody.WriteString(fmt.Sprintf(
|
||||||
|
"Error making API request.\n\n"+
|
||||||
|
"URL: %s %s\n"+
|
||||||
|
"Code: %d. Errors:\n\n",
|
||||||
|
r.Request.Method, r.Request.URL.String(),
|
||||||
|
r.StatusCode))
|
||||||
|
for _, err := range resp.Errors {
|
||||||
|
errBody.WriteString(fmt.Sprintf("* %s", err))
|
||||||
|
}
|
||||||
|
|
||||||
|
return fmt.Errorf(errBody.String())
|
||||||
|
}
|
||||||
|
|
||||||
|
// ErrorResponse is the raw structure of errors when they're returned by the
|
||||||
|
// HTTP API.
|
||||||
|
type ErrorResponse struct {
|
||||||
|
Errors []string
|
||||||
|
}
|
|
@ -0,0 +1,320 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/hashicorp/errwrap"
|
||||||
|
"github.com/hashicorp/vault/helper/jsonutil"
|
||||||
|
"github.com/hashicorp/vault/helper/parseutil"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Secret is the structure returned for every secret within Vault.
|
||||||
|
type Secret struct {
|
||||||
|
// The request ID that generated this response
|
||||||
|
RequestID string `json:"request_id"`
|
||||||
|
|
||||||
|
LeaseID string `json:"lease_id"`
|
||||||
|
LeaseDuration int `json:"lease_duration"`
|
||||||
|
Renewable bool `json:"renewable"`
|
||||||
|
|
||||||
|
// Data is the actual contents of the secret. The format of the data
|
||||||
|
// is arbitrary and up to the secret backend.
|
||||||
|
Data map[string]interface{} `json:"data"`
|
||||||
|
|
||||||
|
// Warnings contains any warnings related to the operation. These
|
||||||
|
// are not issues that caused the command to fail, but that the
|
||||||
|
// client should be aware of.
|
||||||
|
Warnings []string `json:"warnings"`
|
||||||
|
|
||||||
|
// Auth, if non-nil, means that there was authentication information
|
||||||
|
// attached to this response.
|
||||||
|
Auth *SecretAuth `json:"auth,omitempty"`
|
||||||
|
|
||||||
|
// WrapInfo, if non-nil, means that the initial response was wrapped in the
|
||||||
|
// cubbyhole of the given token (which has a TTL of the given number of
|
||||||
|
// seconds)
|
||||||
|
WrapInfo *SecretWrapInfo `json:"wrap_info,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// TokenID returns the standardized token ID (token) for the given secret.
|
||||||
|
func (s *Secret) TokenID() (string, error) {
|
||||||
|
if s == nil {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if s.Auth != nil && len(s.Auth.ClientToken) > 0 {
|
||||||
|
return s.Auth.ClientToken, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if s.Data == nil || s.Data["id"] == nil {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
id, ok := s.Data["id"].(string)
|
||||||
|
if !ok {
|
||||||
|
return "", fmt.Errorf("token found but in the wrong format")
|
||||||
|
}
|
||||||
|
|
||||||
|
return id, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// TokenAccessor returns the standardized token accessor for the given secret.
|
||||||
|
// If the secret is nil or does not contain an accessor, this returns the empty
|
||||||
|
// string.
|
||||||
|
func (s *Secret) TokenAccessor() (string, error) {
|
||||||
|
if s == nil {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if s.Auth != nil && len(s.Auth.Accessor) > 0 {
|
||||||
|
return s.Auth.Accessor, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if s.Data == nil || s.Data["accessor"] == nil {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
|
||||||
|
accessor, ok := s.Data["accessor"].(string)
|
||||||
|
if !ok {
|
||||||
|
return "", fmt.Errorf("token found but in the wrong format")
|
||||||
|
}
|
||||||
|
|
||||||
|
return accessor, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// TokenRemainingUses returns the standardized remaining uses for the given
|
||||||
|
// secret. If the secret is nil or does not contain the "num_uses", this
|
||||||
|
// returns -1. On error, this will return -1 and a non-nil error.
|
||||||
|
func (s *Secret) TokenRemainingUses() (int, error) {
|
||||||
|
if s == nil || s.Data == nil || s.Data["num_uses"] == nil {
|
||||||
|
return -1, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
uses, err := parseutil.ParseInt(s.Data["num_uses"])
|
||||||
|
if err != nil {
|
||||||
|
return 0, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return int(uses), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// TokenPolicies returns the standardized list of policies for the given secret.
|
||||||
|
// If the secret is nil or does not contain any policies, this returns nil. It
|
||||||
|
// also populates the secret's Auth info with identity/token policy info.
|
||||||
|
func (s *Secret) TokenPolicies() ([]string, error) {
|
||||||
|
if s == nil {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if s.Auth != nil && len(s.Auth.Policies) > 0 {
|
||||||
|
return s.Auth.Policies, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if s.Data == nil || s.Data["policies"] == nil {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
var tokenPolicies []string
|
||||||
|
|
||||||
|
// Token policies
|
||||||
|
{
|
||||||
|
_, ok := s.Data["policies"]
|
||||||
|
if !ok {
|
||||||
|
goto TOKEN_DONE
|
||||||
|
}
|
||||||
|
|
||||||
|
sList, ok := s.Data["policies"].([]string)
|
||||||
|
if ok {
|
||||||
|
tokenPolicies = sList
|
||||||
|
goto TOKEN_DONE
|
||||||
|
}
|
||||||
|
|
||||||
|
list, ok := s.Data["policies"].([]interface{})
|
||||||
|
if !ok {
|
||||||
|
return nil, fmt.Errorf("unable to convert token policies to expected format")
|
||||||
|
}
|
||||||
|
for _, v := range list {
|
||||||
|
p, ok := v.(string)
|
||||||
|
if !ok {
|
||||||
|
return nil, fmt.Errorf("unable to convert policy %v to string", v)
|
||||||
|
}
|
||||||
|
tokenPolicies = append(tokenPolicies, p)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
TOKEN_DONE:
|
||||||
|
var identityPolicies []string
|
||||||
|
|
||||||
|
// Identity policies
|
||||||
|
{
|
||||||
|
_, ok := s.Data["identity_policies"]
|
||||||
|
if !ok {
|
||||||
|
goto DONE
|
||||||
|
}
|
||||||
|
|
||||||
|
sList, ok := s.Data["identity_policies"].([]string)
|
||||||
|
if ok {
|
||||||
|
identityPolicies = sList
|
||||||
|
goto DONE
|
||||||
|
}
|
||||||
|
|
||||||
|
list, ok := s.Data["identity_policies"].([]interface{})
|
||||||
|
if !ok {
|
||||||
|
return nil, fmt.Errorf("unable to convert identity policies to expected format")
|
||||||
|
}
|
||||||
|
for _, v := range list {
|
||||||
|
p, ok := v.(string)
|
||||||
|
if !ok {
|
||||||
|
return nil, fmt.Errorf("unable to convert policy %v to string", v)
|
||||||
|
}
|
||||||
|
identityPolicies = append(identityPolicies, p)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
DONE:
|
||||||
|
|
||||||
|
if s.Auth == nil {
|
||||||
|
s.Auth = &SecretAuth{}
|
||||||
|
}
|
||||||
|
|
||||||
|
policies := append(tokenPolicies, identityPolicies...)
|
||||||
|
|
||||||
|
s.Auth.TokenPolicies = tokenPolicies
|
||||||
|
s.Auth.IdentityPolicies = identityPolicies
|
||||||
|
s.Auth.Policies = policies
|
||||||
|
|
||||||
|
return policies, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// TokenMetadata returns the map of metadata associated with this token, if any
|
||||||
|
// exists. If the secret is nil or does not contain the "metadata" key, this
|
||||||
|
// returns nil.
|
||||||
|
func (s *Secret) TokenMetadata() (map[string]string, error) {
|
||||||
|
if s == nil {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if s.Auth != nil && len(s.Auth.Metadata) > 0 {
|
||||||
|
return s.Auth.Metadata, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if s.Data == nil || (s.Data["metadata"] == nil && s.Data["meta"] == nil) {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
data, ok := s.Data["metadata"].(map[string]interface{})
|
||||||
|
if !ok {
|
||||||
|
data, ok = s.Data["meta"].(map[string]interface{})
|
||||||
|
if !ok {
|
||||||
|
return nil, fmt.Errorf("unable to convert metadata field to expected format")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
metadata := make(map[string]string, len(data))
|
||||||
|
for k, v := range data {
|
||||||
|
typed, ok := v.(string)
|
||||||
|
if !ok {
|
||||||
|
return nil, fmt.Errorf("unable to convert metadata value %v to string", v)
|
||||||
|
}
|
||||||
|
metadata[k] = typed
|
||||||
|
}
|
||||||
|
|
||||||
|
return metadata, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// TokenIsRenewable returns the standardized token renewability for the given
|
||||||
|
// secret. If the secret is nil or does not contain the "renewable" key, this
|
||||||
|
// returns false.
|
||||||
|
func (s *Secret) TokenIsRenewable() (bool, error) {
|
||||||
|
if s == nil {
|
||||||
|
return false, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if s.Auth != nil && s.Auth.Renewable {
|
||||||
|
return s.Auth.Renewable, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if s.Data == nil || s.Data["renewable"] == nil {
|
||||||
|
return false, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
renewable, err := parseutil.ParseBool(s.Data["renewable"])
|
||||||
|
if err != nil {
|
||||||
|
return false, errwrap.Wrapf("could not convert renewable value to a boolean: {{err}}", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return renewable, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// TokenTTL returns the standardized remaining token TTL for the given secret.
|
||||||
|
// If the secret is nil or does not contain a TTL, this returns 0.
|
||||||
|
func (s *Secret) TokenTTL() (time.Duration, error) {
|
||||||
|
if s == nil {
|
||||||
|
return 0, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if s.Auth != nil && s.Auth.LeaseDuration > 0 {
|
||||||
|
return time.Duration(s.Auth.LeaseDuration) * time.Second, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if s.Data == nil || s.Data["ttl"] == nil {
|
||||||
|
return 0, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
ttl, err := parseutil.ParseDurationSecond(s.Data["ttl"])
|
||||||
|
if err != nil {
|
||||||
|
return 0, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return ttl, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// SecretWrapInfo contains wrapping information if we have it. If what is
|
||||||
|
// contained is an authentication token, the accessor for the token will be
|
||||||
|
// available in WrappedAccessor.
|
||||||
|
type SecretWrapInfo struct {
|
||||||
|
Token string `json:"token"`
|
||||||
|
Accessor string `json:"accessor"`
|
||||||
|
TTL int `json:"ttl"`
|
||||||
|
CreationTime time.Time `json:"creation_time"`
|
||||||
|
CreationPath string `json:"creation_path"`
|
||||||
|
WrappedAccessor string `json:"wrapped_accessor"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// SecretAuth is the structure containing auth information if we have it.
|
||||||
|
type SecretAuth struct {
|
||||||
|
ClientToken string `json:"client_token"`
|
||||||
|
Accessor string `json:"accessor"`
|
||||||
|
Policies []string `json:"policies"`
|
||||||
|
TokenPolicies []string `json:"token_policies"`
|
||||||
|
IdentityPolicies []string `json:"identity_policies"`
|
||||||
|
Metadata map[string]string `json:"metadata"`
|
||||||
|
|
||||||
|
LeaseDuration int `json:"lease_duration"`
|
||||||
|
Renewable bool `json:"renewable"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// ParseSecret is used to parse a secret value from JSON from an io.Reader.
|
||||||
|
func ParseSecret(r io.Reader) (*Secret, error) {
|
||||||
|
// First read the data into a buffer. Not super efficient but we want to
|
||||||
|
// know if we actually have a body or not.
|
||||||
|
var buf bytes.Buffer
|
||||||
|
_, err := buf.ReadFrom(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if buf.Len() == 0 {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// First decode the JSON into a map[string]interface{}
|
||||||
|
var secret Secret
|
||||||
|
if err := jsonutil.DecodeJSONFromReader(&buf, &secret); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return &secret, nil
|
||||||
|
}
|
|
@ -0,0 +1,55 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import "fmt"
|
||||||
|
|
||||||
|
// SSH is used to return a client to invoke operations on SSH backend.
|
||||||
|
type SSH struct {
|
||||||
|
c *Client
|
||||||
|
MountPoint string
|
||||||
|
}
|
||||||
|
|
||||||
|
// SSH returns the client for logical-backend API calls.
|
||||||
|
func (c *Client) SSH() *SSH {
|
||||||
|
return c.SSHWithMountPoint(SSHHelperDefaultMountPoint)
|
||||||
|
}
|
||||||
|
|
||||||
|
// SSHWithMountPoint returns the client with specific SSH mount point.
|
||||||
|
func (c *Client) SSHWithMountPoint(mountPoint string) *SSH {
|
||||||
|
return &SSH{
|
||||||
|
c: c,
|
||||||
|
MountPoint: mountPoint,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Credential invokes the SSH backend API to create a credential to establish an SSH session.
|
||||||
|
func (c *SSH) Credential(role string, data map[string]interface{}) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("PUT", fmt.Sprintf("/v1/%s/creds/%s", c.MountPoint, role))
|
||||||
|
if err := r.SetJSONBody(data); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
// SignKey signs the given public key and returns a signed public key to pass
|
||||||
|
// along with the SSH request.
|
||||||
|
func (c *SSH) SignKey(role string, data map[string]interface{}) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("PUT", fmt.Sprintf("/v1/%s/sign/%s", c.MountPoint, role))
|
||||||
|
if err := r.SetJSONBody(data); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
|
@ -0,0 +1,231 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import (
|
||||||
|
"crypto/tls"
|
||||||
|
"crypto/x509"
|
||||||
|
"fmt"
|
||||||
|
"io/ioutil"
|
||||||
|
"os"
|
||||||
|
|
||||||
|
"github.com/hashicorp/errwrap"
|
||||||
|
"github.com/hashicorp/go-cleanhttp"
|
||||||
|
"github.com/hashicorp/go-multierror"
|
||||||
|
"github.com/hashicorp/go-rootcerts"
|
||||||
|
"github.com/hashicorp/hcl"
|
||||||
|
"github.com/hashicorp/hcl/hcl/ast"
|
||||||
|
"github.com/hashicorp/vault/helper/hclutil"
|
||||||
|
"github.com/mitchellh/mapstructure"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// SSHHelperDefaultMountPoint is the default path at which SSH backend will be
|
||||||
|
// mounted in the Vault server.
|
||||||
|
SSHHelperDefaultMountPoint = "ssh"
|
||||||
|
|
||||||
|
// VerifyEchoRequest is the echo request message sent as OTP by the helper.
|
||||||
|
VerifyEchoRequest = "verify-echo-request"
|
||||||
|
|
||||||
|
// VerifyEchoResponse is the echo response message sent as a response to OTP
|
||||||
|
// matching echo request.
|
||||||
|
VerifyEchoResponse = "verify-echo-response"
|
||||||
|
)
|
||||||
|
|
||||||
|
// SSHHelper is a structure representing a vault-ssh-helper which can talk to vault server
|
||||||
|
// in order to verify the OTP entered by the user. It contains the path at which
|
||||||
|
// SSH backend is mounted at the server.
|
||||||
|
type SSHHelper struct {
|
||||||
|
c *Client
|
||||||
|
MountPoint string
|
||||||
|
}
|
||||||
|
|
||||||
|
// SSHVerifyResponse is a structure representing the fields in Vault server's
|
||||||
|
// response.
|
||||||
|
type SSHVerifyResponse struct {
|
||||||
|
// Usually empty. If the request OTP is echo request message, this will
|
||||||
|
// be set to the corresponding echo response message.
|
||||||
|
Message string `json:"message" mapstructure:"message"`
|
||||||
|
|
||||||
|
// Username associated with the OTP
|
||||||
|
Username string `json:"username" mapstructure:"username"`
|
||||||
|
|
||||||
|
// IP associated with the OTP
|
||||||
|
IP string `json:"ip" mapstructure:"ip"`
|
||||||
|
|
||||||
|
// Name of the role against which the OTP was issued
|
||||||
|
RoleName string `json:"role_name" mapstructure:"role_name"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// SSHHelperConfig is a structure which represents the entries from the vault-ssh-helper's configuration file.
|
||||||
|
type SSHHelperConfig struct {
|
||||||
|
VaultAddr string `hcl:"vault_addr"`
|
||||||
|
SSHMountPoint string `hcl:"ssh_mount_point"`
|
||||||
|
CACert string `hcl:"ca_cert"`
|
||||||
|
CAPath string `hcl:"ca_path"`
|
||||||
|
AllowedCidrList string `hcl:"allowed_cidr_list"`
|
||||||
|
AllowedRoles string `hcl:"allowed_roles"`
|
||||||
|
TLSSkipVerify bool `hcl:"tls_skip_verify"`
|
||||||
|
TLSServerName string `hcl:"tls_server_name"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetTLSParameters sets the TLS parameters for this SSH agent.
|
||||||
|
func (c *SSHHelperConfig) SetTLSParameters(clientConfig *Config, certPool *x509.CertPool) {
|
||||||
|
tlsConfig := &tls.Config{
|
||||||
|
InsecureSkipVerify: c.TLSSkipVerify,
|
||||||
|
MinVersion: tls.VersionTLS12,
|
||||||
|
RootCAs: certPool,
|
||||||
|
ServerName: c.TLSServerName,
|
||||||
|
}
|
||||||
|
|
||||||
|
transport := cleanhttp.DefaultTransport()
|
||||||
|
transport.TLSClientConfig = tlsConfig
|
||||||
|
clientConfig.HttpClient.Transport = transport
|
||||||
|
}
|
||||||
|
|
||||||
|
// Returns true if any of the following conditions are true:
|
||||||
|
// * CA cert is configured
|
||||||
|
// * CA path is configured
|
||||||
|
// * configured to skip certificate verification
|
||||||
|
// * TLS server name is configured
|
||||||
|
//
|
||||||
|
func (c *SSHHelperConfig) shouldSetTLSParameters() bool {
|
||||||
|
return c.CACert != "" || c.CAPath != "" || c.TLSServerName != "" || c.TLSSkipVerify
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewClient returns a new client for the configuration. This client will be used by the
|
||||||
|
// vault-ssh-helper to communicate with Vault server and verify the OTP entered by user.
|
||||||
|
// If the configuration supplies Vault SSL certificates, then the client will
|
||||||
|
// have TLS configured in its transport.
|
||||||
|
func (c *SSHHelperConfig) NewClient() (*Client, error) {
|
||||||
|
// Creating a default client configuration for communicating with vault server.
|
||||||
|
clientConfig := DefaultConfig()
|
||||||
|
|
||||||
|
// Pointing the client to the actual address of vault server.
|
||||||
|
clientConfig.Address = c.VaultAddr
|
||||||
|
|
||||||
|
// Check if certificates are provided via config file.
|
||||||
|
if c.shouldSetTLSParameters() {
|
||||||
|
rootConfig := &rootcerts.Config{
|
||||||
|
CAFile: c.CACert,
|
||||||
|
CAPath: c.CAPath,
|
||||||
|
}
|
||||||
|
certPool, err := rootcerts.LoadCACerts(rootConfig)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
// Enable TLS on the HTTP client information
|
||||||
|
c.SetTLSParameters(clientConfig, certPool)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Creating the client object for the given configuration
|
||||||
|
client, err := NewClient(clientConfig)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return client, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// LoadSSHHelperConfig loads ssh-helper's configuration from the file and populates the corresponding
|
||||||
|
// in-memory structure.
|
||||||
|
//
|
||||||
|
// Vault address is a required parameter.
|
||||||
|
// Mount point defaults to "ssh".
|
||||||
|
func LoadSSHHelperConfig(path string) (*SSHHelperConfig, error) {
|
||||||
|
contents, err := ioutil.ReadFile(path)
|
||||||
|
if err != nil && !os.IsNotExist(err) {
|
||||||
|
return nil, multierror.Prefix(err, "ssh_helper:")
|
||||||
|
}
|
||||||
|
return ParseSSHHelperConfig(string(contents))
|
||||||
|
}
|
||||||
|
|
||||||
|
// ParseSSHHelperConfig parses the given contents as a string for the SSHHelper
|
||||||
|
// configuration.
|
||||||
|
func ParseSSHHelperConfig(contents string) (*SSHHelperConfig, error) {
|
||||||
|
root, err := hcl.Parse(string(contents))
|
||||||
|
if err != nil {
|
||||||
|
return nil, errwrap.Wrapf("error parsing config: {{err}}", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
list, ok := root.Node.(*ast.ObjectList)
|
||||||
|
if !ok {
|
||||||
|
return nil, fmt.Errorf("error parsing config: file doesn't contain a root object")
|
||||||
|
}
|
||||||
|
|
||||||
|
valid := []string{
|
||||||
|
"vault_addr",
|
||||||
|
"ssh_mount_point",
|
||||||
|
"ca_cert",
|
||||||
|
"ca_path",
|
||||||
|
"allowed_cidr_list",
|
||||||
|
"allowed_roles",
|
||||||
|
"tls_skip_verify",
|
||||||
|
"tls_server_name",
|
||||||
|
}
|
||||||
|
if err := hclutil.CheckHCLKeys(list, valid); err != nil {
|
||||||
|
return nil, multierror.Prefix(err, "ssh_helper:")
|
||||||
|
}
|
||||||
|
|
||||||
|
var c SSHHelperConfig
|
||||||
|
c.SSHMountPoint = SSHHelperDefaultMountPoint
|
||||||
|
if err := hcl.DecodeObject(&c, list); err != nil {
|
||||||
|
return nil, multierror.Prefix(err, "ssh_helper:")
|
||||||
|
}
|
||||||
|
|
||||||
|
if c.VaultAddr == "" {
|
||||||
|
return nil, fmt.Errorf(`missing config "vault_addr"`)
|
||||||
|
}
|
||||||
|
return &c, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// SSHHelper creates an SSHHelper object which can talk to Vault server with SSH backend
|
||||||
|
// mounted at default path ("ssh").
|
||||||
|
func (c *Client) SSHHelper() *SSHHelper {
|
||||||
|
return c.SSHHelperWithMountPoint(SSHHelperDefaultMountPoint)
|
||||||
|
}
|
||||||
|
|
||||||
|
// SSHHelperWithMountPoint creates an SSHHelper object which can talk to Vault server with SSH backend
|
||||||
|
// mounted at a specific mount point.
|
||||||
|
func (c *Client) SSHHelperWithMountPoint(mountPoint string) *SSHHelper {
|
||||||
|
return &SSHHelper{
|
||||||
|
c: c,
|
||||||
|
MountPoint: mountPoint,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Verify verifies if the key provided by user is present in Vault server. The response
|
||||||
|
// will contain the IP address and username associated with the OTP. In case the
|
||||||
|
// OTP matches the echo request message, instead of searching an entry for the OTP,
|
||||||
|
// an echo response message is returned. This feature is used by ssh-helper to verify if
|
||||||
|
// its configured correctly.
|
||||||
|
func (c *SSHHelper) Verify(otp string) (*SSHVerifyResponse, error) {
|
||||||
|
data := map[string]interface{}{
|
||||||
|
"otp": otp,
|
||||||
|
}
|
||||||
|
verifyPath := fmt.Sprintf("/v1/%s/verify", c.MountPoint)
|
||||||
|
r := c.c.NewRequest("PUT", verifyPath)
|
||||||
|
if err := r.SetJSONBody(data); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
secret, err := ParseSecret(resp.Body)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if secret.Data == nil {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
var verifyResp SSHVerifyResponse
|
||||||
|
err = mapstructure.Decode(secret.Data, &verifyResp)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return &verifyResp, nil
|
||||||
|
}
|
|
@ -0,0 +1,11 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
// Sys is used to perform system-related operations on Vault.
|
||||||
|
type Sys struct {
|
||||||
|
c *Client
|
||||||
|
}
|
||||||
|
|
||||||
|
// Sys is used to return the client for sys-related API calls.
|
||||||
|
func (c *Client) Sys() *Sys {
|
||||||
|
return &Sys{c: c}
|
||||||
|
}
|
|
@ -0,0 +1,125 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/mitchellh/mapstructure"
|
||||||
|
)
|
||||||
|
|
||||||
|
func (c *Sys) AuditHash(path string, input string) (string, error) {
|
||||||
|
body := map[string]interface{}{
|
||||||
|
"input": input,
|
||||||
|
}
|
||||||
|
|
||||||
|
r := c.c.NewRequest("PUT", fmt.Sprintf("/v1/sys/audit-hash/%s", path))
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
type d struct {
|
||||||
|
Hash string `json:"hash"`
|
||||||
|
}
|
||||||
|
|
||||||
|
var result d
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
return result.Hash, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) ListAudit() (map[string]*Audit, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/audit")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result map[string]interface{}
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
mounts := map[string]*Audit{}
|
||||||
|
for k, v := range result {
|
||||||
|
switch v.(type) {
|
||||||
|
case map[string]interface{}:
|
||||||
|
default:
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
var res Audit
|
||||||
|
err = mapstructure.Decode(v, &res)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
// Not a mount, some other api.Secret data
|
||||||
|
if res.Type == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
mounts[k] = &res
|
||||||
|
}
|
||||||
|
|
||||||
|
return mounts, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// DEPRECATED: Use EnableAuditWithOptions instead
|
||||||
|
func (c *Sys) EnableAudit(
|
||||||
|
path string, auditType string, desc string, opts map[string]string) error {
|
||||||
|
return c.EnableAuditWithOptions(path, &EnableAuditOptions{
|
||||||
|
Type: auditType,
|
||||||
|
Description: desc,
|
||||||
|
Options: opts,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) EnableAuditWithOptions(path string, options *EnableAuditOptions) error {
|
||||||
|
r := c.c.NewRequest("PUT", fmt.Sprintf("/v1/sys/audit/%s", path))
|
||||||
|
if err := r.SetJSONBody(options); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) DisableAudit(path string) error {
|
||||||
|
r := c.c.NewRequest("DELETE", fmt.Sprintf("/v1/sys/audit/%s", path))
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Structures for the requests/resposne are all down here. They aren't
|
||||||
|
// individually documented because the map almost directly to the raw HTTP API
|
||||||
|
// documentation. Please refer to that documentation for more details.
|
||||||
|
|
||||||
|
type EnableAuditOptions struct {
|
||||||
|
Type string `json:"type"`
|
||||||
|
Description string `json:"description"`
|
||||||
|
Options map[string]string `json:"options"`
|
||||||
|
Local bool `json:"local"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type Audit struct {
|
||||||
|
Path string
|
||||||
|
Type string
|
||||||
|
Description string
|
||||||
|
Options map[string]string
|
||||||
|
Local bool
|
||||||
|
}
|
|
@ -0,0 +1,119 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/mitchellh/mapstructure"
|
||||||
|
)
|
||||||
|
|
||||||
|
func (c *Sys) ListAuth() (map[string]*AuthMount, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/auth")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result map[string]interface{}
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
mounts := map[string]*AuthMount{}
|
||||||
|
for k, v := range result {
|
||||||
|
switch v.(type) {
|
||||||
|
case map[string]interface{}:
|
||||||
|
default:
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
var res AuthMount
|
||||||
|
err = mapstructure.Decode(v, &res)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
// Not a mount, some other api.Secret data
|
||||||
|
if res.Type == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
mounts[k] = &res
|
||||||
|
}
|
||||||
|
|
||||||
|
return mounts, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// DEPRECATED: Use EnableAuthWithOptions instead
|
||||||
|
func (c *Sys) EnableAuth(path, authType, desc string) error {
|
||||||
|
return c.EnableAuthWithOptions(path, &EnableAuthOptions{
|
||||||
|
Type: authType,
|
||||||
|
Description: desc,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) EnableAuthWithOptions(path string, options *EnableAuthOptions) error {
|
||||||
|
r := c.c.NewRequest("POST", fmt.Sprintf("/v1/sys/auth/%s", path))
|
||||||
|
if err := r.SetJSONBody(options); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) DisableAuth(path string) error {
|
||||||
|
r := c.c.NewRequest("DELETE", fmt.Sprintf("/v1/sys/auth/%s", path))
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Structures for the requests/resposne are all down here. They aren't
|
||||||
|
// individually documented because the map almost directly to the raw HTTP API
|
||||||
|
// documentation. Please refer to that documentation for more details.
|
||||||
|
|
||||||
|
type EnableAuthOptions struct {
|
||||||
|
Type string `json:"type"`
|
||||||
|
Description string `json:"description"`
|
||||||
|
Config AuthConfigInput `json:"config"`
|
||||||
|
Local bool `json:"local"`
|
||||||
|
PluginName string `json:"plugin_name,omitempty"`
|
||||||
|
SealWrap bool `json:"seal_wrap" mapstructure:"seal_wrap"`
|
||||||
|
Options map[string]string `json:"options" mapstructure:"options"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type AuthConfigInput struct {
|
||||||
|
DefaultLeaseTTL string `json:"default_lease_ttl" mapstructure:"default_lease_ttl"`
|
||||||
|
MaxLeaseTTL string `json:"max_lease_ttl" mapstructure:"max_lease_ttl"`
|
||||||
|
PluginName string `json:"plugin_name,omitempty" mapstructure:"plugin_name"`
|
||||||
|
AuditNonHMACRequestKeys []string `json:"audit_non_hmac_request_keys,omitempty" mapstructure:"audit_non_hmac_request_keys"`
|
||||||
|
AuditNonHMACResponseKeys []string `json:"audit_non_hmac_response_keys,omitempty" mapstructure:"audit_non_hmac_response_keys"`
|
||||||
|
ListingVisibility string `json:"listing_visibility,omitempty" mapstructure:"listing_visibility"`
|
||||||
|
PassthroughRequestHeaders []string `json:"passthrough_request_headers,omitempty" mapstructure:"passthrough_request_headers"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type AuthMount struct {
|
||||||
|
Type string `json:"type" mapstructure:"type"`
|
||||||
|
Description string `json:"description" mapstructure:"description"`
|
||||||
|
Accessor string `json:"accessor" mapstructure:"accessor"`
|
||||||
|
Config AuthConfigOutput `json:"config" mapstructure:"config"`
|
||||||
|
Local bool `json:"local" mapstructure:"local"`
|
||||||
|
SealWrap bool `json:"seal_wrap" mapstructure:"seal_wrap"`
|
||||||
|
Options map[string]string `json:"options" mapstructure:"options"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type AuthConfigOutput struct {
|
||||||
|
DefaultLeaseTTL int `json:"default_lease_ttl" mapstructure:"default_lease_ttl"`
|
||||||
|
MaxLeaseTTL int `json:"max_lease_ttl" mapstructure:"max_lease_ttl"`
|
||||||
|
PluginName string `json:"plugin_name,omitempty" mapstructure:"plugin_name"`
|
||||||
|
AuditNonHMACRequestKeys []string `json:"audit_non_hmac_request_keys,omitempty" mapstructure:"audit_non_hmac_request_keys"`
|
||||||
|
AuditNonHMACResponseKeys []string `json:"audit_non_hmac_response_keys,omitempty" mapstructure:"audit_non_hmac_response_keys"`
|
||||||
|
ListingVisibility string `json:"listing_visibility,omitempty" mapstructure:"listing_visibility"`
|
||||||
|
PassthroughRequestHeaders []string `json:"passthrough_request_headers,omitempty" mapstructure:"passthrough_request_headers"`
|
||||||
|
}
|
|
@ -0,0 +1,49 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import "fmt"
|
||||||
|
|
||||||
|
func (c *Sys) CapabilitiesSelf(path string) ([]string, error) {
|
||||||
|
return c.Capabilities(c.c.Token(), path)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) Capabilities(token, path string) ([]string, error) {
|
||||||
|
body := map[string]string{
|
||||||
|
"token": token,
|
||||||
|
"path": path,
|
||||||
|
}
|
||||||
|
|
||||||
|
reqPath := "/v1/sys/capabilities"
|
||||||
|
if token == c.c.Token() {
|
||||||
|
reqPath = fmt.Sprintf("%s-self", reqPath)
|
||||||
|
}
|
||||||
|
|
||||||
|
r := c.c.NewRequest("POST", reqPath)
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result map[string]interface{}
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if result["capabilities"] == nil {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
var capabilities []string
|
||||||
|
capabilitiesRaw, ok := result["capabilities"].([]interface{})
|
||||||
|
if !ok {
|
||||||
|
return nil, fmt.Errorf("error interpreting returned capabilities")
|
||||||
|
}
|
||||||
|
for _, capability := range capabilitiesRaw {
|
||||||
|
capabilities = append(capabilities, capability.(string))
|
||||||
|
}
|
||||||
|
return capabilities, nil
|
||||||
|
}
|
|
@ -0,0 +1,56 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
func (c *Sys) CORSStatus() (*CORSResponse, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/config/cors")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result CORSResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) ConfigureCORS(req *CORSRequest) (*CORSResponse, error) {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/config/cors")
|
||||||
|
if err := r.SetJSONBody(req); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result CORSResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) DisableCORS() (*CORSResponse, error) {
|
||||||
|
r := c.c.NewRequest("DELETE", "/v1/sys/config/cors")
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result CORSResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
type CORSRequest struct {
|
||||||
|
AllowedOrigins string `json:"allowed_origins"`
|
||||||
|
Enabled bool `json:"enabled"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type CORSResponse struct {
|
||||||
|
AllowedOrigins string `json:"allowed_origins"`
|
||||||
|
Enabled bool `json:"enabled"`
|
||||||
|
}
|
|
@ -0,0 +1,110 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
func (c *Sys) GenerateRootStatus() (*GenerateRootStatusResponse, error) {
|
||||||
|
return c.generateRootStatusCommon("/v1/sys/generate-root/attempt")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) GenerateDROperationTokenStatus() (*GenerateRootStatusResponse, error) {
|
||||||
|
return c.generateRootStatusCommon("/v1/sys/replication/dr/secondary/generate-operation-token/attempt")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) generateRootStatusCommon(path string) (*GenerateRootStatusResponse, error) {
|
||||||
|
r := c.c.NewRequest("GET", path)
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result GenerateRootStatusResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) GenerateRootInit(otp, pgpKey string) (*GenerateRootStatusResponse, error) {
|
||||||
|
return c.generateRootInitCommon("/v1/sys/generate-root/attempt", otp, pgpKey)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) GenerateDROperationTokenInit(otp, pgpKey string) (*GenerateRootStatusResponse, error) {
|
||||||
|
return c.generateRootInitCommon("/v1/sys/replication/dr/secondary/generate-operation-token/attempt", otp, pgpKey)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) generateRootInitCommon(path, otp, pgpKey string) (*GenerateRootStatusResponse, error) {
|
||||||
|
body := map[string]interface{}{
|
||||||
|
"otp": otp,
|
||||||
|
"pgp_key": pgpKey,
|
||||||
|
}
|
||||||
|
|
||||||
|
r := c.c.NewRequest("PUT", path)
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result GenerateRootStatusResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) GenerateRootCancel() error {
|
||||||
|
return c.generateRootCancelCommon("/v1/sys/generate-root/attempt")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) GenerateDROperationTokenCancel() error {
|
||||||
|
return c.generateRootCancelCommon("/v1/sys/replication/dr/secondary/generate-operation-token/attempt")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) generateRootCancelCommon(path string) error {
|
||||||
|
r := c.c.NewRequest("DELETE", path)
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) GenerateRootUpdate(shard, nonce string) (*GenerateRootStatusResponse, error) {
|
||||||
|
return c.generateRootUpdateCommon("/v1/sys/generate-root/update", shard, nonce)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) GenerateDROperationTokenUpdate(shard, nonce string) (*GenerateRootStatusResponse, error) {
|
||||||
|
return c.generateRootUpdateCommon("/v1/sys/replication/dr/secondary/generate-operation-token/update", shard, nonce)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) generateRootUpdateCommon(path, shard, nonce string) (*GenerateRootStatusResponse, error) {
|
||||||
|
body := map[string]interface{}{
|
||||||
|
"key": shard,
|
||||||
|
"nonce": nonce,
|
||||||
|
}
|
||||||
|
|
||||||
|
r := c.c.NewRequest("PUT", path)
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result GenerateRootStatusResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
type GenerateRootStatusResponse struct {
|
||||||
|
Nonce string `json:"nonce"`
|
||||||
|
Started bool `json:"started"`
|
||||||
|
Progress int `json:"progress"`
|
||||||
|
Required int `json:"required"`
|
||||||
|
Complete bool `json:"complete"`
|
||||||
|
EncodedToken string `json:"encoded_token"`
|
||||||
|
EncodedRootToken string `json:"encoded_root_token"`
|
||||||
|
PGPFingerprint string `json:"pgp_fingerprint"`
|
||||||
|
}
|
|
@ -0,0 +1,33 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
func (c *Sys) Health() (*HealthResponse, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/health")
|
||||||
|
// If the code is 400 or above it will automatically turn into an error,
|
||||||
|
// but the sys/health API defaults to returning 5xx when not sealed or
|
||||||
|
// inited, so we force this code to be something else so we parse correctly
|
||||||
|
r.Params.Add("uninitcode", "299")
|
||||||
|
r.Params.Add("sealedcode", "299")
|
||||||
|
r.Params.Add("standbycode", "299")
|
||||||
|
r.Params.Add("drsecondarycode", "299")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result HealthResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
type HealthResponse struct {
|
||||||
|
Initialized bool `json:"initialized"`
|
||||||
|
Sealed bool `json:"sealed"`
|
||||||
|
Standby bool `json:"standby"`
|
||||||
|
ReplicationPerformanceMode string `json:"replication_performance_mode"`
|
||||||
|
ReplicationDRMode string `json:"replication_dr_mode"`
|
||||||
|
ServerTimeUTC int64 `json:"server_time_utc"`
|
||||||
|
Version string `json:"version"`
|
||||||
|
ClusterName string `json:"cluster_name,omitempty"`
|
||||||
|
ClusterID string `json:"cluster_id,omitempty"`
|
||||||
|
}
|
|
@ -0,0 +1,54 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
func (c *Sys) InitStatus() (bool, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/init")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result InitStatusResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return result.Initialized, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) Init(opts *InitRequest) (*InitResponse, error) {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/init")
|
||||||
|
if err := r.SetJSONBody(opts); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result InitResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
type InitRequest struct {
|
||||||
|
SecretShares int `json:"secret_shares"`
|
||||||
|
SecretThreshold int `json:"secret_threshold"`
|
||||||
|
StoredShares int `json:"stored_shares"`
|
||||||
|
PGPKeys []string `json:"pgp_keys"`
|
||||||
|
RecoveryShares int `json:"recovery_shares"`
|
||||||
|
RecoveryThreshold int `json:"recovery_threshold"`
|
||||||
|
RecoveryPGPKeys []string `json:"recovery_pgp_keys"`
|
||||||
|
RootTokenPGPKey string `json:"root_token_pgp_key"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type InitStatusResponse struct {
|
||||||
|
Initialized bool
|
||||||
|
}
|
||||||
|
|
||||||
|
type InitResponse struct {
|
||||||
|
Keys []string `json:"keys"`
|
||||||
|
KeysB64 []string `json:"keys_base64"`
|
||||||
|
RecoveryKeys []string `json:"recovery_keys"`
|
||||||
|
RecoveryKeysB64 []string `json:"recovery_keys_base64"`
|
||||||
|
RootToken string `json:"root_token"`
|
||||||
|
}
|
|
@ -0,0 +1,21 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
func (c *Sys) Leader() (*LeaderResponse, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/leader")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result LeaderResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
type LeaderResponse struct {
|
||||||
|
HAEnabled bool `json:"ha_enabled"`
|
||||||
|
IsSelf bool `json:"is_self"`
|
||||||
|
LeaderAddress string `json:"leader_address"`
|
||||||
|
LeaderClusterAddress string `json:"leader_cluster_address"`
|
||||||
|
}
|
|
@ -0,0 +1,89 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import "errors"
|
||||||
|
|
||||||
|
func (c *Sys) Renew(id string, increment int) (*Secret, error) {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/leases/renew")
|
||||||
|
|
||||||
|
body := map[string]interface{}{
|
||||||
|
"increment": increment,
|
||||||
|
"lease_id": id,
|
||||||
|
}
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return ParseSecret(resp.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) Revoke(id string) error {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/leases/revoke/"+id)
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RevokePrefix(id string) error {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/leases/revoke-prefix/"+id)
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RevokeForce(id string) error {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/leases/revoke-force/"+id)
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RevokeWithOptions(opts *RevokeOptions) error {
|
||||||
|
if opts == nil {
|
||||||
|
return errors.New("nil options provided")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Construct path
|
||||||
|
path := "/v1/sys/leases/revoke/"
|
||||||
|
switch {
|
||||||
|
case opts.Force:
|
||||||
|
path = "/v1/sys/leases/revoke-force/"
|
||||||
|
case opts.Prefix:
|
||||||
|
path = "/v1/sys/leases/revoke-prefix/"
|
||||||
|
}
|
||||||
|
path += opts.LeaseID
|
||||||
|
|
||||||
|
r := c.c.NewRequest("PUT", path)
|
||||||
|
if !opts.Force {
|
||||||
|
body := map[string]interface{}{
|
||||||
|
"sync": opts.Sync,
|
||||||
|
}
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
type RevokeOptions struct {
|
||||||
|
LeaseID string
|
||||||
|
Force bool
|
||||||
|
Prefix bool
|
||||||
|
Sync bool
|
||||||
|
}
|
|
@ -0,0 +1,160 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/mitchellh/mapstructure"
|
||||||
|
)
|
||||||
|
|
||||||
|
func (c *Sys) ListMounts() (map[string]*MountOutput, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/mounts")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result map[string]interface{}
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
mounts := map[string]*MountOutput{}
|
||||||
|
for k, v := range result {
|
||||||
|
switch v.(type) {
|
||||||
|
case map[string]interface{}:
|
||||||
|
default:
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
var res MountOutput
|
||||||
|
err = mapstructure.Decode(v, &res)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
// Not a mount, some other api.Secret data
|
||||||
|
if res.Type == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
mounts[k] = &res
|
||||||
|
}
|
||||||
|
|
||||||
|
return mounts, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) Mount(path string, mountInfo *MountInput) error {
|
||||||
|
r := c.c.NewRequest("POST", fmt.Sprintf("/v1/sys/mounts/%s", path))
|
||||||
|
if err := r.SetJSONBody(mountInfo); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) Unmount(path string) error {
|
||||||
|
r := c.c.NewRequest("DELETE", fmt.Sprintf("/v1/sys/mounts/%s", path))
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) Remount(from, to string) error {
|
||||||
|
body := map[string]interface{}{
|
||||||
|
"from": from,
|
||||||
|
"to": to,
|
||||||
|
}
|
||||||
|
|
||||||
|
r := c.c.NewRequest("POST", "/v1/sys/remount")
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) TuneMount(path string, config MountConfigInput) error {
|
||||||
|
r := c.c.NewRequest("POST", fmt.Sprintf("/v1/sys/mounts/%s/tune", path))
|
||||||
|
if err := r.SetJSONBody(config); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) MountConfig(path string) (*MountConfigOutput, error) {
|
||||||
|
r := c.c.NewRequest("GET", fmt.Sprintf("/v1/sys/mounts/%s/tune", path))
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result MountConfigOutput
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
type MountInput struct {
|
||||||
|
Type string `json:"type"`
|
||||||
|
Description string `json:"description"`
|
||||||
|
Config MountConfigInput `json:"config"`
|
||||||
|
Options map[string]string `json:"options"`
|
||||||
|
Local bool `json:"local"`
|
||||||
|
PluginName string `json:"plugin_name,omitempty"`
|
||||||
|
SealWrap bool `json:"seal_wrap" mapstructure:"seal_wrap"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type MountConfigInput struct {
|
||||||
|
Options map[string]string `json:"options" mapstructure:"options"`
|
||||||
|
DefaultLeaseTTL string `json:"default_lease_ttl" mapstructure:"default_lease_ttl"`
|
||||||
|
Description *string `json:"description,omitempty" mapstructure:"description"`
|
||||||
|
MaxLeaseTTL string `json:"max_lease_ttl" mapstructure:"max_lease_ttl"`
|
||||||
|
ForceNoCache bool `json:"force_no_cache" mapstructure:"force_no_cache"`
|
||||||
|
PluginName string `json:"plugin_name,omitempty" mapstructure:"plugin_name"`
|
||||||
|
AuditNonHMACRequestKeys []string `json:"audit_non_hmac_request_keys,omitempty" mapstructure:"audit_non_hmac_request_keys"`
|
||||||
|
AuditNonHMACResponseKeys []string `json:"audit_non_hmac_response_keys,omitempty" mapstructure:"audit_non_hmac_response_keys"`
|
||||||
|
ListingVisibility string `json:"listing_visibility,omitempty" mapstructure:"listing_visibility"`
|
||||||
|
PassthroughRequestHeaders []string `json:"passthrough_request_headers,omitempty" mapstructure:"passthrough_request_headers"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type MountOutput struct {
|
||||||
|
Type string `json:"type"`
|
||||||
|
Description string `json:"description"`
|
||||||
|
Accessor string `json:"accessor"`
|
||||||
|
Config MountConfigOutput `json:"config"`
|
||||||
|
Options map[string]string `json:"options"`
|
||||||
|
Local bool `json:"local"`
|
||||||
|
SealWrap bool `json:"seal_wrap" mapstructure:"seal_wrap"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type MountConfigOutput struct {
|
||||||
|
DefaultLeaseTTL int `json:"default_lease_ttl" mapstructure:"default_lease_ttl"`
|
||||||
|
MaxLeaseTTL int `json:"max_lease_ttl" mapstructure:"max_lease_ttl"`
|
||||||
|
ForceNoCache bool `json:"force_no_cache" mapstructure:"force_no_cache"`
|
||||||
|
PluginName string `json:"plugin_name,omitempty" mapstructure:"plugin_name"`
|
||||||
|
AuditNonHMACRequestKeys []string `json:"audit_non_hmac_request_keys,omitempty" mapstructure:"audit_non_hmac_request_keys"`
|
||||||
|
AuditNonHMACResponseKeys []string `json:"audit_non_hmac_response_keys,omitempty" mapstructure:"audit_non_hmac_response_keys"`
|
||||||
|
ListingVisibility string `json:"listing_visibility,omitempty" mapstructure:"listing_visibility"`
|
||||||
|
PassthroughRequestHeaders []string `json:"passthrough_request_headers,omitempty" mapstructure:"passthrough_request_headers"`
|
||||||
|
}
|
|
@ -0,0 +1,119 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"net/http"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ListPluginsInput is used as input to the ListPlugins function.
|
||||||
|
type ListPluginsInput struct{}
|
||||||
|
|
||||||
|
// ListPluginsResponse is the response from the ListPlugins call.
|
||||||
|
type ListPluginsResponse struct {
|
||||||
|
// Names is the list of names of the plugins.
|
||||||
|
Names []string
|
||||||
|
}
|
||||||
|
|
||||||
|
// ListPlugins lists all plugins in the catalog and returns their names as a
|
||||||
|
// list of strings.
|
||||||
|
func (c *Sys) ListPlugins(i *ListPluginsInput) (*ListPluginsResponse, error) {
|
||||||
|
path := "/v1/sys/plugins/catalog"
|
||||||
|
req := c.c.NewRequest("LIST", path)
|
||||||
|
resp, err := c.c.RawRequest(req)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result struct {
|
||||||
|
Data struct {
|
||||||
|
Keys []string `json:"keys"`
|
||||||
|
} `json:"data"`
|
||||||
|
}
|
||||||
|
if err := resp.DecodeJSON(&result); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return &ListPluginsResponse{Names: result.Data.Keys}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetPluginInput is used as input to the GetPlugin function.
|
||||||
|
type GetPluginInput struct {
|
||||||
|
Name string `json:"-"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetPluginResponse is the response from the GetPlugin call.
|
||||||
|
type GetPluginResponse struct {
|
||||||
|
Args []string `json:"args"`
|
||||||
|
Builtin bool `json:"builtin"`
|
||||||
|
Command string `json:"command"`
|
||||||
|
Name string `json:"name"`
|
||||||
|
SHA256 string `json:"sha256"`
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) GetPlugin(i *GetPluginInput) (*GetPluginResponse, error) {
|
||||||
|
path := fmt.Sprintf("/v1/sys/plugins/catalog/%s", i.Name)
|
||||||
|
req := c.c.NewRequest(http.MethodGet, path)
|
||||||
|
resp, err := c.c.RawRequest(req)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result struct {
|
||||||
|
Data GetPluginResponse
|
||||||
|
}
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return &result.Data, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// RegisterPluginInput is used as input to the RegisterPlugin function.
|
||||||
|
type RegisterPluginInput struct {
|
||||||
|
// Name is the name of the plugin. Required.
|
||||||
|
Name string `json:"-"`
|
||||||
|
|
||||||
|
// Args is the list of args to spawn the process with.
|
||||||
|
Args []string `json:"args,omitempty"`
|
||||||
|
|
||||||
|
// Command is the command to run.
|
||||||
|
Command string `json:"command,omitempty"`
|
||||||
|
|
||||||
|
// SHA256 is the shasum of the plugin.
|
||||||
|
SHA256 string `json:"sha256,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// RegisterPlugin registers the plugin with the given information.
|
||||||
|
func (c *Sys) RegisterPlugin(i *RegisterPluginInput) error {
|
||||||
|
path := fmt.Sprintf("/v1/sys/plugins/catalog/%s", i.Name)
|
||||||
|
req := c.c.NewRequest(http.MethodPut, path)
|
||||||
|
if err := req.SetJSONBody(i); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(req)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// DeregisterPluginInput is used as input to the DeregisterPlugin function.
|
||||||
|
type DeregisterPluginInput struct {
|
||||||
|
// Name is the name of the plugin. Required.
|
||||||
|
Name string `json:"-"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// DeregisterPlugin removes the plugin with the given name from the plugin
|
||||||
|
// catalog.
|
||||||
|
func (c *Sys) DeregisterPlugin(i *DeregisterPluginInput) error {
|
||||||
|
path := fmt.Sprintf("/v1/sys/plugins/catalog/%s", i.Name)
|
||||||
|
req := c.c.NewRequest(http.MethodDelete, path)
|
||||||
|
resp, err := c.c.RawRequest(req)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
|
@ -0,0 +1,97 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import "fmt"
|
||||||
|
|
||||||
|
func (c *Sys) ListPolicies() ([]string, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/policy")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result map[string]interface{}
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
var ok bool
|
||||||
|
if _, ok = result["policies"]; !ok {
|
||||||
|
return nil, fmt.Errorf("policies not found in response")
|
||||||
|
}
|
||||||
|
|
||||||
|
listRaw := result["policies"].([]interface{})
|
||||||
|
var policies []string
|
||||||
|
|
||||||
|
for _, val := range listRaw {
|
||||||
|
policies = append(policies, val.(string))
|
||||||
|
}
|
||||||
|
|
||||||
|
return policies, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) GetPolicy(name string) (string, error) {
|
||||||
|
r := c.c.NewRequest("GET", fmt.Sprintf("/v1/sys/policy/%s", name))
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if resp != nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
if resp.StatusCode == 404 {
|
||||||
|
return "", nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
var result map[string]interface{}
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
|
||||||
|
if rulesRaw, ok := result["rules"]; ok {
|
||||||
|
return rulesRaw.(string), nil
|
||||||
|
}
|
||||||
|
if policyRaw, ok := result["policy"]; ok {
|
||||||
|
return policyRaw.(string), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
return "", fmt.Errorf("no policy found in response")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) PutPolicy(name, rules string) error {
|
||||||
|
body := map[string]string{
|
||||||
|
"rules": rules,
|
||||||
|
}
|
||||||
|
|
||||||
|
r := c.c.NewRequest("PUT", fmt.Sprintf("/v1/sys/policy/%s", name))
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) DeletePolicy(name string) error {
|
||||||
|
r := c.c.NewRequest("DELETE", fmt.Sprintf("/v1/sys/policy/%s", name))
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
type getPoliciesResp struct {
|
||||||
|
Rules string `json:"rules"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type listPoliciesResp struct {
|
||||||
|
Policies []string `json:"policies"`
|
||||||
|
}
|
|
@ -0,0 +1,309 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
func (c *Sys) RekeyStatus() (*RekeyStatusResponse, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/rekey/init")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result RekeyStatusResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyRecoveryKeyStatus() (*RekeyStatusResponse, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/rekey-recovery-key/init")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result RekeyStatusResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyVerificationStatus() (*RekeyVerificationStatusResponse, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/rekey/verify")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result RekeyVerificationStatusResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyRecoveryKeyVerificationStatus() (*RekeyVerificationStatusResponse, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/rekey-recovery-key/verify")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result RekeyVerificationStatusResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyInit(config *RekeyInitRequest) (*RekeyStatusResponse, error) {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/rekey/init")
|
||||||
|
if err := r.SetJSONBody(config); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result RekeyStatusResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyRecoveryKeyInit(config *RekeyInitRequest) (*RekeyStatusResponse, error) {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/rekey-recovery-key/init")
|
||||||
|
if err := r.SetJSONBody(config); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result RekeyStatusResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyCancel() error {
|
||||||
|
r := c.c.NewRequest("DELETE", "/v1/sys/rekey/init")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyRecoveryKeyCancel() error {
|
||||||
|
r := c.c.NewRequest("DELETE", "/v1/sys/rekey-recovery-key/init")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyVerificationCancel() error {
|
||||||
|
r := c.c.NewRequest("DELETE", "/v1/sys/rekey/verify")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyRecoveryKeyVerificationCancel() error {
|
||||||
|
r := c.c.NewRequest("DELETE", "/v1/sys/rekey-recovery-key/verify")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyUpdate(shard, nonce string) (*RekeyUpdateResponse, error) {
|
||||||
|
body := map[string]interface{}{
|
||||||
|
"key": shard,
|
||||||
|
"nonce": nonce,
|
||||||
|
}
|
||||||
|
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/rekey/update")
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result RekeyUpdateResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyRecoveryKeyUpdate(shard, nonce string) (*RekeyUpdateResponse, error) {
|
||||||
|
body := map[string]interface{}{
|
||||||
|
"key": shard,
|
||||||
|
"nonce": nonce,
|
||||||
|
}
|
||||||
|
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/rekey-recovery-key/update")
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result RekeyUpdateResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyRetrieveBackup() (*RekeyRetrieveResponse, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/rekey/backup")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result RekeyRetrieveResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyRetrieveRecoveryBackup() (*RekeyRetrieveResponse, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/rekey/recovery-backup")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result RekeyRetrieveResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyDeleteBackup() error {
|
||||||
|
r := c.c.NewRequest("DELETE", "/v1/sys/rekey/backup")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyDeleteRecoveryBackup() error {
|
||||||
|
r := c.c.NewRequest("DELETE", "/v1/sys/rekey/recovery-backup")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyVerificationUpdate(shard, nonce string) (*RekeyVerificationUpdateResponse, error) {
|
||||||
|
body := map[string]interface{}{
|
||||||
|
"key": shard,
|
||||||
|
"nonce": nonce,
|
||||||
|
}
|
||||||
|
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/rekey/verify")
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result RekeyVerificationUpdateResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) RekeyRecoveryKeyVerificationUpdate(shard, nonce string) (*RekeyVerificationUpdateResponse, error) {
|
||||||
|
body := map[string]interface{}{
|
||||||
|
"key": shard,
|
||||||
|
"nonce": nonce,
|
||||||
|
}
|
||||||
|
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/rekey-recovery-key/verify")
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result RekeyVerificationUpdateResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
type RekeyInitRequest struct {
|
||||||
|
SecretShares int `json:"secret_shares"`
|
||||||
|
SecretThreshold int `json:"secret_threshold"`
|
||||||
|
StoredShares int `json:"stored_shares"`
|
||||||
|
PGPKeys []string `json:"pgp_keys"`
|
||||||
|
Backup bool
|
||||||
|
RequireVerification bool `json:"require_verification"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type RekeyStatusResponse struct {
|
||||||
|
Nonce string `json:"nonce"`
|
||||||
|
Started bool `json:"started"`
|
||||||
|
T int `json:"t"`
|
||||||
|
N int `json:"n"`
|
||||||
|
Progress int `json:"progress"`
|
||||||
|
Required int `json:"required"`
|
||||||
|
PGPFingerprints []string `json:"pgp_fingerprints"`
|
||||||
|
Backup bool `json:"backup"`
|
||||||
|
VerificationRequired bool `json:"verification_required"`
|
||||||
|
VerificationNonce string `json:"verification_nonce"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type RekeyUpdateResponse struct {
|
||||||
|
Nonce string `json:"nonce"`
|
||||||
|
Complete bool `json:"complete"`
|
||||||
|
Keys []string `json:"keys"`
|
||||||
|
KeysB64 []string `json:"keys_base64"`
|
||||||
|
PGPFingerprints []string `json:"pgp_fingerprints"`
|
||||||
|
Backup bool `json:"backup"`
|
||||||
|
VerificationRequired bool `json:"verification_required"`
|
||||||
|
VerificationNonce string `json:"verification_nonce,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type RekeyRetrieveResponse struct {
|
||||||
|
Nonce string `json:"nonce"`
|
||||||
|
Keys map[string][]string `json:"keys"`
|
||||||
|
KeysB64 map[string][]string `json:"keys_base64"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type RekeyVerificationStatusResponse struct {
|
||||||
|
Nonce string `json:"nonce"`
|
||||||
|
Started bool `json:"started"`
|
||||||
|
T int `json:"t"`
|
||||||
|
N int `json:"n"`
|
||||||
|
Progress int `json:"progress"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type RekeyVerificationUpdateResponse struct {
|
||||||
|
Nonce string `json:"nonce"`
|
||||||
|
Complete bool `json:"complete"`
|
||||||
|
}
|
|
@ -0,0 +1,30 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
import "time"
|
||||||
|
|
||||||
|
func (c *Sys) Rotate() error {
|
||||||
|
r := c.c.NewRequest("POST", "/v1/sys/rotate")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) KeyStatus() (*KeyStatus, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/key-status")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
result := new(KeyStatus)
|
||||||
|
err = resp.DecodeJSON(result)
|
||||||
|
return result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
type KeyStatus struct {
|
||||||
|
Term int `json:"term"`
|
||||||
|
InstallTime time.Time `json:"install_time"`
|
||||||
|
}
|
|
@ -0,0 +1,62 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
func (c *Sys) SealStatus() (*SealStatusResponse, error) {
|
||||||
|
r := c.c.NewRequest("GET", "/v1/sys/seal-status")
|
||||||
|
return sealStatusRequest(c, r)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) Seal() error {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/seal")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) ResetUnsealProcess() (*SealStatusResponse, error) {
|
||||||
|
body := map[string]interface{}{"reset": true}
|
||||||
|
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/unseal")
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return sealStatusRequest(c, r)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Sys) Unseal(shard string) (*SealStatusResponse, error) {
|
||||||
|
body := map[string]interface{}{"key": shard}
|
||||||
|
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/unseal")
|
||||||
|
if err := r.SetJSONBody(body); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return sealStatusRequest(c, r)
|
||||||
|
}
|
||||||
|
|
||||||
|
func sealStatusRequest(c *Sys, r *Request) (*SealStatusResponse, error) {
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
var result SealStatusResponse
|
||||||
|
err = resp.DecodeJSON(&result)
|
||||||
|
return &result, err
|
||||||
|
}
|
||||||
|
|
||||||
|
type SealStatusResponse struct {
|
||||||
|
Type string `json:"type"`
|
||||||
|
Sealed bool `json:"sealed"`
|
||||||
|
T int `json:"t"`
|
||||||
|
N int `json:"n"`
|
||||||
|
Progress int `json:"progress"`
|
||||||
|
Nonce string `json:"nonce"`
|
||||||
|
Version string `json:"version"`
|
||||||
|
ClusterName string `json:"cluster_name,omitempty"`
|
||||||
|
ClusterID string `json:"cluster_id,omitempty"`
|
||||||
|
RecoverySeal bool `json:"recovery_seal"`
|
||||||
|
}
|
|
@ -0,0 +1,10 @@
|
||||||
|
package api
|
||||||
|
|
||||||
|
func (c *Sys) StepDown() error {
|
||||||
|
r := c.c.NewRequest("PUT", "/v1/sys/step-down")
|
||||||
|
resp, err := c.c.RawRequest(r)
|
||||||
|
if err == nil {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
192
vendor/github.com/hashicorp/vault/helper/compressutil/compress.go
generated
vendored
Normal file
192
vendor/github.com/hashicorp/vault/helper/compressutil/compress.go
generated
vendored
Normal file
|
@ -0,0 +1,192 @@
|
||||||
|
package compressutil
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"compress/gzip"
|
||||||
|
"compress/lzw"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
|
||||||
|
"github.com/golang/snappy"
|
||||||
|
"github.com/hashicorp/errwrap"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// A byte value used as a canary prefix for the compressed information
|
||||||
|
// which is used to distinguish if a JSON input is compressed or not.
|
||||||
|
// The value of this constant should not be a first character of any
|
||||||
|
// valid JSON string.
|
||||||
|
|
||||||
|
// Byte value used as canary when using Gzip format
|
||||||
|
CompressionCanaryGzip byte = 'G'
|
||||||
|
|
||||||
|
// Byte value used as canary when using Lzw format
|
||||||
|
CompressionCanaryLzw byte = 'L'
|
||||||
|
|
||||||
|
// Byte value used as canary when using Snappy format
|
||||||
|
CompressionCanarySnappy byte = 'S'
|
||||||
|
|
||||||
|
CompressionTypeLzw = "lzw"
|
||||||
|
|
||||||
|
CompressionTypeGzip = "gzip"
|
||||||
|
|
||||||
|
CompressionTypeSnappy = "snappy"
|
||||||
|
)
|
||||||
|
|
||||||
|
// SnappyReadCloser embeds the snappy reader which implements the io.Reader
|
||||||
|
// interface. The decompress procedure in this utility expects an
|
||||||
|
// io.ReadCloser. This type implements the io.Closer interface to retain the
|
||||||
|
// generic way of decompression.
|
||||||
|
type SnappyReadCloser struct {
|
||||||
|
*snappy.Reader
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close is a noop method implemented only to satisfy the io.Closer interface
|
||||||
|
func (s *SnappyReadCloser) Close() error {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// CompressionConfig is used to select a compression type to be performed by
|
||||||
|
// Compress and Decompress utilities.
|
||||||
|
// Supported types are:
|
||||||
|
// * CompressionTypeLzw
|
||||||
|
// * CompressionTypeGzip
|
||||||
|
// * CompressionTypeSnappy
|
||||||
|
//
|
||||||
|
// When using CompressionTypeGzip, the compression levels can also be chosen:
|
||||||
|
// * gzip.DefaultCompression
|
||||||
|
// * gzip.BestSpeed
|
||||||
|
// * gzip.BestCompression
|
||||||
|
type CompressionConfig struct {
|
||||||
|
// Type of the compression algorithm to be used
|
||||||
|
Type string
|
||||||
|
|
||||||
|
// When using Gzip format, the compression level to employ
|
||||||
|
GzipCompressionLevel int
|
||||||
|
}
|
||||||
|
|
||||||
|
// Compress places the canary byte in a buffer and uses the same buffer to fill
|
||||||
|
// in the compressed information of the given input. The configuration supports
|
||||||
|
// two type of compression: LZW and Gzip. When using Gzip compression format,
|
||||||
|
// if GzipCompressionLevel is not specified, the 'gzip.DefaultCompression' will
|
||||||
|
// be assumed.
|
||||||
|
func Compress(data []byte, config *CompressionConfig) ([]byte, error) {
|
||||||
|
var buf bytes.Buffer
|
||||||
|
var writer io.WriteCloser
|
||||||
|
var err error
|
||||||
|
|
||||||
|
if config == nil {
|
||||||
|
return nil, fmt.Errorf("config is nil")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Write the canary into the buffer and create writer to compress the
|
||||||
|
// input data based on the configured type
|
||||||
|
switch config.Type {
|
||||||
|
case CompressionTypeLzw:
|
||||||
|
buf.Write([]byte{CompressionCanaryLzw})
|
||||||
|
|
||||||
|
writer = lzw.NewWriter(&buf, lzw.LSB, 8)
|
||||||
|
case CompressionTypeGzip:
|
||||||
|
buf.Write([]byte{CompressionCanaryGzip})
|
||||||
|
|
||||||
|
switch {
|
||||||
|
case config.GzipCompressionLevel == gzip.BestCompression,
|
||||||
|
config.GzipCompressionLevel == gzip.BestSpeed,
|
||||||
|
config.GzipCompressionLevel == gzip.DefaultCompression:
|
||||||
|
// These are valid compression levels
|
||||||
|
default:
|
||||||
|
// If compression level is set to NoCompression or to
|
||||||
|
// any invalid value, fallback to Defaultcompression
|
||||||
|
config.GzipCompressionLevel = gzip.DefaultCompression
|
||||||
|
}
|
||||||
|
writer, err = gzip.NewWriterLevel(&buf, config.GzipCompressionLevel)
|
||||||
|
case CompressionTypeSnappy:
|
||||||
|
buf.Write([]byte{CompressionCanarySnappy})
|
||||||
|
writer = snappy.NewBufferedWriter(&buf)
|
||||||
|
default:
|
||||||
|
return nil, fmt.Errorf("unsupported compression type")
|
||||||
|
}
|
||||||
|
|
||||||
|
if err != nil {
|
||||||
|
return nil, errwrap.Wrapf("failed to create a compression writer: {{err}}", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if writer == nil {
|
||||||
|
return nil, fmt.Errorf("failed to create a compression writer")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Compress the input and place it in the same buffer containing the
|
||||||
|
// canary byte.
|
||||||
|
if _, err = writer.Write(data); err != nil {
|
||||||
|
return nil, errwrap.Wrapf("failed to compress input data: err: {{err}}", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close the io.WriteCloser
|
||||||
|
if err = writer.Close(); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Return the compressed bytes with canary byte at the start
|
||||||
|
return buf.Bytes(), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Decompress checks if the first byte in the input matches the canary byte.
|
||||||
|
// If the first byte is a canary byte, then the input past the canary byte
|
||||||
|
// will be decompressed using the method specified in the given configuration.
|
||||||
|
// If the first byte isn't a canary byte, then the utility returns a boolean
|
||||||
|
// value indicating that the input was not compressed.
|
||||||
|
func Decompress(data []byte) ([]byte, bool, error) {
|
||||||
|
var err error
|
||||||
|
var reader io.ReadCloser
|
||||||
|
if data == nil || len(data) == 0 {
|
||||||
|
return nil, false, fmt.Errorf("'data' being decompressed is empty")
|
||||||
|
}
|
||||||
|
|
||||||
|
switch {
|
||||||
|
// If the first byte matches the canary byte, remove the canary
|
||||||
|
// byte and try to decompress the data that is after the canary.
|
||||||
|
case data[0] == CompressionCanaryGzip:
|
||||||
|
if len(data) < 2 {
|
||||||
|
return nil, false, fmt.Errorf("invalid 'data' after the canary")
|
||||||
|
}
|
||||||
|
data = data[1:]
|
||||||
|
reader, err = gzip.NewReader(bytes.NewReader(data))
|
||||||
|
case data[0] == CompressionCanaryLzw:
|
||||||
|
if len(data) < 2 {
|
||||||
|
return nil, false, fmt.Errorf("invalid 'data' after the canary")
|
||||||
|
}
|
||||||
|
data = data[1:]
|
||||||
|
reader = lzw.NewReader(bytes.NewReader(data), lzw.LSB, 8)
|
||||||
|
|
||||||
|
case data[0] == CompressionCanarySnappy:
|
||||||
|
if len(data) < 2 {
|
||||||
|
return nil, false, fmt.Errorf("invalid 'data' after the canary")
|
||||||
|
}
|
||||||
|
data = data[1:]
|
||||||
|
reader = &SnappyReadCloser{
|
||||||
|
Reader: snappy.NewReader(bytes.NewReader(data)),
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
// If the first byte doesn't match the canary byte, it means
|
||||||
|
// that the content was not compressed at all. Indicate the
|
||||||
|
// caller that the input was not compressed.
|
||||||
|
return nil, true, nil
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return nil, false, errwrap.Wrapf("failed to create a compression reader: {{err}}", err)
|
||||||
|
}
|
||||||
|
if reader == nil {
|
||||||
|
return nil, false, fmt.Errorf("failed to create a compression reader")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Close the io.ReadCloser
|
||||||
|
defer reader.Close()
|
||||||
|
|
||||||
|
// Read all the compressed data into a buffer
|
||||||
|
var buf bytes.Buffer
|
||||||
|
if _, err = io.Copy(&buf, reader); err != nil {
|
||||||
|
return nil, false, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return buf.Bytes(), false, nil
|
||||||
|
}
|
|
@ -0,0 +1,36 @@
|
||||||
|
package hclutil
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
multierror "github.com/hashicorp/go-multierror"
|
||||||
|
"github.com/hashicorp/hcl/hcl/ast"
|
||||||
|
)
|
||||||
|
|
||||||
|
// CheckHCLKeys checks whether the keys in the AST list contains any of the valid keys provided.
|
||||||
|
func CheckHCLKeys(node ast.Node, valid []string) error {
|
||||||
|
var list *ast.ObjectList
|
||||||
|
switch n := node.(type) {
|
||||||
|
case *ast.ObjectList:
|
||||||
|
list = n
|
||||||
|
case *ast.ObjectType:
|
||||||
|
list = n.List
|
||||||
|
default:
|
||||||
|
return fmt.Errorf("cannot check HCL keys of type %T", n)
|
||||||
|
}
|
||||||
|
|
||||||
|
validMap := make(map[string]struct{}, len(valid))
|
||||||
|
for _, v := range valid {
|
||||||
|
validMap[v] = struct{}{}
|
||||||
|
}
|
||||||
|
|
||||||
|
var result error
|
||||||
|
for _, item := range list.Items {
|
||||||
|
key := item.Keys[0].Token.Value().(string)
|
||||||
|
if _, ok := validMap[key]; !ok {
|
||||||
|
result = multierror.Append(result, fmt.Errorf("invalid key %q on line %d", key, item.Assign.Line))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return result
|
||||||
|
}
|
|
@ -0,0 +1,100 @@
|
||||||
|
package jsonutil
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"compress/gzip"
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
|
||||||
|
"github.com/hashicorp/errwrap"
|
||||||
|
"github.com/hashicorp/vault/helper/compressutil"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Encodes/Marshals the given object into JSON
|
||||||
|
func EncodeJSON(in interface{}) ([]byte, error) {
|
||||||
|
if in == nil {
|
||||||
|
return nil, fmt.Errorf("input for encoding is nil")
|
||||||
|
}
|
||||||
|
var buf bytes.Buffer
|
||||||
|
enc := json.NewEncoder(&buf)
|
||||||
|
if err := enc.Encode(in); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return buf.Bytes(), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// EncodeJSONAndCompress encodes the given input into JSON and compresses the
|
||||||
|
// encoded value (using Gzip format BestCompression level, by default). A
|
||||||
|
// canary byte is placed at the beginning of the returned bytes for the logic
|
||||||
|
// in decompression method to identify compressed input.
|
||||||
|
func EncodeJSONAndCompress(in interface{}, config *compressutil.CompressionConfig) ([]byte, error) {
|
||||||
|
if in == nil {
|
||||||
|
return nil, fmt.Errorf("input for encoding is nil")
|
||||||
|
}
|
||||||
|
|
||||||
|
// First JSON encode the given input
|
||||||
|
encodedBytes, err := EncodeJSON(in)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if config == nil {
|
||||||
|
config = &compressutil.CompressionConfig{
|
||||||
|
Type: compressutil.CompressionTypeGzip,
|
||||||
|
GzipCompressionLevel: gzip.BestCompression,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return compressutil.Compress(encodedBytes, config)
|
||||||
|
}
|
||||||
|
|
||||||
|
// DecodeJSON tries to decompress the given data. The call to decompress, fails
|
||||||
|
// if the content was not compressed in the first place, which is identified by
|
||||||
|
// a canary byte before the compressed data. If the data is not compressed, it
|
||||||
|
// is JSON decoded directly. Otherwise the decompressed data will be JSON
|
||||||
|
// decoded.
|
||||||
|
func DecodeJSON(data []byte, out interface{}) error {
|
||||||
|
if data == nil || len(data) == 0 {
|
||||||
|
return fmt.Errorf("'data' being decoded is nil")
|
||||||
|
}
|
||||||
|
if out == nil {
|
||||||
|
return fmt.Errorf("output parameter 'out' is nil")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Decompress the data if it was compressed in the first place
|
||||||
|
decompressedBytes, uncompressed, err := compressutil.Decompress(data)
|
||||||
|
if err != nil {
|
||||||
|
return errwrap.Wrapf("failed to decompress JSON: {{err}}", err)
|
||||||
|
}
|
||||||
|
if !uncompressed && (decompressedBytes == nil || len(decompressedBytes) == 0) {
|
||||||
|
return fmt.Errorf("decompressed data being decoded is invalid")
|
||||||
|
}
|
||||||
|
|
||||||
|
// If the input supplied failed to contain the compression canary, it
|
||||||
|
// will be notified by the compression utility. Decode the decompressed
|
||||||
|
// input.
|
||||||
|
if !uncompressed {
|
||||||
|
data = decompressedBytes
|
||||||
|
}
|
||||||
|
|
||||||
|
return DecodeJSONFromReader(bytes.NewReader(data), out)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Decodes/Unmarshals the given io.Reader pointing to a JSON, into a desired object
|
||||||
|
func DecodeJSONFromReader(r io.Reader, out interface{}) error {
|
||||||
|
if r == nil {
|
||||||
|
return fmt.Errorf("'io.Reader' being decoded is nil")
|
||||||
|
}
|
||||||
|
if out == nil {
|
||||||
|
return fmt.Errorf("output parameter 'out' is nil")
|
||||||
|
}
|
||||||
|
|
||||||
|
dec := json.NewDecoder(r)
|
||||||
|
|
||||||
|
// While decoding JSON values, interpret the integer values as `json.Number`s instead of `float64`.
|
||||||
|
dec.UseNumber()
|
||||||
|
|
||||||
|
// Since 'out' is an interface representing a pointer, pass it to the decoder without an '&'
|
||||||
|
return dec.Decode(out)
|
||||||
|
}
|
|
@ -0,0 +1,163 @@
|
||||||
|
package parseutil
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/json"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/hashicorp/errwrap"
|
||||||
|
sockaddr "github.com/hashicorp/go-sockaddr"
|
||||||
|
"github.com/hashicorp/vault/helper/strutil"
|
||||||
|
"github.com/mitchellh/mapstructure"
|
||||||
|
)
|
||||||
|
|
||||||
|
func ParseDurationSecond(in interface{}) (time.Duration, error) {
|
||||||
|
var dur time.Duration
|
||||||
|
jsonIn, ok := in.(json.Number)
|
||||||
|
if ok {
|
||||||
|
in = jsonIn.String()
|
||||||
|
}
|
||||||
|
switch in.(type) {
|
||||||
|
case string:
|
||||||
|
inp := in.(string)
|
||||||
|
if inp == "" {
|
||||||
|
return time.Duration(0), nil
|
||||||
|
}
|
||||||
|
var err error
|
||||||
|
// Look for a suffix otherwise its a plain second value
|
||||||
|
if strings.HasSuffix(inp, "s") || strings.HasSuffix(inp, "m") || strings.HasSuffix(inp, "h") || strings.HasSuffix(inp, "ms") {
|
||||||
|
dur, err = time.ParseDuration(inp)
|
||||||
|
if err != nil {
|
||||||
|
return dur, err
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// Plain integer
|
||||||
|
secs, err := strconv.ParseInt(inp, 10, 64)
|
||||||
|
if err != nil {
|
||||||
|
return dur, err
|
||||||
|
}
|
||||||
|
dur = time.Duration(secs) * time.Second
|
||||||
|
}
|
||||||
|
case int:
|
||||||
|
dur = time.Duration(in.(int)) * time.Second
|
||||||
|
case int32:
|
||||||
|
dur = time.Duration(in.(int32)) * time.Second
|
||||||
|
case int64:
|
||||||
|
dur = time.Duration(in.(int64)) * time.Second
|
||||||
|
case uint:
|
||||||
|
dur = time.Duration(in.(uint)) * time.Second
|
||||||
|
case uint32:
|
||||||
|
dur = time.Duration(in.(uint32)) * time.Second
|
||||||
|
case uint64:
|
||||||
|
dur = time.Duration(in.(uint64)) * time.Second
|
||||||
|
default:
|
||||||
|
return 0, errors.New("could not parse duration from input")
|
||||||
|
}
|
||||||
|
|
||||||
|
return dur, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func ParseInt(in interface{}) (int64, error) {
|
||||||
|
var ret int64
|
||||||
|
jsonIn, ok := in.(json.Number)
|
||||||
|
if ok {
|
||||||
|
in = jsonIn.String()
|
||||||
|
}
|
||||||
|
switch in.(type) {
|
||||||
|
case string:
|
||||||
|
inp := in.(string)
|
||||||
|
if inp == "" {
|
||||||
|
return 0, nil
|
||||||
|
}
|
||||||
|
var err error
|
||||||
|
left, err := strconv.ParseInt(inp, 10, 64)
|
||||||
|
if err != nil {
|
||||||
|
return ret, err
|
||||||
|
}
|
||||||
|
ret = left
|
||||||
|
case int:
|
||||||
|
ret = int64(in.(int))
|
||||||
|
case int32:
|
||||||
|
ret = int64(in.(int32))
|
||||||
|
case int64:
|
||||||
|
ret = in.(int64)
|
||||||
|
case uint:
|
||||||
|
ret = int64(in.(uint))
|
||||||
|
case uint32:
|
||||||
|
ret = int64(in.(uint32))
|
||||||
|
case uint64:
|
||||||
|
ret = int64(in.(uint64))
|
||||||
|
default:
|
||||||
|
return 0, errors.New("could not parse value from input")
|
||||||
|
}
|
||||||
|
|
||||||
|
return ret, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func ParseBool(in interface{}) (bool, error) {
|
||||||
|
var result bool
|
||||||
|
if err := mapstructure.WeakDecode(in, &result); err != nil {
|
||||||
|
return false, err
|
||||||
|
}
|
||||||
|
return result, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func ParseCommaStringSlice(in interface{}) ([]string, error) {
|
||||||
|
var result []string
|
||||||
|
config := &mapstructure.DecoderConfig{
|
||||||
|
Result: &result,
|
||||||
|
WeaklyTypedInput: true,
|
||||||
|
DecodeHook: mapstructure.StringToSliceHookFunc(","),
|
||||||
|
}
|
||||||
|
decoder, err := mapstructure.NewDecoder(config)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if err := decoder.Decode(in); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return strutil.TrimStrings(result), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func ParseAddrs(addrs interface{}) ([]*sockaddr.SockAddrMarshaler, error) {
|
||||||
|
out := make([]*sockaddr.SockAddrMarshaler, 0)
|
||||||
|
stringAddrs := make([]string, 0)
|
||||||
|
|
||||||
|
switch addrs.(type) {
|
||||||
|
case string:
|
||||||
|
stringAddrs = strutil.ParseArbitraryStringSlice(addrs.(string), ",")
|
||||||
|
if len(stringAddrs) == 0 {
|
||||||
|
return nil, fmt.Errorf("unable to parse addresses from %v", addrs)
|
||||||
|
}
|
||||||
|
|
||||||
|
case []string:
|
||||||
|
stringAddrs = addrs.([]string)
|
||||||
|
|
||||||
|
case []interface{}:
|
||||||
|
for _, v := range addrs.([]interface{}) {
|
||||||
|
stringAddr, ok := v.(string)
|
||||||
|
if !ok {
|
||||||
|
return nil, fmt.Errorf("error parsing %v as string", v)
|
||||||
|
}
|
||||||
|
stringAddrs = append(stringAddrs, stringAddr)
|
||||||
|
}
|
||||||
|
|
||||||
|
default:
|
||||||
|
return nil, fmt.Errorf("unknown address input type %T", addrs)
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, addr := range stringAddrs {
|
||||||
|
sa, err := sockaddr.NewSockAddr(addr)
|
||||||
|
if err != nil {
|
||||||
|
return nil, errwrap.Wrapf(fmt.Sprintf("error parsing address %q: {{err}}", addr), err)
|
||||||
|
}
|
||||||
|
out = append(out, &sockaddr.SockAddrMarshaler{
|
||||||
|
SockAddr: sa,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
return out, nil
|
||||||
|
}
|
|
@ -0,0 +1,327 @@
|
||||||
|
package strutil
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/base64"
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
"sort"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/hashicorp/errwrap"
|
||||||
|
glob "github.com/ryanuber/go-glob"
|
||||||
|
)
|
||||||
|
|
||||||
|
// StrListContainsGlob looks for a string in a list of strings and allows
|
||||||
|
// globs.
|
||||||
|
func StrListContainsGlob(haystack []string, needle string) bool {
|
||||||
|
for _, item := range haystack {
|
||||||
|
if glob.Glob(item, needle) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// StrListContains looks for a string in a list of strings.
|
||||||
|
func StrListContains(haystack []string, needle string) bool {
|
||||||
|
for _, item := range haystack {
|
||||||
|
if item == needle {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// StrListSubset checks if a given list is a subset
|
||||||
|
// of another set
|
||||||
|
func StrListSubset(super, sub []string) bool {
|
||||||
|
for _, item := range sub {
|
||||||
|
if !StrListContains(super, item) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parses a comma separated list of strings into a slice of strings.
|
||||||
|
// The return slice will be sorted and will not contain duplicate or
|
||||||
|
// empty items.
|
||||||
|
func ParseDedupAndSortStrings(input string, sep string) []string {
|
||||||
|
input = strings.TrimSpace(input)
|
||||||
|
parsed := []string{}
|
||||||
|
if input == "" {
|
||||||
|
// Don't return nil
|
||||||
|
return parsed
|
||||||
|
}
|
||||||
|
return RemoveDuplicates(strings.Split(input, sep), false)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parses a comma separated list of strings into a slice of strings.
|
||||||
|
// The return slice will be sorted and will not contain duplicate or
|
||||||
|
// empty items. The values will be converted to lower case.
|
||||||
|
func ParseDedupLowercaseAndSortStrings(input string, sep string) []string {
|
||||||
|
input = strings.TrimSpace(input)
|
||||||
|
parsed := []string{}
|
||||||
|
if input == "" {
|
||||||
|
// Don't return nil
|
||||||
|
return parsed
|
||||||
|
}
|
||||||
|
return RemoveDuplicates(strings.Split(input, sep), true)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parses a comma separated list of `<key>=<value>` tuples into a
|
||||||
|
// map[string]string.
|
||||||
|
func ParseKeyValues(input string, out map[string]string, sep string) error {
|
||||||
|
if out == nil {
|
||||||
|
return fmt.Errorf("'out is nil")
|
||||||
|
}
|
||||||
|
|
||||||
|
keyValues := ParseDedupLowercaseAndSortStrings(input, sep)
|
||||||
|
if len(keyValues) == 0 {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, keyValue := range keyValues {
|
||||||
|
shards := strings.Split(keyValue, "=")
|
||||||
|
if len(shards) != 2 {
|
||||||
|
return fmt.Errorf("invalid <key,value> format")
|
||||||
|
}
|
||||||
|
|
||||||
|
key := strings.TrimSpace(shards[0])
|
||||||
|
value := strings.TrimSpace(shards[1])
|
||||||
|
if key == "" || value == "" {
|
||||||
|
return fmt.Errorf("invalid <key,value> pair: key: %q value: %q", key, value)
|
||||||
|
}
|
||||||
|
out[key] = value
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parses arbitrary <key,value> tuples. The input can be one of
|
||||||
|
// the following:
|
||||||
|
// * JSON string
|
||||||
|
// * Base64 encoded JSON string
|
||||||
|
// * Comma separated list of `<key>=<value>` pairs
|
||||||
|
// * Base64 encoded string containing comma separated list of
|
||||||
|
// `<key>=<value>` pairs
|
||||||
|
//
|
||||||
|
// Input will be parsed into the output parameter, which should
|
||||||
|
// be a non-nil map[string]string.
|
||||||
|
func ParseArbitraryKeyValues(input string, out map[string]string, sep string) error {
|
||||||
|
input = strings.TrimSpace(input)
|
||||||
|
if input == "" {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
if out == nil {
|
||||||
|
return fmt.Errorf("'out' is nil")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Try to base64 decode the input. If successful, consider the decoded
|
||||||
|
// value as input.
|
||||||
|
inputBytes, err := base64.StdEncoding.DecodeString(input)
|
||||||
|
if err == nil {
|
||||||
|
input = string(inputBytes)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Try to JSON unmarshal the input. If successful, consider that the
|
||||||
|
// metadata was supplied as JSON input.
|
||||||
|
err = json.Unmarshal([]byte(input), &out)
|
||||||
|
if err != nil {
|
||||||
|
// If JSON unmarshalling fails, consider that the input was
|
||||||
|
// supplied as a comma separated string of 'key=value' pairs.
|
||||||
|
if err = ParseKeyValues(input, out, sep); err != nil {
|
||||||
|
return errwrap.Wrapf("failed to parse the input: {{err}}", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Validate the parsed input
|
||||||
|
for key, value := range out {
|
||||||
|
if key != "" && value == "" {
|
||||||
|
return fmt.Errorf("invalid value for key %q", key)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parses a `sep`-separated list of strings into a
|
||||||
|
// []string.
|
||||||
|
//
|
||||||
|
// The output will always be a valid slice but may be of length zero.
|
||||||
|
func ParseStringSlice(input string, sep string) []string {
|
||||||
|
input = strings.TrimSpace(input)
|
||||||
|
if input == "" {
|
||||||
|
return []string{}
|
||||||
|
}
|
||||||
|
|
||||||
|
splitStr := strings.Split(input, sep)
|
||||||
|
ret := make([]string, len(splitStr))
|
||||||
|
for i, val := range splitStr {
|
||||||
|
ret[i] = val
|
||||||
|
}
|
||||||
|
|
||||||
|
return ret
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parses arbitrary string slice. The input can be one of
|
||||||
|
// the following:
|
||||||
|
// * JSON string
|
||||||
|
// * Base64 encoded JSON string
|
||||||
|
// * `sep` separated list of values
|
||||||
|
// * Base64-encoded string containing a `sep` separated list of values
|
||||||
|
//
|
||||||
|
// Note that the separator is ignored if the input is found to already be in a
|
||||||
|
// structured format (e.g., JSON)
|
||||||
|
//
|
||||||
|
// The output will always be a valid slice but may be of length zero.
|
||||||
|
func ParseArbitraryStringSlice(input string, sep string) []string {
|
||||||
|
input = strings.TrimSpace(input)
|
||||||
|
if input == "" {
|
||||||
|
return []string{}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Try to base64 decode the input. If successful, consider the decoded
|
||||||
|
// value as input.
|
||||||
|
inputBytes, err := base64.StdEncoding.DecodeString(input)
|
||||||
|
if err == nil {
|
||||||
|
input = string(inputBytes)
|
||||||
|
}
|
||||||
|
|
||||||
|
ret := []string{}
|
||||||
|
|
||||||
|
// Try to JSON unmarshal the input. If successful, consider that the
|
||||||
|
// metadata was supplied as JSON input.
|
||||||
|
err = json.Unmarshal([]byte(input), &ret)
|
||||||
|
if err != nil {
|
||||||
|
// If JSON unmarshalling fails, consider that the input was
|
||||||
|
// supplied as a separated string of values.
|
||||||
|
return ParseStringSlice(input, sep)
|
||||||
|
}
|
||||||
|
|
||||||
|
if ret == nil {
|
||||||
|
return []string{}
|
||||||
|
}
|
||||||
|
|
||||||
|
return ret
|
||||||
|
}
|
||||||
|
|
||||||
|
// TrimStrings takes a slice of strings and returns a slice of strings
|
||||||
|
// with trimmed spaces
|
||||||
|
func TrimStrings(items []string) []string {
|
||||||
|
ret := make([]string, len(items))
|
||||||
|
for i, item := range items {
|
||||||
|
ret[i] = strings.TrimSpace(item)
|
||||||
|
}
|
||||||
|
return ret
|
||||||
|
}
|
||||||
|
|
||||||
|
// Removes duplicate and empty elements from a slice of strings. This also may
|
||||||
|
// convert the items in the slice to lower case and returns a sorted slice.
|
||||||
|
func RemoveDuplicates(items []string, lowercase bool) []string {
|
||||||
|
itemsMap := map[string]bool{}
|
||||||
|
for _, item := range items {
|
||||||
|
item = strings.TrimSpace(item)
|
||||||
|
if lowercase {
|
||||||
|
item = strings.ToLower(item)
|
||||||
|
}
|
||||||
|
if item == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
itemsMap[item] = true
|
||||||
|
}
|
||||||
|
items = make([]string, 0, len(itemsMap))
|
||||||
|
for item, _ := range itemsMap {
|
||||||
|
items = append(items, item)
|
||||||
|
}
|
||||||
|
sort.Strings(items)
|
||||||
|
return items
|
||||||
|
}
|
||||||
|
|
||||||
|
// EquivalentSlices checks whether the given string sets are equivalent, as in,
|
||||||
|
// they contain the same values.
|
||||||
|
func EquivalentSlices(a, b []string) bool {
|
||||||
|
if a == nil && b == nil {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
if a == nil || b == nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// First we'll build maps to ensure unique values
|
||||||
|
mapA := map[string]bool{}
|
||||||
|
mapB := map[string]bool{}
|
||||||
|
for _, keyA := range a {
|
||||||
|
mapA[keyA] = true
|
||||||
|
}
|
||||||
|
for _, keyB := range b {
|
||||||
|
mapB[keyB] = true
|
||||||
|
}
|
||||||
|
|
||||||
|
// Now we'll build our checking slices
|
||||||
|
var sortedA, sortedB []string
|
||||||
|
for keyA, _ := range mapA {
|
||||||
|
sortedA = append(sortedA, keyA)
|
||||||
|
}
|
||||||
|
for keyB, _ := range mapB {
|
||||||
|
sortedB = append(sortedB, keyB)
|
||||||
|
}
|
||||||
|
sort.Strings(sortedA)
|
||||||
|
sort.Strings(sortedB)
|
||||||
|
|
||||||
|
// Finally, compare
|
||||||
|
if len(sortedA) != len(sortedB) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
for i := range sortedA {
|
||||||
|
if sortedA[i] != sortedB[i] {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// StrListDelete removes the first occurrence of the given item from the slice
|
||||||
|
// of strings if the item exists.
|
||||||
|
func StrListDelete(s []string, d string) []string {
|
||||||
|
if s == nil {
|
||||||
|
return s
|
||||||
|
}
|
||||||
|
|
||||||
|
for index, element := range s {
|
||||||
|
if element == d {
|
||||||
|
return append(s[:index], s[index+1:]...)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return s
|
||||||
|
}
|
||||||
|
|
||||||
|
func GlobbedStringsMatch(item, val string) bool {
|
||||||
|
if len(item) < 2 {
|
||||||
|
return val == item
|
||||||
|
}
|
||||||
|
|
||||||
|
hasPrefix := strings.HasPrefix(item, "*")
|
||||||
|
hasSuffix := strings.HasSuffix(item, "*")
|
||||||
|
|
||||||
|
if hasPrefix && hasSuffix {
|
||||||
|
return strings.Contains(val, item[1:len(item)-1])
|
||||||
|
} else if hasPrefix {
|
||||||
|
return strings.HasSuffix(val, item[1:])
|
||||||
|
} else if hasSuffix {
|
||||||
|
return strings.HasPrefix(val, item[:len(item)-1])
|
||||||
|
}
|
||||||
|
|
||||||
|
return val == item
|
||||||
|
}
|
||||||
|
|
||||||
|
// AppendIfMissing adds a string to a slice if the given string is not present
|
||||||
|
func AppendIfMissing(slice []string, i string) []string {
|
||||||
|
if StrListContains(slice, i) {
|
||||||
|
return slice
|
||||||
|
}
|
||||||
|
return append(slice, i)
|
||||||
|
}
|
|
@ -0,0 +1,21 @@
|
||||||
|
The MIT License (MIT)
|
||||||
|
|
||||||
|
Copyright (c) 2014 Ryan Uber
|
||||||
|
|
||||||
|
Permission is hereby granted, free of charge, to any person obtaining a copy
|
||||||
|
of this software and associated documentation files (the "Software"), to deal
|
||||||
|
in the Software without restriction, including without limitation the rights
|
||||||
|
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
||||||
|
copies of the Software, and to permit persons to whom the Software is
|
||||||
|
furnished to do so, subject to the following conditions:
|
||||||
|
|
||||||
|
The above copyright notice and this permission notice shall be included in all
|
||||||
|
copies or substantial portions of the Software.
|
||||||
|
|
||||||
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||||
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||||
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
||||||
|
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||||
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
||||||
|
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
||||||
|
SOFTWARE.
|
|
@ -0,0 +1,29 @@
|
||||||
|
# String globbing in golang [![Build Status](https://travis-ci.org/ryanuber/go-glob.svg)](https://travis-ci.org/ryanuber/go-glob)
|
||||||
|
|
||||||
|
`go-glob` is a single-function library implementing basic string glob support.
|
||||||
|
|
||||||
|
Globs are an extremely user-friendly way of supporting string matching without
|
||||||
|
requiring knowledge of regular expressions or Go's particular regex engine. Most
|
||||||
|
people understand that if you put a `*` character somewhere in a string, it is
|
||||||
|
treated as a wildcard. Surprisingly, this functionality isn't found in Go's
|
||||||
|
standard library, except for `path.Match`, which is intended to be used while
|
||||||
|
comparing paths (not arbitrary strings), and contains specialized logic for this
|
||||||
|
use case. A better solution might be a POSIX basic (non-ERE) regular expression
|
||||||
|
engine for Go, which doesn't exist currently.
|
||||||
|
|
||||||
|
Example
|
||||||
|
=======
|
||||||
|
|
||||||
|
```
|
||||||
|
package main
|
||||||
|
|
||||||
|
import "github.com/ryanuber/go-glob"
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
glob.Glob("*World!", "Hello, World!") // true
|
||||||
|
glob.Glob("Hello,*", "Hello, World!") // true
|
||||||
|
glob.Glob("*ello,*", "Hello, World!") // true
|
||||||
|
glob.Glob("World!", "Hello, World!") // false
|
||||||
|
glob.Glob("/home/*", "/home/ryanuber/.bashrc") // true
|
||||||
|
}
|
||||||
|
```
|
|
@ -0,0 +1,56 @@
|
||||||
|
package glob
|
||||||
|
|
||||||
|
import "strings"
|
||||||
|
|
||||||
|
// The character which is treated like a glob
|
||||||
|
const GLOB = "*"
|
||||||
|
|
||||||
|
// Glob will test a string pattern, potentially containing globs, against a
|
||||||
|
// subject string. The result is a simple true/false, determining whether or
|
||||||
|
// not the glob pattern matched the subject text.
|
||||||
|
func Glob(pattern, subj string) bool {
|
||||||
|
// Empty pattern can only match empty subject
|
||||||
|
if pattern == "" {
|
||||||
|
return subj == pattern
|
||||||
|
}
|
||||||
|
|
||||||
|
// If the pattern _is_ a glob, it matches everything
|
||||||
|
if pattern == GLOB {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
parts := strings.Split(pattern, GLOB)
|
||||||
|
|
||||||
|
if len(parts) == 1 {
|
||||||
|
// No globs in pattern, so test for equality
|
||||||
|
return subj == pattern
|
||||||
|
}
|
||||||
|
|
||||||
|
leadingGlob := strings.HasPrefix(pattern, GLOB)
|
||||||
|
trailingGlob := strings.HasSuffix(pattern, GLOB)
|
||||||
|
end := len(parts) - 1
|
||||||
|
|
||||||
|
// Go over the leading parts and ensure they match.
|
||||||
|
for i := 0; i < end; i++ {
|
||||||
|
idx := strings.Index(subj, parts[i])
|
||||||
|
|
||||||
|
switch i {
|
||||||
|
case 0:
|
||||||
|
// Check the first section. Requires special handling.
|
||||||
|
if !leadingGlob && idx != 0 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
// Check that the middle parts match.
|
||||||
|
if idx < 0 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Trim evaluated text from subj as we loop over the pattern.
|
||||||
|
subj = subj[idx+len(parts[i]):]
|
||||||
|
}
|
||||||
|
|
||||||
|
// Reached the last section. Requires special handling.
|
||||||
|
return trailingGlob || strings.HasSuffix(subj, parts[end])
|
||||||
|
}
|
|
@ -0,0 +1,50 @@
|
||||||
|
// Copyright 2018 The Go Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
// Package httpguts provides functions implementing various details
|
||||||
|
// of the HTTP specification.
|
||||||
|
//
|
||||||
|
// This package is shared by the standard library (which vendors it)
|
||||||
|
// and x/net/http2. It comes with no API stability promise.
|
||||||
|
package httpguts
|
||||||
|
|
||||||
|
import (
|
||||||
|
"net/textproto"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
// ValidTrailerHeader reports whether name is a valid header field name to appear
|
||||||
|
// in trailers.
|
||||||
|
// See RFC 7230, Section 4.1.2
|
||||||
|
func ValidTrailerHeader(name string) bool {
|
||||||
|
name = textproto.CanonicalMIMEHeaderKey(name)
|
||||||
|
if strings.HasPrefix(name, "If-") || badTrailer[name] {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
var badTrailer = map[string]bool{
|
||||||
|
"Authorization": true,
|
||||||
|
"Cache-Control": true,
|
||||||
|
"Connection": true,
|
||||||
|
"Content-Encoding": true,
|
||||||
|
"Content-Length": true,
|
||||||
|
"Content-Range": true,
|
||||||
|
"Content-Type": true,
|
||||||
|
"Expect": true,
|
||||||
|
"Host": true,
|
||||||
|
"Keep-Alive": true,
|
||||||
|
"Max-Forwards": true,
|
||||||
|
"Pragma": true,
|
||||||
|
"Proxy-Authenticate": true,
|
||||||
|
"Proxy-Authorization": true,
|
||||||
|
"Proxy-Connection": true,
|
||||||
|
"Range": true,
|
||||||
|
"Realm": true,
|
||||||
|
"Te": true,
|
||||||
|
"Trailer": true,
|
||||||
|
"Transfer-Encoding": true,
|
||||||
|
"Www-Authenticate": true,
|
||||||
|
}
|
|
@ -0,0 +1,346 @@
|
||||||
|
// Copyright 2016 The Go Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a BSD-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
package httpguts
|
||||||
|
|
||||||
|
import (
|
||||||
|
"net"
|
||||||
|
"strings"
|
||||||
|
"unicode/utf8"
|
||||||
|
|
||||||
|
"golang.org/x/net/idna"
|
||||||
|
)
|
||||||
|
|
||||||
|
var isTokenTable = [127]bool{
|
||||||
|
'!': true,
|
||||||
|
'#': true,
|
||||||
|
'$': true,
|
||||||
|
'%': true,
|
||||||
|
'&': true,
|
||||||
|
'\'': true,
|
||||||
|
'*': true,
|
||||||
|
'+': true,
|
||||||
|
'-': true,
|
||||||
|
'.': true,
|
||||||
|
'0': true,
|
||||||
|
'1': true,
|
||||||
|
'2': true,
|
||||||
|
'3': true,
|
||||||
|
'4': true,
|
||||||
|
'5': true,
|
||||||
|
'6': true,
|
||||||
|
'7': true,
|
||||||
|
'8': true,
|
||||||
|
'9': true,
|
||||||
|
'A': true,
|
||||||
|
'B': true,
|
||||||
|
'C': true,
|
||||||
|
'D': true,
|
||||||
|
'E': true,
|
||||||
|
'F': true,
|
||||||
|
'G': true,
|
||||||
|
'H': true,
|
||||||
|
'I': true,
|
||||||
|
'J': true,
|
||||||
|
'K': true,
|
||||||
|
'L': true,
|
||||||
|
'M': true,
|
||||||
|
'N': true,
|
||||||
|
'O': true,
|
||||||
|
'P': true,
|
||||||
|
'Q': true,
|
||||||
|
'R': true,
|
||||||
|
'S': true,
|
||||||
|
'T': true,
|
||||||
|
'U': true,
|
||||||
|
'W': true,
|
||||||
|
'V': true,
|
||||||
|
'X': true,
|
||||||
|
'Y': true,
|
||||||
|
'Z': true,
|
||||||
|
'^': true,
|
||||||
|
'_': true,
|
||||||
|
'`': true,
|
||||||
|
'a': true,
|
||||||
|
'b': true,
|
||||||
|
'c': true,
|
||||||
|
'd': true,
|
||||||
|
'e': true,
|
||||||
|
'f': true,
|
||||||
|
'g': true,
|
||||||
|
'h': true,
|
||||||
|
'i': true,
|
||||||
|
'j': true,
|
||||||
|
'k': true,
|
||||||
|
'l': true,
|
||||||
|
'm': true,
|
||||||
|
'n': true,
|
||||||
|
'o': true,
|
||||||
|
'p': true,
|
||||||
|
'q': true,
|
||||||
|
'r': true,
|
||||||
|
's': true,
|
||||||
|
't': true,
|
||||||
|
'u': true,
|
||||||
|
'v': true,
|
||||||
|
'w': true,
|
||||||
|
'x': true,
|
||||||
|
'y': true,
|
||||||
|
'z': true,
|
||||||
|
'|': true,
|
||||||
|
'~': true,
|
||||||
|
}
|
||||||
|
|
||||||
|
func IsTokenRune(r rune) bool {
|
||||||
|
i := int(r)
|
||||||
|
return i < len(isTokenTable) && isTokenTable[i]
|
||||||
|
}
|
||||||
|
|
||||||
|
func isNotToken(r rune) bool {
|
||||||
|
return !IsTokenRune(r)
|
||||||
|
}
|
||||||
|
|
||||||
|
// HeaderValuesContainsToken reports whether any string in values
|
||||||
|
// contains the provided token, ASCII case-insensitively.
|
||||||
|
func HeaderValuesContainsToken(values []string, token string) bool {
|
||||||
|
for _, v := range values {
|
||||||
|
if headerValueContainsToken(v, token) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// isOWS reports whether b is an optional whitespace byte, as defined
|
||||||
|
// by RFC 7230 section 3.2.3.
|
||||||
|
func isOWS(b byte) bool { return b == ' ' || b == '\t' }
|
||||||
|
|
||||||
|
// trimOWS returns x with all optional whitespace removes from the
|
||||||
|
// beginning and end.
|
||||||
|
func trimOWS(x string) string {
|
||||||
|
// TODO: consider using strings.Trim(x, " \t") instead,
|
||||||
|
// if and when it's fast enough. See issue 10292.
|
||||||
|
// But this ASCII-only code will probably always beat UTF-8
|
||||||
|
// aware code.
|
||||||
|
for len(x) > 0 && isOWS(x[0]) {
|
||||||
|
x = x[1:]
|
||||||
|
}
|
||||||
|
for len(x) > 0 && isOWS(x[len(x)-1]) {
|
||||||
|
x = x[:len(x)-1]
|
||||||
|
}
|
||||||
|
return x
|
||||||
|
}
|
||||||
|
|
||||||
|
// headerValueContainsToken reports whether v (assumed to be a
|
||||||
|
// 0#element, in the ABNF extension described in RFC 7230 section 7)
|
||||||
|
// contains token amongst its comma-separated tokens, ASCII
|
||||||
|
// case-insensitively.
|
||||||
|
func headerValueContainsToken(v string, token string) bool {
|
||||||
|
v = trimOWS(v)
|
||||||
|
if comma := strings.IndexByte(v, ','); comma != -1 {
|
||||||
|
return tokenEqual(trimOWS(v[:comma]), token) || headerValueContainsToken(v[comma+1:], token)
|
||||||
|
}
|
||||||
|
return tokenEqual(v, token)
|
||||||
|
}
|
||||||
|
|
||||||
|
// lowerASCII returns the ASCII lowercase version of b.
|
||||||
|
func lowerASCII(b byte) byte {
|
||||||
|
if 'A' <= b && b <= 'Z' {
|
||||||
|
return b + ('a' - 'A')
|
||||||
|
}
|
||||||
|
return b
|
||||||
|
}
|
||||||
|
|
||||||
|
// tokenEqual reports whether t1 and t2 are equal, ASCII case-insensitively.
|
||||||
|
func tokenEqual(t1, t2 string) bool {
|
||||||
|
if len(t1) != len(t2) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
for i, b := range t1 {
|
||||||
|
if b >= utf8.RuneSelf {
|
||||||
|
// No UTF-8 or non-ASCII allowed in tokens.
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
if lowerASCII(byte(b)) != lowerASCII(t2[i]) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// isLWS reports whether b is linear white space, according
|
||||||
|
// to http://www.w3.org/Protocols/rfc2616/rfc2616-sec2.html#sec2.2
|
||||||
|
// LWS = [CRLF] 1*( SP | HT )
|
||||||
|
func isLWS(b byte) bool { return b == ' ' || b == '\t' }
|
||||||
|
|
||||||
|
// isCTL reports whether b is a control byte, according
|
||||||
|
// to http://www.w3.org/Protocols/rfc2616/rfc2616-sec2.html#sec2.2
|
||||||
|
// CTL = <any US-ASCII control character
|
||||||
|
// (octets 0 - 31) and DEL (127)>
|
||||||
|
func isCTL(b byte) bool {
|
||||||
|
const del = 0x7f // a CTL
|
||||||
|
return b < ' ' || b == del
|
||||||
|
}
|
||||||
|
|
||||||
|
// ValidHeaderFieldName reports whether v is a valid HTTP/1.x header name.
|
||||||
|
// HTTP/2 imposes the additional restriction that uppercase ASCII
|
||||||
|
// letters are not allowed.
|
||||||
|
//
|
||||||
|
// RFC 7230 says:
|
||||||
|
// header-field = field-name ":" OWS field-value OWS
|
||||||
|
// field-name = token
|
||||||
|
// token = 1*tchar
|
||||||
|
// tchar = "!" / "#" / "$" / "%" / "&" / "'" / "*" / "+" / "-" / "." /
|
||||||
|
// "^" / "_" / "`" / "|" / "~" / DIGIT / ALPHA
|
||||||
|
func ValidHeaderFieldName(v string) bool {
|
||||||
|
if len(v) == 0 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
for _, r := range v {
|
||||||
|
if !IsTokenRune(r) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// ValidHostHeader reports whether h is a valid host header.
|
||||||
|
func ValidHostHeader(h string) bool {
|
||||||
|
// The latest spec is actually this:
|
||||||
|
//
|
||||||
|
// http://tools.ietf.org/html/rfc7230#section-5.4
|
||||||
|
// Host = uri-host [ ":" port ]
|
||||||
|
//
|
||||||
|
// Where uri-host is:
|
||||||
|
// http://tools.ietf.org/html/rfc3986#section-3.2.2
|
||||||
|
//
|
||||||
|
// But we're going to be much more lenient for now and just
|
||||||
|
// search for any byte that's not a valid byte in any of those
|
||||||
|
// expressions.
|
||||||
|
for i := 0; i < len(h); i++ {
|
||||||
|
if !validHostByte[h[i]] {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// See the validHostHeader comment.
|
||||||
|
var validHostByte = [256]bool{
|
||||||
|
'0': true, '1': true, '2': true, '3': true, '4': true, '5': true, '6': true, '7': true,
|
||||||
|
'8': true, '9': true,
|
||||||
|
|
||||||
|
'a': true, 'b': true, 'c': true, 'd': true, 'e': true, 'f': true, 'g': true, 'h': true,
|
||||||
|
'i': true, 'j': true, 'k': true, 'l': true, 'm': true, 'n': true, 'o': true, 'p': true,
|
||||||
|
'q': true, 'r': true, 's': true, 't': true, 'u': true, 'v': true, 'w': true, 'x': true,
|
||||||
|
'y': true, 'z': true,
|
||||||
|
|
||||||
|
'A': true, 'B': true, 'C': true, 'D': true, 'E': true, 'F': true, 'G': true, 'H': true,
|
||||||
|
'I': true, 'J': true, 'K': true, 'L': true, 'M': true, 'N': true, 'O': true, 'P': true,
|
||||||
|
'Q': true, 'R': true, 'S': true, 'T': true, 'U': true, 'V': true, 'W': true, 'X': true,
|
||||||
|
'Y': true, 'Z': true,
|
||||||
|
|
||||||
|
'!': true, // sub-delims
|
||||||
|
'$': true, // sub-delims
|
||||||
|
'%': true, // pct-encoded (and used in IPv6 zones)
|
||||||
|
'&': true, // sub-delims
|
||||||
|
'(': true, // sub-delims
|
||||||
|
')': true, // sub-delims
|
||||||
|
'*': true, // sub-delims
|
||||||
|
'+': true, // sub-delims
|
||||||
|
',': true, // sub-delims
|
||||||
|
'-': true, // unreserved
|
||||||
|
'.': true, // unreserved
|
||||||
|
':': true, // IPv6address + Host expression's optional port
|
||||||
|
';': true, // sub-delims
|
||||||
|
'=': true, // sub-delims
|
||||||
|
'[': true,
|
||||||
|
'\'': true, // sub-delims
|
||||||
|
']': true,
|
||||||
|
'_': true, // unreserved
|
||||||
|
'~': true, // unreserved
|
||||||
|
}
|
||||||
|
|
||||||
|
// ValidHeaderFieldValue reports whether v is a valid "field-value" according to
|
||||||
|
// http://www.w3.org/Protocols/rfc2616/rfc2616-sec4.html#sec4.2 :
|
||||||
|
//
|
||||||
|
// message-header = field-name ":" [ field-value ]
|
||||||
|
// field-value = *( field-content | LWS )
|
||||||
|
// field-content = <the OCTETs making up the field-value
|
||||||
|
// and consisting of either *TEXT or combinations
|
||||||
|
// of token, separators, and quoted-string>
|
||||||
|
//
|
||||||
|
// http://www.w3.org/Protocols/rfc2616/rfc2616-sec2.html#sec2.2 :
|
||||||
|
//
|
||||||
|
// TEXT = <any OCTET except CTLs,
|
||||||
|
// but including LWS>
|
||||||
|
// LWS = [CRLF] 1*( SP | HT )
|
||||||
|
// CTL = <any US-ASCII control character
|
||||||
|
// (octets 0 - 31) and DEL (127)>
|
||||||
|
//
|
||||||
|
// RFC 7230 says:
|
||||||
|
// field-value = *( field-content / obs-fold )
|
||||||
|
// obj-fold = N/A to http2, and deprecated
|
||||||
|
// field-content = field-vchar [ 1*( SP / HTAB ) field-vchar ]
|
||||||
|
// field-vchar = VCHAR / obs-text
|
||||||
|
// obs-text = %x80-FF
|
||||||
|
// VCHAR = "any visible [USASCII] character"
|
||||||
|
//
|
||||||
|
// http2 further says: "Similarly, HTTP/2 allows header field values
|
||||||
|
// that are not valid. While most of the values that can be encoded
|
||||||
|
// will not alter header field parsing, carriage return (CR, ASCII
|
||||||
|
// 0xd), line feed (LF, ASCII 0xa), and the zero character (NUL, ASCII
|
||||||
|
// 0x0) might be exploited by an attacker if they are translated
|
||||||
|
// verbatim. Any request or response that contains a character not
|
||||||
|
// permitted in a header field value MUST be treated as malformed
|
||||||
|
// (Section 8.1.2.6). Valid characters are defined by the
|
||||||
|
// field-content ABNF rule in Section 3.2 of [RFC7230]."
|
||||||
|
//
|
||||||
|
// This function does not (yet?) properly handle the rejection of
|
||||||
|
// strings that begin or end with SP or HTAB.
|
||||||
|
func ValidHeaderFieldValue(v string) bool {
|
||||||
|
for i := 0; i < len(v); i++ {
|
||||||
|
b := v[i]
|
||||||
|
if isCTL(b) && !isLWS(b) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
func isASCII(s string) bool {
|
||||||
|
for i := 0; i < len(s); i++ {
|
||||||
|
if s[i] >= utf8.RuneSelf {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// PunycodeHostPort returns the IDNA Punycode version
|
||||||
|
// of the provided "host" or "host:port" string.
|
||||||
|
func PunycodeHostPort(v string) (string, error) {
|
||||||
|
if isASCII(v) {
|
||||||
|
return v, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
host, port, err := net.SplitHostPort(v)
|
||||||
|
if err != nil {
|
||||||
|
// The input 'v' argument was just a "host" argument,
|
||||||
|
// without a port. This error should not be returned
|
||||||
|
// to the caller.
|
||||||
|
host = v
|
||||||
|
port = ""
|
||||||
|
}
|
||||||
|
host, err = idna.ToASCII(host)
|
||||||
|
if err != nil {
|
||||||
|
// Non-UTF-8? Not representable in Punycode, in any
|
||||||
|
// case.
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
if port == "" {
|
||||||
|
return host, nil
|
||||||
|
}
|
||||||
|
return net.JoinHostPort(host, port), nil
|
||||||
|
}
|
|
@ -0,0 +1,51 @@
|
||||||
|
#
|
||||||
|
# This Dockerfile builds a recent curl with HTTP/2 client support, using
|
||||||
|
# a recent nghttp2 build.
|
||||||
|
#
|
||||||
|
# See the Makefile for how to tag it. If Docker and that image is found, the
|
||||||
|
# Go tests use this curl binary for integration tests.
|
||||||
|
#
|
||||||
|
|
||||||
|
FROM ubuntu:trusty
|
||||||
|
|
||||||
|
RUN apt-get update && \
|
||||||
|
apt-get upgrade -y && \
|
||||||
|
apt-get install -y git-core build-essential wget
|
||||||
|
|
||||||
|
RUN apt-get install -y --no-install-recommends \
|
||||||
|
autotools-dev libtool pkg-config zlib1g-dev \
|
||||||
|
libcunit1-dev libssl-dev libxml2-dev libevent-dev \
|
||||||
|
automake autoconf
|
||||||
|
|
||||||
|
# The list of packages nghttp2 recommends for h2load:
|
||||||
|
RUN apt-get install -y --no-install-recommends make binutils \
|
||||||
|
autoconf automake autotools-dev \
|
||||||
|
libtool pkg-config zlib1g-dev libcunit1-dev libssl-dev libxml2-dev \
|
||||||
|
libev-dev libevent-dev libjansson-dev libjemalloc-dev \
|
||||||
|
cython python3.4-dev python-setuptools
|
||||||
|
|
||||||
|
# Note: setting NGHTTP2_VER before the git clone, so an old git clone isn't cached:
|
||||||
|
ENV NGHTTP2_VER 895da9a
|
||||||
|
RUN cd /root && git clone https://github.com/tatsuhiro-t/nghttp2.git
|
||||||
|
|
||||||
|
WORKDIR /root/nghttp2
|
||||||
|
RUN git reset --hard $NGHTTP2_VER
|
||||||
|
RUN autoreconf -i
|
||||||
|
RUN automake
|
||||||
|
RUN autoconf
|
||||||
|
RUN ./configure
|
||||||
|
RUN make
|
||||||
|
RUN make install
|
||||||
|
|
||||||
|
WORKDIR /root
|
||||||
|
RUN wget http://curl.haxx.se/download/curl-7.45.0.tar.gz
|
||||||
|
RUN tar -zxvf curl-7.45.0.tar.gz
|
||||||
|
WORKDIR /root/curl-7.45.0
|
||||||
|
RUN ./configure --with-ssl --with-nghttp2=/usr/local
|
||||||
|
RUN make
|
||||||
|
RUN make install
|
||||||
|
RUN ldconfig
|
||||||
|
|
||||||
|
CMD ["-h"]
|
||||||
|
ENTRYPOINT ["/usr/local/bin/curl"]
|
||||||
|
|
|
@ -0,0 +1,3 @@
|
||||||
|
curlimage:
|
||||||
|
docker build -t gohttp2/curl .
|
||||||
|
|
|
@ -0,0 +1,20 @@
|
||||||
|
This is a work-in-progress HTTP/2 implementation for Go.
|
||||||
|
|
||||||
|
It will eventually live in the Go standard library and won't require
|
||||||
|
any changes to your code to use. It will just be automatic.
|
||||||
|
|
||||||
|
Status:
|
||||||
|
|
||||||
|
* The server support is pretty good. A few things are missing
|
||||||
|
but are being worked on.
|
||||||
|
* The client work has just started but shares a lot of code
|
||||||
|
is coming along much quicker.
|
||||||
|
|
||||||
|
Docs are at https://godoc.org/golang.org/x/net/http2
|
||||||
|
|
||||||
|
Demo test server at https://http2.golang.org/
|
||||||
|
|
||||||
|
Help & bug reports welcome!
|
||||||
|
|
||||||
|
Contributing: https://golang.org/doc/contribute.html
|
||||||
|
Bugs: https://golang.org/issue/new?title=x/net/http2:+
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue