mirror of
https://github.com/Luzifer/duplicity-backup.git
synced 2024-11-09 15:40:06 +00:00
Deps: Update dependencies
Signed-off-by: Knut Ahlers <knut@ahlers.me>
This commit is contained in:
parent
67397a6acb
commit
6275a80c3f
527 changed files with 372072 additions and 2780 deletions
151
Gopkg.lock
generated
151
Gopkg.lock
generated
|
@ -2,46 +2,64 @@
|
||||||
|
|
||||||
|
|
||||||
[[projects]]
|
[[projects]]
|
||||||
digest = "1:db2d1318c12a5e0df8a212bf3448f507075479fccad35dd14d4cfa50e2f4f599"
|
digest = "1:34e09c50d26400b5e70ade7e041edb75b6aea1b1af428a8d608772555e1bbf92"
|
||||||
name = "github.com/Luzifer/go_helpers"
|
name = "github.com/Luzifer/go_helpers"
|
||||||
packages = [
|
packages = [
|
||||||
"str",
|
"str",
|
||||||
"which",
|
"which",
|
||||||
]
|
]
|
||||||
pruneopts = "UT"
|
pruneopts = "NUT"
|
||||||
revision = "6abbbafaada02b63dd8f9e185921fd8b3c35b6c2"
|
revision = "c3bea85c97943065c31d13b2193a9ef8c8488fb2"
|
||||||
version = "v1.3.0"
|
version = "v2.8.0"
|
||||||
|
|
||||||
[[projects]]
|
[[projects]]
|
||||||
digest = "1:eb4501e44a0a8035e18fc922d237c7a1535f58c0a78ec2c9eb0a458a1a0379fa"
|
digest = "1:f6cc072a289a686fda22819d871cd1b0407640141b2f6616dfbab957c96bf6c3"
|
||||||
name = "github.com/Luzifer/rconfig"
|
name = "github.com/Luzifer/rconfig"
|
||||||
packages = ["."]
|
packages = ["."]
|
||||||
pruneopts = "UT"
|
pruneopts = "NUT"
|
||||||
revision = "26776536e61487fdffbd3ce87f827177a5903f98"
|
revision = "5b80190bff90ccb9899db31e45baac7b1bede03b"
|
||||||
|
version = "v2.2.0"
|
||||||
|
|
||||||
[[projects]]
|
[[projects]]
|
||||||
digest = "1:edc30b45e5ffe48e6405aaad2542011346c541a598d48ba67a275f73692f15bf"
|
digest = "1:5a23cd3a5496a0b2da7e3b348d14e77b11a210738c398200d7d6f04ea8cf3bd8"
|
||||||
name = "github.com/asaskevich/govalidator"
|
name = "github.com/asaskevich/govalidator"
|
||||||
packages = ["."]
|
packages = ["."]
|
||||||
pruneopts = "UT"
|
pruneopts = "NUT"
|
||||||
revision = "df81827fdd59d8b4fb93d8910b286ab7a3919520"
|
revision = "ccb8e960c48f04d6935e72476ae4a51028f9e22f"
|
||||||
|
version = "v9"
|
||||||
|
|
||||||
[[projects]]
|
[[projects]]
|
||||||
digest = "1:f048897d11e3e08c3ef8b257688e4dd44207ef16f10a9061bb39a5498b3d7d89"
|
branch = "master"
|
||||||
|
digest = "1:8d485687c449a7db7d4e8e1d6526036fe38e433232416010fe3970c929134073"
|
||||||
|
name = "github.com/hpcloud/tail"
|
||||||
|
packages = [
|
||||||
|
".",
|
||||||
|
"ratelimiter",
|
||||||
|
"util",
|
||||||
|
"watch",
|
||||||
|
"winfile",
|
||||||
|
]
|
||||||
|
pruneopts = "NUT"
|
||||||
|
revision = "a1dbeea552b7c8df4b542c66073e393de198a800"
|
||||||
|
|
||||||
|
[[projects]]
|
||||||
|
digest = "1:a4df73029d2c42fabcb6b41e327d2f87e685284ec03edf76921c267d9cfc9c23"
|
||||||
name = "github.com/mitchellh/go-homedir"
|
name = "github.com/mitchellh/go-homedir"
|
||||||
packages = ["."]
|
packages = ["."]
|
||||||
pruneopts = "UT"
|
pruneopts = "NUT"
|
||||||
revision = "981ab348d865cf048eb7d17e78ac7192632d8415"
|
revision = "ae18d6b8b3205b561c79e8e5f69bff09736185f4"
|
||||||
|
version = "v1.0.0"
|
||||||
|
|
||||||
[[projects]]
|
[[projects]]
|
||||||
digest = "1:61115f0d16dc54abaaacce64409a63f1745bf07d099f5c9d79ebdf247b460847"
|
branch = "master"
|
||||||
|
digest = "1:26217ee135b8157549e648efe97dff3282e4bd597a912d784db964df41067f29"
|
||||||
name = "github.com/nightlyone/lockfile"
|
name = "github.com/nightlyone/lockfile"
|
||||||
packages = ["."]
|
packages = ["."]
|
||||||
pruneopts = "UT"
|
pruneopts = "NUT"
|
||||||
revision = "b30dcbfa86e3a1eaa4e6622de2ce57be2c138c10"
|
revision = "0ad87eef1443f64d3d8c50da647e2b1552851124"
|
||||||
|
|
||||||
[[projects]]
|
[[projects]]
|
||||||
digest = "1:3725923e811d894e732664790100e0bdf58fc6bd67209103f685d2fc7a78b8df"
|
digest = "1:0be1cd4c73d5e22a30edcf32a18e9809a370a7a4a52c4f41a86070b34da93fef"
|
||||||
name = "github.com/onsi/ginkgo"
|
name = "github.com/onsi/ginkgo"
|
||||||
packages = [
|
packages = [
|
||||||
".",
|
".",
|
||||||
|
@ -52,25 +70,30 @@
|
||||||
"internal/leafnodes",
|
"internal/leafnodes",
|
||||||
"internal/remote",
|
"internal/remote",
|
||||||
"internal/spec",
|
"internal/spec",
|
||||||
|
"internal/spec_iterator",
|
||||||
"internal/specrunner",
|
"internal/specrunner",
|
||||||
"internal/suite",
|
"internal/suite",
|
||||||
"internal/testingtproxy",
|
"internal/testingtproxy",
|
||||||
"internal/writer",
|
"internal/writer",
|
||||||
"reporters",
|
"reporters",
|
||||||
"reporters/stenographer",
|
"reporters/stenographer",
|
||||||
|
"reporters/stenographer/support/go-colorable",
|
||||||
|
"reporters/stenographer/support/go-isatty",
|
||||||
"types",
|
"types",
|
||||||
]
|
]
|
||||||
pruneopts = "UT"
|
pruneopts = "NUT"
|
||||||
revision = "1b59c57df76ede42c08590546916e6a18685857d"
|
revision = "3774a09d95489ccaa16032e0770d08ea77ba6184"
|
||||||
|
version = "v1.6.0"
|
||||||
|
|
||||||
[[projects]]
|
[[projects]]
|
||||||
digest = "1:4560887a9c590c4141776e84258f10b99094580f629de5437a619dea8a9ebd04"
|
digest = "1:95f40a9db820078d1795c7ba2d476016aca05dc4267eaf6752a925e437cb351f"
|
||||||
name = "github.com/onsi/gomega"
|
name = "github.com/onsi/gomega"
|
||||||
packages = [
|
packages = [
|
||||||
".",
|
".",
|
||||||
"format",
|
"format",
|
||||||
"internal/assertion",
|
"internal/assertion",
|
||||||
"internal/asyncassertion",
|
"internal/asyncassertion",
|
||||||
|
"internal/oraclematcher",
|
||||||
"internal/testingtsupport",
|
"internal/testingtsupport",
|
||||||
"matchers",
|
"matchers",
|
||||||
"matchers/support/goraph/bipartitegraph",
|
"matchers/support/goraph/bipartitegraph",
|
||||||
|
@ -79,22 +102,94 @@
|
||||||
"matchers/support/goraph/util",
|
"matchers/support/goraph/util",
|
||||||
"types",
|
"types",
|
||||||
]
|
]
|
||||||
pruneopts = "UT"
|
pruneopts = "NUT"
|
||||||
revision = "6331bf5a5b5e7a832348789eb3cedff7a6917103"
|
revision = "7615b9433f86a8bdf29709bf288bc4fd0636a369"
|
||||||
|
version = "v1.4.2"
|
||||||
|
|
||||||
[[projects]]
|
[[projects]]
|
||||||
digest = "1:3bfae3cafaa74e12f74f2d1c648a74fba3ca0764d566cef503072e68b927c642"
|
digest = "1:9d8420bbf131d1618bde6530af37c3799340d3762cc47210c1d9532a4c3a2779"
|
||||||
name = "github.com/spf13/pflag"
|
name = "github.com/spf13/pflag"
|
||||||
packages = ["."]
|
packages = ["."]
|
||||||
pruneopts = "UT"
|
pruneopts = "NUT"
|
||||||
revision = "b084184666e02084b8ccb9b704bf0d79c466eb1d"
|
revision = "298182f68c66c05229eb03ac171abe6e309ee79a"
|
||||||
|
version = "v1.0.3"
|
||||||
|
|
||||||
[[projects]]
|
[[projects]]
|
||||||
digest = "1:c27797c5f42d349e2a604510822df7d037415aae58bf1e6fd35624eda757c0aa"
|
branch = "master"
|
||||||
|
digest = "1:5193d913046443e59093d66a97a40c51f4a5ea4ceba60f3b3ecf89694de5d16f"
|
||||||
|
name = "golang.org/x/net"
|
||||||
|
packages = [
|
||||||
|
"html",
|
||||||
|
"html/atom",
|
||||||
|
"html/charset",
|
||||||
|
]
|
||||||
|
pruneopts = "NUT"
|
||||||
|
revision = "146acd28ed5894421fb5aac80ca93bc1b1f46f87"
|
||||||
|
|
||||||
|
[[projects]]
|
||||||
|
branch = "master"
|
||||||
|
digest = "1:a9bc64e1206f3d5ce2ac71e397ddfe81065ac706929d34fcb3f0a48e74cd6e3b"
|
||||||
|
name = "golang.org/x/sys"
|
||||||
|
packages = ["unix"]
|
||||||
|
pruneopts = "NUT"
|
||||||
|
revision = "4497e2df6f9e69048a54498c7affbbec3294ad47"
|
||||||
|
|
||||||
|
[[projects]]
|
||||||
|
digest = "1:7c61a813b250ba8bb02bebb0382e6c3d00e04da2b577dc58985b86312fb89ffd"
|
||||||
|
name = "golang.org/x/text"
|
||||||
|
packages = [
|
||||||
|
"encoding",
|
||||||
|
"encoding/charmap",
|
||||||
|
"encoding/htmlindex",
|
||||||
|
"encoding/internal",
|
||||||
|
"encoding/internal/identifier",
|
||||||
|
"encoding/japanese",
|
||||||
|
"encoding/korean",
|
||||||
|
"encoding/simplifiedchinese",
|
||||||
|
"encoding/traditionalchinese",
|
||||||
|
"encoding/unicode",
|
||||||
|
"internal/gen",
|
||||||
|
"internal/tag",
|
||||||
|
"internal/utf8internal",
|
||||||
|
"language",
|
||||||
|
"runes",
|
||||||
|
"transform",
|
||||||
|
"unicode/cldr",
|
||||||
|
]
|
||||||
|
pruneopts = "NUT"
|
||||||
|
revision = "f21a4dfb5e38f5895301dc265a8def02365cc3d0"
|
||||||
|
version = "v0.3.0"
|
||||||
|
|
||||||
|
[[projects]]
|
||||||
|
digest = "1:1b91ae0dc69a41d4c2ed23ea5cffb721ea63f5037ca4b81e6d6771fbb8f45129"
|
||||||
|
name = "gopkg.in/fsnotify/fsnotify.v1"
|
||||||
|
packages = ["."]
|
||||||
|
pruneopts = "NUT"
|
||||||
|
revision = "c2828203cd70a50dcccfb2761f8b1f8ceef9a8e9"
|
||||||
|
version = "v1.4.7"
|
||||||
|
|
||||||
|
[[projects]]
|
||||||
|
digest = "1:63119d850a4a2aea8370aa74537a7d50923131ed207fc9aa12ddaf0b9ee94773"
|
||||||
|
name = "gopkg.in/tomb.v1"
|
||||||
|
packages = ["."]
|
||||||
|
pruneopts = "NUT"
|
||||||
|
revision = "c131134a1947e9afd9cecfe11f4c6dff0732ae58"
|
||||||
|
|
||||||
|
[[projects]]
|
||||||
|
branch = "v2"
|
||||||
|
digest = "1:1ab6db2d2bd353449c5d1e976ba7a92a0ece6e83aaab3e6674f8f2f1faebb85a"
|
||||||
|
name = "gopkg.in/validator.v2"
|
||||||
|
packages = ["."]
|
||||||
|
pruneopts = "NUT"
|
||||||
|
revision = "135c24b11c19e52befcae2ec3fca5d9b78c4e98e"
|
||||||
|
|
||||||
|
[[projects]]
|
||||||
|
digest = "1:7c95b35057a0ff2e19f707173cc1a947fa43a6eb5c4d300d196ece0334046082"
|
||||||
name = "gopkg.in/yaml.v2"
|
name = "gopkg.in/yaml.v2"
|
||||||
packages = ["."]
|
packages = ["."]
|
||||||
pruneopts = "UT"
|
pruneopts = "NUT"
|
||||||
revision = "53feefa2559fb8dfa8d81baad31be332c97d6c77"
|
revision = "5420a8b6744d3b0345ab293f6fcba19c978f1183"
|
||||||
|
version = "v2.2.1"
|
||||||
|
|
||||||
[solve-meta]
|
[solve-meta]
|
||||||
analyzer-name = "dep"
|
analyzer-name = "dep"
|
||||||
|
|
31
Gopkg.toml
31
Gopkg.toml
|
@ -27,8 +27,37 @@
|
||||||
|
|
||||||
[[constraint]]
|
[[constraint]]
|
||||||
name = "github.com/Luzifer/go_helpers"
|
name = "github.com/Luzifer/go_helpers"
|
||||||
version = "1.3.0"
|
version = "2.8.0"
|
||||||
|
|
||||||
|
[[constraint]]
|
||||||
|
name = "github.com/Luzifer/rconfig"
|
||||||
|
version = "2.2.0"
|
||||||
|
|
||||||
|
[[constraint]]
|
||||||
|
name = "github.com/asaskevich/govalidator"
|
||||||
|
version = "9.0.0"
|
||||||
|
|
||||||
|
[[constraint]]
|
||||||
|
name = "github.com/mitchellh/go-homedir"
|
||||||
|
version = "1.0.0"
|
||||||
|
|
||||||
|
[[constraint]]
|
||||||
|
branch = "master"
|
||||||
|
name = "github.com/nightlyone/lockfile"
|
||||||
|
|
||||||
|
[[constraint]]
|
||||||
|
name = "github.com/onsi/ginkgo"
|
||||||
|
version = "1.6.0"
|
||||||
|
|
||||||
|
[[constraint]]
|
||||||
|
name = "github.com/onsi/gomega"
|
||||||
|
version = "1.4.2"
|
||||||
|
|
||||||
|
[[constraint]]
|
||||||
|
name = "gopkg.in/yaml.v2"
|
||||||
|
version = "2.2.1"
|
||||||
|
|
||||||
[prune]
|
[prune]
|
||||||
|
non-go = true
|
||||||
go-tests = true
|
go-tests = true
|
||||||
unused-packages = true
|
unused-packages = true
|
||||||
|
|
202
vendor/github.com/Luzifer/go_helpers/LICENSE
generated
vendored
Normal file
202
vendor/github.com/Luzifer/go_helpers/LICENSE
generated
vendored
Normal file
|
@ -0,0 +1,202 @@
|
||||||
|
Apache License
|
||||||
|
Version 2.0, January 2004
|
||||||
|
http://www.apache.org/licenses/
|
||||||
|
|
||||||
|
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
||||||
|
|
||||||
|
1. Definitions.
|
||||||
|
|
||||||
|
"License" shall mean the terms and conditions for use, reproduction,
|
||||||
|
and distribution as defined by Sections 1 through 9 of this document.
|
||||||
|
|
||||||
|
"Licensor" shall mean the copyright owner or entity authorized by
|
||||||
|
the copyright owner that is granting the License.
|
||||||
|
|
||||||
|
"Legal Entity" shall mean the union of the acting entity and all
|
||||||
|
other entities that control, are controlled by, or are under common
|
||||||
|
control with that entity. For the purposes of this definition,
|
||||||
|
"control" means (i) the power, direct or indirect, to cause the
|
||||||
|
direction or management of such entity, whether by contract or
|
||||||
|
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
||||||
|
outstanding shares, or (iii) beneficial ownership of such entity.
|
||||||
|
|
||||||
|
"You" (or "Your") shall mean an individual or Legal Entity
|
||||||
|
exercising permissions granted by this License.
|
||||||
|
|
||||||
|
"Source" form shall mean the preferred form for making modifications,
|
||||||
|
including but not limited to software source code, documentation
|
||||||
|
source, and configuration files.
|
||||||
|
|
||||||
|
"Object" form shall mean any form resulting from mechanical
|
||||||
|
transformation or translation of a Source form, including but
|
||||||
|
not limited to compiled object code, generated documentation,
|
||||||
|
and conversions to other media types.
|
||||||
|
|
||||||
|
"Work" shall mean the work of authorship, whether in Source or
|
||||||
|
Object form, made available under the License, as indicated by a
|
||||||
|
copyright notice that is included in or attached to the work
|
||||||
|
(an example is provided in the Appendix below).
|
||||||
|
|
||||||
|
"Derivative Works" shall mean any work, whether in Source or Object
|
||||||
|
form, that is based on (or derived from) the Work and for which the
|
||||||
|
editorial revisions, annotations, elaborations, or other modifications
|
||||||
|
represent, as a whole, an original work of authorship. For the purposes
|
||||||
|
of this License, Derivative Works shall not include works that remain
|
||||||
|
separable from, or merely link (or bind by name) to the interfaces of,
|
||||||
|
the Work and Derivative Works thereof.
|
||||||
|
|
||||||
|
"Contribution" shall mean any work of authorship, including
|
||||||
|
the original version of the Work and any modifications or additions
|
||||||
|
to that Work or Derivative Works thereof, that is intentionally
|
||||||
|
submitted to Licensor for inclusion in the Work by the copyright owner
|
||||||
|
or by an individual or Legal Entity authorized to submit on behalf of
|
||||||
|
the copyright owner. For the purposes of this definition, "submitted"
|
||||||
|
means any form of electronic, verbal, or written communication sent
|
||||||
|
to the Licensor or its representatives, including but not limited to
|
||||||
|
communication on electronic mailing lists, source code control systems,
|
||||||
|
and issue tracking systems that are managed by, or on behalf of, the
|
||||||
|
Licensor for the purpose of discussing and improving the Work, but
|
||||||
|
excluding communication that is conspicuously marked or otherwise
|
||||||
|
designated in writing by the copyright owner as "Not a Contribution."
|
||||||
|
|
||||||
|
"Contributor" shall mean Licensor and any individual or Legal Entity
|
||||||
|
on behalf of whom a Contribution has been received by Licensor and
|
||||||
|
subsequently incorporated within the Work.
|
||||||
|
|
||||||
|
2. Grant of Copyright License. Subject to the terms and conditions of
|
||||||
|
this License, each Contributor hereby grants to You a perpetual,
|
||||||
|
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||||
|
copyright license to reproduce, prepare Derivative Works of,
|
||||||
|
publicly display, publicly perform, sublicense, and distribute the
|
||||||
|
Work and such Derivative Works in Source or Object form.
|
||||||
|
|
||||||
|
3. Grant of Patent License. Subject to the terms and conditions of
|
||||||
|
this License, each Contributor hereby grants to You a perpetual,
|
||||||
|
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||||
|
(except as stated in this section) patent license to make, have made,
|
||||||
|
use, offer to sell, sell, import, and otherwise transfer the Work,
|
||||||
|
where such license applies only to those patent claims licensable
|
||||||
|
by such Contributor that are necessarily infringed by their
|
||||||
|
Contribution(s) alone or by combination of their Contribution(s)
|
||||||
|
with the Work to which such Contribution(s) was submitted. If You
|
||||||
|
institute patent litigation against any entity (including a
|
||||||
|
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
||||||
|
or a Contribution incorporated within the Work constitutes direct
|
||||||
|
or contributory patent infringement, then any patent licenses
|
||||||
|
granted to You under this License for that Work shall terminate
|
||||||
|
as of the date such litigation is filed.
|
||||||
|
|
||||||
|
4. Redistribution. You may reproduce and distribute copies of the
|
||||||
|
Work or Derivative Works thereof in any medium, with or without
|
||||||
|
modifications, and in Source or Object form, provided that You
|
||||||
|
meet the following conditions:
|
||||||
|
|
||||||
|
(a) You must give any other recipients of the Work or
|
||||||
|
Derivative Works a copy of this License; and
|
||||||
|
|
||||||
|
(b) You must cause any modified files to carry prominent notices
|
||||||
|
stating that You changed the files; and
|
||||||
|
|
||||||
|
(c) You must retain, in the Source form of any Derivative Works
|
||||||
|
that You distribute, all copyright, patent, trademark, and
|
||||||
|
attribution notices from the Source form of the Work,
|
||||||
|
excluding those notices that do not pertain to any part of
|
||||||
|
the Derivative Works; and
|
||||||
|
|
||||||
|
(d) If the Work includes a "NOTICE" text file as part of its
|
||||||
|
distribution, then any Derivative Works that You distribute must
|
||||||
|
include a readable copy of the attribution notices contained
|
||||||
|
within such NOTICE file, excluding those notices that do not
|
||||||
|
pertain to any part of the Derivative Works, in at least one
|
||||||
|
of the following places: within a NOTICE text file distributed
|
||||||
|
as part of the Derivative Works; within the Source form or
|
||||||
|
documentation, if provided along with the Derivative Works; or,
|
||||||
|
within a display generated by the Derivative Works, if and
|
||||||
|
wherever such third-party notices normally appear. The contents
|
||||||
|
of the NOTICE file are for informational purposes only and
|
||||||
|
do not modify the License. You may add Your own attribution
|
||||||
|
notices within Derivative Works that You distribute, alongside
|
||||||
|
or as an addendum to the NOTICE text from the Work, provided
|
||||||
|
that such additional attribution notices cannot be construed
|
||||||
|
as modifying the License.
|
||||||
|
|
||||||
|
You may add Your own copyright statement to Your modifications and
|
||||||
|
may provide additional or different license terms and conditions
|
||||||
|
for use, reproduction, or distribution of Your modifications, or
|
||||||
|
for any such Derivative Works as a whole, provided Your use,
|
||||||
|
reproduction, and distribution of the Work otherwise complies with
|
||||||
|
the conditions stated in this License.
|
||||||
|
|
||||||
|
5. Submission of Contributions. Unless You explicitly state otherwise,
|
||||||
|
any Contribution intentionally submitted for inclusion in the Work
|
||||||
|
by You to the Licensor shall be under the terms and conditions of
|
||||||
|
this License, without any additional terms or conditions.
|
||||||
|
Notwithstanding the above, nothing herein shall supersede or modify
|
||||||
|
the terms of any separate license agreement you may have executed
|
||||||
|
with Licensor regarding such Contributions.
|
||||||
|
|
||||||
|
6. Trademarks. This License does not grant permission to use the trade
|
||||||
|
names, trademarks, service marks, or product names of the Licensor,
|
||||||
|
except as required for reasonable and customary use in describing the
|
||||||
|
origin of the Work and reproducing the content of the NOTICE file.
|
||||||
|
|
||||||
|
7. Disclaimer of Warranty. Unless required by applicable law or
|
||||||
|
agreed to in writing, Licensor provides the Work (and each
|
||||||
|
Contributor provides its Contributions) on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||||
|
implied, including, without limitation, any warranties or conditions
|
||||||
|
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
||||||
|
PARTICULAR PURPOSE. You are solely responsible for determining the
|
||||||
|
appropriateness of using or redistributing the Work and assume any
|
||||||
|
risks associated with Your exercise of permissions under this License.
|
||||||
|
|
||||||
|
8. Limitation of Liability. In no event and under no legal theory,
|
||||||
|
whether in tort (including negligence), contract, or otherwise,
|
||||||
|
unless required by applicable law (such as deliberate and grossly
|
||||||
|
negligent acts) or agreed to in writing, shall any Contributor be
|
||||||
|
liable to You for damages, including any direct, indirect, special,
|
||||||
|
incidental, or consequential damages of any character arising as a
|
||||||
|
result of this License or out of the use or inability to use the
|
||||||
|
Work (including but not limited to damages for loss of goodwill,
|
||||||
|
work stoppage, computer failure or malfunction, or any and all
|
||||||
|
other commercial damages or losses), even if such Contributor
|
||||||
|
has been advised of the possibility of such damages.
|
||||||
|
|
||||||
|
9. Accepting Warranty or Additional Liability. While redistributing
|
||||||
|
the Work or Derivative Works thereof, You may choose to offer,
|
||||||
|
and charge a fee for, acceptance of support, warranty, indemnity,
|
||||||
|
or other liability obligations and/or rights consistent with this
|
||||||
|
License. However, in accepting such obligations, You may act only
|
||||||
|
on Your own behalf and on Your sole responsibility, not on behalf
|
||||||
|
of any other Contributor, and only if You agree to indemnify,
|
||||||
|
defend, and hold each Contributor harmless for any liability
|
||||||
|
incurred by, or claims asserted against, such Contributor by reason
|
||||||
|
of your accepting any such warranty or additional liability.
|
||||||
|
|
||||||
|
END OF TERMS AND CONDITIONS
|
||||||
|
|
||||||
|
APPENDIX: How to apply the Apache License to your work.
|
||||||
|
|
||||||
|
To apply the Apache License to your work, attach the following
|
||||||
|
boilerplate notice, with the fields enclosed by brackets "[]"
|
||||||
|
replaced with your own identifying information. (Don't include
|
||||||
|
the brackets!) The text should be enclosed in the appropriate
|
||||||
|
comment syntax for the file format. We also recommend that a
|
||||||
|
file or class name and description of purpose be included on the
|
||||||
|
same "printed page" as the copyright notice for easier
|
||||||
|
identification within third-party archives.
|
||||||
|
|
||||||
|
Copyright 2016- Knut Ahlers <knut@ahlers.me>
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
|
8
vendor/github.com/Luzifer/rconfig/.travis.yml
generated
vendored
8
vendor/github.com/Luzifer/rconfig/.travis.yml
generated
vendored
|
@ -1,8 +0,0 @@
|
||||||
language: go
|
|
||||||
|
|
||||||
go:
|
|
||||||
- 1.4
|
|
||||||
- 1.5
|
|
||||||
- tip
|
|
||||||
|
|
||||||
script: go test -v -race -cover ./...
|
|
209
vendor/github.com/Luzifer/rconfig/LICENSE
generated
vendored
209
vendor/github.com/Luzifer/rconfig/LICENSE
generated
vendored
|
@ -1,13 +1,202 @@
|
||||||
Copyright 2015 Knut Ahlers <knut@ahlers.me>
|
Apache License
|
||||||
|
Version 2.0, January 2004
|
||||||
|
http://www.apache.org/licenses/
|
||||||
|
|
||||||
Licensed under the Apache License, Version 2.0 (the "License");
|
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
||||||
you may not use this file except in compliance with the License.
|
|
||||||
You may obtain a copy of the License at
|
|
||||||
|
|
||||||
http://www.apache.org/licenses/LICENSE-2.0
|
1. Definitions.
|
||||||
|
|
||||||
|
"License" shall mean the terms and conditions for use, reproduction,
|
||||||
|
and distribution as defined by Sections 1 through 9 of this document.
|
||||||
|
|
||||||
|
"Licensor" shall mean the copyright owner or entity authorized by
|
||||||
|
the copyright owner that is granting the License.
|
||||||
|
|
||||||
|
"Legal Entity" shall mean the union of the acting entity and all
|
||||||
|
other entities that control, are controlled by, or are under common
|
||||||
|
control with that entity. For the purposes of this definition,
|
||||||
|
"control" means (i) the power, direct or indirect, to cause the
|
||||||
|
direction or management of such entity, whether by contract or
|
||||||
|
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
||||||
|
outstanding shares, or (iii) beneficial ownership of such entity.
|
||||||
|
|
||||||
|
"You" (or "Your") shall mean an individual or Legal Entity
|
||||||
|
exercising permissions granted by this License.
|
||||||
|
|
||||||
|
"Source" form shall mean the preferred form for making modifications,
|
||||||
|
including but not limited to software source code, documentation
|
||||||
|
source, and configuration files.
|
||||||
|
|
||||||
|
"Object" form shall mean any form resulting from mechanical
|
||||||
|
transformation or translation of a Source form, including but
|
||||||
|
not limited to compiled object code, generated documentation,
|
||||||
|
and conversions to other media types.
|
||||||
|
|
||||||
|
"Work" shall mean the work of authorship, whether in Source or
|
||||||
|
Object form, made available under the License, as indicated by a
|
||||||
|
copyright notice that is included in or attached to the work
|
||||||
|
(an example is provided in the Appendix below).
|
||||||
|
|
||||||
|
"Derivative Works" shall mean any work, whether in Source or Object
|
||||||
|
form, that is based on (or derived from) the Work and for which the
|
||||||
|
editorial revisions, annotations, elaborations, or other modifications
|
||||||
|
represent, as a whole, an original work of authorship. For the purposes
|
||||||
|
of this License, Derivative Works shall not include works that remain
|
||||||
|
separable from, or merely link (or bind by name) to the interfaces of,
|
||||||
|
the Work and Derivative Works thereof.
|
||||||
|
|
||||||
|
"Contribution" shall mean any work of authorship, including
|
||||||
|
the original version of the Work and any modifications or additions
|
||||||
|
to that Work or Derivative Works thereof, that is intentionally
|
||||||
|
submitted to Licensor for inclusion in the Work by the copyright owner
|
||||||
|
or by an individual or Legal Entity authorized to submit on behalf of
|
||||||
|
the copyright owner. For the purposes of this definition, "submitted"
|
||||||
|
means any form of electronic, verbal, or written communication sent
|
||||||
|
to the Licensor or its representatives, including but not limited to
|
||||||
|
communication on electronic mailing lists, source code control systems,
|
||||||
|
and issue tracking systems that are managed by, or on behalf of, the
|
||||||
|
Licensor for the purpose of discussing and improving the Work, but
|
||||||
|
excluding communication that is conspicuously marked or otherwise
|
||||||
|
designated in writing by the copyright owner as "Not a Contribution."
|
||||||
|
|
||||||
|
"Contributor" shall mean Licensor and any individual or Legal Entity
|
||||||
|
on behalf of whom a Contribution has been received by Licensor and
|
||||||
|
subsequently incorporated within the Work.
|
||||||
|
|
||||||
|
2. Grant of Copyright License. Subject to the terms and conditions of
|
||||||
|
this License, each Contributor hereby grants to You a perpetual,
|
||||||
|
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||||
|
copyright license to reproduce, prepare Derivative Works of,
|
||||||
|
publicly display, publicly perform, sublicense, and distribute the
|
||||||
|
Work and such Derivative Works in Source or Object form.
|
||||||
|
|
||||||
|
3. Grant of Patent License. Subject to the terms and conditions of
|
||||||
|
this License, each Contributor hereby grants to You a perpetual,
|
||||||
|
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||||
|
(except as stated in this section) patent license to make, have made,
|
||||||
|
use, offer to sell, sell, import, and otherwise transfer the Work,
|
||||||
|
where such license applies only to those patent claims licensable
|
||||||
|
by such Contributor that are necessarily infringed by their
|
||||||
|
Contribution(s) alone or by combination of their Contribution(s)
|
||||||
|
with the Work to which such Contribution(s) was submitted. If You
|
||||||
|
institute patent litigation against any entity (including a
|
||||||
|
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
||||||
|
or a Contribution incorporated within the Work constitutes direct
|
||||||
|
or contributory patent infringement, then any patent licenses
|
||||||
|
granted to You under this License for that Work shall terminate
|
||||||
|
as of the date such litigation is filed.
|
||||||
|
|
||||||
|
4. Redistribution. You may reproduce and distribute copies of the
|
||||||
|
Work or Derivative Works thereof in any medium, with or without
|
||||||
|
modifications, and in Source or Object form, provided that You
|
||||||
|
meet the following conditions:
|
||||||
|
|
||||||
|
(a) You must give any other recipients of the Work or
|
||||||
|
Derivative Works a copy of this License; and
|
||||||
|
|
||||||
|
(b) You must cause any modified files to carry prominent notices
|
||||||
|
stating that You changed the files; and
|
||||||
|
|
||||||
|
(c) You must retain, in the Source form of any Derivative Works
|
||||||
|
that You distribute, all copyright, patent, trademark, and
|
||||||
|
attribution notices from the Source form of the Work,
|
||||||
|
excluding those notices that do not pertain to any part of
|
||||||
|
the Derivative Works; and
|
||||||
|
|
||||||
|
(d) If the Work includes a "NOTICE" text file as part of its
|
||||||
|
distribution, then any Derivative Works that You distribute must
|
||||||
|
include a readable copy of the attribution notices contained
|
||||||
|
within such NOTICE file, excluding those notices that do not
|
||||||
|
pertain to any part of the Derivative Works, in at least one
|
||||||
|
of the following places: within a NOTICE text file distributed
|
||||||
|
as part of the Derivative Works; within the Source form or
|
||||||
|
documentation, if provided along with the Derivative Works; or,
|
||||||
|
within a display generated by the Derivative Works, if and
|
||||||
|
wherever such third-party notices normally appear. The contents
|
||||||
|
of the NOTICE file are for informational purposes only and
|
||||||
|
do not modify the License. You may add Your own attribution
|
||||||
|
notices within Derivative Works that You distribute, alongside
|
||||||
|
or as an addendum to the NOTICE text from the Work, provided
|
||||||
|
that such additional attribution notices cannot be construed
|
||||||
|
as modifying the License.
|
||||||
|
|
||||||
|
You may add Your own copyright statement to Your modifications and
|
||||||
|
may provide additional or different license terms and conditions
|
||||||
|
for use, reproduction, or distribution of Your modifications, or
|
||||||
|
for any such Derivative Works as a whole, provided Your use,
|
||||||
|
reproduction, and distribution of the Work otherwise complies with
|
||||||
|
the conditions stated in this License.
|
||||||
|
|
||||||
|
5. Submission of Contributions. Unless You explicitly state otherwise,
|
||||||
|
any Contribution intentionally submitted for inclusion in the Work
|
||||||
|
by You to the Licensor shall be under the terms and conditions of
|
||||||
|
this License, without any additional terms or conditions.
|
||||||
|
Notwithstanding the above, nothing herein shall supersede or modify
|
||||||
|
the terms of any separate license agreement you may have executed
|
||||||
|
with Licensor regarding such Contributions.
|
||||||
|
|
||||||
|
6. Trademarks. This License does not grant permission to use the trade
|
||||||
|
names, trademarks, service marks, or product names of the Licensor,
|
||||||
|
except as required for reasonable and customary use in describing the
|
||||||
|
origin of the Work and reproducing the content of the NOTICE file.
|
||||||
|
|
||||||
|
7. Disclaimer of Warranty. Unless required by applicable law or
|
||||||
|
agreed to in writing, Licensor provides the Work (and each
|
||||||
|
Contributor provides its Contributions) on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||||
|
implied, including, without limitation, any warranties or conditions
|
||||||
|
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
||||||
|
PARTICULAR PURPOSE. You are solely responsible for determining the
|
||||||
|
appropriateness of using or redistributing the Work and assume any
|
||||||
|
risks associated with Your exercise of permissions under this License.
|
||||||
|
|
||||||
|
8. Limitation of Liability. In no event and under no legal theory,
|
||||||
|
whether in tort (including negligence), contract, or otherwise,
|
||||||
|
unless required by applicable law (such as deliberate and grossly
|
||||||
|
negligent acts) or agreed to in writing, shall any Contributor be
|
||||||
|
liable to You for damages, including any direct, indirect, special,
|
||||||
|
incidental, or consequential damages of any character arising as a
|
||||||
|
result of this License or out of the use or inability to use the
|
||||||
|
Work (including but not limited to damages for loss of goodwill,
|
||||||
|
work stoppage, computer failure or malfunction, or any and all
|
||||||
|
other commercial damages or losses), even if such Contributor
|
||||||
|
has been advised of the possibility of such damages.
|
||||||
|
|
||||||
|
9. Accepting Warranty or Additional Liability. While redistributing
|
||||||
|
the Work or Derivative Works thereof, You may choose to offer,
|
||||||
|
and charge a fee for, acceptance of support, warranty, indemnity,
|
||||||
|
or other liability obligations and/or rights consistent with this
|
||||||
|
License. However, in accepting such obligations, You may act only
|
||||||
|
on Your own behalf and on Your sole responsibility, not on behalf
|
||||||
|
of any other Contributor, and only if You agree to indemnify,
|
||||||
|
defend, and hold each Contributor harmless for any liability
|
||||||
|
incurred by, or claims asserted against, such Contributor by reason
|
||||||
|
of your accepting any such warranty or additional liability.
|
||||||
|
|
||||||
|
END OF TERMS AND CONDITIONS
|
||||||
|
|
||||||
|
APPENDIX: How to apply the Apache License to your work.
|
||||||
|
|
||||||
|
To apply the Apache License to your work, attach the following
|
||||||
|
boilerplate notice, with the fields enclosed by brackets "[]"
|
||||||
|
replaced with your own identifying information. (Don't include
|
||||||
|
the brackets!) The text should be enclosed in the appropriate
|
||||||
|
comment syntax for the file format. We also recommend that a
|
||||||
|
file or class name and description of purpose be included on the
|
||||||
|
same "printed page" as the copyright notice for easier
|
||||||
|
identification within third-party archives.
|
||||||
|
|
||||||
|
Copyright 2015- Knut Ahlers <knut@ahlers.me>
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
|
||||||
Unless required by applicable law or agreed to in writing, software
|
|
||||||
distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
See the License for the specific language governing permissions and
|
|
||||||
limitations under the License.
|
|
||||||
|
|
94
vendor/github.com/Luzifer/rconfig/README.md
generated
vendored
94
vendor/github.com/Luzifer/rconfig/README.md
generated
vendored
|
@ -1,94 +0,0 @@
|
||||||
[![Build Status](https://travis-ci.org/Luzifer/rconfig.svg?branch=master)](https://travis-ci.org/Luzifer/rconfig)
|
|
||||||
[![License: Apache v2.0](https://badge.luzifer.io/v1/badge?color=5d79b5&title=license&text=Apache+v2.0)](http://www.apache.org/licenses/LICENSE-2.0)
|
|
||||||
[![Documentation](https://badge.luzifer.io/v1/badge?title=godoc&text=reference)](https://godoc.org/github.com/Luzifer/rconfig)
|
|
||||||
[![Go Report](http://goreportcard.com/badge/Luzifer/rconfig)](http://goreportcard.com/report/Luzifer/rconfig)
|
|
||||||
|
|
||||||
## Description
|
|
||||||
|
|
||||||
> Package rconfig implements a CLI configuration reader with struct-embedded defaults, environment variables and posix compatible flag parsing using the [pflag](https://github.com/spf13/pflag) library.
|
|
||||||
|
|
||||||
## Installation
|
|
||||||
|
|
||||||
Install by running:
|
|
||||||
|
|
||||||
```
|
|
||||||
go get -u github.com/Luzifer/rconfig
|
|
||||||
```
|
|
||||||
|
|
||||||
OR fetch a specific version:
|
|
||||||
|
|
||||||
```
|
|
||||||
go get -u gopkg.in/luzifer/rconfig.v1
|
|
||||||
```
|
|
||||||
|
|
||||||
Run tests by running:
|
|
||||||
|
|
||||||
```
|
|
||||||
go test -v -race -cover github.com/Luzifer/rconfig
|
|
||||||
```
|
|
||||||
|
|
||||||
## Usage
|
|
||||||
|
|
||||||
As a first step define a struct holding your configuration:
|
|
||||||
|
|
||||||
```go
|
|
||||||
type config struct {
|
|
||||||
Username string `default:"unknown" flag:"user" description:"Your name"`
|
|
||||||
Details struct {
|
|
||||||
Age int `default:"25" flag:"age" env:"age" description:"Your age"`
|
|
||||||
}
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
Next create an instance of that struct and let `rconfig` fill that config:
|
|
||||||
|
|
||||||
```go
|
|
||||||
var cfg config
|
|
||||||
func init() {
|
|
||||||
cfg = config{}
|
|
||||||
rconfig.Parse(&cfg)
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
You're ready to access your configuration:
|
|
||||||
|
|
||||||
```go
|
|
||||||
func main() {
|
|
||||||
fmt.Printf("Hello %s, happy birthday for your %dth birthday.",
|
|
||||||
cfg.Username,
|
|
||||||
cfg.Details.Age)
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
### Provide variable defaults by using a file
|
|
||||||
|
|
||||||
Given you have a file `~/.myapp.yml` containing some secrets or usernames (for the example below username is assumed to be "luzifer") as a default configuration for your application you can use this source code to load the defaults from that file using the `vardefault` tag in your configuration struct.
|
|
||||||
|
|
||||||
The order of the directives (lower number = higher precedence):
|
|
||||||
|
|
||||||
1. Flags provided in command line
|
|
||||||
1. Environment variables
|
|
||||||
1. Variable defaults (`vardefault` tag in the struct)
|
|
||||||
1. `default` tag in the struct
|
|
||||||
|
|
||||||
```go
|
|
||||||
type config struct {
|
|
||||||
Username string `vardefault:"username" flag:"username" description:"Your username"`
|
|
||||||
}
|
|
||||||
|
|
||||||
var cfg = config{}
|
|
||||||
|
|
||||||
func init() {
|
|
||||||
rconfig.SetVariableDefaults(rconfig.VarDefaultsFromYAMLFile("~/.myapp.yml"))
|
|
||||||
rconfig.Parse(&cfg)
|
|
||||||
}
|
|
||||||
|
|
||||||
func main() {
|
|
||||||
fmt.Printf("Username = %s", cfg.Username)
|
|
||||||
// Output: Username = luzifer
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
## More info
|
|
||||||
|
|
||||||
You can see the full reference documentation of the rconfig package [at godoc.org](https://godoc.org/github.com/Luzifer/rconfig), or through go's standard documentation system by running `godoc -http=:6060` and browsing to [http://localhost:6060/pkg/github.com/Luzifer/rconfig](http://localhost:6060/pkg/github.com/Luzifer/rconfig) after installation.
|
|
64
vendor/github.com/Luzifer/rconfig/autoenv.go
generated
vendored
Normal file
64
vendor/github.com/Luzifer/rconfig/autoenv.go
generated
vendored
Normal file
|
@ -0,0 +1,64 @@
|
||||||
|
package rconfig
|
||||||
|
|
||||||
|
import "strings"
|
||||||
|
|
||||||
|
type characterClass [2]rune
|
||||||
|
|
||||||
|
func (c characterClass) Contains(r rune) bool {
|
||||||
|
return c[0] <= r && c[1] >= r
|
||||||
|
}
|
||||||
|
|
||||||
|
type characterClasses []characterClass
|
||||||
|
|
||||||
|
func (c characterClasses) Contains(r rune) bool {
|
||||||
|
for _, cc := range c {
|
||||||
|
if cc.Contains(r) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
charGroupUpperLetter = characterClass{'A', 'Z'}
|
||||||
|
charGroupLowerLetter = characterClass{'a', 'z'}
|
||||||
|
charGroupNumber = characterClass{'0', '9'}
|
||||||
|
charGroupLowerNumber = characterClasses{charGroupLowerLetter, charGroupNumber}
|
||||||
|
)
|
||||||
|
|
||||||
|
func deriveEnvVarName(s string) string {
|
||||||
|
var (
|
||||||
|
words []string
|
||||||
|
word []rune
|
||||||
|
)
|
||||||
|
|
||||||
|
for _, l := range s {
|
||||||
|
switch {
|
||||||
|
case charGroupUpperLetter.Contains(l):
|
||||||
|
if len(word) > 0 && charGroupLowerNumber.Contains(word[len(word)-1]) {
|
||||||
|
words = append(words, string(word))
|
||||||
|
word = []rune{}
|
||||||
|
}
|
||||||
|
word = append(word, l)
|
||||||
|
|
||||||
|
case charGroupLowerLetter.Contains(l):
|
||||||
|
if len(word) > 1 && charGroupUpperLetter.Contains(word[len(word)-1]) {
|
||||||
|
words = append(words, string(word[0:len(word)-1]))
|
||||||
|
word = word[len(word)-1:]
|
||||||
|
}
|
||||||
|
word = append(word, l)
|
||||||
|
|
||||||
|
case charGroupNumber.Contains(l):
|
||||||
|
word = append(word, l)
|
||||||
|
|
||||||
|
default:
|
||||||
|
if len(word) > 0 {
|
||||||
|
words = append(words, string(word))
|
||||||
|
}
|
||||||
|
word = []rune{}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
words = append(words, string(word))
|
||||||
|
|
||||||
|
return strings.ToUpper(strings.Join(words, "_"))
|
||||||
|
}
|
172
vendor/github.com/Luzifer/rconfig/config.go
generated
vendored
172
vendor/github.com/Luzifer/rconfig/config.go
generated
vendored
|
@ -10,13 +10,31 @@ import (
|
||||||
"reflect"
|
"reflect"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/spf13/pflag"
|
"github.com/spf13/pflag"
|
||||||
|
validator "gopkg.in/validator.v2"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
type afterFunc func() error
|
||||||
|
|
||||||
var (
|
var (
|
||||||
|
autoEnv bool
|
||||||
fs *pflag.FlagSet
|
fs *pflag.FlagSet
|
||||||
variableDefaults map[string]string
|
variableDefaults map[string]string
|
||||||
|
|
||||||
|
timeParserFormats = []string{
|
||||||
|
// Default constants
|
||||||
|
time.RFC3339Nano, time.RFC3339,
|
||||||
|
time.RFC1123Z, time.RFC1123,
|
||||||
|
time.RFC822Z, time.RFC822,
|
||||||
|
time.RFC850, time.RubyDate, time.UnixDate, time.ANSIC,
|
||||||
|
"2006-01-02 15:04:05.999999999 -0700 MST",
|
||||||
|
// More uncommon time formats
|
||||||
|
"2006-01-02 15:04:05", "2006-01-02 15:04:05Z07:00", // Simplified ISO time format
|
||||||
|
"01/02/2006 15:04:05", "01/02/2006 15:04:05Z07:00", // US time format
|
||||||
|
"02.01.2006 15:04:05", "02.01.2006 15:04:05Z07:00", // DE time format
|
||||||
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
|
@ -44,11 +62,32 @@ func Parse(config interface{}) error {
|
||||||
return parse(config, nil)
|
return parse(config, nil)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ParseAndValidate works exactly like Parse but implements an additional run of
|
||||||
|
// the go-validator package on the configuration struct. Therefore additonal struct
|
||||||
|
// tags are supported like described in the readme file of the go-validator package:
|
||||||
|
//
|
||||||
|
// https://github.com/go-validator/validator/tree/v2#usage
|
||||||
|
func ParseAndValidate(config interface{}) error {
|
||||||
|
return parseAndValidate(config, nil)
|
||||||
|
}
|
||||||
|
|
||||||
// Args returns the non-flag command-line arguments.
|
// Args returns the non-flag command-line arguments.
|
||||||
func Args() []string {
|
func Args() []string {
|
||||||
return fs.Args()
|
return fs.Args()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// AddTimeParserFormats adds custom formats to parse time.Time fields
|
||||||
|
func AddTimeParserFormats(f ...string) {
|
||||||
|
timeParserFormats = append(timeParserFormats, f...)
|
||||||
|
}
|
||||||
|
|
||||||
|
// AutoEnv enables or disables automated env variable guessing. If no `env` struct
|
||||||
|
// tag was set and AutoEnv is enabled the env variable name is derived from the
|
||||||
|
// name of the field: `MyFieldName` will get `MY_FIELD_NAME`
|
||||||
|
func AutoEnv(enable bool) {
|
||||||
|
autoEnv = enable
|
||||||
|
}
|
||||||
|
|
||||||
// Usage prints a basic usage with the corresponding defaults for the flags to
|
// Usage prints a basic usage with the corresponding defaults for the flags to
|
||||||
// os.Stdout. The defaults are derived from the `default` struct-tag and the ENV.
|
// os.Stdout. The defaults are derived from the `default` struct-tag and the ENV.
|
||||||
func Usage() {
|
func Usage() {
|
||||||
|
@ -64,28 +103,51 @@ func SetVariableDefaults(defaults map[string]string) {
|
||||||
variableDefaults = defaults
|
variableDefaults = defaults
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func parseAndValidate(in interface{}, args []string) error {
|
||||||
|
if err := parse(in, args); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
return validator.Validate(in)
|
||||||
|
}
|
||||||
|
|
||||||
func parse(in interface{}, args []string) error {
|
func parse(in interface{}, args []string) error {
|
||||||
if args == nil {
|
if args == nil {
|
||||||
args = os.Args
|
args = os.Args
|
||||||
}
|
}
|
||||||
|
|
||||||
fs = pflag.NewFlagSet(os.Args[0], pflag.ExitOnError)
|
fs = pflag.NewFlagSet(os.Args[0], pflag.ExitOnError)
|
||||||
if err := execTags(in, fs); err != nil {
|
afterFuncs, err := execTags(in, fs)
|
||||||
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
return fs.Parse(args)
|
if err := fs.Parse(args); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
if afterFuncs != nil {
|
||||||
|
for _, f := range afterFuncs {
|
||||||
|
if err := f(); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func execTags(in interface{}, fs *pflag.FlagSet) error {
|
func execTags(in interface{}, fs *pflag.FlagSet) ([]afterFunc, error) {
|
||||||
if reflect.TypeOf(in).Kind() != reflect.Ptr {
|
if reflect.TypeOf(in).Kind() != reflect.Ptr {
|
||||||
return errors.New("Calling parser with non-pointer")
|
return nil, errors.New("Calling parser with non-pointer")
|
||||||
}
|
}
|
||||||
|
|
||||||
if reflect.ValueOf(in).Elem().Kind() != reflect.Struct {
|
if reflect.ValueOf(in).Elem().Kind() != reflect.Struct {
|
||||||
return errors.New("Calling parser with pointer to non-struct")
|
return nil, errors.New("Calling parser with pointer to non-struct")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
afterFuncs := []afterFunc{}
|
||||||
|
|
||||||
st := reflect.ValueOf(in).Elem()
|
st := reflect.ValueOf(in).Elem()
|
||||||
for i := 0; i < st.NumField(); i++ {
|
for i := 0; i < st.NumField(); i++ {
|
||||||
valField := st.Field(i)
|
valField := st.Field(i)
|
||||||
|
@ -97,9 +159,79 @@ func execTags(in interface{}, fs *pflag.FlagSet) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
value := varDefault(typeField.Tag.Get("vardefault"), typeField.Tag.Get("default"))
|
value := varDefault(typeField.Tag.Get("vardefault"), typeField.Tag.Get("default"))
|
||||||
value = envDefault(typeField.Tag.Get("env"), value)
|
value = envDefault(typeField, value)
|
||||||
parts := strings.Split(typeField.Tag.Get("flag"), ",")
|
parts := strings.Split(typeField.Tag.Get("flag"), ",")
|
||||||
|
|
||||||
|
switch typeField.Type {
|
||||||
|
case reflect.TypeOf(time.Duration(0)):
|
||||||
|
v, err := time.ParseDuration(value)
|
||||||
|
if err != nil {
|
||||||
|
if value == "" {
|
||||||
|
v = time.Duration(0)
|
||||||
|
} else {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if typeField.Tag.Get("flag") != "" {
|
||||||
|
if len(parts) == 1 {
|
||||||
|
fs.DurationVar(valField.Addr().Interface().(*time.Duration), parts[0], v, typeField.Tag.Get("description"))
|
||||||
|
} else {
|
||||||
|
fs.DurationVarP(valField.Addr().Interface().(*time.Duration), parts[0], parts[1], v, typeField.Tag.Get("description"))
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
valField.Set(reflect.ValueOf(v))
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
|
||||||
|
case reflect.TypeOf(time.Time{}):
|
||||||
|
var sVar string
|
||||||
|
|
||||||
|
if typeField.Tag.Get("flag") != "" {
|
||||||
|
if len(parts) == 1 {
|
||||||
|
fs.StringVar(&sVar, parts[0], value, typeField.Tag.Get("description"))
|
||||||
|
} else {
|
||||||
|
fs.StringVarP(&sVar, parts[0], parts[1], value, typeField.Tag.Get("description"))
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
sVar = value
|
||||||
|
}
|
||||||
|
|
||||||
|
afterFuncs = append(afterFuncs, func(valField reflect.Value, sVar *string) func() error {
|
||||||
|
return func() error {
|
||||||
|
if *sVar == "" {
|
||||||
|
// No time, no problem
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check whether we could have a timestamp
|
||||||
|
if ts, err := strconv.ParseInt(*sVar, 10, 64); err == nil {
|
||||||
|
t := time.Unix(ts, 0)
|
||||||
|
valField.Set(reflect.ValueOf(t))
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// We haven't so lets walk through possible time formats
|
||||||
|
matched := false
|
||||||
|
for _, tf := range timeParserFormats {
|
||||||
|
if t, err := time.Parse(tf, *sVar); err == nil {
|
||||||
|
matched = true
|
||||||
|
valField.Set(reflect.ValueOf(t))
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if !matched {
|
||||||
|
return fmt.Errorf("Value %q did not match expected time formats", *sVar)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}(valField, &sVar))
|
||||||
|
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
switch typeField.Type.Kind() {
|
switch typeField.Type.Kind() {
|
||||||
case reflect.String:
|
case reflect.String:
|
||||||
if typeField.Tag.Get("flag") != "" {
|
if typeField.Tag.Get("flag") != "" {
|
||||||
|
@ -130,7 +262,7 @@ func execTags(in interface{}, fs *pflag.FlagSet) error {
|
||||||
if value == "" {
|
if value == "" {
|
||||||
vt = 0
|
vt = 0
|
||||||
} else {
|
} else {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if typeField.Tag.Get("flag") != "" {
|
if typeField.Tag.Get("flag") != "" {
|
||||||
|
@ -145,7 +277,7 @@ func execTags(in interface{}, fs *pflag.FlagSet) error {
|
||||||
if value == "" {
|
if value == "" {
|
||||||
vt = 0
|
vt = 0
|
||||||
} else {
|
} else {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if typeField.Tag.Get("flag") != "" {
|
if typeField.Tag.Get("flag") != "" {
|
||||||
|
@ -160,7 +292,7 @@ func execTags(in interface{}, fs *pflag.FlagSet) error {
|
||||||
if value == "" {
|
if value == "" {
|
||||||
vt = 0.0
|
vt = 0.0
|
||||||
} else {
|
} else {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if typeField.Tag.Get("flag") != "" {
|
if typeField.Tag.Get("flag") != "" {
|
||||||
|
@ -170,9 +302,11 @@ func execTags(in interface{}, fs *pflag.FlagSet) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
case reflect.Struct:
|
case reflect.Struct:
|
||||||
if err := execTags(valField.Addr().Interface(), fs); err != nil {
|
afs, err := execTags(valField.Addr().Interface(), fs)
|
||||||
return err
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
}
|
}
|
||||||
|
afterFuncs = append(afterFuncs, afs...)
|
||||||
|
|
||||||
case reflect.Slice:
|
case reflect.Slice:
|
||||||
switch typeField.Type.Elem().Kind() {
|
switch typeField.Type.Elem().Kind() {
|
||||||
|
@ -181,7 +315,7 @@ func execTags(in interface{}, fs *pflag.FlagSet) error {
|
||||||
for _, v := range strings.Split(value, ",") {
|
for _, v := range strings.Split(value, ",") {
|
||||||
it, err := strconv.ParseInt(strings.TrimSpace(v), 10, 64)
|
it, err := strconv.ParseInt(strings.TrimSpace(v), 10, 64)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return nil, err
|
||||||
}
|
}
|
||||||
def = append(def, int(it))
|
def = append(def, int(it))
|
||||||
}
|
}
|
||||||
|
@ -195,7 +329,10 @@ func execTags(in interface{}, fs *pflag.FlagSet) error {
|
||||||
if len(del) == 0 {
|
if len(del) == 0 {
|
||||||
del = ","
|
del = ","
|
||||||
}
|
}
|
||||||
def := strings.Split(value, del)
|
var def = []string{}
|
||||||
|
if value != "" {
|
||||||
|
def = strings.Split(value, del)
|
||||||
|
}
|
||||||
if len(parts) == 1 {
|
if len(parts) == 1 {
|
||||||
fs.StringSliceVar(valField.Addr().Interface().(*[]string), parts[0], def, typeField.Tag.Get("description"))
|
fs.StringSliceVar(valField.Addr().Interface().(*[]string), parts[0], def, typeField.Tag.Get("description"))
|
||||||
} else {
|
} else {
|
||||||
|
@ -205,7 +342,7 @@ func execTags(in interface{}, fs *pflag.FlagSet) error {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil
|
return afterFuncs, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func registerFlagFloat(t reflect.Kind, fs *pflag.FlagSet, field interface{}, parts []string, vt float64, desc string) {
|
func registerFlagFloat(t reflect.Kind, fs *pflag.FlagSet, field interface{}, parts []string, vt float64, desc string) {
|
||||||
|
@ -289,9 +426,14 @@ func registerFlagUint(t reflect.Kind, fs *pflag.FlagSet, field interface{}, part
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func envDefault(env, def string) string {
|
func envDefault(field reflect.StructField, def string) string {
|
||||||
value := def
|
value := def
|
||||||
|
|
||||||
|
env := field.Tag.Get("env")
|
||||||
|
if env == "" && autoEnv {
|
||||||
|
env = deriveEnvVarName(field.Name)
|
||||||
|
}
|
||||||
|
|
||||||
if env != "" {
|
if env != "" {
|
||||||
if e := os.Getenv(env); e != "" {
|
if e := os.Getenv(env); e != "" {
|
||||||
value = e
|
value = e
|
||||||
|
|
14
vendor/github.com/asaskevich/govalidator/.travis.yml
generated
vendored
14
vendor/github.com/asaskevich/govalidator/.travis.yml
generated
vendored
|
@ -1,14 +0,0 @@
|
||||||
language: go
|
|
||||||
|
|
||||||
go:
|
|
||||||
- 1.1
|
|
||||||
- 1.2
|
|
||||||
- 1.3
|
|
||||||
- 1.4
|
|
||||||
- 1.5
|
|
||||||
- 1.6
|
|
||||||
- tip
|
|
||||||
|
|
||||||
notifications:
|
|
||||||
email:
|
|
||||||
- bwatas@gmail.com
|
|
398
vendor/github.com/asaskevich/govalidator/README.md
generated
vendored
398
vendor/github.com/asaskevich/govalidator/README.md
generated
vendored
|
@ -1,398 +0,0 @@
|
||||||
govalidator
|
|
||||||
===========
|
|
||||||
[![Gitter](https://badges.gitter.im/Join%20Chat.svg)](https://gitter.im/asaskevich/govalidator?utm_source=badge&utm_medium=badge&utm_campaign=pr-badge) [![GoDoc](https://godoc.org/github.com/asaskevich/govalidator?status.png)](https://godoc.org/github.com/asaskevich/govalidator) [![Coverage Status](https://img.shields.io/coveralls/asaskevich/govalidator.svg)](https://coveralls.io/r/asaskevich/govalidator?branch=master) [![wercker status](https://app.wercker.com/status/1ec990b09ea86c910d5f08b0e02c6043/s "wercker status")](https://app.wercker.com/project/bykey/1ec990b09ea86c910d5f08b0e02c6043)
|
|
||||||
[![Build Status](https://travis-ci.org/asaskevich/govalidator.svg?branch=master)](https://travis-ci.org/asaskevich/govalidator)
|
|
||||||
|
|
||||||
A package of validators and sanitizers for strings, structs and collections. Based on [validator.js](https://github.com/chriso/validator.js).
|
|
||||||
|
|
||||||
#### Installation
|
|
||||||
Make sure that Go is installed on your computer.
|
|
||||||
Type the following command in your terminal:
|
|
||||||
|
|
||||||
go get github.com/asaskevich/govalidator
|
|
||||||
|
|
||||||
or you can get specified release of the package with `gopkg.in`:
|
|
||||||
|
|
||||||
go get gopkg.in/asaskevich/govalidator.v4
|
|
||||||
|
|
||||||
After it the package is ready to use.
|
|
||||||
|
|
||||||
|
|
||||||
#### Import package in your project
|
|
||||||
Add following line in your `*.go` file:
|
|
||||||
```go
|
|
||||||
import "github.com/asaskevich/govalidator"
|
|
||||||
```
|
|
||||||
If you are unhappy to use long `govalidator`, you can do something like this:
|
|
||||||
```go
|
|
||||||
import (
|
|
||||||
valid "github.com/asaskevich/govalidator"
|
|
||||||
)
|
|
||||||
```
|
|
||||||
|
|
||||||
#### Activate behavior to require all fields have a validation tag by default
|
|
||||||
`SetFieldsRequiredByDefault` causes validation to fail when struct fields do not include validations or are not explicitly marked as exempt (using `valid:"-"` or `valid:"email,optional"`). A good place to activate this is a package init function or the main() function.
|
|
||||||
|
|
||||||
```go
|
|
||||||
import "github.com/asaskevich/govalidator"
|
|
||||||
|
|
||||||
func init() {
|
|
||||||
govalidator.SetFieldsRequiredByDefault(true)
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
Here's some code to explain it:
|
|
||||||
```go
|
|
||||||
// this struct definition will fail govalidator.ValidateStruct() (and the field values do not matter):
|
|
||||||
type exampleStruct struct {
|
|
||||||
Name string ``
|
|
||||||
Email string `valid:"email"`
|
|
||||||
|
|
||||||
// this, however, will only fail when Email is empty or an invalid email address:
|
|
||||||
type exampleStruct2 struct {
|
|
||||||
Name string `valid:"-"`
|
|
||||||
Email string `valid:"email"`
|
|
||||||
|
|
||||||
// lastly, this will only fail when Email is an invalid email address but not when it's empty:
|
|
||||||
type exampleStruct2 struct {
|
|
||||||
Name string `valid:"-"`
|
|
||||||
Email string `valid:"email,optional"`
|
|
||||||
```
|
|
||||||
|
|
||||||
#### Recent breaking changes (see [#123](https://github.com/asaskevich/govalidator/pull/123))
|
|
||||||
##### Custom validator function signature
|
|
||||||
A context was added as the second parameter, for structs this is the object being validated – this makes dependent validation possible.
|
|
||||||
```go
|
|
||||||
import "github.com/asaskevich/govalidator"
|
|
||||||
|
|
||||||
// old signature
|
|
||||||
func(i interface{}) bool
|
|
||||||
|
|
||||||
// new signature
|
|
||||||
func(i interface{}, o interface{}) bool
|
|
||||||
```
|
|
||||||
|
|
||||||
##### Adding a custom validator
|
|
||||||
This was changed to prevent data races when accessing custom validators.
|
|
||||||
```go
|
|
||||||
import "github.com/asaskevich/govalidator"
|
|
||||||
|
|
||||||
// before
|
|
||||||
govalidator.CustomTypeTagMap["customByteArrayValidator"] = CustomTypeValidator(func(i interface{}, o interface{}) bool {
|
|
||||||
// ...
|
|
||||||
})
|
|
||||||
|
|
||||||
// after
|
|
||||||
govalidator.CustomTypeTagMap.Set("customByteArrayValidator", CustomTypeValidator(func(i interface{}, o interface{}) bool {
|
|
||||||
// ...
|
|
||||||
}))
|
|
||||||
```
|
|
||||||
|
|
||||||
#### List of functions:
|
|
||||||
```go
|
|
||||||
func Abs(value float64) float64
|
|
||||||
func BlackList(str, chars string) string
|
|
||||||
func ByteLength(str string, params ...string) bool
|
|
||||||
func StringLength(str string, params ...string) bool
|
|
||||||
func StringMatches(s string, params ...string) bool
|
|
||||||
func CamelCaseToUnderscore(str string) string
|
|
||||||
func Contains(str, substring string) bool
|
|
||||||
func Count(array []interface{}, iterator ConditionIterator) int
|
|
||||||
func Each(array []interface{}, iterator Iterator)
|
|
||||||
func ErrorByField(e error, field string) string
|
|
||||||
func Filter(array []interface{}, iterator ConditionIterator) []interface{}
|
|
||||||
func Find(array []interface{}, iterator ConditionIterator) interface{}
|
|
||||||
func GetLine(s string, index int) (string, error)
|
|
||||||
func GetLines(s string) []string
|
|
||||||
func IsHost(s string) bool
|
|
||||||
func InRange(value, left, right float64) bool
|
|
||||||
func IsASCII(str string) bool
|
|
||||||
func IsAlpha(str string) bool
|
|
||||||
func IsAlphanumeric(str string) bool
|
|
||||||
func IsBase64(str string) bool
|
|
||||||
func IsByteLength(str string, min, max int) bool
|
|
||||||
func IsCreditCard(str string) bool
|
|
||||||
func IsDataURI(str string) bool
|
|
||||||
func IsDialString(str string) bool
|
|
||||||
func IsDNSName(str string) bool
|
|
||||||
func IsDivisibleBy(str, num string) bool
|
|
||||||
func IsEmail(str string) bool
|
|
||||||
func IsFilePath(str string) (bool, int)
|
|
||||||
func IsFloat(str string) bool
|
|
||||||
func IsFullWidth(str string) bool
|
|
||||||
func IsHalfWidth(str string) bool
|
|
||||||
func IsHexadecimal(str string) bool
|
|
||||||
func IsHexcolor(str string) bool
|
|
||||||
func IsIP(str string) bool
|
|
||||||
func IsIPv4(str string) bool
|
|
||||||
func IsIPv6(str string) bool
|
|
||||||
func IsISBN(str string, version int) bool
|
|
||||||
func IsISBN10(str string) bool
|
|
||||||
func IsISBN13(str string) bool
|
|
||||||
func IsISO3166Alpha2(str string) bool
|
|
||||||
func IsISO3166Alpha3(str string) bool
|
|
||||||
func IsInt(str string) bool
|
|
||||||
func IsJSON(str string) bool
|
|
||||||
func IsLatitude(str string) bool
|
|
||||||
func IsLongitude(str string) bool
|
|
||||||
func IsLowerCase(str string) bool
|
|
||||||
func IsMAC(str string) bool
|
|
||||||
func IsMongoID(str string) bool
|
|
||||||
func IsMultibyte(str string) bool
|
|
||||||
func IsNatural(value float64) bool
|
|
||||||
func IsNegative(value float64) bool
|
|
||||||
func IsNonNegative(value float64) bool
|
|
||||||
func IsNonPositive(value float64) bool
|
|
||||||
func IsNull(str string) bool
|
|
||||||
func IsNumeric(str string) bool
|
|
||||||
func IsPort(str string) bool
|
|
||||||
func IsPositive(value float64) bool
|
|
||||||
func IsPrintableASCII(str string) bool
|
|
||||||
func IsRGBcolor(str string) bool
|
|
||||||
func IsRequestURI(rawurl string) bool
|
|
||||||
func IsRequestURL(rawurl string) bool
|
|
||||||
func IsSSN(str string) bool
|
|
||||||
func IsSemver(str string) bool
|
|
||||||
func IsURL(str string) bool
|
|
||||||
func IsUTFDigit(str string) bool
|
|
||||||
func IsUTFLetter(str string) bool
|
|
||||||
func IsUTFLetterNumeric(str string) bool
|
|
||||||
func IsUTFNumeric(str string) bool
|
|
||||||
func IsUUID(str string) bool
|
|
||||||
func IsUUIDv3(str string) bool
|
|
||||||
func IsUUIDv4(str string) bool
|
|
||||||
func IsUUIDv5(str string) bool
|
|
||||||
func IsUpperCase(str string) bool
|
|
||||||
func IsVariableWidth(str string) bool
|
|
||||||
func IsWhole(value float64) bool
|
|
||||||
func LeftTrim(str, chars string) string
|
|
||||||
func Map(array []interface{}, iterator ResultIterator) []interface{}
|
|
||||||
func Matches(str, pattern string) bool
|
|
||||||
func NormalizeEmail(str string) (string, error)
|
|
||||||
func RemoveTags(s string) string
|
|
||||||
func ReplacePattern(str, pattern, replace string) string
|
|
||||||
func Reverse(s string) string
|
|
||||||
func RightTrim(str, chars string) string
|
|
||||||
func SafeFileName(str string) string
|
|
||||||
func Sign(value float64) float64
|
|
||||||
func StripLow(str string, keepNewLines bool) string
|
|
||||||
func ToBoolean(str string) (bool, error)
|
|
||||||
func ToFloat(str string) (float64, error)
|
|
||||||
func ToInt(str string) (int64, error)
|
|
||||||
func ToJSON(obj interface{}) (string, error)
|
|
||||||
func ToString(obj interface{}) string
|
|
||||||
func Trim(str, chars string) string
|
|
||||||
func Truncate(str string, length int, ending string) string
|
|
||||||
func UnderscoreToCamelCase(s string) string
|
|
||||||
func ValidateStruct(s interface{}) (bool, error)
|
|
||||||
func WhiteList(str, chars string) string
|
|
||||||
type ConditionIterator
|
|
||||||
type Error
|
|
||||||
func (e Error) Error() string
|
|
||||||
type Errors
|
|
||||||
func (es Errors) Error() string
|
|
||||||
type ISO3166Entry
|
|
||||||
type Iterator
|
|
||||||
type ParamValidator
|
|
||||||
type ResultIterator
|
|
||||||
type UnsupportedTypeError
|
|
||||||
func (e *UnsupportedTypeError) Error() string
|
|
||||||
type Validator
|
|
||||||
```
|
|
||||||
|
|
||||||
#### Examples
|
|
||||||
###### IsURL
|
|
||||||
```go
|
|
||||||
println(govalidator.IsURL(`http://user@pass:domain.com/path/page`))
|
|
||||||
```
|
|
||||||
###### ToString
|
|
||||||
```go
|
|
||||||
type User struct {
|
|
||||||
FirstName string
|
|
||||||
LastName string
|
|
||||||
}
|
|
||||||
|
|
||||||
str := govalidator.ToString(&User{"John", "Juan"})
|
|
||||||
println(str)
|
|
||||||
```
|
|
||||||
###### Each, Map, Filter, Count for slices
|
|
||||||
Each iterates over the slice/array and calls Iterator for every item
|
|
||||||
```go
|
|
||||||
data := []interface{}{1, 2, 3, 4, 5}
|
|
||||||
var fn govalidator.Iterator = func(value interface{}, index int) {
|
|
||||||
println(value.(int))
|
|
||||||
}
|
|
||||||
govalidator.Each(data, fn)
|
|
||||||
```
|
|
||||||
```go
|
|
||||||
data := []interface{}{1, 2, 3, 4, 5}
|
|
||||||
var fn govalidator.ResultIterator = func(value interface{}, index int) interface{} {
|
|
||||||
return value.(int) * 3
|
|
||||||
}
|
|
||||||
_ = govalidator.Map(data, fn) // result = []interface{}{1, 6, 9, 12, 15}
|
|
||||||
```
|
|
||||||
```go
|
|
||||||
data := []interface{}{1, 2, 3, 4, 5, 6, 7, 8, 9, 10}
|
|
||||||
var fn govalidator.ConditionIterator = func(value interface{}, index int) bool {
|
|
||||||
return value.(int)%2 == 0
|
|
||||||
}
|
|
||||||
_ = govalidator.Filter(data, fn) // result = []interface{}{2, 4, 6, 8, 10}
|
|
||||||
_ = govalidator.Count(data, fn) // result = 5
|
|
||||||
```
|
|
||||||
###### ValidateStruct [#2](https://github.com/asaskevich/govalidator/pull/2)
|
|
||||||
If you want to validate structs, you can use tag `valid` for any field in your structure. All validators used with this field in one tag are separated by comma. If you want to skip validation, place `-` in your tag. If you need a validator that is not on the list below, you can add it like this:
|
|
||||||
```go
|
|
||||||
govalidator.TagMap["duck"] = govalidator.Validator(func(str string) bool {
|
|
||||||
return str == "duck"
|
|
||||||
})
|
|
||||||
```
|
|
||||||
For completely custom validators (interface-based), see below.
|
|
||||||
|
|
||||||
Here is a list of available validators for struct fields (validator - used function):
|
|
||||||
```go
|
|
||||||
"alpha": IsAlpha,
|
|
||||||
"alphanum": IsAlphanumeric,
|
|
||||||
"ascii": IsASCII,
|
|
||||||
"base64": IsBase64,
|
|
||||||
"creditcard": IsCreditCard,
|
|
||||||
"datauri": IsDataURI,
|
|
||||||
"dialstring": IsDialString,
|
|
||||||
"dns": IsDNSName,
|
|
||||||
"email": IsEmail,
|
|
||||||
"float": IsFloat,
|
|
||||||
"fullwidth": IsFullWidth,
|
|
||||||
"halfwidth": IsHalfWidth,
|
|
||||||
"hexadecimal": IsHexadecimal,
|
|
||||||
"hexcolor": IsHexcolor,
|
|
||||||
"host": IsHost,
|
|
||||||
"int": IsInt,
|
|
||||||
"ip": IsIP,
|
|
||||||
"ipv4": IsIPv4,
|
|
||||||
"ipv6": IsIPv6,
|
|
||||||
"isbn10": IsISBN10,
|
|
||||||
"isbn13": IsISBN13,
|
|
||||||
"json": IsJSON,
|
|
||||||
"latitude": IsLatitude,
|
|
||||||
"longitude": IsLongitude,
|
|
||||||
"lowercase": IsLowerCase,
|
|
||||||
"mac": IsMAC,
|
|
||||||
"multibyte": IsMultibyte,
|
|
||||||
"null": IsNull,
|
|
||||||
"numeric": IsNumeric,
|
|
||||||
"port": IsPort,
|
|
||||||
"printableascii": IsPrintableASCII,
|
|
||||||
"requri": IsRequestURI,
|
|
||||||
"requrl": IsRequestURL,
|
|
||||||
"rgbcolor": IsRGBcolor,
|
|
||||||
"ssn": IsSSN,
|
|
||||||
"semver": IsSemver,
|
|
||||||
"uppercase": IsUpperCase,
|
|
||||||
"url": IsURL,
|
|
||||||
"utfdigit": IsUTFDigit,
|
|
||||||
"utfletter": IsUTFLetter,
|
|
||||||
"utfletternum": IsUTFLetterNumeric,
|
|
||||||
"utfnumeric": IsUTFNumeric,
|
|
||||||
"uuid": IsUUID,
|
|
||||||
"uuidv3": IsUUIDv3,
|
|
||||||
"uuidv4": IsUUIDv4,
|
|
||||||
"uuidv5": IsUUIDv5,
|
|
||||||
"variablewidth": IsVariableWidth,
|
|
||||||
```
|
|
||||||
Validators with parameters
|
|
||||||
|
|
||||||
```go
|
|
||||||
"length(min|max)": ByteLength,
|
|
||||||
"matches(pattern)": StringMatches,
|
|
||||||
```
|
|
||||||
|
|
||||||
And here is small example of usage:
|
|
||||||
```go
|
|
||||||
type Post struct {
|
|
||||||
Title string `valid:"alphanum,required"`
|
|
||||||
Message string `valid:"duck,ascii"`
|
|
||||||
AuthorIP string `valid:"ipv4"`
|
|
||||||
Date string `valid:"-"`
|
|
||||||
}
|
|
||||||
post := &Post{
|
|
||||||
Title: "My Example Post",
|
|
||||||
Message: "duck",
|
|
||||||
AuthorIP: "123.234.54.3",
|
|
||||||
}
|
|
||||||
|
|
||||||
// Add your own struct validation tags
|
|
||||||
govalidator.TagMap["duck"] = govalidator.Validator(func(str string) bool {
|
|
||||||
return str == "duck"
|
|
||||||
})
|
|
||||||
|
|
||||||
result, err := govalidator.ValidateStruct(post)
|
|
||||||
if err != nil {
|
|
||||||
println("error: " + err.Error())
|
|
||||||
}
|
|
||||||
println(result)
|
|
||||||
```
|
|
||||||
###### WhiteList
|
|
||||||
```go
|
|
||||||
// Remove all characters from string ignoring characters between "a" and "z"
|
|
||||||
println(govalidator.WhiteList("a3a43a5a4a3a2a23a4a5a4a3a4", "a-z") == "aaaaaaaaaaaa")
|
|
||||||
```
|
|
||||||
|
|
||||||
###### Custom validation functions
|
|
||||||
Custom validation using your own domain specific validators is also available - here's an example of how to use it:
|
|
||||||
```go
|
|
||||||
import "github.com/asaskevich/govalidator"
|
|
||||||
|
|
||||||
type CustomByteArray [6]byte // custom types are supported and can be validated
|
|
||||||
|
|
||||||
type StructWithCustomByteArray struct {
|
|
||||||
ID CustomByteArray `valid:"customByteArrayValidator,customMinLengthValidator"` // multiple custom validators are possible as well and will be evaluated in sequence
|
|
||||||
Email string `valid:"email"`
|
|
||||||
CustomMinLength int `valid:"-"`
|
|
||||||
}
|
|
||||||
|
|
||||||
govalidator.CustomTypeTagMap.Set("customByteArrayValidator", CustomTypeValidator(func(i interface{}, context interface{}) bool {
|
|
||||||
switch v := context.(type) { // you can type switch on the context interface being validated
|
|
||||||
case StructWithCustomByteArray:
|
|
||||||
// you can check and validate against some other field in the context,
|
|
||||||
// return early or not validate against the context at all – your choice
|
|
||||||
case SomeOtherType:
|
|
||||||
// ...
|
|
||||||
default:
|
|
||||||
// expecting some other type? Throw/panic here or continue
|
|
||||||
}
|
|
||||||
|
|
||||||
switch v := i.(type) { // type switch on the struct field being validated
|
|
||||||
case CustomByteArray:
|
|
||||||
for _, e := range v { // this validator checks that the byte array is not empty, i.e. not all zeroes
|
|
||||||
if e != 0 {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return false
|
|
||||||
}))
|
|
||||||
govalidator.CustomTypeTagMap.Set("customMinLengthValidator", CustomTypeValidator(func(i interface{}, context interface{}) bool {
|
|
||||||
switch v := context.(type) { // this validates a field against the value in another field, i.e. dependent validation
|
|
||||||
case StructWithCustomByteArray:
|
|
||||||
return len(v.ID) >= v.CustomMinLength
|
|
||||||
}
|
|
||||||
return false
|
|
||||||
}))
|
|
||||||
```
|
|
||||||
|
|
||||||
#### Notes
|
|
||||||
Documentation is available here: [godoc.org](https://godoc.org/github.com/asaskevich/govalidator).
|
|
||||||
Full information about code coverage is also available here: [govalidator on gocover.io](http://gocover.io/github.com/asaskevich/govalidator).
|
|
||||||
|
|
||||||
#### Support
|
|
||||||
If you do have a contribution for the package feel free to put up a Pull Request or open Issue.
|
|
||||||
|
|
||||||
#### Special thanks to [contributors](https://github.com/asaskevich/govalidator/graphs/contributors)
|
|
||||||
* [Daniel Lohse](https://github.com/annismckenzie)
|
|
||||||
* [Attila Oláh](https://github.com/attilaolah)
|
|
||||||
* [Daniel Korner](https://github.com/Dadie)
|
|
||||||
* [Steven Wilkin](https://github.com/stevenwilkin)
|
|
||||||
* [Deiwin Sarjas](https://github.com/deiwin)
|
|
||||||
* [Noah Shibley](https://github.com/slugmobile)
|
|
||||||
* [Nathan Davies](https://github.com/nathj07)
|
|
||||||
* [Matt Sanford](https://github.com/mzsanford)
|
|
||||||
* [Simon ccl1115](https://github.com/ccl1115)
|
|
35
vendor/github.com/asaskevich/govalidator/converter.go
generated
vendored
35
vendor/github.com/asaskevich/govalidator/converter.go
generated
vendored
|
@ -3,6 +3,7 @@ package govalidator
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"reflect"
|
||||||
"strconv"
|
"strconv"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -30,20 +31,34 @@ func ToFloat(str string) (float64, error) {
|
||||||
return res, err
|
return res, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// ToInt convert the input string to an integer, or 0 if the input is not an integer.
|
// ToInt convert the input string or any int type to an integer type 64, or 0 if the input is not an integer.
|
||||||
func ToInt(str string) (int64, error) {
|
func ToInt(value interface{}) (res int64, err error) {
|
||||||
res, err := strconv.ParseInt(str, 0, 64)
|
val := reflect.ValueOf(value)
|
||||||
if err != nil {
|
|
||||||
|
switch value.(type) {
|
||||||
|
case int, int8, int16, int32, int64:
|
||||||
|
res = val.Int()
|
||||||
|
case uint, uint8, uint16, uint32, uint64:
|
||||||
|
res = int64(val.Uint())
|
||||||
|
case string:
|
||||||
|
if IsInt(val.String()) {
|
||||||
|
res, err = strconv.ParseInt(val.String(), 0, 64)
|
||||||
|
if err != nil {
|
||||||
|
res = 0
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
err = fmt.Errorf("math: square root of negative number %g", value)
|
||||||
|
res = 0
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
err = fmt.Errorf("math: square root of negative number %g", value)
|
||||||
res = 0
|
res = 0
|
||||||
}
|
}
|
||||||
return res, err
|
|
||||||
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// ToBoolean convert the input string to a boolean.
|
// ToBoolean convert the input string to a boolean.
|
||||||
func ToBoolean(str string) (bool, error) {
|
func ToBoolean(str string) (bool, error) {
|
||||||
res, err := strconv.ParseBool(str)
|
return strconv.ParseBool(str)
|
||||||
if err != nil {
|
|
||||||
res = false
|
|
||||||
}
|
|
||||||
return res, err
|
|
||||||
}
|
}
|
||||||
|
|
11
vendor/github.com/asaskevich/govalidator/error.go
generated
vendored
11
vendor/github.com/asaskevich/govalidator/error.go
generated
vendored
|
@ -1,5 +1,7 @@
|
||||||
package govalidator
|
package govalidator
|
||||||
|
|
||||||
|
import "strings"
|
||||||
|
|
||||||
// Errors is an array of multiple errors and conforms to the error interface.
|
// Errors is an array of multiple errors and conforms to the error interface.
|
||||||
type Errors []error
|
type Errors []error
|
||||||
|
|
||||||
|
@ -9,11 +11,11 @@ func (es Errors) Errors() []error {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (es Errors) Error() string {
|
func (es Errors) Error() string {
|
||||||
var err string
|
var errs []string
|
||||||
for _, e := range es {
|
for _, e := range es {
|
||||||
err += e.Error() + ";"
|
errs = append(errs, e.Error())
|
||||||
}
|
}
|
||||||
return err
|
return strings.Join(errs, ";")
|
||||||
}
|
}
|
||||||
|
|
||||||
// Error encapsulates a name, an error and whether there's a custom error message or not.
|
// Error encapsulates a name, an error and whether there's a custom error message or not.
|
||||||
|
@ -21,6 +23,9 @@ type Error struct {
|
||||||
Name string
|
Name string
|
||||||
Err error
|
Err error
|
||||||
CustomErrorMessageExists bool
|
CustomErrorMessageExists bool
|
||||||
|
|
||||||
|
// Validator indicates the name of the validator that failed
|
||||||
|
Validator string
|
||||||
}
|
}
|
||||||
|
|
||||||
func (e Error) Error() string {
|
func (e Error) Error() string {
|
||||||
|
|
48
vendor/github.com/asaskevich/govalidator/numerics.go
generated
vendored
48
vendor/github.com/asaskevich/govalidator/numerics.go
generated
vendored
|
@ -1,10 +1,13 @@
|
||||||
package govalidator
|
package govalidator
|
||||||
|
|
||||||
import "math"
|
import (
|
||||||
|
"math"
|
||||||
|
"reflect"
|
||||||
|
)
|
||||||
|
|
||||||
// Abs returns absolute value of number
|
// Abs returns absolute value of number
|
||||||
func Abs(value float64) float64 {
|
func Abs(value float64) float64 {
|
||||||
return value * Sign(value)
|
return math.Abs(value)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Sign returns signum of number: 1 in case of value > 0, -1 in case of value < 0, 0 otherwise
|
// Sign returns signum of number: 1 in case of value > 0, -1 in case of value < 0, 0 otherwise
|
||||||
|
@ -39,16 +42,53 @@ func IsNonPositive(value float64) bool {
|
||||||
}
|
}
|
||||||
|
|
||||||
// InRange returns true if value lies between left and right border
|
// InRange returns true if value lies between left and right border
|
||||||
func InRange(value, left, right float64) bool {
|
func InRangeInt(value, left, right interface{}) bool {
|
||||||
|
value64, _ := ToInt(value)
|
||||||
|
left64, _ := ToInt(left)
|
||||||
|
right64, _ := ToInt(right)
|
||||||
|
if left64 > right64 {
|
||||||
|
left64, right64 = right64, left64
|
||||||
|
}
|
||||||
|
return value64 >= left64 && value64 <= right64
|
||||||
|
}
|
||||||
|
|
||||||
|
// InRange returns true if value lies between left and right border
|
||||||
|
func InRangeFloat32(value, left, right float32) bool {
|
||||||
if left > right {
|
if left > right {
|
||||||
left, right = right, left
|
left, right = right, left
|
||||||
}
|
}
|
||||||
return value >= left && value <= right
|
return value >= left && value <= right
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// InRange returns true if value lies between left and right border
|
||||||
|
func InRangeFloat64(value, left, right float64) bool {
|
||||||
|
if left > right {
|
||||||
|
left, right = right, left
|
||||||
|
}
|
||||||
|
return value >= left && value <= right
|
||||||
|
}
|
||||||
|
|
||||||
|
// InRange returns true if value lies between left and right border, generic type to handle int, float32 or float64, all types must the same type
|
||||||
|
func InRange(value interface{}, left interface{}, right interface{}) bool {
|
||||||
|
|
||||||
|
reflectValue := reflect.TypeOf(value).Kind()
|
||||||
|
reflectLeft := reflect.TypeOf(left).Kind()
|
||||||
|
reflectRight := reflect.TypeOf(right).Kind()
|
||||||
|
|
||||||
|
if reflectValue == reflect.Int && reflectLeft == reflect.Int && reflectRight == reflect.Int {
|
||||||
|
return InRangeInt(value.(int), left.(int), right.(int))
|
||||||
|
} else if reflectValue == reflect.Float32 && reflectLeft == reflect.Float32 && reflectRight == reflect.Float32 {
|
||||||
|
return InRangeFloat32(value.(float32), left.(float32), right.(float32))
|
||||||
|
} else if reflectValue == reflect.Float64 && reflectLeft == reflect.Float64 && reflectRight == reflect.Float64 {
|
||||||
|
return InRangeFloat64(value.(float64), left.(float64), right.(float64))
|
||||||
|
} else {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// IsWhole returns true if value is whole number
|
// IsWhole returns true if value is whole number
|
||||||
func IsWhole(value float64) bool {
|
func IsWhole(value float64) bool {
|
||||||
return Abs(math.Remainder(value, 1)) == 0
|
return math.Remainder(value, 1) == 0
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsNatural returns true if value is natural number (positive and whole)
|
// IsNatural returns true if value is natural number (positive and whole)
|
||||||
|
|
26
vendor/github.com/asaskevich/govalidator/patterns.go
generated
vendored
26
vendor/github.com/asaskevich/govalidator/patterns.go
generated
vendored
|
@ -4,7 +4,7 @@ import "regexp"
|
||||||
|
|
||||||
// Basic regular expressions for validating strings
|
// Basic regular expressions for validating strings
|
||||||
const (
|
const (
|
||||||
Email string = "^(((([a-zA-Z]|\\d|[!#\\$%&'\\*\\+\\-\\/=\\?\\^_`{\\|}~]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])+(\\.([a-zA-Z]|\\d|[!#\\$%&'\\*\\+\\-\\/=\\?\\^_`{\\|}~]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])+)*)|((\\x22)((((\\x20|\\x09)*(\\x0d\\x0a))?(\\x20|\\x09)+)?(([\\x01-\\x08\\x0b\\x0c\\x0e-\\x1f\\x7f]|\\x21|[\\x23-\\x5b]|[\\x5d-\\x7e]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])|(\\([\\x01-\\x09\\x0b\\x0c\\x0d-\\x7f]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}]))))*(((\\x20|\\x09)*(\\x0d\\x0a))?(\\x20|\\x09)+)?(\\x22)))@((([a-zA-Z]|\\d|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])|(([a-zA-Z]|\\d|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])([a-zA-Z]|\\d|-|\\.|_|~|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])*([a-zA-Z]|\\d|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])))\\.)+(([a-zA-Z]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])|(([a-zA-Z]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])([a-zA-Z]|\\d|-|\\.|_|~|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])*([a-zA-Z]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])))\\.?$"
|
//Email string = "^(((([a-zA-Z]|\\d|[!#\\$%&'\\*\\+\\-\\/=\\?\\^_`{\\|}~]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])+(\\.([a-zA-Z]|\\d|[!#\\$%&'\\*\\+\\-\\/=\\?\\^_`{\\|}~]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])+)*)|((\\x22)((((\\x20|\\x09)*(\\x0d\\x0a))?(\\x20|\\x09)+)?(([\\x01-\\x08\\x0b\\x0c\\x0e-\\x1f\\x7f]|\\x21|[\\x23-\\x5b]|[\\x5d-\\x7e]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])|(\\([\\x01-\\x09\\x0b\\x0c\\x0d-\\x7f]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}]))))*(((\\x20|\\x09)*(\\x0d\\x0a))?(\\x20|\\x09)+)?(\\x22)))@((([a-zA-Z]|\\d|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])|(([a-zA-Z]|\\d|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])([a-zA-Z]|\\d|-|\\.|_|~|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])*([a-zA-Z]|\\d|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])))\\.)+(([a-zA-Z]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])|(([a-zA-Z]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])([a-zA-Z]|\\d|-|_|~|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])*([a-zA-Z]|[\\x{00A0}-\\x{D7FF}\\x{F900}-\\x{FDCF}\\x{FDF0}-\\x{FFEF}])))\\.?$"
|
||||||
CreditCard string = "^(?:4[0-9]{12}(?:[0-9]{3})?|5[1-5][0-9]{14}|6(?:011|5[0-9][0-9])[0-9]{12}|3[47][0-9]{13}|3(?:0[0-5]|[68][0-9])[0-9]{11}|(?:2131|1800|35\\d{3})\\d{11})$"
|
CreditCard string = "^(?:4[0-9]{12}(?:[0-9]{3})?|5[1-5][0-9]{14}|6(?:011|5[0-9][0-9])[0-9]{12}|3[47][0-9]{13}|3(?:0[0-5]|[68][0-9])[0-9]{11}|(?:2131|1800|35\\d{3})\\d{11})$"
|
||||||
ISBN10 string = "^(?:[0-9]{9}X|[0-9]{10})$"
|
ISBN10 string = "^(?:[0-9]{9}X|[0-9]{10})$"
|
||||||
ISBN13 string = "^(?:[0-9]{13})$"
|
ISBN13 string = "^(?:[0-9]{13})$"
|
||||||
|
@ -14,7 +14,7 @@ const (
|
||||||
UUID string = "^[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}$"
|
UUID string = "^[0-9a-f]{8}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{4}-[0-9a-f]{12}$"
|
||||||
Alpha string = "^[a-zA-Z]+$"
|
Alpha string = "^[a-zA-Z]+$"
|
||||||
Alphanumeric string = "^[a-zA-Z0-9]+$"
|
Alphanumeric string = "^[a-zA-Z0-9]+$"
|
||||||
Numeric string = "^[-+]?[0-9]+$"
|
Numeric string = "^[0-9]+$"
|
||||||
Int string = "^(?:[-+]?(?:0|[1-9][0-9]*))$"
|
Int string = "^(?:[-+]?(?:0|[1-9][0-9]*))$"
|
||||||
Float string = "^(?:[-+]?(?:[0-9]+))?(?:\\.[0-9]*)?(?:[eE][\\+\\-]?(?:[0-9]+))?$"
|
Float string = "^(?:[-+]?(?:[0-9]+))?(?:\\.[0-9]*)?(?:[eE][\\+\\-]?(?:[0-9]+))?$"
|
||||||
Hexadecimal string = "^[0-9a-fA-F]+$"
|
Hexadecimal string = "^[0-9a-fA-F]+$"
|
||||||
|
@ -29,13 +29,22 @@ const (
|
||||||
DataURI string = "^data:.+\\/(.+);base64$"
|
DataURI string = "^data:.+\\/(.+);base64$"
|
||||||
Latitude string = "^[-+]?([1-8]?\\d(\\.\\d+)?|90(\\.0+)?)$"
|
Latitude string = "^[-+]?([1-8]?\\d(\\.\\d+)?|90(\\.0+)?)$"
|
||||||
Longitude string = "^[-+]?(180(\\.0+)?|((1[0-7]\\d)|([1-9]?\\d))(\\.\\d+)?)$"
|
Longitude string = "^[-+]?(180(\\.0+)?|((1[0-7]\\d)|([1-9]?\\d))(\\.\\d+)?)$"
|
||||||
DNSName string = `^([a-zA-Z0-9]{1}[a-zA-Z0-9_-]{1,62}){1}(\.[a-zA-Z0-9]{1}[a-zA-Z0-9_-]{1,62})*$`
|
DNSName string = `^([a-zA-Z0-9_]{1}[a-zA-Z0-9_-]{0,62}){1}(\.[a-zA-Z0-9_]{1}[a-zA-Z0-9_-]{0,62})*[\._]?$`
|
||||||
URL string = `^((ftp|https?):\/\/)?(\S+(:\S*)?@)?((([1-9]\d?|1\d\d|2[01]\d|22[0-3])(\.(1?\d{1,2}|2[0-4]\d|25[0-5])){2}(?:\.([0-9]\d?|1\d\d|2[0-4]\d|25[0-4]))|(([a-zA-Z0-9]+([-\.][a-zA-Z0-9]+)*)|((www\.)?))?(([a-zA-Z\x{00a1}-\x{ffff}0-9]+-?-?)*[a-zA-Z\x{00a1}-\x{ffff}0-9]+)(?:\.([a-zA-Z\x{00a1}-\x{ffff}]{1,}))?))(:(\d{1,5}))?((\/|\?|#)[^\s]*)?$`
|
IP string = `(([0-9a-fA-F]{1,4}:){7,7}[0-9a-fA-F]{1,4}|([0-9a-fA-F]{1,4}:){1,7}:|([0-9a-fA-F]{1,4}:){1,6}:[0-9a-fA-F]{1,4}|([0-9a-fA-F]{1,4}:){1,5}(:[0-9a-fA-F]{1,4}){1,2}|([0-9a-fA-F]{1,4}:){1,4}(:[0-9a-fA-F]{1,4}){1,3}|([0-9a-fA-F]{1,4}:){1,3}(:[0-9a-fA-F]{1,4}){1,4}|([0-9a-fA-F]{1,4}:){1,2}(:[0-9a-fA-F]{1,4}){1,5}|[0-9a-fA-F]{1,4}:((:[0-9a-fA-F]{1,4}){1,6})|:((:[0-9a-fA-F]{1,4}){1,7}|:)|fe80:(:[0-9a-fA-F]{0,4}){0,4}%[0-9a-zA-Z]{1,}|::(ffff(:0{1,4}){0,1}:){0,1}((25[0-5]|(2[0-4]|1{0,1}[0-9]){0,1}[0-9])\.){3,3}(25[0-5]|(2[0-4]|1{0,1}[0-9]){0,1}[0-9])|([0-9a-fA-F]{1,4}:){1,4}:((25[0-5]|(2[0-4]|1{0,1}[0-9]){0,1}[0-9])\.){3,3}(25[0-5]|(2[0-4]|1{0,1}[0-9]){0,1}[0-9]))`
|
||||||
|
URLSchema string = `((ftp|tcp|udp|wss?|https?):\/\/)`
|
||||||
|
URLUsername string = `(\S+(:\S*)?@)`
|
||||||
|
URLPath string = `((\/|\?|#)[^\s]*)`
|
||||||
|
URLPort string = `(:(\d{1,5}))`
|
||||||
|
URLIP string = `([1-9]\d?|1\d\d|2[01]\d|22[0-3])(\.(1?\d{1,2}|2[0-4]\d|25[0-5])){2}(?:\.([0-9]\d?|1\d\d|2[0-4]\d|25[0-4]))`
|
||||||
|
URLSubdomain string = `((www\.)|([a-zA-Z0-9]([-\.][-\._a-zA-Z0-9]+)*))`
|
||||||
|
URL string = `^` + URLSchema + `?` + URLUsername + `?` + `((` + URLIP + `|(\[` + IP + `\])|(([a-zA-Z0-9]([a-zA-Z0-9-_]+)?[a-zA-Z0-9]([-\.][a-zA-Z0-9]+)*)|(` + URLSubdomain + `?))?(([a-zA-Z\x{00a1}-\x{ffff}0-9]+-?-?)*[a-zA-Z\x{00a1}-\x{ffff}0-9]+)(?:\.([a-zA-Z\x{00a1}-\x{ffff}]{1,}))?))\.?` + URLPort + `?` + URLPath + `?$`
|
||||||
SSN string = `^\d{3}[- ]?\d{2}[- ]?\d{4}$`
|
SSN string = `^\d{3}[- ]?\d{2}[- ]?\d{4}$`
|
||||||
WinPath string = `^[a-zA-Z]:\\(?:[^\\/:*?"<>|\r\n]+\\)*[^\\/:*?"<>|\r\n]*$`
|
WinPath string = `^[a-zA-Z]:\\(?:[^\\/:*?"<>|\r\n]+\\)*[^\\/:*?"<>|\r\n]*$`
|
||||||
UnixPath string = `^((?:\/[a-zA-Z0-9\.\:]+(?:_[a-zA-Z0-9\:\.]+)*(?:\-[\:a-zA-Z0-9\.]+)*)+\/?)$`
|
UnixPath string = `^(/[^/\x00]*)+/?$`
|
||||||
Semver string = "^v?(?:0|[1-9]\\d*)\\.(?:0|[1-9]\\d*)\\.(?:0|[1-9]\\d*)(-(0|[1-9]\\d*|\\d*[a-zA-Z-][0-9a-zA-Z-]*)(\\.(0|[1-9]\\d*|\\d*[a-zA-Z-][0-9a-zA-Z-]*))*)?(\\+[0-9a-zA-Z-]+(\\.[0-9a-zA-Z-]+)*)?$"
|
Semver string = "^v?(?:0|[1-9]\\d*)\\.(?:0|[1-9]\\d*)\\.(?:0|[1-9]\\d*)(-(0|[1-9]\\d*|\\d*[a-zA-Z-][0-9a-zA-Z-]*)(\\.(0|[1-9]\\d*|\\d*[a-zA-Z-][0-9a-zA-Z-]*))*)?(\\+[0-9a-zA-Z-]+(\\.[0-9a-zA-Z-]+)*)?$"
|
||||||
tagName string = "valid"
|
tagName string = "valid"
|
||||||
|
hasLowerCase string = ".*[[:lower:]]"
|
||||||
|
hasUpperCase string = ".*[[:upper:]]"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Used by IsFilePath func
|
// Used by IsFilePath func
|
||||||
|
@ -49,7 +58,10 @@ const (
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
rxEmail = regexp.MustCompile(Email)
|
userRegexp = regexp.MustCompile("^[a-zA-Z0-9!#$%&'*+/=?^_`{|}~.-]+$")
|
||||||
|
hostRegexp = regexp.MustCompile("^[^\\s]+\\.[^\\s]+$")
|
||||||
|
userDotRegexp = regexp.MustCompile("(^[.]{1})|([.]{1}$)|([.]{2,})")
|
||||||
|
//rxEmail = regexp.MustCompile(Email)
|
||||||
rxCreditCard = regexp.MustCompile(CreditCard)
|
rxCreditCard = regexp.MustCompile(CreditCard)
|
||||||
rxISBN10 = regexp.MustCompile(ISBN10)
|
rxISBN10 = regexp.MustCompile(ISBN10)
|
||||||
rxISBN13 = regexp.MustCompile(ISBN13)
|
rxISBN13 = regexp.MustCompile(ISBN13)
|
||||||
|
@ -80,4 +92,6 @@ var (
|
||||||
rxWinPath = regexp.MustCompile(WinPath)
|
rxWinPath = regexp.MustCompile(WinPath)
|
||||||
rxUnixPath = regexp.MustCompile(UnixPath)
|
rxUnixPath = regexp.MustCompile(UnixPath)
|
||||||
rxSemver = regexp.MustCompile(Semver)
|
rxSemver = regexp.MustCompile(Semver)
|
||||||
|
rxHasLowerCase = regexp.MustCompile(hasLowerCase)
|
||||||
|
rxHasUpperCase = regexp.MustCompile(hasUpperCase)
|
||||||
)
|
)
|
||||||
|
|
334
vendor/github.com/asaskevich/govalidator/types.go
generated
vendored
334
vendor/github.com/asaskevich/govalidator/types.go
generated
vendored
|
@ -29,15 +29,23 @@ type stringValues []reflect.Value
|
||||||
// ParamTagMap is a map of functions accept variants parameters
|
// ParamTagMap is a map of functions accept variants parameters
|
||||||
var ParamTagMap = map[string]ParamValidator{
|
var ParamTagMap = map[string]ParamValidator{
|
||||||
"length": ByteLength,
|
"length": ByteLength,
|
||||||
|
"range": Range,
|
||||||
|
"runelength": RuneLength,
|
||||||
"stringlength": StringLength,
|
"stringlength": StringLength,
|
||||||
"matches": StringMatches,
|
"matches": StringMatches,
|
||||||
|
"in": isInRaw,
|
||||||
|
"rsapub": IsRsaPub,
|
||||||
}
|
}
|
||||||
|
|
||||||
// ParamTagRegexMap maps param tags to their respective regexes.
|
// ParamTagRegexMap maps param tags to their respective regexes.
|
||||||
var ParamTagRegexMap = map[string]*regexp.Regexp{
|
var ParamTagRegexMap = map[string]*regexp.Regexp{
|
||||||
|
"range": regexp.MustCompile("^range\\((\\d+)\\|(\\d+)\\)$"),
|
||||||
"length": regexp.MustCompile("^length\\((\\d+)\\|(\\d+)\\)$"),
|
"length": regexp.MustCompile("^length\\((\\d+)\\|(\\d+)\\)$"),
|
||||||
|
"runelength": regexp.MustCompile("^runelength\\((\\d+)\\|(\\d+)\\)$"),
|
||||||
"stringlength": regexp.MustCompile("^stringlength\\((\\d+)\\|(\\d+)\\)$"),
|
"stringlength": regexp.MustCompile("^stringlength\\((\\d+)\\|(\\d+)\\)$"),
|
||||||
"matches": regexp.MustCompile(`matches\(([^)]+)\)`),
|
"in": regexp.MustCompile(`^in\((.*)\)`),
|
||||||
|
"matches": regexp.MustCompile(`^matches\((.+)\)$`),
|
||||||
|
"rsapub": regexp.MustCompile("^rsapub\\((\\d+)\\)$"),
|
||||||
}
|
}
|
||||||
|
|
||||||
type customTypeTagMap struct {
|
type customTypeTagMap struct {
|
||||||
|
@ -66,53 +74,58 @@ var CustomTypeTagMap = &customTypeTagMap{validators: make(map[string]CustomTypeV
|
||||||
|
|
||||||
// TagMap is a map of functions, that can be used as tags for ValidateStruct function.
|
// TagMap is a map of functions, that can be used as tags for ValidateStruct function.
|
||||||
var TagMap = map[string]Validator{
|
var TagMap = map[string]Validator{
|
||||||
"email": IsEmail,
|
"email": IsEmail,
|
||||||
"url": IsURL,
|
"url": IsURL,
|
||||||
"dialstring": IsDialString,
|
"dialstring": IsDialString,
|
||||||
"requrl": IsRequestURL,
|
"requrl": IsRequestURL,
|
||||||
"requri": IsRequestURI,
|
"requri": IsRequestURI,
|
||||||
"alpha": IsAlpha,
|
"alpha": IsAlpha,
|
||||||
"utfletter": IsUTFLetter,
|
"utfletter": IsUTFLetter,
|
||||||
"alphanum": IsAlphanumeric,
|
"alphanum": IsAlphanumeric,
|
||||||
"utfletternum": IsUTFLetterNumeric,
|
"utfletternum": IsUTFLetterNumeric,
|
||||||
"numeric": IsNumeric,
|
"numeric": IsNumeric,
|
||||||
"utfnumeric": IsUTFNumeric,
|
"utfnumeric": IsUTFNumeric,
|
||||||
"utfdigit": IsUTFDigit,
|
"utfdigit": IsUTFDigit,
|
||||||
"hexadecimal": IsHexadecimal,
|
"hexadecimal": IsHexadecimal,
|
||||||
"hexcolor": IsHexcolor,
|
"hexcolor": IsHexcolor,
|
||||||
"rgbcolor": IsRGBcolor,
|
"rgbcolor": IsRGBcolor,
|
||||||
"lowercase": IsLowerCase,
|
"lowercase": IsLowerCase,
|
||||||
"uppercase": IsUpperCase,
|
"uppercase": IsUpperCase,
|
||||||
"int": IsInt,
|
"int": IsInt,
|
||||||
"float": IsFloat,
|
"float": IsFloat,
|
||||||
"null": IsNull,
|
"null": IsNull,
|
||||||
"uuid": IsUUID,
|
"uuid": IsUUID,
|
||||||
"uuidv3": IsUUIDv3,
|
"uuidv3": IsUUIDv3,
|
||||||
"uuidv4": IsUUIDv4,
|
"uuidv4": IsUUIDv4,
|
||||||
"uuidv5": IsUUIDv5,
|
"uuidv5": IsUUIDv5,
|
||||||
"creditcard": IsCreditCard,
|
"creditcard": IsCreditCard,
|
||||||
"isbn10": IsISBN10,
|
"isbn10": IsISBN10,
|
||||||
"isbn13": IsISBN13,
|
"isbn13": IsISBN13,
|
||||||
"json": IsJSON,
|
"json": IsJSON,
|
||||||
"multibyte": IsMultibyte,
|
"multibyte": IsMultibyte,
|
||||||
"ascii": IsASCII,
|
"ascii": IsASCII,
|
||||||
"printableascii": IsPrintableASCII,
|
"printableascii": IsPrintableASCII,
|
||||||
"fullwidth": IsFullWidth,
|
"fullwidth": IsFullWidth,
|
||||||
"halfwidth": IsHalfWidth,
|
"halfwidth": IsHalfWidth,
|
||||||
"variablewidth": IsVariableWidth,
|
"variablewidth": IsVariableWidth,
|
||||||
"base64": IsBase64,
|
"base64": IsBase64,
|
||||||
"datauri": IsDataURI,
|
"datauri": IsDataURI,
|
||||||
"ip": IsIP,
|
"ip": IsIP,
|
||||||
"port": IsPort,
|
"port": IsPort,
|
||||||
"ipv4": IsIPv4,
|
"ipv4": IsIPv4,
|
||||||
"ipv6": IsIPv6,
|
"ipv6": IsIPv6,
|
||||||
"dns": IsDNSName,
|
"dns": IsDNSName,
|
||||||
"host": IsHost,
|
"host": IsHost,
|
||||||
"mac": IsMAC,
|
"mac": IsMAC,
|
||||||
"latitude": IsLatitude,
|
"latitude": IsLatitude,
|
||||||
"longitude": IsLongitude,
|
"longitude": IsLongitude,
|
||||||
"ssn": IsSSN,
|
"ssn": IsSSN,
|
||||||
"semver": IsSemver,
|
"semver": IsSemver,
|
||||||
|
"rfc3339": IsRFC3339,
|
||||||
|
"rfc3339WithoutZone": IsRFC3339WithoutZone,
|
||||||
|
"ISO3166Alpha2": IsISO3166Alpha2,
|
||||||
|
"ISO3166Alpha3": IsISO3166Alpha3,
|
||||||
|
"ISO4217": IsISO4217,
|
||||||
}
|
}
|
||||||
|
|
||||||
// ISO3166Entry stores country codes
|
// ISO3166Entry stores country codes
|
||||||
|
@ -376,3 +389,228 @@ var ISO3166List = []ISO3166Entry{
|
||||||
{"Yemen", "Yémen (le)", "YE", "YEM", "887"},
|
{"Yemen", "Yémen (le)", "YE", "YEM", "887"},
|
||||||
{"Zambia", "Zambie (la)", "ZM", "ZMB", "894"},
|
{"Zambia", "Zambie (la)", "ZM", "ZMB", "894"},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ISO4217List is the list of ISO currency codes
|
||||||
|
var ISO4217List = []string{
|
||||||
|
"AED", "AFN", "ALL", "AMD", "ANG", "AOA", "ARS", "AUD", "AWG", "AZN",
|
||||||
|
"BAM", "BBD", "BDT", "BGN", "BHD", "BIF", "BMD", "BND", "BOB", "BOV", "BRL", "BSD", "BTN", "BWP", "BYN", "BZD",
|
||||||
|
"CAD", "CDF", "CHE", "CHF", "CHW", "CLF", "CLP", "CNY", "COP", "COU", "CRC", "CUC", "CUP", "CVE", "CZK",
|
||||||
|
"DJF", "DKK", "DOP", "DZD",
|
||||||
|
"EGP", "ERN", "ETB", "EUR",
|
||||||
|
"FJD", "FKP",
|
||||||
|
"GBP", "GEL", "GHS", "GIP", "GMD", "GNF", "GTQ", "GYD",
|
||||||
|
"HKD", "HNL", "HRK", "HTG", "HUF",
|
||||||
|
"IDR", "ILS", "INR", "IQD", "IRR", "ISK",
|
||||||
|
"JMD", "JOD", "JPY",
|
||||||
|
"KES", "KGS", "KHR", "KMF", "KPW", "KRW", "KWD", "KYD", "KZT",
|
||||||
|
"LAK", "LBP", "LKR", "LRD", "LSL", "LYD",
|
||||||
|
"MAD", "MDL", "MGA", "MKD", "MMK", "MNT", "MOP", "MRO", "MUR", "MVR", "MWK", "MXN", "MXV", "MYR", "MZN",
|
||||||
|
"NAD", "NGN", "NIO", "NOK", "NPR", "NZD",
|
||||||
|
"OMR",
|
||||||
|
"PAB", "PEN", "PGK", "PHP", "PKR", "PLN", "PYG",
|
||||||
|
"QAR",
|
||||||
|
"RON", "RSD", "RUB", "RWF",
|
||||||
|
"SAR", "SBD", "SCR", "SDG", "SEK", "SGD", "SHP", "SLL", "SOS", "SRD", "SSP", "STD", "SVC", "SYP", "SZL",
|
||||||
|
"THB", "TJS", "TMT", "TND", "TOP", "TRY", "TTD", "TWD", "TZS",
|
||||||
|
"UAH", "UGX", "USD", "USN", "UYI", "UYU", "UZS",
|
||||||
|
"VEF", "VND", "VUV",
|
||||||
|
"WST",
|
||||||
|
"XAF", "XAG", "XAU", "XBA", "XBB", "XBC", "XBD", "XCD", "XDR", "XOF", "XPD", "XPF", "XPT", "XSU", "XTS", "XUA", "XXX",
|
||||||
|
"YER",
|
||||||
|
"ZAR", "ZMW", "ZWL",
|
||||||
|
}
|
||||||
|
|
||||||
|
// ISO693Entry stores ISO language codes
|
||||||
|
type ISO693Entry struct {
|
||||||
|
Alpha3bCode string
|
||||||
|
Alpha2Code string
|
||||||
|
English string
|
||||||
|
}
|
||||||
|
|
||||||
|
//ISO693List based on http://data.okfn.org/data/core/language-codes/r/language-codes-3b2.json
|
||||||
|
var ISO693List = []ISO693Entry{
|
||||||
|
{Alpha3bCode: "aar", Alpha2Code: "aa", English: "Afar"},
|
||||||
|
{Alpha3bCode: "abk", Alpha2Code: "ab", English: "Abkhazian"},
|
||||||
|
{Alpha3bCode: "afr", Alpha2Code: "af", English: "Afrikaans"},
|
||||||
|
{Alpha3bCode: "aka", Alpha2Code: "ak", English: "Akan"},
|
||||||
|
{Alpha3bCode: "alb", Alpha2Code: "sq", English: "Albanian"},
|
||||||
|
{Alpha3bCode: "amh", Alpha2Code: "am", English: "Amharic"},
|
||||||
|
{Alpha3bCode: "ara", Alpha2Code: "ar", English: "Arabic"},
|
||||||
|
{Alpha3bCode: "arg", Alpha2Code: "an", English: "Aragonese"},
|
||||||
|
{Alpha3bCode: "arm", Alpha2Code: "hy", English: "Armenian"},
|
||||||
|
{Alpha3bCode: "asm", Alpha2Code: "as", English: "Assamese"},
|
||||||
|
{Alpha3bCode: "ava", Alpha2Code: "av", English: "Avaric"},
|
||||||
|
{Alpha3bCode: "ave", Alpha2Code: "ae", English: "Avestan"},
|
||||||
|
{Alpha3bCode: "aym", Alpha2Code: "ay", English: "Aymara"},
|
||||||
|
{Alpha3bCode: "aze", Alpha2Code: "az", English: "Azerbaijani"},
|
||||||
|
{Alpha3bCode: "bak", Alpha2Code: "ba", English: "Bashkir"},
|
||||||
|
{Alpha3bCode: "bam", Alpha2Code: "bm", English: "Bambara"},
|
||||||
|
{Alpha3bCode: "baq", Alpha2Code: "eu", English: "Basque"},
|
||||||
|
{Alpha3bCode: "bel", Alpha2Code: "be", English: "Belarusian"},
|
||||||
|
{Alpha3bCode: "ben", Alpha2Code: "bn", English: "Bengali"},
|
||||||
|
{Alpha3bCode: "bih", Alpha2Code: "bh", English: "Bihari languages"},
|
||||||
|
{Alpha3bCode: "bis", Alpha2Code: "bi", English: "Bislama"},
|
||||||
|
{Alpha3bCode: "bos", Alpha2Code: "bs", English: "Bosnian"},
|
||||||
|
{Alpha3bCode: "bre", Alpha2Code: "br", English: "Breton"},
|
||||||
|
{Alpha3bCode: "bul", Alpha2Code: "bg", English: "Bulgarian"},
|
||||||
|
{Alpha3bCode: "bur", Alpha2Code: "my", English: "Burmese"},
|
||||||
|
{Alpha3bCode: "cat", Alpha2Code: "ca", English: "Catalan; Valencian"},
|
||||||
|
{Alpha3bCode: "cha", Alpha2Code: "ch", English: "Chamorro"},
|
||||||
|
{Alpha3bCode: "che", Alpha2Code: "ce", English: "Chechen"},
|
||||||
|
{Alpha3bCode: "chi", Alpha2Code: "zh", English: "Chinese"},
|
||||||
|
{Alpha3bCode: "chu", Alpha2Code: "cu", English: "Church Slavic; Old Slavonic; Church Slavonic; Old Bulgarian; Old Church Slavonic"},
|
||||||
|
{Alpha3bCode: "chv", Alpha2Code: "cv", English: "Chuvash"},
|
||||||
|
{Alpha3bCode: "cor", Alpha2Code: "kw", English: "Cornish"},
|
||||||
|
{Alpha3bCode: "cos", Alpha2Code: "co", English: "Corsican"},
|
||||||
|
{Alpha3bCode: "cre", Alpha2Code: "cr", English: "Cree"},
|
||||||
|
{Alpha3bCode: "cze", Alpha2Code: "cs", English: "Czech"},
|
||||||
|
{Alpha3bCode: "dan", Alpha2Code: "da", English: "Danish"},
|
||||||
|
{Alpha3bCode: "div", Alpha2Code: "dv", English: "Divehi; Dhivehi; Maldivian"},
|
||||||
|
{Alpha3bCode: "dut", Alpha2Code: "nl", English: "Dutch; Flemish"},
|
||||||
|
{Alpha3bCode: "dzo", Alpha2Code: "dz", English: "Dzongkha"},
|
||||||
|
{Alpha3bCode: "eng", Alpha2Code: "en", English: "English"},
|
||||||
|
{Alpha3bCode: "epo", Alpha2Code: "eo", English: "Esperanto"},
|
||||||
|
{Alpha3bCode: "est", Alpha2Code: "et", English: "Estonian"},
|
||||||
|
{Alpha3bCode: "ewe", Alpha2Code: "ee", English: "Ewe"},
|
||||||
|
{Alpha3bCode: "fao", Alpha2Code: "fo", English: "Faroese"},
|
||||||
|
{Alpha3bCode: "fij", Alpha2Code: "fj", English: "Fijian"},
|
||||||
|
{Alpha3bCode: "fin", Alpha2Code: "fi", English: "Finnish"},
|
||||||
|
{Alpha3bCode: "fre", Alpha2Code: "fr", English: "French"},
|
||||||
|
{Alpha3bCode: "fry", Alpha2Code: "fy", English: "Western Frisian"},
|
||||||
|
{Alpha3bCode: "ful", Alpha2Code: "ff", English: "Fulah"},
|
||||||
|
{Alpha3bCode: "geo", Alpha2Code: "ka", English: "Georgian"},
|
||||||
|
{Alpha3bCode: "ger", Alpha2Code: "de", English: "German"},
|
||||||
|
{Alpha3bCode: "gla", Alpha2Code: "gd", English: "Gaelic; Scottish Gaelic"},
|
||||||
|
{Alpha3bCode: "gle", Alpha2Code: "ga", English: "Irish"},
|
||||||
|
{Alpha3bCode: "glg", Alpha2Code: "gl", English: "Galician"},
|
||||||
|
{Alpha3bCode: "glv", Alpha2Code: "gv", English: "Manx"},
|
||||||
|
{Alpha3bCode: "gre", Alpha2Code: "el", English: "Greek, Modern (1453-)"},
|
||||||
|
{Alpha3bCode: "grn", Alpha2Code: "gn", English: "Guarani"},
|
||||||
|
{Alpha3bCode: "guj", Alpha2Code: "gu", English: "Gujarati"},
|
||||||
|
{Alpha3bCode: "hat", Alpha2Code: "ht", English: "Haitian; Haitian Creole"},
|
||||||
|
{Alpha3bCode: "hau", Alpha2Code: "ha", English: "Hausa"},
|
||||||
|
{Alpha3bCode: "heb", Alpha2Code: "he", English: "Hebrew"},
|
||||||
|
{Alpha3bCode: "her", Alpha2Code: "hz", English: "Herero"},
|
||||||
|
{Alpha3bCode: "hin", Alpha2Code: "hi", English: "Hindi"},
|
||||||
|
{Alpha3bCode: "hmo", Alpha2Code: "ho", English: "Hiri Motu"},
|
||||||
|
{Alpha3bCode: "hrv", Alpha2Code: "hr", English: "Croatian"},
|
||||||
|
{Alpha3bCode: "hun", Alpha2Code: "hu", English: "Hungarian"},
|
||||||
|
{Alpha3bCode: "ibo", Alpha2Code: "ig", English: "Igbo"},
|
||||||
|
{Alpha3bCode: "ice", Alpha2Code: "is", English: "Icelandic"},
|
||||||
|
{Alpha3bCode: "ido", Alpha2Code: "io", English: "Ido"},
|
||||||
|
{Alpha3bCode: "iii", Alpha2Code: "ii", English: "Sichuan Yi; Nuosu"},
|
||||||
|
{Alpha3bCode: "iku", Alpha2Code: "iu", English: "Inuktitut"},
|
||||||
|
{Alpha3bCode: "ile", Alpha2Code: "ie", English: "Interlingue; Occidental"},
|
||||||
|
{Alpha3bCode: "ina", Alpha2Code: "ia", English: "Interlingua (International Auxiliary Language Association)"},
|
||||||
|
{Alpha3bCode: "ind", Alpha2Code: "id", English: "Indonesian"},
|
||||||
|
{Alpha3bCode: "ipk", Alpha2Code: "ik", English: "Inupiaq"},
|
||||||
|
{Alpha3bCode: "ita", Alpha2Code: "it", English: "Italian"},
|
||||||
|
{Alpha3bCode: "jav", Alpha2Code: "jv", English: "Javanese"},
|
||||||
|
{Alpha3bCode: "jpn", Alpha2Code: "ja", English: "Japanese"},
|
||||||
|
{Alpha3bCode: "kal", Alpha2Code: "kl", English: "Kalaallisut; Greenlandic"},
|
||||||
|
{Alpha3bCode: "kan", Alpha2Code: "kn", English: "Kannada"},
|
||||||
|
{Alpha3bCode: "kas", Alpha2Code: "ks", English: "Kashmiri"},
|
||||||
|
{Alpha3bCode: "kau", Alpha2Code: "kr", English: "Kanuri"},
|
||||||
|
{Alpha3bCode: "kaz", Alpha2Code: "kk", English: "Kazakh"},
|
||||||
|
{Alpha3bCode: "khm", Alpha2Code: "km", English: "Central Khmer"},
|
||||||
|
{Alpha3bCode: "kik", Alpha2Code: "ki", English: "Kikuyu; Gikuyu"},
|
||||||
|
{Alpha3bCode: "kin", Alpha2Code: "rw", English: "Kinyarwanda"},
|
||||||
|
{Alpha3bCode: "kir", Alpha2Code: "ky", English: "Kirghiz; Kyrgyz"},
|
||||||
|
{Alpha3bCode: "kom", Alpha2Code: "kv", English: "Komi"},
|
||||||
|
{Alpha3bCode: "kon", Alpha2Code: "kg", English: "Kongo"},
|
||||||
|
{Alpha3bCode: "kor", Alpha2Code: "ko", English: "Korean"},
|
||||||
|
{Alpha3bCode: "kua", Alpha2Code: "kj", English: "Kuanyama; Kwanyama"},
|
||||||
|
{Alpha3bCode: "kur", Alpha2Code: "ku", English: "Kurdish"},
|
||||||
|
{Alpha3bCode: "lao", Alpha2Code: "lo", English: "Lao"},
|
||||||
|
{Alpha3bCode: "lat", Alpha2Code: "la", English: "Latin"},
|
||||||
|
{Alpha3bCode: "lav", Alpha2Code: "lv", English: "Latvian"},
|
||||||
|
{Alpha3bCode: "lim", Alpha2Code: "li", English: "Limburgan; Limburger; Limburgish"},
|
||||||
|
{Alpha3bCode: "lin", Alpha2Code: "ln", English: "Lingala"},
|
||||||
|
{Alpha3bCode: "lit", Alpha2Code: "lt", English: "Lithuanian"},
|
||||||
|
{Alpha3bCode: "ltz", Alpha2Code: "lb", English: "Luxembourgish; Letzeburgesch"},
|
||||||
|
{Alpha3bCode: "lub", Alpha2Code: "lu", English: "Luba-Katanga"},
|
||||||
|
{Alpha3bCode: "lug", Alpha2Code: "lg", English: "Ganda"},
|
||||||
|
{Alpha3bCode: "mac", Alpha2Code: "mk", English: "Macedonian"},
|
||||||
|
{Alpha3bCode: "mah", Alpha2Code: "mh", English: "Marshallese"},
|
||||||
|
{Alpha3bCode: "mal", Alpha2Code: "ml", English: "Malayalam"},
|
||||||
|
{Alpha3bCode: "mao", Alpha2Code: "mi", English: "Maori"},
|
||||||
|
{Alpha3bCode: "mar", Alpha2Code: "mr", English: "Marathi"},
|
||||||
|
{Alpha3bCode: "may", Alpha2Code: "ms", English: "Malay"},
|
||||||
|
{Alpha3bCode: "mlg", Alpha2Code: "mg", English: "Malagasy"},
|
||||||
|
{Alpha3bCode: "mlt", Alpha2Code: "mt", English: "Maltese"},
|
||||||
|
{Alpha3bCode: "mon", Alpha2Code: "mn", English: "Mongolian"},
|
||||||
|
{Alpha3bCode: "nau", Alpha2Code: "na", English: "Nauru"},
|
||||||
|
{Alpha3bCode: "nav", Alpha2Code: "nv", English: "Navajo; Navaho"},
|
||||||
|
{Alpha3bCode: "nbl", Alpha2Code: "nr", English: "Ndebele, South; South Ndebele"},
|
||||||
|
{Alpha3bCode: "nde", Alpha2Code: "nd", English: "Ndebele, North; North Ndebele"},
|
||||||
|
{Alpha3bCode: "ndo", Alpha2Code: "ng", English: "Ndonga"},
|
||||||
|
{Alpha3bCode: "nep", Alpha2Code: "ne", English: "Nepali"},
|
||||||
|
{Alpha3bCode: "nno", Alpha2Code: "nn", English: "Norwegian Nynorsk; Nynorsk, Norwegian"},
|
||||||
|
{Alpha3bCode: "nob", Alpha2Code: "nb", English: "Bokmål, Norwegian; Norwegian Bokmål"},
|
||||||
|
{Alpha3bCode: "nor", Alpha2Code: "no", English: "Norwegian"},
|
||||||
|
{Alpha3bCode: "nya", Alpha2Code: "ny", English: "Chichewa; Chewa; Nyanja"},
|
||||||
|
{Alpha3bCode: "oci", Alpha2Code: "oc", English: "Occitan (post 1500); Provençal"},
|
||||||
|
{Alpha3bCode: "oji", Alpha2Code: "oj", English: "Ojibwa"},
|
||||||
|
{Alpha3bCode: "ori", Alpha2Code: "or", English: "Oriya"},
|
||||||
|
{Alpha3bCode: "orm", Alpha2Code: "om", English: "Oromo"},
|
||||||
|
{Alpha3bCode: "oss", Alpha2Code: "os", English: "Ossetian; Ossetic"},
|
||||||
|
{Alpha3bCode: "pan", Alpha2Code: "pa", English: "Panjabi; Punjabi"},
|
||||||
|
{Alpha3bCode: "per", Alpha2Code: "fa", English: "Persian"},
|
||||||
|
{Alpha3bCode: "pli", Alpha2Code: "pi", English: "Pali"},
|
||||||
|
{Alpha3bCode: "pol", Alpha2Code: "pl", English: "Polish"},
|
||||||
|
{Alpha3bCode: "por", Alpha2Code: "pt", English: "Portuguese"},
|
||||||
|
{Alpha3bCode: "pus", Alpha2Code: "ps", English: "Pushto; Pashto"},
|
||||||
|
{Alpha3bCode: "que", Alpha2Code: "qu", English: "Quechua"},
|
||||||
|
{Alpha3bCode: "roh", Alpha2Code: "rm", English: "Romansh"},
|
||||||
|
{Alpha3bCode: "rum", Alpha2Code: "ro", English: "Romanian; Moldavian; Moldovan"},
|
||||||
|
{Alpha3bCode: "run", Alpha2Code: "rn", English: "Rundi"},
|
||||||
|
{Alpha3bCode: "rus", Alpha2Code: "ru", English: "Russian"},
|
||||||
|
{Alpha3bCode: "sag", Alpha2Code: "sg", English: "Sango"},
|
||||||
|
{Alpha3bCode: "san", Alpha2Code: "sa", English: "Sanskrit"},
|
||||||
|
{Alpha3bCode: "sin", Alpha2Code: "si", English: "Sinhala; Sinhalese"},
|
||||||
|
{Alpha3bCode: "slo", Alpha2Code: "sk", English: "Slovak"},
|
||||||
|
{Alpha3bCode: "slv", Alpha2Code: "sl", English: "Slovenian"},
|
||||||
|
{Alpha3bCode: "sme", Alpha2Code: "se", English: "Northern Sami"},
|
||||||
|
{Alpha3bCode: "smo", Alpha2Code: "sm", English: "Samoan"},
|
||||||
|
{Alpha3bCode: "sna", Alpha2Code: "sn", English: "Shona"},
|
||||||
|
{Alpha3bCode: "snd", Alpha2Code: "sd", English: "Sindhi"},
|
||||||
|
{Alpha3bCode: "som", Alpha2Code: "so", English: "Somali"},
|
||||||
|
{Alpha3bCode: "sot", Alpha2Code: "st", English: "Sotho, Southern"},
|
||||||
|
{Alpha3bCode: "spa", Alpha2Code: "es", English: "Spanish; Castilian"},
|
||||||
|
{Alpha3bCode: "srd", Alpha2Code: "sc", English: "Sardinian"},
|
||||||
|
{Alpha3bCode: "srp", Alpha2Code: "sr", English: "Serbian"},
|
||||||
|
{Alpha3bCode: "ssw", Alpha2Code: "ss", English: "Swati"},
|
||||||
|
{Alpha3bCode: "sun", Alpha2Code: "su", English: "Sundanese"},
|
||||||
|
{Alpha3bCode: "swa", Alpha2Code: "sw", English: "Swahili"},
|
||||||
|
{Alpha3bCode: "swe", Alpha2Code: "sv", English: "Swedish"},
|
||||||
|
{Alpha3bCode: "tah", Alpha2Code: "ty", English: "Tahitian"},
|
||||||
|
{Alpha3bCode: "tam", Alpha2Code: "ta", English: "Tamil"},
|
||||||
|
{Alpha3bCode: "tat", Alpha2Code: "tt", English: "Tatar"},
|
||||||
|
{Alpha3bCode: "tel", Alpha2Code: "te", English: "Telugu"},
|
||||||
|
{Alpha3bCode: "tgk", Alpha2Code: "tg", English: "Tajik"},
|
||||||
|
{Alpha3bCode: "tgl", Alpha2Code: "tl", English: "Tagalog"},
|
||||||
|
{Alpha3bCode: "tha", Alpha2Code: "th", English: "Thai"},
|
||||||
|
{Alpha3bCode: "tib", Alpha2Code: "bo", English: "Tibetan"},
|
||||||
|
{Alpha3bCode: "tir", Alpha2Code: "ti", English: "Tigrinya"},
|
||||||
|
{Alpha3bCode: "ton", Alpha2Code: "to", English: "Tonga (Tonga Islands)"},
|
||||||
|
{Alpha3bCode: "tsn", Alpha2Code: "tn", English: "Tswana"},
|
||||||
|
{Alpha3bCode: "tso", Alpha2Code: "ts", English: "Tsonga"},
|
||||||
|
{Alpha3bCode: "tuk", Alpha2Code: "tk", English: "Turkmen"},
|
||||||
|
{Alpha3bCode: "tur", Alpha2Code: "tr", English: "Turkish"},
|
||||||
|
{Alpha3bCode: "twi", Alpha2Code: "tw", English: "Twi"},
|
||||||
|
{Alpha3bCode: "uig", Alpha2Code: "ug", English: "Uighur; Uyghur"},
|
||||||
|
{Alpha3bCode: "ukr", Alpha2Code: "uk", English: "Ukrainian"},
|
||||||
|
{Alpha3bCode: "urd", Alpha2Code: "ur", English: "Urdu"},
|
||||||
|
{Alpha3bCode: "uzb", Alpha2Code: "uz", English: "Uzbek"},
|
||||||
|
{Alpha3bCode: "ven", Alpha2Code: "ve", English: "Venda"},
|
||||||
|
{Alpha3bCode: "vie", Alpha2Code: "vi", English: "Vietnamese"},
|
||||||
|
{Alpha3bCode: "vol", Alpha2Code: "vo", English: "Volapük"},
|
||||||
|
{Alpha3bCode: "wel", Alpha2Code: "cy", English: "Welsh"},
|
||||||
|
{Alpha3bCode: "wln", Alpha2Code: "wa", English: "Walloon"},
|
||||||
|
{Alpha3bCode: "wol", Alpha2Code: "wo", English: "Wolof"},
|
||||||
|
{Alpha3bCode: "xho", Alpha2Code: "xh", English: "Xhosa"},
|
||||||
|
{Alpha3bCode: "yid", Alpha2Code: "yi", English: "Yiddish"},
|
||||||
|
{Alpha3bCode: "yor", Alpha2Code: "yo", English: "Yoruba"},
|
||||||
|
{Alpha3bCode: "zha", Alpha2Code: "za", English: "Zhuang; Chuang"},
|
||||||
|
{Alpha3bCode: "zul", Alpha2Code: "zu", English: "Zulu"},
|
||||||
|
}
|
||||||
|
|
83
vendor/github.com/asaskevich/govalidator/utils.go
generated
vendored
83
vendor/github.com/asaskevich/govalidator/utils.go
generated
vendored
|
@ -4,10 +4,12 @@ import (
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"html"
|
"html"
|
||||||
|
"math"
|
||||||
"path"
|
"path"
|
||||||
"regexp"
|
"regexp"
|
||||||
"strings"
|
"strings"
|
||||||
"unicode"
|
"unicode"
|
||||||
|
"unicode/utf8"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Contains check if the string contains the substring.
|
// Contains check if the string contains the substring.
|
||||||
|
@ -25,27 +27,21 @@ func Matches(str, pattern string) bool {
|
||||||
// LeftTrim trim characters from the left-side of the input.
|
// LeftTrim trim characters from the left-side of the input.
|
||||||
// If second argument is empty, it's will be remove leading spaces.
|
// If second argument is empty, it's will be remove leading spaces.
|
||||||
func LeftTrim(str, chars string) string {
|
func LeftTrim(str, chars string) string {
|
||||||
pattern := ""
|
|
||||||
if chars == "" {
|
if chars == "" {
|
||||||
pattern = "^\\s+"
|
return strings.TrimLeftFunc(str, unicode.IsSpace)
|
||||||
} else {
|
|
||||||
pattern = "^[" + chars + "]+"
|
|
||||||
}
|
}
|
||||||
r, _ := regexp.Compile(pattern)
|
r, _ := regexp.Compile("^[" + chars + "]+")
|
||||||
return string(r.ReplaceAll([]byte(str), []byte("")))
|
return r.ReplaceAllString(str, "")
|
||||||
}
|
}
|
||||||
|
|
||||||
// RightTrim trim characters from the right-side of the input.
|
// RightTrim trim characters from the right-side of the input.
|
||||||
// If second argument is empty, it's will be remove spaces.
|
// If second argument is empty, it's will be remove spaces.
|
||||||
func RightTrim(str, chars string) string {
|
func RightTrim(str, chars string) string {
|
||||||
pattern := ""
|
|
||||||
if chars == "" {
|
if chars == "" {
|
||||||
pattern = "\\s+$"
|
return strings.TrimRightFunc(str, unicode.IsSpace)
|
||||||
} else {
|
|
||||||
pattern = "[" + chars + "]+$"
|
|
||||||
}
|
}
|
||||||
r, _ := regexp.Compile(pattern)
|
r, _ := regexp.Compile("[" + chars + "]+$")
|
||||||
return string(r.ReplaceAll([]byte(str), []byte("")))
|
return r.ReplaceAllString(str, "")
|
||||||
}
|
}
|
||||||
|
|
||||||
// Trim trim characters from both sides of the input.
|
// Trim trim characters from both sides of the input.
|
||||||
|
@ -58,14 +54,14 @@ func Trim(str, chars string) string {
|
||||||
func WhiteList(str, chars string) string {
|
func WhiteList(str, chars string) string {
|
||||||
pattern := "[^" + chars + "]+"
|
pattern := "[^" + chars + "]+"
|
||||||
r, _ := regexp.Compile(pattern)
|
r, _ := regexp.Compile(pattern)
|
||||||
return string(r.ReplaceAll([]byte(str), []byte("")))
|
return r.ReplaceAllString(str, "")
|
||||||
}
|
}
|
||||||
|
|
||||||
// BlackList remove characters that appear in the blacklist.
|
// BlackList remove characters that appear in the blacklist.
|
||||||
func BlackList(str, chars string) string {
|
func BlackList(str, chars string) string {
|
||||||
pattern := "[" + chars + "]+"
|
pattern := "[" + chars + "]+"
|
||||||
r, _ := regexp.Compile(pattern)
|
r, _ := regexp.Compile(pattern)
|
||||||
return string(r.ReplaceAll([]byte(str), []byte("")))
|
return r.ReplaceAllString(str, "")
|
||||||
}
|
}
|
||||||
|
|
||||||
// StripLow remove characters with a numerical value < 32 and 127, mostly control characters.
|
// StripLow remove characters with a numerical value < 32 and 127, mostly control characters.
|
||||||
|
@ -83,7 +79,7 @@ func StripLow(str string, keepNewLines bool) string {
|
||||||
// ReplacePattern replace regular expression pattern in string
|
// ReplacePattern replace regular expression pattern in string
|
||||||
func ReplacePattern(str, pattern, replace string) string {
|
func ReplacePattern(str, pattern, replace string) string {
|
||||||
r, _ := regexp.Compile(pattern)
|
r, _ := regexp.Compile(pattern)
|
||||||
return string(r.ReplaceAll([]byte(str), []byte(replace)))
|
return r.ReplaceAllString(str, replace)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Escape replace <, >, & and " with HTML entities.
|
// Escape replace <, >, & and " with HTML entities.
|
||||||
|
@ -112,7 +108,9 @@ func CamelCaseToUnderscore(str string) string {
|
||||||
var output []rune
|
var output []rune
|
||||||
var segment []rune
|
var segment []rune
|
||||||
for _, r := range str {
|
for _, r := range str {
|
||||||
if !unicode.IsLower(r) {
|
|
||||||
|
// not treat number as separate segment
|
||||||
|
if !unicode.IsLower(r) && string(r) != "_" && !unicode.IsNumber(r) {
|
||||||
output = addSegment(output, segment)
|
output = addSegment(output, segment)
|
||||||
segment = nil
|
segment = nil
|
||||||
}
|
}
|
||||||
|
@ -211,3 +209,56 @@ func Truncate(str string, length int, ending string) string {
|
||||||
|
|
||||||
return str
|
return str
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// PadLeft pad left side of string if size of string is less then indicated pad length
|
||||||
|
func PadLeft(str string, padStr string, padLen int) string {
|
||||||
|
return buildPadStr(str, padStr, padLen, true, false)
|
||||||
|
}
|
||||||
|
|
||||||
|
// PadRight pad right side of string if size of string is less then indicated pad length
|
||||||
|
func PadRight(str string, padStr string, padLen int) string {
|
||||||
|
return buildPadStr(str, padStr, padLen, false, true)
|
||||||
|
}
|
||||||
|
|
||||||
|
// PadBoth pad sides of string if size of string is less then indicated pad length
|
||||||
|
func PadBoth(str string, padStr string, padLen int) string {
|
||||||
|
return buildPadStr(str, padStr, padLen, true, true)
|
||||||
|
}
|
||||||
|
|
||||||
|
// PadString either left, right or both sides, not the padding string can be unicode and more then one
|
||||||
|
// character
|
||||||
|
func buildPadStr(str string, padStr string, padLen int, padLeft bool, padRight bool) string {
|
||||||
|
|
||||||
|
// When padded length is less then the current string size
|
||||||
|
if padLen < utf8.RuneCountInString(str) {
|
||||||
|
return str
|
||||||
|
}
|
||||||
|
|
||||||
|
padLen -= utf8.RuneCountInString(str)
|
||||||
|
|
||||||
|
targetLen := padLen
|
||||||
|
|
||||||
|
targetLenLeft := targetLen
|
||||||
|
targetLenRight := targetLen
|
||||||
|
if padLeft && padRight {
|
||||||
|
targetLenLeft = padLen / 2
|
||||||
|
targetLenRight = padLen - targetLenLeft
|
||||||
|
}
|
||||||
|
|
||||||
|
strToRepeatLen := utf8.RuneCountInString(padStr)
|
||||||
|
|
||||||
|
repeatTimes := int(math.Ceil(float64(targetLen) / float64(strToRepeatLen)))
|
||||||
|
repeatedString := strings.Repeat(padStr, repeatTimes)
|
||||||
|
|
||||||
|
leftSide := ""
|
||||||
|
if padLeft {
|
||||||
|
leftSide = repeatedString[0:targetLenLeft]
|
||||||
|
}
|
||||||
|
|
||||||
|
rightSide := ""
|
||||||
|
if padRight {
|
||||||
|
rightSide = repeatedString[0:targetLenRight]
|
||||||
|
}
|
||||||
|
|
||||||
|
return leftSide + str + rightSide
|
||||||
|
}
|
||||||
|
|
512
vendor/github.com/asaskevich/govalidator/validator.go
generated
vendored
512
vendor/github.com/asaskevich/govalidator/validator.go
generated
vendored
|
@ -2,8 +2,14 @@
|
||||||
package govalidator
|
package govalidator
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"bytes"
|
||||||
|
"crypto/rsa"
|
||||||
|
"crypto/x509"
|
||||||
|
"encoding/base64"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
|
"encoding/pem"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"io/ioutil"
|
||||||
"net"
|
"net"
|
||||||
"net/url"
|
"net/url"
|
||||||
"reflect"
|
"reflect"
|
||||||
|
@ -11,12 +17,21 @@ import (
|
||||||
"sort"
|
"sort"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
|
"time"
|
||||||
"unicode"
|
"unicode"
|
||||||
"unicode/utf8"
|
"unicode/utf8"
|
||||||
)
|
)
|
||||||
|
|
||||||
var fieldsRequiredByDefault bool
|
var (
|
||||||
|
fieldsRequiredByDefault bool
|
||||||
|
notNumberRegexp = regexp.MustCompile("[^0-9]+")
|
||||||
|
whiteSpacesAndMinus = regexp.MustCompile("[\\s-]+")
|
||||||
|
paramsRegexp = regexp.MustCompile("\\(.*\\)$")
|
||||||
|
)
|
||||||
|
|
||||||
|
const maxURLRuneCount = 2083
|
||||||
|
const minURLRuneCount = 3
|
||||||
|
const RF3339WithoutZone = "2006-01-02T15:04:05"
|
||||||
|
|
||||||
// SetFieldsRequiredByDefault causes validation to fail when struct fields
|
// SetFieldsRequiredByDefault causes validation to fail when struct fields
|
||||||
// do not include validations or are not explicitly marked as exempt (using `valid:"-"` or `valid:"email,optional"`).
|
// do not include validations or are not explicitly marked as exempt (using `valid:"-"` or `valid:"email,optional"`).
|
||||||
|
@ -37,17 +52,47 @@ func SetFieldsRequiredByDefault(value bool) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsEmail check if the string is an email.
|
// IsEmail check if the string is an email.
|
||||||
func IsEmail(str string) bool {
|
func IsEmail(email string) bool {
|
||||||
// TODO uppercase letters are not supported
|
if len(email) < 6 || len(email) > 254 {
|
||||||
return rxEmail.MatchString(str)
|
return false
|
||||||
|
}
|
||||||
|
at := strings.LastIndex(email, "@")
|
||||||
|
if at <= 0 || at > len(email)-3 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
user := email[:at]
|
||||||
|
host := email[at+1:]
|
||||||
|
if len(user) > 64 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
if userDotRegexp.MatchString(user) || !userRegexp.MatchString(user) || !hostRegexp.MatchString(host) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
switch host {
|
||||||
|
case "localhost", "example.com":
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
if _, err := net.LookupMX(host); err != nil {
|
||||||
|
if _, err := net.LookupIP(host); err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsURL check if the string is an URL.
|
// IsURL check if the string is an URL.
|
||||||
func IsURL(str string) bool {
|
func IsURL(str string) bool {
|
||||||
if str == "" || len(str) >= 2083 || len(str) <= 3 || strings.HasPrefix(str, ".") {
|
if str == "" || utf8.RuneCountInString(str) >= maxURLRuneCount || len(str) <= minURLRuneCount || strings.HasPrefix(str, ".") {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
u, err := url.Parse(str)
|
strTemp := str
|
||||||
|
if strings.Index(str, ":") >= 0 && strings.Index(str, "://") == -1 {
|
||||||
|
// support no indicated urlscheme but with colon for port number
|
||||||
|
// http:// is appended so url.Parse will succeed, strTemp used so it does not impact rxURL.MatchString
|
||||||
|
strTemp = "http://" + str
|
||||||
|
}
|
||||||
|
u, err := url.Parse(strTemp)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
@ -58,11 +103,10 @@ func IsURL(str string) bool {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
return rxURL.MatchString(str)
|
return rxURL.MatchString(str)
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsRequestURL check if the string rawurl, assuming
|
// IsRequestURL check if the string rawurl, assuming
|
||||||
// it was recieved in an HTTP request, is a valid
|
// it was received in an HTTP request, is a valid
|
||||||
// URL confirm to RFC 3986
|
// URL confirm to RFC 3986
|
||||||
func IsRequestURL(rawurl string) bool {
|
func IsRequestURL(rawurl string) bool {
|
||||||
url, err := url.ParseRequestURI(rawurl)
|
url, err := url.ParseRequestURI(rawurl)
|
||||||
|
@ -76,7 +120,7 @@ func IsRequestURL(rawurl string) bool {
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsRequestURI check if the string rawurl, assuming
|
// IsRequestURI check if the string rawurl, assuming
|
||||||
// it was recieved in an HTTP request, is an
|
// it was received in an HTTP request, is an
|
||||||
// absolute URI or an absolute path.
|
// absolute URI or an absolute path.
|
||||||
func IsRequestURI(rawurl string) bool {
|
func IsRequestURI(rawurl string) bool {
|
||||||
_, err := url.ParseRequestURI(rawurl)
|
_, err := url.ParseRequestURI(rawurl)
|
||||||
|
@ -211,6 +255,22 @@ func IsUpperCase(str string) bool {
|
||||||
return str == strings.ToUpper(str)
|
return str == strings.ToUpper(str)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// HasLowerCase check if the string contains at least 1 lowercase. Empty string is valid.
|
||||||
|
func HasLowerCase(str string) bool {
|
||||||
|
if IsNull(str) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
return rxHasLowerCase.MatchString(str)
|
||||||
|
}
|
||||||
|
|
||||||
|
// HasUpperCase check if the string contians as least 1 uppercase. Empty string is valid.
|
||||||
|
func HasUpperCase(str string) bool {
|
||||||
|
if IsNull(str) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
return rxHasUpperCase.MatchString(str)
|
||||||
|
}
|
||||||
|
|
||||||
// IsInt check if the string is an integer. Empty string is valid.
|
// IsInt check if the string is an integer. Empty string is valid.
|
||||||
func IsInt(str string) bool {
|
func IsInt(str string) bool {
|
||||||
if IsNull(str) {
|
if IsNull(str) {
|
||||||
|
@ -269,9 +329,8 @@ func IsUUID(str string) bool {
|
||||||
|
|
||||||
// IsCreditCard check if the string is a credit card.
|
// IsCreditCard check if the string is a credit card.
|
||||||
func IsCreditCard(str string) bool {
|
func IsCreditCard(str string) bool {
|
||||||
r, _ := regexp.Compile("[^0-9]+")
|
sanitized := notNumberRegexp.ReplaceAllString(str, "")
|
||||||
sanitized := r.ReplaceAll([]byte(str), []byte(""))
|
if !rxCreditCard.MatchString(sanitized) {
|
||||||
if !rxCreditCard.MatchString(string(sanitized)) {
|
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
var sum int64
|
var sum int64
|
||||||
|
@ -279,7 +338,7 @@ func IsCreditCard(str string) bool {
|
||||||
var tmpNum int64
|
var tmpNum int64
|
||||||
var shouldDouble bool
|
var shouldDouble bool
|
||||||
for i := len(sanitized) - 1; i >= 0; i-- {
|
for i := len(sanitized) - 1; i >= 0; i-- {
|
||||||
digit = string(sanitized[i:(i + 1)])
|
digit = sanitized[i:(i + 1)]
|
||||||
tmpNum, _ = ToInt(digit)
|
tmpNum, _ = ToInt(digit)
|
||||||
if shouldDouble {
|
if shouldDouble {
|
||||||
tmpNum *= 2
|
tmpNum *= 2
|
||||||
|
@ -313,12 +372,11 @@ func IsISBN13(str string) bool {
|
||||||
// IsISBN check if the string is an ISBN (version 10 or 13).
|
// IsISBN check if the string is an ISBN (version 10 or 13).
|
||||||
// If version value is not equal to 10 or 13, it will be check both variants.
|
// If version value is not equal to 10 or 13, it will be check both variants.
|
||||||
func IsISBN(str string, version int) bool {
|
func IsISBN(str string, version int) bool {
|
||||||
r, _ := regexp.Compile("[\\s-]+")
|
sanitized := whiteSpacesAndMinus.ReplaceAllString(str, "")
|
||||||
sanitized := r.ReplaceAll([]byte(str), []byte(""))
|
|
||||||
var checksum int32
|
var checksum int32
|
||||||
var i int32
|
var i int32
|
||||||
if version == 10 {
|
if version == 10 {
|
||||||
if !rxISBN10.MatchString(string(sanitized)) {
|
if !rxISBN10.MatchString(sanitized) {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
for i = 0; i < 9; i++ {
|
for i = 0; i < 9; i++ {
|
||||||
|
@ -334,7 +392,7 @@ func IsISBN(str string, version int) bool {
|
||||||
}
|
}
|
||||||
return false
|
return false
|
||||||
} else if version == 13 {
|
} else if version == 13 {
|
||||||
if !rxISBN13.MatchString(string(sanitized)) {
|
if !rxISBN13.MatchString(sanitized) {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
factor := []int32{1, 3}
|
factor := []int32{1, 3}
|
||||||
|
@ -452,13 +510,60 @@ func IsISO3166Alpha3(str string) bool {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// IsISO693Alpha2 checks if a string is valid two-letter language code
|
||||||
|
func IsISO693Alpha2(str string) bool {
|
||||||
|
for _, entry := range ISO693List {
|
||||||
|
if str == entry.Alpha2Code {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsISO693Alpha3b checks if a string is valid three-letter language code
|
||||||
|
func IsISO693Alpha3b(str string) bool {
|
||||||
|
for _, entry := range ISO693List {
|
||||||
|
if str == entry.Alpha3bCode {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
// IsDNSName will validate the given string as a DNS name
|
// IsDNSName will validate the given string as a DNS name
|
||||||
func IsDNSName(str string) bool {
|
func IsDNSName(str string) bool {
|
||||||
if str == "" || len(strings.Replace(str, ".", "", -1)) > 255 {
|
if str == "" || len(strings.Replace(str, ".", "", -1)) > 255 {
|
||||||
// constraints already violated
|
// constraints already violated
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
return rxDNSName.MatchString(str)
|
return !IsIP(str) && rxDNSName.MatchString(str)
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsHash checks if a string is a hash of type algorithm.
|
||||||
|
// Algorithm is one of ['md4', 'md5', 'sha1', 'sha256', 'sha384', 'sha512', 'ripemd128', 'ripemd160', 'tiger128', 'tiger160', 'tiger192', 'crc32', 'crc32b']
|
||||||
|
func IsHash(str string, algorithm string) bool {
|
||||||
|
len := "0"
|
||||||
|
algo := strings.ToLower(algorithm)
|
||||||
|
|
||||||
|
if algo == "crc32" || algo == "crc32b" {
|
||||||
|
len = "8"
|
||||||
|
} else if algo == "md5" || algo == "md4" || algo == "ripemd128" || algo == "tiger128" {
|
||||||
|
len = "32"
|
||||||
|
} else if algo == "sha1" || algo == "ripemd160" || algo == "tiger160" {
|
||||||
|
len = "40"
|
||||||
|
} else if algo == "tiger192" {
|
||||||
|
len = "48"
|
||||||
|
} else if algo == "sha256" {
|
||||||
|
len = "64"
|
||||||
|
} else if algo == "sha384" {
|
||||||
|
len = "96"
|
||||||
|
} else if algo == "sha512" {
|
||||||
|
len = "128"
|
||||||
|
} else {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
return Matches(str, "^[a-f0-9]{"+len+"}$")
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsDialString validates the given string for usage with the various Dial() functions
|
// IsDialString validates the given string for usage with the various Dial() functions
|
||||||
|
@ -496,6 +601,12 @@ func IsIPv6(str string) bool {
|
||||||
return ip != nil && strings.Contains(str, ":")
|
return ip != nil && strings.Contains(str, ":")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// IsCIDR check if the string is an valid CIDR notiation (IPV4 & IPV6)
|
||||||
|
func IsCIDR(str string) bool {
|
||||||
|
_, _, err := net.ParseCIDR(str)
|
||||||
|
return err == nil
|
||||||
|
}
|
||||||
|
|
||||||
// IsMAC check if a string is valid MAC address.
|
// IsMAC check if a string is valid MAC address.
|
||||||
// Possible MAC formats:
|
// Possible MAC formats:
|
||||||
// 01:23:45:67:89:ab
|
// 01:23:45:67:89:ab
|
||||||
|
@ -529,6 +640,60 @@ func IsLongitude(str string) bool {
|
||||||
return rxLongitude.MatchString(str)
|
return rxLongitude.MatchString(str)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// IsRsaPublicKey check if a string is valid public key with provided length
|
||||||
|
func IsRsaPublicKey(str string, keylen int) bool {
|
||||||
|
bb := bytes.NewBufferString(str)
|
||||||
|
pemBytes, err := ioutil.ReadAll(bb)
|
||||||
|
if err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
block, _ := pem.Decode(pemBytes)
|
||||||
|
if block != nil && block.Type != "PUBLIC KEY" {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
var der []byte
|
||||||
|
|
||||||
|
if block != nil {
|
||||||
|
der = block.Bytes
|
||||||
|
} else {
|
||||||
|
der, err = base64.StdEncoding.DecodeString(str)
|
||||||
|
if err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
key, err := x509.ParsePKIXPublicKey(der)
|
||||||
|
if err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
pubkey, ok := key.(*rsa.PublicKey)
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
bitlen := len(pubkey.N.Bytes()) * 8
|
||||||
|
return bitlen == int(keylen)
|
||||||
|
}
|
||||||
|
|
||||||
|
func toJSONName(tag string) string {
|
||||||
|
if tag == "" {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
// JSON name always comes first. If there's no options then split[0] is
|
||||||
|
// JSON name, if JSON name is not set, then split[0] is an empty string.
|
||||||
|
split := strings.SplitN(tag, ",", 2)
|
||||||
|
|
||||||
|
name := split[0]
|
||||||
|
|
||||||
|
// However it is possible that the field is skipped when
|
||||||
|
// (de-)serializing from/to JSON, in which case assume that there is no
|
||||||
|
// tag name to use
|
||||||
|
if name == "-" {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
return name
|
||||||
|
}
|
||||||
|
|
||||||
// ValidateStruct use tags for fields.
|
// ValidateStruct use tags for fields.
|
||||||
// result will be equal to `false` if there are any errors.
|
// result will be equal to `false` if there are any errors.
|
||||||
func ValidateStruct(s interface{}) (bool, error) {
|
func ValidateStruct(s interface{}) (bool, error) {
|
||||||
|
@ -552,11 +717,42 @@ func ValidateStruct(s interface{}) (bool, error) {
|
||||||
if typeField.PkgPath != "" {
|
if typeField.PkgPath != "" {
|
||||||
continue // Private field
|
continue // Private field
|
||||||
}
|
}
|
||||||
resultField, err2 := typeCheck(valueField, typeField, val)
|
structResult := true
|
||||||
|
if (valueField.Kind() == reflect.Struct ||
|
||||||
|
(valueField.Kind() == reflect.Ptr && valueField.Elem().Kind() == reflect.Struct)) &&
|
||||||
|
typeField.Tag.Get(tagName) != "-" {
|
||||||
|
var err error
|
||||||
|
structResult, err = ValidateStruct(valueField.Interface())
|
||||||
|
if err != nil {
|
||||||
|
errs = append(errs, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
resultField, err2 := typeCheck(valueField, typeField, val, nil)
|
||||||
if err2 != nil {
|
if err2 != nil {
|
||||||
|
|
||||||
|
// Replace structure name with JSON name if there is a tag on the variable
|
||||||
|
jsonTag := toJSONName(typeField.Tag.Get("json"))
|
||||||
|
if jsonTag != "" {
|
||||||
|
switch jsonError := err2.(type) {
|
||||||
|
case Error:
|
||||||
|
jsonError.Name = jsonTag
|
||||||
|
err2 = jsonError
|
||||||
|
case Errors:
|
||||||
|
for i2, err3 := range jsonError {
|
||||||
|
switch customErr := err3.(type) {
|
||||||
|
case Error:
|
||||||
|
customErr.Name = jsonTag
|
||||||
|
jsonError[i2] = customErr
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
err2 = jsonError
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
errs = append(errs, err2)
|
errs = append(errs, err2)
|
||||||
}
|
}
|
||||||
result = result && resultField
|
result = result && resultField && structResult
|
||||||
}
|
}
|
||||||
if len(errs) > 0 {
|
if len(errs) > 0 {
|
||||||
err = errs
|
err = errs
|
||||||
|
@ -567,8 +763,11 @@ func ValidateStruct(s interface{}) (bool, error) {
|
||||||
// parseTagIntoMap parses a struct tag `valid:required~Some error message,length(2|3)` into map[string]string{"required": "Some error message", "length(2|3)": ""}
|
// parseTagIntoMap parses a struct tag `valid:required~Some error message,length(2|3)` into map[string]string{"required": "Some error message", "length(2|3)": ""}
|
||||||
func parseTagIntoMap(tag string) tagOptionsMap {
|
func parseTagIntoMap(tag string) tagOptionsMap {
|
||||||
optionsMap := make(tagOptionsMap)
|
optionsMap := make(tagOptionsMap)
|
||||||
options := strings.SplitN(tag, ",", -1)
|
options := strings.Split(tag, ",")
|
||||||
|
|
||||||
for _, option := range options {
|
for _, option := range options {
|
||||||
|
option = strings.TrimSpace(option)
|
||||||
|
|
||||||
validationOptions := strings.Split(option, "~")
|
validationOptions := strings.Split(option, "~")
|
||||||
if !isValidTag(validationOptions[0]) {
|
if !isValidTag(validationOptions[0]) {
|
||||||
continue
|
continue
|
||||||
|
@ -588,7 +787,7 @@ func isValidTag(s string) bool {
|
||||||
}
|
}
|
||||||
for _, c := range s {
|
for _, c := range s {
|
||||||
switch {
|
switch {
|
||||||
case strings.ContainsRune("!#$%&()*+-./:<=>?@[]^_{|}~ ", c):
|
case strings.ContainsRune("\\'\"!#$%&()*+-./:<=>?@[]^_{|}~ ", c):
|
||||||
// Backslash and quote chars are reserved, but
|
// Backslash and quote chars are reserved, but
|
||||||
// otherwise any punctuation chars are allowed
|
// otherwise any punctuation chars are allowed
|
||||||
// in a tag name.
|
// in a tag name.
|
||||||
|
@ -614,6 +813,33 @@ func IsSemver(str string) bool {
|
||||||
return rxSemver.MatchString(str)
|
return rxSemver.MatchString(str)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// IsTime check if string is valid according to given format
|
||||||
|
func IsTime(str string, format string) bool {
|
||||||
|
_, err := time.Parse(format, str)
|
||||||
|
return err == nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsRFC3339 check if string is valid timestamp value according to RFC3339
|
||||||
|
func IsRFC3339(str string) bool {
|
||||||
|
return IsTime(str, time.RFC3339)
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsRFC3339WithoutZone check if string is valid timestamp value according to RFC3339 which excludes the timezone.
|
||||||
|
func IsRFC3339WithoutZone(str string) bool {
|
||||||
|
return IsTime(str, RF3339WithoutZone)
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsISO4217 check if string is valid ISO currency code
|
||||||
|
func IsISO4217(str string) bool {
|
||||||
|
for _, currency := range ISO4217List {
|
||||||
|
if str == currency {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
// ByteLength check string's length
|
// ByteLength check string's length
|
||||||
func ByteLength(str string, params ...string) bool {
|
func ByteLength(str string, params ...string) bool {
|
||||||
if len(params) == 2 {
|
if len(params) == 2 {
|
||||||
|
@ -625,6 +851,23 @@ func ByteLength(str string, params ...string) bool {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// RuneLength check string's length
|
||||||
|
// Alias for StringLength
|
||||||
|
func RuneLength(str string, params ...string) bool {
|
||||||
|
return StringLength(str, params...)
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsRsaPub check whether string is valid RSA key
|
||||||
|
// Alias for IsRsaPublicKey
|
||||||
|
func IsRsaPub(str string, params ...string) bool {
|
||||||
|
if len(params) == 1 {
|
||||||
|
len, _ := ToInt(params[0])
|
||||||
|
return IsRsaPublicKey(str, int(len))
|
||||||
|
}
|
||||||
|
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
// StringMatches checks if a string matches a given pattern.
|
// StringMatches checks if a string matches a given pattern.
|
||||||
func StringMatches(s string, params ...string) bool {
|
func StringMatches(s string, params ...string) bool {
|
||||||
if len(params) == 1 {
|
if len(params) == 1 {
|
||||||
|
@ -647,20 +890,55 @@ func StringLength(str string, params ...string) bool {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Range check string's length
|
||||||
|
func Range(str string, params ...string) bool {
|
||||||
|
if len(params) == 2 {
|
||||||
|
value, _ := ToFloat(str)
|
||||||
|
min, _ := ToFloat(params[0])
|
||||||
|
max, _ := ToFloat(params[1])
|
||||||
|
return InRange(value, min, max)
|
||||||
|
}
|
||||||
|
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
func isInRaw(str string, params ...string) bool {
|
||||||
|
if len(params) == 1 {
|
||||||
|
rawParams := params[0]
|
||||||
|
|
||||||
|
parsedParams := strings.Split(rawParams, "|")
|
||||||
|
|
||||||
|
return IsIn(str, parsedParams...)
|
||||||
|
}
|
||||||
|
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsIn check if string str is a member of the set of strings params
|
||||||
|
func IsIn(str string, params ...string) bool {
|
||||||
|
for _, param := range params {
|
||||||
|
if str == param {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
func checkRequired(v reflect.Value, t reflect.StructField, options tagOptionsMap) (bool, error) {
|
func checkRequired(v reflect.Value, t reflect.StructField, options tagOptionsMap) (bool, error) {
|
||||||
if requiredOption, isRequired := options["required"]; isRequired {
|
if requiredOption, isRequired := options["required"]; isRequired {
|
||||||
if len(requiredOption) > 0 {
|
if len(requiredOption) > 0 {
|
||||||
return false, Error{t.Name, fmt.Errorf(requiredOption), true}
|
return false, Error{t.Name, fmt.Errorf(requiredOption), true, "required"}
|
||||||
}
|
}
|
||||||
return false, Error{t.Name, fmt.Errorf("non zero value required"), false}
|
return false, Error{t.Name, fmt.Errorf("non zero value required"), false, "required"}
|
||||||
} else if _, isOptional := options["optional"]; fieldsRequiredByDefault && !isOptional {
|
} else if _, isOptional := options["optional"]; fieldsRequiredByDefault && !isOptional {
|
||||||
return false, Error{t.Name, fmt.Errorf("All fields are required to at least have one validation defined"), false}
|
return false, Error{t.Name, fmt.Errorf("Missing required field"), false, "required"}
|
||||||
}
|
}
|
||||||
// not required and empty is valid
|
// not required and empty is valid
|
||||||
return true, nil
|
return true, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func typeCheck(v reflect.Value, t reflect.StructField, o reflect.Value) (bool, error) {
|
func typeCheck(v reflect.Value, t reflect.StructField, o reflect.Value, options tagOptionsMap) (isValid bool, resultErr error) {
|
||||||
if !v.IsValid() {
|
if !v.IsValid() {
|
||||||
return false, nil
|
return false, nil
|
||||||
}
|
}
|
||||||
|
@ -673,31 +951,15 @@ func typeCheck(v reflect.Value, t reflect.StructField, o reflect.Value) (bool, e
|
||||||
if !fieldsRequiredByDefault {
|
if !fieldsRequiredByDefault {
|
||||||
return true, nil
|
return true, nil
|
||||||
}
|
}
|
||||||
return false, Error{t.Name, fmt.Errorf("All fields are required to at least have one validation defined"), false}
|
return false, Error{t.Name, fmt.Errorf("All fields are required to at least have one validation defined"), false, "required"}
|
||||||
case "-":
|
case "-":
|
||||||
return true, nil
|
return true, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
options := parseTagIntoMap(tag)
|
isRootType := false
|
||||||
var customTypeErrors Errors
|
if options == nil {
|
||||||
var customTypeValidatorsExist bool
|
isRootType = true
|
||||||
for validatorName, customErrorMessage := range options {
|
options = parseTagIntoMap(tag)
|
||||||
if validatefunc, ok := CustomTypeTagMap.Get(validatorName); ok {
|
|
||||||
customTypeValidatorsExist = true
|
|
||||||
if result := validatefunc(v.Interface(), o.Interface()); !result {
|
|
||||||
if len(customErrorMessage) > 0 {
|
|
||||||
customTypeErrors = append(customTypeErrors, Error{Name: t.Name, Err: fmt.Errorf(customErrorMessage), CustomErrorMessageExists: true})
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
customTypeErrors = append(customTypeErrors, Error{Name: t.Name, Err: fmt.Errorf("%s does not validate as %s", fmt.Sprint(v), validatorName), CustomErrorMessageExists: false})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if customTypeValidatorsExist {
|
|
||||||
if len(customTypeErrors.Errors()) > 0 {
|
|
||||||
return false, customTypeErrors
|
|
||||||
}
|
|
||||||
return true, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if isEmptyValue(v) {
|
if isEmptyValue(v) {
|
||||||
|
@ -705,6 +967,42 @@ func typeCheck(v reflect.Value, t reflect.StructField, o reflect.Value) (bool, e
|
||||||
return checkRequired(v, t, options)
|
return checkRequired(v, t, options)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var customTypeErrors Errors
|
||||||
|
for validatorName, customErrorMessage := range options {
|
||||||
|
if validatefunc, ok := CustomTypeTagMap.Get(validatorName); ok {
|
||||||
|
delete(options, validatorName)
|
||||||
|
|
||||||
|
if result := validatefunc(v.Interface(), o.Interface()); !result {
|
||||||
|
if len(customErrorMessage) > 0 {
|
||||||
|
customTypeErrors = append(customTypeErrors, Error{Name: t.Name, Err: fmt.Errorf(customErrorMessage), CustomErrorMessageExists: true, Validator: stripParams(validatorName)})
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
customTypeErrors = append(customTypeErrors, Error{Name: t.Name, Err: fmt.Errorf("%s does not validate as %s", fmt.Sprint(v), validatorName), CustomErrorMessageExists: false, Validator: stripParams(validatorName)})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(customTypeErrors.Errors()) > 0 {
|
||||||
|
return false, customTypeErrors
|
||||||
|
}
|
||||||
|
|
||||||
|
if isRootType {
|
||||||
|
// Ensure that we've checked the value by all specified validators before report that the value is valid
|
||||||
|
defer func() {
|
||||||
|
delete(options, "optional")
|
||||||
|
delete(options, "required")
|
||||||
|
|
||||||
|
if isValid && resultErr == nil && len(options) != 0 {
|
||||||
|
for validator := range options {
|
||||||
|
isValid = false
|
||||||
|
resultErr = Error{t.Name, fmt.Errorf(
|
||||||
|
"The following validator is invalid or can't be applied to the field: %q", validator), false, stripParams(validator)}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
switch v.Kind() {
|
switch v.Kind() {
|
||||||
case reflect.Bool,
|
case reflect.Bool,
|
||||||
reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64,
|
reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64,
|
||||||
|
@ -712,75 +1010,72 @@ func typeCheck(v reflect.Value, t reflect.StructField, o reflect.Value) (bool, e
|
||||||
reflect.Float32, reflect.Float64,
|
reflect.Float32, reflect.Float64,
|
||||||
reflect.String:
|
reflect.String:
|
||||||
// for each tag option check the map of validator functions
|
// for each tag option check the map of validator functions
|
||||||
for validator, customErrorMessage := range options {
|
for validatorSpec, customErrorMessage := range options {
|
||||||
var negate bool
|
var negate bool
|
||||||
customMsgExists := (len(customErrorMessage) > 0)
|
validator := validatorSpec
|
||||||
// Check wether the tag looks like '!something' or 'something'
|
customMsgExists := len(customErrorMessage) > 0
|
||||||
|
|
||||||
|
// Check whether the tag looks like '!something' or 'something'
|
||||||
if validator[0] == '!' {
|
if validator[0] == '!' {
|
||||||
validator = string(validator[1:])
|
validator = validator[1:]
|
||||||
negate = true
|
negate = true
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check for param validators
|
// Check for param validators
|
||||||
for key, value := range ParamTagRegexMap {
|
for key, value := range ParamTagRegexMap {
|
||||||
ps := value.FindStringSubmatch(validator)
|
ps := value.FindStringSubmatch(validator)
|
||||||
if len(ps) > 0 {
|
if len(ps) == 0 {
|
||||||
if validatefunc, ok := ParamTagMap[key]; ok {
|
continue
|
||||||
switch v.Kind() {
|
}
|
||||||
case reflect.String:
|
|
||||||
field := fmt.Sprint(v) // make value into string, then validate with regex
|
|
||||||
if result := validatefunc(field, ps[1:]...); (!result && !negate) || (result && negate) {
|
|
||||||
var err error
|
|
||||||
if !negate {
|
|
||||||
if customMsgExists {
|
|
||||||
err = fmt.Errorf(customErrorMessage)
|
|
||||||
} else {
|
|
||||||
err = fmt.Errorf("%s does not validate as %s", field, validator)
|
|
||||||
}
|
|
||||||
|
|
||||||
} else {
|
validatefunc, ok := ParamTagMap[key]
|
||||||
if customMsgExists {
|
if !ok {
|
||||||
err = fmt.Errorf(customErrorMessage)
|
continue
|
||||||
} else {
|
}
|
||||||
err = fmt.Errorf("%s does validate as %s", field, validator)
|
|
||||||
}
|
delete(options, validatorSpec)
|
||||||
}
|
|
||||||
return false, Error{t.Name, err, customMsgExists}
|
switch v.Kind() {
|
||||||
}
|
case reflect.String,
|
||||||
default:
|
reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64,
|
||||||
// type not yet supported, fail
|
reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64,
|
||||||
return false, Error{t.Name, fmt.Errorf("Validator %s doesn't support kind %s", validator, v.Kind()), false}
|
reflect.Float32, reflect.Float64:
|
||||||
|
|
||||||
|
field := fmt.Sprint(v) // make value into string, then validate with regex
|
||||||
|
if result := validatefunc(field, ps[1:]...); (!result && !negate) || (result && negate) {
|
||||||
|
if customMsgExists {
|
||||||
|
return false, Error{t.Name, fmt.Errorf(customErrorMessage), customMsgExists, stripParams(validatorSpec)}
|
||||||
}
|
}
|
||||||
|
if negate {
|
||||||
|
return false, Error{t.Name, fmt.Errorf("%s does validate as %s", field, validator), customMsgExists, stripParams(validatorSpec)}
|
||||||
|
}
|
||||||
|
return false, Error{t.Name, fmt.Errorf("%s does not validate as %s", field, validator), customMsgExists, stripParams(validatorSpec)}
|
||||||
}
|
}
|
||||||
|
default:
|
||||||
|
// type not yet supported, fail
|
||||||
|
return false, Error{t.Name, fmt.Errorf("Validator %s doesn't support kind %s", validator, v.Kind()), false, stripParams(validatorSpec)}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if validatefunc, ok := TagMap[validator]; ok {
|
if validatefunc, ok := TagMap[validator]; ok {
|
||||||
|
delete(options, validatorSpec)
|
||||||
|
|
||||||
switch v.Kind() {
|
switch v.Kind() {
|
||||||
case reflect.String:
|
case reflect.String:
|
||||||
field := fmt.Sprint(v) // make value into string, then validate with regex
|
field := fmt.Sprint(v) // make value into string, then validate with regex
|
||||||
if result := validatefunc(field); !result && !negate || result && negate {
|
if result := validatefunc(field); !result && !negate || result && negate {
|
||||||
var err error
|
if customMsgExists {
|
||||||
|
return false, Error{t.Name, fmt.Errorf(customErrorMessage), customMsgExists, stripParams(validatorSpec)}
|
||||||
if !negate {
|
|
||||||
if customMsgExists {
|
|
||||||
err = fmt.Errorf(customErrorMessage)
|
|
||||||
} else {
|
|
||||||
err = fmt.Errorf("%s does not validate as %s", field, validator)
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
if customMsgExists {
|
|
||||||
err = fmt.Errorf(customErrorMessage)
|
|
||||||
} else {
|
|
||||||
err = fmt.Errorf("%s does validate as %s", field, validator)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
return false, Error{t.Name, err, customMsgExists}
|
if negate {
|
||||||
|
return false, Error{t.Name, fmt.Errorf("%s does validate as %s", field, validator), customMsgExists, stripParams(validatorSpec)}
|
||||||
|
}
|
||||||
|
return false, Error{t.Name, fmt.Errorf("%s does not validate as %s", field, validator), customMsgExists, stripParams(validatorSpec)}
|
||||||
}
|
}
|
||||||
default:
|
default:
|
||||||
//Not Yet Supported Types (Fail here!)
|
//Not Yet Supported Types (Fail here!)
|
||||||
err := fmt.Errorf("Validator %s doesn't support kind %s for value %v", validator, v.Kind(), v)
|
err := fmt.Errorf("Validator %s doesn't support kind %s for value %v", validator, v.Kind(), v)
|
||||||
return false, Error{t.Name, err, false}
|
return false, Error{t.Name, err, false, stripParams(validatorSpec)}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -794,25 +1089,15 @@ func typeCheck(v reflect.Value, t reflect.StructField, o reflect.Value) (bool, e
|
||||||
sort.Sort(sv)
|
sort.Sort(sv)
|
||||||
result := true
|
result := true
|
||||||
for _, k := range sv {
|
for _, k := range sv {
|
||||||
resultItem, err := ValidateStruct(v.MapIndex(k).Interface())
|
|
||||||
if err != nil {
|
|
||||||
return false, err
|
|
||||||
}
|
|
||||||
result = result && resultItem
|
|
||||||
}
|
|
||||||
return result, nil
|
|
||||||
case reflect.Slice:
|
|
||||||
result := true
|
|
||||||
for i := 0; i < v.Len(); i++ {
|
|
||||||
var resultItem bool
|
var resultItem bool
|
||||||
var err error
|
var err error
|
||||||
if v.Index(i).Kind() != reflect.Struct {
|
if v.MapIndex(k).Kind() != reflect.Struct {
|
||||||
resultItem, err = typeCheck(v.Index(i), t, o)
|
resultItem, err = typeCheck(v.MapIndex(k), t, o, options)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return false, err
|
return false, err
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
resultItem, err = ValidateStruct(v.Index(i).Interface())
|
resultItem, err = ValidateStruct(v.MapIndex(k).Interface())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return false, err
|
return false, err
|
||||||
}
|
}
|
||||||
|
@ -820,13 +1105,13 @@ func typeCheck(v reflect.Value, t reflect.StructField, o reflect.Value) (bool, e
|
||||||
result = result && resultItem
|
result = result && resultItem
|
||||||
}
|
}
|
||||||
return result, nil
|
return result, nil
|
||||||
case reflect.Array:
|
case reflect.Slice, reflect.Array:
|
||||||
result := true
|
result := true
|
||||||
for i := 0; i < v.Len(); i++ {
|
for i := 0; i < v.Len(); i++ {
|
||||||
var resultItem bool
|
var resultItem bool
|
||||||
var err error
|
var err error
|
||||||
if v.Index(i).Kind() != reflect.Struct {
|
if v.Index(i).Kind() != reflect.Struct {
|
||||||
resultItem, err = typeCheck(v.Index(i), t, o)
|
resultItem, err = typeCheck(v.Index(i), t, o, options)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return false, err
|
return false, err
|
||||||
}
|
}
|
||||||
|
@ -850,7 +1135,7 @@ func typeCheck(v reflect.Value, t reflect.StructField, o reflect.Value) (bool, e
|
||||||
if v.IsNil() {
|
if v.IsNil() {
|
||||||
return true, nil
|
return true, nil
|
||||||
}
|
}
|
||||||
return typeCheck(v.Elem(), t, o)
|
return typeCheck(v.Elem(), t, o, options)
|
||||||
case reflect.Struct:
|
case reflect.Struct:
|
||||||
return ValidateStruct(v.Interface())
|
return ValidateStruct(v.Interface())
|
||||||
default:
|
default:
|
||||||
|
@ -858,6 +1143,10 @@ func typeCheck(v reflect.Value, t reflect.StructField, o reflect.Value) (bool, e
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func stripParams(validatorString string) string {
|
||||||
|
return paramsRegexp.ReplaceAllString(validatorString, "")
|
||||||
|
}
|
||||||
|
|
||||||
func isEmptyValue(v reflect.Value) bool {
|
func isEmptyValue(v reflect.Value) bool {
|
||||||
switch v.Kind() {
|
switch v.Kind() {
|
||||||
case reflect.String, reflect.Array:
|
case reflect.String, reflect.Array:
|
||||||
|
@ -903,7 +1192,10 @@ func ErrorsByField(e error) map[string]string {
|
||||||
m[e.(Error).Name] = e.(Error).Err.Error()
|
m[e.(Error).Name] = e.(Error).Err.Error()
|
||||||
case Errors:
|
case Errors:
|
||||||
for _, item := range e.(Errors).Errors() {
|
for _, item := range e.(Errors).Errors() {
|
||||||
m[item.(Error).Name] = item.(Error).Err.Error()
|
n := ErrorsByField(item)
|
||||||
|
for k, v := range n {
|
||||||
|
m[k] = v
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
15
vendor/github.com/asaskevich/govalidator/wercker.yml
generated
vendored
15
vendor/github.com/asaskevich/govalidator/wercker.yml
generated
vendored
|
@ -1,15 +0,0 @@
|
||||||
box: wercker/golang
|
|
||||||
build:
|
|
||||||
steps:
|
|
||||||
- setup-go-workspace
|
|
||||||
|
|
||||||
- script:
|
|
||||||
name: go get
|
|
||||||
code: |
|
|
||||||
go version
|
|
||||||
go get -t ./...
|
|
||||||
|
|
||||||
- script:
|
|
||||||
name: go test
|
|
||||||
code: |
|
|
||||||
go test -race ./...
|
|
21
vendor/github.com/hpcloud/tail/LICENSE.txt
generated
vendored
Normal file
21
vendor/github.com/hpcloud/tail/LICENSE.txt
generated
vendored
Normal file
|
@ -0,0 +1,21 @@
|
||||||
|
# The MIT License (MIT)
|
||||||
|
|
||||||
|
# © Copyright 2015 Hewlett Packard Enterprise Development LP
|
||||||
|
Copyright (c) 2014 ActiveState
|
||||||
|
|
||||||
|
Permission is hereby granted, free of charge, to any person obtaining a copy
|
||||||
|
of this software and associated documentation files (the "Software"), to deal
|
||||||
|
in the Software without restriction, including without limitation the rights
|
||||||
|
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
||||||
|
copies of the Software, and to permit persons to whom the Software is
|
||||||
|
furnished to do so, subject to the following conditions:
|
||||||
|
The above copyright notice and this permission notice shall be included in all
|
||||||
|
copies or substantial portions of the Software.
|
||||||
|
|
||||||
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||||
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||||
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
||||||
|
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||||
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
||||||
|
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
||||||
|
SOFTWARE.
|
7
vendor/github.com/hpcloud/tail/ratelimiter/Licence
generated
vendored
Normal file
7
vendor/github.com/hpcloud/tail/ratelimiter/Licence
generated
vendored
Normal file
|
@ -0,0 +1,7 @@
|
||||||
|
Copyright (C) 2013 99designs
|
||||||
|
|
||||||
|
Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:
|
||||||
|
|
||||||
|
The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.
|
||||||
|
|
||||||
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
97
vendor/github.com/hpcloud/tail/ratelimiter/leakybucket.go
generated
vendored
Normal file
97
vendor/github.com/hpcloud/tail/ratelimiter/leakybucket.go
generated
vendored
Normal file
|
@ -0,0 +1,97 @@
|
||||||
|
// Package ratelimiter implements the Leaky Bucket ratelimiting algorithm with memcached and in-memory backends.
|
||||||
|
package ratelimiter
|
||||||
|
|
||||||
|
import (
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
type LeakyBucket struct {
|
||||||
|
Size uint16
|
||||||
|
Fill float64
|
||||||
|
LeakInterval time.Duration // time.Duration for 1 unit of size to leak
|
||||||
|
Lastupdate time.Time
|
||||||
|
Now func() time.Time
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewLeakyBucket(size uint16, leakInterval time.Duration) *LeakyBucket {
|
||||||
|
bucket := LeakyBucket{
|
||||||
|
Size: size,
|
||||||
|
Fill: 0,
|
||||||
|
LeakInterval: leakInterval,
|
||||||
|
Now: time.Now,
|
||||||
|
Lastupdate: time.Now(),
|
||||||
|
}
|
||||||
|
|
||||||
|
return &bucket
|
||||||
|
}
|
||||||
|
|
||||||
|
func (b *LeakyBucket) updateFill() {
|
||||||
|
now := b.Now()
|
||||||
|
if b.Fill > 0 {
|
||||||
|
elapsed := now.Sub(b.Lastupdate)
|
||||||
|
|
||||||
|
b.Fill -= float64(elapsed) / float64(b.LeakInterval)
|
||||||
|
if b.Fill < 0 {
|
||||||
|
b.Fill = 0
|
||||||
|
}
|
||||||
|
}
|
||||||
|
b.Lastupdate = now
|
||||||
|
}
|
||||||
|
|
||||||
|
func (b *LeakyBucket) Pour(amount uint16) bool {
|
||||||
|
b.updateFill()
|
||||||
|
|
||||||
|
var newfill float64 = b.Fill + float64(amount)
|
||||||
|
|
||||||
|
if newfill > float64(b.Size) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
b.Fill = newfill
|
||||||
|
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// The time at which this bucket will be completely drained
|
||||||
|
func (b *LeakyBucket) DrainedAt() time.Time {
|
||||||
|
return b.Lastupdate.Add(time.Duration(b.Fill * float64(b.LeakInterval)))
|
||||||
|
}
|
||||||
|
|
||||||
|
// The duration until this bucket is completely drained
|
||||||
|
func (b *LeakyBucket) TimeToDrain() time.Duration {
|
||||||
|
return b.DrainedAt().Sub(b.Now())
|
||||||
|
}
|
||||||
|
|
||||||
|
func (b *LeakyBucket) TimeSinceLastUpdate() time.Duration {
|
||||||
|
return b.Now().Sub(b.Lastupdate)
|
||||||
|
}
|
||||||
|
|
||||||
|
type LeakyBucketSer struct {
|
||||||
|
Size uint16
|
||||||
|
Fill float64
|
||||||
|
LeakInterval time.Duration // time.Duration for 1 unit of size to leak
|
||||||
|
Lastupdate time.Time
|
||||||
|
}
|
||||||
|
|
||||||
|
func (b *LeakyBucket) Serialise() *LeakyBucketSer {
|
||||||
|
bucket := LeakyBucketSer{
|
||||||
|
Size: b.Size,
|
||||||
|
Fill: b.Fill,
|
||||||
|
LeakInterval: b.LeakInterval,
|
||||||
|
Lastupdate: b.Lastupdate,
|
||||||
|
}
|
||||||
|
|
||||||
|
return &bucket
|
||||||
|
}
|
||||||
|
|
||||||
|
func (b *LeakyBucketSer) DeSerialise() *LeakyBucket {
|
||||||
|
bucket := LeakyBucket{
|
||||||
|
Size: b.Size,
|
||||||
|
Fill: b.Fill,
|
||||||
|
LeakInterval: b.LeakInterval,
|
||||||
|
Lastupdate: b.Lastupdate,
|
||||||
|
Now: time.Now,
|
||||||
|
}
|
||||||
|
|
||||||
|
return &bucket
|
||||||
|
}
|
60
vendor/github.com/hpcloud/tail/ratelimiter/memory.go
generated
vendored
Normal file
60
vendor/github.com/hpcloud/tail/ratelimiter/memory.go
generated
vendored
Normal file
|
@ -0,0 +1,60 @@
|
||||||
|
package ratelimiter
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
GC_SIZE int = 100
|
||||||
|
GC_PERIOD time.Duration = 60 * time.Second
|
||||||
|
)
|
||||||
|
|
||||||
|
type Memory struct {
|
||||||
|
store map[string]LeakyBucket
|
||||||
|
lastGCCollected time.Time
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewMemory() *Memory {
|
||||||
|
m := new(Memory)
|
||||||
|
m.store = make(map[string]LeakyBucket)
|
||||||
|
m.lastGCCollected = time.Now()
|
||||||
|
return m
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *Memory) GetBucketFor(key string) (*LeakyBucket, error) {
|
||||||
|
|
||||||
|
bucket, ok := m.store[key]
|
||||||
|
if !ok {
|
||||||
|
return nil, errors.New("miss")
|
||||||
|
}
|
||||||
|
|
||||||
|
return &bucket, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *Memory) SetBucketFor(key string, bucket LeakyBucket) error {
|
||||||
|
|
||||||
|
if len(m.store) > GC_SIZE {
|
||||||
|
m.GarbageCollect()
|
||||||
|
}
|
||||||
|
|
||||||
|
m.store[key] = bucket
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *Memory) GarbageCollect() {
|
||||||
|
now := time.Now()
|
||||||
|
|
||||||
|
// rate limit GC to once per minute
|
||||||
|
if now.Unix() >= m.lastGCCollected.Add(GC_PERIOD).Unix() {
|
||||||
|
for key, bucket := range m.store {
|
||||||
|
// if the bucket is drained, then GC
|
||||||
|
if bucket.DrainedAt().Unix() < now.Unix() {
|
||||||
|
delete(m.store, key)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
m.lastGCCollected = now
|
||||||
|
}
|
||||||
|
}
|
6
vendor/github.com/hpcloud/tail/ratelimiter/storage.go
generated
vendored
Normal file
6
vendor/github.com/hpcloud/tail/ratelimiter/storage.go
generated
vendored
Normal file
|
@ -0,0 +1,6 @@
|
||||||
|
package ratelimiter
|
||||||
|
|
||||||
|
type Storage interface {
|
||||||
|
GetBucketFor(string) (*LeakyBucket, error)
|
||||||
|
SetBucketFor(string, LeakyBucket) error
|
||||||
|
}
|
437
vendor/github.com/hpcloud/tail/tail.go
generated
vendored
Normal file
437
vendor/github.com/hpcloud/tail/tail.go
generated
vendored
Normal file
|
@ -0,0 +1,437 @@
|
||||||
|
// Copyright (c) 2015 HPE Software Inc. All rights reserved.
|
||||||
|
// Copyright (c) 2013 ActiveState Software Inc. All rights reserved.
|
||||||
|
|
||||||
|
package tail
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bufio"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"io/ioutil"
|
||||||
|
"log"
|
||||||
|
"os"
|
||||||
|
"strings"
|
||||||
|
"sync"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/hpcloud/tail/ratelimiter"
|
||||||
|
"github.com/hpcloud/tail/util"
|
||||||
|
"github.com/hpcloud/tail/watch"
|
||||||
|
"gopkg.in/tomb.v1"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
ErrStop = errors.New("tail should now stop")
|
||||||
|
)
|
||||||
|
|
||||||
|
type Line struct {
|
||||||
|
Text string
|
||||||
|
Time time.Time
|
||||||
|
Err error // Error from tail
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewLine returns a Line with present time.
|
||||||
|
func NewLine(text string) *Line {
|
||||||
|
return &Line{text, time.Now(), nil}
|
||||||
|
}
|
||||||
|
|
||||||
|
// SeekInfo represents arguments to `os.Seek`
|
||||||
|
type SeekInfo struct {
|
||||||
|
Offset int64
|
||||||
|
Whence int // os.SEEK_*
|
||||||
|
}
|
||||||
|
|
||||||
|
type logger interface {
|
||||||
|
Fatal(v ...interface{})
|
||||||
|
Fatalf(format string, v ...interface{})
|
||||||
|
Fatalln(v ...interface{})
|
||||||
|
Panic(v ...interface{})
|
||||||
|
Panicf(format string, v ...interface{})
|
||||||
|
Panicln(v ...interface{})
|
||||||
|
Print(v ...interface{})
|
||||||
|
Printf(format string, v ...interface{})
|
||||||
|
Println(v ...interface{})
|
||||||
|
}
|
||||||
|
|
||||||
|
// Config is used to specify how a file must be tailed.
|
||||||
|
type Config struct {
|
||||||
|
// File-specifc
|
||||||
|
Location *SeekInfo // Seek to this location before tailing
|
||||||
|
ReOpen bool // Reopen recreated files (tail -F)
|
||||||
|
MustExist bool // Fail early if the file does not exist
|
||||||
|
Poll bool // Poll for file changes instead of using inotify
|
||||||
|
Pipe bool // Is a named pipe (mkfifo)
|
||||||
|
RateLimiter *ratelimiter.LeakyBucket
|
||||||
|
|
||||||
|
// Generic IO
|
||||||
|
Follow bool // Continue looking for new lines (tail -f)
|
||||||
|
MaxLineSize int // If non-zero, split longer lines into multiple lines
|
||||||
|
|
||||||
|
// Logger, when nil, is set to tail.DefaultLogger
|
||||||
|
// To disable logging: set field to tail.DiscardingLogger
|
||||||
|
Logger logger
|
||||||
|
}
|
||||||
|
|
||||||
|
type Tail struct {
|
||||||
|
Filename string
|
||||||
|
Lines chan *Line
|
||||||
|
Config
|
||||||
|
|
||||||
|
file *os.File
|
||||||
|
reader *bufio.Reader
|
||||||
|
|
||||||
|
watcher watch.FileWatcher
|
||||||
|
changes *watch.FileChanges
|
||||||
|
|
||||||
|
tomb.Tomb // provides: Done, Kill, Dying
|
||||||
|
|
||||||
|
lk sync.Mutex
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
// DefaultLogger is used when Config.Logger == nil
|
||||||
|
DefaultLogger = log.New(os.Stderr, "", log.LstdFlags)
|
||||||
|
// DiscardingLogger can be used to disable logging output
|
||||||
|
DiscardingLogger = log.New(ioutil.Discard, "", 0)
|
||||||
|
)
|
||||||
|
|
||||||
|
// TailFile begins tailing the file. Output stream is made available
|
||||||
|
// via the `Tail.Lines` channel. To handle errors during tailing,
|
||||||
|
// invoke the `Wait` or `Err` method after finishing reading from the
|
||||||
|
// `Lines` channel.
|
||||||
|
func TailFile(filename string, config Config) (*Tail, error) {
|
||||||
|
if config.ReOpen && !config.Follow {
|
||||||
|
util.Fatal("cannot set ReOpen without Follow.")
|
||||||
|
}
|
||||||
|
|
||||||
|
t := &Tail{
|
||||||
|
Filename: filename,
|
||||||
|
Lines: make(chan *Line),
|
||||||
|
Config: config,
|
||||||
|
}
|
||||||
|
|
||||||
|
// when Logger was not specified in config, use default logger
|
||||||
|
if t.Logger == nil {
|
||||||
|
t.Logger = log.New(os.Stderr, "", log.LstdFlags)
|
||||||
|
}
|
||||||
|
|
||||||
|
if t.Poll {
|
||||||
|
t.watcher = watch.NewPollingFileWatcher(filename)
|
||||||
|
} else {
|
||||||
|
t.watcher = watch.NewInotifyFileWatcher(filename)
|
||||||
|
}
|
||||||
|
|
||||||
|
if t.MustExist {
|
||||||
|
var err error
|
||||||
|
t.file, err = OpenFile(t.Filename)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
go t.tailFileSync()
|
||||||
|
|
||||||
|
return t, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Return the file's current position, like stdio's ftell().
|
||||||
|
// But this value is not very accurate.
|
||||||
|
// it may readed one line in the chan(tail.Lines),
|
||||||
|
// so it may lost one line.
|
||||||
|
func (tail *Tail) Tell() (offset int64, err error) {
|
||||||
|
if tail.file == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
offset, err = tail.file.Seek(0, os.SEEK_CUR)
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
tail.lk.Lock()
|
||||||
|
defer tail.lk.Unlock()
|
||||||
|
if tail.reader == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
offset -= int64(tail.reader.Buffered())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Stop stops the tailing activity.
|
||||||
|
func (tail *Tail) Stop() error {
|
||||||
|
tail.Kill(nil)
|
||||||
|
return tail.Wait()
|
||||||
|
}
|
||||||
|
|
||||||
|
// StopAtEOF stops tailing as soon as the end of the file is reached.
|
||||||
|
func (tail *Tail) StopAtEOF() error {
|
||||||
|
tail.Kill(errStopAtEOF)
|
||||||
|
return tail.Wait()
|
||||||
|
}
|
||||||
|
|
||||||
|
var errStopAtEOF = errors.New("tail: stop at eof")
|
||||||
|
|
||||||
|
func (tail *Tail) close() {
|
||||||
|
close(tail.Lines)
|
||||||
|
tail.closeFile()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (tail *Tail) closeFile() {
|
||||||
|
if tail.file != nil {
|
||||||
|
tail.file.Close()
|
||||||
|
tail.file = nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (tail *Tail) reopen() error {
|
||||||
|
tail.closeFile()
|
||||||
|
for {
|
||||||
|
var err error
|
||||||
|
tail.file, err = OpenFile(tail.Filename)
|
||||||
|
if err != nil {
|
||||||
|
if os.IsNotExist(err) {
|
||||||
|
tail.Logger.Printf("Waiting for %s to appear...", tail.Filename)
|
||||||
|
if err := tail.watcher.BlockUntilExists(&tail.Tomb); err != nil {
|
||||||
|
if err == tomb.ErrDying {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
return fmt.Errorf("Failed to detect creation of %s: %s", tail.Filename, err)
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
return fmt.Errorf("Unable to open file %s: %s", tail.Filename, err)
|
||||||
|
}
|
||||||
|
break
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (tail *Tail) readLine() (string, error) {
|
||||||
|
tail.lk.Lock()
|
||||||
|
line, err := tail.reader.ReadString('\n')
|
||||||
|
tail.lk.Unlock()
|
||||||
|
if err != nil {
|
||||||
|
// Note ReadString "returns the data read before the error" in
|
||||||
|
// case of an error, including EOF, so we return it as is. The
|
||||||
|
// caller is expected to process it if err is EOF.
|
||||||
|
return line, err
|
||||||
|
}
|
||||||
|
|
||||||
|
line = strings.TrimRight(line, "\n")
|
||||||
|
|
||||||
|
return line, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (tail *Tail) tailFileSync() {
|
||||||
|
defer tail.Done()
|
||||||
|
defer tail.close()
|
||||||
|
|
||||||
|
if !tail.MustExist {
|
||||||
|
// deferred first open.
|
||||||
|
err := tail.reopen()
|
||||||
|
if err != nil {
|
||||||
|
if err != tomb.ErrDying {
|
||||||
|
tail.Kill(err)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Seek to requested location on first open of the file.
|
||||||
|
if tail.Location != nil {
|
||||||
|
_, err := tail.file.Seek(tail.Location.Offset, tail.Location.Whence)
|
||||||
|
tail.Logger.Printf("Seeked %s - %+v\n", tail.Filename, tail.Location)
|
||||||
|
if err != nil {
|
||||||
|
tail.Killf("Seek error on %s: %s", tail.Filename, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
tail.openReader()
|
||||||
|
|
||||||
|
var offset int64
|
||||||
|
var err error
|
||||||
|
|
||||||
|
// Read line by line.
|
||||||
|
for {
|
||||||
|
// do not seek in named pipes
|
||||||
|
if !tail.Pipe {
|
||||||
|
// grab the position in case we need to back up in the event of a half-line
|
||||||
|
offset, err = tail.Tell()
|
||||||
|
if err != nil {
|
||||||
|
tail.Kill(err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
line, err := tail.readLine()
|
||||||
|
|
||||||
|
// Process `line` even if err is EOF.
|
||||||
|
if err == nil {
|
||||||
|
cooloff := !tail.sendLine(line)
|
||||||
|
if cooloff {
|
||||||
|
// Wait a second before seeking till the end of
|
||||||
|
// file when rate limit is reached.
|
||||||
|
msg := ("Too much log activity; waiting a second " +
|
||||||
|
"before resuming tailing")
|
||||||
|
tail.Lines <- &Line{msg, time.Now(), errors.New(msg)}
|
||||||
|
select {
|
||||||
|
case <-time.After(time.Second):
|
||||||
|
case <-tail.Dying():
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if err := tail.seekEnd(); err != nil {
|
||||||
|
tail.Kill(err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else if err == io.EOF {
|
||||||
|
if !tail.Follow {
|
||||||
|
if line != "" {
|
||||||
|
tail.sendLine(line)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if tail.Follow && line != "" {
|
||||||
|
// this has the potential to never return the last line if
|
||||||
|
// it's not followed by a newline; seems a fair trade here
|
||||||
|
err := tail.seekTo(SeekInfo{Offset: offset, Whence: 0})
|
||||||
|
if err != nil {
|
||||||
|
tail.Kill(err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// When EOF is reached, wait for more data to become
|
||||||
|
// available. Wait strategy is based on the `tail.watcher`
|
||||||
|
// implementation (inotify or polling).
|
||||||
|
err := tail.waitForChanges()
|
||||||
|
if err != nil {
|
||||||
|
if err != ErrStop {
|
||||||
|
tail.Kill(err)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// non-EOF error
|
||||||
|
tail.Killf("Error reading %s: %s", tail.Filename, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-tail.Dying():
|
||||||
|
if tail.Err() == errStopAtEOF {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
return
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// waitForChanges waits until the file has been appended, deleted,
|
||||||
|
// moved or truncated. When moved or deleted - the file will be
|
||||||
|
// reopened if ReOpen is true. Truncated files are always reopened.
|
||||||
|
func (tail *Tail) waitForChanges() error {
|
||||||
|
if tail.changes == nil {
|
||||||
|
pos, err := tail.file.Seek(0, os.SEEK_CUR)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
tail.changes, err = tail.watcher.ChangeEvents(&tail.Tomb, pos)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-tail.changes.Modified:
|
||||||
|
return nil
|
||||||
|
case <-tail.changes.Deleted:
|
||||||
|
tail.changes = nil
|
||||||
|
if tail.ReOpen {
|
||||||
|
// XXX: we must not log from a library.
|
||||||
|
tail.Logger.Printf("Re-opening moved/deleted file %s ...", tail.Filename)
|
||||||
|
if err := tail.reopen(); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
tail.Logger.Printf("Successfully reopened %s", tail.Filename)
|
||||||
|
tail.openReader()
|
||||||
|
return nil
|
||||||
|
} else {
|
||||||
|
tail.Logger.Printf("Stopping tail as file no longer exists: %s", tail.Filename)
|
||||||
|
return ErrStop
|
||||||
|
}
|
||||||
|
case <-tail.changes.Truncated:
|
||||||
|
// Always reopen truncated files (Follow is true)
|
||||||
|
tail.Logger.Printf("Re-opening truncated file %s ...", tail.Filename)
|
||||||
|
if err := tail.reopen(); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
tail.Logger.Printf("Successfully reopened truncated %s", tail.Filename)
|
||||||
|
tail.openReader()
|
||||||
|
return nil
|
||||||
|
case <-tail.Dying():
|
||||||
|
return ErrStop
|
||||||
|
}
|
||||||
|
panic("unreachable")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (tail *Tail) openReader() {
|
||||||
|
if tail.MaxLineSize > 0 {
|
||||||
|
// add 2 to account for newline characters
|
||||||
|
tail.reader = bufio.NewReaderSize(tail.file, tail.MaxLineSize+2)
|
||||||
|
} else {
|
||||||
|
tail.reader = bufio.NewReader(tail.file)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (tail *Tail) seekEnd() error {
|
||||||
|
return tail.seekTo(SeekInfo{Offset: 0, Whence: os.SEEK_END})
|
||||||
|
}
|
||||||
|
|
||||||
|
func (tail *Tail) seekTo(pos SeekInfo) error {
|
||||||
|
_, err := tail.file.Seek(pos.Offset, pos.Whence)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("Seek error on %s: %s", tail.Filename, err)
|
||||||
|
}
|
||||||
|
// Reset the read buffer whenever the file is re-seek'ed
|
||||||
|
tail.reader.Reset(tail.file)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// sendLine sends the line(s) to Lines channel, splitting longer lines
|
||||||
|
// if necessary. Return false if rate limit is reached.
|
||||||
|
func (tail *Tail) sendLine(line string) bool {
|
||||||
|
now := time.Now()
|
||||||
|
lines := []string{line}
|
||||||
|
|
||||||
|
// Split longer lines
|
||||||
|
if tail.MaxLineSize > 0 && len(line) > tail.MaxLineSize {
|
||||||
|
lines = util.PartitionString(line, tail.MaxLineSize)
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, line := range lines {
|
||||||
|
tail.Lines <- &Line{line, now, nil}
|
||||||
|
}
|
||||||
|
|
||||||
|
if tail.Config.RateLimiter != nil {
|
||||||
|
ok := tail.Config.RateLimiter.Pour(uint16(len(lines)))
|
||||||
|
if !ok {
|
||||||
|
tail.Logger.Printf("Leaky bucket full (%v); entering 1s cooloff period.\n",
|
||||||
|
tail.Filename)
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// Cleanup removes inotify watches added by the tail package. This function is
|
||||||
|
// meant to be invoked from a process's exit handler. Linux kernel may not
|
||||||
|
// automatically remove inotify watches after the process exits.
|
||||||
|
func (tail *Tail) Cleanup() {
|
||||||
|
watch.Cleanup(tail.Filename)
|
||||||
|
}
|
11
vendor/github.com/hpcloud/tail/tail_posix.go
generated
vendored
Normal file
11
vendor/github.com/hpcloud/tail/tail_posix.go
generated
vendored
Normal file
|
@ -0,0 +1,11 @@
|
||||||
|
// +build linux darwin freebsd netbsd openbsd
|
||||||
|
|
||||||
|
package tail
|
||||||
|
|
||||||
|
import (
|
||||||
|
"os"
|
||||||
|
)
|
||||||
|
|
||||||
|
func OpenFile(name string) (file *os.File, err error) {
|
||||||
|
return os.Open(name)
|
||||||
|
}
|
12
vendor/github.com/hpcloud/tail/tail_windows.go
generated
vendored
Normal file
12
vendor/github.com/hpcloud/tail/tail_windows.go
generated
vendored
Normal file
|
@ -0,0 +1,12 @@
|
||||||
|
// +build windows
|
||||||
|
|
||||||
|
package tail
|
||||||
|
|
||||||
|
import (
|
||||||
|
"github.com/hpcloud/tail/winfile"
|
||||||
|
"os"
|
||||||
|
)
|
||||||
|
|
||||||
|
func OpenFile(name string) (file *os.File, err error) {
|
||||||
|
return winfile.OpenFile(name, os.O_RDONLY, 0)
|
||||||
|
}
|
48
vendor/github.com/hpcloud/tail/util/util.go
generated
vendored
Normal file
48
vendor/github.com/hpcloud/tail/util/util.go
generated
vendored
Normal file
|
@ -0,0 +1,48 @@
|
||||||
|
// Copyright (c) 2015 HPE Software Inc. All rights reserved.
|
||||||
|
// Copyright (c) 2013 ActiveState Software Inc. All rights reserved.
|
||||||
|
|
||||||
|
package util
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"log"
|
||||||
|
"os"
|
||||||
|
"runtime/debug"
|
||||||
|
)
|
||||||
|
|
||||||
|
type Logger struct {
|
||||||
|
*log.Logger
|
||||||
|
}
|
||||||
|
|
||||||
|
var LOGGER = &Logger{log.New(os.Stderr, "", log.LstdFlags)}
|
||||||
|
|
||||||
|
// fatal is like panic except it displays only the current goroutine's stack.
|
||||||
|
func Fatal(format string, v ...interface{}) {
|
||||||
|
// https://github.com/hpcloud/log/blob/master/log.go#L45
|
||||||
|
LOGGER.Output(2, fmt.Sprintf("FATAL -- "+format, v...)+"\n"+string(debug.Stack()))
|
||||||
|
os.Exit(1)
|
||||||
|
}
|
||||||
|
|
||||||
|
// partitionString partitions the string into chunks of given size,
|
||||||
|
// with the last chunk of variable size.
|
||||||
|
func PartitionString(s string, chunkSize int) []string {
|
||||||
|
if chunkSize <= 0 {
|
||||||
|
panic("invalid chunkSize")
|
||||||
|
}
|
||||||
|
length := len(s)
|
||||||
|
chunks := 1 + length/chunkSize
|
||||||
|
start := 0
|
||||||
|
end := chunkSize
|
||||||
|
parts := make([]string, 0, chunks)
|
||||||
|
for {
|
||||||
|
if end > length {
|
||||||
|
end = length
|
||||||
|
}
|
||||||
|
parts = append(parts, s[start:end])
|
||||||
|
if end == length {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
start, end = end, end+chunkSize
|
||||||
|
}
|
||||||
|
return parts
|
||||||
|
}
|
36
vendor/github.com/hpcloud/tail/watch/filechanges.go
generated
vendored
Normal file
36
vendor/github.com/hpcloud/tail/watch/filechanges.go
generated
vendored
Normal file
|
@ -0,0 +1,36 @@
|
||||||
|
package watch
|
||||||
|
|
||||||
|
type FileChanges struct {
|
||||||
|
Modified chan bool // Channel to get notified of modifications
|
||||||
|
Truncated chan bool // Channel to get notified of truncations
|
||||||
|
Deleted chan bool // Channel to get notified of deletions/renames
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewFileChanges() *FileChanges {
|
||||||
|
return &FileChanges{
|
||||||
|
make(chan bool, 1), make(chan bool, 1), make(chan bool, 1)}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (fc *FileChanges) NotifyModified() {
|
||||||
|
sendOnlyIfEmpty(fc.Modified)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (fc *FileChanges) NotifyTruncated() {
|
||||||
|
sendOnlyIfEmpty(fc.Truncated)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (fc *FileChanges) NotifyDeleted() {
|
||||||
|
sendOnlyIfEmpty(fc.Deleted)
|
||||||
|
}
|
||||||
|
|
||||||
|
// sendOnlyIfEmpty sends on a bool channel only if the channel has no
|
||||||
|
// backlog to be read by other goroutines. This concurrency pattern
|
||||||
|
// can be used to notify other goroutines if and only if they are
|
||||||
|
// looking for it (i.e., subsequent notifications can be compressed
|
||||||
|
// into one).
|
||||||
|
func sendOnlyIfEmpty(ch chan bool) {
|
||||||
|
select {
|
||||||
|
case ch <- true:
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
}
|
135
vendor/github.com/hpcloud/tail/watch/inotify.go
generated
vendored
Normal file
135
vendor/github.com/hpcloud/tail/watch/inotify.go
generated
vendored
Normal file
|
@ -0,0 +1,135 @@
|
||||||
|
// Copyright (c) 2015 HPE Software Inc. All rights reserved.
|
||||||
|
// Copyright (c) 2013 ActiveState Software Inc. All rights reserved.
|
||||||
|
|
||||||
|
package watch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
|
|
||||||
|
"github.com/hpcloud/tail/util"
|
||||||
|
|
||||||
|
"gopkg.in/fsnotify/fsnotify.v1"
|
||||||
|
"gopkg.in/tomb.v1"
|
||||||
|
)
|
||||||
|
|
||||||
|
// InotifyFileWatcher uses inotify to monitor file changes.
|
||||||
|
type InotifyFileWatcher struct {
|
||||||
|
Filename string
|
||||||
|
Size int64
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewInotifyFileWatcher(filename string) *InotifyFileWatcher {
|
||||||
|
fw := &InotifyFileWatcher{filepath.Clean(filename), 0}
|
||||||
|
return fw
|
||||||
|
}
|
||||||
|
|
||||||
|
func (fw *InotifyFileWatcher) BlockUntilExists(t *tomb.Tomb) error {
|
||||||
|
err := WatchCreate(fw.Filename)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
defer RemoveWatchCreate(fw.Filename)
|
||||||
|
|
||||||
|
// Do a real check now as the file might have been created before
|
||||||
|
// calling `WatchFlags` above.
|
||||||
|
if _, err = os.Stat(fw.Filename); !os.IsNotExist(err) {
|
||||||
|
// file exists, or stat returned an error.
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
events := Events(fw.Filename)
|
||||||
|
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case evt, ok := <-events:
|
||||||
|
if !ok {
|
||||||
|
return fmt.Errorf("inotify watcher has been closed")
|
||||||
|
}
|
||||||
|
evtName, err := filepath.Abs(evt.Name)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
fwFilename, err := filepath.Abs(fw.Filename)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if evtName == fwFilename {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
case <-t.Dying():
|
||||||
|
return tomb.ErrDying
|
||||||
|
}
|
||||||
|
}
|
||||||
|
panic("unreachable")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (fw *InotifyFileWatcher) ChangeEvents(t *tomb.Tomb, pos int64) (*FileChanges, error) {
|
||||||
|
err := Watch(fw.Filename)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
changes := NewFileChanges()
|
||||||
|
fw.Size = pos
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
|
||||||
|
events := Events(fw.Filename)
|
||||||
|
|
||||||
|
for {
|
||||||
|
prevSize := fw.Size
|
||||||
|
|
||||||
|
var evt fsnotify.Event
|
||||||
|
var ok bool
|
||||||
|
|
||||||
|
select {
|
||||||
|
case evt, ok = <-events:
|
||||||
|
if !ok {
|
||||||
|
RemoveWatch(fw.Filename)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
case <-t.Dying():
|
||||||
|
RemoveWatch(fw.Filename)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
switch {
|
||||||
|
case evt.Op&fsnotify.Remove == fsnotify.Remove:
|
||||||
|
fallthrough
|
||||||
|
|
||||||
|
case evt.Op&fsnotify.Rename == fsnotify.Rename:
|
||||||
|
RemoveWatch(fw.Filename)
|
||||||
|
changes.NotifyDeleted()
|
||||||
|
return
|
||||||
|
|
||||||
|
//With an open fd, unlink(fd) - inotify returns IN_ATTRIB (==fsnotify.Chmod)
|
||||||
|
case evt.Op&fsnotify.Chmod == fsnotify.Chmod:
|
||||||
|
fallthrough
|
||||||
|
|
||||||
|
case evt.Op&fsnotify.Write == fsnotify.Write:
|
||||||
|
fi, err := os.Stat(fw.Filename)
|
||||||
|
if err != nil {
|
||||||
|
if os.IsNotExist(err) {
|
||||||
|
RemoveWatch(fw.Filename)
|
||||||
|
changes.NotifyDeleted()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
// XXX: report this error back to the user
|
||||||
|
util.Fatal("Failed to stat file %v: %v", fw.Filename, err)
|
||||||
|
}
|
||||||
|
fw.Size = fi.Size()
|
||||||
|
|
||||||
|
if prevSize > 0 && prevSize > fw.Size {
|
||||||
|
changes.NotifyTruncated()
|
||||||
|
} else {
|
||||||
|
changes.NotifyModified()
|
||||||
|
}
|
||||||
|
prevSize = fw.Size
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
return changes, nil
|
||||||
|
}
|
248
vendor/github.com/hpcloud/tail/watch/inotify_tracker.go
generated
vendored
Normal file
248
vendor/github.com/hpcloud/tail/watch/inotify_tracker.go
generated
vendored
Normal file
|
@ -0,0 +1,248 @@
|
||||||
|
// Copyright (c) 2015 HPE Software Inc. All rights reserved.
|
||||||
|
// Copyright (c) 2013 ActiveState Software Inc. All rights reserved.
|
||||||
|
|
||||||
|
package watch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"log"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
|
"sync"
|
||||||
|
"syscall"
|
||||||
|
|
||||||
|
"github.com/hpcloud/tail/util"
|
||||||
|
|
||||||
|
"gopkg.in/fsnotify/fsnotify.v1"
|
||||||
|
)
|
||||||
|
|
||||||
|
type InotifyTracker struct {
|
||||||
|
mux sync.Mutex
|
||||||
|
watcher *fsnotify.Watcher
|
||||||
|
chans map[string]chan fsnotify.Event
|
||||||
|
done map[string]chan bool
|
||||||
|
watchNums map[string]int
|
||||||
|
watch chan *watchInfo
|
||||||
|
remove chan *watchInfo
|
||||||
|
error chan error
|
||||||
|
}
|
||||||
|
|
||||||
|
type watchInfo struct {
|
||||||
|
op fsnotify.Op
|
||||||
|
fname string
|
||||||
|
}
|
||||||
|
|
||||||
|
func (this *watchInfo) isCreate() bool {
|
||||||
|
return this.op == fsnotify.Create
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
// globally shared InotifyTracker; ensures only one fsnotify.Watcher is used
|
||||||
|
shared *InotifyTracker
|
||||||
|
|
||||||
|
// these are used to ensure the shared InotifyTracker is run exactly once
|
||||||
|
once = sync.Once{}
|
||||||
|
goRun = func() {
|
||||||
|
shared = &InotifyTracker{
|
||||||
|
mux: sync.Mutex{},
|
||||||
|
chans: make(map[string]chan fsnotify.Event),
|
||||||
|
done: make(map[string]chan bool),
|
||||||
|
watchNums: make(map[string]int),
|
||||||
|
watch: make(chan *watchInfo),
|
||||||
|
remove: make(chan *watchInfo),
|
||||||
|
error: make(chan error),
|
||||||
|
}
|
||||||
|
go shared.run()
|
||||||
|
}
|
||||||
|
|
||||||
|
logger = log.New(os.Stderr, "", log.LstdFlags)
|
||||||
|
)
|
||||||
|
|
||||||
|
// Watch signals the run goroutine to begin watching the input filename
|
||||||
|
func Watch(fname string) error {
|
||||||
|
return watch(&watchInfo{
|
||||||
|
fname: fname,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
// Watch create signals the run goroutine to begin watching the input filename
|
||||||
|
// if call the WatchCreate function, don't call the Cleanup, call the RemoveWatchCreate
|
||||||
|
func WatchCreate(fname string) error {
|
||||||
|
return watch(&watchInfo{
|
||||||
|
op: fsnotify.Create,
|
||||||
|
fname: fname,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func watch(winfo *watchInfo) error {
|
||||||
|
// start running the shared InotifyTracker if not already running
|
||||||
|
once.Do(goRun)
|
||||||
|
|
||||||
|
winfo.fname = filepath.Clean(winfo.fname)
|
||||||
|
shared.watch <- winfo
|
||||||
|
return <-shared.error
|
||||||
|
}
|
||||||
|
|
||||||
|
// RemoveWatch signals the run goroutine to remove the watch for the input filename
|
||||||
|
func RemoveWatch(fname string) error {
|
||||||
|
return remove(&watchInfo{
|
||||||
|
fname: fname,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
// RemoveWatch create signals the run goroutine to remove the watch for the input filename
|
||||||
|
func RemoveWatchCreate(fname string) error {
|
||||||
|
return remove(&watchInfo{
|
||||||
|
op: fsnotify.Create,
|
||||||
|
fname: fname,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func remove(winfo *watchInfo) error {
|
||||||
|
// start running the shared InotifyTracker if not already running
|
||||||
|
once.Do(goRun)
|
||||||
|
|
||||||
|
winfo.fname = filepath.Clean(winfo.fname)
|
||||||
|
shared.mux.Lock()
|
||||||
|
done := shared.done[winfo.fname]
|
||||||
|
if done != nil {
|
||||||
|
delete(shared.done, winfo.fname)
|
||||||
|
close(done)
|
||||||
|
}
|
||||||
|
shared.mux.Unlock()
|
||||||
|
|
||||||
|
shared.remove <- winfo
|
||||||
|
return <-shared.error
|
||||||
|
}
|
||||||
|
|
||||||
|
// Events returns a channel to which FileEvents corresponding to the input filename
|
||||||
|
// will be sent. This channel will be closed when removeWatch is called on this
|
||||||
|
// filename.
|
||||||
|
func Events(fname string) <-chan fsnotify.Event {
|
||||||
|
shared.mux.Lock()
|
||||||
|
defer shared.mux.Unlock()
|
||||||
|
|
||||||
|
return shared.chans[fname]
|
||||||
|
}
|
||||||
|
|
||||||
|
// Cleanup removes the watch for the input filename if necessary.
|
||||||
|
func Cleanup(fname string) error {
|
||||||
|
return RemoveWatch(fname)
|
||||||
|
}
|
||||||
|
|
||||||
|
// watchFlags calls fsnotify.WatchFlags for the input filename and flags, creating
|
||||||
|
// a new Watcher if the previous Watcher was closed.
|
||||||
|
func (shared *InotifyTracker) addWatch(winfo *watchInfo) error {
|
||||||
|
shared.mux.Lock()
|
||||||
|
defer shared.mux.Unlock()
|
||||||
|
|
||||||
|
if shared.chans[winfo.fname] == nil {
|
||||||
|
shared.chans[winfo.fname] = make(chan fsnotify.Event)
|
||||||
|
}
|
||||||
|
if shared.done[winfo.fname] == nil {
|
||||||
|
shared.done[winfo.fname] = make(chan bool)
|
||||||
|
}
|
||||||
|
|
||||||
|
fname := winfo.fname
|
||||||
|
if winfo.isCreate() {
|
||||||
|
// Watch for new files to be created in the parent directory.
|
||||||
|
fname = filepath.Dir(fname)
|
||||||
|
}
|
||||||
|
|
||||||
|
var err error
|
||||||
|
// already in inotify watch
|
||||||
|
if shared.watchNums[fname] == 0 {
|
||||||
|
err = shared.watcher.Add(fname)
|
||||||
|
}
|
||||||
|
if err == nil {
|
||||||
|
shared.watchNums[fname]++
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// removeWatch calls fsnotify.RemoveWatch for the input filename and closes the
|
||||||
|
// corresponding events channel.
|
||||||
|
func (shared *InotifyTracker) removeWatch(winfo *watchInfo) error {
|
||||||
|
shared.mux.Lock()
|
||||||
|
|
||||||
|
ch := shared.chans[winfo.fname]
|
||||||
|
if ch != nil {
|
||||||
|
delete(shared.chans, winfo.fname)
|
||||||
|
close(ch)
|
||||||
|
}
|
||||||
|
|
||||||
|
fname := winfo.fname
|
||||||
|
if winfo.isCreate() {
|
||||||
|
// Watch for new files to be created in the parent directory.
|
||||||
|
fname = filepath.Dir(fname)
|
||||||
|
}
|
||||||
|
shared.watchNums[fname]--
|
||||||
|
watchNum := shared.watchNums[fname]
|
||||||
|
if watchNum == 0 {
|
||||||
|
delete(shared.watchNums, fname)
|
||||||
|
}
|
||||||
|
shared.mux.Unlock()
|
||||||
|
|
||||||
|
var err error
|
||||||
|
// If we were the last ones to watch this file, unsubscribe from inotify.
|
||||||
|
// This needs to happen after releasing the lock because fsnotify waits
|
||||||
|
// synchronously for the kernel to acknowledge the removal of the watch
|
||||||
|
// for this file, which causes us to deadlock if we still held the lock.
|
||||||
|
if watchNum == 0 {
|
||||||
|
err = shared.watcher.Remove(fname)
|
||||||
|
}
|
||||||
|
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// sendEvent sends the input event to the appropriate Tail.
|
||||||
|
func (shared *InotifyTracker) sendEvent(event fsnotify.Event) {
|
||||||
|
name := filepath.Clean(event.Name)
|
||||||
|
|
||||||
|
shared.mux.Lock()
|
||||||
|
ch := shared.chans[name]
|
||||||
|
done := shared.done[name]
|
||||||
|
shared.mux.Unlock()
|
||||||
|
|
||||||
|
if ch != nil && done != nil {
|
||||||
|
select {
|
||||||
|
case ch <- event:
|
||||||
|
case <-done:
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// run starts the goroutine in which the shared struct reads events from its
|
||||||
|
// Watcher's Event channel and sends the events to the appropriate Tail.
|
||||||
|
func (shared *InotifyTracker) run() {
|
||||||
|
watcher, err := fsnotify.NewWatcher()
|
||||||
|
if err != nil {
|
||||||
|
util.Fatal("failed to create Watcher")
|
||||||
|
}
|
||||||
|
shared.watcher = watcher
|
||||||
|
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case winfo := <-shared.watch:
|
||||||
|
shared.error <- shared.addWatch(winfo)
|
||||||
|
|
||||||
|
case winfo := <-shared.remove:
|
||||||
|
shared.error <- shared.removeWatch(winfo)
|
||||||
|
|
||||||
|
case event, open := <-shared.watcher.Events:
|
||||||
|
if !open {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
shared.sendEvent(event)
|
||||||
|
|
||||||
|
case err, open := <-shared.watcher.Errors:
|
||||||
|
if !open {
|
||||||
|
return
|
||||||
|
} else if err != nil {
|
||||||
|
sysErr, ok := err.(*os.SyscallError)
|
||||||
|
if !ok || sysErr.Err != syscall.EINTR {
|
||||||
|
logger.Printf("Error in Watcher Error channel: %s", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
118
vendor/github.com/hpcloud/tail/watch/polling.go
generated
vendored
Normal file
118
vendor/github.com/hpcloud/tail/watch/polling.go
generated
vendored
Normal file
|
@ -0,0 +1,118 @@
|
||||||
|
// Copyright (c) 2015 HPE Software Inc. All rights reserved.
|
||||||
|
// Copyright (c) 2013 ActiveState Software Inc. All rights reserved.
|
||||||
|
|
||||||
|
package watch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"os"
|
||||||
|
"runtime"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/hpcloud/tail/util"
|
||||||
|
"gopkg.in/tomb.v1"
|
||||||
|
)
|
||||||
|
|
||||||
|
// PollingFileWatcher polls the file for changes.
|
||||||
|
type PollingFileWatcher struct {
|
||||||
|
Filename string
|
||||||
|
Size int64
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewPollingFileWatcher(filename string) *PollingFileWatcher {
|
||||||
|
fw := &PollingFileWatcher{filename, 0}
|
||||||
|
return fw
|
||||||
|
}
|
||||||
|
|
||||||
|
var POLL_DURATION time.Duration
|
||||||
|
|
||||||
|
func (fw *PollingFileWatcher) BlockUntilExists(t *tomb.Tomb) error {
|
||||||
|
for {
|
||||||
|
if _, err := os.Stat(fw.Filename); err == nil {
|
||||||
|
return nil
|
||||||
|
} else if !os.IsNotExist(err) {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
select {
|
||||||
|
case <-time.After(POLL_DURATION):
|
||||||
|
continue
|
||||||
|
case <-t.Dying():
|
||||||
|
return tomb.ErrDying
|
||||||
|
}
|
||||||
|
}
|
||||||
|
panic("unreachable")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (fw *PollingFileWatcher) ChangeEvents(t *tomb.Tomb, pos int64) (*FileChanges, error) {
|
||||||
|
origFi, err := os.Stat(fw.Filename)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
changes := NewFileChanges()
|
||||||
|
var prevModTime time.Time
|
||||||
|
|
||||||
|
// XXX: use tomb.Tomb to cleanly manage these goroutines. replace
|
||||||
|
// the fatal (below) with tomb's Kill.
|
||||||
|
|
||||||
|
fw.Size = pos
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
prevSize := fw.Size
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-t.Dying():
|
||||||
|
return
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
|
||||||
|
time.Sleep(POLL_DURATION)
|
||||||
|
fi, err := os.Stat(fw.Filename)
|
||||||
|
if err != nil {
|
||||||
|
// Windows cannot delete a file if a handle is still open (tail keeps one open)
|
||||||
|
// so it gives access denied to anything trying to read it until all handles are released.
|
||||||
|
if os.IsNotExist(err) || (runtime.GOOS == "windows" && os.IsPermission(err)) {
|
||||||
|
// File does not exist (has been deleted).
|
||||||
|
changes.NotifyDeleted()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// XXX: report this error back to the user
|
||||||
|
util.Fatal("Failed to stat file %v: %v", fw.Filename, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// File got moved/renamed?
|
||||||
|
if !os.SameFile(origFi, fi) {
|
||||||
|
changes.NotifyDeleted()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// File got truncated?
|
||||||
|
fw.Size = fi.Size()
|
||||||
|
if prevSize > 0 && prevSize > fw.Size {
|
||||||
|
changes.NotifyTruncated()
|
||||||
|
prevSize = fw.Size
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// File got bigger?
|
||||||
|
if prevSize > 0 && prevSize < fw.Size {
|
||||||
|
changes.NotifyModified()
|
||||||
|
prevSize = fw.Size
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
prevSize = fw.Size
|
||||||
|
|
||||||
|
// File was appended to (changed)?
|
||||||
|
modTime := fi.ModTime()
|
||||||
|
if modTime != prevModTime {
|
||||||
|
prevModTime = modTime
|
||||||
|
changes.NotifyModified()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
return changes, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
POLL_DURATION = 250 * time.Millisecond
|
||||||
|
}
|
20
vendor/github.com/hpcloud/tail/watch/watch.go
generated
vendored
Normal file
20
vendor/github.com/hpcloud/tail/watch/watch.go
generated
vendored
Normal file
|
@ -0,0 +1,20 @@
|
||||||
|
// Copyright (c) 2015 HPE Software Inc. All rights reserved.
|
||||||
|
// Copyright (c) 2013 ActiveState Software Inc. All rights reserved.
|
||||||
|
|
||||||
|
package watch
|
||||||
|
|
||||||
|
import "gopkg.in/tomb.v1"
|
||||||
|
|
||||||
|
// FileWatcher monitors file-level events.
|
||||||
|
type FileWatcher interface {
|
||||||
|
// BlockUntilExists blocks until the file comes into existence.
|
||||||
|
BlockUntilExists(*tomb.Tomb) error
|
||||||
|
|
||||||
|
// ChangeEvents reports on changes to a file, be it modification,
|
||||||
|
// deletion, renames or truncations. Returned FileChanges group of
|
||||||
|
// channels will be closed, thus become unusable, after a deletion
|
||||||
|
// or truncation event.
|
||||||
|
// In order to properly report truncations, ChangeEvents requires
|
||||||
|
// the caller to pass their current offset in the file.
|
||||||
|
ChangeEvents(*tomb.Tomb, int64) (*FileChanges, error)
|
||||||
|
}
|
92
vendor/github.com/hpcloud/tail/winfile/winfile.go
generated
vendored
Normal file
92
vendor/github.com/hpcloud/tail/winfile/winfile.go
generated
vendored
Normal file
|
@ -0,0 +1,92 @@
|
||||||
|
// +build windows
|
||||||
|
|
||||||
|
package winfile
|
||||||
|
|
||||||
|
import (
|
||||||
|
"os"
|
||||||
|
"syscall"
|
||||||
|
"unsafe"
|
||||||
|
)
|
||||||
|
|
||||||
|
// issue also described here
|
||||||
|
//https://codereview.appspot.com/8203043/
|
||||||
|
|
||||||
|
// https://github.com/jnwhiteh/golang/blob/master/src/pkg/syscall/syscall_windows.go#L218
|
||||||
|
func Open(path string, mode int, perm uint32) (fd syscall.Handle, err error) {
|
||||||
|
if len(path) == 0 {
|
||||||
|
return syscall.InvalidHandle, syscall.ERROR_FILE_NOT_FOUND
|
||||||
|
}
|
||||||
|
pathp, err := syscall.UTF16PtrFromString(path)
|
||||||
|
if err != nil {
|
||||||
|
return syscall.InvalidHandle, err
|
||||||
|
}
|
||||||
|
var access uint32
|
||||||
|
switch mode & (syscall.O_RDONLY | syscall.O_WRONLY | syscall.O_RDWR) {
|
||||||
|
case syscall.O_RDONLY:
|
||||||
|
access = syscall.GENERIC_READ
|
||||||
|
case syscall.O_WRONLY:
|
||||||
|
access = syscall.GENERIC_WRITE
|
||||||
|
case syscall.O_RDWR:
|
||||||
|
access = syscall.GENERIC_READ | syscall.GENERIC_WRITE
|
||||||
|
}
|
||||||
|
if mode&syscall.O_CREAT != 0 {
|
||||||
|
access |= syscall.GENERIC_WRITE
|
||||||
|
}
|
||||||
|
if mode&syscall.O_APPEND != 0 {
|
||||||
|
access &^= syscall.GENERIC_WRITE
|
||||||
|
access |= syscall.FILE_APPEND_DATA
|
||||||
|
}
|
||||||
|
sharemode := uint32(syscall.FILE_SHARE_READ | syscall.FILE_SHARE_WRITE | syscall.FILE_SHARE_DELETE)
|
||||||
|
var sa *syscall.SecurityAttributes
|
||||||
|
if mode&syscall.O_CLOEXEC == 0 {
|
||||||
|
sa = makeInheritSa()
|
||||||
|
}
|
||||||
|
var createmode uint32
|
||||||
|
switch {
|
||||||
|
case mode&(syscall.O_CREAT|syscall.O_EXCL) == (syscall.O_CREAT | syscall.O_EXCL):
|
||||||
|
createmode = syscall.CREATE_NEW
|
||||||
|
case mode&(syscall.O_CREAT|syscall.O_TRUNC) == (syscall.O_CREAT | syscall.O_TRUNC):
|
||||||
|
createmode = syscall.CREATE_ALWAYS
|
||||||
|
case mode&syscall.O_CREAT == syscall.O_CREAT:
|
||||||
|
createmode = syscall.OPEN_ALWAYS
|
||||||
|
case mode&syscall.O_TRUNC == syscall.O_TRUNC:
|
||||||
|
createmode = syscall.TRUNCATE_EXISTING
|
||||||
|
default:
|
||||||
|
createmode = syscall.OPEN_EXISTING
|
||||||
|
}
|
||||||
|
h, e := syscall.CreateFile(pathp, access, sharemode, sa, createmode, syscall.FILE_ATTRIBUTE_NORMAL, 0)
|
||||||
|
return h, e
|
||||||
|
}
|
||||||
|
|
||||||
|
// https://github.com/jnwhiteh/golang/blob/master/src/pkg/syscall/syscall_windows.go#L211
|
||||||
|
func makeInheritSa() *syscall.SecurityAttributes {
|
||||||
|
var sa syscall.SecurityAttributes
|
||||||
|
sa.Length = uint32(unsafe.Sizeof(sa))
|
||||||
|
sa.InheritHandle = 1
|
||||||
|
return &sa
|
||||||
|
}
|
||||||
|
|
||||||
|
// https://github.com/jnwhiteh/golang/blob/master/src/pkg/os/file_windows.go#L133
|
||||||
|
func OpenFile(name string, flag int, perm os.FileMode) (file *os.File, err error) {
|
||||||
|
r, e := Open(name, flag|syscall.O_CLOEXEC, syscallMode(perm))
|
||||||
|
if e != nil {
|
||||||
|
return nil, e
|
||||||
|
}
|
||||||
|
return os.NewFile(uintptr(r), name), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// https://github.com/jnwhiteh/golang/blob/master/src/pkg/os/file_posix.go#L61
|
||||||
|
func syscallMode(i os.FileMode) (o uint32) {
|
||||||
|
o |= uint32(i.Perm())
|
||||||
|
if i&os.ModeSetuid != 0 {
|
||||||
|
o |= syscall.S_ISUID
|
||||||
|
}
|
||||||
|
if i&os.ModeSetgid != 0 {
|
||||||
|
o |= syscall.S_ISGID
|
||||||
|
}
|
||||||
|
if i&os.ModeSticky != 0 {
|
||||||
|
o |= syscall.S_ISVTX
|
||||||
|
}
|
||||||
|
// No mapping for Go's ModeTemporary (plan9 only).
|
||||||
|
return
|
||||||
|
}
|
14
vendor/github.com/mitchellh/go-homedir/README.md
generated
vendored
14
vendor/github.com/mitchellh/go-homedir/README.md
generated
vendored
|
@ -1,14 +0,0 @@
|
||||||
# go-homedir
|
|
||||||
|
|
||||||
This is a Go library for detecting the user's home directory without
|
|
||||||
the use of cgo, so the library can be used in cross-compilation environments.
|
|
||||||
|
|
||||||
Usage is incredibly simple, just call `homedir.Dir()` to get the home directory
|
|
||||||
for a user, and `homedir.Expand()` to expand the `~` in a path to the home
|
|
||||||
directory.
|
|
||||||
|
|
||||||
**Why not just use `os/user`?** The built-in `os/user` package requires
|
|
||||||
cgo on Darwin systems. This means that any Go code that uses that package
|
|
||||||
cannot cross compile. But 99% of the time the use for `os/user` is just to
|
|
||||||
retrieve the home directory, which we can do for the current user without
|
|
||||||
cgo. This library does that, enabling cross-compilation.
|
|
61
vendor/github.com/mitchellh/go-homedir/homedir.go
generated
vendored
61
vendor/github.com/mitchellh/go-homedir/homedir.go
generated
vendored
|
@ -77,33 +77,51 @@ func Expand(path string) (string, error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func dirUnix() (string, error) {
|
func dirUnix() (string, error) {
|
||||||
|
homeEnv := "HOME"
|
||||||
|
if runtime.GOOS == "plan9" {
|
||||||
|
// On plan9, env vars are lowercase.
|
||||||
|
homeEnv = "home"
|
||||||
|
}
|
||||||
|
|
||||||
// First prefer the HOME environmental variable
|
// First prefer the HOME environmental variable
|
||||||
if home := os.Getenv("HOME"); home != "" {
|
if home := os.Getenv(homeEnv); home != "" {
|
||||||
return home, nil
|
return home, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// If that fails, try getent
|
|
||||||
var stdout bytes.Buffer
|
var stdout bytes.Buffer
|
||||||
cmd := exec.Command("getent", "passwd", strconv.Itoa(os.Getuid()))
|
|
||||||
cmd.Stdout = &stdout
|
// If that fails, try OS specific commands
|
||||||
if err := cmd.Run(); err != nil {
|
if runtime.GOOS == "darwin" {
|
||||||
// If "getent" is missing, ignore it
|
cmd := exec.Command("sh", "-c", `dscl -q . -read /Users/"$(whoami)" NFSHomeDirectory | sed 's/^[^ ]*: //'`)
|
||||||
if err == exec.ErrNotFound {
|
cmd.Stdout = &stdout
|
||||||
return "", err
|
if err := cmd.Run(); err == nil {
|
||||||
|
result := strings.TrimSpace(stdout.String())
|
||||||
|
if result != "" {
|
||||||
|
return result, nil
|
||||||
|
}
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
if passwd := strings.TrimSpace(stdout.String()); passwd != "" {
|
cmd := exec.Command("getent", "passwd", strconv.Itoa(os.Getuid()))
|
||||||
// username:password:uid:gid:gecos:home:shell
|
cmd.Stdout = &stdout
|
||||||
passwdParts := strings.SplitN(passwd, ":", 7)
|
if err := cmd.Run(); err != nil {
|
||||||
if len(passwdParts) > 5 {
|
// If the error is ErrNotFound, we ignore it. Otherwise, return it.
|
||||||
return passwdParts[5], nil
|
if err != exec.ErrNotFound {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if passwd := strings.TrimSpace(stdout.String()); passwd != "" {
|
||||||
|
// username:password:uid:gid:gecos:home:shell
|
||||||
|
passwdParts := strings.SplitN(passwd, ":", 7)
|
||||||
|
if len(passwdParts) > 5 {
|
||||||
|
return passwdParts[5], nil
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// If all else fails, try the shell
|
// If all else fails, try the shell
|
||||||
stdout.Reset()
|
stdout.Reset()
|
||||||
cmd = exec.Command("sh", "-c", "cd && pwd")
|
cmd := exec.Command("sh", "-c", "cd && pwd")
|
||||||
cmd.Stdout = &stdout
|
cmd.Stdout = &stdout
|
||||||
if err := cmd.Run(); err != nil {
|
if err := cmd.Run(); err != nil {
|
||||||
return "", err
|
return "", err
|
||||||
|
@ -118,14 +136,21 @@ func dirUnix() (string, error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func dirWindows() (string, error) {
|
func dirWindows() (string, error) {
|
||||||
|
// First prefer the HOME environmental variable
|
||||||
|
if home := os.Getenv("HOME"); home != "" {
|
||||||
|
return home, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Prefer standard environment variable USERPROFILE
|
||||||
|
if home := os.Getenv("USERPROFILE"); home != "" {
|
||||||
|
return home, nil
|
||||||
|
}
|
||||||
|
|
||||||
drive := os.Getenv("HOMEDRIVE")
|
drive := os.Getenv("HOMEDRIVE")
|
||||||
path := os.Getenv("HOMEPATH")
|
path := os.Getenv("HOMEPATH")
|
||||||
home := drive + path
|
home := drive + path
|
||||||
if drive == "" || path == "" {
|
if drive == "" || path == "" {
|
||||||
home = os.Getenv("USERPROFILE")
|
return "", errors.New("HOMEDRIVE, HOMEPATH, or USERPROFILE are blank")
|
||||||
}
|
|
||||||
if home == "" {
|
|
||||||
return "", errors.New("HOMEDRIVE, HOMEPATH, and USERPROFILE are blank")
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return home, nil
|
return home, nil
|
||||||
|
|
27
vendor/github.com/nightlyone/lockfile/.gitignore
generated
vendored
27
vendor/github.com/nightlyone/lockfile/.gitignore
generated
vendored
|
@ -1,27 +0,0 @@
|
||||||
# Compiled Object files, Static and Dynamic libs (Shared Objects)
|
|
||||||
*.o
|
|
||||||
*.a
|
|
||||||
*.so
|
|
||||||
|
|
||||||
# Folders
|
|
||||||
_obj
|
|
||||||
_test
|
|
||||||
|
|
||||||
# popular temporaries
|
|
||||||
.err
|
|
||||||
.out
|
|
||||||
.diff
|
|
||||||
|
|
||||||
# Architecture specific extensions/prefixes
|
|
||||||
*.[568vq]
|
|
||||||
[568vq].out
|
|
||||||
|
|
||||||
*.cgo1.go
|
|
||||||
*.cgo2.c
|
|
||||||
_cgo_defun.c
|
|
||||||
_cgo_gotypes.go
|
|
||||||
_cgo_export.*
|
|
||||||
|
|
||||||
_testmain.go
|
|
||||||
|
|
||||||
*.exe
|
|
3
vendor/github.com/nightlyone/lockfile/.gitmodules
generated
vendored
3
vendor/github.com/nightlyone/lockfile/.gitmodules
generated
vendored
|
@ -1,3 +0,0 @@
|
||||||
[submodule "git-hooks"]
|
|
||||||
path = git-hooks
|
|
||||||
url = https://github.com/nightlyone/git-hooks
|
|
2
vendor/github.com/nightlyone/lockfile/.travis.yml
generated
vendored
2
vendor/github.com/nightlyone/lockfile/.travis.yml
generated
vendored
|
@ -1,2 +0,0 @@
|
||||||
language: go
|
|
||||||
|
|
52
vendor/github.com/nightlyone/lockfile/README.md
generated
vendored
52
vendor/github.com/nightlyone/lockfile/README.md
generated
vendored
|
@ -1,52 +0,0 @@
|
||||||
lockfile
|
|
||||||
=========
|
|
||||||
Handle locking via pid files.
|
|
||||||
|
|
||||||
[![Build Status Unix][1]][2]
|
|
||||||
[![Build status Windows][3]][4]
|
|
||||||
|
|
||||||
[1]: https://secure.travis-ci.org/nightlyone/lockfile.png
|
|
||||||
[2]: https://travis-ci.org/nightlyone/lockfile
|
|
||||||
[3]: https://ci.appveyor.com/api/projects/status/7mojkmauj81uvp8u/branch/master?svg=true
|
|
||||||
[4]: https://ci.appveyor.com/project/nightlyone/lockfile/branch/master
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
install
|
|
||||||
-------
|
|
||||||
Install [Go 1][5], either [from source][6] or [with a prepackaged binary][7].
|
|
||||||
For Windows suport, Go 1.4 or newer is required.
|
|
||||||
|
|
||||||
Then run
|
|
||||||
|
|
||||||
go get github.com/nightlyone/lockfile
|
|
||||||
|
|
||||||
[5]: http://golang.org
|
|
||||||
[6]: http://golang.org/doc/install/source
|
|
||||||
[7]: http://golang.org/doc/install
|
|
||||||
|
|
||||||
LICENSE
|
|
||||||
-------
|
|
||||||
BSD
|
|
||||||
|
|
||||||
documentation
|
|
||||||
-------------
|
|
||||||
[package documentation at godoc.org](http://godoc.org/github.com/nightlyone/lockfile)
|
|
||||||
|
|
||||||
install
|
|
||||||
-------------------
|
|
||||||
go get github.com/nightlyone/lockfile
|
|
||||||
|
|
||||||
|
|
||||||
contributing
|
|
||||||
============
|
|
||||||
|
|
||||||
Contributions are welcome. Please open an issue or send me a pull request for a dedicated branch.
|
|
||||||
Make sure the git commit hooks show it works.
|
|
||||||
|
|
||||||
git commit hooks
|
|
||||||
-----------------------
|
|
||||||
enable commit hooks via
|
|
||||||
|
|
||||||
cd .git ; rm -rf hooks; ln -s ../git-hooks hooks ; cd ..
|
|
||||||
|
|
12
vendor/github.com/nightlyone/lockfile/appveyor.yml
generated
vendored
12
vendor/github.com/nightlyone/lockfile/appveyor.yml
generated
vendored
|
@ -1,12 +0,0 @@
|
||||||
clone_folder: c:\gopath\src\github.com\nightlyone\lockfile
|
|
||||||
|
|
||||||
environment:
|
|
||||||
GOPATH: c:\gopath
|
|
||||||
|
|
||||||
install:
|
|
||||||
- go version
|
|
||||||
- go env
|
|
||||||
- go get -v -t ./...
|
|
||||||
|
|
||||||
build_script:
|
|
||||||
- go test -v ./...
|
|
162
vendor/github.com/nightlyone/lockfile/lockfile.go
generated
vendored
162
vendor/github.com/nightlyone/lockfile/lockfile.go
generated
vendored
|
@ -1,24 +1,46 @@
|
||||||
// Handle pid file based locking.
|
// Package lockfile handles pid file based locking.
|
||||||
|
// While a sync.Mutex helps against concurrency issues within a single process,
|
||||||
|
// this package is designed to help against concurrency issues between cooperating processes
|
||||||
|
// or serializing multiple invocations of the same process. You can also combine sync.Mutex
|
||||||
|
// with Lockfile in order to serialize an action between different goroutines in a single program
|
||||||
|
// and also multiple invocations of this program.
|
||||||
package lockfile
|
package lockfile
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"io"
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// Lockfile is a pid file which can be locked
|
||||||
type Lockfile string
|
type Lockfile string
|
||||||
|
|
||||||
|
// TemporaryError is a type of error where a retry after a random amount of sleep should help to mitigate it.
|
||||||
|
type TemporaryError string
|
||||||
|
|
||||||
|
func (t TemporaryError) Error() string { return string(t) }
|
||||||
|
|
||||||
|
// Temporary returns always true.
|
||||||
|
// It exists, so you can detect it via
|
||||||
|
// if te, ok := err.(interface{ Temporary() bool }); ok {
|
||||||
|
// fmt.Println("I am a temporay error situation, so wait and retry")
|
||||||
|
// }
|
||||||
|
func (t TemporaryError) Temporary() bool { return true }
|
||||||
|
|
||||||
|
// Various errors returned by this package
|
||||||
var (
|
var (
|
||||||
ErrBusy = errors.New("Locked by other process") // If you get this, retry after a short sleep might help
|
ErrBusy = TemporaryError("Locked by other process") // If you get this, retry after a short sleep might help
|
||||||
ErrNeedAbsPath = errors.New("Lockfiles must be given as absolute path names")
|
ErrNotExist = TemporaryError("Lockfile created, but doesn't exist") // If you get this, retry after a short sleep might help
|
||||||
ErrInvalidPid = errors.New("Lockfile contains invalid pid for system")
|
ErrNeedAbsPath = errors.New("Lockfiles must be given as absolute path names")
|
||||||
ErrDeadOwner = errors.New("Lockfile contains pid of process not existent on this system anymore")
|
ErrInvalidPid = errors.New("Lockfile contains invalid pid for system")
|
||||||
|
ErrDeadOwner = errors.New("Lockfile contains pid of process not existent on this system anymore")
|
||||||
|
ErrRogueDeletion = errors.New("Lockfile owned by me has been removed unexpectedly")
|
||||||
)
|
)
|
||||||
|
|
||||||
// Describe a new filename located at path. It is expected to be an absolute path
|
// New describes a new filename located at the given absolute path.
|
||||||
func New(path string) (Lockfile, error) {
|
func New(path string) (Lockfile, error) {
|
||||||
if !filepath.IsAbs(path) {
|
if !filepath.IsAbs(path) {
|
||||||
return Lockfile(""), ErrNeedAbsPath
|
return Lockfile(""), ErrNeedAbsPath
|
||||||
|
@ -26,7 +48,7 @@ func New(path string) (Lockfile, error) {
|
||||||
return Lockfile(path), nil
|
return Lockfile(path), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Who owns the lockfile?
|
// GetOwner returns who owns the lockfile.
|
||||||
func (l Lockfile) GetOwner() (*os.Process, error) {
|
func (l Lockfile) GetOwner() (*os.Process, error) {
|
||||||
name := string(l)
|
name := string(l)
|
||||||
|
|
||||||
|
@ -36,53 +58,68 @@ func (l Lockfile) GetOwner() (*os.Process, error) {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
var pid int
|
// try hard for pids. If no pid, the lockfile is junk anyway and we delete it.
|
||||||
_, err = fmt.Sscanln(string(content), &pid)
|
pid, err := scanPidLine(content)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, ErrInvalidPid
|
return nil, err
|
||||||
|
}
|
||||||
|
running, err := isRunning(pid)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// try hard for pids. If no pid, the lockfile is junk anyway and we delete it.
|
if running {
|
||||||
if pid > 0 {
|
proc, err := os.FindProcess(pid)
|
||||||
p, err := os.FindProcess(pid)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
return p, isProcessAlive(p)
|
return proc, nil
|
||||||
} else {
|
|
||||||
return nil, ErrInvalidPid
|
|
||||||
}
|
}
|
||||||
panic("Not reached")
|
return nil, ErrDeadOwner
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Try to get Lockfile lock. Returns nil, if successful and and error describing the reason, it didn't work out.
|
// TryLock tries to own the lock.
|
||||||
// Please note, that existing lockfiles containing pids of dead processes and lockfiles containing no pid at all
|
// It Returns nil, if successful and and error describing the reason, it didn't work out.
|
||||||
// are deleted.
|
// Please note, that existing lockfiles containing pids of dead processes
|
||||||
|
// and lockfiles containing no pid at all are simply deleted.
|
||||||
func (l Lockfile) TryLock() error {
|
func (l Lockfile) TryLock() error {
|
||||||
name := string(l)
|
name := string(l)
|
||||||
|
|
||||||
// This has been checked by New already. If we trigger here,
|
// This has been checked by New already. If we trigger here,
|
||||||
// the caller didn't use New and re-implemented it's functionality badly.
|
// the caller didn't use New and re-implemented it's functionality badly.
|
||||||
// So panic, that he might find this easily during testing.
|
// So panic, that he might find this easily during testing.
|
||||||
if !filepath.IsAbs(string(name)) {
|
if !filepath.IsAbs(name) {
|
||||||
panic(ErrNeedAbsPath)
|
panic(ErrNeedAbsPath)
|
||||||
}
|
}
|
||||||
|
|
||||||
tmplock, err := ioutil.TempFile(filepath.Dir(name), "")
|
tmplock, err := ioutil.TempFile(filepath.Dir(name), filepath.Base(name)+".")
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
} else {
|
|
||||||
defer tmplock.Close()
|
|
||||||
defer os.Remove(tmplock.Name())
|
|
||||||
}
|
|
||||||
|
|
||||||
_, err = tmplock.WriteString(fmt.Sprintf("%d\n", os.Getpid()))
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
// return value intentionally ignored, as ignoring it is part of the algorithm
|
cleanup := func() {
|
||||||
_ = os.Link(tmplock.Name(), name)
|
_ = tmplock.Close()
|
||||||
|
_ = os.Remove(tmplock.Name())
|
||||||
|
}
|
||||||
|
defer cleanup()
|
||||||
|
|
||||||
|
if err := writePidLine(tmplock, os.Getpid()); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
// EEXIST and similiar error codes, caught by os.IsExist, are intentionally ignored,
|
||||||
|
// as it means that someone was faster creating this link
|
||||||
|
// and ignoring this kind of error is part of the algorithm.
|
||||||
|
// The we will probably fail the pid owner check later, if this process is still alive.
|
||||||
|
// We cannot ignore ALL errors, since failure to support hard links, disk full
|
||||||
|
// as well as many other errors can happen to a filesystem operation
|
||||||
|
// and we really want to abort on those.
|
||||||
|
if err := os.Link(tmplock.Name(), name); err != nil {
|
||||||
|
if !os.IsExist(err) {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
fiTmp, err := os.Lstat(tmplock.Name())
|
fiTmp, err := os.Lstat(tmplock.Name())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -90,6 +127,10 @@ func (l Lockfile) TryLock() error {
|
||||||
}
|
}
|
||||||
fiLock, err := os.Lstat(name)
|
fiLock, err := os.Lstat(name)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
// tell user that a retry would be a good idea
|
||||||
|
if os.IsNotExist(err) {
|
||||||
|
return ErrNotExist
|
||||||
|
}
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -98,13 +139,15 @@ func (l Lockfile) TryLock() error {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
_, err = l.GetOwner()
|
proc, err := l.GetOwner()
|
||||||
switch err {
|
switch err {
|
||||||
default:
|
default:
|
||||||
// Other errors -> defensively fail and let caller handle this
|
// Other errors -> defensively fail and let caller handle this
|
||||||
return err
|
return err
|
||||||
case nil:
|
case nil:
|
||||||
return ErrBusy
|
if proc.Pid != os.Getpid() {
|
||||||
|
return ErrBusy
|
||||||
|
}
|
||||||
case ErrDeadOwner, ErrInvalidPid:
|
case ErrDeadOwner, ErrInvalidPid:
|
||||||
// cases we can fix below
|
// cases we can fix below
|
||||||
}
|
}
|
||||||
|
@ -112,14 +155,57 @@ func (l Lockfile) TryLock() error {
|
||||||
// clean stale/invalid lockfile
|
// clean stale/invalid lockfile
|
||||||
err = os.Remove(name)
|
err = os.Remove(name)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
// If it doesn't exist, then it doesn't matter who removed it.
|
||||||
|
if !os.IsNotExist(err) {
|
||||||
|
return err
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// now that we cleaned up the stale lockfile, let's recurse
|
// now that the stale lockfile is gone, let's recurse
|
||||||
return l.TryLock()
|
return l.TryLock()
|
||||||
}
|
}
|
||||||
|
|
||||||
// Release a lock again. Returns any error that happend during release of lock.
|
// Unlock a lock again, if we owned it. Returns any error that happend during release of lock.
|
||||||
func (l Lockfile) Unlock() error {
|
func (l Lockfile) Unlock() error {
|
||||||
return os.Remove(string(l))
|
proc, err := l.GetOwner()
|
||||||
|
switch err {
|
||||||
|
case ErrInvalidPid, ErrDeadOwner:
|
||||||
|
return ErrRogueDeletion
|
||||||
|
case nil:
|
||||||
|
if proc.Pid == os.Getpid() {
|
||||||
|
// we really own it, so let's remove it.
|
||||||
|
return os.Remove(string(l))
|
||||||
|
}
|
||||||
|
// Not owned by me, so don't delete it.
|
||||||
|
return ErrRogueDeletion
|
||||||
|
default:
|
||||||
|
// This is an application error or system error.
|
||||||
|
// So give a better error for logging here.
|
||||||
|
if os.IsNotExist(err) {
|
||||||
|
return ErrRogueDeletion
|
||||||
|
}
|
||||||
|
// Other errors -> defensively fail and let caller handle this
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func writePidLine(w io.Writer, pid int) error {
|
||||||
|
_, err := io.WriteString(w, fmt.Sprintf("%d\n", pid))
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
func scanPidLine(content []byte) (int, error) {
|
||||||
|
if len(content) == 0 {
|
||||||
|
return 0, ErrInvalidPid
|
||||||
|
}
|
||||||
|
|
||||||
|
var pid int
|
||||||
|
if _, err := fmt.Sscanln(string(content), &pid); err != nil {
|
||||||
|
return 0, ErrInvalidPid
|
||||||
|
}
|
||||||
|
|
||||||
|
if pid <= 0 {
|
||||||
|
return 0, ErrInvalidPid
|
||||||
|
}
|
||||||
|
return pid, nil
|
||||||
}
|
}
|
||||||
|
|
24
vendor/github.com/nightlyone/lockfile/lockfile_unix.go
generated
vendored
24
vendor/github.com/nightlyone/lockfile/lockfile_unix.go
generated
vendored
|
@ -1,4 +1,4 @@
|
||||||
// +build darwin dragonfly freebsd linux nacl netbsd openbsd solaris
|
// +build darwin dragonfly freebsd linux nacl netbsd openbsd solaris aix
|
||||||
|
|
||||||
package lockfile
|
package lockfile
|
||||||
|
|
||||||
|
@ -7,22 +7,14 @@ import (
|
||||||
"syscall"
|
"syscall"
|
||||||
)
|
)
|
||||||
|
|
||||||
func isProcessAlive(p *os.Process) error {
|
func isRunning(pid int) (bool, error) {
|
||||||
err := p.Signal(os.Signal(syscall.Signal(0)))
|
proc, err := os.FindProcess(pid)
|
||||||
if err == nil {
|
if err != nil {
|
||||||
return nil
|
return false, err
|
||||||
}
|
|
||||||
errno, ok := err.(syscall.Errno)
|
|
||||||
if !ok {
|
|
||||||
return ErrDeadOwner
|
|
||||||
}
|
}
|
||||||
|
|
||||||
switch errno {
|
if err := proc.Signal(syscall.Signal(0)); err != nil {
|
||||||
case syscall.ESRCH:
|
return false, nil
|
||||||
return ErrDeadOwner
|
|
||||||
case syscall.EPERM:
|
|
||||||
return nil
|
|
||||||
default:
|
|
||||||
return err
|
|
||||||
}
|
}
|
||||||
|
return true, nil
|
||||||
}
|
}
|
||||||
|
|
36
vendor/github.com/nightlyone/lockfile/lockfile_windows.go
generated
vendored
36
vendor/github.com/nightlyone/lockfile/lockfile_windows.go
generated
vendored
|
@ -1,32 +1,30 @@
|
||||||
package lockfile
|
package lockfile
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"os"
|
|
||||||
"reflect"
|
|
||||||
"syscall"
|
"syscall"
|
||||||
)
|
)
|
||||||
|
|
||||||
func isProcessAlive(p *os.Process) error {
|
//For some reason these consts don't exist in syscall.
|
||||||
// Extract handle value from the os.Process struct to avoid the need
|
const (
|
||||||
// of a second, manually opened process handle.
|
error_invalid_parameter = 87
|
||||||
value := reflect.ValueOf(p)
|
code_still_active = 259
|
||||||
// Dereference *os.Process to os.Process
|
)
|
||||||
value = value.Elem()
|
|
||||||
field := value.FieldByName("handle")
|
|
||||||
|
|
||||||
handle := syscall.Handle(field.Uint())
|
func isRunning(pid int) (bool, error) {
|
||||||
|
procHnd, err := syscall.OpenProcess(syscall.PROCESS_QUERY_INFORMATION, true, uint32(pid))
|
||||||
|
if err != nil {
|
||||||
|
if scerr, ok := err.(syscall.Errno); ok {
|
||||||
|
if uintptr(scerr) == error_invalid_parameter {
|
||||||
|
return false, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
var code uint32
|
var code uint32
|
||||||
err := syscall.GetExitCodeProcess(handle, &code)
|
err = syscall.GetExitCodeProcess(procHnd, &code)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return false, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// code will contain the exit code of the process or 259 (STILL_ALIVE)
|
return code == code_still_active, nil
|
||||||
// if the process has not exited yet.
|
|
||||||
if code == 259 {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
return ErrDeadOwner
|
|
||||||
}
|
}
|
||||||
|
|
4
vendor/github.com/onsi/ginkgo/.gitignore
generated
vendored
4
vendor/github.com/onsi/ginkgo/.gitignore
generated
vendored
|
@ -1,4 +0,0 @@
|
||||||
.DS_Store
|
|
||||||
TODO
|
|
||||||
tmp/**/*
|
|
||||||
*.coverprofile
|
|
15
vendor/github.com/onsi/ginkgo/.travis.yml
generated
vendored
15
vendor/github.com/onsi/ginkgo/.travis.yml
generated
vendored
|
@ -1,15 +0,0 @@
|
||||||
language: go
|
|
||||||
go:
|
|
||||||
- 1.3
|
|
||||||
- 1.4
|
|
||||||
- 1.5
|
|
||||||
- tip
|
|
||||||
|
|
||||||
install:
|
|
||||||
- go get -v ./...
|
|
||||||
- go get golang.org/x/tools/cmd/cover
|
|
||||||
- go get github.com/onsi/gomega
|
|
||||||
- go install github.com/onsi/ginkgo/ginkgo
|
|
||||||
- export PATH=$PATH:$HOME/gopath/bin
|
|
||||||
|
|
||||||
script: $HOME/gopath/bin/ginkgo -r --randomizeAllSpecs --randomizeSuites --race --trace
|
|
136
vendor/github.com/onsi/ginkgo/CHANGELOG.md
generated
vendored
136
vendor/github.com/onsi/ginkgo/CHANGELOG.md
generated
vendored
|
@ -1,136 +0,0 @@
|
||||||
## HEAD
|
|
||||||
|
|
||||||
Improvements:
|
|
||||||
|
|
||||||
- `Skip(message)` can be used to skip the current test.
|
|
||||||
- Added `extensions/table` - a Ginkgo DSL for [Table Driven Tests](http://onsi.github.io/ginkgo/#table-driven-tests)
|
|
||||||
|
|
||||||
Bug Fixes:
|
|
||||||
|
|
||||||
- Ginkgo tests now fail when you `panic(nil)` (#167)
|
|
||||||
|
|
||||||
## 1.2.0 5/31/2015
|
|
||||||
|
|
||||||
Improvements
|
|
||||||
|
|
||||||
- `ginkgo -coverpkg` calls down to `go test -coverpkg` (#160)
|
|
||||||
- `ginkgo -afterSuiteHook COMMAND` invokes the passed-in `COMMAND` after a test suite completes (#152)
|
|
||||||
- Relaxed requirement for Go 1.4+. `ginkgo` now works with Go v1.3+ (#166)
|
|
||||||
|
|
||||||
## 1.2.0-beta
|
|
||||||
|
|
||||||
Ginkgo now requires Go 1.4+
|
|
||||||
|
|
||||||
Improvements:
|
|
||||||
|
|
||||||
- Call reporters in reverse order when announcing spec completion -- allows custom reporters to emit output before the default reporter does.
|
|
||||||
- Improved focus behavior. Now, this:
|
|
||||||
|
|
||||||
```golang
|
|
||||||
FDescribe("Some describe", func() {
|
|
||||||
It("A", func() {})
|
|
||||||
|
|
||||||
FIt("B", func() {})
|
|
||||||
})
|
|
||||||
```
|
|
||||||
|
|
||||||
will run `B` but *not* `A`. This tends to be a common usage pattern when in the thick of writing and debugging tests.
|
|
||||||
- When `SIGINT` is received, Ginkgo will emit the contents of the `GinkgoWriter` before running the `AfterSuite`. Useful for debugging stuck tests.
|
|
||||||
- When `--progress` is set, Ginkgo will write test progress (in particular, Ginkgo will say when it is about to run a BeforeEach, AfterEach, It, etc...) to the `GinkgoWriter`. This is useful for debugging stuck tests and tests that generate many logs.
|
|
||||||
- Improved output when an error occurs in a setup or teardown block.
|
|
||||||
- When `--dryRun` is set, Ginkgo will walk the spec tree and emit to its reporter *without* actually running anything. Best paired with `-v` to understand which specs will run in which order.
|
|
||||||
- Add `By` to help document long `It`s. `By` simply writes to the `GinkgoWriter`.
|
|
||||||
- Add support for precompiled tests:
|
|
||||||
- `ginkgo build <path-to-package>` will now compile the package, producing a file named `package.test`
|
|
||||||
- The compiled `package.test` file can be run directly. This runs the tests in series.
|
|
||||||
- To run precompiled tests in parallel, you can run: `ginkgo -p package.test`
|
|
||||||
- Support `bootstrap`ping and `generate`ing [Agouti](http://agouti.org) specs.
|
|
||||||
- `ginkgo generate` and `ginkgo bootstrap` now honor the package name already defined in a given directory
|
|
||||||
- The `ginkgo` CLI ignores `SIGQUIT`. Prevents its stack dump from interlacing with the underlying test suite's stack dump.
|
|
||||||
- The `ginkgo` CLI now compiles tests into a temporary directory instead of the package directory. This necessitates upgrading to Go v1.4+.
|
|
||||||
- `ginkgo -notify` now works on Linux
|
|
||||||
|
|
||||||
Bug Fixes:
|
|
||||||
|
|
||||||
- If --skipPackages is used and all packages are skipped, Ginkgo should exit 0.
|
|
||||||
- Fix tempfile leak when running in parallel
|
|
||||||
- Fix incorrect failure message when a panic occurs during a parallel test run
|
|
||||||
- Fixed an issue where a pending test within a focused context (or a focused test within a pending context) would skip all other tests.
|
|
||||||
- Be more consistent about handling SIGTERM as well as SIGINT
|
|
||||||
- When interupted while concurrently compiling test suites in the background, Ginkgo now cleans up the compiled artifacts.
|
|
||||||
- Fixed a long standing bug where `ginkgo -p` would hang if a process spawned by one of the Ginkgo parallel nodes does not exit. (Hooray!)
|
|
||||||
|
|
||||||
## 1.1.0 (8/2/2014)
|
|
||||||
|
|
||||||
No changes, just dropping the beta.
|
|
||||||
|
|
||||||
## 1.1.0-beta (7/22/2014)
|
|
||||||
New Features:
|
|
||||||
|
|
||||||
- `ginkgo watch` now monitors packages *and their dependencies* for changes. The depth of the dependency tree can be modified with the `-depth` flag.
|
|
||||||
- Test suites with a programmatic focus (`FIt`, `FDescribe`, etc...) exit with non-zero status code, evne when they pass. This allows CI systems to detect accidental commits of focused test suites.
|
|
||||||
- `ginkgo -p` runs the testsuite in parallel with an auto-detected number of nodes.
|
|
||||||
- `ginkgo -tags=TAG_LIST` passes a list of tags down to the `go build` command.
|
|
||||||
- `ginkgo --failFast` aborts the test suite after the first failure.
|
|
||||||
- `ginkgo generate file_1 file_2` can take multiple file arguments.
|
|
||||||
- Ginkgo now summarizes any spec failures that occured at the end of the test run.
|
|
||||||
- `ginkgo --randomizeSuites` will run tests *suites* in random order using the generated/passed-in seed.
|
|
||||||
|
|
||||||
Improvements:
|
|
||||||
|
|
||||||
- `ginkgo -skipPackage` now takes a comma-separated list of strings. If the *relative path* to a package matches one of the entries in the comma-separated list, that package is skipped.
|
|
||||||
- `ginkgo --untilItFails` no longer recompiles between attempts.
|
|
||||||
- Ginkgo now panics when a runnable node (`It`, `BeforeEach`, `JustBeforeEach`, `AfterEach`, `Measure`) is nested within another runnable node. This is always a mistake. Any test suites that panic because of this change should be fixed.
|
|
||||||
|
|
||||||
Bug Fixes:
|
|
||||||
|
|
||||||
- `ginkgo boostrap` and `ginkgo generate` no longer fail when dealing with `hyphen-separated-packages`.
|
|
||||||
- parallel specs are now better distributed across nodes - fixed a crashing bug where (for example) distributing 11 tests across 7 nodes would panic
|
|
||||||
|
|
||||||
## 1.0.0 (5/24/2014)
|
|
||||||
New Features:
|
|
||||||
|
|
||||||
- Add `GinkgoParallelNode()` - shorthand for `config.GinkgoConfig.ParallelNode`
|
|
||||||
|
|
||||||
Improvements:
|
|
||||||
|
|
||||||
- When compilation fails, the compilation output is rewritten to present a correct *relative* path. Allows ⌘-clicking in iTerm open the file in your text editor.
|
|
||||||
- `--untilItFails` and `ginkgo watch` now generate new random seeds between test runs, unless a particular random seed is specified.
|
|
||||||
|
|
||||||
Bug Fixes:
|
|
||||||
|
|
||||||
- `-cover` now generates a correctly combined coverprofile when running with in parallel with multiple `-node`s.
|
|
||||||
- Print out the contents of the `GinkgoWriter` when `BeforeSuite` or `AfterSuite` fail.
|
|
||||||
- Fix all remaining race conditions in Ginkgo's test suite.
|
|
||||||
|
|
||||||
## 1.0.0-beta (4/14/2014)
|
|
||||||
Breaking changes:
|
|
||||||
|
|
||||||
- `thirdparty/gomocktestreporter` is gone. Use `GinkgoT()` instead
|
|
||||||
- Modified the Reporter interface
|
|
||||||
- `watch` is now a subcommand, not a flag.
|
|
||||||
|
|
||||||
DSL changes:
|
|
||||||
|
|
||||||
- `BeforeSuite` and `AfterSuite` for setting up and tearing down test suites.
|
|
||||||
- `AfterSuite` is triggered on interrupt (`^C`) as well as exit.
|
|
||||||
- `SynchronizedBeforeSuite` and `SynchronizedAfterSuite` for setting up and tearing down singleton resources across parallel nodes.
|
|
||||||
|
|
||||||
CLI changes:
|
|
||||||
|
|
||||||
- `watch` is now a subcommand, not a flag
|
|
||||||
- `--nodot` flag can be passed to `ginkgo generate` and `ginkgo bootstrap` to avoid dot imports. This explicitly imports all exported identifiers in Ginkgo and Gomega. Refreshing this list can be done by running `ginkgo nodot`
|
|
||||||
- Additional arguments can be passed to specs. Pass them after the `--` separator
|
|
||||||
- `--skipPackage` flag takes a regexp and ignores any packages with package names passing said regexp.
|
|
||||||
- `--trace` flag prints out full stack traces when errors occur, not just the line at which the error occurs.
|
|
||||||
|
|
||||||
Misc:
|
|
||||||
|
|
||||||
- Start using semantic versioning
|
|
||||||
- Start maintaining changelog
|
|
||||||
|
|
||||||
Major refactor:
|
|
||||||
|
|
||||||
- Pull out Ginkgo's internal to `internal`
|
|
||||||
- Rename `example` everywhere to `spec`
|
|
||||||
- Much more!
|
|
115
vendor/github.com/onsi/ginkgo/README.md
generated
vendored
115
vendor/github.com/onsi/ginkgo/README.md
generated
vendored
|
@ -1,115 +0,0 @@
|
||||||
![Ginkgo: A Golang BDD Testing Framework](http://onsi.github.io/ginkgo/images/ginkgo.png)
|
|
||||||
|
|
||||||
[![Build Status](https://travis-ci.org/onsi/ginkgo.png)](https://travis-ci.org/onsi/ginkgo)
|
|
||||||
|
|
||||||
Jump to the [docs](http://onsi.github.io/ginkgo/) to learn more. To start rolling your Ginkgo tests *now* [keep reading](#set-me-up)!
|
|
||||||
|
|
||||||
To discuss Ginkgo and get updates, join the [google group](https://groups.google.com/d/forum/ginkgo-and-gomega).
|
|
||||||
|
|
||||||
## Feature List
|
|
||||||
|
|
||||||
- Ginkgo uses Go's `testing` package and can live alongside your existing `testing` tests. It's easy to [bootstrap](http://onsi.github.io/ginkgo/#bootstrapping-a-suite) and start writing your [first tests](http://onsi.github.io/ginkgo/#adding-specs-to-a-suite)
|
|
||||||
|
|
||||||
- Structure your BDD-style tests expressively:
|
|
||||||
- Nestable [`Describe` and `Context` container blocks](http://onsi.github.io/ginkgo/#organizing-specs-with-containers-describe-and-context)
|
|
||||||
- [`BeforeEach` and `AfterEach` blocks](http://onsi.github.io/ginkgo/#extracting-common-setup-beforeeach) for setup and teardown
|
|
||||||
- [`It` blocks](http://onsi.github.io/ginkgo/#individual-specs-) that hold your assertions
|
|
||||||
- [`JustBeforeEach` blocks](http://onsi.github.io/ginkgo/#separating-creation-and-configuration-justbeforeeach) that separate creation from configuration (also known as the subject action pattern).
|
|
||||||
- [`BeforeSuite` and `AfterSuite` blocks](http://onsi.github.io/ginkgo/#global-setup-and-teardown-beforesuite-and-aftersuite) to prep for and cleanup after a suite.
|
|
||||||
|
|
||||||
- A comprehensive test runner that lets you:
|
|
||||||
- Mark specs as [pending](http://onsi.github.io/ginkgo/#pending-specs)
|
|
||||||
- [Focus](http://onsi.github.io/ginkgo/#focused-specs) individual specs, and groups of specs, either programmatically or on the command line
|
|
||||||
- Run your tests in [random order](http://onsi.github.io/ginkgo/#spec-permutation), and then reuse random seeds to replicate the same order.
|
|
||||||
- Break up your test suite into parallel processes for straightforward [test parallelization](http://onsi.github.io/ginkgo/#parallel-specs)
|
|
||||||
|
|
||||||
- `ginkgo`: a command line interface with plenty of handy command line arguments for [running your tests](http://onsi.github.io/ginkgo/#running-tests) and [generating](http://onsi.github.io/ginkgo/#generators) test files. Here are a few choice examples:
|
|
||||||
- `ginkgo -nodes=N` runs your tests in `N` parallel processes and print out coherent output in realtime
|
|
||||||
- `ginkgo -cover` runs your tests using Golang's code coverage tool
|
|
||||||
- `ginkgo convert` converts an XUnit-style `testing` package to a Ginkgo-style package
|
|
||||||
- `ginkgo -focus="REGEXP"` and `ginkgo -skip="REGEXP"` allow you to specify a subset of tests to run via regular expression
|
|
||||||
- `ginkgo -r` runs all tests suites under the current directory
|
|
||||||
- `ginkgo -v` prints out identifying information for each tests just before it runs
|
|
||||||
|
|
||||||
And much more: run `ginkgo help` for details!
|
|
||||||
|
|
||||||
The `ginkgo` CLI is convenient, but purely optional -- Ginkgo works just fine with `go test`
|
|
||||||
|
|
||||||
- `ginkgo watch` [watches](https://onsi.github.io/ginkgo/#watching-for-changes) packages *and their dependencies* for changes, then reruns tests. Run tests immediately as you develop!
|
|
||||||
|
|
||||||
- Built-in support for testing [asynchronicity](http://onsi.github.io/ginkgo/#asynchronous-tests)
|
|
||||||
|
|
||||||
- Built-in support for [benchmarking](http://onsi.github.io/ginkgo/#benchmark-tests) your code. Control the number of benchmark samples as you gather runtimes and other, arbitrary, bits of numerical information about your code.
|
|
||||||
|
|
||||||
- [Completions for Sublime Text](https://github.com/onsi/ginkgo-sublime-completions): just use [Package Control](https://sublime.wbond.net/) to install `Ginkgo Completions`.
|
|
||||||
|
|
||||||
- Straightforward support for third-party testing libraries such as [Gomock](https://code.google.com/p/gomock/) and [Testify](https://github.com/stretchr/testify). Check out the [docs](http://onsi.github.io/ginkgo/#third-party-integrations) for details.
|
|
||||||
|
|
||||||
- A modular architecture that lets you easily:
|
|
||||||
- Write [custom reporters](http://onsi.github.io/ginkgo/#writing-custom-reporters) (for example, Ginkgo comes with a [JUnit XML reporter](http://onsi.github.io/ginkgo/#generating-junit-xml-output) and a TeamCity reporter).
|
|
||||||
- [Adapt an existing matcher library (or write your own!)](http://onsi.github.io/ginkgo/#using-other-matcher-libraries) to work with Ginkgo
|
|
||||||
|
|
||||||
## [Gomega](http://github.com/onsi/gomega): Ginkgo's Preferred Matcher Library
|
|
||||||
|
|
||||||
Ginkgo is best paired with Gomega. Learn more about Gomega [here](http://onsi.github.io/gomega/)
|
|
||||||
|
|
||||||
## [Agouti](http://github.com/sclevine/agouti): A Golang Acceptance Testing Framework
|
|
||||||
|
|
||||||
Agouti allows you run WebDriver integration tests. Learn more about Agouti [here](http://agouti.org)
|
|
||||||
|
|
||||||
## Set Me Up!
|
|
||||||
|
|
||||||
You'll need Golang v1.3+ (Ubuntu users: you probably have Golang v1.0 -- you'll need to upgrade!)
|
|
||||||
|
|
||||||
```bash
|
|
||||||
|
|
||||||
go get github.com/onsi/ginkgo/ginkgo # installs the ginkgo CLI
|
|
||||||
go get github.com/onsi/gomega # fetches the matcher library
|
|
||||||
|
|
||||||
cd path/to/package/you/want/to/test
|
|
||||||
|
|
||||||
ginkgo bootstrap # set up a new ginkgo suite
|
|
||||||
ginkgo generate # will create a sample test file. edit this file and add your tests then...
|
|
||||||
|
|
||||||
go test # to run your tests
|
|
||||||
|
|
||||||
ginkgo # also runs your tests
|
|
||||||
|
|
||||||
```
|
|
||||||
|
|
||||||
## I'm new to Go: What are my testing options?
|
|
||||||
|
|
||||||
Of course, I heartily recommend [Ginkgo](https://github.com/onsi/ginkgo) and [Gomega](https://github.com/onsi/gomega). Both packages are seeing heavy, daily, production use on a number of projects and boast a mature and comprehensive feature-set.
|
|
||||||
|
|
||||||
With that said, it's great to know what your options are :)
|
|
||||||
|
|
||||||
### What Golang gives you out of the box
|
|
||||||
|
|
||||||
Testing is a first class citizen in Golang, however Go's built-in testing primitives are somewhat limited: The [testing](http://golang.org/pkg/testing) package provides basic XUnit style tests and no assertion library.
|
|
||||||
|
|
||||||
### Matcher libraries for Golang's XUnit style tests
|
|
||||||
|
|
||||||
A number of matcher libraries have been written to augment Go's built-in XUnit style tests. Here are two that have gained traction:
|
|
||||||
|
|
||||||
- [testify](https://github.com/stretchr/testify)
|
|
||||||
- [gocheck](http://labix.org/gocheck)
|
|
||||||
|
|
||||||
You can also use Ginkgo's matcher library [Gomega](https://github.com/onsi/gomega) in [XUnit style tests](http://onsi.github.io/gomega/#using-gomega-with-golangs-xunitstyle-tests)
|
|
||||||
|
|
||||||
### BDD style testing frameworks
|
|
||||||
|
|
||||||
There are a handful of BDD-style testing frameworks written for Golang. Here are a few:
|
|
||||||
|
|
||||||
- [Ginkgo](https://github.com/onsi/ginkgo) ;)
|
|
||||||
- [GoConvey](https://github.com/smartystreets/goconvey)
|
|
||||||
- [Goblin](https://github.com/franela/goblin)
|
|
||||||
- [Mao](https://github.com/azer/mao)
|
|
||||||
- [Zen](https://github.com/pranavraja/zen)
|
|
||||||
|
|
||||||
Finally, @shageman has [put together](https://github.com/shageman/gotestit) a comprehensive comparison of golang testing libraries.
|
|
||||||
|
|
||||||
Go explore!
|
|
||||||
|
|
||||||
## License
|
|
||||||
|
|
||||||
Ginkgo is MIT-Licensed
|
|
54
vendor/github.com/onsi/ginkgo/config/config.go
generated
vendored
54
vendor/github.com/onsi/ginkgo/config/config.go
generated
vendored
|
@ -20,18 +20,21 @@ import (
|
||||||
"fmt"
|
"fmt"
|
||||||
)
|
)
|
||||||
|
|
||||||
const VERSION = "1.2.0"
|
const VERSION = "1.6.0"
|
||||||
|
|
||||||
type GinkgoConfigType struct {
|
type GinkgoConfigType struct {
|
||||||
RandomSeed int64
|
RandomSeed int64
|
||||||
RandomizeAllSpecs bool
|
RandomizeAllSpecs bool
|
||||||
FocusString string
|
RegexScansFilePath bool
|
||||||
SkipString string
|
FocusString string
|
||||||
SkipMeasurements bool
|
SkipString string
|
||||||
FailOnPending bool
|
SkipMeasurements bool
|
||||||
FailFast bool
|
FailOnPending bool
|
||||||
EmitSpecProgress bool
|
FailFast bool
|
||||||
DryRun bool
|
FlakeAttempts int
|
||||||
|
EmitSpecProgress bool
|
||||||
|
DryRun bool
|
||||||
|
DebugParallel bool
|
||||||
|
|
||||||
ParallelNode int
|
ParallelNode int
|
||||||
ParallelTotal int
|
ParallelTotal int
|
||||||
|
@ -45,6 +48,7 @@ type DefaultReporterConfigType struct {
|
||||||
NoColor bool
|
NoColor bool
|
||||||
SlowSpecThreshold float64
|
SlowSpecThreshold float64
|
||||||
NoisyPendings bool
|
NoisyPendings bool
|
||||||
|
NoisySkippings bool
|
||||||
Succinct bool
|
Succinct bool
|
||||||
Verbose bool
|
Verbose bool
|
||||||
FullTrace bool
|
FullTrace bool
|
||||||
|
@ -62,15 +66,24 @@ func processPrefix(prefix string) string {
|
||||||
func Flags(flagSet *flag.FlagSet, prefix string, includeParallelFlags bool) {
|
func Flags(flagSet *flag.FlagSet, prefix string, includeParallelFlags bool) {
|
||||||
prefix = processPrefix(prefix)
|
prefix = processPrefix(prefix)
|
||||||
flagSet.Int64Var(&(GinkgoConfig.RandomSeed), prefix+"seed", time.Now().Unix(), "The seed used to randomize the spec suite.")
|
flagSet.Int64Var(&(GinkgoConfig.RandomSeed), prefix+"seed", time.Now().Unix(), "The seed used to randomize the spec suite.")
|
||||||
flagSet.BoolVar(&(GinkgoConfig.RandomizeAllSpecs), prefix+"randomizeAllSpecs", false, "If set, ginkgo will randomize all specs together. By default, ginkgo only randomizes the top level Describe/Context groups.")
|
flagSet.BoolVar(&(GinkgoConfig.RandomizeAllSpecs), prefix+"randomizeAllSpecs", false, "If set, ginkgo will randomize all specs together. By default, ginkgo only randomizes the top level Describe, Context and When groups.")
|
||||||
flagSet.BoolVar(&(GinkgoConfig.SkipMeasurements), prefix+"skipMeasurements", false, "If set, ginkgo will skip any measurement specs.")
|
flagSet.BoolVar(&(GinkgoConfig.SkipMeasurements), prefix+"skipMeasurements", false, "If set, ginkgo will skip any measurement specs.")
|
||||||
flagSet.BoolVar(&(GinkgoConfig.FailOnPending), prefix+"failOnPending", false, "If set, ginkgo will mark the test suite as failed if any specs are pending.")
|
flagSet.BoolVar(&(GinkgoConfig.FailOnPending), prefix+"failOnPending", false, "If set, ginkgo will mark the test suite as failed if any specs are pending.")
|
||||||
flagSet.BoolVar(&(GinkgoConfig.FailFast), prefix+"failFast", false, "If set, ginkgo will stop running a test suite after a failure occurs.")
|
flagSet.BoolVar(&(GinkgoConfig.FailFast), prefix+"failFast", false, "If set, ginkgo will stop running a test suite after a failure occurs.")
|
||||||
|
|
||||||
flagSet.BoolVar(&(GinkgoConfig.DryRun), prefix+"dryRun", false, "If set, ginkgo will walk the test hierarchy without actually running anything. Best paired with -v.")
|
flagSet.BoolVar(&(GinkgoConfig.DryRun), prefix+"dryRun", false, "If set, ginkgo will walk the test hierarchy without actually running anything. Best paired with -v.")
|
||||||
|
|
||||||
flagSet.StringVar(&(GinkgoConfig.FocusString), prefix+"focus", "", "If set, ginkgo will only run specs that match this regular expression.")
|
flagSet.StringVar(&(GinkgoConfig.FocusString), prefix+"focus", "", "If set, ginkgo will only run specs that match this regular expression.")
|
||||||
flagSet.StringVar(&(GinkgoConfig.SkipString), prefix+"skip", "", "If set, ginkgo will only run specs that do not match this regular expression.")
|
flagSet.StringVar(&(GinkgoConfig.SkipString), prefix+"skip", "", "If set, ginkgo will only run specs that do not match this regular expression.")
|
||||||
|
|
||||||
|
flagSet.BoolVar(&(GinkgoConfig.RegexScansFilePath), prefix+"regexScansFilePath", false, "If set, ginkgo regex matching also will look at the file path (code location).")
|
||||||
|
|
||||||
|
flagSet.IntVar(&(GinkgoConfig.FlakeAttempts), prefix+"flakeAttempts", 1, "Make up to this many attempts to run each spec. Please note that if any of the attempts succeed, the suite will not be failed. But any failures will still be recorded.")
|
||||||
|
|
||||||
flagSet.BoolVar(&(GinkgoConfig.EmitSpecProgress), prefix+"progress", false, "If set, ginkgo will emit progress information as each spec runs to the GinkgoWriter.")
|
flagSet.BoolVar(&(GinkgoConfig.EmitSpecProgress), prefix+"progress", false, "If set, ginkgo will emit progress information as each spec runs to the GinkgoWriter.")
|
||||||
|
|
||||||
|
flagSet.BoolVar(&(GinkgoConfig.DebugParallel), prefix+"debug", false, "If set, ginkgo will emit node output to files when running in parallel.")
|
||||||
|
|
||||||
if includeParallelFlags {
|
if includeParallelFlags {
|
||||||
flagSet.IntVar(&(GinkgoConfig.ParallelNode), prefix+"parallel.node", 1, "This worker node's (one-indexed) node number. For running specs in parallel.")
|
flagSet.IntVar(&(GinkgoConfig.ParallelNode), prefix+"parallel.node", 1, "This worker node's (one-indexed) node number. For running specs in parallel.")
|
||||||
flagSet.IntVar(&(GinkgoConfig.ParallelTotal), prefix+"parallel.total", 1, "The total number of worker nodes. For running specs in parallel.")
|
flagSet.IntVar(&(GinkgoConfig.ParallelTotal), prefix+"parallel.total", 1, "The total number of worker nodes. For running specs in parallel.")
|
||||||
|
@ -79,8 +92,9 @@ func Flags(flagSet *flag.FlagSet, prefix string, includeParallelFlags bool) {
|
||||||
}
|
}
|
||||||
|
|
||||||
flagSet.BoolVar(&(DefaultReporterConfig.NoColor), prefix+"noColor", false, "If set, suppress color output in default reporter.")
|
flagSet.BoolVar(&(DefaultReporterConfig.NoColor), prefix+"noColor", false, "If set, suppress color output in default reporter.")
|
||||||
flagSet.Float64Var(&(DefaultReporterConfig.SlowSpecThreshold), prefix+"slowSpecThreshold", 5.0, "(in seconds) Specs that take longer to run than this threshold are flagged as slow by the default reporter (default: 5 seconds).")
|
flagSet.Float64Var(&(DefaultReporterConfig.SlowSpecThreshold), prefix+"slowSpecThreshold", 5.0, "(in seconds) Specs that take longer to run than this threshold are flagged as slow by the default reporter.")
|
||||||
flagSet.BoolVar(&(DefaultReporterConfig.NoisyPendings), prefix+"noisyPendings", true, "If set, default reporter will shout about pending tests.")
|
flagSet.BoolVar(&(DefaultReporterConfig.NoisyPendings), prefix+"noisyPendings", true, "If set, default reporter will shout about pending tests.")
|
||||||
|
flagSet.BoolVar(&(DefaultReporterConfig.NoisySkippings), prefix+"noisySkippings", true, "If set, default reporter will shout about skipping tests.")
|
||||||
flagSet.BoolVar(&(DefaultReporterConfig.Verbose), prefix+"v", false, "If set, default reporter print out all specs as they begin.")
|
flagSet.BoolVar(&(DefaultReporterConfig.Verbose), prefix+"v", false, "If set, default reporter print out all specs as they begin.")
|
||||||
flagSet.BoolVar(&(DefaultReporterConfig.Succinct), prefix+"succinct", false, "If set, default reporter prints out a very succinct report")
|
flagSet.BoolVar(&(DefaultReporterConfig.Succinct), prefix+"succinct", false, "If set, default reporter prints out a very succinct report")
|
||||||
flagSet.BoolVar(&(DefaultReporterConfig.FullTrace), prefix+"trace", false, "If set, default reporter prints out the full stack trace when a failure occurs")
|
flagSet.BoolVar(&(DefaultReporterConfig.FullTrace), prefix+"trace", false, "If set, default reporter prints out the full stack trace when a failure occurs")
|
||||||
|
@ -122,10 +136,18 @@ func BuildFlagArgs(prefix string, ginkgo GinkgoConfigType, reporter DefaultRepor
|
||||||
result = append(result, fmt.Sprintf("--%sskip=%s", prefix, ginkgo.SkipString))
|
result = append(result, fmt.Sprintf("--%sskip=%s", prefix, ginkgo.SkipString))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if ginkgo.FlakeAttempts > 1 {
|
||||||
|
result = append(result, fmt.Sprintf("--%sflakeAttempts=%d", prefix, ginkgo.FlakeAttempts))
|
||||||
|
}
|
||||||
|
|
||||||
if ginkgo.EmitSpecProgress {
|
if ginkgo.EmitSpecProgress {
|
||||||
result = append(result, fmt.Sprintf("--%sprogress", prefix))
|
result = append(result, fmt.Sprintf("--%sprogress", prefix))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if ginkgo.DebugParallel {
|
||||||
|
result = append(result, fmt.Sprintf("--%sdebug", prefix))
|
||||||
|
}
|
||||||
|
|
||||||
if ginkgo.ParallelNode != 0 {
|
if ginkgo.ParallelNode != 0 {
|
||||||
result = append(result, fmt.Sprintf("--%sparallel.node=%d", prefix, ginkgo.ParallelNode))
|
result = append(result, fmt.Sprintf("--%sparallel.node=%d", prefix, ginkgo.ParallelNode))
|
||||||
}
|
}
|
||||||
|
@ -142,6 +164,10 @@ func BuildFlagArgs(prefix string, ginkgo GinkgoConfigType, reporter DefaultRepor
|
||||||
result = append(result, fmt.Sprintf("--%sparallel.synchost=%s", prefix, ginkgo.SyncHost))
|
result = append(result, fmt.Sprintf("--%sparallel.synchost=%s", prefix, ginkgo.SyncHost))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if ginkgo.RegexScansFilePath {
|
||||||
|
result = append(result, fmt.Sprintf("--%sregexScansFilePath", prefix))
|
||||||
|
}
|
||||||
|
|
||||||
if reporter.NoColor {
|
if reporter.NoColor {
|
||||||
result = append(result, fmt.Sprintf("--%snoColor", prefix))
|
result = append(result, fmt.Sprintf("--%snoColor", prefix))
|
||||||
}
|
}
|
||||||
|
@ -154,6 +180,10 @@ func BuildFlagArgs(prefix string, ginkgo GinkgoConfigType, reporter DefaultRepor
|
||||||
result = append(result, fmt.Sprintf("--%snoisyPendings=false", prefix))
|
result = append(result, fmt.Sprintf("--%snoisyPendings=false", prefix))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if !reporter.NoisySkippings {
|
||||||
|
result = append(result, fmt.Sprintf("--%snoisySkippings=false", prefix))
|
||||||
|
}
|
||||||
|
|
||||||
if reporter.Verbose {
|
if reporter.Verbose {
|
||||||
result = append(result, fmt.Sprintf("--%sv", prefix))
|
result = append(result, fmt.Sprintf("--%sv", prefix))
|
||||||
}
|
}
|
||||||
|
|
93
vendor/github.com/onsi/ginkgo/ginkgo_dsl.go
generated
vendored
93
vendor/github.com/onsi/ginkgo/ginkgo_dsl.go
generated
vendored
|
@ -29,6 +29,7 @@ import (
|
||||||
"github.com/onsi/ginkgo/internal/writer"
|
"github.com/onsi/ginkgo/internal/writer"
|
||||||
"github.com/onsi/ginkgo/reporters"
|
"github.com/onsi/ginkgo/reporters"
|
||||||
"github.com/onsi/ginkgo/reporters/stenographer"
|
"github.com/onsi/ginkgo/reporters/stenographer"
|
||||||
|
colorable "github.com/onsi/ginkgo/reporters/stenographer/support/go-colorable"
|
||||||
"github.com/onsi/ginkgo/types"
|
"github.com/onsi/ginkgo/types"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -68,6 +69,14 @@ type GinkgoTestingT interface {
|
||||||
Fail()
|
Fail()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
//GinkgoRandomSeed returns the seed used to randomize spec execution order. It is
|
||||||
|
//useful for seeding your own pseudorandom number generators (PRNGs) to ensure
|
||||||
|
//consistent executions from run to run, where your tests contain variability (for
|
||||||
|
//example, when selecting random test data).
|
||||||
|
func GinkgoRandomSeed() int64 {
|
||||||
|
return config.GinkgoConfig.RandomSeed
|
||||||
|
}
|
||||||
|
|
||||||
//GinkgoParallelNode returns the parallel node number for the current ginkgo process
|
//GinkgoParallelNode returns the parallel node number for the current ginkgo process
|
||||||
//The node number is 1-indexed
|
//The node number is 1-indexed
|
||||||
func GinkgoParallelNode() int {
|
func GinkgoParallelNode() int {
|
||||||
|
@ -141,7 +150,8 @@ type GinkgoTestDescription struct {
|
||||||
FileName string
|
FileName string
|
||||||
LineNumber int
|
LineNumber int
|
||||||
|
|
||||||
Failed bool
|
Failed bool
|
||||||
|
Duration time.Duration
|
||||||
}
|
}
|
||||||
|
|
||||||
//CurrentGinkgoTestDescripton returns information about the current running test.
|
//CurrentGinkgoTestDescripton returns information about the current running test.
|
||||||
|
@ -161,6 +171,7 @@ func CurrentGinkgoTestDescription() GinkgoTestDescription {
|
||||||
FileName: subjectCodeLocation.FileName,
|
FileName: subjectCodeLocation.FileName,
|
||||||
LineNumber: subjectCodeLocation.LineNumber,
|
LineNumber: subjectCodeLocation.LineNumber,
|
||||||
Failed: summary.HasFailureState(),
|
Failed: summary.HasFailureState(),
|
||||||
|
Duration: summary.RunTime,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -168,6 +179,8 @@ func CurrentGinkgoTestDescription() GinkgoTestDescription {
|
||||||
//
|
//
|
||||||
//You use the Time() function to time how long the passed in body function takes to run
|
//You use the Time() function to time how long the passed in body function takes to run
|
||||||
//You use the RecordValue() function to track arbitrary numerical measurements.
|
//You use the RecordValue() function to track arbitrary numerical measurements.
|
||||||
|
//The RecordValueWithPrecision() function can be used alternatively to provide the unit
|
||||||
|
//and resolution of the numeric measurement.
|
||||||
//The optional info argument is passed to the test reporter and can be used to
|
//The optional info argument is passed to the test reporter and can be used to
|
||||||
// provide the measurement data to a custom reporter with context.
|
// provide the measurement data to a custom reporter with context.
|
||||||
//
|
//
|
||||||
|
@ -175,6 +188,7 @@ func CurrentGinkgoTestDescription() GinkgoTestDescription {
|
||||||
type Benchmarker interface {
|
type Benchmarker interface {
|
||||||
Time(name string, body func(), info ...interface{}) (elapsedTime time.Duration)
|
Time(name string, body func(), info ...interface{}) (elapsedTime time.Duration)
|
||||||
RecordValue(name string, value float64, info ...interface{})
|
RecordValue(name string, value float64, info ...interface{})
|
||||||
|
RecordValueWithPrecision(name string, value float64, units string, precision int, info ...interface{})
|
||||||
}
|
}
|
||||||
|
|
||||||
//RunSpecs is the entry point for the Ginkgo test runner.
|
//RunSpecs is the entry point for the Ginkgo test runner.
|
||||||
|
@ -191,7 +205,7 @@ func RunSpecs(t GinkgoTestingT, description string) bool {
|
||||||
//To run your tests with Ginkgo's default reporter and your custom reporter(s), replace
|
//To run your tests with Ginkgo's default reporter and your custom reporter(s), replace
|
||||||
//RunSpecs() with this method.
|
//RunSpecs() with this method.
|
||||||
func RunSpecsWithDefaultAndCustomReporters(t GinkgoTestingT, description string, specReporters []Reporter) bool {
|
func RunSpecsWithDefaultAndCustomReporters(t GinkgoTestingT, description string, specReporters []Reporter) bool {
|
||||||
specReporters = append([]Reporter{buildDefaultReporter()}, specReporters...)
|
specReporters = append(specReporters, buildDefaultReporter())
|
||||||
return RunSpecsWithCustomReporters(t, description, specReporters)
|
return RunSpecsWithCustomReporters(t, description, specReporters)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -205,7 +219,7 @@ func RunSpecsWithCustomReporters(t GinkgoTestingT, description string, specRepor
|
||||||
reporters[i] = reporter
|
reporters[i] = reporter
|
||||||
}
|
}
|
||||||
passed, hasFocusedTests := globalSuite.Run(t, description, reporters, writer, config.GinkgoConfig)
|
passed, hasFocusedTests := globalSuite.Run(t, description, reporters, writer, config.GinkgoConfig)
|
||||||
if passed && hasFocusedTests {
|
if passed && hasFocusedTests && strings.TrimSpace(os.Getenv("GINKGO_EDITOR_INTEGRATION")) == "" {
|
||||||
fmt.Println("PASS | FOCUSED")
|
fmt.Println("PASS | FOCUSED")
|
||||||
os.Exit(types.GINKGO_FOCUS_EXIT_CODE)
|
os.Exit(types.GINKGO_FOCUS_EXIT_CODE)
|
||||||
}
|
}
|
||||||
|
@ -215,14 +229,18 @@ func RunSpecsWithCustomReporters(t GinkgoTestingT, description string, specRepor
|
||||||
func buildDefaultReporter() Reporter {
|
func buildDefaultReporter() Reporter {
|
||||||
remoteReportingServer := config.GinkgoConfig.StreamHost
|
remoteReportingServer := config.GinkgoConfig.StreamHost
|
||||||
if remoteReportingServer == "" {
|
if remoteReportingServer == "" {
|
||||||
stenographer := stenographer.New(!config.DefaultReporterConfig.NoColor)
|
stenographer := stenographer.New(!config.DefaultReporterConfig.NoColor, config.GinkgoConfig.FlakeAttempts > 1, colorable.NewColorableStdout())
|
||||||
return reporters.NewDefaultReporter(config.DefaultReporterConfig, stenographer)
|
return reporters.NewDefaultReporter(config.DefaultReporterConfig, stenographer)
|
||||||
} else {
|
} else {
|
||||||
return remote.NewForwardingReporter(remoteReportingServer, &http.Client{}, remote.NewOutputInterceptor())
|
debugFile := ""
|
||||||
|
if config.GinkgoConfig.DebugParallel {
|
||||||
|
debugFile = fmt.Sprintf("ginkgo-node-%d.log", config.GinkgoConfig.ParallelNode)
|
||||||
|
}
|
||||||
|
return remote.NewForwardingReporter(config.DefaultReporterConfig, remoteReportingServer, &http.Client{}, remote.NewOutputInterceptor(), GinkgoWriter.(*writer.Writer), debugFile)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
//Skip notifies Ginkgo that the current spec should be skipped.
|
//Skip notifies Ginkgo that the current spec was skipped.
|
||||||
func Skip(message string, callerSkip ...int) {
|
func Skip(message string, callerSkip ...int) {
|
||||||
skip := 0
|
skip := 0
|
||||||
if len(callerSkip) > 0 {
|
if len(callerSkip) > 0 {
|
||||||
|
@ -264,9 +282,9 @@ func GinkgoRecover() {
|
||||||
//Describe blocks allow you to organize your specs. A Describe block can contain any number of
|
//Describe blocks allow you to organize your specs. A Describe block can contain any number of
|
||||||
//BeforeEach, AfterEach, JustBeforeEach, It, and Measurement blocks.
|
//BeforeEach, AfterEach, JustBeforeEach, It, and Measurement blocks.
|
||||||
//
|
//
|
||||||
//In addition you can nest Describe and Context blocks. Describe and Context blocks are functionally
|
//In addition you can nest Describe, Context and When blocks. Describe, Context and When blocks are functionally
|
||||||
//equivalent. The difference is purely semantic -- you typical Describe the behavior of an object
|
//equivalent. The difference is purely semantic -- you typical Describe the behavior of an object
|
||||||
//or method and, within that Describe, outline a number of Contexts.
|
//or method and, within that Describe, outline a number of Contexts and Whens.
|
||||||
func Describe(text string, body func()) bool {
|
func Describe(text string, body func()) bool {
|
||||||
globalSuite.PushContainerNode(text, body, types.FlagTypeNone, codelocation.New(1))
|
globalSuite.PushContainerNode(text, body, types.FlagTypeNone, codelocation.New(1))
|
||||||
return true
|
return true
|
||||||
|
@ -293,9 +311,9 @@ func XDescribe(text string, body func()) bool {
|
||||||
//Context blocks allow you to organize your specs. A Context block can contain any number of
|
//Context blocks allow you to organize your specs. A Context block can contain any number of
|
||||||
//BeforeEach, AfterEach, JustBeforeEach, It, and Measurement blocks.
|
//BeforeEach, AfterEach, JustBeforeEach, It, and Measurement blocks.
|
||||||
//
|
//
|
||||||
//In addition you can nest Describe and Context blocks. Describe and Context blocks are functionally
|
//In addition you can nest Describe, Context and When blocks. Describe, Context and When blocks are functionally
|
||||||
//equivalent. The difference is purely semantic -- you typical Describe the behavior of an object
|
//equivalent. The difference is purely semantic -- you typical Describe the behavior of an object
|
||||||
//or method and, within that Describe, outline a number of Contexts.
|
//or method and, within that Describe, outline a number of Contexts and Whens.
|
||||||
func Context(text string, body func()) bool {
|
func Context(text string, body func()) bool {
|
||||||
globalSuite.PushContainerNode(text, body, types.FlagTypeNone, codelocation.New(1))
|
globalSuite.PushContainerNode(text, body, types.FlagTypeNone, codelocation.New(1))
|
||||||
return true
|
return true
|
||||||
|
@ -319,6 +337,35 @@ func XContext(text string, body func()) bool {
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
//When blocks allow you to organize your specs. A When block can contain any number of
|
||||||
|
//BeforeEach, AfterEach, JustBeforeEach, It, and Measurement blocks.
|
||||||
|
//
|
||||||
|
//In addition you can nest Describe, Context and When blocks. Describe, Context and When blocks are functionally
|
||||||
|
//equivalent. The difference is purely semantic -- you typical Describe the behavior of an object
|
||||||
|
//or method and, within that Describe, outline a number of Contexts and Whens.
|
||||||
|
func When(text string, body func()) bool {
|
||||||
|
globalSuite.PushContainerNode("when "+text, body, types.FlagTypeNone, codelocation.New(1))
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
//You can focus the tests within a describe block using FWhen
|
||||||
|
func FWhen(text string, body func()) bool {
|
||||||
|
globalSuite.PushContainerNode("when "+text, body, types.FlagTypeFocused, codelocation.New(1))
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
//You can mark the tests within a describe block as pending using PWhen
|
||||||
|
func PWhen(text string, body func()) bool {
|
||||||
|
globalSuite.PushContainerNode("when "+text, body, types.FlagTypePending, codelocation.New(1))
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
//You can mark the tests within a describe block as pending using XWhen
|
||||||
|
func XWhen(text string, body func()) bool {
|
||||||
|
globalSuite.PushContainerNode("when "+text, body, types.FlagTypePending, codelocation.New(1))
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
//It blocks contain your test code and assertions. You cannot nest any other Ginkgo blocks
|
//It blocks contain your test code and assertions. You cannot nest any other Ginkgo blocks
|
||||||
//within an It block.
|
//within an It block.
|
||||||
//
|
//
|
||||||
|
@ -347,6 +394,32 @@ func XIt(text string, _ ...interface{}) bool {
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
//Specify blocks are aliases for It blocks and allow for more natural wording in situations
|
||||||
|
//which "It" does not fit into a natural sentence flow. All the same protocols apply for Specify blocks
|
||||||
|
//which apply to It blocks.
|
||||||
|
func Specify(text string, body interface{}, timeout ...float64) bool {
|
||||||
|
globalSuite.PushItNode(text, body, types.FlagTypeNone, codelocation.New(1), parseTimeout(timeout...))
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
//You can focus individual Specifys using FSpecify
|
||||||
|
func FSpecify(text string, body interface{}, timeout ...float64) bool {
|
||||||
|
globalSuite.PushItNode(text, body, types.FlagTypeFocused, codelocation.New(1), parseTimeout(timeout...))
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
//You can mark Specifys as pending using PSpecify
|
||||||
|
func PSpecify(text string, is ...interface{}) bool {
|
||||||
|
globalSuite.PushItNode(text, func() {}, types.FlagTypePending, codelocation.New(1), 0)
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
//You can mark Specifys as pending using XSpecify
|
||||||
|
func XSpecify(text string, is ...interface{}) bool {
|
||||||
|
globalSuite.PushItNode(text, func() {}, types.FlagTypePending, codelocation.New(1), 0)
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
//By allows you to better document large Its.
|
//By allows you to better document large Its.
|
||||||
//
|
//
|
||||||
//Generally you should try to keep your Its short and to the point. This is not always possible, however,
|
//Generally you should try to keep your Its short and to the point. This is not always possible, however,
|
||||||
|
|
14
vendor/github.com/onsi/ginkgo/internal/leafnodes/benchmarker.go
generated
vendored
14
vendor/github.com/onsi/ginkgo/internal/leafnodes/benchmarker.go
generated
vendored
|
@ -28,20 +28,27 @@ func (b *benchmarker) Time(name string, body func(), info ...interface{}) (elaps
|
||||||
|
|
||||||
b.mu.Lock()
|
b.mu.Lock()
|
||||||
defer b.mu.Unlock()
|
defer b.mu.Unlock()
|
||||||
measurement := b.getMeasurement(name, "Fastest Time", "Slowest Time", "Average Time", "s", info...)
|
measurement := b.getMeasurement(name, "Fastest Time", "Slowest Time", "Average Time", "s", 3, info...)
|
||||||
measurement.Results = append(measurement.Results, elapsedTime.Seconds())
|
measurement.Results = append(measurement.Results, elapsedTime.Seconds())
|
||||||
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
func (b *benchmarker) RecordValue(name string, value float64, info ...interface{}) {
|
func (b *benchmarker) RecordValue(name string, value float64, info ...interface{}) {
|
||||||
measurement := b.getMeasurement(name, "Smallest", " Largest", " Average", "", info...)
|
|
||||||
b.mu.Lock()
|
b.mu.Lock()
|
||||||
|
measurement := b.getMeasurement(name, "Smallest", " Largest", " Average", "", 3, info...)
|
||||||
defer b.mu.Unlock()
|
defer b.mu.Unlock()
|
||||||
measurement.Results = append(measurement.Results, value)
|
measurement.Results = append(measurement.Results, value)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (b *benchmarker) getMeasurement(name string, smallestLabel string, largestLabel string, averageLabel string, units string, info ...interface{}) *types.SpecMeasurement {
|
func (b *benchmarker) RecordValueWithPrecision(name string, value float64, units string, precision int, info ...interface{}) {
|
||||||
|
b.mu.Lock()
|
||||||
|
measurement := b.getMeasurement(name, "Smallest", " Largest", " Average", units, precision, info...)
|
||||||
|
defer b.mu.Unlock()
|
||||||
|
measurement.Results = append(measurement.Results, value)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (b *benchmarker) getMeasurement(name string, smallestLabel string, largestLabel string, averageLabel string, units string, precision int, info ...interface{}) *types.SpecMeasurement {
|
||||||
measurement, ok := b.measurements[name]
|
measurement, ok := b.measurements[name]
|
||||||
if !ok {
|
if !ok {
|
||||||
var computedInfo interface{}
|
var computedInfo interface{}
|
||||||
|
@ -57,6 +64,7 @@ func (b *benchmarker) getMeasurement(name string, smallestLabel string, largestL
|
||||||
LargestLabel: largestLabel,
|
LargestLabel: largestLabel,
|
||||||
AverageLabel: averageLabel,
|
AverageLabel: averageLabel,
|
||||||
Units: units,
|
Units: units,
|
||||||
|
Precision: precision,
|
||||||
Results: make([]float64, 0),
|
Results: make([]float64, 0),
|
||||||
}
|
}
|
||||||
b.measurements[name] = measurement
|
b.measurements[name] = measurement
|
||||||
|
|
3
vendor/github.com/onsi/ginkgo/internal/leafnodes/it_node.go
generated
vendored
3
vendor/github.com/onsi/ginkgo/internal/leafnodes/it_node.go
generated
vendored
|
@ -1,9 +1,10 @@
|
||||||
package leafnodes
|
package leafnodes
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/onsi/ginkgo/internal/failer"
|
"github.com/onsi/ginkgo/internal/failer"
|
||||||
"github.com/onsi/ginkgo/types"
|
"github.com/onsi/ginkgo/types"
|
||||||
"time"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
type ItNode struct {
|
type ItNode struct {
|
||||||
|
|
3
vendor/github.com/onsi/ginkgo/internal/leafnodes/measure_node.go
generated
vendored
3
vendor/github.com/onsi/ginkgo/internal/leafnodes/measure_node.go
generated
vendored
|
@ -1,9 +1,10 @@
|
||||||
package leafnodes
|
package leafnodes
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"reflect"
|
||||||
|
|
||||||
"github.com/onsi/ginkgo/internal/failer"
|
"github.com/onsi/ginkgo/internal/failer"
|
||||||
"github.com/onsi/ginkgo/types"
|
"github.com/onsi/ginkgo/types"
|
||||||
"reflect"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
type MeasureNode struct {
|
type MeasureNode struct {
|
||||||
|
|
16
vendor/github.com/onsi/ginkgo/internal/leafnodes/runner.go
generated
vendored
16
vendor/github.com/onsi/ginkgo/internal/leafnodes/runner.go
generated
vendored
|
@ -2,11 +2,12 @@ package leafnodes
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"reflect"
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/onsi/ginkgo/internal/codelocation"
|
"github.com/onsi/ginkgo/internal/codelocation"
|
||||||
"github.com/onsi/ginkgo/internal/failer"
|
"github.com/onsi/ginkgo/internal/failer"
|
||||||
"github.com/onsi/ginkgo/types"
|
"github.com/onsi/ginkgo/types"
|
||||||
"reflect"
|
|
||||||
"time"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
type runner struct {
|
type runner struct {
|
||||||
|
@ -68,7 +69,7 @@ func (r *runner) runAsync() (outcome types.SpecState, failure types.SpecFailure)
|
||||||
done := make(chan interface{}, 1)
|
done := make(chan interface{}, 1)
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
finished := false
|
finished := false
|
||||||
|
|
||||||
defer func() {
|
defer func() {
|
||||||
if e := recover(); e != nil || !finished {
|
if e := recover(); e != nil || !finished {
|
||||||
|
@ -83,9 +84,12 @@ func (r *runner) runAsync() (outcome types.SpecState, failure types.SpecFailure)
|
||||||
}()
|
}()
|
||||||
|
|
||||||
r.asyncFunc(done)
|
r.asyncFunc(done)
|
||||||
finished = true
|
finished = true
|
||||||
}()
|
}()
|
||||||
|
|
||||||
|
// If this goroutine gets no CPU time before the select block,
|
||||||
|
// the <-done case may complete even if the test took longer than the timeoutThreshold.
|
||||||
|
// This can cause flaky behaviour, but we haven't seen it in the wild.
|
||||||
select {
|
select {
|
||||||
case <-done:
|
case <-done:
|
||||||
case <-time.After(r.timeoutThreshold):
|
case <-time.After(r.timeoutThreshold):
|
||||||
|
@ -96,7 +100,7 @@ func (r *runner) runAsync() (outcome types.SpecState, failure types.SpecFailure)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
func (r *runner) runSync() (outcome types.SpecState, failure types.SpecFailure) {
|
func (r *runner) runSync() (outcome types.SpecState, failure types.SpecFailure) {
|
||||||
finished := false
|
finished := false
|
||||||
|
|
||||||
defer func() {
|
defer func() {
|
||||||
if e := recover(); e != nil || !finished {
|
if e := recover(); e != nil || !finished {
|
||||||
|
@ -107,7 +111,7 @@ func (r *runner) runSync() (outcome types.SpecState, failure types.SpecFailure)
|
||||||
}()
|
}()
|
||||||
|
|
||||||
r.syncFunc()
|
r.syncFunc()
|
||||||
finished = true
|
finished = true
|
||||||
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
3
vendor/github.com/onsi/ginkgo/internal/leafnodes/setup_nodes.go
generated
vendored
3
vendor/github.com/onsi/ginkgo/internal/leafnodes/setup_nodes.go
generated
vendored
|
@ -1,9 +1,10 @@
|
||||||
package leafnodes
|
package leafnodes
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/onsi/ginkgo/internal/failer"
|
"github.com/onsi/ginkgo/internal/failer"
|
||||||
"github.com/onsi/ginkgo/types"
|
"github.com/onsi/ginkgo/types"
|
||||||
"time"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
type SetupNode struct {
|
type SetupNode struct {
|
||||||
|
|
3
vendor/github.com/onsi/ginkgo/internal/leafnodes/suite_nodes.go
generated
vendored
3
vendor/github.com/onsi/ginkgo/internal/leafnodes/suite_nodes.go
generated
vendored
|
@ -1,9 +1,10 @@
|
||||||
package leafnodes
|
package leafnodes
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/onsi/ginkgo/internal/failer"
|
"github.com/onsi/ginkgo/internal/failer"
|
||||||
"github.com/onsi/ginkgo/types"
|
"github.com/onsi/ginkgo/types"
|
||||||
"time"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
type SuiteNode interface {
|
type SuiteNode interface {
|
||||||
|
|
5
vendor/github.com/onsi/ginkgo/internal/leafnodes/synchronized_after_suite_node.go
generated
vendored
5
vendor/github.com/onsi/ginkgo/internal/leafnodes/synchronized_after_suite_node.go
generated
vendored
|
@ -2,11 +2,12 @@ package leafnodes
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"github.com/onsi/ginkgo/internal/failer"
|
|
||||||
"github.com/onsi/ginkgo/types"
|
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"net/http"
|
"net/http"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/internal/failer"
|
||||||
|
"github.com/onsi/ginkgo/types"
|
||||||
)
|
)
|
||||||
|
|
||||||
type synchronizedAfterSuiteNode struct {
|
type synchronizedAfterSuiteNode struct {
|
||||||
|
|
7
vendor/github.com/onsi/ginkgo/internal/leafnodes/synchronized_before_suite_node.go
generated
vendored
7
vendor/github.com/onsi/ginkgo/internal/leafnodes/synchronized_before_suite_node.go
generated
vendored
|
@ -3,12 +3,13 @@ package leafnodes
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"github.com/onsi/ginkgo/internal/failer"
|
|
||||||
"github.com/onsi/ginkgo/types"
|
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"net/http"
|
"net/http"
|
||||||
"reflect"
|
"reflect"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/internal/failer"
|
||||||
|
"github.com/onsi/ginkgo/types"
|
||||||
)
|
)
|
||||||
|
|
||||||
type synchronizedBeforeSuiteNode struct {
|
type synchronizedBeforeSuiteNode struct {
|
||||||
|
@ -109,8 +110,6 @@ func (node *synchronizedBeforeSuiteNode) waitForA(syncHost string) (types.SpecSt
|
||||||
|
|
||||||
time.Sleep(50 * time.Millisecond)
|
time.Sleep(50 * time.Millisecond)
|
||||||
}
|
}
|
||||||
|
|
||||||
return types.SpecStateFailed, failure("Shouldn't get here!")
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (node *synchronizedBeforeSuiteNode) Passed() bool {
|
func (node *synchronizedBeforeSuiteNode) Passed() bool {
|
||||||
|
|
11
vendor/github.com/onsi/ginkgo/internal/remote/aggregator.go
generated
vendored
11
vendor/github.com/onsi/ginkgo/internal/remote/aggregator.go
generated
vendored
|
@ -125,14 +125,12 @@ func (aggregator *Aggregator) registerSuiteBeginning(configAndSuite configAndSui
|
||||||
|
|
||||||
aggregator.stenographer.AnnounceSuite(configAndSuite.summary.SuiteDescription, configAndSuite.config.RandomSeed, configAndSuite.config.RandomizeAllSpecs, aggregator.config.Succinct)
|
aggregator.stenographer.AnnounceSuite(configAndSuite.summary.SuiteDescription, configAndSuite.config.RandomSeed, configAndSuite.config.RandomizeAllSpecs, aggregator.config.Succinct)
|
||||||
|
|
||||||
numberOfSpecsToRun := 0
|
|
||||||
totalNumberOfSpecs := 0
|
totalNumberOfSpecs := 0
|
||||||
for _, configAndSuite := range aggregator.aggregatedSuiteBeginnings {
|
if len(aggregator.aggregatedSuiteBeginnings) > 0 {
|
||||||
numberOfSpecsToRun += configAndSuite.summary.NumberOfSpecsThatWillBeRun
|
totalNumberOfSpecs = configAndSuite.summary.NumberOfSpecsBeforeParallelization
|
||||||
totalNumberOfSpecs += configAndSuite.summary.NumberOfTotalSpecs
|
|
||||||
}
|
}
|
||||||
|
|
||||||
aggregator.stenographer.AnnounceNumberOfSpecs(numberOfSpecsToRun, totalNumberOfSpecs, aggregator.config.Succinct)
|
aggregator.stenographer.AnnounceTotalNumberOfSpecs(totalNumberOfSpecs, aggregator.config.Succinct)
|
||||||
aggregator.stenographer.AnnounceAggregatedParallelRun(aggregator.nodeCount, aggregator.config.Succinct)
|
aggregator.stenographer.AnnounceAggregatedParallelRun(aggregator.nodeCount, aggregator.config.Succinct)
|
||||||
aggregator.flushCompletedSpecs()
|
aggregator.flushCompletedSpecs()
|
||||||
}
|
}
|
||||||
|
@ -209,7 +207,7 @@ func (aggregator *Aggregator) announceSpec(specSummary *types.SpecSummary) {
|
||||||
case types.SpecStatePending:
|
case types.SpecStatePending:
|
||||||
aggregator.stenographer.AnnouncePendingSpec(specSummary, aggregator.config.NoisyPendings && !aggregator.config.Succinct)
|
aggregator.stenographer.AnnouncePendingSpec(specSummary, aggregator.config.NoisyPendings && !aggregator.config.Succinct)
|
||||||
case types.SpecStateSkipped:
|
case types.SpecStateSkipped:
|
||||||
aggregator.stenographer.AnnounceSkippedSpec(specSummary, aggregator.config.Succinct, aggregator.config.FullTrace)
|
aggregator.stenographer.AnnounceSkippedSpec(specSummary, aggregator.config.Succinct || !aggregator.config.NoisySkippings, aggregator.config.FullTrace)
|
||||||
case types.SpecStateTimedOut:
|
case types.SpecStateTimedOut:
|
||||||
aggregator.stenographer.AnnounceSpecTimedOut(specSummary, aggregator.config.Succinct, aggregator.config.FullTrace)
|
aggregator.stenographer.AnnounceSpecTimedOut(specSummary, aggregator.config.Succinct, aggregator.config.FullTrace)
|
||||||
case types.SpecStatePanicked:
|
case types.SpecStatePanicked:
|
||||||
|
@ -239,6 +237,7 @@ func (aggregator *Aggregator) registerSuiteEnding(suite *types.SuiteSummary) (fi
|
||||||
aggregatedSuiteSummary.NumberOfFailedSpecs += suiteSummary.NumberOfFailedSpecs
|
aggregatedSuiteSummary.NumberOfFailedSpecs += suiteSummary.NumberOfFailedSpecs
|
||||||
aggregatedSuiteSummary.NumberOfPendingSpecs += suiteSummary.NumberOfPendingSpecs
|
aggregatedSuiteSummary.NumberOfPendingSpecs += suiteSummary.NumberOfPendingSpecs
|
||||||
aggregatedSuiteSummary.NumberOfSkippedSpecs += suiteSummary.NumberOfSkippedSpecs
|
aggregatedSuiteSummary.NumberOfSkippedSpecs += suiteSummary.NumberOfSkippedSpecs
|
||||||
|
aggregatedSuiteSummary.NumberOfFlakedSpecs += suiteSummary.NumberOfFlakedSpecs
|
||||||
}
|
}
|
||||||
|
|
||||||
aggregatedSuiteSummary.RunTime = time.Since(aggregator.startTime)
|
aggregatedSuiteSummary.RunTime = time.Since(aggregator.startTime)
|
||||||
|
|
61
vendor/github.com/onsi/ginkgo/internal/remote/forwarding_reporter.go
generated
vendored
61
vendor/github.com/onsi/ginkgo/internal/remote/forwarding_reporter.go
generated
vendored
|
@ -3,8 +3,14 @@ package remote
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"os"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/internal/writer"
|
||||||
|
"github.com/onsi/ginkgo/reporters"
|
||||||
|
"github.com/onsi/ginkgo/reporters/stenographer"
|
||||||
|
|
||||||
"github.com/onsi/ginkgo/config"
|
"github.com/onsi/ginkgo/config"
|
||||||
"github.com/onsi/ginkgo/types"
|
"github.com/onsi/ginkgo/types"
|
||||||
|
@ -30,14 +36,41 @@ type ForwardingReporter struct {
|
||||||
serverHost string
|
serverHost string
|
||||||
poster Poster
|
poster Poster
|
||||||
outputInterceptor OutputInterceptor
|
outputInterceptor OutputInterceptor
|
||||||
|
debugMode bool
|
||||||
|
debugFile *os.File
|
||||||
|
nestedReporter *reporters.DefaultReporter
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewForwardingReporter(serverHost string, poster Poster, outputInterceptor OutputInterceptor) *ForwardingReporter {
|
func NewForwardingReporter(config config.DefaultReporterConfigType, serverHost string, poster Poster, outputInterceptor OutputInterceptor, ginkgoWriter *writer.Writer, debugFile string) *ForwardingReporter {
|
||||||
return &ForwardingReporter{
|
reporter := &ForwardingReporter{
|
||||||
serverHost: serverHost,
|
serverHost: serverHost,
|
||||||
poster: poster,
|
poster: poster,
|
||||||
outputInterceptor: outputInterceptor,
|
outputInterceptor: outputInterceptor,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if debugFile != "" {
|
||||||
|
var err error
|
||||||
|
reporter.debugMode = true
|
||||||
|
reporter.debugFile, err = os.Create(debugFile)
|
||||||
|
if err != nil {
|
||||||
|
fmt.Println(err.Error())
|
||||||
|
os.Exit(1)
|
||||||
|
}
|
||||||
|
|
||||||
|
if !config.Verbose {
|
||||||
|
//if verbose is true then the GinkgoWriter emits to stdout. Don't _also_ redirect GinkgoWriter output as that will result in duplication.
|
||||||
|
ginkgoWriter.AndRedirectTo(reporter.debugFile)
|
||||||
|
}
|
||||||
|
outputInterceptor.StreamTo(reporter.debugFile) //This is not working
|
||||||
|
|
||||||
|
stenographer := stenographer.New(false, true, reporter.debugFile)
|
||||||
|
config.Succinct = false
|
||||||
|
config.Verbose = true
|
||||||
|
config.FullTrace = true
|
||||||
|
reporter.nestedReporter = reporters.NewDefaultReporter(config, stenographer)
|
||||||
|
}
|
||||||
|
|
||||||
|
return reporter
|
||||||
}
|
}
|
||||||
|
|
||||||
func (reporter *ForwardingReporter) post(path string, data interface{}) {
|
func (reporter *ForwardingReporter) post(path string, data interface{}) {
|
||||||
|
@ -56,6 +89,10 @@ func (reporter *ForwardingReporter) SpecSuiteWillBegin(conf config.GinkgoConfigT
|
||||||
}
|
}
|
||||||
|
|
||||||
reporter.outputInterceptor.StartInterceptingOutput()
|
reporter.outputInterceptor.StartInterceptingOutput()
|
||||||
|
if reporter.debugMode {
|
||||||
|
reporter.nestedReporter.SpecSuiteWillBegin(conf, summary)
|
||||||
|
reporter.debugFile.Sync()
|
||||||
|
}
|
||||||
reporter.post("/SpecSuiteWillBegin", data)
|
reporter.post("/SpecSuiteWillBegin", data)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -63,10 +100,18 @@ func (reporter *ForwardingReporter) BeforeSuiteDidRun(setupSummary *types.SetupS
|
||||||
output, _ := reporter.outputInterceptor.StopInterceptingAndReturnOutput()
|
output, _ := reporter.outputInterceptor.StopInterceptingAndReturnOutput()
|
||||||
reporter.outputInterceptor.StartInterceptingOutput()
|
reporter.outputInterceptor.StartInterceptingOutput()
|
||||||
setupSummary.CapturedOutput = output
|
setupSummary.CapturedOutput = output
|
||||||
|
if reporter.debugMode {
|
||||||
|
reporter.nestedReporter.BeforeSuiteDidRun(setupSummary)
|
||||||
|
reporter.debugFile.Sync()
|
||||||
|
}
|
||||||
reporter.post("/BeforeSuiteDidRun", setupSummary)
|
reporter.post("/BeforeSuiteDidRun", setupSummary)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (reporter *ForwardingReporter) SpecWillRun(specSummary *types.SpecSummary) {
|
func (reporter *ForwardingReporter) SpecWillRun(specSummary *types.SpecSummary) {
|
||||||
|
if reporter.debugMode {
|
||||||
|
reporter.nestedReporter.SpecWillRun(specSummary)
|
||||||
|
reporter.debugFile.Sync()
|
||||||
|
}
|
||||||
reporter.post("/SpecWillRun", specSummary)
|
reporter.post("/SpecWillRun", specSummary)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -74,6 +119,10 @@ func (reporter *ForwardingReporter) SpecDidComplete(specSummary *types.SpecSumma
|
||||||
output, _ := reporter.outputInterceptor.StopInterceptingAndReturnOutput()
|
output, _ := reporter.outputInterceptor.StopInterceptingAndReturnOutput()
|
||||||
reporter.outputInterceptor.StartInterceptingOutput()
|
reporter.outputInterceptor.StartInterceptingOutput()
|
||||||
specSummary.CapturedOutput = output
|
specSummary.CapturedOutput = output
|
||||||
|
if reporter.debugMode {
|
||||||
|
reporter.nestedReporter.SpecDidComplete(specSummary)
|
||||||
|
reporter.debugFile.Sync()
|
||||||
|
}
|
||||||
reporter.post("/SpecDidComplete", specSummary)
|
reporter.post("/SpecDidComplete", specSummary)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -81,10 +130,18 @@ func (reporter *ForwardingReporter) AfterSuiteDidRun(setupSummary *types.SetupSu
|
||||||
output, _ := reporter.outputInterceptor.StopInterceptingAndReturnOutput()
|
output, _ := reporter.outputInterceptor.StopInterceptingAndReturnOutput()
|
||||||
reporter.outputInterceptor.StartInterceptingOutput()
|
reporter.outputInterceptor.StartInterceptingOutput()
|
||||||
setupSummary.CapturedOutput = output
|
setupSummary.CapturedOutput = output
|
||||||
|
if reporter.debugMode {
|
||||||
|
reporter.nestedReporter.AfterSuiteDidRun(setupSummary)
|
||||||
|
reporter.debugFile.Sync()
|
||||||
|
}
|
||||||
reporter.post("/AfterSuiteDidRun", setupSummary)
|
reporter.post("/AfterSuiteDidRun", setupSummary)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (reporter *ForwardingReporter) SpecSuiteDidEnd(summary *types.SuiteSummary) {
|
func (reporter *ForwardingReporter) SpecSuiteDidEnd(summary *types.SuiteSummary) {
|
||||||
reporter.outputInterceptor.StopInterceptingAndReturnOutput()
|
reporter.outputInterceptor.StopInterceptingAndReturnOutput()
|
||||||
|
if reporter.debugMode {
|
||||||
|
reporter.nestedReporter.SpecSuiteDidEnd(summary)
|
||||||
|
reporter.debugFile.Sync()
|
||||||
|
}
|
||||||
reporter.post("/SpecSuiteDidEnd", summary)
|
reporter.post("/SpecSuiteDidEnd", summary)
|
||||||
}
|
}
|
||||||
|
|
3
vendor/github.com/onsi/ginkgo/internal/remote/output_interceptor.go
generated
vendored
3
vendor/github.com/onsi/ginkgo/internal/remote/output_interceptor.go
generated
vendored
|
@ -1,5 +1,7 @@
|
||||||
package remote
|
package remote
|
||||||
|
|
||||||
|
import "os"
|
||||||
|
|
||||||
/*
|
/*
|
||||||
The OutputInterceptor is used by the ForwardingReporter to
|
The OutputInterceptor is used by the ForwardingReporter to
|
||||||
intercept and capture all stdin and stderr output during a test run.
|
intercept and capture all stdin and stderr output during a test run.
|
||||||
|
@ -7,4 +9,5 @@ intercept and capture all stdin and stderr output during a test run.
|
||||||
type OutputInterceptor interface {
|
type OutputInterceptor interface {
|
||||||
StartInterceptingOutput() error
|
StartInterceptingOutput() error
|
||||||
StopInterceptingAndReturnOutput() (string, error)
|
StopInterceptingAndReturnOutput() (string, error)
|
||||||
|
StreamTo(*os.File)
|
||||||
}
|
}
|
||||||
|
|
39
vendor/github.com/onsi/ginkgo/internal/remote/output_interceptor_unix.go
generated
vendored
39
vendor/github.com/onsi/ginkgo/internal/remote/output_interceptor_unix.go
generated
vendored
|
@ -1,4 +1,4 @@
|
||||||
// +build freebsd openbsd netbsd dragonfly darwin linux
|
// +build freebsd openbsd netbsd dragonfly darwin linux solaris
|
||||||
|
|
||||||
package remote
|
package remote
|
||||||
|
|
||||||
|
@ -6,7 +6,8 @@ import (
|
||||||
"errors"
|
"errors"
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"os"
|
"os"
|
||||||
"syscall"
|
|
||||||
|
"github.com/hpcloud/tail"
|
||||||
)
|
)
|
||||||
|
|
||||||
func NewOutputInterceptor() OutputInterceptor {
|
func NewOutputInterceptor() OutputInterceptor {
|
||||||
|
@ -15,7 +16,10 @@ func NewOutputInterceptor() OutputInterceptor {
|
||||||
|
|
||||||
type outputInterceptor struct {
|
type outputInterceptor struct {
|
||||||
redirectFile *os.File
|
redirectFile *os.File
|
||||||
|
streamTarget *os.File
|
||||||
intercepting bool
|
intercepting bool
|
||||||
|
tailer *tail.Tail
|
||||||
|
doneTailing chan bool
|
||||||
}
|
}
|
||||||
|
|
||||||
func (interceptor *outputInterceptor) StartInterceptingOutput() error {
|
func (interceptor *outputInterceptor) StartInterceptingOutput() error {
|
||||||
|
@ -31,8 +35,24 @@ func (interceptor *outputInterceptor) StartInterceptingOutput() error {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
syscall.Dup2(int(interceptor.redirectFile.Fd()), 1)
|
// Call a function in ./syscall_dup_*.go
|
||||||
syscall.Dup2(int(interceptor.redirectFile.Fd()), 2)
|
// If building for everything other than linux_arm64,
|
||||||
|
// use a "normal" syscall.Dup2(oldfd, newfd) call. If building for linux_arm64 (which doesn't have syscall.Dup2)
|
||||||
|
// call syscall.Dup3(oldfd, newfd, 0). They are nearly identical, see: http://linux.die.net/man/2/dup3
|
||||||
|
syscallDup(int(interceptor.redirectFile.Fd()), 1)
|
||||||
|
syscallDup(int(interceptor.redirectFile.Fd()), 2)
|
||||||
|
|
||||||
|
if interceptor.streamTarget != nil {
|
||||||
|
interceptor.tailer, _ = tail.TailFile(interceptor.redirectFile.Name(), tail.Config{Follow: true})
|
||||||
|
interceptor.doneTailing = make(chan bool)
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
for line := range interceptor.tailer.Lines {
|
||||||
|
interceptor.streamTarget.Write([]byte(line.Text + "\n"))
|
||||||
|
}
|
||||||
|
close(interceptor.doneTailing)
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@ -48,5 +68,16 @@ func (interceptor *outputInterceptor) StopInterceptingAndReturnOutput() (string,
|
||||||
|
|
||||||
interceptor.intercepting = false
|
interceptor.intercepting = false
|
||||||
|
|
||||||
|
if interceptor.streamTarget != nil {
|
||||||
|
interceptor.tailer.Stop()
|
||||||
|
interceptor.tailer.Cleanup()
|
||||||
|
<-interceptor.doneTailing
|
||||||
|
interceptor.streamTarget.Sync()
|
||||||
|
}
|
||||||
|
|
||||||
return string(output), err
|
return string(output), err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (interceptor *outputInterceptor) StreamTo(out *os.File) {
|
||||||
|
interceptor.streamTarget = out
|
||||||
|
}
|
||||||
|
|
3
vendor/github.com/onsi/ginkgo/internal/remote/output_interceptor_win.go
generated
vendored
3
vendor/github.com/onsi/ginkgo/internal/remote/output_interceptor_win.go
generated
vendored
|
@ -4,6 +4,7 @@ package remote
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"errors"
|
"errors"
|
||||||
|
"os"
|
||||||
)
|
)
|
||||||
|
|
||||||
func NewOutputInterceptor() OutputInterceptor {
|
func NewOutputInterceptor() OutputInterceptor {
|
||||||
|
@ -31,3 +32,5 @@ func (interceptor *outputInterceptor) StopInterceptingAndReturnOutput() (string,
|
||||||
|
|
||||||
return "", nil
|
return "", nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (interceptor *outputInterceptor) StreamTo(*os.File) {}
|
||||||
|
|
28
vendor/github.com/onsi/ginkgo/internal/remote/server.go
generated
vendored
28
vendor/github.com/onsi/ginkgo/internal/remote/server.go
generated
vendored
|
@ -9,13 +9,16 @@ package remote
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"github.com/onsi/ginkgo/config"
|
|
||||||
"github.com/onsi/ginkgo/reporters"
|
|
||||||
"github.com/onsi/ginkgo/types"
|
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"net"
|
"net"
|
||||||
"net/http"
|
"net/http"
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/internal/spec_iterator"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/config"
|
||||||
|
"github.com/onsi/ginkgo/reporters"
|
||||||
|
"github.com/onsi/ginkgo/types"
|
||||||
)
|
)
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -29,6 +32,7 @@ type Server struct {
|
||||||
lock *sync.Mutex
|
lock *sync.Mutex
|
||||||
beforeSuiteData types.RemoteBeforeSuiteData
|
beforeSuiteData types.RemoteBeforeSuiteData
|
||||||
parallelTotal int
|
parallelTotal int
|
||||||
|
counter int
|
||||||
}
|
}
|
||||||
|
|
||||||
//Create a new server, automatically selecting a port
|
//Create a new server, automatically selecting a port
|
||||||
|
@ -41,7 +45,7 @@ func NewServer(parallelTotal int) (*Server, error) {
|
||||||
listener: listener,
|
listener: listener,
|
||||||
lock: &sync.Mutex{},
|
lock: &sync.Mutex{},
|
||||||
alives: make([]func() bool, parallelTotal),
|
alives: make([]func() bool, parallelTotal),
|
||||||
beforeSuiteData: types.RemoteBeforeSuiteData{nil, types.RemoteBeforeSuiteStatePending},
|
beforeSuiteData: types.RemoteBeforeSuiteData{Data: nil, State: types.RemoteBeforeSuiteStatePending},
|
||||||
parallelTotal: parallelTotal,
|
parallelTotal: parallelTotal,
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
@ -63,6 +67,8 @@ func (server *Server) Start() {
|
||||||
//synchronization endpoints
|
//synchronization endpoints
|
||||||
mux.HandleFunc("/BeforeSuiteState", server.handleBeforeSuiteState)
|
mux.HandleFunc("/BeforeSuiteState", server.handleBeforeSuiteState)
|
||||||
mux.HandleFunc("/RemoteAfterSuiteData", server.handleRemoteAfterSuiteData)
|
mux.HandleFunc("/RemoteAfterSuiteData", server.handleRemoteAfterSuiteData)
|
||||||
|
mux.HandleFunc("/counter", server.handleCounter)
|
||||||
|
mux.HandleFunc("/has-counter", server.handleHasCounter) //for backward compatibility
|
||||||
|
|
||||||
go httpServer.Serve(server.listener)
|
go httpServer.Serve(server.listener)
|
||||||
}
|
}
|
||||||
|
@ -202,3 +208,17 @@ func (server *Server) handleRemoteAfterSuiteData(writer http.ResponseWriter, req
|
||||||
enc := json.NewEncoder(writer)
|
enc := json.NewEncoder(writer)
|
||||||
enc.Encode(afterSuiteData)
|
enc.Encode(afterSuiteData)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (server *Server) handleCounter(writer http.ResponseWriter, request *http.Request) {
|
||||||
|
c := spec_iterator.Counter{}
|
||||||
|
server.lock.Lock()
|
||||||
|
c.Index = server.counter
|
||||||
|
server.counter = server.counter + 1
|
||||||
|
server.lock.Unlock()
|
||||||
|
|
||||||
|
json.NewEncoder(writer).Encode(c)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (server *Server) handleHasCounter(writer http.ResponseWriter, request *http.Request) {
|
||||||
|
writer.Write([]byte(""))
|
||||||
|
}
|
||||||
|
|
11
vendor/github.com/onsi/ginkgo/internal/remote/syscall_dup_linux_arm64.go
generated
vendored
Normal file
11
vendor/github.com/onsi/ginkgo/internal/remote/syscall_dup_linux_arm64.go
generated
vendored
Normal file
|
@ -0,0 +1,11 @@
|
||||||
|
// +build linux,arm64
|
||||||
|
|
||||||
|
package remote
|
||||||
|
|
||||||
|
import "syscall"
|
||||||
|
|
||||||
|
// linux_arm64 doesn't have syscall.Dup2 which ginkgo uses, so
|
||||||
|
// use the nearly identical syscall.Dup3 instead
|
||||||
|
func syscallDup(oldfd int, newfd int) (err error) {
|
||||||
|
return syscall.Dup3(oldfd, newfd, 0)
|
||||||
|
}
|
9
vendor/github.com/onsi/ginkgo/internal/remote/syscall_dup_solaris.go
generated
vendored
Normal file
9
vendor/github.com/onsi/ginkgo/internal/remote/syscall_dup_solaris.go
generated
vendored
Normal file
|
@ -0,0 +1,9 @@
|
||||||
|
// +build solaris
|
||||||
|
|
||||||
|
package remote
|
||||||
|
|
||||||
|
import "golang.org/x/sys/unix"
|
||||||
|
|
||||||
|
func syscallDup(oldfd int, newfd int) (err error) {
|
||||||
|
return unix.Dup2(oldfd, newfd)
|
||||||
|
}
|
11
vendor/github.com/onsi/ginkgo/internal/remote/syscall_dup_unix.go
generated
vendored
Normal file
11
vendor/github.com/onsi/ginkgo/internal/remote/syscall_dup_unix.go
generated
vendored
Normal file
|
@ -0,0 +1,11 @@
|
||||||
|
// +build !linux !arm64
|
||||||
|
// +build !windows
|
||||||
|
// +build !solaris
|
||||||
|
|
||||||
|
package remote
|
||||||
|
|
||||||
|
import "syscall"
|
||||||
|
|
||||||
|
func syscallDup(oldfd int, newfd int) (err error) {
|
||||||
|
return syscall.Dup2(oldfd, newfd)
|
||||||
|
}
|
82
vendor/github.com/onsi/ginkgo/internal/spec/spec.go
generated
vendored
82
vendor/github.com/onsi/ginkgo/internal/spec/spec.go
generated
vendored
|
@ -5,6 +5,8 @@ import (
|
||||||
"io"
|
"io"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"sync"
|
||||||
|
|
||||||
"github.com/onsi/ginkgo/internal/containernode"
|
"github.com/onsi/ginkgo/internal/containernode"
|
||||||
"github.com/onsi/ginkgo/internal/leafnodes"
|
"github.com/onsi/ginkgo/internal/leafnodes"
|
||||||
"github.com/onsi/ginkgo/types"
|
"github.com/onsi/ginkgo/types"
|
||||||
|
@ -17,9 +19,13 @@ type Spec struct {
|
||||||
|
|
||||||
containers []*containernode.ContainerNode
|
containers []*containernode.ContainerNode
|
||||||
|
|
||||||
state types.SpecState
|
state types.SpecState
|
||||||
runTime time.Duration
|
runTime time.Duration
|
||||||
failure types.SpecFailure
|
startTime time.Time
|
||||||
|
failure types.SpecFailure
|
||||||
|
previousFailures bool
|
||||||
|
|
||||||
|
stateMutex *sync.Mutex
|
||||||
}
|
}
|
||||||
|
|
||||||
func New(subject leafnodes.SubjectNode, containers []*containernode.ContainerNode, announceProgress bool) *Spec {
|
func New(subject leafnodes.SubjectNode, containers []*containernode.ContainerNode, announceProgress bool) *Spec {
|
||||||
|
@ -28,6 +34,7 @@ func New(subject leafnodes.SubjectNode, containers []*containernode.ContainerNod
|
||||||
containers: containers,
|
containers: containers,
|
||||||
focused: subject.Flag() == types.FlagTypeFocused,
|
focused: subject.Flag() == types.FlagTypeFocused,
|
||||||
announceProgress: announceProgress,
|
announceProgress: announceProgress,
|
||||||
|
stateMutex: &sync.Mutex{},
|
||||||
}
|
}
|
||||||
|
|
||||||
spec.processFlag(subject.Flag())
|
spec.processFlag(subject.Flag())
|
||||||
|
@ -42,28 +49,32 @@ func (spec *Spec) processFlag(flag types.FlagType) {
|
||||||
if flag == types.FlagTypeFocused {
|
if flag == types.FlagTypeFocused {
|
||||||
spec.focused = true
|
spec.focused = true
|
||||||
} else if flag == types.FlagTypePending {
|
} else if flag == types.FlagTypePending {
|
||||||
spec.state = types.SpecStatePending
|
spec.setState(types.SpecStatePending)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (spec *Spec) Skip() {
|
func (spec *Spec) Skip() {
|
||||||
spec.state = types.SpecStateSkipped
|
spec.setState(types.SpecStateSkipped)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (spec *Spec) Failed() bool {
|
func (spec *Spec) Failed() bool {
|
||||||
return spec.state == types.SpecStateFailed || spec.state == types.SpecStatePanicked || spec.state == types.SpecStateTimedOut
|
return spec.getState() == types.SpecStateFailed || spec.getState() == types.SpecStatePanicked || spec.getState() == types.SpecStateTimedOut
|
||||||
}
|
}
|
||||||
|
|
||||||
func (spec *Spec) Passed() bool {
|
func (spec *Spec) Passed() bool {
|
||||||
return spec.state == types.SpecStatePassed
|
return spec.getState() == types.SpecStatePassed
|
||||||
|
}
|
||||||
|
|
||||||
|
func (spec *Spec) Flaked() bool {
|
||||||
|
return spec.getState() == types.SpecStatePassed && spec.previousFailures
|
||||||
}
|
}
|
||||||
|
|
||||||
func (spec *Spec) Pending() bool {
|
func (spec *Spec) Pending() bool {
|
||||||
return spec.state == types.SpecStatePending
|
return spec.getState() == types.SpecStatePending
|
||||||
}
|
}
|
||||||
|
|
||||||
func (spec *Spec) Skipped() bool {
|
func (spec *Spec) Skipped() bool {
|
||||||
return spec.state == types.SpecStateSkipped
|
return spec.getState() == types.SpecStateSkipped
|
||||||
}
|
}
|
||||||
|
|
||||||
func (spec *Spec) Focused() bool {
|
func (spec *Spec) Focused() bool {
|
||||||
|
@ -86,13 +97,18 @@ func (spec *Spec) Summary(suiteID string) *types.SpecSummary {
|
||||||
componentTexts[len(spec.containers)] = spec.subject.Text()
|
componentTexts[len(spec.containers)] = spec.subject.Text()
|
||||||
componentCodeLocations[len(spec.containers)] = spec.subject.CodeLocation()
|
componentCodeLocations[len(spec.containers)] = spec.subject.CodeLocation()
|
||||||
|
|
||||||
|
runTime := spec.runTime
|
||||||
|
if runTime == 0 && !spec.startTime.IsZero() {
|
||||||
|
runTime = time.Since(spec.startTime)
|
||||||
|
}
|
||||||
|
|
||||||
return &types.SpecSummary{
|
return &types.SpecSummary{
|
||||||
IsMeasurement: spec.IsMeasurement(),
|
IsMeasurement: spec.IsMeasurement(),
|
||||||
NumberOfSamples: spec.subject.Samples(),
|
NumberOfSamples: spec.subject.Samples(),
|
||||||
ComponentTexts: componentTexts,
|
ComponentTexts: componentTexts,
|
||||||
ComponentCodeLocations: componentCodeLocations,
|
ComponentCodeLocations: componentCodeLocations,
|
||||||
State: spec.state,
|
State: spec.getState(),
|
||||||
RunTime: spec.runTime,
|
RunTime: runTime,
|
||||||
Failure: spec.failure,
|
Failure: spec.failure,
|
||||||
Measurements: spec.measurementsReport(),
|
Measurements: spec.measurementsReport(),
|
||||||
SuiteID: suiteID,
|
SuiteID: suiteID,
|
||||||
|
@ -109,22 +125,38 @@ func (spec *Spec) ConcatenatedString() string {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (spec *Spec) Run(writer io.Writer) {
|
func (spec *Spec) Run(writer io.Writer) {
|
||||||
startTime := time.Now()
|
if spec.getState() == types.SpecStateFailed {
|
||||||
|
spec.previousFailures = true
|
||||||
|
}
|
||||||
|
|
||||||
|
spec.startTime = time.Now()
|
||||||
defer func() {
|
defer func() {
|
||||||
spec.runTime = time.Since(startTime)
|
spec.runTime = time.Since(spec.startTime)
|
||||||
}()
|
}()
|
||||||
|
|
||||||
for sample := 0; sample < spec.subject.Samples(); sample++ {
|
for sample := 0; sample < spec.subject.Samples(); sample++ {
|
||||||
spec.runSample(sample, writer)
|
spec.runSample(sample, writer)
|
||||||
|
|
||||||
if spec.state != types.SpecStatePassed {
|
if spec.getState() != types.SpecStatePassed {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (spec *Spec) getState() types.SpecState {
|
||||||
|
spec.stateMutex.Lock()
|
||||||
|
defer spec.stateMutex.Unlock()
|
||||||
|
return spec.state
|
||||||
|
}
|
||||||
|
|
||||||
|
func (spec *Spec) setState(state types.SpecState) {
|
||||||
|
spec.stateMutex.Lock()
|
||||||
|
defer spec.stateMutex.Unlock()
|
||||||
|
spec.state = state
|
||||||
|
}
|
||||||
|
|
||||||
func (spec *Spec) runSample(sample int, writer io.Writer) {
|
func (spec *Spec) runSample(sample int, writer io.Writer) {
|
||||||
spec.state = types.SpecStatePassed
|
spec.setState(types.SpecStatePassed)
|
||||||
spec.failure = types.SpecFailure{}
|
spec.failure = types.SpecFailure{}
|
||||||
innerMostContainerIndexToUnwind := -1
|
innerMostContainerIndexToUnwind := -1
|
||||||
|
|
||||||
|
@ -134,8 +166,8 @@ func (spec *Spec) runSample(sample int, writer io.Writer) {
|
||||||
for _, afterEach := range container.SetupNodesOfType(types.SpecComponentTypeAfterEach) {
|
for _, afterEach := range container.SetupNodesOfType(types.SpecComponentTypeAfterEach) {
|
||||||
spec.announceSetupNode(writer, "AfterEach", container, afterEach)
|
spec.announceSetupNode(writer, "AfterEach", container, afterEach)
|
||||||
afterEachState, afterEachFailure := afterEach.Run()
|
afterEachState, afterEachFailure := afterEach.Run()
|
||||||
if afterEachState != types.SpecStatePassed && spec.state == types.SpecStatePassed {
|
if afterEachState != types.SpecStatePassed && spec.getState() == types.SpecStatePassed {
|
||||||
spec.state = afterEachState
|
spec.setState(afterEachState)
|
||||||
spec.failure = afterEachFailure
|
spec.failure = afterEachFailure
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -146,8 +178,10 @@ func (spec *Spec) runSample(sample int, writer io.Writer) {
|
||||||
innerMostContainerIndexToUnwind = i
|
innerMostContainerIndexToUnwind = i
|
||||||
for _, beforeEach := range container.SetupNodesOfType(types.SpecComponentTypeBeforeEach) {
|
for _, beforeEach := range container.SetupNodesOfType(types.SpecComponentTypeBeforeEach) {
|
||||||
spec.announceSetupNode(writer, "BeforeEach", container, beforeEach)
|
spec.announceSetupNode(writer, "BeforeEach", container, beforeEach)
|
||||||
spec.state, spec.failure = beforeEach.Run()
|
s, f := beforeEach.Run()
|
||||||
if spec.state != types.SpecStatePassed {
|
spec.failure = f
|
||||||
|
spec.setState(s)
|
||||||
|
if spec.getState() != types.SpecStatePassed {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -156,15 +190,19 @@ func (spec *Spec) runSample(sample int, writer io.Writer) {
|
||||||
for _, container := range spec.containers {
|
for _, container := range spec.containers {
|
||||||
for _, justBeforeEach := range container.SetupNodesOfType(types.SpecComponentTypeJustBeforeEach) {
|
for _, justBeforeEach := range container.SetupNodesOfType(types.SpecComponentTypeJustBeforeEach) {
|
||||||
spec.announceSetupNode(writer, "JustBeforeEach", container, justBeforeEach)
|
spec.announceSetupNode(writer, "JustBeforeEach", container, justBeforeEach)
|
||||||
spec.state, spec.failure = justBeforeEach.Run()
|
s, f := justBeforeEach.Run()
|
||||||
if spec.state != types.SpecStatePassed {
|
spec.failure = f
|
||||||
|
spec.setState(s)
|
||||||
|
if spec.getState() != types.SpecStatePassed {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
spec.announceSubject(writer, spec.subject)
|
spec.announceSubject(writer, spec.subject)
|
||||||
spec.state, spec.failure = spec.subject.Run()
|
s, f := spec.subject.Run()
|
||||||
|
spec.failure = f
|
||||||
|
spec.setState(s)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (spec *Spec) announceSetupNode(writer io.Writer, nodeType string, container *containernode.ContainerNode, setupNode leafnodes.BasicNode) {
|
func (spec *Spec) announceSetupNode(writer io.Writer, nodeType string, container *containernode.ContainerNode, setupNode leafnodes.BasicNode) {
|
||||||
|
|
41
vendor/github.com/onsi/ginkgo/internal/spec/specs.go
generated
vendored
41
vendor/github.com/onsi/ginkgo/internal/spec/specs.go
generated
vendored
|
@ -7,15 +7,14 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
type Specs struct {
|
type Specs struct {
|
||||||
specs []*Spec
|
specs []*Spec
|
||||||
numberOfOriginalSpecs int
|
hasProgrammaticFocus bool
|
||||||
hasProgrammaticFocus bool
|
RegexScansFilePath bool
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewSpecs(specs []*Spec) *Specs {
|
func NewSpecs(specs []*Spec) *Specs {
|
||||||
return &Specs{
|
return &Specs{
|
||||||
specs: specs,
|
specs: specs,
|
||||||
numberOfOriginalSpecs: len(specs),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -23,10 +22,6 @@ func (e *Specs) Specs() []*Spec {
|
||||||
return e.specs
|
return e.specs
|
||||||
}
|
}
|
||||||
|
|
||||||
func (e *Specs) NumberOfOriginalSpecs() int {
|
|
||||||
return e.numberOfOriginalSpecs
|
|
||||||
}
|
|
||||||
|
|
||||||
func (e *Specs) HasProgrammaticFocus() bool {
|
func (e *Specs) HasProgrammaticFocus() bool {
|
||||||
return e.hasProgrammaticFocus
|
return e.hasProgrammaticFocus
|
||||||
}
|
}
|
||||||
|
@ -45,7 +40,7 @@ func (e *Specs) ApplyFocus(description string, focusString string, skipString st
|
||||||
if focusString == "" && skipString == "" {
|
if focusString == "" && skipString == "" {
|
||||||
e.applyProgrammaticFocus()
|
e.applyProgrammaticFocus()
|
||||||
} else {
|
} else {
|
||||||
e.applyRegExpFocus(description, focusString, skipString)
|
e.applyRegExpFocusAndSkip(description, focusString, skipString)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -67,12 +62,27 @@ func (e *Specs) applyProgrammaticFocus() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (e *Specs) applyRegExpFocus(description string, focusString string, skipString string) {
|
// toMatch returns a byte[] to be used by regex matchers. When adding new behaviours to the matching function,
|
||||||
|
// this is the place which we append to.
|
||||||
|
func (e *Specs) toMatch(description string, spec *Spec) []byte {
|
||||||
|
if e.RegexScansFilePath {
|
||||||
|
return []byte(
|
||||||
|
description + " " +
|
||||||
|
spec.ConcatenatedString() + " " +
|
||||||
|
spec.subject.CodeLocation().FileName)
|
||||||
|
} else {
|
||||||
|
return []byte(
|
||||||
|
description + " " +
|
||||||
|
spec.ConcatenatedString())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (e *Specs) applyRegExpFocusAndSkip(description string, focusString string, skipString string) {
|
||||||
for _, spec := range e.specs {
|
for _, spec := range e.specs {
|
||||||
matchesFocus := true
|
matchesFocus := true
|
||||||
matchesSkip := false
|
matchesSkip := false
|
||||||
|
|
||||||
toMatch := []byte(description + " " + spec.ConcatenatedString())
|
toMatch := e.toMatch(description, spec)
|
||||||
|
|
||||||
if focusString != "" {
|
if focusString != "" {
|
||||||
focusFilter := regexp.MustCompile(focusString)
|
focusFilter := regexp.MustCompile(focusString)
|
||||||
|
@ -98,15 +108,6 @@ func (e *Specs) SkipMeasurements() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (e *Specs) TrimForParallelization(total int, node int) {
|
|
||||||
startIndex, count := ParallelizedIndexRange(len(e.specs), total, node)
|
|
||||||
if count == 0 {
|
|
||||||
e.specs = make([]*Spec, 0)
|
|
||||||
} else {
|
|
||||||
e.specs = e.specs[startIndex : startIndex+count]
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
//sort.Interface
|
//sort.Interface
|
||||||
|
|
||||||
func (e *Specs) Len() int {
|
func (e *Specs) Len() int {
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
package spec
|
package spec_iterator
|
||||||
|
|
||||||
func ParallelizedIndexRange(length int, parallelTotal int, parallelNode int) (startIndex int, count int) {
|
func ParallelizedIndexRange(length int, parallelTotal int, parallelNode int) (startIndex int, count int) {
|
||||||
if length == 0 {
|
if length == 0 {
|
59
vendor/github.com/onsi/ginkgo/internal/spec_iterator/parallel_spec_iterator.go
generated
vendored
Normal file
59
vendor/github.com/onsi/ginkgo/internal/spec_iterator/parallel_spec_iterator.go
generated
vendored
Normal file
|
@ -0,0 +1,59 @@
|
||||||
|
package spec_iterator
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
"net/http"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/internal/spec"
|
||||||
|
)
|
||||||
|
|
||||||
|
type ParallelIterator struct {
|
||||||
|
specs []*spec.Spec
|
||||||
|
host string
|
||||||
|
client *http.Client
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewParallelIterator(specs []*spec.Spec, host string) *ParallelIterator {
|
||||||
|
return &ParallelIterator{
|
||||||
|
specs: specs,
|
||||||
|
host: host,
|
||||||
|
client: &http.Client{},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *ParallelIterator) Next() (*spec.Spec, error) {
|
||||||
|
resp, err := s.client.Get(s.host + "/counter")
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
if resp.StatusCode != http.StatusOK {
|
||||||
|
return nil, fmt.Errorf("unexpected status code %d", resp.StatusCode)
|
||||||
|
}
|
||||||
|
|
||||||
|
var counter Counter
|
||||||
|
err = json.NewDecoder(resp.Body).Decode(&counter)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if counter.Index >= len(s.specs) {
|
||||||
|
return nil, ErrClosed
|
||||||
|
}
|
||||||
|
|
||||||
|
return s.specs[counter.Index], nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *ParallelIterator) NumberOfSpecsPriorToIteration() int {
|
||||||
|
return len(s.specs)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *ParallelIterator) NumberOfSpecsToProcessIfKnown() (int, bool) {
|
||||||
|
return -1, false
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *ParallelIterator) NumberOfSpecsThatWillBeRunIfKnown() (int, bool) {
|
||||||
|
return -1, false
|
||||||
|
}
|
45
vendor/github.com/onsi/ginkgo/internal/spec_iterator/serial_spec_iterator.go
generated
vendored
Normal file
45
vendor/github.com/onsi/ginkgo/internal/spec_iterator/serial_spec_iterator.go
generated
vendored
Normal file
|
@ -0,0 +1,45 @@
|
||||||
|
package spec_iterator
|
||||||
|
|
||||||
|
import (
|
||||||
|
"github.com/onsi/ginkgo/internal/spec"
|
||||||
|
)
|
||||||
|
|
||||||
|
type SerialIterator struct {
|
||||||
|
specs []*spec.Spec
|
||||||
|
index int
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewSerialIterator(specs []*spec.Spec) *SerialIterator {
|
||||||
|
return &SerialIterator{
|
||||||
|
specs: specs,
|
||||||
|
index: 0,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *SerialIterator) Next() (*spec.Spec, error) {
|
||||||
|
if s.index >= len(s.specs) {
|
||||||
|
return nil, ErrClosed
|
||||||
|
}
|
||||||
|
|
||||||
|
spec := s.specs[s.index]
|
||||||
|
s.index += 1
|
||||||
|
return spec, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *SerialIterator) NumberOfSpecsPriorToIteration() int {
|
||||||
|
return len(s.specs)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *SerialIterator) NumberOfSpecsToProcessIfKnown() (int, bool) {
|
||||||
|
return len(s.specs), true
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *SerialIterator) NumberOfSpecsThatWillBeRunIfKnown() (int, bool) {
|
||||||
|
count := 0
|
||||||
|
for _, s := range s.specs {
|
||||||
|
if !s.Skipped() && !s.Pending() {
|
||||||
|
count += 1
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return count, true
|
||||||
|
}
|
47
vendor/github.com/onsi/ginkgo/internal/spec_iterator/sharded_parallel_spec_iterator.go
generated
vendored
Normal file
47
vendor/github.com/onsi/ginkgo/internal/spec_iterator/sharded_parallel_spec_iterator.go
generated
vendored
Normal file
|
@ -0,0 +1,47 @@
|
||||||
|
package spec_iterator
|
||||||
|
|
||||||
|
import "github.com/onsi/ginkgo/internal/spec"
|
||||||
|
|
||||||
|
type ShardedParallelIterator struct {
|
||||||
|
specs []*spec.Spec
|
||||||
|
index int
|
||||||
|
maxIndex int
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewShardedParallelIterator(specs []*spec.Spec, total int, node int) *ShardedParallelIterator {
|
||||||
|
startIndex, count := ParallelizedIndexRange(len(specs), total, node)
|
||||||
|
|
||||||
|
return &ShardedParallelIterator{
|
||||||
|
specs: specs,
|
||||||
|
index: startIndex,
|
||||||
|
maxIndex: startIndex + count,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *ShardedParallelIterator) Next() (*spec.Spec, error) {
|
||||||
|
if s.index >= s.maxIndex {
|
||||||
|
return nil, ErrClosed
|
||||||
|
}
|
||||||
|
|
||||||
|
spec := s.specs[s.index]
|
||||||
|
s.index += 1
|
||||||
|
return spec, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *ShardedParallelIterator) NumberOfSpecsPriorToIteration() int {
|
||||||
|
return len(s.specs)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *ShardedParallelIterator) NumberOfSpecsToProcessIfKnown() (int, bool) {
|
||||||
|
return s.maxIndex - s.index, true
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *ShardedParallelIterator) NumberOfSpecsThatWillBeRunIfKnown() (int, bool) {
|
||||||
|
count := 0
|
||||||
|
for i := s.index; i < s.maxIndex; i += 1 {
|
||||||
|
if !s.specs[i].Skipped() && !s.specs[i].Pending() {
|
||||||
|
count += 1
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return count, true
|
||||||
|
}
|
20
vendor/github.com/onsi/ginkgo/internal/spec_iterator/spec_iterator.go
generated
vendored
Normal file
20
vendor/github.com/onsi/ginkgo/internal/spec_iterator/spec_iterator.go
generated
vendored
Normal file
|
@ -0,0 +1,20 @@
|
||||||
|
package spec_iterator
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/internal/spec"
|
||||||
|
)
|
||||||
|
|
||||||
|
var ErrClosed = errors.New("no more specs to run")
|
||||||
|
|
||||||
|
type SpecIterator interface {
|
||||||
|
Next() (*spec.Spec, error)
|
||||||
|
NumberOfSpecsPriorToIteration() int
|
||||||
|
NumberOfSpecsToProcessIfKnown() (int, bool)
|
||||||
|
NumberOfSpecsThatWillBeRunIfKnown() (int, bool)
|
||||||
|
}
|
||||||
|
|
||||||
|
type Counter struct {
|
||||||
|
Index int `json:"index"`
|
||||||
|
}
|
141
vendor/github.com/onsi/ginkgo/internal/specrunner/spec_runner.go
generated
vendored
141
vendor/github.com/onsi/ginkgo/internal/specrunner/spec_runner.go
generated
vendored
|
@ -7,6 +7,8 @@ import (
|
||||||
"sync"
|
"sync"
|
||||||
"syscall"
|
"syscall"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/internal/spec_iterator"
|
||||||
|
|
||||||
"github.com/onsi/ginkgo/config"
|
"github.com/onsi/ginkgo/config"
|
||||||
"github.com/onsi/ginkgo/internal/leafnodes"
|
"github.com/onsi/ginkgo/internal/leafnodes"
|
||||||
"github.com/onsi/ginkgo/internal/spec"
|
"github.com/onsi/ginkgo/internal/spec"
|
||||||
|
@ -20,7 +22,7 @@ import (
|
||||||
type SpecRunner struct {
|
type SpecRunner struct {
|
||||||
description string
|
description string
|
||||||
beforeSuiteNode leafnodes.SuiteNode
|
beforeSuiteNode leafnodes.SuiteNode
|
||||||
specs *spec.Specs
|
iterator spec_iterator.SpecIterator
|
||||||
afterSuiteNode leafnodes.SuiteNode
|
afterSuiteNode leafnodes.SuiteNode
|
||||||
reporters []reporters.Reporter
|
reporters []reporters.Reporter
|
||||||
startTime time.Time
|
startTime time.Time
|
||||||
|
@ -29,14 +31,15 @@ type SpecRunner struct {
|
||||||
writer Writer.WriterInterface
|
writer Writer.WriterInterface
|
||||||
config config.GinkgoConfigType
|
config config.GinkgoConfigType
|
||||||
interrupted bool
|
interrupted bool
|
||||||
|
processedSpecs []*spec.Spec
|
||||||
lock *sync.Mutex
|
lock *sync.Mutex
|
||||||
}
|
}
|
||||||
|
|
||||||
func New(description string, beforeSuiteNode leafnodes.SuiteNode, specs *spec.Specs, afterSuiteNode leafnodes.SuiteNode, reporters []reporters.Reporter, writer Writer.WriterInterface, config config.GinkgoConfigType) *SpecRunner {
|
func New(description string, beforeSuiteNode leafnodes.SuiteNode, iterator spec_iterator.SpecIterator, afterSuiteNode leafnodes.SuiteNode, reporters []reporters.Reporter, writer Writer.WriterInterface, config config.GinkgoConfigType) *SpecRunner {
|
||||||
return &SpecRunner{
|
return &SpecRunner{
|
||||||
description: description,
|
description: description,
|
||||||
beforeSuiteNode: beforeSuiteNode,
|
beforeSuiteNode: beforeSuiteNode,
|
||||||
specs: specs,
|
iterator: iterator,
|
||||||
afterSuiteNode: afterSuiteNode,
|
afterSuiteNode: afterSuiteNode,
|
||||||
reporters: reporters,
|
reporters: reporters,
|
||||||
writer: writer,
|
writer: writer,
|
||||||
|
@ -53,7 +56,9 @@ func (runner *SpecRunner) Run() bool {
|
||||||
}
|
}
|
||||||
|
|
||||||
runner.reportSuiteWillBegin()
|
runner.reportSuiteWillBegin()
|
||||||
go runner.registerForInterrupts()
|
signalRegistered := make(chan struct{})
|
||||||
|
go runner.registerForInterrupts(signalRegistered)
|
||||||
|
<-signalRegistered
|
||||||
|
|
||||||
suitePassed := runner.runBeforeSuite()
|
suitePassed := runner.runBeforeSuite()
|
||||||
|
|
||||||
|
@ -79,7 +84,18 @@ func (runner *SpecRunner) performDryRun() {
|
||||||
runner.reportBeforeSuite(summary)
|
runner.reportBeforeSuite(summary)
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, spec := range runner.specs.Specs() {
|
for {
|
||||||
|
spec, err := runner.iterator.Next()
|
||||||
|
if err == spec_iterator.ErrClosed {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
fmt.Println("failed to iterate over tests:\n" + err.Error())
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
runner.processedSpecs = append(runner.processedSpecs, spec)
|
||||||
|
|
||||||
summary := spec.Summary(runner.suiteID)
|
summary := spec.Summary(runner.suiteID)
|
||||||
runner.reportSpecWillRun(summary)
|
runner.reportSpecWillRun(summary)
|
||||||
if summary.State == types.SpecStateInvalid {
|
if summary.State == types.SpecStateInvalid {
|
||||||
|
@ -130,28 +146,39 @@ func (runner *SpecRunner) runAfterSuite() bool {
|
||||||
func (runner *SpecRunner) runSpecs() bool {
|
func (runner *SpecRunner) runSpecs() bool {
|
||||||
suiteFailed := false
|
suiteFailed := false
|
||||||
skipRemainingSpecs := false
|
skipRemainingSpecs := false
|
||||||
for _, spec := range runner.specs.Specs() {
|
for {
|
||||||
|
spec, err := runner.iterator.Next()
|
||||||
|
if err == spec_iterator.ErrClosed {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
fmt.Println("failed to iterate over tests:\n" + err.Error())
|
||||||
|
suiteFailed = true
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
runner.processedSpecs = append(runner.processedSpecs, spec)
|
||||||
|
|
||||||
if runner.wasInterrupted() {
|
if runner.wasInterrupted() {
|
||||||
return suiteFailed
|
break
|
||||||
}
|
}
|
||||||
if skipRemainingSpecs {
|
if skipRemainingSpecs {
|
||||||
spec.Skip()
|
spec.Skip()
|
||||||
}
|
}
|
||||||
runner.reportSpecWillRun(spec.Summary(runner.suiteID))
|
|
||||||
|
|
||||||
if !spec.Skipped() && !spec.Pending() {
|
if !spec.Skipped() && !spec.Pending() {
|
||||||
runner.runningSpec = spec
|
if passed := runner.runSpec(spec); !passed {
|
||||||
spec.Run(runner.writer)
|
|
||||||
runner.runningSpec = nil
|
|
||||||
if spec.Failed() {
|
|
||||||
suiteFailed = true
|
suiteFailed = true
|
||||||
}
|
}
|
||||||
} else if spec.Pending() && runner.config.FailOnPending {
|
} else if spec.Pending() && runner.config.FailOnPending {
|
||||||
|
runner.reportSpecWillRun(spec.Summary(runner.suiteID))
|
||||||
suiteFailed = true
|
suiteFailed = true
|
||||||
|
runner.reportSpecDidComplete(spec.Summary(runner.suiteID), spec.Failed())
|
||||||
|
} else {
|
||||||
|
runner.reportSpecWillRun(spec.Summary(runner.suiteID))
|
||||||
|
runner.reportSpecDidComplete(spec.Summary(runner.suiteID), spec.Failed())
|
||||||
}
|
}
|
||||||
|
|
||||||
runner.reportSpecDidComplete(spec.Summary(runner.suiteID), spec.Failed())
|
|
||||||
|
|
||||||
if spec.Failed() && runner.config.FailFast {
|
if spec.Failed() && runner.config.FailFast {
|
||||||
skipRemainingSpecs = true
|
skipRemainingSpecs = true
|
||||||
}
|
}
|
||||||
|
@ -160,6 +187,26 @@ func (runner *SpecRunner) runSpecs() bool {
|
||||||
return !suiteFailed
|
return !suiteFailed
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (runner *SpecRunner) runSpec(spec *spec.Spec) (passed bool) {
|
||||||
|
maxAttempts := 1
|
||||||
|
if runner.config.FlakeAttempts > 0 {
|
||||||
|
// uninitialized configs count as 1
|
||||||
|
maxAttempts = runner.config.FlakeAttempts
|
||||||
|
}
|
||||||
|
|
||||||
|
for i := 0; i < maxAttempts; i++ {
|
||||||
|
runner.reportSpecWillRun(spec.Summary(runner.suiteID))
|
||||||
|
runner.runningSpec = spec
|
||||||
|
spec.Run(runner.writer)
|
||||||
|
runner.runningSpec = nil
|
||||||
|
runner.reportSpecDidComplete(spec.Summary(runner.suiteID), spec.Failed())
|
||||||
|
if !spec.Failed() {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
func (runner *SpecRunner) CurrentSpecSummary() (*types.SpecSummary, bool) {
|
func (runner *SpecRunner) CurrentSpecSummary() (*types.SpecSummary, bool) {
|
||||||
if runner.runningSpec == nil {
|
if runner.runningSpec == nil {
|
||||||
return nil, false
|
return nil, false
|
||||||
|
@ -168,9 +215,10 @@ func (runner *SpecRunner) CurrentSpecSummary() (*types.SpecSummary, bool) {
|
||||||
return runner.runningSpec.Summary(runner.suiteID), true
|
return runner.runningSpec.Summary(runner.suiteID), true
|
||||||
}
|
}
|
||||||
|
|
||||||
func (runner *SpecRunner) registerForInterrupts() {
|
func (runner *SpecRunner) registerForInterrupts(signalRegistered chan struct{}) {
|
||||||
c := make(chan os.Signal, 1)
|
c := make(chan os.Signal, 1)
|
||||||
signal.Notify(c, os.Interrupt, syscall.SIGTERM)
|
signal.Notify(c, os.Interrupt, syscall.SIGTERM)
|
||||||
|
close(signalRegistered)
|
||||||
|
|
||||||
<-c
|
<-c
|
||||||
signal.Stop(c)
|
signal.Stop(c)
|
||||||
|
@ -225,7 +273,7 @@ func (runner *SpecRunner) wasInterrupted() bool {
|
||||||
|
|
||||||
func (runner *SpecRunner) reportSuiteWillBegin() {
|
func (runner *SpecRunner) reportSuiteWillBegin() {
|
||||||
runner.startTime = time.Now()
|
runner.startTime = time.Now()
|
||||||
summary := runner.summary(true)
|
summary := runner.suiteWillBeginSummary()
|
||||||
for _, reporter := range runner.reporters {
|
for _, reporter := range runner.reporters {
|
||||||
reporter.SpecSuiteWillBegin(runner.config, summary)
|
reporter.SpecSuiteWillBegin(runner.config, summary)
|
||||||
}
|
}
|
||||||
|
@ -252,6 +300,9 @@ func (runner *SpecRunner) reportSpecWillRun(summary *types.SpecSummary) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (runner *SpecRunner) reportSpecDidComplete(summary *types.SpecSummary, failed bool) {
|
func (runner *SpecRunner) reportSpecDidComplete(summary *types.SpecSummary, failed bool) {
|
||||||
|
if failed && len(summary.CapturedOutput) == 0 {
|
||||||
|
summary.CapturedOutput = string(runner.writer.Bytes())
|
||||||
|
}
|
||||||
for i := len(runner.reporters) - 1; i >= 1; i-- {
|
for i := len(runner.reporters) - 1; i >= 1; i-- {
|
||||||
runner.reporters[i].SpecDidComplete(summary)
|
runner.reporters[i].SpecDidComplete(summary)
|
||||||
}
|
}
|
||||||
|
@ -264,17 +315,17 @@ func (runner *SpecRunner) reportSpecDidComplete(summary *types.SpecSummary, fail
|
||||||
}
|
}
|
||||||
|
|
||||||
func (runner *SpecRunner) reportSuiteDidEnd(success bool) {
|
func (runner *SpecRunner) reportSuiteDidEnd(success bool) {
|
||||||
summary := runner.summary(success)
|
summary := runner.suiteDidEndSummary(success)
|
||||||
summary.RunTime = time.Since(runner.startTime)
|
summary.RunTime = time.Since(runner.startTime)
|
||||||
for _, reporter := range runner.reporters {
|
for _, reporter := range runner.reporters {
|
||||||
reporter.SpecSuiteDidEnd(summary)
|
reporter.SpecSuiteDidEnd(summary)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (runner *SpecRunner) countSpecsSatisfying(filter func(ex *spec.Spec) bool) (count int) {
|
func (runner *SpecRunner) countSpecsThatRanSatisfying(filter func(ex *spec.Spec) bool) (count int) {
|
||||||
count = 0
|
count = 0
|
||||||
|
|
||||||
for _, spec := range runner.specs.Specs() {
|
for _, spec := range runner.processedSpecs {
|
||||||
if filter(spec) {
|
if filter(spec) {
|
||||||
count++
|
count++
|
||||||
}
|
}
|
||||||
|
@ -283,28 +334,37 @@ func (runner *SpecRunner) countSpecsSatisfying(filter func(ex *spec.Spec) bool)
|
||||||
return count
|
return count
|
||||||
}
|
}
|
||||||
|
|
||||||
func (runner *SpecRunner) summary(success bool) *types.SuiteSummary {
|
func (runner *SpecRunner) suiteDidEndSummary(success bool) *types.SuiteSummary {
|
||||||
numberOfSpecsThatWillBeRun := runner.countSpecsSatisfying(func(ex *spec.Spec) bool {
|
numberOfSpecsThatWillBeRun := runner.countSpecsThatRanSatisfying(func(ex *spec.Spec) bool {
|
||||||
return !ex.Skipped() && !ex.Pending()
|
return !ex.Skipped() && !ex.Pending()
|
||||||
})
|
})
|
||||||
|
|
||||||
numberOfPendingSpecs := runner.countSpecsSatisfying(func(ex *spec.Spec) bool {
|
numberOfPendingSpecs := runner.countSpecsThatRanSatisfying(func(ex *spec.Spec) bool {
|
||||||
return ex.Pending()
|
return ex.Pending()
|
||||||
})
|
})
|
||||||
|
|
||||||
numberOfSkippedSpecs := runner.countSpecsSatisfying(func(ex *spec.Spec) bool {
|
numberOfSkippedSpecs := runner.countSpecsThatRanSatisfying(func(ex *spec.Spec) bool {
|
||||||
return ex.Skipped()
|
return ex.Skipped()
|
||||||
})
|
})
|
||||||
|
|
||||||
numberOfPassedSpecs := runner.countSpecsSatisfying(func(ex *spec.Spec) bool {
|
numberOfPassedSpecs := runner.countSpecsThatRanSatisfying(func(ex *spec.Spec) bool {
|
||||||
return ex.Passed()
|
return ex.Passed()
|
||||||
})
|
})
|
||||||
|
|
||||||
numberOfFailedSpecs := runner.countSpecsSatisfying(func(ex *spec.Spec) bool {
|
numberOfFlakedSpecs := runner.countSpecsThatRanSatisfying(func(ex *spec.Spec) bool {
|
||||||
|
return ex.Flaked()
|
||||||
|
})
|
||||||
|
|
||||||
|
numberOfFailedSpecs := runner.countSpecsThatRanSatisfying(func(ex *spec.Spec) bool {
|
||||||
return ex.Failed()
|
return ex.Failed()
|
||||||
})
|
})
|
||||||
|
|
||||||
if runner.beforeSuiteNode != nil && !runner.beforeSuiteNode.Passed() && !runner.config.DryRun {
|
if runner.beforeSuiteNode != nil && !runner.beforeSuiteNode.Passed() && !runner.config.DryRun {
|
||||||
|
var known bool
|
||||||
|
numberOfSpecsThatWillBeRun, known = runner.iterator.NumberOfSpecsThatWillBeRunIfKnown()
|
||||||
|
if !known {
|
||||||
|
numberOfSpecsThatWillBeRun = runner.iterator.NumberOfSpecsPriorToIteration()
|
||||||
|
}
|
||||||
numberOfFailedSpecs = numberOfSpecsThatWillBeRun
|
numberOfFailedSpecs = numberOfSpecsThatWillBeRun
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -313,12 +373,39 @@ func (runner *SpecRunner) summary(success bool) *types.SuiteSummary {
|
||||||
SuiteSucceeded: success,
|
SuiteSucceeded: success,
|
||||||
SuiteID: runner.suiteID,
|
SuiteID: runner.suiteID,
|
||||||
|
|
||||||
NumberOfSpecsBeforeParallelization: runner.specs.NumberOfOriginalSpecs(),
|
NumberOfSpecsBeforeParallelization: runner.iterator.NumberOfSpecsPriorToIteration(),
|
||||||
NumberOfTotalSpecs: len(runner.specs.Specs()),
|
NumberOfTotalSpecs: len(runner.processedSpecs),
|
||||||
NumberOfSpecsThatWillBeRun: numberOfSpecsThatWillBeRun,
|
NumberOfSpecsThatWillBeRun: numberOfSpecsThatWillBeRun,
|
||||||
NumberOfPendingSpecs: numberOfPendingSpecs,
|
NumberOfPendingSpecs: numberOfPendingSpecs,
|
||||||
NumberOfSkippedSpecs: numberOfSkippedSpecs,
|
NumberOfSkippedSpecs: numberOfSkippedSpecs,
|
||||||
NumberOfPassedSpecs: numberOfPassedSpecs,
|
NumberOfPassedSpecs: numberOfPassedSpecs,
|
||||||
NumberOfFailedSpecs: numberOfFailedSpecs,
|
NumberOfFailedSpecs: numberOfFailedSpecs,
|
||||||
|
NumberOfFlakedSpecs: numberOfFlakedSpecs,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (runner *SpecRunner) suiteWillBeginSummary() *types.SuiteSummary {
|
||||||
|
numTotal, known := runner.iterator.NumberOfSpecsToProcessIfKnown()
|
||||||
|
if !known {
|
||||||
|
numTotal = -1
|
||||||
|
}
|
||||||
|
|
||||||
|
numToRun, known := runner.iterator.NumberOfSpecsThatWillBeRunIfKnown()
|
||||||
|
if !known {
|
||||||
|
numToRun = -1
|
||||||
|
}
|
||||||
|
|
||||||
|
return &types.SuiteSummary{
|
||||||
|
SuiteDescription: runner.description,
|
||||||
|
SuiteID: runner.suiteID,
|
||||||
|
|
||||||
|
NumberOfSpecsBeforeParallelization: runner.iterator.NumberOfSpecsPriorToIteration(),
|
||||||
|
NumberOfTotalSpecs: numTotal,
|
||||||
|
NumberOfSpecsThatWillBeRun: numToRun,
|
||||||
|
NumberOfPendingSpecs: -1,
|
||||||
|
NumberOfSkippedSpecs: -1,
|
||||||
|
NumberOfPassedSpecs: -1,
|
||||||
|
NumberOfFailedSpecs: -1,
|
||||||
|
NumberOfFlakedSpecs: -1,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
34
vendor/github.com/onsi/ginkgo/internal/suite/suite.go
generated
vendored
34
vendor/github.com/onsi/ginkgo/internal/suite/suite.go
generated
vendored
|
@ -2,8 +2,11 @@ package suite
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"math/rand"
|
"math/rand"
|
||||||
|
"net/http"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/internal/spec_iterator"
|
||||||
|
|
||||||
"github.com/onsi/ginkgo/config"
|
"github.com/onsi/ginkgo/config"
|
||||||
"github.com/onsi/ginkgo/internal/containernode"
|
"github.com/onsi/ginkgo/internal/containernode"
|
||||||
"github.com/onsi/ginkgo/internal/failer"
|
"github.com/onsi/ginkgo/internal/failer"
|
||||||
|
@ -52,18 +55,18 @@ func (suite *Suite) Run(t ginkgoTestingT, description string, reporters []report
|
||||||
|
|
||||||
r := rand.New(rand.NewSource(config.RandomSeed))
|
r := rand.New(rand.NewSource(config.RandomSeed))
|
||||||
suite.topLevelContainer.Shuffle(r)
|
suite.topLevelContainer.Shuffle(r)
|
||||||
specs := suite.generateSpecs(description, config)
|
iterator, hasProgrammaticFocus := suite.generateSpecsIterator(description, config)
|
||||||
suite.runner = specrunner.New(description, suite.beforeSuiteNode, specs, suite.afterSuiteNode, reporters, writer, config)
|
suite.runner = specrunner.New(description, suite.beforeSuiteNode, iterator, suite.afterSuiteNode, reporters, writer, config)
|
||||||
|
|
||||||
suite.running = true
|
suite.running = true
|
||||||
success := suite.runner.Run()
|
success := suite.runner.Run()
|
||||||
if !success {
|
if !success {
|
||||||
t.Fail()
|
t.Fail()
|
||||||
}
|
}
|
||||||
return success, specs.HasProgrammaticFocus()
|
return success, hasProgrammaticFocus
|
||||||
}
|
}
|
||||||
|
|
||||||
func (suite *Suite) generateSpecs(description string, config config.GinkgoConfigType) *spec.Specs {
|
func (suite *Suite) generateSpecsIterator(description string, config config.GinkgoConfigType) (spec_iterator.SpecIterator, bool) {
|
||||||
specsSlice := []*spec.Spec{}
|
specsSlice := []*spec.Spec{}
|
||||||
suite.topLevelContainer.BackPropagateProgrammaticFocus()
|
suite.topLevelContainer.BackPropagateProgrammaticFocus()
|
||||||
for _, collatedNodes := range suite.topLevelContainer.Collate() {
|
for _, collatedNodes := range suite.topLevelContainer.Collate() {
|
||||||
|
@ -71,6 +74,7 @@ func (suite *Suite) generateSpecs(description string, config config.GinkgoConfig
|
||||||
}
|
}
|
||||||
|
|
||||||
specs := spec.NewSpecs(specsSlice)
|
specs := spec.NewSpecs(specsSlice)
|
||||||
|
specs.RegexScansFilePath = config.RegexScansFilePath
|
||||||
|
|
||||||
if config.RandomizeAllSpecs {
|
if config.RandomizeAllSpecs {
|
||||||
specs.Shuffle(rand.New(rand.NewSource(config.RandomSeed)))
|
specs.Shuffle(rand.New(rand.NewSource(config.RandomSeed)))
|
||||||
|
@ -82,11 +86,19 @@ func (suite *Suite) generateSpecs(description string, config config.GinkgoConfig
|
||||||
specs.SkipMeasurements()
|
specs.SkipMeasurements()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var iterator spec_iterator.SpecIterator
|
||||||
|
|
||||||
if config.ParallelTotal > 1 {
|
if config.ParallelTotal > 1 {
|
||||||
specs.TrimForParallelization(config.ParallelTotal, config.ParallelNode)
|
iterator = spec_iterator.NewParallelIterator(specs.Specs(), config.SyncHost)
|
||||||
|
resp, err := http.Get(config.SyncHost + "/has-counter")
|
||||||
|
if err != nil || resp.StatusCode != http.StatusOK {
|
||||||
|
iterator = spec_iterator.NewShardedParallelIterator(specs.Specs(), config.ParallelTotal, config.ParallelNode)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
iterator = spec_iterator.NewSerialIterator(specs.Specs())
|
||||||
}
|
}
|
||||||
|
|
||||||
return specs
|
return iterator, specs.HasProgrammaticFocus()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (suite *Suite) CurrentRunningSpecSummary() (*types.SpecSummary, bool) {
|
func (suite *Suite) CurrentRunningSpecSummary() (*types.SpecSummary, bool) {
|
||||||
|
@ -137,35 +149,35 @@ func (suite *Suite) PushContainerNode(text string, body func(), flag types.FlagT
|
||||||
|
|
||||||
func (suite *Suite) PushItNode(text string, body interface{}, flag types.FlagType, codeLocation types.CodeLocation, timeout time.Duration) {
|
func (suite *Suite) PushItNode(text string, body interface{}, flag types.FlagType, codeLocation types.CodeLocation, timeout time.Duration) {
|
||||||
if suite.running {
|
if suite.running {
|
||||||
suite.failer.Fail("You may only call It from within a Describe or Context", codeLocation)
|
suite.failer.Fail("You may only call It from within a Describe, Context or When", codeLocation)
|
||||||
}
|
}
|
||||||
suite.currentContainer.PushSubjectNode(leafnodes.NewItNode(text, body, flag, codeLocation, timeout, suite.failer, suite.containerIndex))
|
suite.currentContainer.PushSubjectNode(leafnodes.NewItNode(text, body, flag, codeLocation, timeout, suite.failer, suite.containerIndex))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (suite *Suite) PushMeasureNode(text string, body interface{}, flag types.FlagType, codeLocation types.CodeLocation, samples int) {
|
func (suite *Suite) PushMeasureNode(text string, body interface{}, flag types.FlagType, codeLocation types.CodeLocation, samples int) {
|
||||||
if suite.running {
|
if suite.running {
|
||||||
suite.failer.Fail("You may only call Measure from within a Describe or Context", codeLocation)
|
suite.failer.Fail("You may only call Measure from within a Describe, Context or When", codeLocation)
|
||||||
}
|
}
|
||||||
suite.currentContainer.PushSubjectNode(leafnodes.NewMeasureNode(text, body, flag, codeLocation, samples, suite.failer, suite.containerIndex))
|
suite.currentContainer.PushSubjectNode(leafnodes.NewMeasureNode(text, body, flag, codeLocation, samples, suite.failer, suite.containerIndex))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (suite *Suite) PushBeforeEachNode(body interface{}, codeLocation types.CodeLocation, timeout time.Duration) {
|
func (suite *Suite) PushBeforeEachNode(body interface{}, codeLocation types.CodeLocation, timeout time.Duration) {
|
||||||
if suite.running {
|
if suite.running {
|
||||||
suite.failer.Fail("You may only call BeforeEach from within a Describe or Context", codeLocation)
|
suite.failer.Fail("You may only call BeforeEach from within a Describe, Context or When", codeLocation)
|
||||||
}
|
}
|
||||||
suite.currentContainer.PushSetupNode(leafnodes.NewBeforeEachNode(body, codeLocation, timeout, suite.failer, suite.containerIndex))
|
suite.currentContainer.PushSetupNode(leafnodes.NewBeforeEachNode(body, codeLocation, timeout, suite.failer, suite.containerIndex))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (suite *Suite) PushJustBeforeEachNode(body interface{}, codeLocation types.CodeLocation, timeout time.Duration) {
|
func (suite *Suite) PushJustBeforeEachNode(body interface{}, codeLocation types.CodeLocation, timeout time.Duration) {
|
||||||
if suite.running {
|
if suite.running {
|
||||||
suite.failer.Fail("You may only call JustBeforeEach from within a Describe or Context", codeLocation)
|
suite.failer.Fail("You may only call JustBeforeEach from within a Describe, Context or When", codeLocation)
|
||||||
}
|
}
|
||||||
suite.currentContainer.PushSetupNode(leafnodes.NewJustBeforeEachNode(body, codeLocation, timeout, suite.failer, suite.containerIndex))
|
suite.currentContainer.PushSetupNode(leafnodes.NewJustBeforeEachNode(body, codeLocation, timeout, suite.failer, suite.containerIndex))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (suite *Suite) PushAfterEachNode(body interface{}, codeLocation types.CodeLocation, timeout time.Duration) {
|
func (suite *Suite) PushAfterEachNode(body interface{}, codeLocation types.CodeLocation, timeout time.Duration) {
|
||||||
if suite.running {
|
if suite.running {
|
||||||
suite.failer.Fail("You may only call AfterEach from within a Describe or Context", codeLocation)
|
suite.failer.Fail("You may only call AfterEach from within a Describe, Context or When", codeLocation)
|
||||||
}
|
}
|
||||||
suite.currentContainer.PushSetupNode(leafnodes.NewAfterEachNode(body, codeLocation, timeout, suite.failer, suite.containerIndex))
|
suite.currentContainer.PushSetupNode(leafnodes.NewAfterEachNode(body, codeLocation, timeout, suite.failer, suite.containerIndex))
|
||||||
}
|
}
|
||||||
|
|
4
vendor/github.com/onsi/ginkgo/internal/testingtproxy/testing_t_proxy.go
generated
vendored
4
vendor/github.com/onsi/ginkgo/internal/testingtproxy/testing_t_proxy.go
generated
vendored
|
@ -50,7 +50,7 @@ func (t *ginkgoTestingTProxy) Log(args ...interface{}) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (t *ginkgoTestingTProxy) Logf(format string, args ...interface{}) {
|
func (t *ginkgoTestingTProxy) Logf(format string, args ...interface{}) {
|
||||||
fmt.Fprintf(t.writer, format, args...)
|
t.Log(fmt.Sprintf(format, args...))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (t *ginkgoTestingTProxy) Failed() bool {
|
func (t *ginkgoTestingTProxy) Failed() bool {
|
||||||
|
@ -65,7 +65,7 @@ func (t *ginkgoTestingTProxy) Skip(args ...interface{}) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (t *ginkgoTestingTProxy) Skipf(format string, args ...interface{}) {
|
func (t *ginkgoTestingTProxy) Skipf(format string, args ...interface{}) {
|
||||||
fmt.Printf(format, args...)
|
t.Skip(fmt.Sprintf(format, args...))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (t *ginkgoTestingTProxy) SkipNow() {
|
func (t *ginkgoTestingTProxy) SkipNow() {
|
||||||
|
|
5
vendor/github.com/onsi/ginkgo/internal/writer/fake_writer.go
generated
vendored
5
vendor/github.com/onsi/ginkgo/internal/writer/fake_writer.go
generated
vendored
|
@ -26,6 +26,11 @@ func (writer *FakeGinkgoWriter) DumpOutWithHeader(header string) {
|
||||||
writer.EventStream = append(writer.EventStream, "DUMP_WITH_HEADER: "+header)
|
writer.EventStream = append(writer.EventStream, "DUMP_WITH_HEADER: "+header)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (writer *FakeGinkgoWriter) Bytes() []byte {
|
||||||
|
writer.EventStream = append(writer.EventStream, "BYTES")
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
func (writer *FakeGinkgoWriter) Write(data []byte) (n int, err error) {
|
func (writer *FakeGinkgoWriter) Write(data []byte) (n int, err error) {
|
||||||
return 0, nil
|
return 0, nil
|
||||||
}
|
}
|
||||||
|
|
30
vendor/github.com/onsi/ginkgo/internal/writer/writer.go
generated
vendored
30
vendor/github.com/onsi/ginkgo/internal/writer/writer.go
generated
vendored
|
@ -12,13 +12,15 @@ type WriterInterface interface {
|
||||||
Truncate()
|
Truncate()
|
||||||
DumpOut()
|
DumpOut()
|
||||||
DumpOutWithHeader(header string)
|
DumpOutWithHeader(header string)
|
||||||
|
Bytes() []byte
|
||||||
}
|
}
|
||||||
|
|
||||||
type Writer struct {
|
type Writer struct {
|
||||||
buffer *bytes.Buffer
|
buffer *bytes.Buffer
|
||||||
outWriter io.Writer
|
outWriter io.Writer
|
||||||
lock *sync.Mutex
|
lock *sync.Mutex
|
||||||
stream bool
|
stream bool
|
||||||
|
redirector io.Writer
|
||||||
}
|
}
|
||||||
|
|
||||||
func New(outWriter io.Writer) *Writer {
|
func New(outWriter io.Writer) *Writer {
|
||||||
|
@ -30,6 +32,10 @@ func New(outWriter io.Writer) *Writer {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (w *Writer) AndRedirectTo(writer io.Writer) {
|
||||||
|
w.redirector = writer
|
||||||
|
}
|
||||||
|
|
||||||
func (w *Writer) SetStream(stream bool) {
|
func (w *Writer) SetStream(stream bool) {
|
||||||
w.lock.Lock()
|
w.lock.Lock()
|
||||||
defer w.lock.Unlock()
|
defer w.lock.Unlock()
|
||||||
|
@ -40,11 +46,14 @@ func (w *Writer) Write(b []byte) (n int, err error) {
|
||||||
w.lock.Lock()
|
w.lock.Lock()
|
||||||
defer w.lock.Unlock()
|
defer w.lock.Unlock()
|
||||||
|
|
||||||
|
n, err = w.buffer.Write(b)
|
||||||
|
if w.redirector != nil {
|
||||||
|
w.redirector.Write(b)
|
||||||
|
}
|
||||||
if w.stream {
|
if w.stream {
|
||||||
return w.outWriter.Write(b)
|
return w.outWriter.Write(b)
|
||||||
} else {
|
|
||||||
return w.buffer.Write(b)
|
|
||||||
}
|
}
|
||||||
|
return n, err
|
||||||
}
|
}
|
||||||
|
|
||||||
func (w *Writer) Truncate() {
|
func (w *Writer) Truncate() {
|
||||||
|
@ -61,6 +70,15 @@ func (w *Writer) DumpOut() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (w *Writer) Bytes() []byte {
|
||||||
|
w.lock.Lock()
|
||||||
|
defer w.lock.Unlock()
|
||||||
|
b := w.buffer.Bytes()
|
||||||
|
copied := make([]byte, len(b))
|
||||||
|
copy(copied, b)
|
||||||
|
return copied
|
||||||
|
}
|
||||||
|
|
||||||
func (w *Writer) DumpOutWithHeader(header string) {
|
func (w *Writer) DumpOutWithHeader(header string) {
|
||||||
w.lock.Lock()
|
w.lock.Lock()
|
||||||
defer w.lock.Unlock()
|
defer w.lock.Unlock()
|
||||||
|
|
7
vendor/github.com/onsi/ginkgo/reporters/default_reporter.go
generated
vendored
7
vendor/github.com/onsi/ginkgo/reporters/default_reporter.go
generated
vendored
|
@ -29,9 +29,10 @@ func NewDefaultReporter(config config.DefaultReporterConfigType, stenographer st
|
||||||
func (reporter *DefaultReporter) SpecSuiteWillBegin(config config.GinkgoConfigType, summary *types.SuiteSummary) {
|
func (reporter *DefaultReporter) SpecSuiteWillBegin(config config.GinkgoConfigType, summary *types.SuiteSummary) {
|
||||||
reporter.stenographer.AnnounceSuite(summary.SuiteDescription, config.RandomSeed, config.RandomizeAllSpecs, reporter.config.Succinct)
|
reporter.stenographer.AnnounceSuite(summary.SuiteDescription, config.RandomSeed, config.RandomizeAllSpecs, reporter.config.Succinct)
|
||||||
if config.ParallelTotal > 1 {
|
if config.ParallelTotal > 1 {
|
||||||
reporter.stenographer.AnnounceParallelRun(config.ParallelNode, config.ParallelTotal, summary.NumberOfTotalSpecs, summary.NumberOfSpecsBeforeParallelization, reporter.config.Succinct)
|
reporter.stenographer.AnnounceParallelRun(config.ParallelNode, config.ParallelTotal, reporter.config.Succinct)
|
||||||
|
} else {
|
||||||
|
reporter.stenographer.AnnounceNumberOfSpecs(summary.NumberOfSpecsThatWillBeRun, summary.NumberOfTotalSpecs, reporter.config.Succinct)
|
||||||
}
|
}
|
||||||
reporter.stenographer.AnnounceNumberOfSpecs(summary.NumberOfSpecsThatWillBeRun, summary.NumberOfTotalSpecs, reporter.config.Succinct)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (reporter *DefaultReporter) BeforeSuiteDidRun(setupSummary *types.SetupSummary) {
|
func (reporter *DefaultReporter) BeforeSuiteDidRun(setupSummary *types.SetupSummary) {
|
||||||
|
@ -65,7 +66,7 @@ func (reporter *DefaultReporter) SpecDidComplete(specSummary *types.SpecSummary)
|
||||||
case types.SpecStatePending:
|
case types.SpecStatePending:
|
||||||
reporter.stenographer.AnnouncePendingSpec(specSummary, reporter.config.NoisyPendings && !reporter.config.Succinct)
|
reporter.stenographer.AnnouncePendingSpec(specSummary, reporter.config.NoisyPendings && !reporter.config.Succinct)
|
||||||
case types.SpecStateSkipped:
|
case types.SpecStateSkipped:
|
||||||
reporter.stenographer.AnnounceSkippedSpec(specSummary, reporter.config.Succinct, reporter.config.FullTrace)
|
reporter.stenographer.AnnounceSkippedSpec(specSummary, reporter.config.Succinct || !reporter.config.NoisySkippings, reporter.config.FullTrace)
|
||||||
case types.SpecStateTimedOut:
|
case types.SpecStateTimedOut:
|
||||||
reporter.stenographer.AnnounceSpecTimedOut(specSummary, reporter.config.Succinct, reporter.config.FullTrace)
|
reporter.stenographer.AnnounceSpecTimedOut(specSummary, reporter.config.Succinct, reporter.config.FullTrace)
|
||||||
case types.SpecStatePanicked:
|
case types.SpecStatePanicked:
|
||||||
|
|
25
vendor/github.com/onsi/ginkgo/reporters/junit_reporter.go
generated
vendored
25
vendor/github.com/onsi/ginkgo/reporters/junit_reporter.go
generated
vendored
|
@ -11,17 +11,21 @@ package reporters
|
||||||
import (
|
import (
|
||||||
"encoding/xml"
|
"encoding/xml"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/onsi/ginkgo/config"
|
"math"
|
||||||
"github.com/onsi/ginkgo/types"
|
|
||||||
"os"
|
"os"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/config"
|
||||||
|
"github.com/onsi/ginkgo/types"
|
||||||
)
|
)
|
||||||
|
|
||||||
type JUnitTestSuite struct {
|
type JUnitTestSuite struct {
|
||||||
XMLName xml.Name `xml:"testsuite"`
|
XMLName xml.Name `xml:"testsuite"`
|
||||||
TestCases []JUnitTestCase `xml:"testcase"`
|
TestCases []JUnitTestCase `xml:"testcase"`
|
||||||
|
Name string `xml:"name,attr"`
|
||||||
Tests int `xml:"tests,attr"`
|
Tests int `xml:"tests,attr"`
|
||||||
Failures int `xml:"failures,attr"`
|
Failures int `xml:"failures,attr"`
|
||||||
|
Errors int `xml:"errors,attr"`
|
||||||
Time float64 `xml:"time,attr"`
|
Time float64 `xml:"time,attr"`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -31,6 +35,7 @@ type JUnitTestCase struct {
|
||||||
FailureMessage *JUnitFailureMessage `xml:"failure,omitempty"`
|
FailureMessage *JUnitFailureMessage `xml:"failure,omitempty"`
|
||||||
Skipped *JUnitSkipped `xml:"skipped,omitempty"`
|
Skipped *JUnitSkipped `xml:"skipped,omitempty"`
|
||||||
Time float64 `xml:"time,attr"`
|
Time float64 `xml:"time,attr"`
|
||||||
|
SystemOut string `xml:"system-out,omitempty"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type JUnitFailureMessage struct {
|
type JUnitFailureMessage struct {
|
||||||
|
@ -57,7 +62,7 @@ func NewJUnitReporter(filename string) *JUnitReporter {
|
||||||
|
|
||||||
func (reporter *JUnitReporter) SpecSuiteWillBegin(config config.GinkgoConfigType, summary *types.SuiteSummary) {
|
func (reporter *JUnitReporter) SpecSuiteWillBegin(config config.GinkgoConfigType, summary *types.SuiteSummary) {
|
||||||
reporter.suite = JUnitTestSuite{
|
reporter.suite = JUnitTestSuite{
|
||||||
Tests: summary.NumberOfSpecsThatWillBeRun,
|
Name: summary.SuiteDescription,
|
||||||
TestCases: []JUnitTestCase{},
|
TestCases: []JUnitTestCase{},
|
||||||
}
|
}
|
||||||
reporter.testSuiteName = summary.SuiteDescription
|
reporter.testSuiteName = summary.SuiteDescription
|
||||||
|
@ -74,6 +79,10 @@ func (reporter *JUnitReporter) AfterSuiteDidRun(setupSummary *types.SetupSummary
|
||||||
reporter.handleSetupSummary("AfterSuite", setupSummary)
|
reporter.handleSetupSummary("AfterSuite", setupSummary)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func failureMessage(failure types.SpecFailure) string {
|
||||||
|
return fmt.Sprintf("%s\n%s\n%s", failure.ComponentCodeLocation.String(), failure.Message, failure.Location.String())
|
||||||
|
}
|
||||||
|
|
||||||
func (reporter *JUnitReporter) handleSetupSummary(name string, setupSummary *types.SetupSummary) {
|
func (reporter *JUnitReporter) handleSetupSummary(name string, setupSummary *types.SetupSummary) {
|
||||||
if setupSummary.State != types.SpecStatePassed {
|
if setupSummary.State != types.SpecStatePassed {
|
||||||
testCase := JUnitTestCase{
|
testCase := JUnitTestCase{
|
||||||
|
@ -83,8 +92,9 @@ func (reporter *JUnitReporter) handleSetupSummary(name string, setupSummary *typ
|
||||||
|
|
||||||
testCase.FailureMessage = &JUnitFailureMessage{
|
testCase.FailureMessage = &JUnitFailureMessage{
|
||||||
Type: reporter.failureTypeForState(setupSummary.State),
|
Type: reporter.failureTypeForState(setupSummary.State),
|
||||||
Message: fmt.Sprintf("%s\n%s", setupSummary.Failure.ComponentCodeLocation.String(), setupSummary.Failure.Message),
|
Message: failureMessage(setupSummary.Failure),
|
||||||
}
|
}
|
||||||
|
testCase.SystemOut = setupSummary.CapturedOutput
|
||||||
testCase.Time = setupSummary.RunTime.Seconds()
|
testCase.Time = setupSummary.RunTime.Seconds()
|
||||||
reporter.suite.TestCases = append(reporter.suite.TestCases, testCase)
|
reporter.suite.TestCases = append(reporter.suite.TestCases, testCase)
|
||||||
}
|
}
|
||||||
|
@ -98,8 +108,9 @@ func (reporter *JUnitReporter) SpecDidComplete(specSummary *types.SpecSummary) {
|
||||||
if specSummary.State == types.SpecStateFailed || specSummary.State == types.SpecStateTimedOut || specSummary.State == types.SpecStatePanicked {
|
if specSummary.State == types.SpecStateFailed || specSummary.State == types.SpecStateTimedOut || specSummary.State == types.SpecStatePanicked {
|
||||||
testCase.FailureMessage = &JUnitFailureMessage{
|
testCase.FailureMessage = &JUnitFailureMessage{
|
||||||
Type: reporter.failureTypeForState(specSummary.State),
|
Type: reporter.failureTypeForState(specSummary.State),
|
||||||
Message: fmt.Sprintf("%s\n%s", specSummary.Failure.ComponentCodeLocation.String(), specSummary.Failure.Message),
|
Message: failureMessage(specSummary.Failure),
|
||||||
}
|
}
|
||||||
|
testCase.SystemOut = specSummary.CapturedOutput
|
||||||
}
|
}
|
||||||
if specSummary.State == types.SpecStateSkipped || specSummary.State == types.SpecStatePending {
|
if specSummary.State == types.SpecStateSkipped || specSummary.State == types.SpecStatePending {
|
||||||
testCase.Skipped = &JUnitSkipped{}
|
testCase.Skipped = &JUnitSkipped{}
|
||||||
|
@ -109,8 +120,10 @@ func (reporter *JUnitReporter) SpecDidComplete(specSummary *types.SpecSummary) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (reporter *JUnitReporter) SpecSuiteDidEnd(summary *types.SuiteSummary) {
|
func (reporter *JUnitReporter) SpecSuiteDidEnd(summary *types.SuiteSummary) {
|
||||||
reporter.suite.Time = summary.RunTime.Seconds()
|
reporter.suite.Tests = summary.NumberOfSpecsThatWillBeRun
|
||||||
|
reporter.suite.Time = math.Trunc(summary.RunTime.Seconds() * 1000 / 1000)
|
||||||
reporter.suite.Failures = summary.NumberOfFailedSpecs
|
reporter.suite.Failures = summary.NumberOfFailedSpecs
|
||||||
|
reporter.suite.Errors = 0
|
||||||
file, err := os.Create(reporter.filename)
|
file, err := os.Create(reporter.filename)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
fmt.Printf("Failed to create JUnit report file: %s\n\t%s", reporter.filename, err.Error())
|
fmt.Printf("Failed to create JUnit report file: %s\n\t%s", reporter.filename, err.Error())
|
||||||
|
|
12
vendor/github.com/onsi/ginkgo/reporters/stenographer/console_logging.go
generated
vendored
12
vendor/github.com/onsi/ginkgo/reporters/stenographer/console_logging.go
generated
vendored
|
@ -22,24 +22,24 @@ func (s *consoleStenographer) colorize(colorCode string, format string, args ...
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) printBanner(text string, bannerCharacter string) {
|
func (s *consoleStenographer) printBanner(text string, bannerCharacter string) {
|
||||||
fmt.Println(text)
|
fmt.Fprintln(s.w, text)
|
||||||
fmt.Println(strings.Repeat(bannerCharacter, len(text)))
|
fmt.Fprintln(s.w, strings.Repeat(bannerCharacter, len(text)))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) printNewLine() {
|
func (s *consoleStenographer) printNewLine() {
|
||||||
fmt.Println("")
|
fmt.Fprintln(s.w, "")
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) printDelimiter() {
|
func (s *consoleStenographer) printDelimiter() {
|
||||||
fmt.Println(s.colorize(grayColor, "%s", strings.Repeat("-", 30)))
|
fmt.Fprintln(s.w, s.colorize(grayColor, "%s", strings.Repeat("-", 30)))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) print(indentation int, format string, args ...interface{}) {
|
func (s *consoleStenographer) print(indentation int, format string, args ...interface{}) {
|
||||||
fmt.Print(s.indent(indentation, format, args...))
|
fmt.Fprint(s.w, s.indent(indentation, format, args...))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) println(indentation int, format string, args ...interface{}) {
|
func (s *consoleStenographer) println(indentation int, format string, args ...interface{}) {
|
||||||
fmt.Println(s.indent(indentation, format, args...))
|
fmt.Fprintln(s.w, s.indent(indentation, format, args...))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) indent(indentation int, format string, args ...interface{}) string {
|
func (s *consoleStenographer) indent(indentation int, format string, args ...interface{}) string {
|
||||||
|
|
8
vendor/github.com/onsi/ginkgo/reporters/stenographer/fake_stenographer.go
generated
vendored
8
vendor/github.com/onsi/ginkgo/reporters/stenographer/fake_stenographer.go
generated
vendored
|
@ -74,14 +74,18 @@ func (stenographer *FakeStenographer) AnnounceAggregatedParallelRun(nodes int, s
|
||||||
stenographer.registerCall("AnnounceAggregatedParallelRun", nodes, succinct)
|
stenographer.registerCall("AnnounceAggregatedParallelRun", nodes, succinct)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (stenographer *FakeStenographer) AnnounceParallelRun(node int, nodes int, specsToRun int, totalSpecs int, succinct bool) {
|
func (stenographer *FakeStenographer) AnnounceParallelRun(node int, nodes int, succinct bool) {
|
||||||
stenographer.registerCall("AnnounceParallelRun", node, nodes, specsToRun, totalSpecs, succinct)
|
stenographer.registerCall("AnnounceParallelRun", node, nodes, succinct)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (stenographer *FakeStenographer) AnnounceNumberOfSpecs(specsToRun int, total int, succinct bool) {
|
func (stenographer *FakeStenographer) AnnounceNumberOfSpecs(specsToRun int, total int, succinct bool) {
|
||||||
stenographer.registerCall("AnnounceNumberOfSpecs", specsToRun, total, succinct)
|
stenographer.registerCall("AnnounceNumberOfSpecs", specsToRun, total, succinct)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (stenographer *FakeStenographer) AnnounceTotalNumberOfSpecs(total int, succinct bool) {
|
||||||
|
stenographer.registerCall("AnnounceTotalNumberOfSpecs", total, succinct)
|
||||||
|
}
|
||||||
|
|
||||||
func (stenographer *FakeStenographer) AnnounceSpecRunCompletion(summary *types.SuiteSummary, succinct bool) {
|
func (stenographer *FakeStenographer) AnnounceSpecRunCompletion(summary *types.SuiteSummary, succinct bool) {
|
||||||
stenographer.registerCall("AnnounceSpecRunCompletion", summary, succinct)
|
stenographer.registerCall("AnnounceSpecRunCompletion", summary, succinct)
|
||||||
}
|
}
|
||||||
|
|
82
vendor/github.com/onsi/ginkgo/reporters/stenographer/stenographer.go
generated
vendored
82
vendor/github.com/onsi/ginkgo/reporters/stenographer/stenographer.go
generated
vendored
|
@ -8,6 +8,8 @@ package stenographer
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"runtime"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"github.com/onsi/ginkgo/types"
|
"github.com/onsi/ginkgo/types"
|
||||||
|
@ -34,7 +36,8 @@ const (
|
||||||
type Stenographer interface {
|
type Stenographer interface {
|
||||||
AnnounceSuite(description string, randomSeed int64, randomizingAll bool, succinct bool)
|
AnnounceSuite(description string, randomSeed int64, randomizingAll bool, succinct bool)
|
||||||
AnnounceAggregatedParallelRun(nodes int, succinct bool)
|
AnnounceAggregatedParallelRun(nodes int, succinct bool)
|
||||||
AnnounceParallelRun(node int, nodes int, specsToRun int, totalSpecs int, succinct bool)
|
AnnounceParallelRun(node int, nodes int, succinct bool)
|
||||||
|
AnnounceTotalNumberOfSpecs(total int, succinct bool)
|
||||||
AnnounceNumberOfSpecs(specsToRun int, total int, succinct bool)
|
AnnounceNumberOfSpecs(specsToRun int, total int, succinct bool)
|
||||||
AnnounceSpecRunCompletion(summary *types.SuiteSummary, succinct bool)
|
AnnounceSpecRunCompletion(summary *types.SuiteSummary, succinct bool)
|
||||||
|
|
||||||
|
@ -58,16 +61,26 @@ type Stenographer interface {
|
||||||
SummarizeFailures(summaries []*types.SpecSummary)
|
SummarizeFailures(summaries []*types.SpecSummary)
|
||||||
}
|
}
|
||||||
|
|
||||||
func New(color bool) Stenographer {
|
func New(color bool, enableFlakes bool, writer io.Writer) Stenographer {
|
||||||
|
denoter := "•"
|
||||||
|
if runtime.GOOS == "windows" {
|
||||||
|
denoter = "+"
|
||||||
|
}
|
||||||
return &consoleStenographer{
|
return &consoleStenographer{
|
||||||
color: color,
|
color: color,
|
||||||
cursorState: cursorStateTop,
|
denoter: denoter,
|
||||||
|
cursorState: cursorStateTop,
|
||||||
|
enableFlakes: enableFlakes,
|
||||||
|
w: writer,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
type consoleStenographer struct {
|
type consoleStenographer struct {
|
||||||
color bool
|
color bool
|
||||||
cursorState cursorStateType
|
denoter string
|
||||||
|
cursorState cursorStateType
|
||||||
|
enableFlakes bool
|
||||||
|
w io.Writer
|
||||||
}
|
}
|
||||||
|
|
||||||
var alternatingColors = []string{defaultStyle, grayColor}
|
var alternatingColors = []string{defaultStyle, grayColor}
|
||||||
|
@ -85,17 +98,15 @@ func (s *consoleStenographer) AnnounceSuite(description string, randomSeed int64
|
||||||
s.printNewLine()
|
s.printNewLine()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) AnnounceParallelRun(node int, nodes int, specsToRun int, totalSpecs int, succinct bool) {
|
func (s *consoleStenographer) AnnounceParallelRun(node int, nodes int, succinct bool) {
|
||||||
if succinct {
|
if succinct {
|
||||||
s.print(0, "- node #%d ", node)
|
s.print(0, "- node #%d ", node)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
s.println(0,
|
s.println(0,
|
||||||
"Parallel test node %s/%s. Assigned %s of %s specs.",
|
"Parallel test node %s/%s.",
|
||||||
s.colorize(boldStyle, "%d", node),
|
s.colorize(boldStyle, "%d", node),
|
||||||
s.colorize(boldStyle, "%d", nodes),
|
s.colorize(boldStyle, "%d", nodes),
|
||||||
s.colorize(boldStyle, "%d", specsToRun),
|
|
||||||
s.colorize(boldStyle, "%d", totalSpecs),
|
|
||||||
)
|
)
|
||||||
s.printNewLine()
|
s.printNewLine()
|
||||||
}
|
}
|
||||||
|
@ -127,6 +138,20 @@ func (s *consoleStenographer) AnnounceNumberOfSpecs(specsToRun int, total int, s
|
||||||
s.printNewLine()
|
s.printNewLine()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (s *consoleStenographer) AnnounceTotalNumberOfSpecs(total int, succinct bool) {
|
||||||
|
if succinct {
|
||||||
|
s.print(0, "- %d specs ", total)
|
||||||
|
s.stream()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
s.println(0,
|
||||||
|
"Will run %s specs",
|
||||||
|
s.colorize(boldStyle, "%d", total),
|
||||||
|
)
|
||||||
|
|
||||||
|
s.printNewLine()
|
||||||
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) AnnounceSpecRunCompletion(summary *types.SuiteSummary, succinct bool) {
|
func (s *consoleStenographer) AnnounceSpecRunCompletion(summary *types.SuiteSummary, succinct bool) {
|
||||||
if succinct && summary.SuiteSucceeded {
|
if succinct && summary.SuiteSucceeded {
|
||||||
s.print(0, " %s %s ", s.colorize(greenColor, "SUCCESS!"), summary.RunTime)
|
s.print(0, " %s %s ", s.colorize(greenColor, "SUCCESS!"), summary.RunTime)
|
||||||
|
@ -146,11 +171,16 @@ func (s *consoleStenographer) AnnounceSpecRunCompletion(summary *types.SuiteSumm
|
||||||
status = s.colorize(boldStyle+redColor, "FAIL!")
|
status = s.colorize(boldStyle+redColor, "FAIL!")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
flakes := ""
|
||||||
|
if s.enableFlakes {
|
||||||
|
flakes = " | " + s.colorize(yellowColor+boldStyle, "%d Flaked", summary.NumberOfFlakedSpecs)
|
||||||
|
}
|
||||||
|
|
||||||
s.print(0,
|
s.print(0,
|
||||||
"%s -- %s | %s | %s | %s ",
|
"%s -- %s | %s | %s | %s\n",
|
||||||
status,
|
status,
|
||||||
s.colorize(greenColor+boldStyle, "%d Passed", summary.NumberOfPassedSpecs),
|
s.colorize(greenColor+boldStyle, "%d Passed", summary.NumberOfPassedSpecs),
|
||||||
s.colorize(redColor+boldStyle, "%d Failed", summary.NumberOfFailedSpecs),
|
s.colorize(redColor+boldStyle, "%d Failed", summary.NumberOfFailedSpecs)+flakes,
|
||||||
s.colorize(yellowColor+boldStyle, "%d Pending", summary.NumberOfPendingSpecs),
|
s.colorize(yellowColor+boldStyle, "%d Pending", summary.NumberOfPendingSpecs),
|
||||||
s.colorize(cyanColor+boldStyle, "%d Skipped", summary.NumberOfSkippedSpecs),
|
s.colorize(cyanColor+boldStyle, "%d Skipped", summary.NumberOfSkippedSpecs),
|
||||||
)
|
)
|
||||||
|
@ -216,13 +246,13 @@ func (s *consoleStenographer) AnnounceCapturedOutput(output string) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) AnnounceSuccesfulSpec(spec *types.SpecSummary) {
|
func (s *consoleStenographer) AnnounceSuccesfulSpec(spec *types.SpecSummary) {
|
||||||
s.print(0, s.colorize(greenColor, "•"))
|
s.print(0, s.colorize(greenColor, s.denoter))
|
||||||
s.stream()
|
s.stream()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) AnnounceSuccesfulSlowSpec(spec *types.SpecSummary, succinct bool) {
|
func (s *consoleStenographer) AnnounceSuccesfulSlowSpec(spec *types.SpecSummary, succinct bool) {
|
||||||
s.printBlockWithMessage(
|
s.printBlockWithMessage(
|
||||||
s.colorize(greenColor, "• [SLOW TEST:%.3f seconds]", spec.RunTime.Seconds()),
|
s.colorize(greenColor, "%s [SLOW TEST:%.3f seconds]", s.denoter, spec.RunTime.Seconds()),
|
||||||
"",
|
"",
|
||||||
spec,
|
spec,
|
||||||
succinct,
|
succinct,
|
||||||
|
@ -231,7 +261,7 @@ func (s *consoleStenographer) AnnounceSuccesfulSlowSpec(spec *types.SpecSummary,
|
||||||
|
|
||||||
func (s *consoleStenographer) AnnounceSuccesfulMeasurement(spec *types.SpecSummary, succinct bool) {
|
func (s *consoleStenographer) AnnounceSuccesfulMeasurement(spec *types.SpecSummary, succinct bool) {
|
||||||
s.printBlockWithMessage(
|
s.printBlockWithMessage(
|
||||||
s.colorize(greenColor, "• [MEASUREMENT]"),
|
s.colorize(greenColor, "%s [MEASUREMENT]", s.denoter),
|
||||||
s.measurementReport(spec, succinct),
|
s.measurementReport(spec, succinct),
|
||||||
spec,
|
spec,
|
||||||
succinct,
|
succinct,
|
||||||
|
@ -270,15 +300,15 @@ func (s *consoleStenographer) AnnounceSkippedSpec(spec *types.SpecSummary, succi
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) AnnounceSpecTimedOut(spec *types.SpecSummary, succinct bool, fullTrace bool) {
|
func (s *consoleStenographer) AnnounceSpecTimedOut(spec *types.SpecSummary, succinct bool, fullTrace bool) {
|
||||||
s.printSpecFailure("•... Timeout", spec, succinct, fullTrace)
|
s.printSpecFailure(fmt.Sprintf("%s... Timeout", s.denoter), spec, succinct, fullTrace)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) AnnounceSpecPanicked(spec *types.SpecSummary, succinct bool, fullTrace bool) {
|
func (s *consoleStenographer) AnnounceSpecPanicked(spec *types.SpecSummary, succinct bool, fullTrace bool) {
|
||||||
s.printSpecFailure("•! Panic", spec, succinct, fullTrace)
|
s.printSpecFailure(fmt.Sprintf("%s! Panic", s.denoter), spec, succinct, fullTrace)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) AnnounceSpecFailed(spec *types.SpecSummary, succinct bool, fullTrace bool) {
|
func (s *consoleStenographer) AnnounceSpecFailed(spec *types.SpecSummary, succinct bool, fullTrace bool) {
|
||||||
s.printSpecFailure("• Failure", spec, succinct, fullTrace)
|
s.printSpecFailure(fmt.Sprintf("%s Failure", s.denoter), spec, succinct, fullTrace)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *consoleStenographer) SummarizeFailures(summaries []*types.SpecSummary) {
|
func (s *consoleStenographer) SummarizeFailures(summaries []*types.SpecSummary) {
|
||||||
|
@ -499,15 +529,15 @@ func (s *consoleStenographer) measurementReport(spec *types.SpecSummary, succinc
|
||||||
message = append(message, fmt.Sprintf(" %s - %s: %s%s, %s: %s%s ± %s%s, %s: %s%s",
|
message = append(message, fmt.Sprintf(" %s - %s: %s%s, %s: %s%s ± %s%s, %s: %s%s",
|
||||||
s.colorize(boldStyle, "%s", measurement.Name),
|
s.colorize(boldStyle, "%s", measurement.Name),
|
||||||
measurement.SmallestLabel,
|
measurement.SmallestLabel,
|
||||||
s.colorize(greenColor, "%.3f", measurement.Smallest),
|
s.colorize(greenColor, measurement.PrecisionFmt(), measurement.Smallest),
|
||||||
measurement.Units,
|
measurement.Units,
|
||||||
measurement.AverageLabel,
|
measurement.AverageLabel,
|
||||||
s.colorize(cyanColor, "%.3f", measurement.Average),
|
s.colorize(cyanColor, measurement.PrecisionFmt(), measurement.Average),
|
||||||
measurement.Units,
|
measurement.Units,
|
||||||
s.colorize(cyanColor, "%.3f", measurement.StdDeviation),
|
s.colorize(cyanColor, measurement.PrecisionFmt(), measurement.StdDeviation),
|
||||||
measurement.Units,
|
measurement.Units,
|
||||||
measurement.LargestLabel,
|
measurement.LargestLabel,
|
||||||
s.colorize(redColor, "%.3f", measurement.Largest),
|
s.colorize(redColor, measurement.PrecisionFmt(), measurement.Largest),
|
||||||
measurement.Units,
|
measurement.Units,
|
||||||
))
|
))
|
||||||
}
|
}
|
||||||
|
@ -524,15 +554,15 @@ func (s *consoleStenographer) measurementReport(spec *types.SpecSummary, succinc
|
||||||
s.colorize(boldStyle, "%s", measurement.Name),
|
s.colorize(boldStyle, "%s", measurement.Name),
|
||||||
info,
|
info,
|
||||||
measurement.SmallestLabel,
|
measurement.SmallestLabel,
|
||||||
s.colorize(greenColor, "%.3f", measurement.Smallest),
|
s.colorize(greenColor, measurement.PrecisionFmt(), measurement.Smallest),
|
||||||
measurement.Units,
|
measurement.Units,
|
||||||
measurement.LargestLabel,
|
measurement.LargestLabel,
|
||||||
s.colorize(redColor, "%.3f", measurement.Largest),
|
s.colorize(redColor, measurement.PrecisionFmt(), measurement.Largest),
|
||||||
measurement.Units,
|
measurement.Units,
|
||||||
measurement.AverageLabel,
|
measurement.AverageLabel,
|
||||||
s.colorize(cyanColor, "%.3f", measurement.Average),
|
s.colorize(cyanColor, measurement.PrecisionFmt(), measurement.Average),
|
||||||
measurement.Units,
|
measurement.Units,
|
||||||
s.colorize(cyanColor, "%.3f", measurement.StdDeviation),
|
s.colorize(cyanColor, measurement.PrecisionFmt(), measurement.StdDeviation),
|
||||||
measurement.Units,
|
measurement.Units,
|
||||||
))
|
))
|
||||||
}
|
}
|
||||||
|
|
21
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-colorable/LICENSE
generated
vendored
Normal file
21
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-colorable/LICENSE
generated
vendored
Normal file
|
@ -0,0 +1,21 @@
|
||||||
|
The MIT License (MIT)
|
||||||
|
|
||||||
|
Copyright (c) 2016 Yasuhiro Matsumoto
|
||||||
|
|
||||||
|
Permission is hereby granted, free of charge, to any person obtaining a copy
|
||||||
|
of this software and associated documentation files (the "Software"), to deal
|
||||||
|
in the Software without restriction, including without limitation the rights
|
||||||
|
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
||||||
|
copies of the Software, and to permit persons to whom the Software is
|
||||||
|
furnished to do so, subject to the following conditions:
|
||||||
|
|
||||||
|
The above copyright notice and this permission notice shall be included in all
|
||||||
|
copies or substantial portions of the Software.
|
||||||
|
|
||||||
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||||
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||||
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
||||||
|
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||||
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
||||||
|
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
||||||
|
SOFTWARE.
|
24
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-colorable/colorable_others.go
generated
vendored
Normal file
24
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-colorable/colorable_others.go
generated
vendored
Normal file
|
@ -0,0 +1,24 @@
|
||||||
|
// +build !windows
|
||||||
|
|
||||||
|
package colorable
|
||||||
|
|
||||||
|
import (
|
||||||
|
"io"
|
||||||
|
"os"
|
||||||
|
)
|
||||||
|
|
||||||
|
func NewColorable(file *os.File) io.Writer {
|
||||||
|
if file == nil {
|
||||||
|
panic("nil passed instead of *os.File to NewColorable()")
|
||||||
|
}
|
||||||
|
|
||||||
|
return file
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewColorableStdout() io.Writer {
|
||||||
|
return os.Stdout
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewColorableStderr() io.Writer {
|
||||||
|
return os.Stderr
|
||||||
|
}
|
783
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-colorable/colorable_windows.go
generated
vendored
Normal file
783
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-colorable/colorable_windows.go
generated
vendored
Normal file
|
@ -0,0 +1,783 @@
|
||||||
|
package colorable
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"math"
|
||||||
|
"os"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
"syscall"
|
||||||
|
"unsafe"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
foregroundBlue = 0x1
|
||||||
|
foregroundGreen = 0x2
|
||||||
|
foregroundRed = 0x4
|
||||||
|
foregroundIntensity = 0x8
|
||||||
|
foregroundMask = (foregroundRed | foregroundBlue | foregroundGreen | foregroundIntensity)
|
||||||
|
backgroundBlue = 0x10
|
||||||
|
backgroundGreen = 0x20
|
||||||
|
backgroundRed = 0x40
|
||||||
|
backgroundIntensity = 0x80
|
||||||
|
backgroundMask = (backgroundRed | backgroundBlue | backgroundGreen | backgroundIntensity)
|
||||||
|
)
|
||||||
|
|
||||||
|
type wchar uint16
|
||||||
|
type short int16
|
||||||
|
type dword uint32
|
||||||
|
type word uint16
|
||||||
|
|
||||||
|
type coord struct {
|
||||||
|
x short
|
||||||
|
y short
|
||||||
|
}
|
||||||
|
|
||||||
|
type smallRect struct {
|
||||||
|
left short
|
||||||
|
top short
|
||||||
|
right short
|
||||||
|
bottom short
|
||||||
|
}
|
||||||
|
|
||||||
|
type consoleScreenBufferInfo struct {
|
||||||
|
size coord
|
||||||
|
cursorPosition coord
|
||||||
|
attributes word
|
||||||
|
window smallRect
|
||||||
|
maximumWindowSize coord
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
kernel32 = syscall.NewLazyDLL("kernel32.dll")
|
||||||
|
procGetConsoleScreenBufferInfo = kernel32.NewProc("GetConsoleScreenBufferInfo")
|
||||||
|
procSetConsoleTextAttribute = kernel32.NewProc("SetConsoleTextAttribute")
|
||||||
|
procSetConsoleCursorPosition = kernel32.NewProc("SetConsoleCursorPosition")
|
||||||
|
procFillConsoleOutputCharacter = kernel32.NewProc("FillConsoleOutputCharacterW")
|
||||||
|
procFillConsoleOutputAttribute = kernel32.NewProc("FillConsoleOutputAttribute")
|
||||||
|
)
|
||||||
|
|
||||||
|
type Writer struct {
|
||||||
|
out io.Writer
|
||||||
|
handle syscall.Handle
|
||||||
|
lastbuf bytes.Buffer
|
||||||
|
oldattr word
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewColorable(file *os.File) io.Writer {
|
||||||
|
if file == nil {
|
||||||
|
panic("nil passed instead of *os.File to NewColorable()")
|
||||||
|
}
|
||||||
|
|
||||||
|
if isatty.IsTerminal(file.Fd()) {
|
||||||
|
var csbi consoleScreenBufferInfo
|
||||||
|
handle := syscall.Handle(file.Fd())
|
||||||
|
procGetConsoleScreenBufferInfo.Call(uintptr(handle), uintptr(unsafe.Pointer(&csbi)))
|
||||||
|
return &Writer{out: file, handle: handle, oldattr: csbi.attributes}
|
||||||
|
} else {
|
||||||
|
return file
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewColorableStdout() io.Writer {
|
||||||
|
return NewColorable(os.Stdout)
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewColorableStderr() io.Writer {
|
||||||
|
return NewColorable(os.Stderr)
|
||||||
|
}
|
||||||
|
|
||||||
|
var color256 = map[int]int{
|
||||||
|
0: 0x000000,
|
||||||
|
1: 0x800000,
|
||||||
|
2: 0x008000,
|
||||||
|
3: 0x808000,
|
||||||
|
4: 0x000080,
|
||||||
|
5: 0x800080,
|
||||||
|
6: 0x008080,
|
||||||
|
7: 0xc0c0c0,
|
||||||
|
8: 0x808080,
|
||||||
|
9: 0xff0000,
|
||||||
|
10: 0x00ff00,
|
||||||
|
11: 0xffff00,
|
||||||
|
12: 0x0000ff,
|
||||||
|
13: 0xff00ff,
|
||||||
|
14: 0x00ffff,
|
||||||
|
15: 0xffffff,
|
||||||
|
16: 0x000000,
|
||||||
|
17: 0x00005f,
|
||||||
|
18: 0x000087,
|
||||||
|
19: 0x0000af,
|
||||||
|
20: 0x0000d7,
|
||||||
|
21: 0x0000ff,
|
||||||
|
22: 0x005f00,
|
||||||
|
23: 0x005f5f,
|
||||||
|
24: 0x005f87,
|
||||||
|
25: 0x005faf,
|
||||||
|
26: 0x005fd7,
|
||||||
|
27: 0x005fff,
|
||||||
|
28: 0x008700,
|
||||||
|
29: 0x00875f,
|
||||||
|
30: 0x008787,
|
||||||
|
31: 0x0087af,
|
||||||
|
32: 0x0087d7,
|
||||||
|
33: 0x0087ff,
|
||||||
|
34: 0x00af00,
|
||||||
|
35: 0x00af5f,
|
||||||
|
36: 0x00af87,
|
||||||
|
37: 0x00afaf,
|
||||||
|
38: 0x00afd7,
|
||||||
|
39: 0x00afff,
|
||||||
|
40: 0x00d700,
|
||||||
|
41: 0x00d75f,
|
||||||
|
42: 0x00d787,
|
||||||
|
43: 0x00d7af,
|
||||||
|
44: 0x00d7d7,
|
||||||
|
45: 0x00d7ff,
|
||||||
|
46: 0x00ff00,
|
||||||
|
47: 0x00ff5f,
|
||||||
|
48: 0x00ff87,
|
||||||
|
49: 0x00ffaf,
|
||||||
|
50: 0x00ffd7,
|
||||||
|
51: 0x00ffff,
|
||||||
|
52: 0x5f0000,
|
||||||
|
53: 0x5f005f,
|
||||||
|
54: 0x5f0087,
|
||||||
|
55: 0x5f00af,
|
||||||
|
56: 0x5f00d7,
|
||||||
|
57: 0x5f00ff,
|
||||||
|
58: 0x5f5f00,
|
||||||
|
59: 0x5f5f5f,
|
||||||
|
60: 0x5f5f87,
|
||||||
|
61: 0x5f5faf,
|
||||||
|
62: 0x5f5fd7,
|
||||||
|
63: 0x5f5fff,
|
||||||
|
64: 0x5f8700,
|
||||||
|
65: 0x5f875f,
|
||||||
|
66: 0x5f8787,
|
||||||
|
67: 0x5f87af,
|
||||||
|
68: 0x5f87d7,
|
||||||
|
69: 0x5f87ff,
|
||||||
|
70: 0x5faf00,
|
||||||
|
71: 0x5faf5f,
|
||||||
|
72: 0x5faf87,
|
||||||
|
73: 0x5fafaf,
|
||||||
|
74: 0x5fafd7,
|
||||||
|
75: 0x5fafff,
|
||||||
|
76: 0x5fd700,
|
||||||
|
77: 0x5fd75f,
|
||||||
|
78: 0x5fd787,
|
||||||
|
79: 0x5fd7af,
|
||||||
|
80: 0x5fd7d7,
|
||||||
|
81: 0x5fd7ff,
|
||||||
|
82: 0x5fff00,
|
||||||
|
83: 0x5fff5f,
|
||||||
|
84: 0x5fff87,
|
||||||
|
85: 0x5fffaf,
|
||||||
|
86: 0x5fffd7,
|
||||||
|
87: 0x5fffff,
|
||||||
|
88: 0x870000,
|
||||||
|
89: 0x87005f,
|
||||||
|
90: 0x870087,
|
||||||
|
91: 0x8700af,
|
||||||
|
92: 0x8700d7,
|
||||||
|
93: 0x8700ff,
|
||||||
|
94: 0x875f00,
|
||||||
|
95: 0x875f5f,
|
||||||
|
96: 0x875f87,
|
||||||
|
97: 0x875faf,
|
||||||
|
98: 0x875fd7,
|
||||||
|
99: 0x875fff,
|
||||||
|
100: 0x878700,
|
||||||
|
101: 0x87875f,
|
||||||
|
102: 0x878787,
|
||||||
|
103: 0x8787af,
|
||||||
|
104: 0x8787d7,
|
||||||
|
105: 0x8787ff,
|
||||||
|
106: 0x87af00,
|
||||||
|
107: 0x87af5f,
|
||||||
|
108: 0x87af87,
|
||||||
|
109: 0x87afaf,
|
||||||
|
110: 0x87afd7,
|
||||||
|
111: 0x87afff,
|
||||||
|
112: 0x87d700,
|
||||||
|
113: 0x87d75f,
|
||||||
|
114: 0x87d787,
|
||||||
|
115: 0x87d7af,
|
||||||
|
116: 0x87d7d7,
|
||||||
|
117: 0x87d7ff,
|
||||||
|
118: 0x87ff00,
|
||||||
|
119: 0x87ff5f,
|
||||||
|
120: 0x87ff87,
|
||||||
|
121: 0x87ffaf,
|
||||||
|
122: 0x87ffd7,
|
||||||
|
123: 0x87ffff,
|
||||||
|
124: 0xaf0000,
|
||||||
|
125: 0xaf005f,
|
||||||
|
126: 0xaf0087,
|
||||||
|
127: 0xaf00af,
|
||||||
|
128: 0xaf00d7,
|
||||||
|
129: 0xaf00ff,
|
||||||
|
130: 0xaf5f00,
|
||||||
|
131: 0xaf5f5f,
|
||||||
|
132: 0xaf5f87,
|
||||||
|
133: 0xaf5faf,
|
||||||
|
134: 0xaf5fd7,
|
||||||
|
135: 0xaf5fff,
|
||||||
|
136: 0xaf8700,
|
||||||
|
137: 0xaf875f,
|
||||||
|
138: 0xaf8787,
|
||||||
|
139: 0xaf87af,
|
||||||
|
140: 0xaf87d7,
|
||||||
|
141: 0xaf87ff,
|
||||||
|
142: 0xafaf00,
|
||||||
|
143: 0xafaf5f,
|
||||||
|
144: 0xafaf87,
|
||||||
|
145: 0xafafaf,
|
||||||
|
146: 0xafafd7,
|
||||||
|
147: 0xafafff,
|
||||||
|
148: 0xafd700,
|
||||||
|
149: 0xafd75f,
|
||||||
|
150: 0xafd787,
|
||||||
|
151: 0xafd7af,
|
||||||
|
152: 0xafd7d7,
|
||||||
|
153: 0xafd7ff,
|
||||||
|
154: 0xafff00,
|
||||||
|
155: 0xafff5f,
|
||||||
|
156: 0xafff87,
|
||||||
|
157: 0xafffaf,
|
||||||
|
158: 0xafffd7,
|
||||||
|
159: 0xafffff,
|
||||||
|
160: 0xd70000,
|
||||||
|
161: 0xd7005f,
|
||||||
|
162: 0xd70087,
|
||||||
|
163: 0xd700af,
|
||||||
|
164: 0xd700d7,
|
||||||
|
165: 0xd700ff,
|
||||||
|
166: 0xd75f00,
|
||||||
|
167: 0xd75f5f,
|
||||||
|
168: 0xd75f87,
|
||||||
|
169: 0xd75faf,
|
||||||
|
170: 0xd75fd7,
|
||||||
|
171: 0xd75fff,
|
||||||
|
172: 0xd78700,
|
||||||
|
173: 0xd7875f,
|
||||||
|
174: 0xd78787,
|
||||||
|
175: 0xd787af,
|
||||||
|
176: 0xd787d7,
|
||||||
|
177: 0xd787ff,
|
||||||
|
178: 0xd7af00,
|
||||||
|
179: 0xd7af5f,
|
||||||
|
180: 0xd7af87,
|
||||||
|
181: 0xd7afaf,
|
||||||
|
182: 0xd7afd7,
|
||||||
|
183: 0xd7afff,
|
||||||
|
184: 0xd7d700,
|
||||||
|
185: 0xd7d75f,
|
||||||
|
186: 0xd7d787,
|
||||||
|
187: 0xd7d7af,
|
||||||
|
188: 0xd7d7d7,
|
||||||
|
189: 0xd7d7ff,
|
||||||
|
190: 0xd7ff00,
|
||||||
|
191: 0xd7ff5f,
|
||||||
|
192: 0xd7ff87,
|
||||||
|
193: 0xd7ffaf,
|
||||||
|
194: 0xd7ffd7,
|
||||||
|
195: 0xd7ffff,
|
||||||
|
196: 0xff0000,
|
||||||
|
197: 0xff005f,
|
||||||
|
198: 0xff0087,
|
||||||
|
199: 0xff00af,
|
||||||
|
200: 0xff00d7,
|
||||||
|
201: 0xff00ff,
|
||||||
|
202: 0xff5f00,
|
||||||
|
203: 0xff5f5f,
|
||||||
|
204: 0xff5f87,
|
||||||
|
205: 0xff5faf,
|
||||||
|
206: 0xff5fd7,
|
||||||
|
207: 0xff5fff,
|
||||||
|
208: 0xff8700,
|
||||||
|
209: 0xff875f,
|
||||||
|
210: 0xff8787,
|
||||||
|
211: 0xff87af,
|
||||||
|
212: 0xff87d7,
|
||||||
|
213: 0xff87ff,
|
||||||
|
214: 0xffaf00,
|
||||||
|
215: 0xffaf5f,
|
||||||
|
216: 0xffaf87,
|
||||||
|
217: 0xffafaf,
|
||||||
|
218: 0xffafd7,
|
||||||
|
219: 0xffafff,
|
||||||
|
220: 0xffd700,
|
||||||
|
221: 0xffd75f,
|
||||||
|
222: 0xffd787,
|
||||||
|
223: 0xffd7af,
|
||||||
|
224: 0xffd7d7,
|
||||||
|
225: 0xffd7ff,
|
||||||
|
226: 0xffff00,
|
||||||
|
227: 0xffff5f,
|
||||||
|
228: 0xffff87,
|
||||||
|
229: 0xffffaf,
|
||||||
|
230: 0xffffd7,
|
||||||
|
231: 0xffffff,
|
||||||
|
232: 0x080808,
|
||||||
|
233: 0x121212,
|
||||||
|
234: 0x1c1c1c,
|
||||||
|
235: 0x262626,
|
||||||
|
236: 0x303030,
|
||||||
|
237: 0x3a3a3a,
|
||||||
|
238: 0x444444,
|
||||||
|
239: 0x4e4e4e,
|
||||||
|
240: 0x585858,
|
||||||
|
241: 0x626262,
|
||||||
|
242: 0x6c6c6c,
|
||||||
|
243: 0x767676,
|
||||||
|
244: 0x808080,
|
||||||
|
245: 0x8a8a8a,
|
||||||
|
246: 0x949494,
|
||||||
|
247: 0x9e9e9e,
|
||||||
|
248: 0xa8a8a8,
|
||||||
|
249: 0xb2b2b2,
|
||||||
|
250: 0xbcbcbc,
|
||||||
|
251: 0xc6c6c6,
|
||||||
|
252: 0xd0d0d0,
|
||||||
|
253: 0xdadada,
|
||||||
|
254: 0xe4e4e4,
|
||||||
|
255: 0xeeeeee,
|
||||||
|
}
|
||||||
|
|
||||||
|
func (w *Writer) Write(data []byte) (n int, err error) {
|
||||||
|
var csbi consoleScreenBufferInfo
|
||||||
|
procGetConsoleScreenBufferInfo.Call(uintptr(w.handle), uintptr(unsafe.Pointer(&csbi)))
|
||||||
|
|
||||||
|
er := bytes.NewBuffer(data)
|
||||||
|
loop:
|
||||||
|
for {
|
||||||
|
r1, _, err := procGetConsoleScreenBufferInfo.Call(uintptr(w.handle), uintptr(unsafe.Pointer(&csbi)))
|
||||||
|
if r1 == 0 {
|
||||||
|
break loop
|
||||||
|
}
|
||||||
|
|
||||||
|
c1, _, err := er.ReadRune()
|
||||||
|
if err != nil {
|
||||||
|
break loop
|
||||||
|
}
|
||||||
|
if c1 != 0x1b {
|
||||||
|
fmt.Fprint(w.out, string(c1))
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
c2, _, err := er.ReadRune()
|
||||||
|
if err != nil {
|
||||||
|
w.lastbuf.WriteRune(c1)
|
||||||
|
break loop
|
||||||
|
}
|
||||||
|
if c2 != 0x5b {
|
||||||
|
w.lastbuf.WriteRune(c1)
|
||||||
|
w.lastbuf.WriteRune(c2)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
var buf bytes.Buffer
|
||||||
|
var m rune
|
||||||
|
for {
|
||||||
|
c, _, err := er.ReadRune()
|
||||||
|
if err != nil {
|
||||||
|
w.lastbuf.WriteRune(c1)
|
||||||
|
w.lastbuf.WriteRune(c2)
|
||||||
|
w.lastbuf.Write(buf.Bytes())
|
||||||
|
break loop
|
||||||
|
}
|
||||||
|
if ('a' <= c && c <= 'z') || ('A' <= c && c <= 'Z') || c == '@' {
|
||||||
|
m = c
|
||||||
|
break
|
||||||
|
}
|
||||||
|
buf.Write([]byte(string(c)))
|
||||||
|
}
|
||||||
|
|
||||||
|
var csbi consoleScreenBufferInfo
|
||||||
|
switch m {
|
||||||
|
case 'A':
|
||||||
|
n, err = strconv.Atoi(buf.String())
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
procGetConsoleScreenBufferInfo.Call(uintptr(w.handle), uintptr(unsafe.Pointer(&csbi)))
|
||||||
|
csbi.cursorPosition.y -= short(n)
|
||||||
|
procSetConsoleCursorPosition.Call(uintptr(w.handle), *(*uintptr)(unsafe.Pointer(&csbi.cursorPosition)))
|
||||||
|
case 'B':
|
||||||
|
n, err = strconv.Atoi(buf.String())
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
procGetConsoleScreenBufferInfo.Call(uintptr(w.handle), uintptr(unsafe.Pointer(&csbi)))
|
||||||
|
csbi.cursorPosition.y += short(n)
|
||||||
|
procSetConsoleCursorPosition.Call(uintptr(w.handle), *(*uintptr)(unsafe.Pointer(&csbi.cursorPosition)))
|
||||||
|
case 'C':
|
||||||
|
n, err = strconv.Atoi(buf.String())
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
procGetConsoleScreenBufferInfo.Call(uintptr(w.handle), uintptr(unsafe.Pointer(&csbi)))
|
||||||
|
csbi.cursorPosition.x -= short(n)
|
||||||
|
procSetConsoleCursorPosition.Call(uintptr(w.handle), *(*uintptr)(unsafe.Pointer(&csbi.cursorPosition)))
|
||||||
|
case 'D':
|
||||||
|
n, err = strconv.Atoi(buf.String())
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if n, err = strconv.Atoi(buf.String()); err == nil {
|
||||||
|
var csbi consoleScreenBufferInfo
|
||||||
|
procGetConsoleScreenBufferInfo.Call(uintptr(w.handle), uintptr(unsafe.Pointer(&csbi)))
|
||||||
|
csbi.cursorPosition.x += short(n)
|
||||||
|
procSetConsoleCursorPosition.Call(uintptr(w.handle), *(*uintptr)(unsafe.Pointer(&csbi.cursorPosition)))
|
||||||
|
}
|
||||||
|
case 'E':
|
||||||
|
n, err = strconv.Atoi(buf.String())
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
procGetConsoleScreenBufferInfo.Call(uintptr(w.handle), uintptr(unsafe.Pointer(&csbi)))
|
||||||
|
csbi.cursorPosition.x = 0
|
||||||
|
csbi.cursorPosition.y += short(n)
|
||||||
|
procSetConsoleCursorPosition.Call(uintptr(w.handle), *(*uintptr)(unsafe.Pointer(&csbi.cursorPosition)))
|
||||||
|
case 'F':
|
||||||
|
n, err = strconv.Atoi(buf.String())
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
procGetConsoleScreenBufferInfo.Call(uintptr(w.handle), uintptr(unsafe.Pointer(&csbi)))
|
||||||
|
csbi.cursorPosition.x = 0
|
||||||
|
csbi.cursorPosition.y -= short(n)
|
||||||
|
procSetConsoleCursorPosition.Call(uintptr(w.handle), *(*uintptr)(unsafe.Pointer(&csbi.cursorPosition)))
|
||||||
|
case 'G':
|
||||||
|
n, err = strconv.Atoi(buf.String())
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
procGetConsoleScreenBufferInfo.Call(uintptr(w.handle), uintptr(unsafe.Pointer(&csbi)))
|
||||||
|
csbi.cursorPosition.x = short(n)
|
||||||
|
procSetConsoleCursorPosition.Call(uintptr(w.handle), *(*uintptr)(unsafe.Pointer(&csbi.cursorPosition)))
|
||||||
|
case 'H':
|
||||||
|
token := strings.Split(buf.String(), ";")
|
||||||
|
if len(token) != 2 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
n1, err := strconv.Atoi(token[0])
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
n2, err := strconv.Atoi(token[1])
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
csbi.cursorPosition.x = short(n2)
|
||||||
|
csbi.cursorPosition.x = short(n1)
|
||||||
|
procSetConsoleCursorPosition.Call(uintptr(w.handle), *(*uintptr)(unsafe.Pointer(&csbi.cursorPosition)))
|
||||||
|
case 'J':
|
||||||
|
n, err := strconv.Atoi(buf.String())
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
var cursor coord
|
||||||
|
switch n {
|
||||||
|
case 0:
|
||||||
|
cursor = coord{x: csbi.cursorPosition.x, y: csbi.cursorPosition.y}
|
||||||
|
case 1:
|
||||||
|
cursor = coord{x: csbi.window.left, y: csbi.window.top}
|
||||||
|
case 2:
|
||||||
|
cursor = coord{x: csbi.window.left, y: csbi.window.top}
|
||||||
|
}
|
||||||
|
var count, written dword
|
||||||
|
count = dword(csbi.size.x - csbi.cursorPosition.x + (csbi.size.y-csbi.cursorPosition.y)*csbi.size.x)
|
||||||
|
procFillConsoleOutputCharacter.Call(uintptr(w.handle), uintptr(' '), uintptr(count), *(*uintptr)(unsafe.Pointer(&cursor)), uintptr(unsafe.Pointer(&written)))
|
||||||
|
procFillConsoleOutputAttribute.Call(uintptr(w.handle), uintptr(csbi.attributes), uintptr(count), *(*uintptr)(unsafe.Pointer(&cursor)), uintptr(unsafe.Pointer(&written)))
|
||||||
|
case 'K':
|
||||||
|
n, err := strconv.Atoi(buf.String())
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
var cursor coord
|
||||||
|
switch n {
|
||||||
|
case 0:
|
||||||
|
cursor = coord{x: csbi.cursorPosition.x, y: csbi.cursorPosition.y}
|
||||||
|
case 1:
|
||||||
|
cursor = coord{x: csbi.window.left, y: csbi.window.top + csbi.cursorPosition.y}
|
||||||
|
case 2:
|
||||||
|
cursor = coord{x: csbi.window.left, y: csbi.window.top + csbi.cursorPosition.y}
|
||||||
|
}
|
||||||
|
var count, written dword
|
||||||
|
count = dword(csbi.size.x - csbi.cursorPosition.x)
|
||||||
|
procFillConsoleOutputCharacter.Call(uintptr(w.handle), uintptr(' '), uintptr(count), *(*uintptr)(unsafe.Pointer(&cursor)), uintptr(unsafe.Pointer(&written)))
|
||||||
|
procFillConsoleOutputAttribute.Call(uintptr(w.handle), uintptr(csbi.attributes), uintptr(count), *(*uintptr)(unsafe.Pointer(&cursor)), uintptr(unsafe.Pointer(&written)))
|
||||||
|
case 'm':
|
||||||
|
attr := csbi.attributes
|
||||||
|
cs := buf.String()
|
||||||
|
if cs == "" {
|
||||||
|
procSetConsoleTextAttribute.Call(uintptr(w.handle), uintptr(w.oldattr))
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
token := strings.Split(cs, ";")
|
||||||
|
for i := 0; i < len(token); i += 1 {
|
||||||
|
ns := token[i]
|
||||||
|
if n, err = strconv.Atoi(ns); err == nil {
|
||||||
|
switch {
|
||||||
|
case n == 0 || n == 100:
|
||||||
|
attr = w.oldattr
|
||||||
|
case 1 <= n && n <= 5:
|
||||||
|
attr |= foregroundIntensity
|
||||||
|
case n == 7:
|
||||||
|
attr = ((attr & foregroundMask) << 4) | ((attr & backgroundMask) >> 4)
|
||||||
|
case 22 == n || n == 25 || n == 25:
|
||||||
|
attr |= foregroundIntensity
|
||||||
|
case n == 27:
|
||||||
|
attr = ((attr & foregroundMask) << 4) | ((attr & backgroundMask) >> 4)
|
||||||
|
case 30 <= n && n <= 37:
|
||||||
|
attr = (attr & backgroundMask)
|
||||||
|
if (n-30)&1 != 0 {
|
||||||
|
attr |= foregroundRed
|
||||||
|
}
|
||||||
|
if (n-30)&2 != 0 {
|
||||||
|
attr |= foregroundGreen
|
||||||
|
}
|
||||||
|
if (n-30)&4 != 0 {
|
||||||
|
attr |= foregroundBlue
|
||||||
|
}
|
||||||
|
case n == 38: // set foreground color.
|
||||||
|
if i < len(token)-2 && (token[i+1] == "5" || token[i+1] == "05") {
|
||||||
|
if n256, err := strconv.Atoi(token[i+2]); err == nil {
|
||||||
|
if n256foreAttr == nil {
|
||||||
|
n256setup()
|
||||||
|
}
|
||||||
|
attr &= backgroundMask
|
||||||
|
attr |= n256foreAttr[n256]
|
||||||
|
i += 2
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
attr = attr & (w.oldattr & backgroundMask)
|
||||||
|
}
|
||||||
|
case n == 39: // reset foreground color.
|
||||||
|
attr &= backgroundMask
|
||||||
|
attr |= w.oldattr & foregroundMask
|
||||||
|
case 40 <= n && n <= 47:
|
||||||
|
attr = (attr & foregroundMask)
|
||||||
|
if (n-40)&1 != 0 {
|
||||||
|
attr |= backgroundRed
|
||||||
|
}
|
||||||
|
if (n-40)&2 != 0 {
|
||||||
|
attr |= backgroundGreen
|
||||||
|
}
|
||||||
|
if (n-40)&4 != 0 {
|
||||||
|
attr |= backgroundBlue
|
||||||
|
}
|
||||||
|
case n == 48: // set background color.
|
||||||
|
if i < len(token)-2 && token[i+1] == "5" {
|
||||||
|
if n256, err := strconv.Atoi(token[i+2]); err == nil {
|
||||||
|
if n256backAttr == nil {
|
||||||
|
n256setup()
|
||||||
|
}
|
||||||
|
attr &= foregroundMask
|
||||||
|
attr |= n256backAttr[n256]
|
||||||
|
i += 2
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
attr = attr & (w.oldattr & foregroundMask)
|
||||||
|
}
|
||||||
|
case n == 49: // reset foreground color.
|
||||||
|
attr &= foregroundMask
|
||||||
|
attr |= w.oldattr & backgroundMask
|
||||||
|
case 90 <= n && n <= 97:
|
||||||
|
attr = (attr & backgroundMask)
|
||||||
|
attr |= foregroundIntensity
|
||||||
|
if (n-90)&1 != 0 {
|
||||||
|
attr |= foregroundRed
|
||||||
|
}
|
||||||
|
if (n-90)&2 != 0 {
|
||||||
|
attr |= foregroundGreen
|
||||||
|
}
|
||||||
|
if (n-90)&4 != 0 {
|
||||||
|
attr |= foregroundBlue
|
||||||
|
}
|
||||||
|
case 100 <= n && n <= 107:
|
||||||
|
attr = (attr & foregroundMask)
|
||||||
|
attr |= backgroundIntensity
|
||||||
|
if (n-100)&1 != 0 {
|
||||||
|
attr |= backgroundRed
|
||||||
|
}
|
||||||
|
if (n-100)&2 != 0 {
|
||||||
|
attr |= backgroundGreen
|
||||||
|
}
|
||||||
|
if (n-100)&4 != 0 {
|
||||||
|
attr |= backgroundBlue
|
||||||
|
}
|
||||||
|
}
|
||||||
|
procSetConsoleTextAttribute.Call(uintptr(w.handle), uintptr(attr))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return len(data) - w.lastbuf.Len(), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
type consoleColor struct {
|
||||||
|
rgb int
|
||||||
|
red bool
|
||||||
|
green bool
|
||||||
|
blue bool
|
||||||
|
intensity bool
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c consoleColor) foregroundAttr() (attr word) {
|
||||||
|
if c.red {
|
||||||
|
attr |= foregroundRed
|
||||||
|
}
|
||||||
|
if c.green {
|
||||||
|
attr |= foregroundGreen
|
||||||
|
}
|
||||||
|
if c.blue {
|
||||||
|
attr |= foregroundBlue
|
||||||
|
}
|
||||||
|
if c.intensity {
|
||||||
|
attr |= foregroundIntensity
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c consoleColor) backgroundAttr() (attr word) {
|
||||||
|
if c.red {
|
||||||
|
attr |= backgroundRed
|
||||||
|
}
|
||||||
|
if c.green {
|
||||||
|
attr |= backgroundGreen
|
||||||
|
}
|
||||||
|
if c.blue {
|
||||||
|
attr |= backgroundBlue
|
||||||
|
}
|
||||||
|
if c.intensity {
|
||||||
|
attr |= backgroundIntensity
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
var color16 = []consoleColor{
|
||||||
|
consoleColor{0x000000, false, false, false, false},
|
||||||
|
consoleColor{0x000080, false, false, true, false},
|
||||||
|
consoleColor{0x008000, false, true, false, false},
|
||||||
|
consoleColor{0x008080, false, true, true, false},
|
||||||
|
consoleColor{0x800000, true, false, false, false},
|
||||||
|
consoleColor{0x800080, true, false, true, false},
|
||||||
|
consoleColor{0x808000, true, true, false, false},
|
||||||
|
consoleColor{0xc0c0c0, true, true, true, false},
|
||||||
|
consoleColor{0x808080, false, false, false, true},
|
||||||
|
consoleColor{0x0000ff, false, false, true, true},
|
||||||
|
consoleColor{0x00ff00, false, true, false, true},
|
||||||
|
consoleColor{0x00ffff, false, true, true, true},
|
||||||
|
consoleColor{0xff0000, true, false, false, true},
|
||||||
|
consoleColor{0xff00ff, true, false, true, true},
|
||||||
|
consoleColor{0xffff00, true, true, false, true},
|
||||||
|
consoleColor{0xffffff, true, true, true, true},
|
||||||
|
}
|
||||||
|
|
||||||
|
type hsv struct {
|
||||||
|
h, s, v float32
|
||||||
|
}
|
||||||
|
|
||||||
|
func (a hsv) dist(b hsv) float32 {
|
||||||
|
dh := a.h - b.h
|
||||||
|
switch {
|
||||||
|
case dh > 0.5:
|
||||||
|
dh = 1 - dh
|
||||||
|
case dh < -0.5:
|
||||||
|
dh = -1 - dh
|
||||||
|
}
|
||||||
|
ds := a.s - b.s
|
||||||
|
dv := a.v - b.v
|
||||||
|
return float32(math.Sqrt(float64(dh*dh + ds*ds + dv*dv)))
|
||||||
|
}
|
||||||
|
|
||||||
|
func toHSV(rgb int) hsv {
|
||||||
|
r, g, b := float32((rgb&0xFF0000)>>16)/256.0,
|
||||||
|
float32((rgb&0x00FF00)>>8)/256.0,
|
||||||
|
float32(rgb&0x0000FF)/256.0
|
||||||
|
min, max := minmax3f(r, g, b)
|
||||||
|
h := max - min
|
||||||
|
if h > 0 {
|
||||||
|
if max == r {
|
||||||
|
h = (g - b) / h
|
||||||
|
if h < 0 {
|
||||||
|
h += 6
|
||||||
|
}
|
||||||
|
} else if max == g {
|
||||||
|
h = 2 + (b-r)/h
|
||||||
|
} else {
|
||||||
|
h = 4 + (r-g)/h
|
||||||
|
}
|
||||||
|
}
|
||||||
|
h /= 6.0
|
||||||
|
s := max - min
|
||||||
|
if max != 0 {
|
||||||
|
s /= max
|
||||||
|
}
|
||||||
|
v := max
|
||||||
|
return hsv{h: h, s: s, v: v}
|
||||||
|
}
|
||||||
|
|
||||||
|
type hsvTable []hsv
|
||||||
|
|
||||||
|
func toHSVTable(rgbTable []consoleColor) hsvTable {
|
||||||
|
t := make(hsvTable, len(rgbTable))
|
||||||
|
for i, c := range rgbTable {
|
||||||
|
t[i] = toHSV(c.rgb)
|
||||||
|
}
|
||||||
|
return t
|
||||||
|
}
|
||||||
|
|
||||||
|
func (t hsvTable) find(rgb int) consoleColor {
|
||||||
|
hsv := toHSV(rgb)
|
||||||
|
n := 7
|
||||||
|
l := float32(5.0)
|
||||||
|
for i, p := range t {
|
||||||
|
d := hsv.dist(p)
|
||||||
|
if d < l {
|
||||||
|
l, n = d, i
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return color16[n]
|
||||||
|
}
|
||||||
|
|
||||||
|
func minmax3f(a, b, c float32) (min, max float32) {
|
||||||
|
if a < b {
|
||||||
|
if b < c {
|
||||||
|
return a, c
|
||||||
|
} else if a < c {
|
||||||
|
return a, b
|
||||||
|
} else {
|
||||||
|
return c, b
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if a < c {
|
||||||
|
return b, c
|
||||||
|
} else if b < c {
|
||||||
|
return b, a
|
||||||
|
} else {
|
||||||
|
return c, a
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
var n256foreAttr []word
|
||||||
|
var n256backAttr []word
|
||||||
|
|
||||||
|
func n256setup() {
|
||||||
|
n256foreAttr = make([]word, 256)
|
||||||
|
n256backAttr = make([]word, 256)
|
||||||
|
t := toHSVTable(color16)
|
||||||
|
for i, rgb := range color256 {
|
||||||
|
c := t.find(rgb)
|
||||||
|
n256foreAttr[i] = c.foregroundAttr()
|
||||||
|
n256backAttr[i] = c.backgroundAttr()
|
||||||
|
}
|
||||||
|
}
|
57
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-colorable/noncolorable.go
generated
vendored
Normal file
57
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-colorable/noncolorable.go
generated
vendored
Normal file
|
@ -0,0 +1,57 @@
|
||||||
|
package colorable
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
)
|
||||||
|
|
||||||
|
type NonColorable struct {
|
||||||
|
out io.Writer
|
||||||
|
lastbuf bytes.Buffer
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewNonColorable(w io.Writer) io.Writer {
|
||||||
|
return &NonColorable{out: w}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (w *NonColorable) Write(data []byte) (n int, err error) {
|
||||||
|
er := bytes.NewBuffer(data)
|
||||||
|
loop:
|
||||||
|
for {
|
||||||
|
c1, _, err := er.ReadRune()
|
||||||
|
if err != nil {
|
||||||
|
break loop
|
||||||
|
}
|
||||||
|
if c1 != 0x1b {
|
||||||
|
fmt.Fprint(w.out, string(c1))
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
c2, _, err := er.ReadRune()
|
||||||
|
if err != nil {
|
||||||
|
w.lastbuf.WriteRune(c1)
|
||||||
|
break loop
|
||||||
|
}
|
||||||
|
if c2 != 0x5b {
|
||||||
|
w.lastbuf.WriteRune(c1)
|
||||||
|
w.lastbuf.WriteRune(c2)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
var buf bytes.Buffer
|
||||||
|
for {
|
||||||
|
c, _, err := er.ReadRune()
|
||||||
|
if err != nil {
|
||||||
|
w.lastbuf.WriteRune(c1)
|
||||||
|
w.lastbuf.WriteRune(c2)
|
||||||
|
w.lastbuf.Write(buf.Bytes())
|
||||||
|
break loop
|
||||||
|
}
|
||||||
|
if ('a' <= c && c <= 'z') || ('A' <= c && c <= 'Z') || c == '@' {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
buf.Write([]byte(string(c)))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return len(data) - w.lastbuf.Len(), nil
|
||||||
|
}
|
9
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/LICENSE
generated
vendored
Normal file
9
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/LICENSE
generated
vendored
Normal file
|
@ -0,0 +1,9 @@
|
||||||
|
Copyright (c) Yasuhiro MATSUMOTO <mattn.jp@gmail.com>
|
||||||
|
|
||||||
|
MIT License (Expat)
|
||||||
|
|
||||||
|
Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:
|
||||||
|
|
||||||
|
The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.
|
||||||
|
|
||||||
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
2
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/doc.go
generated
vendored
Normal file
2
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/doc.go
generated
vendored
Normal file
|
@ -0,0 +1,2 @@
|
||||||
|
// Package isatty implements interface to isatty
|
||||||
|
package isatty
|
9
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/isatty_appengine.go
generated
vendored
Normal file
9
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/isatty_appengine.go
generated
vendored
Normal file
|
@ -0,0 +1,9 @@
|
||||||
|
// +build appengine
|
||||||
|
|
||||||
|
package isatty
|
||||||
|
|
||||||
|
// IsTerminal returns true if the file descriptor is terminal which
|
||||||
|
// is always false on on appengine classic which is a sandboxed PaaS.
|
||||||
|
func IsTerminal(fd uintptr) bool {
|
||||||
|
return false
|
||||||
|
}
|
18
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/isatty_bsd.go
generated
vendored
Normal file
18
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/isatty_bsd.go
generated
vendored
Normal file
|
@ -0,0 +1,18 @@
|
||||||
|
// +build darwin freebsd openbsd netbsd
|
||||||
|
// +build !appengine
|
||||||
|
|
||||||
|
package isatty
|
||||||
|
|
||||||
|
import (
|
||||||
|
"syscall"
|
||||||
|
"unsafe"
|
||||||
|
)
|
||||||
|
|
||||||
|
const ioctlReadTermios = syscall.TIOCGETA
|
||||||
|
|
||||||
|
// IsTerminal return true if the file descriptor is terminal.
|
||||||
|
func IsTerminal(fd uintptr) bool {
|
||||||
|
var termios syscall.Termios
|
||||||
|
_, _, err := syscall.Syscall6(syscall.SYS_IOCTL, fd, ioctlReadTermios, uintptr(unsafe.Pointer(&termios)), 0, 0, 0)
|
||||||
|
return err == 0
|
||||||
|
}
|
18
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/isatty_linux.go
generated
vendored
Normal file
18
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/isatty_linux.go
generated
vendored
Normal file
|
@ -0,0 +1,18 @@
|
||||||
|
// +build linux
|
||||||
|
// +build !appengine
|
||||||
|
|
||||||
|
package isatty
|
||||||
|
|
||||||
|
import (
|
||||||
|
"syscall"
|
||||||
|
"unsafe"
|
||||||
|
)
|
||||||
|
|
||||||
|
const ioctlReadTermios = syscall.TCGETS
|
||||||
|
|
||||||
|
// IsTerminal return true if the file descriptor is terminal.
|
||||||
|
func IsTerminal(fd uintptr) bool {
|
||||||
|
var termios syscall.Termios
|
||||||
|
_, _, err := syscall.Syscall6(syscall.SYS_IOCTL, fd, ioctlReadTermios, uintptr(unsafe.Pointer(&termios)), 0, 0, 0)
|
||||||
|
return err == 0
|
||||||
|
}
|
16
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/isatty_solaris.go
generated
vendored
Normal file
16
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/isatty_solaris.go
generated
vendored
Normal file
|
@ -0,0 +1,16 @@
|
||||||
|
// +build solaris
|
||||||
|
// +build !appengine
|
||||||
|
|
||||||
|
package isatty
|
||||||
|
|
||||||
|
import (
|
||||||
|
"golang.org/x/sys/unix"
|
||||||
|
)
|
||||||
|
|
||||||
|
// IsTerminal returns true if the given file descriptor is a terminal.
|
||||||
|
// see: http://src.illumos.org/source/xref/illumos-gate/usr/src/lib/libbc/libc/gen/common/isatty.c
|
||||||
|
func IsTerminal(fd uintptr) bool {
|
||||||
|
var termio unix.Termio
|
||||||
|
err := unix.IoctlSetTermio(int(fd), unix.TCGETA, &termio)
|
||||||
|
return err == nil
|
||||||
|
}
|
19
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/isatty_windows.go
generated
vendored
Normal file
19
vendor/github.com/onsi/ginkgo/reporters/stenographer/support/go-isatty/isatty_windows.go
generated
vendored
Normal file
|
@ -0,0 +1,19 @@
|
||||||
|
// +build windows
|
||||||
|
// +build !appengine
|
||||||
|
|
||||||
|
package isatty
|
||||||
|
|
||||||
|
import (
|
||||||
|
"syscall"
|
||||||
|
"unsafe"
|
||||||
|
)
|
||||||
|
|
||||||
|
var kernel32 = syscall.NewLazyDLL("kernel32.dll")
|
||||||
|
var procGetConsoleMode = kernel32.NewProc("GetConsoleMode")
|
||||||
|
|
||||||
|
// IsTerminal return true if the file descriptor is terminal.
|
||||||
|
func IsTerminal(fd uintptr) bool {
|
||||||
|
var st uint32
|
||||||
|
r, _, e := syscall.Syscall(procGetConsoleMode.Addr(), 2, fd, uintptr(unsafe.Pointer(&st)), 0)
|
||||||
|
return r != 0 && e == 0
|
||||||
|
}
|
5
vendor/github.com/onsi/ginkgo/reporters/teamcity_reporter.go
generated
vendored
5
vendor/github.com/onsi/ginkgo/reporters/teamcity_reporter.go
generated
vendored
|
@ -10,10 +10,11 @@ package reporters
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/onsi/ginkgo/config"
|
|
||||||
"github.com/onsi/ginkgo/types"
|
|
||||||
"io"
|
"io"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
|
"github.com/onsi/ginkgo/config"
|
||||||
|
"github.com/onsi/ginkgo/types"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|
34
vendor/github.com/onsi/ginkgo/types/types.go
generated
vendored
34
vendor/github.com/onsi/ginkgo/types/types.go
generated
vendored
|
@ -1,9 +1,25 @@
|
||||||
package types
|
package types
|
||||||
|
|
||||||
import "time"
|
import (
|
||||||
|
"strconv"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
const GINKGO_FOCUS_EXIT_CODE = 197
|
const GINKGO_FOCUS_EXIT_CODE = 197
|
||||||
|
|
||||||
|
/*
|
||||||
|
SuiteSummary represents the a summary of the test suite and is passed to both
|
||||||
|
Reporter.SpecSuiteWillBegin
|
||||||
|
Reporter.SpecSuiteDidEnd
|
||||||
|
|
||||||
|
this is unfortunate as these two methods should receive different objects. When running in parallel
|
||||||
|
each node does not deterministically know how many specs it will end up running.
|
||||||
|
|
||||||
|
Unfortunately making such a change would break backward compatibility.
|
||||||
|
|
||||||
|
Until Ginkgo 2.0 comes out we will continue to reuse this struct but populate unkown fields
|
||||||
|
with -1.
|
||||||
|
*/
|
||||||
type SuiteSummary struct {
|
type SuiteSummary struct {
|
||||||
SuiteDescription string
|
SuiteDescription string
|
||||||
SuiteSucceeded bool
|
SuiteSucceeded bool
|
||||||
|
@ -16,7 +32,10 @@ type SuiteSummary struct {
|
||||||
NumberOfSkippedSpecs int
|
NumberOfSkippedSpecs int
|
||||||
NumberOfPassedSpecs int
|
NumberOfPassedSpecs int
|
||||||
NumberOfFailedSpecs int
|
NumberOfFailedSpecs int
|
||||||
RunTime time.Duration
|
// Flaked specs are those that failed initially, but then passed on a
|
||||||
|
// subsequent try.
|
||||||
|
NumberOfFlakedSpecs int
|
||||||
|
RunTime time.Duration
|
||||||
}
|
}
|
||||||
|
|
||||||
type SpecSummary struct {
|
type SpecSummary struct {
|
||||||
|
@ -100,6 +119,17 @@ type SpecMeasurement struct {
|
||||||
LargestLabel string
|
LargestLabel string
|
||||||
AverageLabel string
|
AverageLabel string
|
||||||
Units string
|
Units string
|
||||||
|
Precision int
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s SpecMeasurement) PrecisionFmt() string {
|
||||||
|
if s.Precision == 0 {
|
||||||
|
return "%f"
|
||||||
|
}
|
||||||
|
|
||||||
|
str := strconv.Itoa(s.Precision)
|
||||||
|
|
||||||
|
return "%." + str + "f"
|
||||||
}
|
}
|
||||||
|
|
||||||
type SpecState uint
|
type SpecState uint
|
||||||
|
|
3
vendor/github.com/onsi/gomega/.gitignore
generated
vendored
3
vendor/github.com/onsi/gomega/.gitignore
generated
vendored
|
@ -1,3 +0,0 @@
|
||||||
.DS_Store
|
|
||||||
*.test
|
|
||||||
.
|
|
11
vendor/github.com/onsi/gomega/.travis.yml
generated
vendored
11
vendor/github.com/onsi/gomega/.travis.yml
generated
vendored
|
@ -1,11 +0,0 @@
|
||||||
language: go
|
|
||||||
go:
|
|
||||||
- 1.4
|
|
||||||
- 1.5
|
|
||||||
|
|
||||||
install:
|
|
||||||
- go get -v ./...
|
|
||||||
- go get github.com/onsi/ginkgo
|
|
||||||
- go install github.com/onsi/ginkgo/ginkgo
|
|
||||||
|
|
||||||
script: $HOME/gopath/bin/ginkgo -r --randomizeAllSpecs --failOnPending --randomizeSuites --race
|
|
68
vendor/github.com/onsi/gomega/CHANGELOG.md
generated
vendored
68
vendor/github.com/onsi/gomega/CHANGELOG.md
generated
vendored
|
@ -1,68 +0,0 @@
|
||||||
## HEAD
|
|
||||||
|
|
||||||
Improvements:
|
|
||||||
|
|
||||||
- Added `BeSent` which attempts to send a value down a channel and fails if the attempt blocks. Can be paired with `Eventually` to safely send a value down a channel with a timeout.
|
|
||||||
- `Ω`, `Expect`, `Eventually`, and `Consistently` now immediately `panic` if there is no registered fail handler. This is always a mistake that can hide failing tests.
|
|
||||||
- `Receive()` no longer errors when passed a closed channel, it's perfectly fine to attempt to read from a closed channel so Ω(c).Should(Receive()) always fails and Ω(c).ShoudlNot(Receive()) always passes with a closed channel.
|
|
||||||
- Added `HavePrefix` and `HaveSuffix` matchers.
|
|
||||||
- `ghttp` can now handle concurrent requests.
|
|
||||||
- Added `Succeed` which allows one to write `Ω(MyFunction()).Should(Succeed())`.
|
|
||||||
- Improved `ghttp`'s behavior around failing assertions and panics:
|
|
||||||
- If a registered handler makes a failing assertion `ghttp` will return `500`.
|
|
||||||
- If a registered handler panics, `ghttp` will return `500` *and* fail the test. This is new behavior that may cause existing code to break. This code is almost certainly incorrect and creating a false positive.
|
|
||||||
- `ghttp` servers can take an `io.Writer`. `ghttp` will write a line to the writer when each request arrives.
|
|
||||||
|
|
||||||
Bug Fixes:
|
|
||||||
- gexec: `session.Wait` now uses `EventuallyWithOffset` to get the right line number in the failure.
|
|
||||||
- `ContainElement` no longer bails if a passed-in matcher errors.
|
|
||||||
|
|
||||||
## 1.0 (8/2/2014)
|
|
||||||
|
|
||||||
No changes. Dropping "beta" from the version number.
|
|
||||||
|
|
||||||
## 1.0.0-beta (7/8/2014)
|
|
||||||
Breaking Changes:
|
|
||||||
|
|
||||||
- Changed OmegaMatcher interface. Instead of having `Match` return failure messages, two new methods `FailureMessage` and `NegatedFailureMessage` are called instead.
|
|
||||||
- Moved and renamed OmegaFailHandler to types.GomegaFailHandler and OmegaMatcher to types.GomegaMatcher. Any references to OmegaMatcher in any custom matchers will need to be changed to point to types.GomegaMatcher
|
|
||||||
|
|
||||||
New Test-Support Features:
|
|
||||||
|
|
||||||
- `ghttp`: supports testing http clients
|
|
||||||
- Provides a flexible fake http server
|
|
||||||
- Provides a collection of chainable http handlers that perform assertions.
|
|
||||||
- `gbytes`: supports making ordered assertions against streams of data
|
|
||||||
- Provides a `gbytes.Buffer`
|
|
||||||
- Provides a `Say` matcher to perform ordered assertions against output data
|
|
||||||
- `gexec`: supports testing external processes
|
|
||||||
- Provides support for building Go binaries
|
|
||||||
- Wraps and starts `exec.Cmd` commands
|
|
||||||
- Makes it easy to assert against stdout and stderr
|
|
||||||
- Makes it easy to send signals and wait for processes to exit
|
|
||||||
- Provides an `Exit` matcher to assert against exit code.
|
|
||||||
|
|
||||||
DSL Changes:
|
|
||||||
|
|
||||||
- `Eventually` and `Consistently` can accept `time.Duration` interval and polling inputs.
|
|
||||||
- The default timeouts for `Eventually` and `Consistently` are now configurable.
|
|
||||||
|
|
||||||
New Matchers:
|
|
||||||
|
|
||||||
- `ConsistOf`: order-independent assertion against the elements of an array/slice or keys of a map.
|
|
||||||
- `BeTemporally`: like `BeNumerically` but for `time.Time`
|
|
||||||
- `HaveKeyWithValue`: asserts a map has a given key with the given value.
|
|
||||||
|
|
||||||
Updated Matchers:
|
|
||||||
|
|
||||||
- `Receive` matcher can take a matcher as an argument and passes only if the channel under test receives an objet that satisfies the passed-in matcher.
|
|
||||||
- Matchers that implement `MatchMayChangeInTheFuture(actual interface{}) bool` can inform `Eventually` and/or `Consistently` when a match has no chance of changing status in the future. For example, `Receive` returns `false` when a channel is closed.
|
|
||||||
|
|
||||||
Misc:
|
|
||||||
|
|
||||||
- Start using semantic versioning
|
|
||||||
- Start maintaining changelog
|
|
||||||
|
|
||||||
Major refactor:
|
|
||||||
|
|
||||||
- Pull out Gomega's internal to `internal`
|
|
17
vendor/github.com/onsi/gomega/README.md
generated
vendored
17
vendor/github.com/onsi/gomega/README.md
generated
vendored
|
@ -1,17 +0,0 @@
|
||||||
![Gomega: Ginkgo's Preferred Matcher Library](http://onsi.github.io/gomega/images/gomega.png)
|
|
||||||
|
|
||||||
[![Build Status](https://travis-ci.org/onsi/gomega.png)](https://travis-ci.org/onsi/gomega)
|
|
||||||
|
|
||||||
Jump straight to the [docs](http://onsi.github.io/gomega/) to learn about Gomega, including a list of [all available matchers](http://onsi.github.io/gomega/#provided-matchers).
|
|
||||||
|
|
||||||
To discuss Gomega and get updates, join the [google group](https://groups.google.com/d/forum/ginkgo-and-gomega).
|
|
||||||
|
|
||||||
## [Ginkgo](http://github.com/onsi/ginkgo): a BDD Testing Framework for Golang
|
|
||||||
|
|
||||||
Learn more about Ginkgo [here](http://onsi.github.io/ginkgo/)
|
|
||||||
|
|
||||||
## License
|
|
||||||
|
|
||||||
Gomega is MIT-Licensed
|
|
||||||
|
|
||||||
The `ConsistOf` matcher uses [goraph](https://github.com/amitkgupta/goraph) which is embedded in the source to simplify distribution. goraph has an MIT license.
|
|
Some files were not shown because too many files have changed in this diff Show more
Loading…
Reference in a new issue