Rename scripts to build and add revive command as a new build tool command (#10942)
Co-authored-by: techknowlogick <techknowlogick@gitea.io>tokarchuk/v1.17
parent
4af7c47b38
commit
4f63f283c4
@ -0,0 +1,325 @@ |
|||||||
|
// Copyright 2020 The Gitea Authors. All rights reserved.
|
||||||
|
// Copyright (c) 2018 Minko Gechev. All rights reserved.
|
||||||
|
// Use of this source code is governed by a MIT-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
// +build ignore
|
||||||
|
|
||||||
|
package main |
||||||
|
|
||||||
|
import ( |
||||||
|
"flag" |
||||||
|
"fmt" |
||||||
|
"io/ioutil" |
||||||
|
"os" |
||||||
|
"path/filepath" |
||||||
|
"strings" |
||||||
|
|
||||||
|
"github.com/BurntSushi/toml" |
||||||
|
"github.com/mgechev/dots" |
||||||
|
"github.com/mgechev/revive/formatter" |
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
"github.com/mgechev/revive/rule" |
||||||
|
"github.com/mitchellh/go-homedir" |
||||||
|
) |
||||||
|
|
||||||
|
func fail(err string) { |
||||||
|
fmt.Fprintln(os.Stderr, err) |
||||||
|
os.Exit(1) |
||||||
|
} |
||||||
|
|
||||||
|
var defaultRules = []lint.Rule{ |
||||||
|
&rule.VarDeclarationsRule{}, |
||||||
|
&rule.PackageCommentsRule{}, |
||||||
|
&rule.DotImportsRule{}, |
||||||
|
&rule.BlankImportsRule{}, |
||||||
|
&rule.ExportedRule{}, |
||||||
|
&rule.VarNamingRule{}, |
||||||
|
&rule.IndentErrorFlowRule{}, |
||||||
|
&rule.IfReturnRule{}, |
||||||
|
&rule.RangeRule{}, |
||||||
|
&rule.ErrorfRule{}, |
||||||
|
&rule.ErrorNamingRule{}, |
||||||
|
&rule.ErrorStringsRule{}, |
||||||
|
&rule.ReceiverNamingRule{}, |
||||||
|
&rule.IncrementDecrementRule{}, |
||||||
|
&rule.ErrorReturnRule{}, |
||||||
|
&rule.UnexportedReturnRule{}, |
||||||
|
&rule.TimeNamingRule{}, |
||||||
|
&rule.ContextKeysType{}, |
||||||
|
&rule.ContextAsArgumentRule{}, |
||||||
|
} |
||||||
|
|
||||||
|
var allRules = append([]lint.Rule{ |
||||||
|
&rule.ArgumentsLimitRule{}, |
||||||
|
&rule.CyclomaticRule{}, |
||||||
|
&rule.FileHeaderRule{}, |
||||||
|
&rule.EmptyBlockRule{}, |
||||||
|
&rule.SuperfluousElseRule{}, |
||||||
|
&rule.ConfusingNamingRule{}, |
||||||
|
&rule.GetReturnRule{}, |
||||||
|
&rule.ModifiesParamRule{}, |
||||||
|
&rule.ConfusingResultsRule{}, |
||||||
|
&rule.DeepExitRule{}, |
||||||
|
&rule.UnusedParamRule{}, |
||||||
|
&rule.UnreachableCodeRule{}, |
||||||
|
&rule.AddConstantRule{}, |
||||||
|
&rule.FlagParamRule{}, |
||||||
|
&rule.UnnecessaryStmtRule{}, |
||||||
|
&rule.StructTagRule{}, |
||||||
|
&rule.ModifiesValRecRule{}, |
||||||
|
&rule.ConstantLogicalExprRule{}, |
||||||
|
&rule.BoolLiteralRule{}, |
||||||
|
&rule.RedefinesBuiltinIDRule{}, |
||||||
|
&rule.ImportsBlacklistRule{}, |
||||||
|
&rule.FunctionResultsLimitRule{}, |
||||||
|
&rule.MaxPublicStructsRule{}, |
||||||
|
&rule.RangeValInClosureRule{}, |
||||||
|
&rule.RangeValAddress{}, |
||||||
|
&rule.WaitGroupByValueRule{}, |
||||||
|
&rule.AtomicRule{}, |
||||||
|
&rule.EmptyLinesRule{}, |
||||||
|
&rule.LineLengthLimitRule{}, |
||||||
|
&rule.CallToGCRule{}, |
||||||
|
&rule.DuplicatedImportsRule{}, |
||||||
|
&rule.ImportShadowingRule{}, |
||||||
|
&rule.BareReturnRule{}, |
||||||
|
&rule.UnusedReceiverRule{}, |
||||||
|
&rule.UnhandledErrorRule{}, |
||||||
|
&rule.CognitiveComplexityRule{}, |
||||||
|
&rule.StringOfIntRule{}, |
||||||
|
}, defaultRules...) |
||||||
|
|
||||||
|
var allFormatters = []lint.Formatter{ |
||||||
|
&formatter.Stylish{}, |
||||||
|
&formatter.Friendly{}, |
||||||
|
&formatter.JSON{}, |
||||||
|
&formatter.NDJSON{}, |
||||||
|
&formatter.Default{}, |
||||||
|
&formatter.Unix{}, |
||||||
|
&formatter.Checkstyle{}, |
||||||
|
&formatter.Plain{}, |
||||||
|
} |
||||||
|
|
||||||
|
func getFormatters() map[string]lint.Formatter { |
||||||
|
result := map[string]lint.Formatter{} |
||||||
|
for _, f := range allFormatters { |
||||||
|
result[f.Name()] = f |
||||||
|
} |
||||||
|
return result |
||||||
|
} |
||||||
|
|
||||||
|
func getLintingRules(config *lint.Config) []lint.Rule { |
||||||
|
rulesMap := map[string]lint.Rule{} |
||||||
|
for _, r := range allRules { |
||||||
|
rulesMap[r.Name()] = r |
||||||
|
} |
||||||
|
|
||||||
|
lintingRules := []lint.Rule{} |
||||||
|
for name := range config.Rules { |
||||||
|
rule, ok := rulesMap[name] |
||||||
|
if !ok { |
||||||
|
fail("cannot find rule: " + name) |
||||||
|
} |
||||||
|
lintingRules = append(lintingRules, rule) |
||||||
|
} |
||||||
|
|
||||||
|
return lintingRules |
||||||
|
} |
||||||
|
|
||||||
|
func parseConfig(path string) *lint.Config { |
||||||
|
config := &lint.Config{} |
||||||
|
file, err := ioutil.ReadFile(path) |
||||||
|
if err != nil { |
||||||
|
fail("cannot read the config file") |
||||||
|
} |
||||||
|
_, err = toml.Decode(string(file), config) |
||||||
|
if err != nil { |
||||||
|
fail("cannot parse the config file: " + err.Error()) |
||||||
|
} |
||||||
|
return config |
||||||
|
} |
||||||
|
|
||||||
|
func normalizeConfig(config *lint.Config) { |
||||||
|
if config.Confidence == 0 { |
||||||
|
config.Confidence = 0.8 |
||||||
|
} |
||||||
|
severity := config.Severity |
||||||
|
if severity != "" { |
||||||
|
for k, v := range config.Rules { |
||||||
|
if v.Severity == "" { |
||||||
|
v.Severity = severity |
||||||
|
} |
||||||
|
config.Rules[k] = v |
||||||
|
} |
||||||
|
for k, v := range config.Directives { |
||||||
|
if v.Severity == "" { |
||||||
|
v.Severity = severity |
||||||
|
} |
||||||
|
config.Directives[k] = v |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
func getConfig() *lint.Config { |
||||||
|
config := defaultConfig() |
||||||
|
if configPath != "" { |
||||||
|
config = parseConfig(configPath) |
||||||
|
} |
||||||
|
normalizeConfig(config) |
||||||
|
return config |
||||||
|
} |
||||||
|
|
||||||
|
func getFormatter() lint.Formatter { |
||||||
|
formatters := getFormatters() |
||||||
|
formatter := formatters["default"] |
||||||
|
if formatterName != "" { |
||||||
|
f, ok := formatters[formatterName] |
||||||
|
if !ok { |
||||||
|
fail("unknown formatter " + formatterName) |
||||||
|
} |
||||||
|
formatter = f |
||||||
|
} |
||||||
|
return formatter |
||||||
|
} |
||||||
|
|
||||||
|
func buildDefaultConfigPath() string { |
||||||
|
var result string |
||||||
|
if homeDir, err := homedir.Dir(); err == nil { |
||||||
|
result = filepath.Join(homeDir, "revive.toml") |
||||||
|
if _, err := os.Stat(result); err != nil { |
||||||
|
result = "" |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
return result |
||||||
|
} |
||||||
|
|
||||||
|
func defaultConfig() *lint.Config { |
||||||
|
defaultConfig := lint.Config{ |
||||||
|
Confidence: 0.0, |
||||||
|
Severity: lint.SeverityWarning, |
||||||
|
Rules: map[string]lint.RuleConfig{}, |
||||||
|
} |
||||||
|
for _, r := range defaultRules { |
||||||
|
defaultConfig.Rules[r.Name()] = lint.RuleConfig{} |
||||||
|
} |
||||||
|
return &defaultConfig |
||||||
|
} |
||||||
|
|
||||||
|
func normalizeSplit(strs []string) []string { |
||||||
|
res := []string{} |
||||||
|
for _, s := range strs { |
||||||
|
t := strings.Trim(s, " \t") |
||||||
|
if len(t) > 0 { |
||||||
|
res = append(res, t) |
||||||
|
} |
||||||
|
} |
||||||
|
return res |
||||||
|
} |
||||||
|
|
||||||
|
func getPackages() [][]string { |
||||||
|
globs := normalizeSplit(flag.Args()) |
||||||
|
if len(globs) == 0 { |
||||||
|
globs = append(globs, ".") |
||||||
|
} |
||||||
|
|
||||||
|
packages, err := dots.ResolvePackages(globs, normalizeSplit(excludePaths)) |
||||||
|
if err != nil { |
||||||
|
fail(err.Error()) |
||||||
|
} |
||||||
|
|
||||||
|
return packages |
||||||
|
} |
||||||
|
|
||||||
|
type arrayFlags []string |
||||||
|
|
||||||
|
func (i *arrayFlags) String() string { |
||||||
|
return strings.Join([]string(*i), " ") |
||||||
|
} |
||||||
|
|
||||||
|
func (i *arrayFlags) Set(value string) error { |
||||||
|
*i = append(*i, value) |
||||||
|
return nil |
||||||
|
} |
||||||
|
|
||||||
|
var configPath string |
||||||
|
var excludePaths arrayFlags |
||||||
|
var formatterName string |
||||||
|
var help bool |
||||||
|
|
||||||
|
var originalUsage = flag.Usage |
||||||
|
|
||||||
|
func init() { |
||||||
|
flag.Usage = func() { |
||||||
|
originalUsage() |
||||||
|
} |
||||||
|
// command line help strings
|
||||||
|
const ( |
||||||
|
configUsage = "path to the configuration TOML file, defaults to $HOME/revive.toml, if present (i.e. -config myconf.toml)" |
||||||
|
excludeUsage = "list of globs which specify files to be excluded (i.e. -exclude foo/...)" |
||||||
|
formatterUsage = "formatter to be used for the output (i.e. -formatter stylish)" |
||||||
|
) |
||||||
|
|
||||||
|
defaultConfigPath := buildDefaultConfigPath() |
||||||
|
|
||||||
|
flag.StringVar(&configPath, "config", defaultConfigPath, configUsage) |
||||||
|
flag.Var(&excludePaths, "exclude", excludeUsage) |
||||||
|
flag.StringVar(&formatterName, "formatter", "", formatterUsage) |
||||||
|
flag.Parse() |
||||||
|
} |
||||||
|
|
||||||
|
func main() { |
||||||
|
config := getConfig() |
||||||
|
formatter := getFormatter() |
||||||
|
packages := getPackages() |
||||||
|
|
||||||
|
revive := lint.New(func(file string) ([]byte, error) { |
||||||
|
return ioutil.ReadFile(file) |
||||||
|
}) |
||||||
|
|
||||||
|
lintingRules := getLintingRules(config) |
||||||
|
|
||||||
|
failures, err := revive.Lint(packages, lintingRules, *config) |
||||||
|
if err != nil { |
||||||
|
fail(err.Error()) |
||||||
|
} |
||||||
|
|
||||||
|
formatChan := make(chan lint.Failure) |
||||||
|
exitChan := make(chan bool) |
||||||
|
|
||||||
|
var output string |
||||||
|
go (func() { |
||||||
|
output, err = formatter.Format(formatChan, *config) |
||||||
|
if err != nil { |
||||||
|
fail(err.Error()) |
||||||
|
} |
||||||
|
exitChan <- true |
||||||
|
})() |
||||||
|
|
||||||
|
exitCode := 0 |
||||||
|
for f := range failures { |
||||||
|
if f.Confidence < config.Confidence { |
||||||
|
continue |
||||||
|
} |
||||||
|
if exitCode == 0 { |
||||||
|
exitCode = config.WarningCode |
||||||
|
} |
||||||
|
if c, ok := config.Rules[f.RuleName]; ok && c.Severity == lint.SeverityError { |
||||||
|
exitCode = config.ErrorCode |
||||||
|
} |
||||||
|
if c, ok := config.Directives[f.RuleName]; ok && c.Severity == lint.SeverityError { |
||||||
|
exitCode = config.ErrorCode |
||||||
|
} |
||||||
|
|
||||||
|
formatChan <- f |
||||||
|
} |
||||||
|
|
||||||
|
close(formatChan) |
||||||
|
<-exitChan |
||||||
|
if output != "" { |
||||||
|
fmt.Println(output) |
||||||
|
} |
||||||
|
|
||||||
|
os.Exit(exitCode) |
||||||
|
} |
@ -0,0 +1,18 @@ |
|||||||
|
// Copyright 2020 The Gitea Authors. All rights reserved.
|
||||||
|
// Use of this source code is governed by a MIT-style
|
||||||
|
// license that can be found in the LICENSE file.
|
||||||
|
|
||||||
|
package build |
||||||
|
|
||||||
|
import ( |
||||||
|
// for lint
|
||||||
|
_ "github.com/BurntSushi/toml" |
||||||
|
_ "github.com/mgechev/dots" |
||||||
|
_ "github.com/mgechev/revive/formatter" |
||||||
|
_ "github.com/mgechev/revive/lint" |
||||||
|
_ "github.com/mgechev/revive/rule" |
||||||
|
_ "github.com/mitchellh/go-homedir" |
||||||
|
|
||||||
|
// for embed
|
||||||
|
_ "github.com/shurcooL/vfsgen" |
||||||
|
) |
@ -0,0 +1,20 @@ |
|||||||
|
The MIT License (MIT) |
||||||
|
|
||||||
|
Copyright (c) 2013 Fatih Arslan |
||||||
|
|
||||||
|
Permission is hereby granted, free of charge, to any person obtaining a copy of |
||||||
|
this software and associated documentation files (the "Software"), to deal in |
||||||
|
the Software without restriction, including without limitation the rights to |
||||||
|
use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of |
||||||
|
the Software, and to permit persons to whom the Software is furnished to do so, |
||||||
|
subject to the following conditions: |
||||||
|
|
||||||
|
The above copyright notice and this permission notice shall be included in all |
||||||
|
copies or substantial portions of the Software. |
||||||
|
|
||||||
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
||||||
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS |
||||||
|
FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR |
||||||
|
COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER |
||||||
|
IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN |
||||||
|
CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. |
@ -0,0 +1,182 @@ |
|||||||
|
# Archived project. No maintenance. |
||||||
|
|
||||||
|
This project is not maintained anymore and is archived. Feel free to fork and |
||||||
|
make your own changes if needed. For more detail read my blog post: [Taking an indefinite sabbatical from my projects](https://arslan.io/2018/10/09/taking-an-indefinite-sabbatical-from-my-projects/) |
||||||
|
|
||||||
|
Thanks to everyone for their valuable feedback and contributions. |
||||||
|
|
||||||
|
|
||||||
|
# Color [![GoDoc](https://godoc.org/github.com/fatih/color?status.svg)](https://godoc.org/github.com/fatih/color) |
||||||
|
|
||||||
|
Color lets you use colorized outputs in terms of [ANSI Escape |
||||||
|
Codes](http://en.wikipedia.org/wiki/ANSI_escape_code#Colors) in Go (Golang). It |
||||||
|
has support for Windows too! The API can be used in several ways, pick one that |
||||||
|
suits you. |
||||||
|
|
||||||
|
|
||||||
|
![Color](https://i.imgur.com/c1JI0lA.png) |
||||||
|
|
||||||
|
|
||||||
|
## Install |
||||||
|
|
||||||
|
```bash |
||||||
|
go get github.com/fatih/color |
||||||
|
``` |
||||||
|
|
||||||
|
## Examples |
||||||
|
|
||||||
|
### Standard colors |
||||||
|
|
||||||
|
```go |
||||||
|
// Print with default helper functions |
||||||
|
color.Cyan("Prints text in cyan.") |
||||||
|
|
||||||
|
// A newline will be appended automatically |
||||||
|
color.Blue("Prints %s in blue.", "text") |
||||||
|
|
||||||
|
// These are using the default foreground colors |
||||||
|
color.Red("We have red") |
||||||
|
color.Magenta("And many others ..") |
||||||
|
|
||||||
|
``` |
||||||
|
|
||||||
|
### Mix and reuse colors |
||||||
|
|
||||||
|
```go |
||||||
|
// Create a new color object |
||||||
|
c := color.New(color.FgCyan).Add(color.Underline) |
||||||
|
c.Println("Prints cyan text with an underline.") |
||||||
|
|
||||||
|
// Or just add them to New() |
||||||
|
d := color.New(color.FgCyan, color.Bold) |
||||||
|
d.Printf("This prints bold cyan %s\n", "too!.") |
||||||
|
|
||||||
|
// Mix up foreground and background colors, create new mixes! |
||||||
|
red := color.New(color.FgRed) |
||||||
|
|
||||||
|
boldRed := red.Add(color.Bold) |
||||||
|
boldRed.Println("This will print text in bold red.") |
||||||
|
|
||||||
|
whiteBackground := red.Add(color.BgWhite) |
||||||
|
whiteBackground.Println("Red text with white background.") |
||||||
|
``` |
||||||
|
|
||||||
|
### Use your own output (io.Writer) |
||||||
|
|
||||||
|
```go |
||||||
|
// Use your own io.Writer output |
||||||
|
color.New(color.FgBlue).Fprintln(myWriter, "blue color!") |
||||||
|
|
||||||
|
blue := color.New(color.FgBlue) |
||||||
|
blue.Fprint(writer, "This will print text in blue.") |
||||||
|
``` |
||||||
|
|
||||||
|
### Custom print functions (PrintFunc) |
||||||
|
|
||||||
|
```go |
||||||
|
// Create a custom print function for convenience |
||||||
|
red := color.New(color.FgRed).PrintfFunc() |
||||||
|
red("Warning") |
||||||
|
red("Error: %s", err) |
||||||
|
|
||||||
|
// Mix up multiple attributes |
||||||
|
notice := color.New(color.Bold, color.FgGreen).PrintlnFunc() |
||||||
|
notice("Don't forget this...") |
||||||
|
``` |
||||||
|
|
||||||
|
### Custom fprint functions (FprintFunc) |
||||||
|
|
||||||
|
```go |
||||||
|
blue := color.New(FgBlue).FprintfFunc() |
||||||
|
blue(myWriter, "important notice: %s", stars) |
||||||
|
|
||||||
|
// Mix up with multiple attributes |
||||||
|
success := color.New(color.Bold, color.FgGreen).FprintlnFunc() |
||||||
|
success(myWriter, "Don't forget this...") |
||||||
|
``` |
||||||
|
|
||||||
|
### Insert into noncolor strings (SprintFunc) |
||||||
|
|
||||||
|
```go |
||||||
|
// Create SprintXxx functions to mix strings with other non-colorized strings: |
||||||
|
yellow := color.New(color.FgYellow).SprintFunc() |
||||||
|
red := color.New(color.FgRed).SprintFunc() |
||||||
|
fmt.Printf("This is a %s and this is %s.\n", yellow("warning"), red("error")) |
||||||
|
|
||||||
|
info := color.New(color.FgWhite, color.BgGreen).SprintFunc() |
||||||
|
fmt.Printf("This %s rocks!\n", info("package")) |
||||||
|
|
||||||
|
// Use helper functions |
||||||
|
fmt.Println("This", color.RedString("warning"), "should be not neglected.") |
||||||
|
fmt.Printf("%v %v\n", color.GreenString("Info:"), "an important message.") |
||||||
|
|
||||||
|
// Windows supported too! Just don't forget to change the output to color.Output |
||||||
|
fmt.Fprintf(color.Output, "Windows support: %s", color.GreenString("PASS")) |
||||||
|
``` |
||||||
|
|
||||||
|
### Plug into existing code |
||||||
|
|
||||||
|
```go |
||||||
|
// Use handy standard colors |
||||||
|
color.Set(color.FgYellow) |
||||||
|
|
||||||
|
fmt.Println("Existing text will now be in yellow") |
||||||
|
fmt.Printf("This one %s\n", "too") |
||||||
|
|
||||||
|
color.Unset() // Don't forget to unset |
||||||
|
|
||||||
|
// You can mix up parameters |
||||||
|
color.Set(color.FgMagenta, color.Bold) |
||||||
|
defer color.Unset() // Use it in your function |
||||||
|
|
||||||
|
fmt.Println("All text will now be bold magenta.") |
||||||
|
``` |
||||||
|
|
||||||
|
### Disable/Enable color |
||||||
|
|
||||||
|
There might be a case where you want to explicitly disable/enable color output. the |
||||||
|
`go-isatty` package will automatically disable color output for non-tty output streams |
||||||
|
(for example if the output were piped directly to `less`) |
||||||
|
|
||||||
|
`Color` has support to disable/enable colors both globally and for single color |
||||||
|
definitions. For example suppose you have a CLI app and a `--no-color` bool flag. You |
||||||
|
can easily disable the color output with: |
||||||
|
|
||||||
|
```go |
||||||
|
|
||||||
|
var flagNoColor = flag.Bool("no-color", false, "Disable color output") |
||||||
|
|
||||||
|
if *flagNoColor { |
||||||
|
color.NoColor = true // disables colorized output |
||||||
|
} |
||||||
|
``` |
||||||
|
|
||||||
|
It also has support for single color definitions (local). You can |
||||||
|
disable/enable color output on the fly: |
||||||
|
|
||||||
|
```go |
||||||
|
c := color.New(color.FgCyan) |
||||||
|
c.Println("Prints cyan text") |
||||||
|
|
||||||
|
c.DisableColor() |
||||||
|
c.Println("This is printed without any color") |
||||||
|
|
||||||
|
c.EnableColor() |
||||||
|
c.Println("This prints again cyan...") |
||||||
|
``` |
||||||
|
|
||||||
|
## Todo |
||||||
|
|
||||||
|
* Save/Return previous values |
||||||
|
* Evaluate fmt.Formatter interface |
||||||
|
|
||||||
|
|
||||||
|
## Credits |
||||||
|
|
||||||
|
* [Fatih Arslan](https://github.com/fatih) |
||||||
|
* Windows support via @mattn: [colorable](https://github.com/mattn/go-colorable) |
||||||
|
|
||||||
|
## License |
||||||
|
|
||||||
|
The MIT License (MIT) - see [`LICENSE.md`](https://github.com/fatih/color/blob/master/LICENSE.md) for more details |
||||||
|
|
@ -0,0 +1,603 @@ |
|||||||
|
package color |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"io" |
||||||
|
"os" |
||||||
|
"strconv" |
||||||
|
"strings" |
||||||
|
"sync" |
||||||
|
|
||||||
|
"github.com/mattn/go-colorable" |
||||||
|
"github.com/mattn/go-isatty" |
||||||
|
) |
||||||
|
|
||||||
|
var ( |
||||||
|
// NoColor defines if the output is colorized or not. It's dynamically set to
|
||||||
|
// false or true based on the stdout's file descriptor referring to a terminal
|
||||||
|
// or not. This is a global option and affects all colors. For more control
|
||||||
|
// over each color block use the methods DisableColor() individually.
|
||||||
|
NoColor = os.Getenv("TERM") == "dumb" || |
||||||
|
(!isatty.IsTerminal(os.Stdout.Fd()) && !isatty.IsCygwinTerminal(os.Stdout.Fd())) |
||||||
|
|
||||||
|
// Output defines the standard output of the print functions. By default
|
||||||
|
// os.Stdout is used.
|
||||||
|
Output = colorable.NewColorableStdout() |
||||||
|
|
||||||
|
// Error defines a color supporting writer for os.Stderr.
|
||||||
|
Error = colorable.NewColorableStderr() |
||||||
|
|
||||||
|
// colorsCache is used to reduce the count of created Color objects and
|
||||||
|
// allows to reuse already created objects with required Attribute.
|
||||||
|
colorsCache = make(map[Attribute]*Color) |
||||||
|
colorsCacheMu sync.Mutex // protects colorsCache
|
||||||
|
) |
||||||
|
|
||||||
|
// Color defines a custom color object which is defined by SGR parameters.
|
||||||
|
type Color struct { |
||||||
|
params []Attribute |
||||||
|
noColor *bool |
||||||
|
} |
||||||
|
|
||||||
|
// Attribute defines a single SGR Code
|
||||||
|
type Attribute int |
||||||
|
|
||||||
|
const escape = "\x1b" |
||||||
|
|
||||||
|
// Base attributes
|
||||||
|
const ( |
||||||
|
Reset Attribute = iota |
||||||
|
Bold |
||||||
|
Faint |
||||||
|
Italic |
||||||
|
Underline |
||||||
|
BlinkSlow |
||||||
|
BlinkRapid |
||||||
|
ReverseVideo |
||||||
|
Concealed |
||||||
|
CrossedOut |
||||||
|
) |
||||||
|
|
||||||
|
// Foreground text colors
|
||||||
|
const ( |
||||||
|
FgBlack Attribute = iota + 30 |
||||||
|
FgRed |
||||||
|
FgGreen |
||||||
|
FgYellow |
||||||
|
FgBlue |
||||||
|
FgMagenta |
||||||
|
FgCyan |
||||||
|
FgWhite |
||||||
|
) |
||||||
|
|
||||||
|
// Foreground Hi-Intensity text colors
|
||||||
|
const ( |
||||||
|
FgHiBlack Attribute = iota + 90 |
||||||
|
FgHiRed |
||||||
|
FgHiGreen |
||||||
|
FgHiYellow |
||||||
|
FgHiBlue |
||||||
|
FgHiMagenta |
||||||
|
FgHiCyan |
||||||
|
FgHiWhite |
||||||
|
) |
||||||
|
|
||||||
|
// Background text colors
|
||||||
|
const ( |
||||||
|
BgBlack Attribute = iota + 40 |
||||||
|
BgRed |
||||||
|
BgGreen |
||||||
|
BgYellow |
||||||
|
BgBlue |
||||||
|
BgMagenta |
||||||
|
BgCyan |
||||||
|
BgWhite |
||||||
|
) |
||||||
|
|
||||||
|
// Background Hi-Intensity text colors
|
||||||
|
const ( |
||||||
|
BgHiBlack Attribute = iota + 100 |
||||||
|
BgHiRed |
||||||
|
BgHiGreen |
||||||
|
BgHiYellow |
||||||
|
BgHiBlue |
||||||
|
BgHiMagenta |
||||||
|
BgHiCyan |
||||||
|
BgHiWhite |
||||||
|
) |
||||||
|
|
||||||
|
// New returns a newly created color object.
|
||||||
|
func New(value ...Attribute) *Color { |
||||||
|
c := &Color{params: make([]Attribute, 0)} |
||||||
|
c.Add(value...) |
||||||
|
return c |
||||||
|
} |
||||||
|
|
||||||
|
// Set sets the given parameters immediately. It will change the color of
|
||||||
|
// output with the given SGR parameters until color.Unset() is called.
|
||||||
|
func Set(p ...Attribute) *Color { |
||||||
|
c := New(p...) |
||||||
|
c.Set() |
||||||
|
return c |
||||||
|
} |
||||||
|
|
||||||
|
// Unset resets all escape attributes and clears the output. Usually should
|
||||||
|
// be called after Set().
|
||||||
|
func Unset() { |
||||||
|
if NoColor { |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
fmt.Fprintf(Output, "%s[%dm", escape, Reset) |
||||||
|
} |
||||||
|
|
||||||
|
// Set sets the SGR sequence.
|
||||||
|
func (c *Color) Set() *Color { |
||||||
|
if c.isNoColorSet() { |
||||||
|
return c |
||||||
|
} |
||||||
|
|
||||||
|
fmt.Fprintf(Output, c.format()) |
||||||
|
return c |
||||||
|
} |
||||||
|
|
||||||
|
func (c *Color) unset() { |
||||||
|
if c.isNoColorSet() { |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
Unset() |
||||||
|
} |
||||||
|
|
||||||
|
func (c *Color) setWriter(w io.Writer) *Color { |
||||||
|
if c.isNoColorSet() { |
||||||
|
return c |
||||||
|
} |
||||||
|
|
||||||
|
fmt.Fprintf(w, c.format()) |
||||||
|
return c |
||||||
|
} |
||||||
|
|
||||||
|
func (c *Color) unsetWriter(w io.Writer) { |
||||||
|
if c.isNoColorSet() { |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
if NoColor { |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
fmt.Fprintf(w, "%s[%dm", escape, Reset) |
||||||
|
} |
||||||
|
|
||||||
|
// Add is used to chain SGR parameters. Use as many as parameters to combine
|
||||||
|
// and create custom color objects. Example: Add(color.FgRed, color.Underline).
|
||||||
|
func (c *Color) Add(value ...Attribute) *Color { |
||||||
|
c.params = append(c.params, value...) |
||||||
|
return c |
||||||
|
} |
||||||
|
|
||||||
|
func (c *Color) prepend(value Attribute) { |
||||||
|
c.params = append(c.params, 0) |
||||||
|
copy(c.params[1:], c.params[0:]) |
||||||
|
c.params[0] = value |
||||||
|
} |
||||||
|
|
||||||
|
// Fprint formats using the default formats for its operands and writes to w.
|
||||||
|
// Spaces are added between operands when neither is a string.
|
||||||
|
// It returns the number of bytes written and any write error encountered.
|
||||||
|
// On Windows, users should wrap w with colorable.NewColorable() if w is of
|
||||||
|
// type *os.File.
|
||||||
|
func (c *Color) Fprint(w io.Writer, a ...interface{}) (n int, err error) { |
||||||
|
c.setWriter(w) |
||||||
|
defer c.unsetWriter(w) |
||||||
|
|
||||||
|
return fmt.Fprint(w, a...) |
||||||
|
} |
||||||
|
|
||||||
|
// Print formats using the default formats for its operands and writes to
|
||||||
|
// standard output. Spaces are added between operands when neither is a
|
||||||
|
// string. It returns the number of bytes written and any write error
|
||||||
|
// encountered. This is the standard fmt.Print() method wrapped with the given
|
||||||
|
// color.
|
||||||
|
func (c *Color) Print(a ...interface{}) (n int, err error) { |
||||||
|
c.Set() |
||||||
|
defer c.unset() |
||||||
|
|
||||||
|
return fmt.Fprint(Output, a...) |
||||||
|
} |
||||||
|
|
||||||
|
// Fprintf formats according to a format specifier and writes to w.
|
||||||
|
// It returns the number of bytes written and any write error encountered.
|
||||||
|
// On Windows, users should wrap w with colorable.NewColorable() if w is of
|
||||||
|
// type *os.File.
|
||||||
|
func (c *Color) Fprintf(w io.Writer, format string, a ...interface{}) (n int, err error) { |
||||||
|
c.setWriter(w) |
||||||
|
defer c.unsetWriter(w) |
||||||
|
|
||||||
|
return fmt.Fprintf(w, format, a...) |
||||||
|
} |
||||||
|
|
||||||
|
// Printf formats according to a format specifier and writes to standard output.
|
||||||
|
// It returns the number of bytes written and any write error encountered.
|
||||||
|
// This is the standard fmt.Printf() method wrapped with the given color.
|
||||||
|
func (c *Color) Printf(format string, a ...interface{}) (n int, err error) { |
||||||
|
c.Set() |
||||||
|
defer c.unset() |
||||||
|
|
||||||
|
return fmt.Fprintf(Output, format, a...) |
||||||
|
} |
||||||
|
|
||||||
|
// Fprintln formats using the default formats for its operands and writes to w.
|
||||||
|
// Spaces are always added between operands and a newline is appended.
|
||||||
|
// On Windows, users should wrap w with colorable.NewColorable() if w is of
|
||||||
|
// type *os.File.
|
||||||
|
func (c *Color) Fprintln(w io.Writer, a ...interface{}) (n int, err error) { |
||||||
|
c.setWriter(w) |
||||||
|
defer c.unsetWriter(w) |
||||||
|
|
||||||
|
return fmt.Fprintln(w, a...) |
||||||
|
} |
||||||
|
|
||||||
|
// Println formats using the default formats for its operands and writes to
|
||||||
|
// standard output. Spaces are always added between operands and a newline is
|
||||||
|
// appended. It returns the number of bytes written and any write error
|
||||||
|
// encountered. This is the standard fmt.Print() method wrapped with the given
|
||||||
|
// color.
|
||||||
|
func (c *Color) Println(a ...interface{}) (n int, err error) { |
||||||
|
c.Set() |
||||||
|
defer c.unset() |
||||||
|
|
||||||
|
return fmt.Fprintln(Output, a...) |
||||||
|
} |
||||||
|
|
||||||
|
// Sprint is just like Print, but returns a string instead of printing it.
|
||||||
|
func (c *Color) Sprint(a ...interface{}) string { |
||||||
|
return c.wrap(fmt.Sprint(a...)) |
||||||
|
} |
||||||
|
|
||||||
|
// Sprintln is just like Println, but returns a string instead of printing it.
|
||||||
|
func (c *Color) Sprintln(a ...interface{}) string { |
||||||
|
return c.wrap(fmt.Sprintln(a...)) |
||||||
|
} |
||||||
|
|
||||||
|
// Sprintf is just like Printf, but returns a string instead of printing it.
|
||||||
|
func (c *Color) Sprintf(format string, a ...interface{}) string { |
||||||
|
return c.wrap(fmt.Sprintf(format, a...)) |
||||||
|
} |
||||||
|
|
||||||
|
// FprintFunc returns a new function that prints the passed arguments as
|
||||||
|
// colorized with color.Fprint().
|
||||||
|
func (c *Color) FprintFunc() func(w io.Writer, a ...interface{}) { |
||||||
|
return func(w io.Writer, a ...interface{}) { |
||||||
|
c.Fprint(w, a...) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// PrintFunc returns a new function that prints the passed arguments as
|
||||||
|
// colorized with color.Print().
|
||||||
|
func (c *Color) PrintFunc() func(a ...interface{}) { |
||||||
|
return func(a ...interface{}) { |
||||||
|
c.Print(a...) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// FprintfFunc returns a new function that prints the passed arguments as
|
||||||
|
// colorized with color.Fprintf().
|
||||||
|
func (c *Color) FprintfFunc() func(w io.Writer, format string, a ...interface{}) { |
||||||
|
return func(w io.Writer, format string, a ...interface{}) { |
||||||
|
c.Fprintf(w, format, a...) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// PrintfFunc returns a new function that prints the passed arguments as
|
||||||
|
// colorized with color.Printf().
|
||||||
|
func (c *Color) PrintfFunc() func(format string, a ...interface{}) { |
||||||
|
return func(format string, a ...interface{}) { |
||||||
|
c.Printf(format, a...) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// FprintlnFunc returns a new function that prints the passed arguments as
|
||||||
|
// colorized with color.Fprintln().
|
||||||
|
func (c *Color) FprintlnFunc() func(w io.Writer, a ...interface{}) { |
||||||
|
return func(w io.Writer, a ...interface{}) { |
||||||
|
c.Fprintln(w, a...) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// PrintlnFunc returns a new function that prints the passed arguments as
|
||||||
|
// colorized with color.Println().
|
||||||
|
func (c *Color) PrintlnFunc() func(a ...interface{}) { |
||||||
|
return func(a ...interface{}) { |
||||||
|
c.Println(a...) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// SprintFunc returns a new function that returns colorized strings for the
|
||||||
|
// given arguments with fmt.Sprint(). Useful to put into or mix into other
|
||||||
|
// string. Windows users should use this in conjunction with color.Output, example:
|
||||||
|
//
|
||||||
|
// put := New(FgYellow).SprintFunc()
|
||||||
|
// fmt.Fprintf(color.Output, "This is a %s", put("warning"))
|
||||||
|
func (c *Color) SprintFunc() func(a ...interface{}) string { |
||||||
|
return func(a ...interface{}) string { |
||||||
|
return c.wrap(fmt.Sprint(a...)) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// SprintfFunc returns a new function that returns colorized strings for the
|
||||||
|
// given arguments with fmt.Sprintf(). Useful to put into or mix into other
|
||||||
|
// string. Windows users should use this in conjunction with color.Output.
|
||||||
|
func (c *Color) SprintfFunc() func(format string, a ...interface{}) string { |
||||||
|
return func(format string, a ...interface{}) string { |
||||||
|
return c.wrap(fmt.Sprintf(format, a...)) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// SprintlnFunc returns a new function that returns colorized strings for the
|
||||||
|
// given arguments with fmt.Sprintln(). Useful to put into or mix into other
|
||||||
|
// string. Windows users should use this in conjunction with color.Output.
|
||||||
|
func (c *Color) SprintlnFunc() func(a ...interface{}) string { |
||||||
|
return func(a ...interface{}) string { |
||||||
|
return c.wrap(fmt.Sprintln(a...)) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// sequence returns a formatted SGR sequence to be plugged into a "\x1b[...m"
|
||||||
|
// an example output might be: "1;36" -> bold cyan
|
||||||
|
func (c *Color) sequence() string { |
||||||
|
format := make([]string, len(c.params)) |
||||||
|
for i, v := range c.params { |
||||||
|
format[i] = strconv.Itoa(int(v)) |
||||||
|
} |
||||||
|
|
||||||
|
return strings.Join(format, ";") |
||||||
|
} |
||||||
|
|
||||||
|
// wrap wraps the s string with the colors attributes. The string is ready to
|
||||||
|
// be printed.
|
||||||
|
func (c *Color) wrap(s string) string { |
||||||
|
if c.isNoColorSet() { |
||||||
|
return s |
||||||
|
} |
||||||
|
|
||||||
|
return c.format() + s + c.unformat() |
||||||
|
} |
||||||
|
|
||||||
|
func (c *Color) format() string { |
||||||
|
return fmt.Sprintf("%s[%sm", escape, c.sequence()) |
||||||
|
} |
||||||
|
|
||||||
|
func (c *Color) unformat() string { |
||||||
|
return fmt.Sprintf("%s[%dm", escape, Reset) |
||||||
|
} |
||||||
|
|
||||||
|
// DisableColor disables the color output. Useful to not change any existing
|
||||||
|
// code and still being able to output. Can be used for flags like
|
||||||
|
// "--no-color". To enable back use EnableColor() method.
|
||||||
|
func (c *Color) DisableColor() { |
||||||
|
c.noColor = boolPtr(true) |
||||||
|
} |
||||||
|
|
||||||
|
// EnableColor enables the color output. Use it in conjunction with
|
||||||
|
// DisableColor(). Otherwise this method has no side effects.
|
||||||
|
func (c *Color) EnableColor() { |
||||||
|
c.noColor = boolPtr(false) |
||||||
|
} |
||||||
|
|
||||||
|
func (c *Color) isNoColorSet() bool { |
||||||
|
// check first if we have user setted action
|
||||||
|
if c.noColor != nil { |
||||||
|
return *c.noColor |
||||||
|
} |
||||||
|
|
||||||
|
// if not return the global option, which is disabled by default
|
||||||
|
return NoColor |
||||||
|
} |
||||||
|
|
||||||
|
// Equals returns a boolean value indicating whether two colors are equal.
|
||||||
|
func (c *Color) Equals(c2 *Color) bool { |
||||||
|
if len(c.params) != len(c2.params) { |
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
for _, attr := range c.params { |
||||||
|
if !c2.attrExists(attr) { |
||||||
|
return false |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
return true |
||||||
|
} |
||||||
|
|
||||||
|
func (c *Color) attrExists(a Attribute) bool { |
||||||
|
for _, attr := range c.params { |
||||||
|
if attr == a { |
||||||
|
return true |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
func boolPtr(v bool) *bool { |
||||||
|
return &v |
||||||
|
} |
||||||
|
|
||||||
|
func getCachedColor(p Attribute) *Color { |
||||||
|
colorsCacheMu.Lock() |
||||||
|
defer colorsCacheMu.Unlock() |
||||||
|
|
||||||
|
c, ok := colorsCache[p] |
||||||
|
if !ok { |
||||||
|
c = New(p) |
||||||
|
colorsCache[p] = c |
||||||
|
} |
||||||
|
|
||||||
|
return c |
||||||
|
} |
||||||
|
|
||||||
|
func colorPrint(format string, p Attribute, a ...interface{}) { |
||||||
|
c := getCachedColor(p) |
||||||
|
|
||||||
|
if !strings.HasSuffix(format, "\n") { |
||||||
|
format += "\n" |
||||||
|
} |
||||||
|
|
||||||
|
if len(a) == 0 { |
||||||
|
c.Print(format) |
||||||
|
} else { |
||||||
|
c.Printf(format, a...) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
func colorString(format string, p Attribute, a ...interface{}) string { |
||||||
|
c := getCachedColor(p) |
||||||
|
|
||||||
|
if len(a) == 0 { |
||||||
|
return c.SprintFunc()(format) |
||||||
|
} |
||||||
|
|
||||||
|
return c.SprintfFunc()(format, a...) |
||||||
|
} |
||||||
|
|
||||||
|
// Black is a convenient helper function to print with black foreground. A
|
||||||
|
// newline is appended to format by default.
|
||||||
|
func Black(format string, a ...interface{}) { colorPrint(format, FgBlack, a...) } |
||||||
|
|
||||||
|
// Red is a convenient helper function to print with red foreground. A
|
||||||
|
// newline is appended to format by default.
|
||||||
|
func Red(format string, a ...interface{}) { colorPrint(format, FgRed, a...) } |
||||||
|
|
||||||
|
// Green is a convenient helper function to print with green foreground. A
|
||||||
|
// newline is appended to format by default.
|
||||||
|
func Green(format string, a ...interface{}) { colorPrint(format, FgGreen, a...) } |
||||||
|
|
||||||
|
// Yellow is a convenient helper function to print with yellow foreground.
|
||||||
|
// A newline is appended to format by default.
|
||||||
|
func Yellow(format string, a ...interface{}) { colorPrint(format, FgYellow, a...) } |
||||||
|
|
||||||
|
// Blue is a convenient helper function to print with blue foreground. A
|
||||||
|
// newline is appended to format by default.
|
||||||
|
func Blue(format string, a ...interface{}) { colorPrint(format, FgBlue, a...) } |
||||||
|
|
||||||
|
// Magenta is a convenient helper function to print with magenta foreground.
|
||||||
|
// A newline is appended to format by default.
|
||||||
|
func Magenta(format string, a ...interface{}) { colorPrint(format, FgMagenta, a...) } |
||||||
|
|
||||||
|
// Cyan is a convenient helper function to print with cyan foreground. A
|
||||||
|
// newline is appended to format by default.
|
||||||
|
func Cyan(format string, a ...interface{}) { colorPrint(format, FgCyan, a...) } |
||||||
|
|
||||||
|
// White is a convenient helper function to print with white foreground. A
|
||||||
|
// newline is appended to format by default.
|
||||||
|
func White(format string, a ...interface{}) { colorPrint(format, FgWhite, a...) } |
||||||
|
|
||||||
|
// BlackString is a convenient helper function to return a string with black
|
||||||
|
// foreground.
|
||||||
|
func BlackString(format string, a ...interface{}) string { return colorString(format, FgBlack, a...) } |
||||||
|
|
||||||
|
// RedString is a convenient helper function to return a string with red
|
||||||
|
// foreground.
|
||||||
|
func RedString(format string, a ...interface{}) string { return colorString(format, FgRed, a...) } |
||||||
|
|
||||||
|
// GreenString is a convenient helper function to return a string with green
|
||||||
|
// foreground.
|
||||||
|
func GreenString(format string, a ...interface{}) string { return colorString(format, FgGreen, a...) } |
||||||
|
|
||||||
|
// YellowString is a convenient helper function to return a string with yellow
|
||||||
|
// foreground.
|
||||||
|
func YellowString(format string, a ...interface{}) string { return colorString(format, FgYellow, a...) } |
||||||
|
|
||||||
|
// BlueString is a convenient helper function to return a string with blue
|
||||||
|
// foreground.
|
||||||
|
func BlueString(format string, a ...interface{}) string { return colorString(format, FgBlue, a...) } |
||||||
|
|
||||||
|
// MagentaString is a convenient helper function to return a string with magenta
|
||||||
|
// foreground.
|
||||||
|
func MagentaString(format string, a ...interface{}) string { |
||||||
|
return colorString(format, FgMagenta, a...) |
||||||
|
} |
||||||
|
|
||||||
|
// CyanString is a convenient helper function to return a string with cyan
|
||||||
|
// foreground.
|
||||||
|
func CyanString(format string, a ...interface{}) string { return colorString(format, FgCyan, a...) } |
||||||
|
|
||||||
|
// WhiteString is a convenient helper function to return a string with white
|
||||||
|
// foreground.
|
||||||
|
func WhiteString(format string, a ...interface{}) string { return colorString(format, FgWhite, a...) } |
||||||
|
|
||||||
|
// HiBlack is a convenient helper function to print with hi-intensity black foreground. A
|
||||||
|
// newline is appended to format by default.
|
||||||
|
func HiBlack(format string, a ...interface{}) { colorPrint(format, FgHiBlack, a...) } |
||||||
|
|
||||||
|
// HiRed is a convenient helper function to print with hi-intensity red foreground. A
|
||||||
|
// newline is appended to format by default.
|
||||||
|
func HiRed(format string, a ...interface{}) { colorPrint(format, FgHiRed, a...) } |
||||||
|
|
||||||
|
// HiGreen is a convenient helper function to print with hi-intensity green foreground. A
|
||||||
|
// newline is appended to format by default.
|
||||||
|
func HiGreen(format string, a ...interface{}) { colorPrint(format, FgHiGreen, a...) } |
||||||
|
|
||||||
|
// HiYellow is a convenient helper function to print with hi-intensity yellow foreground.
|
||||||
|
// A newline is appended to format by default.
|
||||||
|
func HiYellow(format string, a ...interface{}) { colorPrint(format, FgHiYellow, a...) } |
||||||
|
|
||||||
|
// HiBlue is a convenient helper function to print with hi-intensity blue foreground. A
|
||||||
|
// newline is appended to format by default.
|
||||||
|
func HiBlue(format string, a ...interface{}) { colorPrint(format, FgHiBlue, a...) } |
||||||
|
|
||||||
|
// HiMagenta is a convenient helper function to print with hi-intensity magenta foreground.
|
||||||
|
// A newline is appended to format by default.
|
||||||
|
func HiMagenta(format string, a ...interface{}) { colorPrint(format, FgHiMagenta, a...) } |
||||||
|
|
||||||
|
// HiCyan is a convenient helper function to print with hi-intensity cyan foreground. A
|
||||||
|
// newline is appended to format by default.
|
||||||
|
func HiCyan(format string, a ...interface{}) { colorPrint(format, FgHiCyan, a...) } |
||||||
|
|
||||||
|
// HiWhite is a convenient helper function to print with hi-intensity white foreground. A
|
||||||
|
// newline is appended to format by default.
|
||||||
|
func HiWhite(format string, a ...interface{}) { colorPrint(format, FgHiWhite, a...) } |
||||||
|
|
||||||
|
// HiBlackString is a convenient helper function to return a string with hi-intensity black
|
||||||
|
// foreground.
|
||||||
|
func HiBlackString(format string, a ...interface{}) string { |
||||||
|
return colorString(format, FgHiBlack, a...) |
||||||
|
} |
||||||
|
|
||||||
|
// HiRedString is a convenient helper function to return a string with hi-intensity red
|
||||||
|
// foreground.
|
||||||
|
func HiRedString(format string, a ...interface{}) string { return colorString(format, FgHiRed, a...) } |
||||||
|
|
||||||
|
// HiGreenString is a convenient helper function to return a string with hi-intensity green
|
||||||
|
// foreground.
|
||||||
|
func HiGreenString(format string, a ...interface{}) string { |
||||||
|
return colorString(format, FgHiGreen, a...) |
||||||
|
} |
||||||
|
|
||||||
|
// HiYellowString is a convenient helper function to return a string with hi-intensity yellow
|
||||||
|
// foreground.
|
||||||
|
func HiYellowString(format string, a ...interface{}) string { |
||||||
|
return colorString(format, FgHiYellow, a...) |
||||||
|
} |
||||||
|
|
||||||
|
// HiBlueString is a convenient helper function to return a string with hi-intensity blue
|
||||||
|
// foreground.
|
||||||
|
func HiBlueString(format string, a ...interface{}) string { return colorString(format, FgHiBlue, a...) } |
||||||
|
|
||||||
|
// HiMagentaString is a convenient helper function to return a string with hi-intensity magenta
|
||||||
|
// foreground.
|
||||||
|
func HiMagentaString(format string, a ...interface{}) string { |
||||||
|
return colorString(format, FgHiMagenta, a...) |
||||||
|
} |
||||||
|
|
||||||
|
// HiCyanString is a convenient helper function to return a string with hi-intensity cyan
|
||||||
|
// foreground.
|
||||||
|
func HiCyanString(format string, a ...interface{}) string { return colorString(format, FgHiCyan, a...) } |
||||||
|
|
||||||
|
// HiWhiteString is a convenient helper function to return a string with hi-intensity white
|
||||||
|
// foreground.
|
||||||
|
func HiWhiteString(format string, a ...interface{}) string { |
||||||
|
return colorString(format, FgHiWhite, a...) |
||||||
|
} |
@ -0,0 +1,133 @@ |
|||||||
|
/* |
||||||
|
Package color is an ANSI color package to output colorized or SGR defined |
||||||
|
output to the standard output. The API can be used in several way, pick one |
||||||
|
that suits you. |
||||||
|
|
||||||
|
Use simple and default helper functions with predefined foreground colors: |
||||||
|
|
||||||
|
color.Cyan("Prints text in cyan.") |
||||||
|
|
||||||
|
// a newline will be appended automatically
|
||||||
|
color.Blue("Prints %s in blue.", "text") |
||||||
|
|
||||||
|
// More default foreground colors..
|
||||||
|
color.Red("We have red") |
||||||
|
color.Yellow("Yellow color too!") |
||||||
|
color.Magenta("And many others ..") |
||||||
|
|
||||||
|
// Hi-intensity colors
|
||||||
|
color.HiGreen("Bright green color.") |
||||||
|
color.HiBlack("Bright black means gray..") |
||||||
|
color.HiWhite("Shiny white color!") |
||||||
|
|
||||||
|
However there are times where custom color mixes are required. Below are some |
||||||
|
examples to create custom color objects and use the print functions of each |
||||||
|
separate color object. |
||||||
|
|
||||||
|
// Create a new color object
|
||||||
|
c := color.New(color.FgCyan).Add(color.Underline) |
||||||
|
c.Println("Prints cyan text with an underline.") |
||||||
|
|
||||||
|
// Or just add them to New()
|
||||||
|
d := color.New(color.FgCyan, color.Bold) |
||||||
|
d.Printf("This prints bold cyan %s\n", "too!.") |
||||||
|
|
||||||
|
|
||||||
|
// Mix up foreground and background colors, create new mixes!
|
||||||
|
red := color.New(color.FgRed) |
||||||
|
|
||||||
|
boldRed := red.Add(color.Bold) |
||||||
|
boldRed.Println("This will print text in bold red.") |
||||||
|
|
||||||
|
whiteBackground := red.Add(color.BgWhite) |
||||||
|
whiteBackground.Println("Red text with White background.") |
||||||
|
|
||||||
|
// Use your own io.Writer output
|
||||||
|
color.New(color.FgBlue).Fprintln(myWriter, "blue color!") |
||||||
|
|
||||||
|
blue := color.New(color.FgBlue) |
||||||
|
blue.Fprint(myWriter, "This will print text in blue.") |
||||||
|
|
||||||
|
You can create PrintXxx functions to simplify even more: |
||||||
|
|
||||||
|
// Create a custom print function for convenient
|
||||||
|
red := color.New(color.FgRed).PrintfFunc() |
||||||
|
red("warning") |
||||||
|
red("error: %s", err) |
||||||
|
|
||||||
|
// Mix up multiple attributes
|
||||||
|
notice := color.New(color.Bold, color.FgGreen).PrintlnFunc() |
||||||
|
notice("don't forget this...") |
||||||
|
|
||||||
|
You can also FprintXxx functions to pass your own io.Writer: |
||||||
|
|
||||||
|
blue := color.New(FgBlue).FprintfFunc() |
||||||
|
blue(myWriter, "important notice: %s", stars) |
||||||
|
|
||||||
|
// Mix up with multiple attributes
|
||||||
|
success := color.New(color.Bold, color.FgGreen).FprintlnFunc() |
||||||
|
success(myWriter, don't forget this...") |
||||||
|
|
||||||
|
|
||||||
|
Or create SprintXxx functions to mix strings with other non-colorized strings: |
||||||
|
|
||||||
|
yellow := New(FgYellow).SprintFunc() |
||||||
|
red := New(FgRed).SprintFunc() |
||||||
|
|
||||||
|
fmt.Printf("this is a %s and this is %s.\n", yellow("warning"), red("error")) |
||||||
|
|
||||||
|
info := New(FgWhite, BgGreen).SprintFunc() |
||||||
|
fmt.Printf("this %s rocks!\n", info("package")) |
||||||
|
|
||||||
|
Windows support is enabled by default. All Print functions work as intended. |
||||||
|
However only for color.SprintXXX functions, user should use fmt.FprintXXX and |
||||||
|
set the output to color.Output: |
||||||
|
|
||||||
|
fmt.Fprintf(color.Output, "Windows support: %s", color.GreenString("PASS")) |
||||||
|
|
||||||
|
info := New(FgWhite, BgGreen).SprintFunc() |
||||||
|
fmt.Fprintf(color.Output, "this %s rocks!\n", info("package")) |
||||||
|
|
||||||
|
Using with existing code is possible. Just use the Set() method to set the |
||||||
|
standard output to the given parameters. That way a rewrite of an existing |
||||||
|
code is not required. |
||||||
|
|
||||||
|
// Use handy standard colors.
|
||||||
|
color.Set(color.FgYellow) |
||||||
|
|
||||||
|
fmt.Println("Existing text will be now in Yellow") |
||||||
|
fmt.Printf("This one %s\n", "too") |
||||||
|
|
||||||
|
color.Unset() // don't forget to unset
|
||||||
|
|
||||||
|
// You can mix up parameters
|
||||||
|
color.Set(color.FgMagenta, color.Bold) |
||||||
|
defer color.Unset() // use it in your function
|
||||||
|
|
||||||
|
fmt.Println("All text will be now bold magenta.") |
||||||
|
|
||||||
|
There might be a case where you want to disable color output (for example to |
||||||
|
pipe the standard output of your app to somewhere else). `Color` has support to |
||||||
|
disable colors both globally and for single color definition. For example |
||||||
|
suppose you have a CLI app and a `--no-color` bool flag. You can easily disable |
||||||
|
the color output with: |
||||||
|
|
||||||
|
var flagNoColor = flag.Bool("no-color", false, "Disable color output") |
||||||
|
|
||||||
|
if *flagNoColor { |
||||||
|
color.NoColor = true // disables colorized output
|
||||||
|
} |
||||||
|
|
||||||
|
It also has support for single color definitions (local). You can |
||||||
|
disable/enable color output on the fly: |
||||||
|
|
||||||
|
c := color.New(color.FgCyan) |
||||||
|
c.Println("Prints cyan text") |
||||||
|
|
||||||
|
c.DisableColor() |
||||||
|
c.Println("This is printed without any color") |
||||||
|
|
||||||
|
c.EnableColor() |
||||||
|
c.Println("This prints again cyan...") |
||||||
|
*/ |
||||||
|
package color |
@ -0,0 +1,8 @@ |
|||||||
|
module github.com/fatih/color |
||||||
|
|
||||||
|
go 1.13 |
||||||
|
|
||||||
|
require ( |
||||||
|
github.com/mattn/go-colorable v0.1.4 |
||||||
|
github.com/mattn/go-isatty v0.0.11 |
||||||
|
) |
@ -0,0 +1,8 @@ |
|||||||
|
github.com/mattn/go-colorable v0.1.4 h1:snbPLB8fVfU9iwbbo30TPtbLRzwWu6aJS6Xh4eaaviA= |
||||||
|
github.com/mattn/go-colorable v0.1.4/go.mod h1:U0ppj6V5qS13XJ6of8GYAs25YV2eR4EVcfRqFIhoBtE= |
||||||
|
github.com/mattn/go-isatty v0.0.8/go.mod h1:Iq45c/XA43vh69/j3iqttzPXn0bhXyGjM0Hdxcsrc5s= |
||||||
|
github.com/mattn/go-isatty v0.0.11 h1:FxPOTFNqGkuDUGi3H/qkUbQO4ZiBa2brKq5r0l8TGeM= |
||||||
|
github.com/mattn/go-isatty v0.0.11/go.mod h1:PhnuNfih5lzO57/f3n+odYbM4JtupLOxQOAqxQCu2WE= |
||||||
|
golang.org/x/sys v0.0.0-20190222072716-a9d3bda3a223/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= |
||||||
|
golang.org/x/sys v0.0.0-20191026070338-33540a1f6037 h1:YyJpGZS1sBuBCzLAR1VEpK193GlqGZbnPFnPV/5Rsb4= |
||||||
|
golang.org/x/sys v0.0.0-20191026070338-33540a1f6037/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= |
@ -0,0 +1,60 @@ |
|||||||
|
Copyright (c) 2017, Fatih Arslan |
||||||
|
All rights reserved. |
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without |
||||||
|
modification, are permitted provided that the following conditions are met: |
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright notice, this |
||||||
|
list of conditions and the following disclaimer. |
||||||
|
|
||||||
|
* Redistributions in binary form must reproduce the above copyright notice, |
||||||
|
this list of conditions and the following disclaimer in the documentation |
||||||
|
and/or other materials provided with the distribution. |
||||||
|
|
||||||
|
* Neither the name of structtag nor the names of its |
||||||
|
contributors may be used to endorse or promote products derived from |
||||||
|
this software without specific prior written permission. |
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
||||||
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
||||||
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE |
||||||
|
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE |
||||||
|
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL |
||||||
|
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR |
||||||
|
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER |
||||||
|
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, |
||||||
|
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
||||||
|
|
||||||
|
This software includes some portions from Go. Go is used under the terms of the |
||||||
|
BSD like license. |
||||||
|
|
||||||
|
Copyright (c) 2012 The Go Authors. All rights reserved. |
||||||
|
|
||||||
|
Redistribution and use in source and binary forms, with or without |
||||||
|
modification, are permitted provided that the following conditions are |
||||||
|
met: |
||||||
|
|
||||||
|
* Redistributions of source code must retain the above copyright |
||||||
|
notice, this list of conditions and the following disclaimer. |
||||||
|
* Redistributions in binary form must reproduce the above |
||||||
|
copyright notice, this list of conditions and the following disclaimer |
||||||
|
in the documentation and/or other materials provided with the |
||||||
|
distribution. |
||||||
|
* Neither the name of Google Inc. nor the names of its |
||||||
|
contributors may be used to endorse or promote products derived from |
||||||
|
this software without specific prior written permission. |
||||||
|
|
||||||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
||||||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
||||||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
||||||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
||||||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
||||||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
||||||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
||||||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
||||||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
||||||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
||||||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
||||||
|
|
||||||
|
The Go gopher was designed by Renee French. http://reneefrench.blogspot.com/ The design is licensed under the Creative Commons 3.0 Attributions license. Read this article for more details: https://blog.golang.org/gopher |
@ -0,0 +1,73 @@ |
|||||||
|
# structtag [![GoDoc](http://img.shields.io/badge/go-documentation-blue.svg?style=flat-square)](http://godoc.org/github.com/fatih/structtag) |
||||||
|
|
||||||
|
structtag provides an easy way of parsing and manipulating struct tag fields. |
||||||
|
Please vendor the library as it might change in future versions. |
||||||
|
|
||||||
|
# Install |
||||||
|
|
||||||
|
```bash |
||||||
|
go get github.com/fatih/structtag |
||||||
|
``` |
||||||
|
|
||||||
|
# Example |
||||||
|
|
||||||
|
```go |
||||||
|
package main |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"reflect" |
||||||
|
"sort" |
||||||
|
|
||||||
|
"github.com/fatih/structtag" |
||||||
|
) |
||||||
|
|
||||||
|
func main() { |
||||||
|
type t struct { |
||||||
|
t string `json:"foo,omitempty,string" xml:"foo"` |
||||||
|
} |
||||||
|
|
||||||
|
// get field tag |
||||||
|
tag := reflect.TypeOf(t{}).Field(0).Tag |
||||||
|
|
||||||
|
// ... and start using structtag by parsing the tag |
||||||
|
tags, err := structtag.Parse(string(tag)) |
||||||
|
if err != nil { |
||||||
|
panic(err) |
||||||
|
} |
||||||
|
|
||||||
|
// iterate over all tags |
||||||
|
for _, t := range tags.Tags() { |
||||||
|
fmt.Printf("tag: %+v\n", t) |
||||||
|
} |
||||||
|
|
||||||
|
// get a single tag |
||||||
|
jsonTag, err := tags.Get("json") |
||||||
|
if err != nil { |
||||||
|
panic(err) |
||||||
|
} |
||||||
|
fmt.Println(jsonTag) // Output: json:"foo,omitempty,string" |
||||||
|
fmt.Println(jsonTag.Key) // Output: json |
||||||
|
fmt.Println(jsonTag.Name) // Output: foo |
||||||
|
fmt.Println(jsonTag.Options) // Output: [omitempty string] |
||||||
|
|
||||||
|
// change existing tag |
||||||
|
jsonTag.Name = "foo_bar" |
||||||
|
jsonTag.Options = nil |
||||||
|
tags.Set(jsonTag) |
||||||
|
|
||||||
|
// add new tag |
||||||
|
tags.Set(&structtag.Tag{ |
||||||
|
Key: "hcl", |
||||||
|
Name: "foo", |
||||||
|
Options: []string{"squash"}, |
||||||
|
}) |
||||||
|
|
||||||
|
// print the tags |
||||||
|
fmt.Println(tags) // Output: json:"foo_bar" xml:"foo" hcl:"foo,squash" |
||||||
|
|
||||||
|
// sort tags according to keys |
||||||
|
sort.Sort(tags) |
||||||
|
fmt.Println(tags) // Output: hcl:"foo,squash" json:"foo_bar" xml:"foo" |
||||||
|
} |
||||||
|
``` |
@ -0,0 +1,3 @@ |
|||||||
|
module github.com/fatih/structtag |
||||||
|
|
||||||
|
go 1.12 |
@ -0,0 +1,315 @@ |
|||||||
|
package structtag |
||||||
|
|
||||||
|
import ( |
||||||
|
"bytes" |
||||||
|
"errors" |
||||||
|
"fmt" |
||||||
|
"strconv" |
||||||
|
"strings" |
||||||
|
) |
||||||
|
|
||||||
|
var ( |
||||||
|
errTagSyntax = errors.New("bad syntax for struct tag pair") |
||||||
|
errTagKeySyntax = errors.New("bad syntax for struct tag key") |
||||||
|
errTagValueSyntax = errors.New("bad syntax for struct tag value") |
||||||
|
|
||||||
|
errKeyNotSet = errors.New("tag key does not exist") |
||||||
|
errTagNotExist = errors.New("tag does not exist") |
||||||
|
errTagKeyMismatch = errors.New("mismatch between key and tag.key") |
||||||
|
) |
||||||
|
|
||||||
|
// Tags represent a set of tags from a single struct field
|
||||||
|
type Tags struct { |
||||||
|
tags []*Tag |
||||||
|
} |
||||||
|
|
||||||
|
// Tag defines a single struct's string literal tag
|
||||||
|
type Tag struct { |
||||||
|
// Key is the tag key, such as json, xml, etc..
|
||||||
|
// i.e: `json:"foo,omitempty". Here key is: "json"
|
||||||
|
Key string |
||||||
|
|
||||||
|
// Name is a part of the value
|
||||||
|
// i.e: `json:"foo,omitempty". Here name is: "foo"
|
||||||
|
Name string |
||||||
|
|
||||||
|
// Options is a part of the value. It contains a slice of tag options i.e:
|
||||||
|
// `json:"foo,omitempty". Here options is: ["omitempty"]
|
||||||
|
Options []string |
||||||
|
} |
||||||
|
|
||||||
|
// Parse parses a single struct field tag and returns the set of tags.
|
||||||
|
func Parse(tag string) (*Tags, error) { |
||||||
|
var tags []*Tag |
||||||
|
|
||||||
|
hasTag := tag != "" |
||||||
|
|
||||||
|
// NOTE(arslan) following code is from reflect and vet package with some
|
||||||
|
// modifications to collect all necessary information and extend it with
|
||||||
|
// usable methods
|
||||||
|
for tag != "" { |
||||||
|
// Skip leading space.
|
||||||
|
i := 0 |
||||||
|
for i < len(tag) && tag[i] == ' ' { |
||||||
|
i++ |
||||||
|
} |
||||||
|
tag = tag[i:] |
||||||
|
if tag == "" { |
||||||
|
break |
||||||
|
} |
||||||
|
|
||||||
|
// Scan to colon. A space, a quote or a control character is a syntax
|
||||||
|
// error. Strictly speaking, control chars include the range [0x7f,
|
||||||
|
// 0x9f], not just [0x00, 0x1f], but in practice, we ignore the
|
||||||
|
// multi-byte control characters as it is simpler to inspect the tag's
|
||||||
|
// bytes than the tag's runes.
|
||||||
|
i = 0 |
||||||
|
for i < len(tag) && tag[i] > ' ' && tag[i] != ':' && tag[i] != '"' && tag[i] != 0x7f { |
||||||
|
i++ |
||||||
|
} |
||||||
|
|
||||||
|
if i == 0 { |
||||||
|
return nil, errTagKeySyntax |
||||||
|
} |
||||||
|
if i+1 >= len(tag) || tag[i] != ':' { |
||||||
|
return nil, errTagSyntax |
||||||
|
} |
||||||
|
if tag[i+1] != '"' { |
||||||
|
return nil, errTagValueSyntax |
||||||
|
} |
||||||
|
|
||||||
|
key := string(tag[:i]) |
||||||
|
tag = tag[i+1:] |
||||||
|
|
||||||
|
// Scan quoted string to find value.
|
||||||
|
i = 1 |
||||||
|
for i < len(tag) && tag[i] != '"' { |
||||||
|
if tag[i] == '\\' { |
||||||
|
i++ |
||||||
|
} |
||||||
|
i++ |
||||||
|
} |
||||||
|
if i >= len(tag) { |
||||||
|
return nil, errTagValueSyntax |
||||||
|
} |
||||||
|
|
||||||
|
qvalue := string(tag[:i+1]) |
||||||
|
tag = tag[i+1:] |
||||||
|
|
||||||
|
value, err := strconv.Unquote(qvalue) |
||||||
|
if err != nil { |
||||||
|
return nil, errTagValueSyntax |
||||||
|
} |
||||||
|
|
||||||
|
res := strings.Split(value, ",") |
||||||
|
name := res[0] |
||||||
|
options := res[1:] |
||||||
|
if len(options) == 0 { |
||||||
|
options = nil |
||||||
|
} |
||||||
|
|
||||||
|
tags = append(tags, &Tag{ |
||||||
|
Key: key, |
||||||
|
Name: name, |
||||||
|
Options: options, |
||||||
|
}) |
||||||
|
} |
||||||
|
|
||||||
|
if hasTag && len(tags) == 0 { |
||||||
|
return nil, nil |
||||||
|
} |
||||||
|
|
||||||
|
return &Tags{ |
||||||
|
tags: tags, |
||||||
|
}, nil |
||||||
|
} |
||||||
|
|
||||||
|
// Get returns the tag associated with the given key. If the key is present
|
||||||
|
// in the tag the value (which may be empty) is returned. Otherwise the
|
||||||
|
// returned value will be the empty string. The ok return value reports whether
|
||||||
|
// the tag exists or not (which the return value is nil).
|
||||||
|
func (t *Tags) Get(key string) (*Tag, error) { |
||||||
|
for _, tag := range t.tags { |
||||||
|
if tag.Key == key { |
||||||
|
return tag, nil |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
return nil, errTagNotExist |
||||||
|
} |
||||||
|
|
||||||
|
// Set sets the given tag. If the tag key already exists it'll override it
|
||||||
|
func (t *Tags) Set(tag *Tag) error { |
||||||
|
if tag.Key == "" { |
||||||
|
return errKeyNotSet |
||||||
|
} |
||||||
|
|
||||||
|
added := false |
||||||
|
for i, tg := range t.tags { |
||||||
|
if tg.Key == tag.Key { |
||||||
|
added = true |
||||||
|
t.tags[i] = tag |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
if !added { |
||||||
|
// this means this is a new tag, add it
|
||||||
|
t.tags = append(t.tags, tag) |
||||||
|
} |
||||||
|
|
||||||
|
return nil |
||||||
|
} |
||||||
|
|
||||||
|
// AddOptions adds the given option for the given key. If the option already
|
||||||
|
// exists it doesn't add it again.
|
||||||
|
func (t *Tags) AddOptions(key string, options ...string) { |
||||||
|
for i, tag := range t.tags { |
||||||
|
if tag.Key != key { |
||||||
|
continue |
||||||
|
} |
||||||
|
|
||||||
|
for _, opt := range options { |
||||||
|
if !tag.HasOption(opt) { |
||||||
|
tag.Options = append(tag.Options, opt) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
t.tags[i] = tag |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// DeleteOptions deletes the given options for the given key
|
||||||
|
func (t *Tags) DeleteOptions(key string, options ...string) { |
||||||
|
hasOption := func(option string) bool { |
||||||
|
for _, opt := range options { |
||||||
|
if opt == option { |
||||||
|
return true |
||||||
|
} |
||||||
|
} |
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
for i, tag := range t.tags { |
||||||
|
if tag.Key != key { |
||||||
|
continue |
||||||
|
} |
||||||
|
|
||||||
|
var updated []string |
||||||
|
for _, opt := range tag.Options { |
||||||
|
if !hasOption(opt) { |
||||||
|
updated = append(updated, opt) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
tag.Options = updated |
||||||
|
t.tags[i] = tag |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// Delete deletes the tag for the given keys
|
||||||
|
func (t *Tags) Delete(keys ...string) { |
||||||
|
hasKey := func(key string) bool { |
||||||
|
for _, k := range keys { |
||||||
|
if k == key { |
||||||
|
return true |
||||||
|
} |
||||||
|
} |
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
var updated []*Tag |
||||||
|
for _, tag := range t.tags { |
||||||
|
if !hasKey(tag.Key) { |
||||||
|
updated = append(updated, tag) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
t.tags = updated |
||||||
|
} |
||||||
|
|
||||||
|
// Tags returns a slice of tags. The order is the original tag order unless it
|
||||||
|
// was changed.
|
||||||
|
func (t *Tags) Tags() []*Tag { |
||||||
|
return t.tags |
||||||
|
} |
||||||
|
|
||||||
|
// Tags returns a slice of tags. The order is the original tag order unless it
|
||||||
|
// was changed.
|
||||||
|
func (t *Tags) Keys() []string { |
||||||
|
var keys []string |
||||||
|
for _, tag := range t.tags { |
||||||
|
keys = append(keys, tag.Key) |
||||||
|
} |
||||||
|
return keys |
||||||
|
} |
||||||
|
|
||||||
|
// String reassembles the tags into a valid literal tag field representation
|
||||||
|
func (t *Tags) String() string { |
||||||
|
tags := t.Tags() |
||||||
|
if len(tags) == 0 { |
||||||
|
return "" |
||||||
|
} |
||||||
|
|
||||||
|
var buf bytes.Buffer |
||||||
|
for i, tag := range t.Tags() { |
||||||
|
buf.WriteString(tag.String()) |
||||||
|
if i != len(tags)-1 { |
||||||
|
buf.WriteString(" ") |
||||||
|
} |
||||||
|
} |
||||||
|
return buf.String() |
||||||
|
} |
||||||
|
|
||||||
|
// HasOption returns true if the given option is available in options
|
||||||
|
func (t *Tag) HasOption(opt string) bool { |
||||||
|
for _, tagOpt := range t.Options { |
||||||
|
if tagOpt == opt { |
||||||
|
return true |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
// Value returns the raw value of the tag, i.e. if the tag is
|
||||||
|
// `json:"foo,omitempty", the Value is "foo,omitempty"
|
||||||
|
func (t *Tag) Value() string { |
||||||
|
options := strings.Join(t.Options, ",") |
||||||
|
if options != "" { |
||||||
|
return fmt.Sprintf(`%s,%s`, t.Name, options) |
||||||
|
} |
||||||
|
return t.Name |
||||||
|
} |
||||||
|
|
||||||
|
// String reassembles the tag into a valid tag field representation
|
||||||
|
func (t *Tag) String() string { |
||||||
|
return fmt.Sprintf(`%s:%q`, t.Key, t.Value()) |
||||||
|
} |
||||||
|
|
||||||
|
// GoString implements the fmt.GoStringer interface
|
||||||
|
func (t *Tag) GoString() string { |
||||||
|
template := `{ |
||||||
|
Key: '%s', |
||||||
|
Name: '%s', |
||||||
|
Option: '%s', |
||||||
|
}` |
||||||
|
|
||||||
|
if t.Options == nil { |
||||||
|
return fmt.Sprintf(template, t.Key, t.Name, "nil") |
||||||
|
} |
||||||
|
|
||||||
|
options := strings.Join(t.Options, ",") |
||||||
|
return fmt.Sprintf(template, t.Key, t.Name, options) |
||||||
|
} |
||||||
|
|
||||||
|
func (t *Tags) Len() int { |
||||||
|
return len(t.tags) |
||||||
|
} |
||||||
|
|
||||||
|
func (t *Tags) Less(i int, j int) bool { |
||||||
|
return t.tags[i].Key < t.tags[j].Key |
||||||
|
} |
||||||
|
|
||||||
|
func (t *Tags) Swap(i int, j int) { |
||||||
|
t.tags[i], t.tags[j] = t.tags[j], t.tags[i] |
||||||
|
} |
@ -0,0 +1,9 @@ |
|||||||
|
language: go |
||||||
|
go: |
||||||
|
- tip |
||||||
|
|
||||||
|
before_install: |
||||||
|
- go get github.com/mattn/goveralls |
||||||
|
- go get golang.org/x/tools/cmd/cover |
||||||
|
script: |
||||||
|
- $HOME/gopath/bin/goveralls -repotoken xnXqRGwgW3SXIguzxf90ZSK1GPYZPaGrw |
@ -0,0 +1,21 @@ |
|||||||
|
The MIT License (MIT) |
||||||
|
|
||||||
|
Copyright (c) 2016 Yasuhiro Matsumoto |
||||||
|
|
||||||
|
Permission is hereby granted, free of charge, to any person obtaining a copy |
||||||
|
of this software and associated documentation files (the "Software"), to deal |
||||||
|
in the Software without restriction, including without limitation the rights |
||||||
|
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell |
||||||
|
copies of the Software, and to permit persons to whom the Software is |
||||||
|
furnished to do so, subject to the following conditions: |
||||||
|
|
||||||
|
The above copyright notice and this permission notice shall be included in all |
||||||
|
copies or substantial portions of the Software. |
||||||
|
|
||||||
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
||||||
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
||||||
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE |
||||||
|
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
||||||
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, |
||||||
|
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
||||||
|
SOFTWARE. |
@ -0,0 +1,48 @@ |
|||||||
|
# go-colorable |
||||||
|
|
||||||
|
[![Godoc Reference](https://godoc.org/github.com/mattn/go-colorable?status.svg)](http://godoc.org/github.com/mattn/go-colorable) |
||||||
|
[![Build Status](https://travis-ci.org/mattn/go-colorable.svg?branch=master)](https://travis-ci.org/mattn/go-colorable) |
||||||
|
[![Coverage Status](https://coveralls.io/repos/github/mattn/go-colorable/badge.svg?branch=master)](https://coveralls.io/github/mattn/go-colorable?branch=master) |
||||||
|
[![Go Report Card](https://goreportcard.com/badge/mattn/go-colorable)](https://goreportcard.com/report/mattn/go-colorable) |
||||||
|
|
||||||
|
Colorable writer for windows. |
||||||
|
|
||||||
|
For example, most of logger packages doesn't show colors on windows. (I know we can do it with ansicon. But I don't want.) |
||||||
|
This package is possible to handle escape sequence for ansi color on windows. |
||||||
|
|
||||||
|
## Too Bad! |
||||||
|
|
||||||
|
![](https://raw.githubusercontent.com/mattn/go-colorable/gh-pages/bad.png) |
||||||
|
|
||||||
|
|
||||||
|
## So Good! |
||||||
|
|
||||||
|
![](https://raw.githubusercontent.com/mattn/go-colorable/gh-pages/good.png) |
||||||
|
|
||||||
|
## Usage |
||||||
|
|
||||||
|
```go |
||||||
|
logrus.SetFormatter(&logrus.TextFormatter{ForceColors: true}) |
||||||
|
logrus.SetOutput(colorable.NewColorableStdout()) |
||||||
|
|
||||||
|
logrus.Info("succeeded") |
||||||
|
logrus.Warn("not correct") |
||||||
|
logrus.Error("something error") |
||||||
|
logrus.Fatal("panic") |
||||||
|
``` |
||||||
|
|
||||||
|
You can compile above code on non-windows OSs. |
||||||
|
|
||||||
|
## Installation |
||||||
|
|
||||||
|
``` |
||||||
|
$ go get github.com/mattn/go-colorable |
||||||
|
``` |
||||||
|
|
||||||
|
# License |
||||||
|
|
||||||
|
MIT |
||||||
|
|
||||||
|
# Author |
||||||
|
|
||||||
|
Yasuhiro Matsumoto (a.k.a mattn) |
@ -0,0 +1,29 @@ |
|||||||
|
// +build appengine
|
||||||
|
|
||||||
|
package colorable |
||||||
|
|
||||||
|
import ( |
||||||
|
"io" |
||||||
|
"os" |
||||||
|
|
||||||
|
_ "github.com/mattn/go-isatty" |
||||||
|
) |
||||||
|
|
||||||
|
// NewColorable returns new instance of Writer which handles escape sequence.
|
||||||
|
func NewColorable(file *os.File) io.Writer { |
||||||
|
if file == nil { |
||||||
|
panic("nil passed instead of *os.File to NewColorable()") |
||||||
|
} |
||||||
|
|
||||||
|
return file |
||||||
|
} |
||||||
|
|
||||||
|
// NewColorableStdout returns new instance of Writer which handles escape sequence for stdout.
|
||||||
|
func NewColorableStdout() io.Writer { |
||||||
|
return os.Stdout |
||||||
|
} |
||||||
|
|
||||||
|
// NewColorableStderr returns new instance of Writer which handles escape sequence for stderr.
|
||||||
|
func NewColorableStderr() io.Writer { |
||||||
|
return os.Stderr |
||||||
|
} |
@ -0,0 +1,30 @@ |
|||||||
|
// +build !windows
|
||||||
|
// +build !appengine
|
||||||
|
|
||||||
|
package colorable |
||||||
|
|
||||||
|
import ( |
||||||
|
"io" |
||||||
|
"os" |
||||||
|
|
||||||
|
_ "github.com/mattn/go-isatty" |
||||||
|
) |
||||||
|
|
||||||
|
// NewColorable returns new instance of Writer which handles escape sequence.
|
||||||
|
func NewColorable(file *os.File) io.Writer { |
||||||
|
if file == nil { |
||||||
|
panic("nil passed instead of *os.File to NewColorable()") |
||||||
|
} |
||||||
|
|
||||||
|
return file |
||||||
|
} |
||||||
|
|
||||||
|
// NewColorableStdout returns new instance of Writer which handles escape sequence for stdout.
|
||||||
|
func NewColorableStdout() io.Writer { |
||||||
|
return os.Stdout |
||||||
|
} |
||||||
|
|
||||||
|
// NewColorableStderr returns new instance of Writer which handles escape sequence for stderr.
|
||||||
|
func NewColorableStderr() io.Writer { |
||||||
|
return os.Stderr |
||||||
|
} |
File diff suppressed because it is too large
Load Diff
@ -0,0 +1,3 @@ |
|||||||
|
module github.com/mattn/go-colorable |
||||||
|
|
||||||
|
require github.com/mattn/go-isatty v0.0.8 |
@ -0,0 +1,4 @@ |
|||||||
|
github.com/mattn/go-isatty v0.0.5 h1:tHXDdz1cpzGaovsTB+TVB8q90WEokoVmfMqoVcrLUgw= |
||||||
|
github.com/mattn/go-isatty v0.0.5/go.mod h1:Iq45c/XA43vh69/j3iqttzPXn0bhXyGjM0Hdxcsrc5s= |
||||||
|
golang.org/x/sys v0.0.0-20190222072716-a9d3bda3a223 h1:DH4skfRX4EBpamg7iV4ZlCpblAHI6s6TDM39bFZumv8= |
||||||
|
golang.org/x/sys v0.0.0-20190222072716-a9d3bda3a223/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= |
@ -0,0 +1,55 @@ |
|||||||
|
package colorable |
||||||
|
|
||||||
|
import ( |
||||||
|
"bytes" |
||||||
|
"io" |
||||||
|
) |
||||||
|
|
||||||
|
// NonColorable holds writer but removes escape sequence.
|
||||||
|
type NonColorable struct { |
||||||
|
out io.Writer |
||||||
|
} |
||||||
|
|
||||||
|
// NewNonColorable returns new instance of Writer which removes escape sequence from Writer.
|
||||||
|
func NewNonColorable(w io.Writer) io.Writer { |
||||||
|
return &NonColorable{out: w} |
||||||
|
} |
||||||
|
|
||||||
|
// Write writes data on console
|
||||||
|
func (w *NonColorable) Write(data []byte) (n int, err error) { |
||||||
|
er := bytes.NewReader(data) |
||||||
|
var bw [1]byte |
||||||
|
loop: |
||||||
|
for { |
||||||
|
c1, err := er.ReadByte() |
||||||
|
if err != nil { |
||||||
|
break loop |
||||||
|
} |
||||||
|
if c1 != 0x1b { |
||||||
|
bw[0] = c1 |
||||||
|
w.out.Write(bw[:]) |
||||||
|
continue |
||||||
|
} |
||||||
|
c2, err := er.ReadByte() |
||||||
|
if err != nil { |
||||||
|
break loop |
||||||
|
} |
||||||
|
if c2 != 0x5b { |
||||||
|
continue |
||||||
|
} |
||||||
|
|
||||||
|
var buf bytes.Buffer |
||||||
|
for { |
||||||
|
c, err := er.ReadByte() |
||||||
|
if err != nil { |
||||||
|
break loop |
||||||
|
} |
||||||
|
if ('a' <= c && c <= 'z') || ('A' <= c && c <= 'Z') || c == '@' { |
||||||
|
break |
||||||
|
} |
||||||
|
buf.Write([]byte(string(c))) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
return len(data), nil |
||||||
|
} |
@ -1,3 +1,5 @@ |
|||||||
module github.com/mattn/go-isatty |
module github.com/mattn/go-isatty |
||||||
|
|
||||||
require golang.org/x/sys v0.0.0-20190222072716-a9d3bda3a223 |
go 1.12 |
||||||
|
|
||||||
|
require golang.org/x/sys v0.0.0-20191026070338-33540a1f6037 |
||||||
|
@ -1,2 +1,2 @@ |
|||||||
golang.org/x/sys v0.0.0-20190222072716-a9d3bda3a223 h1:DH4skfRX4EBpamg7iV4ZlCpblAHI6s6TDM39bFZumv8= |
golang.org/x/sys v0.0.0-20191026070338-33540a1f6037 h1:YyJpGZS1sBuBCzLAR1VEpK193GlqGZbnPFnPV/5Rsb4= |
||||||
golang.org/x/sys v0.0.0-20190222072716-a9d3bda3a223/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY= |
golang.org/x/sys v0.0.0-20191026070338-33540a1f6037/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs= |
||||||
|
@ -0,0 +1,22 @@ |
|||||||
|
// +build plan9
|
||||||
|
|
||||||
|
package isatty |
||||||
|
|
||||||
|
import ( |
||||||
|
"syscall" |
||||||
|
) |
||||||
|
|
||||||
|
// IsTerminal returns true if the given file descriptor is a terminal.
|
||||||
|
func IsTerminal(fd uintptr) bool { |
||||||
|
path, err := syscall.Fd2path(int(fd)) |
||||||
|
if err != nil { |
||||||
|
return false |
||||||
|
} |
||||||
|
return path == "/dev/cons" || path == "/mnt/term/dev/cons" |
||||||
|
} |
||||||
|
|
||||||
|
// IsCygwinTerminal return true if the file descriptor is a cygwin or msys2
|
||||||
|
// terminal. This is also always false on this environment.
|
||||||
|
func IsCygwinTerminal(fd uintptr) bool { |
||||||
|
return false |
||||||
|
} |
@ -1,4 +1,4 @@ |
|||||||
// +build linux
|
// +build linux aix
|
||||||
// +build !appengine
|
// +build !appengine
|
||||||
// +build !android
|
// +build !android
|
||||||
|
|
@ -0,0 +1,8 @@ |
|||||||
|
language: go |
||||||
|
go: |
||||||
|
- tip |
||||||
|
before_install: |
||||||
|
- go get github.com/mattn/goveralls |
||||||
|
- go get golang.org/x/tools/cmd/cover |
||||||
|
script: |
||||||
|
- $HOME/gopath/bin/goveralls -repotoken lAKAWPzcGsD3A8yBX3BGGtRUdJ6CaGERL |
@ -0,0 +1,21 @@ |
|||||||
|
The MIT License (MIT) |
||||||
|
|
||||||
|
Copyright (c) 2016 Yasuhiro Matsumoto |
||||||
|
|
||||||
|
Permission is hereby granted, free of charge, to any person obtaining a copy |
||||||
|
of this software and associated documentation files (the "Software"), to deal |
||||||
|
in the Software without restriction, including without limitation the rights |
||||||
|
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell |
||||||
|
copies of the Software, and to permit persons to whom the Software is |
||||||
|
furnished to do so, subject to the following conditions: |
||||||
|
|
||||||
|
The above copyright notice and this permission notice shall be included in all |
||||||
|
copies or substantial portions of the Software. |
||||||
|
|
||||||
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
||||||
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
||||||
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE |
||||||
|
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
||||||
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, |
||||||
|
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
||||||
|
SOFTWARE. |
@ -0,0 +1,27 @@ |
|||||||
|
go-runewidth |
||||||
|
============ |
||||||
|
|
||||||
|
[![Build Status](https://travis-ci.org/mattn/go-runewidth.png?branch=master)](https://travis-ci.org/mattn/go-runewidth) |
||||||
|
[![Coverage Status](https://coveralls.io/repos/mattn/go-runewidth/badge.png?branch=HEAD)](https://coveralls.io/r/mattn/go-runewidth?branch=HEAD) |
||||||
|
[![GoDoc](https://godoc.org/github.com/mattn/go-runewidth?status.svg)](http://godoc.org/github.com/mattn/go-runewidth) |
||||||
|
[![Go Report Card](https://goreportcard.com/badge/github.com/mattn/go-runewidth)](https://goreportcard.com/report/github.com/mattn/go-runewidth) |
||||||
|
|
||||||
|
Provides functions to get fixed width of the character or string. |
||||||
|
|
||||||
|
Usage |
||||||
|
----- |
||||||
|
|
||||||
|
```go |
||||||
|
runewidth.StringWidth("つのだ☆HIRO") == 12 |
||||||
|
``` |
||||||
|
|
||||||
|
|
||||||
|
Author |
||||||
|
------ |
||||||
|
|
||||||
|
Yasuhiro Matsumoto |
||||||
|
|
||||||
|
License |
||||||
|
------- |
||||||
|
|
||||||
|
under the MIT License: http://mattn.mit-license.org/2013 |
@ -0,0 +1,3 @@ |
|||||||
|
module github.com/mattn/go-runewidth |
||||||
|
|
||||||
|
go 1.9 |
@ -0,0 +1,258 @@ |
|||||||
|
package runewidth |
||||||
|
|
||||||
|
import ( |
||||||
|
"os" |
||||||
|
) |
||||||
|
|
||||||
|
//go:generate go run script/generate.go
|
||||||
|
|
||||||
|
var ( |
||||||
|
// EastAsianWidth will be set true if the current locale is CJK
|
||||||
|
EastAsianWidth bool |
||||||
|
|
||||||
|
// ZeroWidthJoiner is flag to set to use UTR#51 ZWJ
|
||||||
|
ZeroWidthJoiner bool |
||||||
|
|
||||||
|
// DefaultCondition is a condition in current locale
|
||||||
|
DefaultCondition = &Condition{} |
||||||
|
) |
||||||
|
|
||||||
|
func init() { |
||||||
|
handleEnv() |
||||||
|
} |
||||||
|
|
||||||
|
func handleEnv() { |
||||||
|
env := os.Getenv("RUNEWIDTH_EASTASIAN") |
||||||
|
if env == "" { |
||||||
|
EastAsianWidth = IsEastAsian() |
||||||
|
} else { |
||||||
|
EastAsianWidth = env == "1" |
||||||
|
} |
||||||
|
// update DefaultCondition
|
||||||
|
DefaultCondition.EastAsianWidth = EastAsianWidth |
||||||
|
DefaultCondition.ZeroWidthJoiner = ZeroWidthJoiner |
||||||
|
} |
||||||
|
|
||||||
|
type interval struct { |
||||||
|
first rune |
||||||
|
last rune |
||||||
|
} |
||||||
|
|
||||||
|
type table []interval |
||||||
|
|
||||||
|
func inTables(r rune, ts ...table) bool { |
||||||
|
for _, t := range ts { |
||||||
|
if inTable(r, t) { |
||||||
|
return true |
||||||
|
} |
||||||
|
} |
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
func inTable(r rune, t table) bool { |
||||||
|
// func (t table) IncludesRune(r rune) bool {
|
||||||
|
if r < t[0].first { |
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
bot := 0 |
||||||
|
top := len(t) - 1 |
||||||
|
for top >= bot { |
||||||
|
mid := (bot + top) >> 1 |
||||||
|
|
||||||
|
switch { |
||||||
|
case t[mid].last < r: |
||||||
|
bot = mid + 1 |
||||||
|
case t[mid].first > r: |
||||||
|
top = mid - 1 |
||||||
|
default: |
||||||
|
return true |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
var private = table{ |
||||||
|
{0x00E000, 0x00F8FF}, {0x0F0000, 0x0FFFFD}, {0x100000, 0x10FFFD}, |
||||||
|
} |
||||||
|
|
||||||
|
var nonprint = table{ |
||||||
|
{0x0000, 0x001F}, {0x007F, 0x009F}, {0x00AD, 0x00AD}, |
||||||
|
{0x070F, 0x070F}, {0x180B, 0x180E}, {0x200B, 0x200F}, |
||||||
|
{0x2028, 0x202E}, {0x206A, 0x206F}, {0xD800, 0xDFFF}, |
||||||
|
{0xFEFF, 0xFEFF}, {0xFFF9, 0xFFFB}, {0xFFFE, 0xFFFF}, |
||||||
|
} |
||||||
|
|
||||||
|
// Condition have flag EastAsianWidth whether the current locale is CJK or not.
|
||||||
|
type Condition struct { |
||||||
|
EastAsianWidth bool |
||||||
|
ZeroWidthJoiner bool |
||||||
|
} |
||||||
|
|
||||||
|
// NewCondition return new instance of Condition which is current locale.
|
||||||
|
func NewCondition() *Condition { |
||||||
|
return &Condition{ |
||||||
|
EastAsianWidth: EastAsianWidth, |
||||||
|
ZeroWidthJoiner: ZeroWidthJoiner, |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// RuneWidth returns the number of cells in r.
|
||||||
|
// See http://www.unicode.org/reports/tr11/
|
||||||
|
func (c *Condition) RuneWidth(r rune) int { |
||||||
|
switch { |
||||||
|
case r < 0 || r > 0x10FFFF || inTables(r, nonprint, combining, notassigned): |
||||||
|
return 0 |
||||||
|
case (c.EastAsianWidth && IsAmbiguousWidth(r)) || inTables(r, doublewidth): |
||||||
|
return 2 |
||||||
|
default: |
||||||
|
return 1 |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
func (c *Condition) stringWidth(s string) (width int) { |
||||||
|
for _, r := range []rune(s) { |
||||||
|
width += c.RuneWidth(r) |
||||||
|
} |
||||||
|
return width |
||||||
|
} |
||||||
|
|
||||||
|
func (c *Condition) stringWidthZeroJoiner(s string) (width int) { |
||||||
|
r1, r2 := rune(0), rune(0) |
||||||
|
for _, r := range []rune(s) { |
||||||
|
if r == 0xFE0E || r == 0xFE0F { |
||||||
|
continue |
||||||
|
} |
||||||
|
w := c.RuneWidth(r) |
||||||
|
if r2 == 0x200D && inTables(r, emoji) && inTables(r1, emoji) { |
||||||
|
if width < w { |
||||||
|
width = w |
||||||
|
} |
||||||
|
} else { |
||||||
|
width += w |
||||||
|
} |
||||||
|
r1, r2 = r2, r |
||||||
|
} |
||||||
|
return width |
||||||
|
} |
||||||
|
|
||||||
|
// StringWidth return width as you can see
|
||||||
|
func (c *Condition) StringWidth(s string) (width int) { |
||||||
|
if c.ZeroWidthJoiner { |
||||||
|
return c.stringWidthZeroJoiner(s) |
||||||
|
} |
||||||
|
return c.stringWidth(s) |
||||||
|
} |
||||||
|
|
||||||
|
// Truncate return string truncated with w cells
|
||||||
|
func (c *Condition) Truncate(s string, w int, tail string) string { |
||||||
|
if c.StringWidth(s) <= w { |
||||||
|
return s |
||||||
|
} |
||||||
|
r := []rune(s) |
||||||
|
tw := c.StringWidth(tail) |
||||||
|
w -= tw |
||||||
|
width := 0 |
||||||
|
i := 0 |
||||||
|
for ; i < len(r); i++ { |
||||||
|
cw := c.RuneWidth(r[i]) |
||||||
|
if width+cw > w { |
||||||
|
break |
||||||
|
} |
||||||
|
width += cw |
||||||
|
} |
||||||
|
return string(r[0:i]) + tail |
||||||
|
} |
||||||
|
|
||||||
|
// Wrap return string wrapped with w cells
|
||||||
|
func (c *Condition) Wrap(s string, w int) string { |
||||||
|
width := 0 |
||||||
|
out := "" |
||||||
|
for _, r := range []rune(s) { |
||||||
|
cw := RuneWidth(r) |
||||||
|
if r == '\n' { |
||||||
|
out += string(r) |
||||||
|
width = 0 |
||||||
|
continue |
||||||
|
} else if width+cw > w { |
||||||
|
out += "\n" |
||||||
|
width = 0 |
||||||
|
out += string(r) |
||||||
|
width += cw |
||||||
|
continue |
||||||
|
} |
||||||
|
out += string(r) |
||||||
|
width += cw |
||||||
|
} |
||||||
|
return out |
||||||
|
} |
||||||
|
|
||||||
|
// FillLeft return string filled in left by spaces in w cells
|
||||||
|
func (c *Condition) FillLeft(s string, w int) string { |
||||||
|
width := c.StringWidth(s) |
||||||
|
count := w - width |
||||||
|
if count > 0 { |
||||||
|
b := make([]byte, count) |
||||||
|
for i := range b { |
||||||
|
b[i] = ' ' |
||||||
|
} |
||||||
|
return string(b) + s |
||||||
|
} |
||||||
|
return s |
||||||
|
} |
||||||
|
|
||||||
|
// FillRight return string filled in left by spaces in w cells
|
||||||
|
func (c *Condition) FillRight(s string, w int) string { |
||||||
|
width := c.StringWidth(s) |
||||||
|
count := w - width |
||||||
|
if count > 0 { |
||||||
|
b := make([]byte, count) |
||||||
|
for i := range b { |
||||||
|
b[i] = ' ' |
||||||
|
} |
||||||
|
return s + string(b) |
||||||
|
} |
||||||
|
return s |
||||||
|
} |
||||||
|
|
||||||
|
// RuneWidth returns the number of cells in r.
|
||||||
|
// See http://www.unicode.org/reports/tr11/
|
||||||
|
func RuneWidth(r rune) int { |
||||||
|
return DefaultCondition.RuneWidth(r) |
||||||
|
} |
||||||
|
|
||||||
|
// IsAmbiguousWidth returns whether is ambiguous width or not.
|
||||||
|
func IsAmbiguousWidth(r rune) bool { |
||||||
|
return inTables(r, private, ambiguous) |
||||||
|
} |
||||||
|
|
||||||
|
// IsNeutralWidth returns whether is neutral width or not.
|
||||||
|
func IsNeutralWidth(r rune) bool { |
||||||
|
return inTable(r, neutral) |
||||||
|
} |
||||||
|
|
||||||
|
// StringWidth return width as you can see
|
||||||
|
func StringWidth(s string) (width int) { |
||||||
|
return DefaultCondition.StringWidth(s) |
||||||
|
} |
||||||
|
|
||||||
|
// Truncate return string truncated with w cells
|
||||||
|
func Truncate(s string, w int, tail string) string { |
||||||
|
return DefaultCondition.Truncate(s, w, tail) |
||||||
|
} |
||||||
|
|
||||||
|
// Wrap return string wrapped with w cells
|
||||||
|
func Wrap(s string, w int) string { |
||||||
|
return DefaultCondition.Wrap(s, w) |
||||||
|
} |
||||||
|
|
||||||
|
// FillLeft return string filled in left by spaces in w cells
|
||||||
|
func FillLeft(s string, w int) string { |
||||||
|
return DefaultCondition.FillLeft(s, w) |
||||||
|
} |
||||||
|
|
||||||
|
// FillRight return string filled in left by spaces in w cells
|
||||||
|
func FillRight(s string, w int) string { |
||||||
|
return DefaultCondition.FillRight(s, w) |
||||||
|
} |
@ -0,0 +1,8 @@ |
|||||||
|
// +build appengine
|
||||||
|
|
||||||
|
package runewidth |
||||||
|
|
||||||
|
// IsEastAsian return true if the current locale is CJK
|
||||||
|
func IsEastAsian() bool { |
||||||
|
return false |
||||||
|
} |
@ -0,0 +1,9 @@ |
|||||||
|
// +build js
|
||||||
|
// +build !appengine
|
||||||
|
|
||||||
|
package runewidth |
||||||
|
|
||||||
|
func IsEastAsian() bool { |
||||||
|
// TODO: Implement this for the web. Detect east asian in a compatible way, and return true.
|
||||||
|
return false |
||||||
|
} |
@ -0,0 +1,79 @@ |
|||||||
|
// +build !windows
|
||||||
|
// +build !js
|
||||||
|
// +build !appengine
|
||||||
|
|
||||||
|
package runewidth |
||||||
|
|
||||||
|
import ( |
||||||
|
"os" |
||||||
|
"regexp" |
||||||
|
"strings" |
||||||
|
) |
||||||
|
|
||||||
|
var reLoc = regexp.MustCompile(`^[a-z][a-z][a-z]?(?:_[A-Z][A-Z])?\.(.+)`) |
||||||
|
|
||||||
|
var mblenTable = map[string]int{ |
||||||
|
"utf-8": 6, |
||||||
|
"utf8": 6, |
||||||
|
"jis": 8, |
||||||
|
"eucjp": 3, |
||||||
|
"euckr": 2, |
||||||
|
"euccn": 2, |
||||||
|
"sjis": 2, |
||||||
|
"cp932": 2, |
||||||
|
"cp51932": 2, |
||||||
|
"cp936": 2, |
||||||
|
"cp949": 2, |
||||||
|
"cp950": 2, |
||||||
|
"big5": 2, |
||||||
|
"gbk": 2, |
||||||
|
"gb2312": 2, |
||||||
|
} |
||||||
|
|
||||||
|
func isEastAsian(locale string) bool { |
||||||
|
charset := strings.ToLower(locale) |
||||||
|
r := reLoc.FindStringSubmatch(locale) |
||||||
|
if len(r) == 2 { |
||||||
|
charset = strings.ToLower(r[1]) |
||||||
|
} |
||||||
|
|
||||||
|
if strings.HasSuffix(charset, "@cjk_narrow") { |
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
for pos, b := range []byte(charset) { |
||||||
|
if b == '@' { |
||||||
|
charset = charset[:pos] |
||||||
|
break |
||||||
|
} |
||||||
|
} |
||||||
|
max := 1 |
||||||
|
if m, ok := mblenTable[charset]; ok { |
||||||
|
max = m |
||||||
|
} |
||||||
|
if max > 1 && (charset[0] != 'u' || |
||||||
|
strings.HasPrefix(locale, "ja") || |
||||||
|
strings.HasPrefix(locale, "ko") || |
||||||
|
strings.HasPrefix(locale, "zh")) { |
||||||
|
return true |
||||||
|
} |
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
// IsEastAsian return true if the current locale is CJK
|
||||||
|
func IsEastAsian() bool { |
||||||
|
locale := os.Getenv("LC_CTYPE") |
||||||
|
if locale == "" { |
||||||
|
locale = os.Getenv("LANG") |
||||||
|
} |
||||||
|
|
||||||
|
// ignore C locale
|
||||||
|
if locale == "POSIX" || locale == "C" { |
||||||
|
return false |
||||||
|
} |
||||||
|
if len(locale) > 1 && locale[0] == 'C' && (locale[1] == '.' || locale[1] == '-') { |
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
return isEastAsian(locale) |
||||||
|
} |
@ -0,0 +1,427 @@ |
|||||||
|
package runewidth |
||||||
|
|
||||||
|
var combining = table{ |
||||||
|
{0x0300, 0x036F}, {0x0483, 0x0489}, {0x07EB, 0x07F3}, |
||||||
|
{0x0C00, 0x0C00}, {0x0C04, 0x0C04}, {0x0D00, 0x0D01}, |
||||||
|
{0x135D, 0x135F}, {0x1A7F, 0x1A7F}, {0x1AB0, 0x1ABE}, |
||||||
|
{0x1B6B, 0x1B73}, {0x1DC0, 0x1DF9}, {0x1DFB, 0x1DFF}, |
||||||
|
{0x20D0, 0x20F0}, {0x2CEF, 0x2CF1}, {0x2DE0, 0x2DFF}, |
||||||
|
{0x3099, 0x309A}, {0xA66F, 0xA672}, {0xA674, 0xA67D}, |
||||||
|
{0xA69E, 0xA69F}, {0xA6F0, 0xA6F1}, {0xA8E0, 0xA8F1}, |
||||||
|
{0xFE20, 0xFE2F}, {0x101FD, 0x101FD}, {0x10376, 0x1037A}, |
||||||
|
{0x10F46, 0x10F50}, {0x11300, 0x11301}, {0x1133B, 0x1133C}, |
||||||
|
{0x11366, 0x1136C}, {0x11370, 0x11374}, {0x16AF0, 0x16AF4}, |
||||||
|
{0x1D165, 0x1D169}, {0x1D16D, 0x1D172}, {0x1D17B, 0x1D182}, |
||||||
|
{0x1D185, 0x1D18B}, {0x1D1AA, 0x1D1AD}, {0x1D242, 0x1D244}, |
||||||
|
{0x1E000, 0x1E006}, {0x1E008, 0x1E018}, {0x1E01B, 0x1E021}, |
||||||
|
{0x1E023, 0x1E024}, {0x1E026, 0x1E02A}, {0x1E8D0, 0x1E8D6}, |
||||||
|
} |
||||||
|
|
||||||
|
var doublewidth = table{ |
||||||
|
{0x1100, 0x115F}, {0x231A, 0x231B}, {0x2329, 0x232A}, |
||||||
|
{0x23E9, 0x23EC}, {0x23F0, 0x23F0}, {0x23F3, 0x23F3}, |
||||||
|
{0x25FD, 0x25FE}, {0x2614, 0x2615}, {0x2648, 0x2653}, |
||||||
|
{0x267F, 0x267F}, {0x2693, 0x2693}, {0x26A1, 0x26A1}, |
||||||
|
{0x26AA, 0x26AB}, {0x26BD, 0x26BE}, {0x26C4, 0x26C5}, |
||||||
|
{0x26CE, 0x26CE}, {0x26D4, 0x26D4}, {0x26EA, 0x26EA}, |
||||||
|
{0x26F2, 0x26F3}, {0x26F5, 0x26F5}, {0x26FA, 0x26FA}, |
||||||
|
{0x26FD, 0x26FD}, {0x2705, 0x2705}, {0x270A, 0x270B}, |
||||||
|
{0x2728, 0x2728}, {0x274C, 0x274C}, {0x274E, 0x274E}, |
||||||
|
{0x2753, 0x2755}, {0x2757, 0x2757}, {0x2795, 0x2797}, |
||||||
|
{0x27B0, 0x27B0}, {0x27BF, 0x27BF}, {0x2B1B, 0x2B1C}, |
||||||
|
{0x2B50, 0x2B50}, {0x2B55, 0x2B55}, {0x2E80, 0x2E99}, |
||||||
|
{0x2E9B, 0x2EF3}, {0x2F00, 0x2FD5}, {0x2FF0, 0x2FFB}, |
||||||
|
{0x3000, 0x303E}, {0x3041, 0x3096}, {0x3099, 0x30FF}, |
||||||
|
{0x3105, 0x312F}, {0x3131, 0x318E}, {0x3190, 0x31BA}, |
||||||
|
{0x31C0, 0x31E3}, {0x31F0, 0x321E}, {0x3220, 0x3247}, |
||||||
|
{0x3250, 0x4DBF}, {0x4E00, 0xA48C}, {0xA490, 0xA4C6}, |
||||||
|
{0xA960, 0xA97C}, {0xAC00, 0xD7A3}, {0xF900, 0xFAFF}, |
||||||
|
{0xFE10, 0xFE19}, {0xFE30, 0xFE52}, {0xFE54, 0xFE66}, |
||||||
|
{0xFE68, 0xFE6B}, {0xFF01, 0xFF60}, {0xFFE0, 0xFFE6}, |
||||||
|
{0x16FE0, 0x16FE3}, {0x17000, 0x187F7}, {0x18800, 0x18AF2}, |
||||||
|
{0x1B000, 0x1B11E}, {0x1B150, 0x1B152}, {0x1B164, 0x1B167}, |
||||||
|
{0x1B170, 0x1B2FB}, {0x1F004, 0x1F004}, {0x1F0CF, 0x1F0CF}, |
||||||
|
{0x1F18E, 0x1F18E}, {0x1F191, 0x1F19A}, {0x1F200, 0x1F202}, |
||||||
|
{0x1F210, 0x1F23B}, {0x1F240, 0x1F248}, {0x1F250, 0x1F251}, |
||||||
|
{0x1F260, 0x1F265}, {0x1F300, 0x1F320}, {0x1F32D, 0x1F335}, |
||||||
|
{0x1F337, 0x1F37C}, {0x1F37E, 0x1F393}, {0x1F3A0, 0x1F3CA}, |
||||||
|
{0x1F3CF, 0x1F3D3}, {0x1F3E0, 0x1F3F0}, {0x1F3F4, 0x1F3F4}, |
||||||
|
{0x1F3F8, 0x1F43E}, {0x1F440, 0x1F440}, {0x1F442, 0x1F4FC}, |
||||||
|
{0x1F4FF, 0x1F53D}, {0x1F54B, 0x1F54E}, {0x1F550, 0x1F567}, |
||||||
|
{0x1F57A, 0x1F57A}, {0x1F595, 0x1F596}, {0x1F5A4, 0x1F5A4}, |
||||||
|
{0x1F5FB, 0x1F64F}, {0x1F680, 0x1F6C5}, {0x1F6CC, 0x1F6CC}, |
||||||
|
{0x1F6D0, 0x1F6D2}, {0x1F6D5, 0x1F6D5}, {0x1F6EB, 0x1F6EC}, |
||||||
|
{0x1F6F4, 0x1F6FA}, {0x1F7E0, 0x1F7EB}, {0x1F90D, 0x1F971}, |
||||||
|
{0x1F973, 0x1F976}, {0x1F97A, 0x1F9A2}, {0x1F9A5, 0x1F9AA}, |
||||||
|
{0x1F9AE, 0x1F9CA}, {0x1F9CD, 0x1F9FF}, {0x1FA70, 0x1FA73}, |
||||||
|
{0x1FA78, 0x1FA7A}, {0x1FA80, 0x1FA82}, {0x1FA90, 0x1FA95}, |
||||||
|
{0x20000, 0x2FFFD}, {0x30000, 0x3FFFD}, |
||||||
|
} |
||||||
|
|
||||||
|
var ambiguous = table{ |
||||||
|
{0x00A1, 0x00A1}, {0x00A4, 0x00A4}, {0x00A7, 0x00A8}, |
||||||
|
{0x00AA, 0x00AA}, {0x00AD, 0x00AE}, {0x00B0, 0x00B4}, |
||||||
|
{0x00B6, 0x00BA}, {0x00BC, 0x00BF}, {0x00C6, 0x00C6}, |
||||||
|
{0x00D0, 0x00D0}, {0x00D7, 0x00D8}, {0x00DE, 0x00E1}, |
||||||
|
{0x00E6, 0x00E6}, {0x00E8, 0x00EA}, {0x00EC, 0x00ED}, |
||||||
|
{0x00F0, 0x00F0}, {0x00F2, 0x00F3}, {0x00F7, 0x00FA}, |
||||||
|
{0x00FC, 0x00FC}, {0x00FE, 0x00FE}, {0x0101, 0x0101}, |
||||||
|
{0x0111, 0x0111}, {0x0113, 0x0113}, {0x011B, 0x011B}, |
||||||
|
{0x0126, 0x0127}, {0x012B, 0x012B}, {0x0131, 0x0133}, |
||||||
|
{0x0138, 0x0138}, {0x013F, 0x0142}, {0x0144, 0x0144}, |
||||||
|
{0x0148, 0x014B}, {0x014D, 0x014D}, {0x0152, 0x0153}, |
||||||
|
{0x0166, 0x0167}, {0x016B, 0x016B}, {0x01CE, 0x01CE}, |
||||||
|
{0x01D0, 0x01D0}, {0x01D2, 0x01D2}, {0x01D4, 0x01D4}, |
||||||
|
{0x01D6, 0x01D6}, {0x01D8, 0x01D8}, {0x01DA, 0x01DA}, |
||||||
|
{0x01DC, 0x01DC}, {0x0251, 0x0251}, {0x0261, 0x0261}, |
||||||
|
{0x02C4, 0x02C4}, {0x02C7, 0x02C7}, {0x02C9, 0x02CB}, |
||||||
|
{0x02CD, 0x02CD}, {0x02D0, 0x02D0}, {0x02D8, 0x02DB}, |
||||||
|
{0x02DD, 0x02DD}, {0x02DF, 0x02DF}, {0x0300, 0x036F}, |
||||||
|
{0x0391, 0x03A1}, {0x03A3, 0x03A9}, {0x03B1, 0x03C1}, |
||||||
|
{0x03C3, 0x03C9}, {0x0401, 0x0401}, {0x0410, 0x044F}, |
||||||
|
{0x0451, 0x0451}, {0x2010, 0x2010}, {0x2013, 0x2016}, |
||||||
|
{0x2018, 0x2019}, {0x201C, 0x201D}, {0x2020, 0x2022}, |
||||||
|
{0x2024, 0x2027}, {0x2030, 0x2030}, {0x2032, 0x2033}, |
||||||
|
{0x2035, 0x2035}, {0x203B, 0x203B}, {0x203E, 0x203E}, |
||||||
|
{0x2074, 0x2074}, {0x207F, 0x207F}, {0x2081, 0x2084}, |
||||||
|
{0x20AC, 0x20AC}, {0x2103, 0x2103}, {0x2105, 0x2105}, |
||||||
|
{0x2109, 0x2109}, {0x2113, 0x2113}, {0x2116, 0x2116}, |
||||||
|
{0x2121, 0x2122}, {0x2126, 0x2126}, {0x212B, 0x212B}, |
||||||
|
{0x2153, 0x2154}, {0x215B, 0x215E}, {0x2160, 0x216B}, |
||||||
|
{0x2170, 0x2179}, {0x2189, 0x2189}, {0x2190, 0x2199}, |
||||||
|
{0x21B8, 0x21B9}, {0x21D2, 0x21D2}, {0x21D4, 0x21D4}, |
||||||
|
{0x21E7, 0x21E7}, {0x2200, 0x2200}, {0x2202, 0x2203}, |
||||||
|
{0x2207, 0x2208}, {0x220B, 0x220B}, {0x220F, 0x220F}, |
||||||
|
{0x2211, 0x2211}, {0x2215, 0x2215}, {0x221A, 0x221A}, |
||||||
|
{0x221D, 0x2220}, {0x2223, 0x2223}, {0x2225, 0x2225}, |
||||||
|
{0x2227, 0x222C}, {0x222E, 0x222E}, {0x2234, 0x2237}, |
||||||
|
{0x223C, 0x223D}, {0x2248, 0x2248}, {0x224C, 0x224C}, |
||||||
|
{0x2252, 0x2252}, {0x2260, 0x2261}, {0x2264, 0x2267}, |
||||||
|
{0x226A, 0x226B}, {0x226E, 0x226F}, {0x2282, 0x2283}, |
||||||
|
{0x2286, 0x2287}, {0x2295, 0x2295}, {0x2299, 0x2299}, |
||||||
|
{0x22A5, 0x22A5}, {0x22BF, 0x22BF}, {0x2312, 0x2312}, |
||||||
|
{0x2460, 0x24E9}, {0x24EB, 0x254B}, {0x2550, 0x2573}, |
||||||
|
{0x2580, 0x258F}, {0x2592, 0x2595}, {0x25A0, 0x25A1}, |
||||||
|
{0x25A3, 0x25A9}, {0x25B2, 0x25B3}, {0x25B6, 0x25B7}, |
||||||
|
{0x25BC, 0x25BD}, {0x25C0, 0x25C1}, {0x25C6, 0x25C8}, |
||||||
|
{0x25CB, 0x25CB}, {0x25CE, 0x25D1}, {0x25E2, 0x25E5}, |
||||||
|
{0x25EF, 0x25EF}, {0x2605, 0x2606}, {0x2609, 0x2609}, |
||||||
|
{0x260E, 0x260F}, {0x261C, 0x261C}, {0x261E, 0x261E}, |
||||||
|
{0x2640, 0x2640}, {0x2642, 0x2642}, {0x2660, 0x2661}, |
||||||
|
{0x2663, 0x2665}, {0x2667, 0x266A}, {0x266C, 0x266D}, |
||||||
|
{0x266F, 0x266F}, {0x269E, 0x269F}, {0x26BF, 0x26BF}, |
||||||
|
{0x26C6, 0x26CD}, {0x26CF, 0x26D3}, {0x26D5, 0x26E1}, |
||||||
|
{0x26E3, 0x26E3}, {0x26E8, 0x26E9}, {0x26EB, 0x26F1}, |
||||||
|
{0x26F4, 0x26F4}, {0x26F6, 0x26F9}, {0x26FB, 0x26FC}, |
||||||
|
{0x26FE, 0x26FF}, {0x273D, 0x273D}, {0x2776, 0x277F}, |
||||||
|
{0x2B56, 0x2B59}, {0x3248, 0x324F}, {0xE000, 0xF8FF}, |
||||||
|
{0xFE00, 0xFE0F}, {0xFFFD, 0xFFFD}, {0x1F100, 0x1F10A}, |
||||||
|
{0x1F110, 0x1F12D}, {0x1F130, 0x1F169}, {0x1F170, 0x1F18D}, |
||||||
|
{0x1F18F, 0x1F190}, {0x1F19B, 0x1F1AC}, {0xE0100, 0xE01EF}, |
||||||
|
{0xF0000, 0xFFFFD}, {0x100000, 0x10FFFD}, |
||||||
|
} |
||||||
|
var notassigned = table{ |
||||||
|
{0x27E6, 0x27ED}, {0x2985, 0x2986}, |
||||||
|
} |
||||||
|
|
||||||
|
var neutral = table{ |
||||||
|
{0x0000, 0x001F}, {0x007F, 0x00A0}, {0x00A9, 0x00A9}, |
||||||
|
{0x00AB, 0x00AB}, {0x00B5, 0x00B5}, {0x00BB, 0x00BB}, |
||||||
|
{0x00C0, 0x00C5}, {0x00C7, 0x00CF}, {0x00D1, 0x00D6}, |
||||||
|
{0x00D9, 0x00DD}, {0x00E2, 0x00E5}, {0x00E7, 0x00E7}, |
||||||
|
{0x00EB, 0x00EB}, {0x00EE, 0x00EF}, {0x00F1, 0x00F1}, |
||||||
|
{0x00F4, 0x00F6}, {0x00FB, 0x00FB}, {0x00FD, 0x00FD}, |
||||||
|
{0x00FF, 0x0100}, {0x0102, 0x0110}, {0x0112, 0x0112}, |
||||||
|
{0x0114, 0x011A}, {0x011C, 0x0125}, {0x0128, 0x012A}, |
||||||
|
{0x012C, 0x0130}, {0x0134, 0x0137}, {0x0139, 0x013E}, |
||||||
|
{0x0143, 0x0143}, {0x0145, 0x0147}, {0x014C, 0x014C}, |
||||||
|
{0x014E, 0x0151}, {0x0154, 0x0165}, {0x0168, 0x016A}, |
||||||
|
{0x016C, 0x01CD}, {0x01CF, 0x01CF}, {0x01D1, 0x01D1}, |
||||||
|
{0x01D3, 0x01D3}, {0x01D5, 0x01D5}, {0x01D7, 0x01D7}, |
||||||
|
{0x01D9, 0x01D9}, {0x01DB, 0x01DB}, {0x01DD, 0x0250}, |
||||||
|
{0x0252, 0x0260}, {0x0262, 0x02C3}, {0x02C5, 0x02C6}, |
||||||
|
{0x02C8, 0x02C8}, {0x02CC, 0x02CC}, {0x02CE, 0x02CF}, |
||||||
|
{0x02D1, 0x02D7}, {0x02DC, 0x02DC}, {0x02DE, 0x02DE}, |
||||||
|
{0x02E0, 0x02FF}, {0x0370, 0x0377}, {0x037A, 0x037F}, |
||||||
|
{0x0384, 0x038A}, {0x038C, 0x038C}, {0x038E, 0x0390}, |
||||||
|
{0x03AA, 0x03B0}, {0x03C2, 0x03C2}, {0x03CA, 0x0400}, |
||||||
|
{0x0402, 0x040F}, {0x0450, 0x0450}, {0x0452, 0x052F}, |
||||||
|
{0x0531, 0x0556}, {0x0559, 0x058A}, {0x058D, 0x058F}, |
||||||
|
{0x0591, 0x05C7}, {0x05D0, 0x05EA}, {0x05EF, 0x05F4}, |
||||||
|
{0x0600, 0x061C}, {0x061E, 0x070D}, {0x070F, 0x074A}, |
||||||
|
{0x074D, 0x07B1}, {0x07C0, 0x07FA}, {0x07FD, 0x082D}, |
||||||
|
{0x0830, 0x083E}, {0x0840, 0x085B}, {0x085E, 0x085E}, |
||||||
|
{0x0860, 0x086A}, {0x08A0, 0x08B4}, {0x08B6, 0x08BD}, |
||||||
|
{0x08D3, 0x0983}, {0x0985, 0x098C}, {0x098F, 0x0990}, |
||||||
|
{0x0993, 0x09A8}, {0x09AA, 0x09B0}, {0x09B2, 0x09B2}, |
||||||
|
{0x09B6, 0x09B9}, {0x09BC, 0x09C4}, {0x09C7, 0x09C8}, |
||||||
|
{0x09CB, 0x09CE}, {0x09D7, 0x09D7}, {0x09DC, 0x09DD}, |
||||||
|
{0x09DF, 0x09E3}, {0x09E6, 0x09FE}, {0x0A01, 0x0A03}, |
||||||
|
{0x0A05, 0x0A0A}, {0x0A0F, 0x0A10}, {0x0A13, 0x0A28}, |
||||||
|
{0x0A2A, 0x0A30}, {0x0A32, 0x0A33}, {0x0A35, 0x0A36}, |
||||||
|
{0x0A38, 0x0A39}, {0x0A3C, 0x0A3C}, {0x0A3E, 0x0A42}, |
||||||
|
{0x0A47, 0x0A48}, {0x0A4B, 0x0A4D}, {0x0A51, 0x0A51}, |
||||||
|
{0x0A59, 0x0A5C}, {0x0A5E, 0x0A5E}, {0x0A66, 0x0A76}, |
||||||
|
{0x0A81, 0x0A83}, {0x0A85, 0x0A8D}, {0x0A8F, 0x0A91}, |
||||||
|
{0x0A93, 0x0AA8}, {0x0AAA, 0x0AB0}, {0x0AB2, 0x0AB3}, |
||||||
|
{0x0AB5, 0x0AB9}, {0x0ABC, 0x0AC5}, {0x0AC7, 0x0AC9}, |
||||||
|
{0x0ACB, 0x0ACD}, {0x0AD0, 0x0AD0}, {0x0AE0, 0x0AE3}, |
||||||
|
{0x0AE6, 0x0AF1}, {0x0AF9, 0x0AFF}, {0x0B01, 0x0B03}, |
||||||
|
{0x0B05, 0x0B0C}, {0x0B0F, 0x0B10}, {0x0B13, 0x0B28}, |
||||||
|
{0x0B2A, 0x0B30}, {0x0B32, 0x0B33}, {0x0B35, 0x0B39}, |
||||||
|
{0x0B3C, 0x0B44}, {0x0B47, 0x0B48}, {0x0B4B, 0x0B4D}, |
||||||
|
{0x0B56, 0x0B57}, {0x0B5C, 0x0B5D}, {0x0B5F, 0x0B63}, |
||||||
|
{0x0B66, 0x0B77}, {0x0B82, 0x0B83}, {0x0B85, 0x0B8A}, |
||||||
|
{0x0B8E, 0x0B90}, {0x0B92, 0x0B95}, {0x0B99, 0x0B9A}, |
||||||
|
{0x0B9C, 0x0B9C}, {0x0B9E, 0x0B9F}, {0x0BA3, 0x0BA4}, |
||||||
|
{0x0BA8, 0x0BAA}, {0x0BAE, 0x0BB9}, {0x0BBE, 0x0BC2}, |
||||||
|
{0x0BC6, 0x0BC8}, {0x0BCA, 0x0BCD}, {0x0BD0, 0x0BD0}, |
||||||
|
{0x0BD7, 0x0BD7}, {0x0BE6, 0x0BFA}, {0x0C00, 0x0C0C}, |
||||||
|
{0x0C0E, 0x0C10}, {0x0C12, 0x0C28}, {0x0C2A, 0x0C39}, |
||||||
|
{0x0C3D, 0x0C44}, {0x0C46, 0x0C48}, {0x0C4A, 0x0C4D}, |
||||||
|
{0x0C55, 0x0C56}, {0x0C58, 0x0C5A}, {0x0C60, 0x0C63}, |
||||||
|
{0x0C66, 0x0C6F}, {0x0C77, 0x0C8C}, {0x0C8E, 0x0C90}, |
||||||
|
{0x0C92, 0x0CA8}, {0x0CAA, 0x0CB3}, {0x0CB5, 0x0CB9}, |
||||||
|
{0x0CBC, 0x0CC4}, {0x0CC6, 0x0CC8}, {0x0CCA, 0x0CCD}, |
||||||
|
{0x0CD5, 0x0CD6}, {0x0CDE, 0x0CDE}, {0x0CE0, 0x0CE3}, |
||||||
|
{0x0CE6, 0x0CEF}, {0x0CF1, 0x0CF2}, {0x0D00, 0x0D03}, |
||||||
|
{0x0D05, 0x0D0C}, {0x0D0E, 0x0D10}, {0x0D12, 0x0D44}, |
||||||
|
{0x0D46, 0x0D48}, {0x0D4A, 0x0D4F}, {0x0D54, 0x0D63}, |
||||||
|
{0x0D66, 0x0D7F}, {0x0D82, 0x0D83}, {0x0D85, 0x0D96}, |
||||||
|
{0x0D9A, 0x0DB1}, {0x0DB3, 0x0DBB}, {0x0DBD, 0x0DBD}, |
||||||
|
{0x0DC0, 0x0DC6}, {0x0DCA, 0x0DCA}, {0x0DCF, 0x0DD4}, |
||||||
|
{0x0DD6, 0x0DD6}, {0x0DD8, 0x0DDF}, {0x0DE6, 0x0DEF}, |
||||||
|
{0x0DF2, 0x0DF4}, {0x0E01, 0x0E3A}, {0x0E3F, 0x0E5B}, |
||||||
|
{0x0E81, 0x0E82}, {0x0E84, 0x0E84}, {0x0E86, 0x0E8A}, |
||||||
|
{0x0E8C, 0x0EA3}, {0x0EA5, 0x0EA5}, {0x0EA7, 0x0EBD}, |
||||||
|
{0x0EC0, 0x0EC4}, {0x0EC6, 0x0EC6}, {0x0EC8, 0x0ECD}, |
||||||
|
{0x0ED0, 0x0ED9}, {0x0EDC, 0x0EDF}, {0x0F00, 0x0F47}, |
||||||
|
{0x0F49, 0x0F6C}, {0x0F71, 0x0F97}, {0x0F99, 0x0FBC}, |
||||||
|
{0x0FBE, 0x0FCC}, {0x0FCE, 0x0FDA}, {0x1000, 0x10C5}, |
||||||
|
{0x10C7, 0x10C7}, {0x10CD, 0x10CD}, {0x10D0, 0x10FF}, |
||||||
|
{0x1160, 0x1248}, {0x124A, 0x124D}, {0x1250, 0x1256}, |
||||||
|
{0x1258, 0x1258}, {0x125A, 0x125D}, {0x1260, 0x1288}, |
||||||
|
{0x128A, 0x128D}, {0x1290, 0x12B0}, {0x12B2, 0x12B5}, |
||||||
|
{0x12B8, 0x12BE}, {0x12C0, 0x12C0}, {0x12C2, 0x12C5}, |
||||||
|
{0x12C8, 0x12D6}, {0x12D8, 0x1310}, {0x1312, 0x1315}, |
||||||
|
{0x1318, 0x135A}, {0x135D, 0x137C}, {0x1380, 0x1399}, |
||||||
|
{0x13A0, 0x13F5}, {0x13F8, 0x13FD}, {0x1400, 0x169C}, |
||||||
|
{0x16A0, 0x16F8}, {0x1700, 0x170C}, {0x170E, 0x1714}, |
||||||
|
{0x1720, 0x1736}, {0x1740, 0x1753}, {0x1760, 0x176C}, |
||||||
|
{0x176E, 0x1770}, {0x1772, 0x1773}, {0x1780, 0x17DD}, |
||||||
|
{0x17E0, 0x17E9}, {0x17F0, 0x17F9}, {0x1800, 0x180E}, |
||||||
|
{0x1810, 0x1819}, {0x1820, 0x1878}, {0x1880, 0x18AA}, |
||||||
|
{0x18B0, 0x18F5}, {0x1900, 0x191E}, {0x1920, 0x192B}, |
||||||
|
{0x1930, 0x193B}, {0x1940, 0x1940}, {0x1944, 0x196D}, |
||||||
|
{0x1970, 0x1974}, {0x1980, 0x19AB}, {0x19B0, 0x19C9}, |
||||||
|
{0x19D0, 0x19DA}, {0x19DE, 0x1A1B}, {0x1A1E, 0x1A5E}, |
||||||
|
{0x1A60, 0x1A7C}, {0x1A7F, 0x1A89}, {0x1A90, 0x1A99}, |
||||||
|
{0x1AA0, 0x1AAD}, {0x1AB0, 0x1ABE}, {0x1B00, 0x1B4B}, |
||||||
|
{0x1B50, 0x1B7C}, {0x1B80, 0x1BF3}, {0x1BFC, 0x1C37}, |
||||||
|
{0x1C3B, 0x1C49}, {0x1C4D, 0x1C88}, {0x1C90, 0x1CBA}, |
||||||
|
{0x1CBD, 0x1CC7}, {0x1CD0, 0x1CFA}, {0x1D00, 0x1DF9}, |
||||||
|
{0x1DFB, 0x1F15}, {0x1F18, 0x1F1D}, {0x1F20, 0x1F45}, |
||||||
|
{0x1F48, 0x1F4D}, {0x1F50, 0x1F57}, {0x1F59, 0x1F59}, |
||||||
|
{0x1F5B, 0x1F5B}, {0x1F5D, 0x1F5D}, {0x1F5F, 0x1F7D}, |
||||||
|
{0x1F80, 0x1FB4}, {0x1FB6, 0x1FC4}, {0x1FC6, 0x1FD3}, |
||||||
|
{0x1FD6, 0x1FDB}, {0x1FDD, 0x1FEF}, {0x1FF2, 0x1FF4}, |
||||||
|
{0x1FF6, 0x1FFE}, {0x2000, 0x200F}, {0x2011, 0x2012}, |
||||||
|
{0x2017, 0x2017}, {0x201A, 0x201B}, {0x201E, 0x201F}, |
||||||
|
{0x2023, 0x2023}, {0x2028, 0x202F}, {0x2031, 0x2031}, |
||||||
|
{0x2034, 0x2034}, {0x2036, 0x203A}, {0x203C, 0x203D}, |
||||||
|
{0x203F, 0x2064}, {0x2066, 0x2071}, {0x2075, 0x207E}, |
||||||
|
{0x2080, 0x2080}, {0x2085, 0x208E}, {0x2090, 0x209C}, |
||||||
|
{0x20A0, 0x20A8}, {0x20AA, 0x20AB}, {0x20AD, 0x20BF}, |
||||||
|
{0x20D0, 0x20F0}, {0x2100, 0x2102}, {0x2104, 0x2104}, |
||||||
|
{0x2106, 0x2108}, {0x210A, 0x2112}, {0x2114, 0x2115}, |
||||||
|
{0x2117, 0x2120}, {0x2123, 0x2125}, {0x2127, 0x212A}, |
||||||
|
{0x212C, 0x2152}, {0x2155, 0x215A}, {0x215F, 0x215F}, |
||||||
|
{0x216C, 0x216F}, {0x217A, 0x2188}, {0x218A, 0x218B}, |
||||||
|
{0x219A, 0x21B7}, {0x21BA, 0x21D1}, {0x21D3, 0x21D3}, |
||||||
|
{0x21D5, 0x21E6}, {0x21E8, 0x21FF}, {0x2201, 0x2201}, |
||||||
|
{0x2204, 0x2206}, {0x2209, 0x220A}, {0x220C, 0x220E}, |
||||||
|
{0x2210, 0x2210}, {0x2212, 0x2214}, {0x2216, 0x2219}, |
||||||
|
{0x221B, 0x221C}, {0x2221, 0x2222}, {0x2224, 0x2224}, |
||||||
|
{0x2226, 0x2226}, {0x222D, 0x222D}, {0x222F, 0x2233}, |
||||||
|
{0x2238, 0x223B}, {0x223E, 0x2247}, {0x2249, 0x224B}, |
||||||
|
{0x224D, 0x2251}, {0x2253, 0x225F}, {0x2262, 0x2263}, |
||||||
|
{0x2268, 0x2269}, {0x226C, 0x226D}, {0x2270, 0x2281}, |
||||||
|
{0x2284, 0x2285}, {0x2288, 0x2294}, {0x2296, 0x2298}, |
||||||
|
{0x229A, 0x22A4}, {0x22A6, 0x22BE}, {0x22C0, 0x2311}, |
||||||
|
{0x2313, 0x2319}, {0x231C, 0x2328}, {0x232B, 0x23E8}, |
||||||
|
{0x23ED, 0x23EF}, {0x23F1, 0x23F2}, {0x23F4, 0x2426}, |
||||||
|
{0x2440, 0x244A}, {0x24EA, 0x24EA}, {0x254C, 0x254F}, |
||||||
|
{0x2574, 0x257F}, {0x2590, 0x2591}, {0x2596, 0x259F}, |
||||||
|
{0x25A2, 0x25A2}, {0x25AA, 0x25B1}, {0x25B4, 0x25B5}, |
||||||
|
{0x25B8, 0x25BB}, {0x25BE, 0x25BF}, {0x25C2, 0x25C5}, |
||||||
|
{0x25C9, 0x25CA}, {0x25CC, 0x25CD}, {0x25D2, 0x25E1}, |
||||||
|
{0x25E6, 0x25EE}, {0x25F0, 0x25FC}, {0x25FF, 0x2604}, |
||||||
|
{0x2607, 0x2608}, {0x260A, 0x260D}, {0x2610, 0x2613}, |
||||||
|
{0x2616, 0x261B}, {0x261D, 0x261D}, {0x261F, 0x263F}, |
||||||
|
{0x2641, 0x2641}, {0x2643, 0x2647}, {0x2654, 0x265F}, |
||||||
|
{0x2662, 0x2662}, {0x2666, 0x2666}, {0x266B, 0x266B}, |
||||||
|
{0x266E, 0x266E}, {0x2670, 0x267E}, {0x2680, 0x2692}, |
||||||
|
{0x2694, 0x269D}, {0x26A0, 0x26A0}, {0x26A2, 0x26A9}, |
||||||
|
{0x26AC, 0x26BC}, {0x26C0, 0x26C3}, {0x26E2, 0x26E2}, |
||||||
|
{0x26E4, 0x26E7}, {0x2700, 0x2704}, {0x2706, 0x2709}, |
||||||
|
{0x270C, 0x2727}, {0x2729, 0x273C}, {0x273E, 0x274B}, |
||||||
|
{0x274D, 0x274D}, {0x274F, 0x2752}, {0x2756, 0x2756}, |
||||||
|
{0x2758, 0x2775}, {0x2780, 0x2794}, {0x2798, 0x27AF}, |
||||||
|
{0x27B1, 0x27BE}, {0x27C0, 0x27E5}, {0x27EE, 0x2984}, |
||||||
|
{0x2987, 0x2B1A}, {0x2B1D, 0x2B4F}, {0x2B51, 0x2B54}, |
||||||
|
{0x2B5A, 0x2B73}, {0x2B76, 0x2B95}, {0x2B98, 0x2C2E}, |
||||||
|
{0x2C30, 0x2C5E}, {0x2C60, 0x2CF3}, {0x2CF9, 0x2D25}, |
||||||
|
{0x2D27, 0x2D27}, {0x2D2D, 0x2D2D}, {0x2D30, 0x2D67}, |
||||||
|
{0x2D6F, 0x2D70}, {0x2D7F, 0x2D96}, {0x2DA0, 0x2DA6}, |
||||||
|
{0x2DA8, 0x2DAE}, {0x2DB0, 0x2DB6}, {0x2DB8, 0x2DBE}, |
||||||
|
{0x2DC0, 0x2DC6}, {0x2DC8, 0x2DCE}, {0x2DD0, 0x2DD6}, |
||||||
|
{0x2DD8, 0x2DDE}, {0x2DE0, 0x2E4F}, {0x303F, 0x303F}, |
||||||
|
{0x4DC0, 0x4DFF}, {0xA4D0, 0xA62B}, {0xA640, 0xA6F7}, |
||||||
|
{0xA700, 0xA7BF}, {0xA7C2, 0xA7C6}, {0xA7F7, 0xA82B}, |
||||||
|
{0xA830, 0xA839}, {0xA840, 0xA877}, {0xA880, 0xA8C5}, |
||||||
|
{0xA8CE, 0xA8D9}, {0xA8E0, 0xA953}, {0xA95F, 0xA95F}, |
||||||
|
{0xA980, 0xA9CD}, {0xA9CF, 0xA9D9}, {0xA9DE, 0xA9FE}, |
||||||
|
{0xAA00, 0xAA36}, {0xAA40, 0xAA4D}, {0xAA50, 0xAA59}, |
||||||
|
{0xAA5C, 0xAAC2}, {0xAADB, 0xAAF6}, {0xAB01, 0xAB06}, |
||||||
|
{0xAB09, 0xAB0E}, {0xAB11, 0xAB16}, {0xAB20, 0xAB26}, |
||||||
|
{0xAB28, 0xAB2E}, {0xAB30, 0xAB67}, {0xAB70, 0xABED}, |
||||||
|
{0xABF0, 0xABF9}, {0xD7B0, 0xD7C6}, {0xD7CB, 0xD7FB}, |
||||||
|
{0xD800, 0xDFFF}, {0xFB00, 0xFB06}, {0xFB13, 0xFB17}, |
||||||
|
{0xFB1D, 0xFB36}, {0xFB38, 0xFB3C}, {0xFB3E, 0xFB3E}, |
||||||
|
{0xFB40, 0xFB41}, {0xFB43, 0xFB44}, {0xFB46, 0xFBC1}, |
||||||
|
{0xFBD3, 0xFD3F}, {0xFD50, 0xFD8F}, {0xFD92, 0xFDC7}, |
||||||
|
{0xFDF0, 0xFDFD}, {0xFE20, 0xFE2F}, {0xFE70, 0xFE74}, |
||||||
|
{0xFE76, 0xFEFC}, {0xFEFF, 0xFEFF}, {0xFFF9, 0xFFFC}, |
||||||
|
{0x10000, 0x1000B}, {0x1000D, 0x10026}, {0x10028, 0x1003A}, |
||||||
|
{0x1003C, 0x1003D}, {0x1003F, 0x1004D}, {0x10050, 0x1005D}, |
||||||
|
{0x10080, 0x100FA}, {0x10100, 0x10102}, {0x10107, 0x10133}, |
||||||
|
{0x10137, 0x1018E}, {0x10190, 0x1019B}, {0x101A0, 0x101A0}, |
||||||
|
{0x101D0, 0x101FD}, {0x10280, 0x1029C}, {0x102A0, 0x102D0}, |
||||||
|
{0x102E0, 0x102FB}, {0x10300, 0x10323}, {0x1032D, 0x1034A}, |
||||||
|
{0x10350, 0x1037A}, {0x10380, 0x1039D}, {0x1039F, 0x103C3}, |
||||||
|
{0x103C8, 0x103D5}, {0x10400, 0x1049D}, {0x104A0, 0x104A9}, |
||||||
|
{0x104B0, 0x104D3}, {0x104D8, 0x104FB}, {0x10500, 0x10527}, |
||||||
|
{0x10530, 0x10563}, {0x1056F, 0x1056F}, {0x10600, 0x10736}, |
||||||
|
{0x10740, 0x10755}, {0x10760, 0x10767}, {0x10800, 0x10805}, |
||||||
|
{0x10808, 0x10808}, {0x1080A, 0x10835}, {0x10837, 0x10838}, |
||||||
|
{0x1083C, 0x1083C}, {0x1083F, 0x10855}, {0x10857, 0x1089E}, |
||||||
|
{0x108A7, 0x108AF}, {0x108E0, 0x108F2}, {0x108F4, 0x108F5}, |
||||||
|
{0x108FB, 0x1091B}, {0x1091F, 0x10939}, {0x1093F, 0x1093F}, |
||||||
|
{0x10980, 0x109B7}, {0x109BC, 0x109CF}, {0x109D2, 0x10A03}, |
||||||
|
{0x10A05, 0x10A06}, {0x10A0C, 0x10A13}, {0x10A15, 0x10A17}, |
||||||
|
{0x10A19, 0x10A35}, {0x10A38, 0x10A3A}, {0x10A3F, 0x10A48}, |
||||||
|
{0x10A50, 0x10A58}, {0x10A60, 0x10A9F}, {0x10AC0, 0x10AE6}, |
||||||
|
{0x10AEB, 0x10AF6}, {0x10B00, 0x10B35}, {0x10B39, 0x10B55}, |
||||||
|
{0x10B58, 0x10B72}, {0x10B78, 0x10B91}, {0x10B99, 0x10B9C}, |
||||||
|
{0x10BA9, 0x10BAF}, {0x10C00, 0x10C48}, {0x10C80, 0x10CB2}, |
||||||
|
{0x10CC0, 0x10CF2}, {0x10CFA, 0x10D27}, {0x10D30, 0x10D39}, |
||||||
|
{0x10E60, 0x10E7E}, {0x10F00, 0x10F27}, {0x10F30, 0x10F59}, |
||||||
|
{0x10FE0, 0x10FF6}, {0x11000, 0x1104D}, {0x11052, 0x1106F}, |
||||||
|
{0x1107F, 0x110C1}, {0x110CD, 0x110CD}, {0x110D0, 0x110E8}, |
||||||
|
{0x110F0, 0x110F9}, {0x11100, 0x11134}, {0x11136, 0x11146}, |
||||||
|
{0x11150, 0x11176}, {0x11180, 0x111CD}, {0x111D0, 0x111DF}, |
||||||
|
{0x111E1, 0x111F4}, {0x11200, 0x11211}, {0x11213, 0x1123E}, |
||||||
|
{0x11280, 0x11286}, {0x11288, 0x11288}, {0x1128A, 0x1128D}, |
||||||
|
{0x1128F, 0x1129D}, {0x1129F, 0x112A9}, {0x112B0, 0x112EA}, |
||||||
|
{0x112F0, 0x112F9}, {0x11300, 0x11303}, {0x11305, 0x1130C}, |
||||||
|
{0x1130F, 0x11310}, {0x11313, 0x11328}, {0x1132A, 0x11330}, |
||||||
|
{0x11332, 0x11333}, {0x11335, 0x11339}, {0x1133B, 0x11344}, |
||||||
|
{0x11347, 0x11348}, {0x1134B, 0x1134D}, {0x11350, 0x11350}, |
||||||
|
{0x11357, 0x11357}, {0x1135D, 0x11363}, {0x11366, 0x1136C}, |
||||||
|
{0x11370, 0x11374}, {0x11400, 0x11459}, {0x1145B, 0x1145B}, |
||||||
|
{0x1145D, 0x1145F}, {0x11480, 0x114C7}, {0x114D0, 0x114D9}, |
||||||
|
{0x11580, 0x115B5}, {0x115B8, 0x115DD}, {0x11600, 0x11644}, |
||||||
|
{0x11650, 0x11659}, {0x11660, 0x1166C}, {0x11680, 0x116B8}, |
||||||
|
{0x116C0, 0x116C9}, {0x11700, 0x1171A}, {0x1171D, 0x1172B}, |
||||||
|
{0x11730, 0x1173F}, {0x11800, 0x1183B}, {0x118A0, 0x118F2}, |
||||||
|
{0x118FF, 0x118FF}, {0x119A0, 0x119A7}, {0x119AA, 0x119D7}, |
||||||
|
{0x119DA, 0x119E4}, {0x11A00, 0x11A47}, {0x11A50, 0x11AA2}, |
||||||
|
{0x11AC0, 0x11AF8}, {0x11C00, 0x11C08}, {0x11C0A, 0x11C36}, |
||||||
|
{0x11C38, 0x11C45}, {0x11C50, 0x11C6C}, {0x11C70, 0x11C8F}, |
||||||
|
{0x11C92, 0x11CA7}, {0x11CA9, 0x11CB6}, {0x11D00, 0x11D06}, |
||||||
|
{0x11D08, 0x11D09}, {0x11D0B, 0x11D36}, {0x11D3A, 0x11D3A}, |
||||||
|
{0x11D3C, 0x11D3D}, {0x11D3F, 0x11D47}, {0x11D50, 0x11D59}, |
||||||
|
{0x11D60, 0x11D65}, {0x11D67, 0x11D68}, {0x11D6A, 0x11D8E}, |
||||||
|
{0x11D90, 0x11D91}, {0x11D93, 0x11D98}, {0x11DA0, 0x11DA9}, |
||||||
|
{0x11EE0, 0x11EF8}, {0x11FC0, 0x11FF1}, {0x11FFF, 0x12399}, |
||||||
|
{0x12400, 0x1246E}, {0x12470, 0x12474}, {0x12480, 0x12543}, |
||||||
|
{0x13000, 0x1342E}, {0x13430, 0x13438}, {0x14400, 0x14646}, |
||||||
|
{0x16800, 0x16A38}, {0x16A40, 0x16A5E}, {0x16A60, 0x16A69}, |
||||||
|
{0x16A6E, 0x16A6F}, {0x16AD0, 0x16AED}, {0x16AF0, 0x16AF5}, |
||||||
|
{0x16B00, 0x16B45}, {0x16B50, 0x16B59}, {0x16B5B, 0x16B61}, |
||||||
|
{0x16B63, 0x16B77}, {0x16B7D, 0x16B8F}, {0x16E40, 0x16E9A}, |
||||||
|
{0x16F00, 0x16F4A}, {0x16F4F, 0x16F87}, {0x16F8F, 0x16F9F}, |
||||||
|
{0x1BC00, 0x1BC6A}, {0x1BC70, 0x1BC7C}, {0x1BC80, 0x1BC88}, |
||||||
|
{0x1BC90, 0x1BC99}, {0x1BC9C, 0x1BCA3}, {0x1D000, 0x1D0F5}, |
||||||
|
{0x1D100, 0x1D126}, {0x1D129, 0x1D1E8}, {0x1D200, 0x1D245}, |
||||||
|
{0x1D2E0, 0x1D2F3}, {0x1D300, 0x1D356}, {0x1D360, 0x1D378}, |
||||||
|
{0x1D400, 0x1D454}, {0x1D456, 0x1D49C}, {0x1D49E, 0x1D49F}, |
||||||
|
{0x1D4A2, 0x1D4A2}, {0x1D4A5, 0x1D4A6}, {0x1D4A9, 0x1D4AC}, |
||||||
|
{0x1D4AE, 0x1D4B9}, {0x1D4BB, 0x1D4BB}, {0x1D4BD, 0x1D4C3}, |
||||||
|
{0x1D4C5, 0x1D505}, {0x1D507, 0x1D50A}, {0x1D50D, 0x1D514}, |
||||||
|
{0x1D516, 0x1D51C}, {0x1D51E, 0x1D539}, {0x1D53B, 0x1D53E}, |
||||||
|
{0x1D540, 0x1D544}, {0x1D546, 0x1D546}, {0x1D54A, 0x1D550}, |
||||||
|
{0x1D552, 0x1D6A5}, {0x1D6A8, 0x1D7CB}, {0x1D7CE, 0x1DA8B}, |
||||||
|
{0x1DA9B, 0x1DA9F}, {0x1DAA1, 0x1DAAF}, {0x1E000, 0x1E006}, |
||||||
|
{0x1E008, 0x1E018}, {0x1E01B, 0x1E021}, {0x1E023, 0x1E024}, |
||||||
|
{0x1E026, 0x1E02A}, {0x1E100, 0x1E12C}, {0x1E130, 0x1E13D}, |
||||||
|
{0x1E140, 0x1E149}, {0x1E14E, 0x1E14F}, {0x1E2C0, 0x1E2F9}, |
||||||
|
{0x1E2FF, 0x1E2FF}, {0x1E800, 0x1E8C4}, {0x1E8C7, 0x1E8D6}, |
||||||
|
{0x1E900, 0x1E94B}, {0x1E950, 0x1E959}, {0x1E95E, 0x1E95F}, |
||||||
|
{0x1EC71, 0x1ECB4}, {0x1ED01, 0x1ED3D}, {0x1EE00, 0x1EE03}, |
||||||
|
{0x1EE05, 0x1EE1F}, {0x1EE21, 0x1EE22}, {0x1EE24, 0x1EE24}, |
||||||
|
{0x1EE27, 0x1EE27}, {0x1EE29, 0x1EE32}, {0x1EE34, 0x1EE37}, |
||||||
|
{0x1EE39, 0x1EE39}, {0x1EE3B, 0x1EE3B}, {0x1EE42, 0x1EE42}, |
||||||
|
{0x1EE47, 0x1EE47}, {0x1EE49, 0x1EE49}, {0x1EE4B, 0x1EE4B}, |
||||||
|
{0x1EE4D, 0x1EE4F}, {0x1EE51, 0x1EE52}, {0x1EE54, 0x1EE54}, |
||||||
|
{0x1EE57, 0x1EE57}, {0x1EE59, 0x1EE59}, {0x1EE5B, 0x1EE5B}, |
||||||
|
{0x1EE5D, 0x1EE5D}, {0x1EE5F, 0x1EE5F}, {0x1EE61, 0x1EE62}, |
||||||
|
{0x1EE64, 0x1EE64}, {0x1EE67, 0x1EE6A}, {0x1EE6C, 0x1EE72}, |
||||||
|
{0x1EE74, 0x1EE77}, {0x1EE79, 0x1EE7C}, {0x1EE7E, 0x1EE7E}, |
||||||
|
{0x1EE80, 0x1EE89}, {0x1EE8B, 0x1EE9B}, {0x1EEA1, 0x1EEA3}, |
||||||
|
{0x1EEA5, 0x1EEA9}, {0x1EEAB, 0x1EEBB}, {0x1EEF0, 0x1EEF1}, |
||||||
|
{0x1F000, 0x1F003}, {0x1F005, 0x1F02B}, {0x1F030, 0x1F093}, |
||||||
|
{0x1F0A0, 0x1F0AE}, {0x1F0B1, 0x1F0BF}, {0x1F0C1, 0x1F0CE}, |
||||||
|
{0x1F0D1, 0x1F0F5}, {0x1F10B, 0x1F10C}, {0x1F12E, 0x1F12F}, |
||||||
|
{0x1F16A, 0x1F16C}, {0x1F1E6, 0x1F1FF}, {0x1F321, 0x1F32C}, |
||||||
|
{0x1F336, 0x1F336}, {0x1F37D, 0x1F37D}, {0x1F394, 0x1F39F}, |
||||||
|
{0x1F3CB, 0x1F3CE}, {0x1F3D4, 0x1F3DF}, {0x1F3F1, 0x1F3F3}, |
||||||
|
{0x1F3F5, 0x1F3F7}, {0x1F43F, 0x1F43F}, {0x1F441, 0x1F441}, |
||||||
|
{0x1F4FD, 0x1F4FE}, {0x1F53E, 0x1F54A}, {0x1F54F, 0x1F54F}, |
||||||
|
{0x1F568, 0x1F579}, {0x1F57B, 0x1F594}, {0x1F597, 0x1F5A3}, |
||||||
|
{0x1F5A5, 0x1F5FA}, {0x1F650, 0x1F67F}, {0x1F6C6, 0x1F6CB}, |
||||||
|
{0x1F6CD, 0x1F6CF}, {0x1F6D3, 0x1F6D4}, {0x1F6E0, 0x1F6EA}, |
||||||
|
{0x1F6F0, 0x1F6F3}, {0x1F700, 0x1F773}, {0x1F780, 0x1F7D8}, |
||||||
|
{0x1F800, 0x1F80B}, {0x1F810, 0x1F847}, {0x1F850, 0x1F859}, |
||||||
|
{0x1F860, 0x1F887}, {0x1F890, 0x1F8AD}, {0x1F900, 0x1F90B}, |
||||||
|
{0x1FA00, 0x1FA53}, {0x1FA60, 0x1FA6D}, {0xE0001, 0xE0001}, |
||||||
|
{0xE0020, 0xE007F}, |
||||||
|
} |
||||||
|
|
||||||
|
var emoji = table{ |
||||||
|
{0x203C, 0x203C}, {0x2049, 0x2049}, {0x2122, 0x2122}, |
||||||
|
{0x2139, 0x2139}, {0x2194, 0x2199}, {0x21A9, 0x21AA}, |
||||||
|
{0x231A, 0x231B}, {0x2328, 0x2328}, {0x2388, 0x2388}, |
||||||
|
{0x23CF, 0x23CF}, {0x23E9, 0x23F3}, {0x23F8, 0x23FA}, |
||||||
|
{0x24C2, 0x24C2}, {0x25AA, 0x25AB}, {0x25B6, 0x25B6}, |
||||||
|
{0x25C0, 0x25C0}, {0x25FB, 0x25FE}, {0x2600, 0x2605}, |
||||||
|
{0x2607, 0x2612}, {0x2614, 0x2685}, {0x2690, 0x2705}, |
||||||
|
{0x2708, 0x2712}, {0x2714, 0x2714}, {0x2716, 0x2716}, |
||||||
|
{0x271D, 0x271D}, {0x2721, 0x2721}, {0x2728, 0x2728}, |
||||||
|
{0x2733, 0x2734}, {0x2744, 0x2744}, {0x2747, 0x2747}, |
||||||
|
{0x274C, 0x274C}, {0x274E, 0x274E}, {0x2753, 0x2755}, |
||||||
|
{0x2757, 0x2757}, {0x2763, 0x2767}, {0x2795, 0x2797}, |
||||||
|
{0x27A1, 0x27A1}, {0x27B0, 0x27B0}, {0x27BF, 0x27BF}, |
||||||
|
{0x2934, 0x2935}, {0x2B05, 0x2B07}, {0x2B1B, 0x2B1C}, |
||||||
|
{0x2B50, 0x2B50}, {0x2B55, 0x2B55}, {0x3030, 0x3030}, |
||||||
|
{0x303D, 0x303D}, {0x3297, 0x3297}, {0x3299, 0x3299}, |
||||||
|
{0x1F000, 0x1F0FF}, {0x1F10D, 0x1F10F}, {0x1F12F, 0x1F12F}, |
||||||
|
{0x1F16C, 0x1F171}, {0x1F17E, 0x1F17F}, {0x1F18E, 0x1F18E}, |
||||||
|
{0x1F191, 0x1F19A}, {0x1F1AD, 0x1F1E5}, {0x1F201, 0x1F20F}, |
||||||
|
{0x1F21A, 0x1F21A}, {0x1F22F, 0x1F22F}, {0x1F232, 0x1F23A}, |
||||||
|
{0x1F23C, 0x1F23F}, {0x1F249, 0x1F3FA}, {0x1F400, 0x1F53D}, |
||||||
|
{0x1F546, 0x1F64F}, {0x1F680, 0x1F6FF}, {0x1F774, 0x1F77F}, |
||||||
|
{0x1F7D5, 0x1F7FF}, {0x1F80C, 0x1F80F}, {0x1F848, 0x1F84F}, |
||||||
|
{0x1F85A, 0x1F85F}, {0x1F888, 0x1F88F}, {0x1F8AE, 0x1F8FF}, |
||||||
|
{0x1F90C, 0x1F93A}, {0x1F93C, 0x1F945}, {0x1F947, 0x1FFFD}, |
||||||
|
} |
@ -0,0 +1,28 @@ |
|||||||
|
// +build windows
|
||||||
|
// +build !appengine
|
||||||
|
|
||||||
|
package runewidth |
||||||
|
|
||||||
|
import ( |
||||||
|
"syscall" |
||||||
|
) |
||||||
|
|
||||||
|
var ( |
||||||
|
kernel32 = syscall.NewLazyDLL("kernel32") |
||||||
|
procGetConsoleOutputCP = kernel32.NewProc("GetConsoleOutputCP") |
||||||
|
) |
||||||
|
|
||||||
|
// IsEastAsian return true if the current locale is CJK
|
||||||
|
func IsEastAsian() bool { |
||||||
|
r1, _, _ := procGetConsoleOutputCP.Call() |
||||||
|
if r1 == 0 { |
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
switch int(r1) { |
||||||
|
case 932, 51932, 936, 949, 950: |
||||||
|
return true |
||||||
|
} |
||||||
|
|
||||||
|
return false |
||||||
|
} |
@ -0,0 +1,2 @@ |
|||||||
|
language: go |
||||||
|
go: master |
@ -0,0 +1,21 @@ |
|||||||
|
MIT License |
||||||
|
|
||||||
|
Copyright (c) 2018 Minko Gechev |
||||||
|
|
||||||
|
Permission is hereby granted, free of charge, to any person obtaining a copy |
||||||
|
of this software and associated documentation files (the "Software"), to deal |
||||||
|
in the Software without restriction, including without limitation the rights |
||||||
|
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell |
||||||
|
copies of the Software, and to permit persons to whom the Software is |
||||||
|
furnished to do so, subject to the following conditions: |
||||||
|
|
||||||
|
The above copyright notice and this permission notice shall be included in all |
||||||
|
copies or substantial portions of the Software. |
||||||
|
|
||||||
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
||||||
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
||||||
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE |
||||||
|
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
||||||
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, |
||||||
|
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
||||||
|
SOFTWARE. |
@ -0,0 +1,456 @@ |
|||||||
|
package dots |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/build" |
||||||
|
"log" |
||||||
|
"os" |
||||||
|
"path" |
||||||
|
"path/filepath" |
||||||
|
"regexp" |
||||||
|
"runtime" |
||||||
|
"strings" |
||||||
|
) |
||||||
|
|
||||||
|
var ( |
||||||
|
buildContext = build.Default |
||||||
|
goroot = filepath.Clean(runtime.GOROOT()) |
||||||
|
gorootSrc = filepath.Join(goroot, "src") |
||||||
|
) |
||||||
|
|
||||||
|
func flatten(arr [][]string) []string { |
||||||
|
var res []string |
||||||
|
for _, e := range arr { |
||||||
|
res = append(res, e...) |
||||||
|
} |
||||||
|
return res |
||||||
|
} |
||||||
|
|
||||||
|
// Resolve accepts a slice of paths with optional "..." placeholder and a slice with paths to be skipped.
|
||||||
|
// The final result is the set of all files from the selected directories subtracted with
|
||||||
|
// the files in the skip slice.
|
||||||
|
func Resolve(includePatterns, skipPatterns []string) ([]string, error) { |
||||||
|
skip, err := resolvePatterns(skipPatterns) |
||||||
|
filter := newPathFilter(flatten(skip)) |
||||||
|
if err != nil { |
||||||
|
return nil, err |
||||||
|
} |
||||||
|
|
||||||
|
pathSet := map[string]bool{} |
||||||
|
includePackages, err := resolvePatterns(includePatterns) |
||||||
|
include := flatten(includePackages) |
||||||
|
if err != nil { |
||||||
|
return nil, err |
||||||
|
} |
||||||
|
|
||||||
|
var result []string |
||||||
|
for _, i := range include { |
||||||
|
if _, ok := pathSet[i]; !ok && !filter(i) { |
||||||
|
pathSet[i] = true |
||||||
|
result = append(result, i) |
||||||
|
} |
||||||
|
} |
||||||
|
return result, err |
||||||
|
} |
||||||
|
|
||||||
|
// ResolvePackages accepts a slice of paths with optional "..." placeholder and a slice with paths to be skipped.
|
||||||
|
// The final result is the set of all files from the selected directories subtracted with
|
||||||
|
// the files in the skip slice. The difference between `Resolve` and `ResolvePackages`
|
||||||
|
// is that `ResolvePackages` preserves the package structure in the nested slices.
|
||||||
|
func ResolvePackages(includePatterns, skipPatterns []string) ([][]string, error) { |
||||||
|
skip, err := resolvePatterns(skipPatterns) |
||||||
|
filter := newPathFilter(flatten(skip)) |
||||||
|
if err != nil { |
||||||
|
return nil, err |
||||||
|
} |
||||||
|
|
||||||
|
pathSet := map[string]bool{} |
||||||
|
include, err := resolvePatterns(includePatterns) |
||||||
|
if err != nil { |
||||||
|
return nil, err |
||||||
|
} |
||||||
|
|
||||||
|
var result [][]string |
||||||
|
for _, p := range include { |
||||||
|
var packageFiles []string |
||||||
|
for _, f := range p { |
||||||
|
if _, ok := pathSet[f]; !ok && !filter(f) { |
||||||
|
pathSet[f] = true |
||||||
|
packageFiles = append(packageFiles, f) |
||||||
|
} |
||||||
|
} |
||||||
|
result = append(result, packageFiles) |
||||||
|
} |
||||||
|
return result, err |
||||||
|
} |
||||||
|
|
||||||
|
func isDir(filename string) bool { |
||||||
|
fi, err := os.Stat(filename) |
||||||
|
return err == nil && fi.IsDir() |
||||||
|
} |
||||||
|
|
||||||
|
func exists(filename string) bool { |
||||||
|
_, err := os.Stat(filename) |
||||||
|
return err == nil |
||||||
|
} |
||||||
|
|
||||||
|
func resolveDir(dirname string) ([]string, error) { |
||||||
|
pkg, err := build.ImportDir(dirname, 0) |
||||||
|
return resolveImportedPackage(pkg, err) |
||||||
|
} |
||||||
|
|
||||||
|
func resolvePackage(pkgname string) ([]string, error) { |
||||||
|
pkg, err := build.Import(pkgname, ".", 0) |
||||||
|
return resolveImportedPackage(pkg, err) |
||||||
|
} |
||||||
|
|
||||||
|
func resolveImportedPackage(pkg *build.Package, err error) ([]string, error) { |
||||||
|
if err != nil { |
||||||
|
if _, nogo := err.(*build.NoGoError); nogo { |
||||||
|
// Don't complain if the failure is due to no Go source files.
|
||||||
|
return nil, nil |
||||||
|
} |
||||||
|
return nil, err |
||||||
|
} |
||||||
|
|
||||||
|
var files []string |
||||||
|
files = append(files, pkg.GoFiles...) |
||||||
|
files = append(files, pkg.CgoFiles...) |
||||||
|
files = append(files, pkg.TestGoFiles...) |
||||||
|
if pkg.Dir != "." { |
||||||
|
for i, f := range files { |
||||||
|
files[i] = filepath.Join(pkg.Dir, f) |
||||||
|
} |
||||||
|
} |
||||||
|
return files, nil |
||||||
|
} |
||||||
|
|
||||||
|
func resolvePatterns(patterns []string) ([][]string, error) { |
||||||
|
var files [][]string |
||||||
|
for _, pattern := range patterns { |
||||||
|
f, err := resolvePattern(pattern) |
||||||
|
if err != nil { |
||||||
|
return nil, err |
||||||
|
} |
||||||
|
files = append(files, f...) |
||||||
|
} |
||||||
|
return files, nil |
||||||
|
} |
||||||
|
|
||||||
|
func resolvePattern(pattern string) ([][]string, error) { |
||||||
|
// dirsRun, filesRun, and pkgsRun indicate whether golint is applied to
|
||||||
|
// directory, file or package targets. The distinction affects which
|
||||||
|
// checks are run. It is no valid to mix target types.
|
||||||
|
var dirsRun, filesRun, pkgsRun int |
||||||
|
var matches []string |
||||||
|
|
||||||
|
if strings.HasSuffix(pattern, "/...") && isDir(pattern[:len(pattern)-len("/...")]) { |
||||||
|
dirsRun = 1 |
||||||
|
for _, dirname := range matchPackagesInFS(pattern) { |
||||||
|
matches = append(matches, dirname) |
||||||
|
} |
||||||
|
} else if isDir(pattern) { |
||||||
|
dirsRun = 1 |
||||||
|
matches = append(matches, pattern) |
||||||
|
} else if exists(pattern) { |
||||||
|
filesRun = 1 |
||||||
|
matches = append(matches, pattern) |
||||||
|
} else { |
||||||
|
pkgsRun = 1 |
||||||
|
matches = append(matches, pattern) |
||||||
|
} |
||||||
|
|
||||||
|
result := [][]string{} |
||||||
|
switch { |
||||||
|
case dirsRun == 1: |
||||||
|
for _, dir := range matches { |
||||||
|
res, err := resolveDir(dir) |
||||||
|
if err != nil { |
||||||
|
return nil, err |
||||||
|
} |
||||||
|
result = append(result, res) |
||||||
|
} |
||||||
|
case filesRun == 1: |
||||||
|
return [][]string{matches}, nil |
||||||
|
case pkgsRun == 1: |
||||||
|
for _, pkg := range importPaths(matches) { |
||||||
|
res, err := resolvePackage(pkg) |
||||||
|
if err != nil { |
||||||
|
return nil, err |
||||||
|
} |
||||||
|
result = append(result, res) |
||||||
|
} |
||||||
|
} |
||||||
|
return result, nil |
||||||
|
} |
||||||
|
|
||||||
|
func newPathFilter(skip []string) func(string) bool { |
||||||
|
filter := map[string]bool{} |
||||||
|
for _, name := range skip { |
||||||
|
filter[name] = true |
||||||
|
} |
||||||
|
|
||||||
|
return func(path string) bool { |
||||||
|
base := filepath.Base(path) |
||||||
|
if filter[base] || filter[path] { |
||||||
|
return true |
||||||
|
} |
||||||
|
return base != "." && base != ".." && strings.ContainsAny(base[0:1], "_.") |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// importPathsNoDotExpansion returns the import paths to use for the given
|
||||||
|
// command line, but it does no ... expansion.
|
||||||
|
func importPathsNoDotExpansion(args []string) []string { |
||||||
|
if len(args) == 0 { |
||||||
|
return []string{"."} |
||||||
|
} |
||||||
|
var out []string |
||||||
|
for _, a := range args { |
||||||
|
// Arguments are supposed to be import paths, but
|
||||||
|
// as a courtesy to Windows developers, rewrite \ to /
|
||||||
|
// in command-line arguments. Handles .\... and so on.
|
||||||
|
if filepath.Separator == '\\' { |
||||||
|
a = strings.Replace(a, `\`, `/`, -1) |
||||||
|
} |
||||||
|
|
||||||
|
// Put argument in canonical form, but preserve leading ./.
|
||||||
|
if strings.HasPrefix(a, "./") { |
||||||
|
a = "./" + path.Clean(a) |
||||||
|
if a == "./." { |
||||||
|
a = "." |
||||||
|
} |
||||||
|
} else { |
||||||
|
a = path.Clean(a) |
||||||
|
} |
||||||
|
if a == "all" || a == "std" { |
||||||
|
out = append(out, matchPackages(a)...) |
||||||
|
continue |
||||||
|
} |
||||||
|
out = append(out, a) |
||||||
|
} |
||||||
|
return out |
||||||
|
} |
||||||
|
|
||||||
|
// importPaths returns the import paths to use for the given command line.
|
||||||
|
func importPaths(args []string) []string { |
||||||
|
args = importPathsNoDotExpansion(args) |
||||||
|
var out []string |
||||||
|
for _, a := range args { |
||||||
|
if strings.Contains(a, "...") { |
||||||
|
if build.IsLocalImport(a) { |
||||||
|
out = append(out, matchPackagesInFS(a)...) |
||||||
|
} else { |
||||||
|
out = append(out, matchPackages(a)...) |
||||||
|
} |
||||||
|
continue |
||||||
|
} |
||||||
|
out = append(out, a) |
||||||
|
} |
||||||
|
return out |
||||||
|
} |
||||||
|
|
||||||
|
// matchPattern(pattern)(name) reports whether
|
||||||
|
// name matches pattern. Pattern is a limited glob
|
||||||
|
// pattern in which '...' means 'any string' and there
|
||||||
|
// is no other special syntax.
|
||||||
|
func matchPattern(pattern string) func(name string) bool { |
||||||
|
re := regexp.QuoteMeta(pattern) |
||||||
|
re = strings.Replace(re, `\.\.\.`, `.*`, -1) |
||||||
|
// Special case: foo/... matches foo too.
|
||||||
|
if strings.HasSuffix(re, `/.*`) { |
||||||
|
re = re[:len(re)-len(`/.*`)] + `(/.*)?` |
||||||
|
} |
||||||
|
reg := regexp.MustCompile(`^` + re + `$`) |
||||||
|
return func(name string) bool { |
||||||
|
return reg.MatchString(name) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// hasPathPrefix reports whether the path s begins with the
|
||||||
|
// elements in prefix.
|
||||||
|
func hasPathPrefix(s, prefix string) bool { |
||||||
|
switch { |
||||||
|
default: |
||||||
|
return false |
||||||
|
case len(s) == len(prefix): |
||||||
|
return s == prefix |
||||||
|
case len(s) > len(prefix): |
||||||
|
if prefix != "" && prefix[len(prefix)-1] == '/' { |
||||||
|
return strings.HasPrefix(s, prefix) |
||||||
|
} |
||||||
|
return s[len(prefix)] == '/' && s[:len(prefix)] == prefix |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// treeCanMatchPattern(pattern)(name) reports whether
|
||||||
|
// name or children of name can possibly match pattern.
|
||||||
|
// Pattern is the same limited glob accepted by matchPattern.
|
||||||
|
func treeCanMatchPattern(pattern string) func(name string) bool { |
||||||
|
wildCard := false |
||||||
|
if i := strings.Index(pattern, "..."); i >= 0 { |
||||||
|
wildCard = true |
||||||
|
pattern = pattern[:i] |
||||||
|
} |
||||||
|
return func(name string) bool { |
||||||
|
return len(name) <= len(pattern) && hasPathPrefix(pattern, name) || |
||||||
|
wildCard && strings.HasPrefix(name, pattern) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
func matchPackages(pattern string) []string { |
||||||
|
match := func(string) bool { return true } |
||||||
|
treeCanMatch := func(string) bool { return true } |
||||||
|
if pattern != "all" && pattern != "std" { |
||||||
|
match = matchPattern(pattern) |
||||||
|
treeCanMatch = treeCanMatchPattern(pattern) |
||||||
|
} |
||||||
|
|
||||||
|
have := map[string]bool{ |
||||||
|
"builtin": true, // ignore pseudo-package that exists only for documentation
|
||||||
|
} |
||||||
|
if !buildContext.CgoEnabled { |
||||||
|
have["runtime/cgo"] = true // ignore during walk
|
||||||
|
} |
||||||
|
var pkgs []string |
||||||
|
|
||||||
|
// Commands
|
||||||
|
cmd := filepath.Join(goroot, "src/cmd") + string(filepath.Separator) |
||||||
|
filepath.Walk(cmd, func(path string, fi os.FileInfo, err error) error { |
||||||
|
if err != nil || !fi.IsDir() || path == cmd { |
||||||
|
return nil |
||||||
|
} |
||||||
|
name := path[len(cmd):] |
||||||
|
if !treeCanMatch(name) { |
||||||
|
return filepath.SkipDir |
||||||
|
} |
||||||
|
// Commands are all in cmd/, not in subdirectories.
|
||||||
|
if strings.Contains(name, string(filepath.Separator)) { |
||||||
|
return filepath.SkipDir |
||||||
|
} |
||||||
|
|
||||||
|
// We use, e.g., cmd/gofmt as the pseudo import path for gofmt.
|
||||||
|
name = "cmd/" + name |
||||||
|
if have[name] { |
||||||
|
return nil |
||||||
|
} |
||||||
|
have[name] = true |
||||||
|
if !match(name) { |
||||||
|
return nil |
||||||
|
} |
||||||
|
_, err = buildContext.ImportDir(path, 0) |
||||||
|
if err != nil { |
||||||
|
if _, noGo := err.(*build.NoGoError); !noGo { |
||||||
|
log.Print(err) |
||||||
|
} |
||||||
|
return nil |
||||||
|
} |
||||||
|
pkgs = append(pkgs, name) |
||||||
|
return nil |
||||||
|
}) |
||||||
|
|
||||||
|
for _, src := range buildContext.SrcDirs() { |
||||||
|
if (pattern == "std" || pattern == "cmd") && src != gorootSrc { |
||||||
|
continue |
||||||
|
} |
||||||
|
src = filepath.Clean(src) + string(filepath.Separator) |
||||||
|
root := src |
||||||
|
if pattern == "cmd" { |
||||||
|
root += "cmd" + string(filepath.Separator) |
||||||
|
} |
||||||
|
filepath.Walk(root, func(path string, fi os.FileInfo, err error) error { |
||||||
|
if err != nil || !fi.IsDir() || path == src { |
||||||
|
return nil |
||||||
|
} |
||||||
|
|
||||||
|
// Avoid .foo, _foo, and testdata directory trees.
|
||||||
|
_, elem := filepath.Split(path) |
||||||
|
if strings.HasPrefix(elem, ".") || strings.HasPrefix(elem, "_") || elem == "testdata" { |
||||||
|
return filepath.SkipDir |
||||||
|
} |
||||||
|
|
||||||
|
name := filepath.ToSlash(path[len(src):]) |
||||||
|
if pattern == "std" && (strings.Contains(name, ".") || name == "cmd") { |
||||||
|
// The name "std" is only the standard library.
|
||||||
|
// If the name is cmd, it's the root of the command tree.
|
||||||
|
return filepath.SkipDir |
||||||
|
} |
||||||
|
if !treeCanMatch(name) { |
||||||
|
return filepath.SkipDir |
||||||
|
} |
||||||
|
if have[name] { |
||||||
|
return nil |
||||||
|
} |
||||||
|
have[name] = true |
||||||
|
if !match(name) { |
||||||
|
return nil |
||||||
|
} |
||||||
|
_, err = buildContext.ImportDir(path, 0) |
||||||
|
if err != nil { |
||||||
|
if _, noGo := err.(*build.NoGoError); noGo { |
||||||
|
return nil |
||||||
|
} |
||||||
|
} |
||||||
|
pkgs = append(pkgs, name) |
||||||
|
return nil |
||||||
|
}) |
||||||
|
} |
||||||
|
return pkgs |
||||||
|
} |
||||||
|
|
||||||
|
func matchPackagesInFS(pattern string) []string { |
||||||
|
// Find directory to begin the scan.
|
||||||
|
// Could be smarter but this one optimization
|
||||||
|
// is enough for now, since ... is usually at the
|
||||||
|
// end of a path.
|
||||||
|
i := strings.Index(pattern, "...") |
||||||
|
dir, _ := path.Split(pattern[:i]) |
||||||
|
|
||||||
|
// pattern begins with ./ or ../.
|
||||||
|
// path.Clean will discard the ./ but not the ../.
|
||||||
|
// We need to preserve the ./ for pattern matching
|
||||||
|
// and in the returned import paths.
|
||||||
|
prefix := "" |
||||||
|
if strings.HasPrefix(pattern, "./") { |
||||||
|
prefix = "./" |
||||||
|
} |
||||||
|
match := matchPattern(pattern) |
||||||
|
|
||||||
|
var pkgs []string |
||||||
|
filepath.Walk(dir, func(path string, fi os.FileInfo, err error) error { |
||||||
|
if err != nil || !fi.IsDir() { |
||||||
|
return nil |
||||||
|
} |
||||||
|
if path == dir { |
||||||
|
// filepath.Walk starts at dir and recurses. For the recursive case,
|
||||||
|
// the path is the result of filepath.Join, which calls filepath.Clean.
|
||||||
|
// The initial case is not Cleaned, though, so we do this explicitly.
|
||||||
|
//
|
||||||
|
// This converts a path like "./io/" to "io". Without this step, running
|
||||||
|
// "cd $GOROOT/src/pkg; go list ./io/..." would incorrectly skip the io
|
||||||
|
// package, because prepending the prefix "./" to the unclean path would
|
||||||
|
// result in "././io", and match("././io") returns false.
|
||||||
|
path = filepath.Clean(path) |
||||||
|
} |
||||||
|
|
||||||
|
// Avoid .foo, _foo, and testdata directory trees, but do not avoid "." or "..".
|
||||||
|
_, elem := filepath.Split(path) |
||||||
|
dot := strings.HasPrefix(elem, ".") && elem != "." && elem != ".." |
||||||
|
if dot || strings.HasPrefix(elem, "_") || elem == "testdata" { |
||||||
|
return filepath.SkipDir |
||||||
|
} |
||||||
|
|
||||||
|
name := prefix + filepath.ToSlash(path) |
||||||
|
if !match(name) { |
||||||
|
return nil |
||||||
|
} |
||||||
|
if _, err = build.ImportDir(path, 0); err != nil { |
||||||
|
if _, noGo := err.(*build.NoGoError); !noGo { |
||||||
|
log.Print(err) |
||||||
|
} |
||||||
|
return nil |
||||||
|
} |
||||||
|
pkgs = append(pkgs, name) |
||||||
|
return nil |
||||||
|
}) |
||||||
|
return pkgs |
||||||
|
} |
@ -0,0 +1,21 @@ |
|||||||
|
MIT License |
||||||
|
|
||||||
|
Copyright (c) 2018 Minko Gechev |
||||||
|
|
||||||
|
Permission is hereby granted, free of charge, to any person obtaining a copy |
||||||
|
of this software and associated documentation files (the "Software"), to deal |
||||||
|
in the Software without restriction, including without limitation the rights |
||||||
|
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell |
||||||
|
copies of the Software, and to permit persons to whom the Software is |
||||||
|
furnished to do so, subject to the following conditions: |
||||||
|
|
||||||
|
The above copyright notice and this permission notice shall be included in all |
||||||
|
copies or substantial portions of the Software. |
||||||
|
|
||||||
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
||||||
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
||||||
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE |
||||||
|
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
||||||
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, |
||||||
|
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
||||||
|
SOFTWARE. |
@ -0,0 +1,76 @@ |
|||||||
|
package formatter |
||||||
|
|
||||||
|
import ( |
||||||
|
"bytes" |
||||||
|
"encoding/xml" |
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
plainTemplate "text/template" |
||||||
|
) |
||||||
|
|
||||||
|
// Checkstyle is an implementation of the Formatter interface
|
||||||
|
// which formats the errors to Checkstyle-like format.
|
||||||
|
type Checkstyle struct { |
||||||
|
Metadata lint.FormatterMetadata |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the name of the formatter
|
||||||
|
func (f *Checkstyle) Name() string { |
||||||
|
return "checkstyle" |
||||||
|
} |
||||||
|
|
||||||
|
type issue struct { |
||||||
|
Line int |
||||||
|
Col int |
||||||
|
What string |
||||||
|
Confidence float64 |
||||||
|
Severity lint.Severity |
||||||
|
RuleName string |
||||||
|
} |
||||||
|
|
||||||
|
// Format formats the failures gotten from the lint.
|
||||||
|
func (f *Checkstyle) Format(failures <-chan lint.Failure, config lint.Config) (string, error) { |
||||||
|
var issues = map[string][]issue{} |
||||||
|
for failure := range failures { |
||||||
|
buf := new(bytes.Buffer) |
||||||
|
xml.Escape(buf, []byte(failure.Failure)) |
||||||
|
what := buf.String() |
||||||
|
iss := issue{ |
||||||
|
Line: failure.Position.Start.Line, |
||||||
|
Col: failure.Position.Start.Column, |
||||||
|
What: what, |
||||||
|
Confidence: failure.Confidence, |
||||||
|
Severity: severity(config, failure), |
||||||
|
RuleName: failure.RuleName, |
||||||
|
} |
||||||
|
fn := failure.GetFilename() |
||||||
|
if issues[fn] == nil { |
||||||
|
issues[fn] = make([]issue, 0) |
||||||
|
} |
||||||
|
issues[fn] = append(issues[fn], iss) |
||||||
|
} |
||||||
|
|
||||||
|
t, err := plainTemplate.New("revive").Parse(checkstyleTemplate) |
||||||
|
if err != nil { |
||||||
|
return "", err |
||||||
|
} |
||||||
|
|
||||||
|
buf := new(bytes.Buffer) |
||||||
|
|
||||||
|
err = t.Execute(buf, issues) |
||||||
|
if err != nil { |
||||||
|
return "", err |
||||||
|
} |
||||||
|
|
||||||
|
return buf.String(), nil |
||||||
|
} |
||||||
|
|
||||||
|
const checkstyleTemplate = `<?xml version='1.0' encoding='UTF-8'?> |
||||||
|
<checkstyle version="5.0"> |
||||||
|
{{- range $k, $v := . }} |
||||||
|
<file name="{{ $k }}"> |
||||||
|
{{- range $i, $issue := $v }} |
||||||
|
<error line="{{ $issue.Line }}" column="{{ $issue.Col }}" message="{{ $issue.What }} (confidence {{ $issue.Confidence}})" severity="{{ $issue.Severity }}" source="revive/{{ $issue.RuleName }}"/> |
||||||
|
{{- end }} |
||||||
|
</file> |
||||||
|
{{- end }} |
||||||
|
</checkstyle>` |
@ -0,0 +1,26 @@ |
|||||||
|
package formatter |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// Default is an implementation of the Formatter interface
|
||||||
|
// which formats the errors to text.
|
||||||
|
type Default struct { |
||||||
|
Metadata lint.FormatterMetadata |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the name of the formatter
|
||||||
|
func (f *Default) Name() string { |
||||||
|
return "default" |
||||||
|
} |
||||||
|
|
||||||
|
// Format formats the failures gotten from the lint.
|
||||||
|
func (f *Default) Format(failures <-chan lint.Failure, _ lint.Config) (string, error) { |
||||||
|
for failure := range failures { |
||||||
|
fmt.Printf("%v: %s\n", failure.Position.Start, failure.Failure) |
||||||
|
} |
||||||
|
return "", nil |
||||||
|
} |
@ -0,0 +1,146 @@ |
|||||||
|
package formatter |
||||||
|
|
||||||
|
import ( |
||||||
|
"bytes" |
||||||
|
"fmt" |
||||||
|
"sort" |
||||||
|
|
||||||
|
"github.com/fatih/color" |
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
"github.com/olekukonko/tablewriter" |
||||||
|
) |
||||||
|
|
||||||
|
var ( |
||||||
|
errorEmoji = color.RedString("✘") |
||||||
|
warningEmoji = color.YellowString("⚠") |
||||||
|
) |
||||||
|
|
||||||
|
var newLines = map[rune]bool{ |
||||||
|
0x000A: true, |
||||||
|
0x000B: true, |
||||||
|
0x000C: true, |
||||||
|
0x000D: true, |
||||||
|
0x0085: true, |
||||||
|
0x2028: true, |
||||||
|
0x2029: true, |
||||||
|
} |
||||||
|
|
||||||
|
// Friendly is an implementation of the Formatter interface
|
||||||
|
// which formats the errors to JSON.
|
||||||
|
type Friendly struct { |
||||||
|
Metadata lint.FormatterMetadata |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the name of the formatter
|
||||||
|
func (f *Friendly) Name() string { |
||||||
|
return "friendly" |
||||||
|
} |
||||||
|
|
||||||
|
// Format formats the failures gotten from the lint.
|
||||||
|
func (f *Friendly) Format(failures <-chan lint.Failure, config lint.Config) (string, error) { |
||||||
|
errorMap := map[string]int{} |
||||||
|
warningMap := map[string]int{} |
||||||
|
totalErrors := 0 |
||||||
|
totalWarnings := 0 |
||||||
|
for failure := range failures { |
||||||
|
sev := severity(config, failure) |
||||||
|
f.printFriendlyFailure(failure, sev) |
||||||
|
if sev == lint.SeverityWarning { |
||||||
|
warningMap[failure.RuleName] = warningMap[failure.RuleName] + 1 |
||||||
|
totalWarnings++ |
||||||
|
} |
||||||
|
if sev == lint.SeverityError { |
||||||
|
errorMap[failure.RuleName] = errorMap[failure.RuleName] + 1 |
||||||
|
totalErrors++ |
||||||
|
} |
||||||
|
} |
||||||
|
f.printSummary(totalErrors, totalWarnings) |
||||||
|
f.printStatistics(color.RedString("Errors:"), errorMap) |
||||||
|
f.printStatistics(color.YellowString("Warnings:"), warningMap) |
||||||
|
return "", nil |
||||||
|
} |
||||||
|
|
||||||
|
func (f *Friendly) printFriendlyFailure(failure lint.Failure, severity lint.Severity) { |
||||||
|
f.printHeaderRow(failure, severity) |
||||||
|
f.printFilePosition(failure) |
||||||
|
fmt.Println() |
||||||
|
fmt.Println() |
||||||
|
} |
||||||
|
|
||||||
|
func (f *Friendly) printHeaderRow(failure lint.Failure, severity lint.Severity) { |
||||||
|
emoji := warningEmoji |
||||||
|
if severity == lint.SeverityError { |
||||||
|
emoji = errorEmoji |
||||||
|
} |
||||||
|
fmt.Print(f.table([][]string{{emoji, "https://revive.run/r#" + failure.RuleName, color.GreenString(failure.Failure)}})) |
||||||
|
} |
||||||
|
|
||||||
|
func (f *Friendly) printFilePosition(failure lint.Failure) { |
||||||
|
fmt.Printf(" %s:%d:%d", failure.GetFilename(), failure.Position.Start.Line, failure.Position.Start.Column) |
||||||
|
} |
||||||
|
|
||||||
|
type statEntry struct { |
||||||
|
name string |
||||||
|
failures int |
||||||
|
} |
||||||
|
|
||||||
|
func (f *Friendly) printSummary(errors, warnings int) { |
||||||
|
emoji := warningEmoji |
||||||
|
if errors > 0 { |
||||||
|
emoji = errorEmoji |
||||||
|
} |
||||||
|
problemsLabel := "problems" |
||||||
|
if errors+warnings == 1 { |
||||||
|
problemsLabel = "problem" |
||||||
|
} |
||||||
|
warningsLabel := "warnings" |
||||||
|
if warnings == 1 { |
||||||
|
warningsLabel = "warning" |
||||||
|
} |
||||||
|
errorsLabel := "errors" |
||||||
|
if errors == 1 { |
||||||
|
errorsLabel = "error" |
||||||
|
} |
||||||
|
str := fmt.Sprintf("%d %s (%d %s, %d %s)", errors+warnings, problemsLabel, errors, errorsLabel, warnings, warningsLabel) |
||||||
|
if errors > 0 { |
||||||
|
fmt.Printf("%s %s\n", emoji, color.RedString(str)) |
||||||
|
fmt.Println() |
||||||
|
return |
||||||
|
} |
||||||
|
if warnings > 0 { |
||||||
|
fmt.Printf("%s %s\n", emoji, color.YellowString(str)) |
||||||
|
fmt.Println() |
||||||
|
return |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
func (f *Friendly) printStatistics(header string, stats map[string]int) { |
||||||
|
if len(stats) == 0 { |
||||||
|
return |
||||||
|
} |
||||||
|
var data []statEntry |
||||||
|
for name, total := range stats { |
||||||
|
data = append(data, statEntry{name, total}) |
||||||
|
} |
||||||
|
sort.Slice(data, func(i, j int) bool { |
||||||
|
return data[i].failures > data[j].failures |
||||||
|
}) |
||||||
|
formatted := [][]string{} |
||||||
|
for _, entry := range data { |
||||||
|
formatted = append(formatted, []string{color.GreenString(fmt.Sprintf("%d", entry.failures)), entry.name}) |
||||||
|
} |
||||||
|
fmt.Println(header) |
||||||
|
fmt.Println(f.table(formatted)) |
||||||
|
} |
||||||
|
|
||||||
|
func (f *Friendly) table(rows [][]string) string { |
||||||
|
buf := new(bytes.Buffer) |
||||||
|
table := tablewriter.NewWriter(buf) |
||||||
|
table.SetBorder(false) |
||||||
|
table.SetColumnSeparator("") |
||||||
|
table.SetRowSeparator("") |
||||||
|
table.SetAutoWrapText(false) |
||||||
|
table.AppendBulk(rows) |
||||||
|
table.Render() |
||||||
|
return buf.String() |
||||||
|
} |
@ -0,0 +1,40 @@ |
|||||||
|
package formatter |
||||||
|
|
||||||
|
import ( |
||||||
|
"encoding/json" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// JSON is an implementation of the Formatter interface
|
||||||
|
// which formats the errors to JSON.
|
||||||
|
type JSON struct { |
||||||
|
Metadata lint.FormatterMetadata |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the name of the formatter
|
||||||
|
func (f *JSON) Name() string { |
||||||
|
return "json" |
||||||
|
} |
||||||
|
|
||||||
|
// jsonObject defines a JSON object of an failure
|
||||||
|
type jsonObject struct { |
||||||
|
Severity lint.Severity |
||||||
|
lint.Failure `json:",inline"` |
||||||
|
} |
||||||
|
|
||||||
|
// Format formats the failures gotten from the lint.
|
||||||
|
func (f *JSON) Format(failures <-chan lint.Failure, config lint.Config) (string, error) { |
||||||
|
var slice []jsonObject |
||||||
|
for failure := range failures { |
||||||
|
obj := jsonObject{} |
||||||
|
obj.Severity = severity(config, failure) |
||||||
|
obj.Failure = failure |
||||||
|
slice = append(slice, obj) |
||||||
|
} |
||||||
|
result, err := json.Marshal(slice) |
||||||
|
if err != nil { |
||||||
|
return "", err |
||||||
|
} |
||||||
|
return string(result), err |
||||||
|
} |
@ -0,0 +1,34 @@ |
|||||||
|
package formatter |
||||||
|
|
||||||
|
import ( |
||||||
|
"encoding/json" |
||||||
|
"os" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// NDJSON is an implementation of the Formatter interface
|
||||||
|
// which formats the errors to NDJSON stream.
|
||||||
|
type NDJSON struct { |
||||||
|
Metadata lint.FormatterMetadata |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the name of the formatter
|
||||||
|
func (f *NDJSON) Name() string { |
||||||
|
return "ndjson" |
||||||
|
} |
||||||
|
|
||||||
|
// Format formats the failures gotten from the lint.
|
||||||
|
func (f *NDJSON) Format(failures <-chan lint.Failure, config lint.Config) (string, error) { |
||||||
|
enc := json.NewEncoder(os.Stdout) |
||||||
|
for failure := range failures { |
||||||
|
obj := jsonObject{} |
||||||
|
obj.Severity = severity(config, failure) |
||||||
|
obj.Failure = failure |
||||||
|
err := enc.Encode(obj) |
||||||
|
if err != nil { |
||||||
|
return "", err |
||||||
|
} |
||||||
|
} |
||||||
|
return "", nil |
||||||
|
} |
@ -0,0 +1,26 @@ |
|||||||
|
package formatter |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// Plain is an implementation of the Formatter interface
|
||||||
|
// which formats the errors to JSON.
|
||||||
|
type Plain struct { |
||||||
|
Metadata lint.FormatterMetadata |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the name of the formatter
|
||||||
|
func (f *Plain) Name() string { |
||||||
|
return "plain" |
||||||
|
} |
||||||
|
|
||||||
|
// Format formats the failures gotten from the lint.
|
||||||
|
func (f *Plain) Format(failures <-chan lint.Failure, _ lint.Config) (string, error) { |
||||||
|
for failure := range failures { |
||||||
|
fmt.Printf("%v: %s %s\n", failure.Position.Start, failure.Failure, "https://revive.run/r#"+failure.RuleName) |
||||||
|
} |
||||||
|
return "", nil |
||||||
|
} |
@ -0,0 +1,13 @@ |
|||||||
|
package formatter |
||||||
|
|
||||||
|
import "github.com/mgechev/revive/lint" |
||||||
|
|
||||||
|
func severity(config lint.Config, failure lint.Failure) lint.Severity { |
||||||
|
if config, ok := config.Rules[failure.RuleName]; ok && config.Severity == lint.SeverityError { |
||||||
|
return lint.SeverityError |
||||||
|
} |
||||||
|
if config, ok := config.Directives[failure.RuleName]; ok && config.Severity == lint.SeverityError { |
||||||
|
return lint.SeverityError |
||||||
|
} |
||||||
|
return lint.SeverityWarning |
||||||
|
} |
@ -0,0 +1,89 @@ |
|||||||
|
package formatter |
||||||
|
|
||||||
|
import ( |
||||||
|
"bytes" |
||||||
|
"fmt" |
||||||
|
|
||||||
|
"github.com/fatih/color" |
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
"github.com/olekukonko/tablewriter" |
||||||
|
) |
||||||
|
|
||||||
|
// Stylish is an implementation of the Formatter interface
|
||||||
|
// which formats the errors to JSON.
|
||||||
|
type Stylish struct { |
||||||
|
Metadata lint.FormatterMetadata |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the name of the formatter
|
||||||
|
func (f *Stylish) Name() string { |
||||||
|
return "stylish" |
||||||
|
} |
||||||
|
|
||||||
|
func formatFailure(failure lint.Failure, severity lint.Severity) []string { |
||||||
|
fString := color.CyanString(failure.Failure) |
||||||
|
fName := color.RedString("https://revive.run/r#" + failure.RuleName) |
||||||
|
lineColumn := failure.Position |
||||||
|
pos := fmt.Sprintf("(%d, %d)", lineColumn.Start.Line, lineColumn.Start.Column) |
||||||
|
if severity == lint.SeverityWarning { |
||||||
|
fName = color.YellowString("https://revive.run/r#" + failure.RuleName) |
||||||
|
} |
||||||
|
return []string{failure.GetFilename(), pos, fName, fString} |
||||||
|
} |
||||||
|
|
||||||
|
// Format formats the failures gotten from the lint.
|
||||||
|
func (f *Stylish) Format(failures <-chan lint.Failure, config lint.Config) (string, error) { |
||||||
|
var result [][]string |
||||||
|
var totalErrors = 0 |
||||||
|
var total = 0 |
||||||
|
|
||||||
|
for f := range failures { |
||||||
|
total++ |
||||||
|
currentType := severity(config, f) |
||||||
|
if currentType == lint.SeverityError { |
||||||
|
totalErrors++ |
||||||
|
} |
||||||
|
result = append(result, formatFailure(f, lint.Severity(currentType))) |
||||||
|
} |
||||||
|
ps := "problems" |
||||||
|
if total == 1 { |
||||||
|
ps = "problem" |
||||||
|
} |
||||||
|
|
||||||
|
fileReport := make(map[string][][]string) |
||||||
|
|
||||||
|
for _, row := range result { |
||||||
|
if _, ok := fileReport[row[0]]; !ok { |
||||||
|
fileReport[row[0]] = [][]string{} |
||||||
|
} |
||||||
|
|
||||||
|
fileReport[row[0]] = append(fileReport[row[0]], []string{row[1], row[2], row[3]}) |
||||||
|
} |
||||||
|
|
||||||
|
output := "" |
||||||
|
for filename, val := range fileReport { |
||||||
|
buf := new(bytes.Buffer) |
||||||
|
table := tablewriter.NewWriter(buf) |
||||||
|
table.SetBorder(false) |
||||||
|
table.SetColumnSeparator("") |
||||||
|
table.SetRowSeparator("") |
||||||
|
table.SetAutoWrapText(false) |
||||||
|
table.AppendBulk(val) |
||||||
|
table.Render() |
||||||
|
c := color.New(color.Underline) |
||||||
|
output += c.SprintfFunc()(filename + "\n") |
||||||
|
output += buf.String() + "\n" |
||||||
|
} |
||||||
|
|
||||||
|
suffix := fmt.Sprintf(" %d %s (%d errors) (%d warnings)", total, ps, totalErrors, total-totalErrors) |
||||||
|
|
||||||
|
if total > 0 && totalErrors > 0 { |
||||||
|
suffix = color.RedString("\n ✖" + suffix) |
||||||
|
} else if total > 0 && totalErrors == 0 { |
||||||
|
suffix = color.YellowString("\n ✖" + suffix) |
||||||
|
} else { |
||||||
|
suffix, output = "", "" |
||||||
|
} |
||||||
|
|
||||||
|
return output + suffix, nil |
||||||
|
} |
@ -0,0 +1,27 @@ |
|||||||
|
package formatter |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// Unix is an implementation of the Formatter interface
|
||||||
|
// which formats the errors to a simple line based error format
|
||||||
|
// main.go:24:9: [errorf] should replace errors.New(fmt.Sprintf(...)) with fmt.Errorf(...)
|
||||||
|
type Unix struct { |
||||||
|
Metadata lint.FormatterMetadata |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the name of the formatter
|
||||||
|
func (f *Unix) Name() string { |
||||||
|
return "unix" |
||||||
|
} |
||||||
|
|
||||||
|
// Format formats the failures gotten from the lint.
|
||||||
|
func (f *Unix) Format(failures <-chan lint.Failure, _ lint.Config) (string, error) { |
||||||
|
for failure := range failures { |
||||||
|
fmt.Printf("%v: [%s] %s\n", failure.Position.Start, failure.RuleName, failure.Failure) |
||||||
|
} |
||||||
|
return "", nil |
||||||
|
} |
@ -0,0 +1,32 @@ |
|||||||
|
package lint |
||||||
|
|
||||||
|
// Arguments is type used for the arguments of a rule.
|
||||||
|
type Arguments = []interface{} |
||||||
|
|
||||||
|
// RuleConfig is type used for the rule configuration.
|
||||||
|
type RuleConfig struct { |
||||||
|
Arguments Arguments |
||||||
|
Severity Severity |
||||||
|
} |
||||||
|
|
||||||
|
// RulesConfig defines the config for all rules.
|
||||||
|
type RulesConfig = map[string]RuleConfig |
||||||
|
|
||||||
|
// DirectiveConfig is type used for the linter directive configuration.
|
||||||
|
type DirectiveConfig struct { |
||||||
|
Severity Severity |
||||||
|
} |
||||||
|
|
||||||
|
// DirectivesConfig defines the config for all directives.
|
||||||
|
type DirectivesConfig = map[string]DirectiveConfig |
||||||
|
|
||||||
|
// Config defines the config of the linter.
|
||||||
|
type Config struct { |
||||||
|
IgnoreGeneratedHeader bool `toml:"ignoreGeneratedHeader"` |
||||||
|
Confidence float64 |
||||||
|
Severity Severity |
||||||
|
Rules RulesConfig `toml:"rule"` |
||||||
|
ErrorCode int `toml:"errorCode"` |
||||||
|
WarningCode int `toml:"warningCode"` |
||||||
|
Directives DirectivesConfig `toml:"directive"` |
||||||
|
} |
@ -0,0 +1,39 @@ |
|||||||
|
package lint |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
"go/token" |
||||||
|
) |
||||||
|
|
||||||
|
const ( |
||||||
|
// SeverityWarning declares failures of type warning
|
||||||
|
SeverityWarning = "warning" |
||||||
|
// SeverityError declares failures of type error.
|
||||||
|
SeverityError = "error" |
||||||
|
) |
||||||
|
|
||||||
|
// Severity is the type for the failure types.
|
||||||
|
type Severity string |
||||||
|
|
||||||
|
// FailurePosition returns the failure position
|
||||||
|
type FailurePosition struct { |
||||||
|
Start token.Position |
||||||
|
End token.Position |
||||||
|
} |
||||||
|
|
||||||
|
// Failure defines a struct for a linting failure.
|
||||||
|
type Failure struct { |
||||||
|
Failure string |
||||||
|
RuleName string |
||||||
|
Category string |
||||||
|
Position FailurePosition |
||||||
|
Node ast.Node `json:"-"` |
||||||
|
Confidence float64 |
||||||
|
// For future use
|
||||||
|
ReplacementLine string |
||||||
|
} |
||||||
|
|
||||||
|
// GetFilename returns the filename.
|
||||||
|
func (f *Failure) GetFilename() string { |
||||||
|
return f.Position.Start.Filename |
||||||
|
} |
@ -0,0 +1,278 @@ |
|||||||
|
package lint |
||||||
|
|
||||||
|
import ( |
||||||
|
"bytes" |
||||||
|
"go/ast" |
||||||
|
"go/parser" |
||||||
|
"go/printer" |
||||||
|
"go/token" |
||||||
|
"go/types" |
||||||
|
"math" |
||||||
|
"regexp" |
||||||
|
"strings" |
||||||
|
) |
||||||
|
|
||||||
|
// File abstraction used for representing files.
|
||||||
|
type File struct { |
||||||
|
Name string |
||||||
|
Pkg *Package |
||||||
|
content []byte |
||||||
|
AST *ast.File |
||||||
|
} |
||||||
|
|
||||||
|
// IsTest returns if the file contains tests.
|
||||||
|
func (f *File) IsTest() bool { return strings.HasSuffix(f.Name, "_test.go") } |
||||||
|
|
||||||
|
// Content returns the file's content.
|
||||||
|
func (f *File) Content() []byte { |
||||||
|
return f.content |
||||||
|
} |
||||||
|
|
||||||
|
// NewFile creates a new file
|
||||||
|
func NewFile(name string, content []byte, pkg *Package) (*File, error) { |
||||||
|
f, err := parser.ParseFile(pkg.fset, name, content, parser.ParseComments) |
||||||
|
if err != nil { |
||||||
|
return nil, err |
||||||
|
} |
||||||
|
return &File{ |
||||||
|
Name: name, |
||||||
|
content: content, |
||||||
|
Pkg: pkg, |
||||||
|
AST: f, |
||||||
|
}, nil |
||||||
|
} |
||||||
|
|
||||||
|
// ToPosition returns line and column for given position.
|
||||||
|
func (f *File) ToPosition(pos token.Pos) token.Position { |
||||||
|
return f.Pkg.fset.Position(pos) |
||||||
|
} |
||||||
|
|
||||||
|
// Render renters a node.
|
||||||
|
func (f *File) Render(x interface{}) string { |
||||||
|
var buf bytes.Buffer |
||||||
|
if err := printer.Fprint(&buf, f.Pkg.fset, x); err != nil { |
||||||
|
panic(err) |
||||||
|
} |
||||||
|
return buf.String() |
||||||
|
} |
||||||
|
|
||||||
|
// CommentMap builds a comment map for the file.
|
||||||
|
func (f *File) CommentMap() ast.CommentMap { |
||||||
|
return ast.NewCommentMap(f.Pkg.fset, f.AST, f.AST.Comments) |
||||||
|
} |
||||||
|
|
||||||
|
var basicTypeKinds = map[types.BasicKind]string{ |
||||||
|
types.UntypedBool: "bool", |
||||||
|
types.UntypedInt: "int", |
||||||
|
types.UntypedRune: "rune", |
||||||
|
types.UntypedFloat: "float64", |
||||||
|
types.UntypedComplex: "complex128", |
||||||
|
types.UntypedString: "string", |
||||||
|
} |
||||||
|
|
||||||
|
// IsUntypedConst reports whether expr is an untyped constant,
|
||||||
|
// and indicates what its default type is.
|
||||||
|
// scope may be nil.
|
||||||
|
func (f *File) IsUntypedConst(expr ast.Expr) (defType string, ok bool) { |
||||||
|
// Re-evaluate expr outside of its context to see if it's untyped.
|
||||||
|
// (An expr evaluated within, for example, an assignment context will get the type of the LHS.)
|
||||||
|
exprStr := f.Render(expr) |
||||||
|
tv, err := types.Eval(f.Pkg.fset, f.Pkg.TypesPkg, expr.Pos(), exprStr) |
||||||
|
if err != nil { |
||||||
|
return "", false |
||||||
|
} |
||||||
|
if b, ok := tv.Type.(*types.Basic); ok { |
||||||
|
if dt, ok := basicTypeKinds[b.Kind()]; ok { |
||||||
|
return dt, true |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
return "", false |
||||||
|
} |
||||||
|
|
||||||
|
func (f *File) isMain() bool { |
||||||
|
if f.AST.Name.Name == "main" { |
||||||
|
return true |
||||||
|
} |
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
const directiveSpecifyDisableReason = "specify-disable-reason" |
||||||
|
|
||||||
|
func (f *File) lint(rules []Rule, config Config, failures chan Failure) { |
||||||
|
rulesConfig := config.Rules |
||||||
|
_, mustSpecifyDisableReason := config.Directives[directiveSpecifyDisableReason] |
||||||
|
disabledIntervals := f.disabledIntervals(rules, mustSpecifyDisableReason, failures) |
||||||
|
for _, currentRule := range rules { |
||||||
|
ruleConfig := rulesConfig[currentRule.Name()] |
||||||
|
currentFailures := currentRule.Apply(f, ruleConfig.Arguments) |
||||||
|
for idx, failure := range currentFailures { |
||||||
|
if failure.RuleName == "" { |
||||||
|
failure.RuleName = currentRule.Name() |
||||||
|
} |
||||||
|
if failure.Node != nil { |
||||||
|
failure.Position = ToFailurePosition(failure.Node.Pos(), failure.Node.End(), f) |
||||||
|
} |
||||||
|
currentFailures[idx] = failure |
||||||
|
} |
||||||
|
currentFailures = f.filterFailures(currentFailures, disabledIntervals) |
||||||
|
for _, failure := range currentFailures { |
||||||
|
if failure.Confidence >= config.Confidence { |
||||||
|
failures <- failure |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
type enableDisableConfig struct { |
||||||
|
enabled bool |
||||||
|
position int |
||||||
|
} |
||||||
|
|
||||||
|
const directiveRE = `^//[\s]*revive:(enable|disable)(?:-(line|next-line))?(?::([^\s]+))?[\s]*(?: (.+))?$` |
||||||
|
const directivePos = 1 |
||||||
|
const modifierPos = 2 |
||||||
|
const rulesPos = 3 |
||||||
|
const reasonPos = 4 |
||||||
|
|
||||||
|
var re = regexp.MustCompile(directiveRE) |
||||||
|
|
||||||
|
func (f *File) disabledIntervals(rules []Rule, mustSpecifyDisableReason bool, failures chan Failure) disabledIntervalsMap { |
||||||
|
enabledDisabledRulesMap := make(map[string][]enableDisableConfig) |
||||||
|
|
||||||
|
getEnabledDisabledIntervals := func() disabledIntervalsMap { |
||||||
|
result := make(disabledIntervalsMap) |
||||||
|
|
||||||
|
for ruleName, disabledArr := range enabledDisabledRulesMap { |
||||||
|
ruleResult := []DisabledInterval{} |
||||||
|
for i := 0; i < len(disabledArr); i++ { |
||||||
|
interval := DisabledInterval{ |
||||||
|
RuleName: ruleName, |
||||||
|
From: token.Position{ |
||||||
|
Filename: f.Name, |
||||||
|
Line: disabledArr[i].position, |
||||||
|
}, |
||||||
|
To: token.Position{ |
||||||
|
Filename: f.Name, |
||||||
|
Line: math.MaxInt32, |
||||||
|
}, |
||||||
|
} |
||||||
|
if i%2 == 0 { |
||||||
|
ruleResult = append(ruleResult, interval) |
||||||
|
} else { |
||||||
|
ruleResult[len(ruleResult)-1].To.Line = disabledArr[i].position |
||||||
|
} |
||||||
|
} |
||||||
|
result[ruleName] = ruleResult |
||||||
|
} |
||||||
|
|
||||||
|
return result |
||||||
|
} |
||||||
|
|
||||||
|
handleConfig := func(isEnabled bool, line int, name string) { |
||||||
|
existing, ok := enabledDisabledRulesMap[name] |
||||||
|
if !ok { |
||||||
|
existing = []enableDisableConfig{} |
||||||
|
enabledDisabledRulesMap[name] = existing |
||||||
|
} |
||||||
|
if (len(existing) > 1 && existing[len(existing)-1].enabled == isEnabled) || |
||||||
|
(len(existing) == 0 && isEnabled) { |
||||||
|
return |
||||||
|
} |
||||||
|
existing = append(existing, enableDisableConfig{ |
||||||
|
enabled: isEnabled, |
||||||
|
position: line, |
||||||
|
}) |
||||||
|
enabledDisabledRulesMap[name] = existing |
||||||
|
} |
||||||
|
|
||||||
|
handleRules := func(filename, modifier string, isEnabled bool, line int, ruleNames []string) []DisabledInterval { |
||||||
|
var result []DisabledInterval |
||||||
|
for _, name := range ruleNames { |
||||||
|
if modifier == "line" { |
||||||
|
handleConfig(isEnabled, line, name) |
||||||
|
handleConfig(!isEnabled, line, name) |
||||||
|
} else if modifier == "next-line" { |
||||||
|
handleConfig(isEnabled, line+1, name) |
||||||
|
handleConfig(!isEnabled, line+1, name) |
||||||
|
} else { |
||||||
|
handleConfig(isEnabled, line, name) |
||||||
|
} |
||||||
|
} |
||||||
|
return result |
||||||
|
} |
||||||
|
|
||||||
|
handleComment := func(filename string, c *ast.CommentGroup, line int) { |
||||||
|
comments := c.List |
||||||
|
for _, c := range comments { |
||||||
|
match := re.FindStringSubmatch(c.Text) |
||||||
|
if len(match) == 0 { |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
ruleNames := []string{} |
||||||
|
tempNames := strings.Split(match[rulesPos], ",") |
||||||
|
for _, name := range tempNames { |
||||||
|
name = strings.Trim(name, "\n") |
||||||
|
if len(name) > 0 { |
||||||
|
ruleNames = append(ruleNames, name) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
mustCheckDisablingReason := mustSpecifyDisableReason && match[directivePos] == "disable" |
||||||
|
if mustCheckDisablingReason && strings.Trim(match[reasonPos], " ") == "" { |
||||||
|
failures <- Failure{ |
||||||
|
Confidence: 1, |
||||||
|
RuleName: directiveSpecifyDisableReason, |
||||||
|
Failure: "reason of lint disabling not found", |
||||||
|
Position: ToFailurePosition(c.Pos(), c.End(), f), |
||||||
|
Node: c, |
||||||
|
} |
||||||
|
continue // skip this linter disabling directive
|
||||||
|
} |
||||||
|
|
||||||
|
// TODO: optimize
|
||||||
|
if len(ruleNames) == 0 { |
||||||
|
for _, rule := range rules { |
||||||
|
ruleNames = append(ruleNames, rule.Name()) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
handleRules(filename, match[modifierPos], match[directivePos] == "enable", line, ruleNames) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
comments := f.AST.Comments |
||||||
|
for _, c := range comments { |
||||||
|
handleComment(f.Name, c, f.ToPosition(c.End()).Line) |
||||||
|
} |
||||||
|
|
||||||
|
return getEnabledDisabledIntervals() |
||||||
|
} |
||||||
|
|
||||||
|
func (f *File) filterFailures(failures []Failure, disabledIntervals disabledIntervalsMap) []Failure { |
||||||
|
result := []Failure{} |
||||||
|
for _, failure := range failures { |
||||||
|
fStart := failure.Position.Start.Line |
||||||
|
fEnd := failure.Position.End.Line |
||||||
|
intervals, ok := disabledIntervals[failure.RuleName] |
||||||
|
if !ok { |
||||||
|
result = append(result, failure) |
||||||
|
} else { |
||||||
|
include := true |
||||||
|
for _, interval := range intervals { |
||||||
|
intStart := interval.From.Line |
||||||
|
intEnd := interval.To.Line |
||||||
|
if (fStart >= intStart && fStart <= intEnd) || |
||||||
|
(fEnd >= intStart && fEnd <= intEnd) { |
||||||
|
include = false |
||||||
|
break |
||||||
|
} |
||||||
|
} |
||||||
|
if include { |
||||||
|
result = append(result, failure) |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
return result |
||||||
|
} |
@ -0,0 +1,14 @@ |
|||||||
|
package lint |
||||||
|
|
||||||
|
// FormatterMetadata configuration of a formatter
|
||||||
|
type FormatterMetadata struct { |
||||||
|
Name string |
||||||
|
Description string |
||||||
|
Sample string |
||||||
|
} |
||||||
|
|
||||||
|
// Formatter defines an interface for failure formatters
|
||||||
|
type Formatter interface { |
||||||
|
Format(<-chan Failure, Config) (string, error) |
||||||
|
Name() string |
||||||
|
} |
@ -0,0 +1,99 @@ |
|||||||
|
package lint |
||||||
|
|
||||||
|
import ( |
||||||
|
"bufio" |
||||||
|
"bytes" |
||||||
|
"fmt" |
||||||
|
"go/token" |
||||||
|
"os" |
||||||
|
"sync" |
||||||
|
) |
||||||
|
|
||||||
|
// ReadFile defines an abstraction for reading files.
|
||||||
|
type ReadFile func(path string) (result []byte, err error) |
||||||
|
|
||||||
|
type disabledIntervalsMap = map[string][]DisabledInterval |
||||||
|
|
||||||
|
// Linter is used for linting set of files.
|
||||||
|
type Linter struct { |
||||||
|
reader ReadFile |
||||||
|
} |
||||||
|
|
||||||
|
// New creates a new Linter
|
||||||
|
func New(reader ReadFile) Linter { |
||||||
|
return Linter{reader: reader} |
||||||
|
} |
||||||
|
|
||||||
|
var ( |
||||||
|
genHdr = []byte("// Code generated ") |
||||||
|
genFtr = []byte(" DO NOT EDIT.") |
||||||
|
) |
||||||
|
|
||||||
|
// Lint lints a set of files with the specified rule.
|
||||||
|
func (l *Linter) Lint(packages [][]string, ruleSet []Rule, config Config) (<-chan Failure, error) { |
||||||
|
failures := make(chan Failure) |
||||||
|
|
||||||
|
var wg sync.WaitGroup |
||||||
|
for _, pkg := range packages { |
||||||
|
wg.Add(1) |
||||||
|
go func(pkg []string) { |
||||||
|
if err := l.lintPackage(pkg, ruleSet, config, failures); err != nil { |
||||||
|
fmt.Fprintln(os.Stderr, err) |
||||||
|
os.Exit(1) |
||||||
|
} |
||||||
|
defer wg.Done() |
||||||
|
}(pkg) |
||||||
|
} |
||||||
|
|
||||||
|
go func() { |
||||||
|
wg.Wait() |
||||||
|
close(failures) |
||||||
|
}() |
||||||
|
|
||||||
|
return failures, nil |
||||||
|
} |
||||||
|
|
||||||
|
func (l *Linter) lintPackage(filenames []string, ruleSet []Rule, config Config, failures chan Failure) error { |
||||||
|
pkg := &Package{ |
||||||
|
fset: token.NewFileSet(), |
||||||
|
files: map[string]*File{}, |
||||||
|
mu: sync.Mutex{}, |
||||||
|
} |
||||||
|
for _, filename := range filenames { |
||||||
|
content, err := l.reader(filename) |
||||||
|
if err != nil { |
||||||
|
return err |
||||||
|
} |
||||||
|
if isGenerated(content) && !config.IgnoreGeneratedHeader { |
||||||
|
continue |
||||||
|
} |
||||||
|
|
||||||
|
file, err := NewFile(filename, content, pkg) |
||||||
|
if err != nil { |
||||||
|
return err |
||||||
|
} |
||||||
|
pkg.files[filename] = file |
||||||
|
} |
||||||
|
|
||||||
|
if len(pkg.files) == 0 { |
||||||
|
return nil |
||||||
|
} |
||||||
|
|
||||||
|
pkg.lint(ruleSet, config, failures) |
||||||
|
|
||||||
|
return nil |
||||||
|
} |
||||||
|
|
||||||
|
// isGenerated reports whether the source file is generated code
|
||||||
|
// according the rules from https://golang.org/s/generatedcode.
|
||||||
|
// This is inherited from the original go lint.
|
||||||
|
func isGenerated(src []byte) bool { |
||||||
|
sc := bufio.NewScanner(bytes.NewReader(src)) |
||||||
|
for sc.Scan() { |
||||||
|
b := sc.Bytes() |
||||||
|
if bytes.HasPrefix(b, genHdr) && bytes.HasSuffix(b, genFtr) && len(b) >= len(genHdr)+len(genFtr) { |
||||||
|
return true |
||||||
|
} |
||||||
|
} |
||||||
|
return false |
||||||
|
} |
@ -0,0 +1,178 @@ |
|||||||
|
package lint |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
"go/token" |
||||||
|
"go/types" |
||||||
|
"sync" |
||||||
|
|
||||||
|
"golang.org/x/tools/go/gcexportdata" |
||||||
|
) |
||||||
|
|
||||||
|
// Package represents a package in the project.
|
||||||
|
type Package struct { |
||||||
|
fset *token.FileSet |
||||||
|
files map[string]*File |
||||||
|
|
||||||
|
TypesPkg *types.Package |
||||||
|
TypesInfo *types.Info |
||||||
|
|
||||||
|
// sortable is the set of types in the package that implement sort.Interface.
|
||||||
|
Sortable map[string]bool |
||||||
|
// main is whether this is a "main" package.
|
||||||
|
main int |
||||||
|
mu sync.Mutex |
||||||
|
} |
||||||
|
|
||||||
|
var newImporter = func(fset *token.FileSet) types.ImporterFrom { |
||||||
|
return gcexportdata.NewImporter(fset, make(map[string]*types.Package)) |
||||||
|
} |
||||||
|
|
||||||
|
var ( |
||||||
|
trueValue = 1 |
||||||
|
falseValue = 2 |
||||||
|
notSet = 3 |
||||||
|
) |
||||||
|
|
||||||
|
// IsMain returns if that's the main package.
|
||||||
|
func (p *Package) IsMain() bool { |
||||||
|
if p.main == trueValue { |
||||||
|
return true |
||||||
|
} else if p.main == falseValue { |
||||||
|
return false |
||||||
|
} |
||||||
|
for _, f := range p.files { |
||||||
|
if f.isMain() { |
||||||
|
p.main = trueValue |
||||||
|
return true |
||||||
|
} |
||||||
|
} |
||||||
|
p.main = falseValue |
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
// TypeCheck performs type checking for given package.
|
||||||
|
func (p *Package) TypeCheck() error { |
||||||
|
p.mu.Lock() |
||||||
|
// If type checking has already been performed
|
||||||
|
// skip it.
|
||||||
|
if p.TypesInfo != nil || p.TypesPkg != nil { |
||||||
|
p.mu.Unlock() |
||||||
|
return nil |
||||||
|
} |
||||||
|
config := &types.Config{ |
||||||
|
// By setting a no-op error reporter, the type checker does as much work as possible.
|
||||||
|
Error: func(error) {}, |
||||||
|
Importer: newImporter(p.fset), |
||||||
|
} |
||||||
|
info := &types.Info{ |
||||||
|
Types: make(map[ast.Expr]types.TypeAndValue), |
||||||
|
Defs: make(map[*ast.Ident]types.Object), |
||||||
|
Uses: make(map[*ast.Ident]types.Object), |
||||||
|
Scopes: make(map[ast.Node]*types.Scope), |
||||||
|
} |
||||||
|
var anyFile *File |
||||||
|
var astFiles []*ast.File |
||||||
|
for _, f := range p.files { |
||||||
|
anyFile = f |
||||||
|
astFiles = append(astFiles, f.AST) |
||||||
|
} |
||||||
|
|
||||||
|
typesPkg, err := check(config, anyFile.AST.Name.Name, p.fset, astFiles, info) |
||||||
|
|
||||||
|
// Remember the typechecking info, even if config.Check failed,
|
||||||
|
// since we will get partial information.
|
||||||
|
p.TypesPkg = typesPkg |
||||||
|
p.TypesInfo = info |
||||||
|
p.mu.Unlock() |
||||||
|
return err |
||||||
|
} |
||||||
|
|
||||||
|
// check function encapsulates the call to go/types.Config.Check method and
|
||||||
|
// recovers if the called method panics (see issue #59)
|
||||||
|
func check(config *types.Config, n string, fset *token.FileSet, astFiles []*ast.File, info *types.Info) (p *types.Package, err error) { |
||||||
|
defer func() { |
||||||
|
if r := recover(); r != nil { |
||||||
|
err, _ = r.(error) |
||||||
|
p = nil |
||||||
|
return |
||||||
|
} |
||||||
|
}() |
||||||
|
|
||||||
|
return config.Check(n, fset, astFiles, info) |
||||||
|
} |
||||||
|
|
||||||
|
// TypeOf returns the type of an expression.
|
||||||
|
func (p *Package) TypeOf(expr ast.Expr) types.Type { |
||||||
|
if p.TypesInfo == nil { |
||||||
|
return nil |
||||||
|
} |
||||||
|
return p.TypesInfo.TypeOf(expr) |
||||||
|
} |
||||||
|
|
||||||
|
type walker struct { |
||||||
|
nmap map[string]int |
||||||
|
has map[string]int |
||||||
|
} |
||||||
|
|
||||||
|
func (w *walker) Visit(n ast.Node) ast.Visitor { |
||||||
|
fn, ok := n.(*ast.FuncDecl) |
||||||
|
if !ok || fn.Recv == nil || len(fn.Recv.List) == 0 { |
||||||
|
return w |
||||||
|
} |
||||||
|
// TODO(dsymonds): We could check the signature to be more precise.
|
||||||
|
recv := receiverType(fn) |
||||||
|
if i, ok := w.nmap[fn.Name.Name]; ok { |
||||||
|
w.has[recv] |= i |
||||||
|
} |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
func (p *Package) scanSortable() { |
||||||
|
p.Sortable = make(map[string]bool) |
||||||
|
|
||||||
|
// bitfield for which methods exist on each type.
|
||||||
|
const ( |
||||||
|
Len = 1 << iota |
||||||
|
Less |
||||||
|
Swap |
||||||
|
) |
||||||
|
nmap := map[string]int{"Len": Len, "Less": Less, "Swap": Swap} |
||||||
|
has := make(map[string]int) |
||||||
|
for _, f := range p.files { |
||||||
|
ast.Walk(&walker{nmap, has}, f.AST) |
||||||
|
} |
||||||
|
for typ, ms := range has { |
||||||
|
if ms == Len|Less|Swap { |
||||||
|
p.Sortable[typ] = true |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// receiverType returns the named type of the method receiver, sans "*",
|
||||||
|
// or "invalid-type" if fn.Recv is ill formed.
|
||||||
|
func receiverType(fn *ast.FuncDecl) string { |
||||||
|
switch e := fn.Recv.List[0].Type.(type) { |
||||||
|
case *ast.Ident: |
||||||
|
return e.Name |
||||||
|
case *ast.StarExpr: |
||||||
|
if id, ok := e.X.(*ast.Ident); ok { |
||||||
|
return id.Name |
||||||
|
} |
||||||
|
} |
||||||
|
// The parser accepts much more than just the legal forms.
|
||||||
|
return "invalid-type" |
||||||
|
} |
||||||
|
|
||||||
|
func (p *Package) lint(rules []Rule, config Config, failures chan Failure) { |
||||||
|
p.scanSortable() |
||||||
|
var wg sync.WaitGroup |
||||||
|
for _, file := range p.files { |
||||||
|
wg.Add(1) |
||||||
|
go (func(file *File) { |
||||||
|
file.lint(rules, config, failures) |
||||||
|
defer wg.Done() |
||||||
|
})(file) |
||||||
|
} |
||||||
|
wg.Wait() |
||||||
|
} |
@ -0,0 +1,31 @@ |
|||||||
|
package lint |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/token" |
||||||
|
) |
||||||
|
|
||||||
|
// DisabledInterval contains a single disabled interval and the associated rule name.
|
||||||
|
type DisabledInterval struct { |
||||||
|
From token.Position |
||||||
|
To token.Position |
||||||
|
RuleName string |
||||||
|
} |
||||||
|
|
||||||
|
// Rule defines an abstract rule interaface
|
||||||
|
type Rule interface { |
||||||
|
Name() string |
||||||
|
Apply(*File, Arguments) []Failure |
||||||
|
} |
||||||
|
|
||||||
|
// AbstractRule defines an abstract rule.
|
||||||
|
type AbstractRule struct { |
||||||
|
Failures []Failure |
||||||
|
} |
||||||
|
|
||||||
|
// ToFailurePosition returns the failure position.
|
||||||
|
func ToFailurePosition(start token.Pos, end token.Pos, file *File) FailurePosition { |
||||||
|
return FailurePosition{ |
||||||
|
Start: file.ToPosition(start), |
||||||
|
End: file.ToPosition(end), |
||||||
|
} |
||||||
|
} |
@ -0,0 +1,128 @@ |
|||||||
|
package lint |
||||||
|
|
||||||
|
import ( |
||||||
|
"strings" |
||||||
|
"unicode" |
||||||
|
) |
||||||
|
|
||||||
|
// Name returns a different name if it should be different.
|
||||||
|
func Name(name string, whitelist, blacklist []string) (should string) { |
||||||
|
// Fast path for simple cases: "_" and all lowercase.
|
||||||
|
if name == "_" { |
||||||
|
return name |
||||||
|
} |
||||||
|
allLower := true |
||||||
|
for _, r := range name { |
||||||
|
if !unicode.IsLower(r) { |
||||||
|
allLower = false |
||||||
|
break |
||||||
|
} |
||||||
|
} |
||||||
|
if allLower { |
||||||
|
return name |
||||||
|
} |
||||||
|
|
||||||
|
// Split camelCase at any lower->upper transition, and split on underscores.
|
||||||
|
// Check each word for common initialisms.
|
||||||
|
runes := []rune(name) |
||||||
|
w, i := 0, 0 // index of start of word, scan
|
||||||
|
for i+1 <= len(runes) { |
||||||
|
eow := false // whether we hit the end of a word
|
||||||
|
if i+1 == len(runes) { |
||||||
|
eow = true |
||||||
|
} else if runes[i+1] == '_' { |
||||||
|
// underscore; shift the remainder forward over any run of underscores
|
||||||
|
eow = true |
||||||
|
n := 1 |
||||||
|
for i+n+1 < len(runes) && runes[i+n+1] == '_' { |
||||||
|
n++ |
||||||
|
} |
||||||
|
|
||||||
|
// Leave at most one underscore if the underscore is between two digits
|
||||||
|
if i+n+1 < len(runes) && unicode.IsDigit(runes[i]) && unicode.IsDigit(runes[i+n+1]) { |
||||||
|
n-- |
||||||
|
} |
||||||
|
|
||||||
|
copy(runes[i+1:], runes[i+n+1:]) |
||||||
|
runes = runes[:len(runes)-n] |
||||||
|
} else if unicode.IsLower(runes[i]) && !unicode.IsLower(runes[i+1]) { |
||||||
|
// lower->non-lower
|
||||||
|
eow = true |
||||||
|
} |
||||||
|
i++ |
||||||
|
if !eow { |
||||||
|
continue |
||||||
|
} |
||||||
|
|
||||||
|
// [w,i) is a word.
|
||||||
|
word := string(runes[w:i]) |
||||||
|
ignoreInitWarnings := map[string]bool{} |
||||||
|
for _, i := range whitelist { |
||||||
|
ignoreInitWarnings[i] = true |
||||||
|
} |
||||||
|
|
||||||
|
extraInits := map[string]bool{} |
||||||
|
for _, i := range blacklist { |
||||||
|
extraInits[i] = true |
||||||
|
} |
||||||
|
|
||||||
|
if u := strings.ToUpper(word); (commonInitialisms[u] || extraInits[u]) && !ignoreInitWarnings[u] { |
||||||
|
// Keep consistent case, which is lowercase only at the start.
|
||||||
|
if w == 0 && unicode.IsLower(runes[w]) { |
||||||
|
u = strings.ToLower(u) |
||||||
|
} |
||||||
|
// All the common initialisms are ASCII,
|
||||||
|
// so we can replace the bytes exactly.
|
||||||
|
copy(runes[w:], []rune(u)) |
||||||
|
} else if w > 0 && strings.ToLower(word) == word { |
||||||
|
// already all lowercase, and not the first word, so uppercase the first character.
|
||||||
|
runes[w] = unicode.ToUpper(runes[w]) |
||||||
|
} |
||||||
|
w = i |
||||||
|
} |
||||||
|
return string(runes) |
||||||
|
} |
||||||
|
|
||||||
|
// commonInitialisms is a set of common initialisms.
|
||||||
|
// Only add entries that are highly unlikely to be non-initialisms.
|
||||||
|
// For instance, "ID" is fine (Freudian code is rare), but "AND" is not.
|
||||||
|
var commonInitialisms = map[string]bool{ |
||||||
|
"ACL": true, |
||||||
|
"API": true, |
||||||
|
"ASCII": true, |
||||||
|
"CPU": true, |
||||||
|
"CSS": true, |
||||||
|
"DNS": true, |
||||||
|
"EOF": true, |
||||||
|
"GUID": true, |
||||||
|
"HTML": true, |
||||||
|
"HTTP": true, |
||||||
|
"HTTPS": true, |
||||||
|
"ID": true, |
||||||
|
"IP": true, |
||||||
|
"JSON": true, |
||||||
|
"LHS": true, |
||||||
|
"QPS": true, |
||||||
|
"RAM": true, |
||||||
|
"RHS": true, |
||||||
|
"RPC": true, |
||||||
|
"SLA": true, |
||||||
|
"SMTP": true, |
||||||
|
"SQL": true, |
||||||
|
"SSH": true, |
||||||
|
"TCP": true, |
||||||
|
"TLS": true, |
||||||
|
"TTL": true, |
||||||
|
"UDP": true, |
||||||
|
"UI": true, |
||||||
|
"UID": true, |
||||||
|
"UUID": true, |
||||||
|
"URI": true, |
||||||
|
"URL": true, |
||||||
|
"UTF8": true, |
||||||
|
"VM": true, |
||||||
|
"XML": true, |
||||||
|
"XMPP": true, |
||||||
|
"XSRF": true, |
||||||
|
"XSS": true, |
||||||
|
} |
@ -0,0 +1,151 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
"go/ast" |
||||||
|
"strconv" |
||||||
|
"strings" |
||||||
|
) |
||||||
|
|
||||||
|
const ( |
||||||
|
defaultStrLitLimit = 2 |
||||||
|
kindFLOAT = "FLOAT" |
||||||
|
kindINT = "INT" |
||||||
|
kindSTRING = "STRING" |
||||||
|
) |
||||||
|
|
||||||
|
type whiteList map[string]map[string]bool |
||||||
|
|
||||||
|
func newWhiteList() whiteList { |
||||||
|
return map[string]map[string]bool{kindINT: map[string]bool{}, kindFLOAT: map[string]bool{}, kindSTRING: map[string]bool{}} |
||||||
|
} |
||||||
|
|
||||||
|
func (wl whiteList) add(kind string, list string) { |
||||||
|
elems := strings.Split(list, ",") |
||||||
|
for _, e := range elems { |
||||||
|
wl[kind][e] = true |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// AddConstantRule lints unused params in functions.
|
||||||
|
type AddConstantRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *AddConstantRule) Apply(file *lint.File, arguments lint.Arguments) []lint.Failure { |
||||||
|
strLitLimit := defaultStrLitLimit |
||||||
|
var whiteList = newWhiteList() |
||||||
|
if len(arguments) > 0 { |
||||||
|
args, ok := arguments[0].(map[string]interface{}) |
||||||
|
if !ok { |
||||||
|
panic(fmt.Sprintf("Invalid argument to the add-constant rule. Expecting a k,v map, got %T", arguments[0])) |
||||||
|
} |
||||||
|
for k, v := range args { |
||||||
|
kind := "" |
||||||
|
switch k { |
||||||
|
case "allowFloats": |
||||||
|
kind = kindFLOAT |
||||||
|
fallthrough |
||||||
|
case "allowInts": |
||||||
|
if kind == "" { |
||||||
|
kind = kindINT |
||||||
|
} |
||||||
|
fallthrough |
||||||
|
case "allowStrs": |
||||||
|
if kind == "" { |
||||||
|
kind = kindSTRING |
||||||
|
} |
||||||
|
list, ok := v.(string) |
||||||
|
if !ok { |
||||||
|
panic(fmt.Sprintf("Invalid argument to the add-constant rule, string expected. Got '%v' (%T)", v, v)) |
||||||
|
} |
||||||
|
whiteList.add(kind, list) |
||||||
|
case "maxLitCount": |
||||||
|
sl, ok := v.(string) |
||||||
|
if !ok { |
||||||
|
panic(fmt.Sprintf("Invalid argument to the add-constant rule, expecting string representation of an integer. Got '%v' (%T)", v, v)) |
||||||
|
} |
||||||
|
|
||||||
|
limit, err := strconv.Atoi(sl) |
||||||
|
if err != nil { |
||||||
|
panic(fmt.Sprintf("Invalid argument to the add-constant rule, expecting string representation of an integer. Got '%v'", v)) |
||||||
|
} |
||||||
|
strLitLimit = limit |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
onFailure := func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
} |
||||||
|
|
||||||
|
w := lintAddConstantRule{onFailure: onFailure, strLits: make(map[string]int, 0), strLitLimit: strLitLimit, whiteLst: whiteList} |
||||||
|
|
||||||
|
ast.Walk(w, file.AST) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *AddConstantRule) Name() string { |
||||||
|
return "add-constant" |
||||||
|
} |
||||||
|
|
||||||
|
type lintAddConstantRule struct { |
||||||
|
onFailure func(lint.Failure) |
||||||
|
strLits map[string]int |
||||||
|
strLitLimit int |
||||||
|
whiteLst whiteList |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintAddConstantRule) Visit(node ast.Node) ast.Visitor { |
||||||
|
switch n := node.(type) { |
||||||
|
case *ast.GenDecl: |
||||||
|
return nil // skip declarations
|
||||||
|
case *ast.BasicLit: |
||||||
|
switch kind := n.Kind.String(); kind { |
||||||
|
case kindFLOAT, kindINT: |
||||||
|
w.checkNumLit(kind, n) |
||||||
|
case kindSTRING: |
||||||
|
w.checkStrLit(n) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
return w |
||||||
|
|
||||||
|
} |
||||||
|
|
||||||
|
func (w lintAddConstantRule) checkStrLit(n *ast.BasicLit) { |
||||||
|
if w.whiteLst[kindSTRING][n.Value] { |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
count := w.strLits[n.Value] |
||||||
|
if count >= 0 { |
||||||
|
w.strLits[n.Value] = count + 1 |
||||||
|
if w.strLits[n.Value] > w.strLitLimit { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: n, |
||||||
|
Category: "style", |
||||||
|
Failure: fmt.Sprintf("string literal %s appears, at least, %d times, create a named constant for it", n.Value, w.strLits[n.Value]), |
||||||
|
}) |
||||||
|
w.strLits[n.Value] = -1 // mark it to avoid failing again on the same literal
|
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintAddConstantRule) checkNumLit(kind string, n *ast.BasicLit) { |
||||||
|
if w.whiteLst[kind][n.Value] { |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: n, |
||||||
|
Category: "style", |
||||||
|
Failure: fmt.Sprintf("avoid magic numbers like '%s', create a named constant for it", n.Value), |
||||||
|
}) |
||||||
|
} |
@ -0,0 +1,67 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"go/ast" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// ArgumentsLimitRule lints given else constructs.
|
||||||
|
type ArgumentsLimitRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *ArgumentsLimitRule) Apply(file *lint.File, arguments lint.Arguments) []lint.Failure { |
||||||
|
if len(arguments) != 1 { |
||||||
|
panic(`invalid configuration for "argument-limit"`) |
||||||
|
} |
||||||
|
|
||||||
|
total, ok := arguments[0].(int64) // Alt. non panicking version
|
||||||
|
if !ok { |
||||||
|
panic(`invalid value passed as argument number to the "argument-list" rule`) |
||||||
|
} |
||||||
|
|
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
walker := lintArgsNum{ |
||||||
|
total: int(total), |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(walker, file.AST) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *ArgumentsLimitRule) Name() string { |
||||||
|
return "argument-limit" |
||||||
|
} |
||||||
|
|
||||||
|
type lintArgsNum struct { |
||||||
|
total int |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintArgsNum) Visit(n ast.Node) ast.Visitor { |
||||||
|
node, ok := n.(*ast.FuncDecl) |
||||||
|
if ok { |
||||||
|
num := 0 |
||||||
|
for _, l := range node.Type.Params.List { |
||||||
|
for range l.Names { |
||||||
|
num++ |
||||||
|
} |
||||||
|
} |
||||||
|
if num > w.total { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Failure: fmt.Sprintf("maximum number of arguments per function exceeded; max %d but got %d", w.total, num), |
||||||
|
Node: node.Type, |
||||||
|
}) |
||||||
|
return w |
||||||
|
} |
||||||
|
} |
||||||
|
return w |
||||||
|
} |
@ -0,0 +1,94 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
"go/token" |
||||||
|
"go/types" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// AtomicRule lints given else constructs.
|
||||||
|
type AtomicRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *AtomicRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
walker := atomic{ |
||||||
|
pkgTypesInfo: file.Pkg.TypesInfo, |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(walker, file.AST) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *AtomicRule) Name() string { |
||||||
|
return "atomic" |
||||||
|
} |
||||||
|
|
||||||
|
type atomic struct { |
||||||
|
pkgTypesInfo *types.Info |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w atomic) Visit(node ast.Node) ast.Visitor { |
||||||
|
n, ok := node.(*ast.AssignStmt) |
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
if len(n.Lhs) != len(n.Rhs) { |
||||||
|
return nil // skip assignment sub-tree
|
||||||
|
} |
||||||
|
if len(n.Lhs) == 1 && n.Tok == token.DEFINE { |
||||||
|
return nil // skip assignment sub-tree
|
||||||
|
} |
||||||
|
|
||||||
|
for i, right := range n.Rhs { |
||||||
|
call, ok := right.(*ast.CallExpr) |
||||||
|
if !ok { |
||||||
|
continue |
||||||
|
} |
||||||
|
sel, ok := call.Fun.(*ast.SelectorExpr) |
||||||
|
if !ok { |
||||||
|
continue |
||||||
|
} |
||||||
|
pkgIdent, _ := sel.X.(*ast.Ident) |
||||||
|
if w.pkgTypesInfo != nil { |
||||||
|
pkgName, ok := w.pkgTypesInfo.Uses[pkgIdent].(*types.PkgName) |
||||||
|
if !ok || pkgName.Imported().Path() != "sync/atomic" { |
||||||
|
continue |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
switch sel.Sel.Name { |
||||||
|
case "AddInt32", "AddInt64", "AddUint32", "AddUint64", "AddUintptr": |
||||||
|
left := n.Lhs[i] |
||||||
|
if len(call.Args) != 2 { |
||||||
|
continue |
||||||
|
} |
||||||
|
arg := call.Args[0] |
||||||
|
broken := false |
||||||
|
|
||||||
|
if uarg, ok := arg.(*ast.UnaryExpr); ok && uarg.Op == token.AND { |
||||||
|
broken = gofmt(left) == gofmt(uarg.X) |
||||||
|
} else if star, ok := left.(*ast.StarExpr); ok { |
||||||
|
broken = gofmt(star.X) == gofmt(arg) |
||||||
|
} |
||||||
|
|
||||||
|
if broken { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Failure: "direct assignment to atomic value", |
||||||
|
Node: n, |
||||||
|
}) |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
return w |
||||||
|
} |
@ -0,0 +1,84 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// BareReturnRule lints given else constructs.
|
||||||
|
type BareReturnRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *BareReturnRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
onFailure := func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
} |
||||||
|
|
||||||
|
w := lintBareReturnRule{onFailure: onFailure} |
||||||
|
ast.Walk(w, file.AST) |
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *BareReturnRule) Name() string { |
||||||
|
return "bare-return" |
||||||
|
} |
||||||
|
|
||||||
|
type lintBareReturnRule struct { |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintBareReturnRule) Visit(node ast.Node) ast.Visitor { |
||||||
|
switch n := node.(type) { |
||||||
|
case *ast.FuncDecl: |
||||||
|
w.checkFunc(n.Type.Results, n.Body) |
||||||
|
case *ast.FuncLit: // to cope with deferred functions and go-routines
|
||||||
|
w.checkFunc(n.Type.Results, n.Body) |
||||||
|
} |
||||||
|
|
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
// checkFunc will verify if the given function has named result and bare returns
|
||||||
|
func (w lintBareReturnRule) checkFunc(results *ast.FieldList, body *ast.BlockStmt) { |
||||||
|
hasNamedResults := results != nil && len(results.List) > 0 && results.List[0].Names != nil |
||||||
|
if !hasNamedResults || body == nil { |
||||||
|
return // nothing to do
|
||||||
|
} |
||||||
|
|
||||||
|
brf := bareReturnFinder{w.onFailure} |
||||||
|
ast.Walk(brf, body) |
||||||
|
} |
||||||
|
|
||||||
|
type bareReturnFinder struct { |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w bareReturnFinder) Visit(node ast.Node) ast.Visitor { |
||||||
|
_, ok := node.(*ast.FuncLit) |
||||||
|
if ok { |
||||||
|
// skip analysing function literals
|
||||||
|
// they will analyzed by the lintBareReturnRule.Visit method
|
||||||
|
return nil |
||||||
|
} |
||||||
|
|
||||||
|
rs, ok := node.(*ast.ReturnStmt) |
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
if len(rs.Results) > 0 { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: rs, |
||||||
|
Failure: "avoid using bare returns, please add return expressions", |
||||||
|
}) |
||||||
|
|
||||||
|
return w |
||||||
|
} |
@ -0,0 +1,74 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// BlankImportsRule lints given else constructs.
|
||||||
|
type BlankImportsRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *BlankImportsRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
fileAst := file.AST |
||||||
|
walker := lintBlankImports{ |
||||||
|
file: file, |
||||||
|
fileAst: fileAst, |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(walker, fileAst) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *BlankImportsRule) Name() string { |
||||||
|
return "blank-imports" |
||||||
|
} |
||||||
|
|
||||||
|
type lintBlankImports struct { |
||||||
|
fileAst *ast.File |
||||||
|
file *lint.File |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintBlankImports) Visit(_ ast.Node) ast.Visitor { |
||||||
|
// In package main and in tests, we don't complain about blank imports.
|
||||||
|
if w.file.Pkg.IsMain() || w.file.IsTest() { |
||||||
|
return nil |
||||||
|
} |
||||||
|
|
||||||
|
// The first element of each contiguous group of blank imports should have
|
||||||
|
// an explanatory comment of some kind.
|
||||||
|
for i, imp := range w.fileAst.Imports { |
||||||
|
pos := w.file.ToPosition(imp.Pos()) |
||||||
|
|
||||||
|
if !isBlank(imp.Name) { |
||||||
|
continue // Ignore non-blank imports.
|
||||||
|
} |
||||||
|
if i > 0 { |
||||||
|
prev := w.fileAst.Imports[i-1] |
||||||
|
prevPos := w.file.ToPosition(prev.Pos()) |
||||||
|
if isBlank(prev.Name) && prevPos.Line+1 == pos.Line { |
||||||
|
continue // A subsequent blank in a group.
|
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// This is the first blank import of a group.
|
||||||
|
if imp.Doc == nil && imp.Comment == nil { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Node: imp, |
||||||
|
Failure: "a blank import should be only in a main or test package, or have a comment justifying it", |
||||||
|
Confidence: 1, |
||||||
|
Category: "imports", |
||||||
|
}) |
||||||
|
} |
||||||
|
} |
||||||
|
return nil |
||||||
|
} |
@ -0,0 +1,73 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
"go/token" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// BoolLiteralRule warns when logic expressions contains Boolean literals.
|
||||||
|
type BoolLiteralRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *BoolLiteralRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
onFailure := func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
} |
||||||
|
|
||||||
|
astFile := file.AST |
||||||
|
w := &lintBoolLiteral{astFile, onFailure} |
||||||
|
ast.Walk(w, astFile) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *BoolLiteralRule) Name() string { |
||||||
|
return "bool-literal-in-expr" |
||||||
|
} |
||||||
|
|
||||||
|
type lintBoolLiteral struct { |
||||||
|
file *ast.File |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w *lintBoolLiteral) Visit(node ast.Node) ast.Visitor { |
||||||
|
switch n := node.(type) { |
||||||
|
case *ast.BinaryExpr: |
||||||
|
if !isBoolOp(n.Op) { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
lexeme, ok := isExprABooleanLit(n.X) |
||||||
|
if !ok { |
||||||
|
lexeme, ok = isExprABooleanLit(n.Y) |
||||||
|
|
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
isConstant := (n.Op == token.LAND && lexeme == "false") || (n.Op == token.LOR && lexeme == "true") |
||||||
|
|
||||||
|
if isConstant { |
||||||
|
w.addFailure(n, "Boolean expression seems to always evaluate to "+lexeme, "logic") |
||||||
|
} else { |
||||||
|
w.addFailure(n, "omit Boolean literal in expression", "style") |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintBoolLiteral) addFailure(node ast.Node, msg string, cat string) { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: node, |
||||||
|
Category: cat, |
||||||
|
Failure: msg, |
||||||
|
}) |
||||||
|
} |
@ -0,0 +1,70 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// CallToGCRule lints calls to the garbage collector.
|
||||||
|
type CallToGCRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *CallToGCRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
onFailure := func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
} |
||||||
|
|
||||||
|
var gcTriggeringFunctions = map[string]map[string]bool{ |
||||||
|
"runtime": map[string]bool{"GC": true}, |
||||||
|
} |
||||||
|
|
||||||
|
w := lintCallToGC{onFailure, gcTriggeringFunctions} |
||||||
|
ast.Walk(w, file.AST) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *CallToGCRule) Name() string { |
||||||
|
return "call-to-gc" |
||||||
|
} |
||||||
|
|
||||||
|
type lintCallToGC struct { |
||||||
|
onFailure func(lint.Failure) |
||||||
|
gcTriggeringFunctions map[string]map[string]bool |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintCallToGC) Visit(node ast.Node) ast.Visitor { |
||||||
|
ce, ok := node.(*ast.CallExpr) |
||||||
|
if !ok { |
||||||
|
return w // nothing to do, the node is not a call
|
||||||
|
} |
||||||
|
|
||||||
|
fc, ok := ce.Fun.(*ast.SelectorExpr) |
||||||
|
if !ok { |
||||||
|
return nil // nothing to do, the call is not of the form pkg.func(...)
|
||||||
|
} |
||||||
|
|
||||||
|
id, ok := fc.X.(*ast.Ident) |
||||||
|
|
||||||
|
if !ok { |
||||||
|
return nil // in case X is not an id (it should be!)
|
||||||
|
} |
||||||
|
|
||||||
|
fn := fc.Sel.Name |
||||||
|
pkg := id.Name |
||||||
|
if !w.gcTriggeringFunctions[pkg][fn] { |
||||||
|
return nil // it isn't a call to a GC triggering function
|
||||||
|
} |
||||||
|
|
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: node, |
||||||
|
Category: "bad practice", |
||||||
|
Failure: "explicit call to the garbage collector", |
||||||
|
}) |
||||||
|
|
||||||
|
return w |
||||||
|
} |
@ -0,0 +1,195 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"go/ast" |
||||||
|
"go/token" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
"golang.org/x/tools/go/ast/astutil" |
||||||
|
) |
||||||
|
|
||||||
|
// CognitiveComplexityRule lints given else constructs.
|
||||||
|
type CognitiveComplexityRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *CognitiveComplexityRule) Apply(file *lint.File, arguments lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
const expectedArgumentsCount = 1 |
||||||
|
if len(arguments) < expectedArgumentsCount { |
||||||
|
panic(fmt.Sprintf("not enough arguments for cognitive-complexity, expected %d, got %d", expectedArgumentsCount, len(arguments))) |
||||||
|
} |
||||||
|
complexity, ok := arguments[0].(int64) |
||||||
|
if !ok { |
||||||
|
panic(fmt.Sprintf("invalid argument type for cognitive-complexity, expected int64, got %T", arguments[0])) |
||||||
|
} |
||||||
|
|
||||||
|
linter := cognitiveComplexityLinter{ |
||||||
|
file: file, |
||||||
|
maxComplexity: int(complexity), |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
linter.lint() |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *CognitiveComplexityRule) Name() string { |
||||||
|
return "cognitive-complexity" |
||||||
|
} |
||||||
|
|
||||||
|
type cognitiveComplexityLinter struct { |
||||||
|
file *lint.File |
||||||
|
maxComplexity int |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w cognitiveComplexityLinter) lint() { |
||||||
|
f := w.file |
||||||
|
for _, decl := range f.AST.Decls { |
||||||
|
if fn, ok := decl.(*ast.FuncDecl); ok { |
||||||
|
v := cognitiveComplexityVisitor{} |
||||||
|
c := v.subTreeComplexity(fn.Body) |
||||||
|
if c > w.maxComplexity { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Category: "maintenance", |
||||||
|
Failure: fmt.Sprintf("function %s has cognitive complexity %d (> max enabled %d)", funcName(fn), c, w.maxComplexity), |
||||||
|
Node: fn, |
||||||
|
}) |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
type cognitiveComplexityVisitor struct { |
||||||
|
complexity int |
||||||
|
nestingLevel int |
||||||
|
} |
||||||
|
|
||||||
|
// subTreeComplexity calculates the cognitive complexity of an AST-subtree.
|
||||||
|
func (v cognitiveComplexityVisitor) subTreeComplexity(n ast.Node) int { |
||||||
|
ast.Walk(&v, n) |
||||||
|
return v.complexity |
||||||
|
} |
||||||
|
|
||||||
|
// Visit implements the ast.Visitor interface.
|
||||||
|
func (v *cognitiveComplexityVisitor) Visit(n ast.Node) ast.Visitor { |
||||||
|
switch n := n.(type) { |
||||||
|
case *ast.IfStmt: |
||||||
|
targets := []ast.Node{n.Cond, n.Body, n.Else} |
||||||
|
v.walk(1, targets...) |
||||||
|
return nil |
||||||
|
case *ast.ForStmt: |
||||||
|
targets := []ast.Node{n.Cond, n.Body} |
||||||
|
v.walk(1, targets...) |
||||||
|
return nil |
||||||
|
case *ast.RangeStmt: |
||||||
|
v.walk(1, n.Body) |
||||||
|
return nil |
||||||
|
case *ast.SelectStmt: |
||||||
|
v.walk(1, n.Body) |
||||||
|
return nil |
||||||
|
case *ast.SwitchStmt: |
||||||
|
v.walk(1, n.Body) |
||||||
|
return nil |
||||||
|
case *ast.TypeSwitchStmt: |
||||||
|
v.walk(1, n.Body) |
||||||
|
return nil |
||||||
|
case *ast.FuncLit: |
||||||
|
v.walk(0, n.Body) // do not increment the complexity, just do the nesting
|
||||||
|
return nil |
||||||
|
case *ast.BinaryExpr: |
||||||
|
v.complexity += v.binExpComplexity(n) |
||||||
|
return nil // skip visiting binexp sub-tree (already visited by binExpComplexity)
|
||||||
|
case *ast.BranchStmt: |
||||||
|
if n.Label != nil { |
||||||
|
v.complexity += 1 |
||||||
|
} |
||||||
|
} |
||||||
|
// TODO handle (at least) direct recursion
|
||||||
|
|
||||||
|
return v |
||||||
|
} |
||||||
|
|
||||||
|
func (v *cognitiveComplexityVisitor) walk(complexityIncrement int, targets ...ast.Node) { |
||||||
|
v.complexity += complexityIncrement + v.nestingLevel |
||||||
|
nesting := v.nestingLevel |
||||||
|
v.nestingLevel++ |
||||||
|
|
||||||
|
for _, t := range targets { |
||||||
|
if t == nil { |
||||||
|
continue |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(v, t) |
||||||
|
} |
||||||
|
|
||||||
|
v.nestingLevel = nesting |
||||||
|
} |
||||||
|
|
||||||
|
func (cognitiveComplexityVisitor) binExpComplexity(n *ast.BinaryExpr) int { |
||||||
|
calculator := binExprComplexityCalculator{opsStack: []token.Token{}} |
||||||
|
|
||||||
|
astutil.Apply(n, calculator.pre, calculator.post) |
||||||
|
|
||||||
|
return calculator.complexity |
||||||
|
} |
||||||
|
|
||||||
|
type binExprComplexityCalculator struct { |
||||||
|
complexity int |
||||||
|
opsStack []token.Token // stack of bool operators
|
||||||
|
subexpStarted bool |
||||||
|
} |
||||||
|
|
||||||
|
func (becc *binExprComplexityCalculator) pre(c *astutil.Cursor) bool { |
||||||
|
switch n := c.Node().(type) { |
||||||
|
case *ast.BinaryExpr: |
||||||
|
isBoolOp := n.Op == token.LAND || n.Op == token.LOR |
||||||
|
if !isBoolOp { |
||||||
|
break |
||||||
|
} |
||||||
|
|
||||||
|
ops := len(becc.opsStack) |
||||||
|
// if
|
||||||
|
// is the first boolop in the expression OR
|
||||||
|
// is the first boolop inside a subexpression (...) OR
|
||||||
|
// is not the same to the previous one
|
||||||
|
// then
|
||||||
|
// increment complexity
|
||||||
|
if ops == 0 || becc.subexpStarted || n.Op != becc.opsStack[ops-1] { |
||||||
|
becc.complexity++ |
||||||
|
becc.subexpStarted = false |
||||||
|
} |
||||||
|
|
||||||
|
becc.opsStack = append(becc.opsStack, n.Op) |
||||||
|
case *ast.ParenExpr: |
||||||
|
becc.subexpStarted = true |
||||||
|
} |
||||||
|
|
||||||
|
return true |
||||||
|
} |
||||||
|
|
||||||
|
func (becc *binExprComplexityCalculator) post(c *astutil.Cursor) bool { |
||||||
|
switch n := c.Node().(type) { |
||||||
|
case *ast.BinaryExpr: |
||||||
|
isBoolOp := n.Op == token.LAND || n.Op == token.LOR |
||||||
|
if !isBoolOp { |
||||||
|
break |
||||||
|
} |
||||||
|
|
||||||
|
ops := len(becc.opsStack) |
||||||
|
if ops > 0 { |
||||||
|
becc.opsStack = becc.opsStack[:ops-1] |
||||||
|
} |
||||||
|
case *ast.ParenExpr: |
||||||
|
becc.subexpStarted = false |
||||||
|
} |
||||||
|
|
||||||
|
return true |
||||||
|
} |
@ -0,0 +1,190 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"go/ast" |
||||||
|
|
||||||
|
"strings" |
||||||
|
"sync" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
type referenceMethod struct { |
||||||
|
fileName string |
||||||
|
id *ast.Ident |
||||||
|
} |
||||||
|
|
||||||
|
type pkgMethods struct { |
||||||
|
pkg *lint.Package |
||||||
|
methods map[string]map[string]*referenceMethod |
||||||
|
mu *sync.Mutex |
||||||
|
} |
||||||
|
|
||||||
|
type packages struct { |
||||||
|
pkgs []pkgMethods |
||||||
|
mu sync.Mutex |
||||||
|
} |
||||||
|
|
||||||
|
func (ps *packages) methodNames(lp *lint.Package) pkgMethods { |
||||||
|
ps.mu.Lock() |
||||||
|
|
||||||
|
for _, pkg := range ps.pkgs { |
||||||
|
if pkg.pkg == lp { |
||||||
|
ps.mu.Unlock() |
||||||
|
return pkg |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
pkgm := pkgMethods{pkg: lp, methods: make(map[string]map[string]*referenceMethod), mu: &sync.Mutex{}} |
||||||
|
ps.pkgs = append(ps.pkgs, pkgm) |
||||||
|
|
||||||
|
ps.mu.Unlock() |
||||||
|
return pkgm |
||||||
|
} |
||||||
|
|
||||||
|
var allPkgs = packages{pkgs: make([]pkgMethods, 1)} |
||||||
|
|
||||||
|
// ConfusingNamingRule lints method names that differ only by capitalization
|
||||||
|
type ConfusingNamingRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *ConfusingNamingRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
fileAst := file.AST |
||||||
|
pkgm := allPkgs.methodNames(file.Pkg) |
||||||
|
walker := lintConfusingNames{ |
||||||
|
fileName: file.Name, |
||||||
|
pkgm: pkgm, |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(&walker, fileAst) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *ConfusingNamingRule) Name() string { |
||||||
|
return "confusing-naming" |
||||||
|
} |
||||||
|
|
||||||
|
//checkMethodName checks if a given method/function name is similar (just case differences) to other method/function of the same struct/file.
|
||||||
|
func checkMethodName(holder string, id *ast.Ident, w *lintConfusingNames) { |
||||||
|
if id.Name == "init" && holder == defaultStructName { |
||||||
|
// ignore init functions
|
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
pkgm := w.pkgm |
||||||
|
name := strings.ToUpper(id.Name) |
||||||
|
|
||||||
|
pkgm.mu.Lock() |
||||||
|
defer pkgm.mu.Unlock() |
||||||
|
|
||||||
|
if pkgm.methods[holder] != nil { |
||||||
|
if pkgm.methods[holder][name] != nil { |
||||||
|
refMethod := pkgm.methods[holder][name] |
||||||
|
// confusing names
|
||||||
|
var kind string |
||||||
|
if holder == defaultStructName { |
||||||
|
kind = "function" |
||||||
|
} else { |
||||||
|
kind = "method" |
||||||
|
} |
||||||
|
var fileName string |
||||||
|
if w.fileName == refMethod.fileName { |
||||||
|
fileName = "the same source file" |
||||||
|
} else { |
||||||
|
fileName = refMethod.fileName |
||||||
|
} |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Failure: fmt.Sprintf("Method '%s' differs only by capitalization to %s '%s' in %s", id.Name, kind, refMethod.id.Name, fileName), |
||||||
|
Confidence: 1, |
||||||
|
Node: id, |
||||||
|
Category: "naming", |
||||||
|
}) |
||||||
|
|
||||||
|
return |
||||||
|
} |
||||||
|
} else { |
||||||
|
pkgm.methods[holder] = make(map[string]*referenceMethod, 1) |
||||||
|
} |
||||||
|
|
||||||
|
// update the black list
|
||||||
|
if pkgm.methods[holder] == nil { |
||||||
|
println("no entry for '", holder, "'") |
||||||
|
} |
||||||
|
pkgm.methods[holder][name] = &referenceMethod{fileName: w.fileName, id: id} |
||||||
|
} |
||||||
|
|
||||||
|
type lintConfusingNames struct { |
||||||
|
fileName string |
||||||
|
pkgm pkgMethods |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
const defaultStructName = "_" // used to map functions
|
||||||
|
|
||||||
|
//getStructName of a function receiver. Defaults to defaultStructName
|
||||||
|
func getStructName(r *ast.FieldList) string { |
||||||
|
result := defaultStructName |
||||||
|
|
||||||
|
if r == nil || len(r.List) < 1 { |
||||||
|
return result |
||||||
|
} |
||||||
|
|
||||||
|
t := r.List[0].Type |
||||||
|
|
||||||
|
if p, _ := t.(*ast.StarExpr); p != nil { // if a pointer receiver => dereference pointer receiver types
|
||||||
|
t = p.X |
||||||
|
} |
||||||
|
|
||||||
|
if p, _ := t.(*ast.Ident); p != nil { |
||||||
|
result = p.Name |
||||||
|
} |
||||||
|
|
||||||
|
return result |
||||||
|
} |
||||||
|
|
||||||
|
func checkStructFields(fields *ast.FieldList, structName string, w *lintConfusingNames) { |
||||||
|
bl := make(map[string]bool, len(fields.List)) |
||||||
|
for _, f := range fields.List { |
||||||
|
for _, id := range f.Names { |
||||||
|
normName := strings.ToUpper(id.Name) |
||||||
|
if bl[normName] { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Failure: fmt.Sprintf("Field '%s' differs only by capitalization to other field in the struct type %s", id.Name, structName), |
||||||
|
Confidence: 1, |
||||||
|
Node: id, |
||||||
|
Category: "naming", |
||||||
|
}) |
||||||
|
} else { |
||||||
|
bl[normName] = true |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
func (w *lintConfusingNames) Visit(n ast.Node) ast.Visitor { |
||||||
|
switch v := n.(type) { |
||||||
|
case *ast.FuncDecl: |
||||||
|
// Exclude naming warnings for functions that are exported to C but
|
||||||
|
// not exported in the Go API.
|
||||||
|
// See https://github.com/golang/lint/issues/144.
|
||||||
|
if ast.IsExported(v.Name.Name) || !isCgoExported(v) { |
||||||
|
checkMethodName(getStructName(v.Recv), v.Name, w) |
||||||
|
} |
||||||
|
case *ast.TypeSpec: |
||||||
|
if s, ok := v.Type.(*ast.StructType); ok { |
||||||
|
checkStructFields(s.Fields, v.Name.Name, w) |
||||||
|
} |
||||||
|
|
||||||
|
default: |
||||||
|
// will add other checks like field names, struct names, etc.
|
||||||
|
} |
||||||
|
|
||||||
|
return w |
||||||
|
} |
@ -0,0 +1,67 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// ConfusingResultsRule lints given function declarations
|
||||||
|
type ConfusingResultsRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *ConfusingResultsRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
fileAst := file.AST |
||||||
|
walker := lintConfusingResults{ |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(walker, fileAst) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *ConfusingResultsRule) Name() string { |
||||||
|
return "confusing-results" |
||||||
|
} |
||||||
|
|
||||||
|
type lintConfusingResults struct { |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintConfusingResults) Visit(n ast.Node) ast.Visitor { |
||||||
|
fn, ok := n.(*ast.FuncDecl) |
||||||
|
if !ok || fn.Type.Results == nil || len(fn.Type.Results.List) < 2 { |
||||||
|
return w |
||||||
|
} |
||||||
|
lastType := "" |
||||||
|
for _, result := range fn.Type.Results.List { |
||||||
|
if len(result.Names) > 0 { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
t, ok := result.Type.(*ast.Ident) |
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
if t.Name == lastType { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Node: n, |
||||||
|
Confidence: 1, |
||||||
|
Category: "naming", |
||||||
|
Failure: "unnamed results of the same type may be confusing, consider using named results", |
||||||
|
}) |
||||||
|
break |
||||||
|
} |
||||||
|
lastType = t.Name |
||||||
|
|
||||||
|
} |
||||||
|
|
||||||
|
return w |
||||||
|
} |
@ -0,0 +1,88 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
"go/ast" |
||||||
|
"go/token" |
||||||
|
) |
||||||
|
|
||||||
|
// ConstantLogicalExprRule warns on constant logical expressions.
|
||||||
|
type ConstantLogicalExprRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *ConstantLogicalExprRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
onFailure := func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
} |
||||||
|
|
||||||
|
astFile := file.AST |
||||||
|
w := &lintConstantLogicalExpr{astFile, onFailure} |
||||||
|
ast.Walk(w, astFile) |
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *ConstantLogicalExprRule) Name() string { |
||||||
|
return "constant-logical-expr" |
||||||
|
} |
||||||
|
|
||||||
|
type lintConstantLogicalExpr struct { |
||||||
|
file *ast.File |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w *lintConstantLogicalExpr) Visit(node ast.Node) ast.Visitor { |
||||||
|
switch n := node.(type) { |
||||||
|
case *ast.BinaryExpr: |
||||||
|
if !w.isOperatorWithLogicalResult(n.Op) { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
if gofmt(n.X) != gofmt(n.Y) { // check if subexpressions are the same
|
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
if n.Op == token.EQL { |
||||||
|
w.newFailure(n, "expression always evaluates to true") |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
if w.isInequalityOperator(n.Op) { |
||||||
|
w.newFailure(n, "expression always evaluates to false") |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
w.newFailure(n, "left and right hand-side sub-expressions are the same") |
||||||
|
} |
||||||
|
|
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
func (w *lintConstantLogicalExpr) isOperatorWithLogicalResult(t token.Token) bool { |
||||||
|
switch t { |
||||||
|
case token.LAND, token.LOR, token.EQL, token.LSS, token.GTR, token.NEQ, token.LEQ, token.GEQ: |
||||||
|
return true |
||||||
|
} |
||||||
|
|
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
func (w *lintConstantLogicalExpr) isInequalityOperator(t token.Token) bool { |
||||||
|
switch t { |
||||||
|
case token.LSS, token.GTR, token.NEQ, token.LEQ, token.GEQ: |
||||||
|
return true |
||||||
|
} |
||||||
|
|
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintConstantLogicalExpr) newFailure(node ast.Node, msg string) { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: node, |
||||||
|
Category: "logic", |
||||||
|
Failure: msg, |
||||||
|
}) |
||||||
|
} |
@ -0,0 +1,60 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// ContextAsArgumentRule lints given else constructs.
|
||||||
|
type ContextAsArgumentRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *ContextAsArgumentRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
fileAst := file.AST |
||||||
|
walker := lintContextArguments{ |
||||||
|
file: file, |
||||||
|
fileAst: fileAst, |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(walker, fileAst) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *ContextAsArgumentRule) Name() string { |
||||||
|
return "context-as-argument" |
||||||
|
} |
||||||
|
|
||||||
|
type lintContextArguments struct { |
||||||
|
file *lint.File |
||||||
|
fileAst *ast.File |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintContextArguments) Visit(n ast.Node) ast.Visitor { |
||||||
|
fn, ok := n.(*ast.FuncDecl) |
||||||
|
if !ok || len(fn.Type.Params.List) <= 1 { |
||||||
|
return w |
||||||
|
} |
||||||
|
// A context.Context should be the first parameter of a function.
|
||||||
|
// Flag any that show up after the first.
|
||||||
|
for _, arg := range fn.Type.Params.List[1:] { |
||||||
|
if isPkgDot(arg.Type, "context", "Context") { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Node: fn, |
||||||
|
Category: "arg-order", |
||||||
|
Failure: "context.Context should be the first parameter of a function", |
||||||
|
Confidence: 0.9, |
||||||
|
}) |
||||||
|
break // only flag one
|
||||||
|
} |
||||||
|
} |
||||||
|
return w |
||||||
|
} |
@ -0,0 +1,81 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"go/ast" |
||||||
|
"go/types" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// ContextKeysType lints given else constructs.
|
||||||
|
type ContextKeysType struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *ContextKeysType) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
fileAst := file.AST |
||||||
|
walker := lintContextKeyTypes{ |
||||||
|
file: file, |
||||||
|
fileAst: fileAst, |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
file.Pkg.TypeCheck() |
||||||
|
ast.Walk(walker, fileAst) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *ContextKeysType) Name() string { |
||||||
|
return "context-keys-type" |
||||||
|
} |
||||||
|
|
||||||
|
type lintContextKeyTypes struct { |
||||||
|
file *lint.File |
||||||
|
fileAst *ast.File |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintContextKeyTypes) Visit(n ast.Node) ast.Visitor { |
||||||
|
switch n := n.(type) { |
||||||
|
case *ast.CallExpr: |
||||||
|
checkContextKeyType(w, n) |
||||||
|
} |
||||||
|
|
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
func checkContextKeyType(w lintContextKeyTypes, x *ast.CallExpr) { |
||||||
|
f := w.file |
||||||
|
sel, ok := x.Fun.(*ast.SelectorExpr) |
||||||
|
if !ok { |
||||||
|
return |
||||||
|
} |
||||||
|
pkg, ok := sel.X.(*ast.Ident) |
||||||
|
if !ok || pkg.Name != "context" { |
||||||
|
return |
||||||
|
} |
||||||
|
if sel.Sel.Name != "WithValue" { |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
// key is second argument to context.WithValue
|
||||||
|
if len(x.Args) != 3 { |
||||||
|
return |
||||||
|
} |
||||||
|
key := f.Pkg.TypesInfo.Types[x.Args[1]] |
||||||
|
|
||||||
|
if ktyp, ok := key.Type.(*types.Basic); ok && ktyp.Kind() != types.Invalid { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: x, |
||||||
|
Category: "content", |
||||||
|
Failure: fmt.Sprintf("should not use basic type %s as key in context.WithValue", key.Type), |
||||||
|
}) |
||||||
|
} |
||||||
|
} |
@ -0,0 +1,115 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"go/ast" |
||||||
|
"go/token" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// Based on https://github.com/fzipp/gocyclo
|
||||||
|
|
||||||
|
// CyclomaticRule lints given else constructs.
|
||||||
|
type CyclomaticRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *CyclomaticRule) Apply(file *lint.File, arguments lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
complexity, ok := arguments[0].(int64) // Alt. non panicking version
|
||||||
|
if !ok { |
||||||
|
panic("invalid argument for cyclomatic complexity") |
||||||
|
} |
||||||
|
|
||||||
|
fileAst := file.AST |
||||||
|
walker := lintCyclomatic{ |
||||||
|
file: file, |
||||||
|
complexity: int(complexity), |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(walker, fileAst) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *CyclomaticRule) Name() string { |
||||||
|
return "cyclomatic" |
||||||
|
} |
||||||
|
|
||||||
|
type lintCyclomatic struct { |
||||||
|
file *lint.File |
||||||
|
complexity int |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintCyclomatic) Visit(_ ast.Node) ast.Visitor { |
||||||
|
f := w.file |
||||||
|
for _, decl := range f.AST.Decls { |
||||||
|
if fn, ok := decl.(*ast.FuncDecl); ok { |
||||||
|
c := complexity(fn) |
||||||
|
if c > w.complexity { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Category: "maintenance", |
||||||
|
Failure: fmt.Sprintf("function %s has cyclomatic complexity %d", funcName(fn), c), |
||||||
|
Node: fn, |
||||||
|
}) |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
return nil |
||||||
|
} |
||||||
|
|
||||||
|
// funcName returns the name representation of a function or method:
|
||||||
|
// "(Type).Name" for methods or simply "Name" for functions.
|
||||||
|
func funcName(fn *ast.FuncDecl) string { |
||||||
|
if fn.Recv != nil { |
||||||
|
if fn.Recv.NumFields() > 0 { |
||||||
|
typ := fn.Recv.List[0].Type |
||||||
|
return fmt.Sprintf("(%s).%s", recvString(typ), fn.Name) |
||||||
|
} |
||||||
|
} |
||||||
|
return fn.Name.Name |
||||||
|
} |
||||||
|
|
||||||
|
// recvString returns a string representation of recv of the
|
||||||
|
// form "T", "*T", or "BADRECV" (if not a proper receiver type).
|
||||||
|
func recvString(recv ast.Expr) string { |
||||||
|
switch t := recv.(type) { |
||||||
|
case *ast.Ident: |
||||||
|
return t.Name |
||||||
|
case *ast.StarExpr: |
||||||
|
return "*" + recvString(t.X) |
||||||
|
} |
||||||
|
return "BADRECV" |
||||||
|
} |
||||||
|
|
||||||
|
// complexity calculates the cyclomatic complexity of a function.
|
||||||
|
func complexity(fn *ast.FuncDecl) int { |
||||||
|
v := complexityVisitor{} |
||||||
|
ast.Walk(&v, fn) |
||||||
|
return v.Complexity |
||||||
|
} |
||||||
|
|
||||||
|
type complexityVisitor struct { |
||||||
|
// Complexity is the cyclomatic complexity
|
||||||
|
Complexity int |
||||||
|
} |
||||||
|
|
||||||
|
// Visit implements the ast.Visitor interface.
|
||||||
|
func (v *complexityVisitor) Visit(n ast.Node) ast.Visitor { |
||||||
|
switch n := n.(type) { |
||||||
|
case *ast.FuncDecl, *ast.IfStmt, *ast.ForStmt, *ast.RangeStmt, *ast.CaseClause, *ast.CommClause: |
||||||
|
v.Complexity++ |
||||||
|
case *ast.BinaryExpr: |
||||||
|
if n.Op == token.LAND || n.Op == token.LOR { |
||||||
|
v.Complexity++ |
||||||
|
} |
||||||
|
} |
||||||
|
return v |
||||||
|
} |
@ -0,0 +1,94 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"go/ast" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// DeepExitRule lints program exit at functions other than main or init.
|
||||||
|
type DeepExitRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *DeepExitRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
onFailure := func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
} |
||||||
|
|
||||||
|
var exitFunctions = map[string]map[string]bool{ |
||||||
|
"os": map[string]bool{"Exit": true}, |
||||||
|
"syscall": map[string]bool{"Exit": true}, |
||||||
|
"log": map[string]bool{ |
||||||
|
"Fatal": true, |
||||||
|
"Fatalf": true, |
||||||
|
"Fatalln": true, |
||||||
|
"Panic": true, |
||||||
|
"Panicf": true, |
||||||
|
"Panicln": true, |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
w := lintDeepExit{onFailure, exitFunctions, file.IsTest()} |
||||||
|
ast.Walk(w, file.AST) |
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *DeepExitRule) Name() string { |
||||||
|
return "deep-exit" |
||||||
|
} |
||||||
|
|
||||||
|
type lintDeepExit struct { |
||||||
|
onFailure func(lint.Failure) |
||||||
|
exitFunctions map[string]map[string]bool |
||||||
|
isTestFile bool |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintDeepExit) Visit(node ast.Node) ast.Visitor { |
||||||
|
if fd, ok := node.(*ast.FuncDecl); ok { |
||||||
|
if w.mustIgnore(fd) { |
||||||
|
return nil // skip analysis of this function
|
||||||
|
} |
||||||
|
|
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
se, ok := node.(*ast.ExprStmt) |
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
ce, ok := se.X.(*ast.CallExpr) |
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
fc, ok := ce.Fun.(*ast.SelectorExpr) |
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
id, ok := fc.X.(*ast.Ident) |
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
fn := fc.Sel.Name |
||||||
|
pkg := id.Name |
||||||
|
if w.exitFunctions[pkg] != nil && w.exitFunctions[pkg][fn] { // it's a call to an exit function
|
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: ce, |
||||||
|
Category: "bad practice", |
||||||
|
Failure: fmt.Sprintf("calls to %s.%s only in main() or init() functions", pkg, fn), |
||||||
|
}) |
||||||
|
} |
||||||
|
|
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
func (w *lintDeepExit) mustIgnore(fd *ast.FuncDecl) bool { |
||||||
|
fn := fd.Name.Name |
||||||
|
|
||||||
|
return fn == "init" || fn == "main" || (w.isTestFile && fn == "TestMain") |
||||||
|
} |
@ -0,0 +1,54 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// DotImportsRule lints given else constructs.
|
||||||
|
type DotImportsRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *DotImportsRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
fileAst := file.AST |
||||||
|
walker := lintImports{ |
||||||
|
file: file, |
||||||
|
fileAst: fileAst, |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(walker, fileAst) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *DotImportsRule) Name() string { |
||||||
|
return "dot-imports" |
||||||
|
} |
||||||
|
|
||||||
|
type lintImports struct { |
||||||
|
file *lint.File |
||||||
|
fileAst *ast.File |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintImports) Visit(_ ast.Node) ast.Visitor { |
||||||
|
for i, is := range w.fileAst.Imports { |
||||||
|
_ = i |
||||||
|
if is.Name != nil && is.Name.Name == "." && !w.file.IsTest() { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Failure: "should not use dot imports", |
||||||
|
Node: is, |
||||||
|
Category: "imports", |
||||||
|
}) |
||||||
|
} |
||||||
|
} |
||||||
|
return nil |
||||||
|
} |
@ -0,0 +1,39 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// DuplicatedImportsRule lints given else constructs.
|
||||||
|
type DuplicatedImportsRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *DuplicatedImportsRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
impPaths := map[string]struct{}{} |
||||||
|
for _, imp := range file.AST.Imports { |
||||||
|
path := imp.Path.Value |
||||||
|
_, ok := impPaths[path] |
||||||
|
if ok { |
||||||
|
failures = append(failures, lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Failure: fmt.Sprintf("Package %s already imported", path), |
||||||
|
Node: imp, |
||||||
|
Category: "imports", |
||||||
|
}) |
||||||
|
continue |
||||||
|
} |
||||||
|
|
||||||
|
impPaths[path] = struct{}{} |
||||||
|
} |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *DuplicatedImportsRule) Name() string { |
||||||
|
return "duplicated-imports" |
||||||
|
} |
@ -0,0 +1,76 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// EmptyBlockRule lints given else constructs.
|
||||||
|
type EmptyBlockRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *EmptyBlockRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
onFailure := func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
} |
||||||
|
|
||||||
|
w := lintEmptyBlock{make([]*ast.BlockStmt, 0), onFailure} |
||||||
|
ast.Walk(w, file.AST) |
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *EmptyBlockRule) Name() string { |
||||||
|
return "empty-block" |
||||||
|
} |
||||||
|
|
||||||
|
type lintEmptyBlock struct { |
||||||
|
ignore []*ast.BlockStmt |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintEmptyBlock) Visit(node ast.Node) ast.Visitor { |
||||||
|
fd, ok := node.(*ast.FuncDecl) |
||||||
|
if ok { |
||||||
|
w.ignore = append(w.ignore, fd.Body) |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
fl, ok := node.(*ast.FuncLit) |
||||||
|
if ok { |
||||||
|
w.ignore = append(w.ignore, fl.Body) |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
block, ok := node.(*ast.BlockStmt) |
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
if mustIgnore(block, w.ignore) { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
if len(block.List) == 0 { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: block, |
||||||
|
Category: "logic", |
||||||
|
Failure: "this block is empty, you can remove it", |
||||||
|
}) |
||||||
|
} |
||||||
|
|
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
func mustIgnore(block *ast.BlockStmt, blackList []*ast.BlockStmt) bool { |
||||||
|
for _, b := range blackList { |
||||||
|
if b == block { |
||||||
|
return true |
||||||
|
} |
||||||
|
} |
||||||
|
return false |
||||||
|
} |
@ -0,0 +1,113 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
"go/token" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// EmptyLinesRule lints empty lines in blocks.
|
||||||
|
type EmptyLinesRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *EmptyLinesRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
onFailure := func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
} |
||||||
|
|
||||||
|
w := lintEmptyLines{file, file.CommentMap(), onFailure} |
||||||
|
ast.Walk(w, file.AST) |
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *EmptyLinesRule) Name() string { |
||||||
|
return "empty-lines" |
||||||
|
} |
||||||
|
|
||||||
|
type lintEmptyLines struct { |
||||||
|
file *lint.File |
||||||
|
cmap ast.CommentMap |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintEmptyLines) Visit(node ast.Node) ast.Visitor { |
||||||
|
block, ok := node.(*ast.BlockStmt) |
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
w.checkStart(block) |
||||||
|
w.checkEnd(block) |
||||||
|
|
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintEmptyLines) checkStart(block *ast.BlockStmt) { |
||||||
|
if len(block.List) == 0 { |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
start := w.position(block.Lbrace) |
||||||
|
firstNode := block.List[0] |
||||||
|
|
||||||
|
if w.commentBetween(start, firstNode) { |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
first := w.position(firstNode.Pos()) |
||||||
|
if first.Line-start.Line > 1 { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: block, |
||||||
|
Category: "style", |
||||||
|
Failure: "extra empty line at the start of a block", |
||||||
|
}) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintEmptyLines) checkEnd(block *ast.BlockStmt) { |
||||||
|
if len(block.List) < 1 { |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
end := w.position(block.Rbrace) |
||||||
|
lastNode := block.List[len(block.List)-1] |
||||||
|
|
||||||
|
if w.commentBetween(end, lastNode) { |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
last := w.position(lastNode.End()) |
||||||
|
if end.Line-last.Line > 1 { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: lastNode, |
||||||
|
Category: "style", |
||||||
|
Failure: "extra empty line at the end of a block", |
||||||
|
}) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintEmptyLines) commentBetween(position token.Position, node ast.Node) bool { |
||||||
|
comments := w.cmap.Filter(node).Comments() |
||||||
|
if len(comments) == 0 { |
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
for _, comment := range comments { |
||||||
|
start, end := w.position(comment.Pos()), w.position(comment.End()) |
||||||
|
if start.Line-position.Line == 1 || position.Line-end.Line == 1 { |
||||||
|
return true |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintEmptyLines) position(pos token.Pos) token.Position { |
||||||
|
return w.file.ToPosition(pos) |
||||||
|
} |
@ -0,0 +1,79 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"go/ast" |
||||||
|
"go/token" |
||||||
|
"strings" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// ErrorNamingRule lints given else constructs.
|
||||||
|
type ErrorNamingRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *ErrorNamingRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
fileAst := file.AST |
||||||
|
walker := lintErrors{ |
||||||
|
file: file, |
||||||
|
fileAst: fileAst, |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(walker, fileAst) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *ErrorNamingRule) Name() string { |
||||||
|
return "error-naming" |
||||||
|
} |
||||||
|
|
||||||
|
type lintErrors struct { |
||||||
|
file *lint.File |
||||||
|
fileAst *ast.File |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintErrors) Visit(_ ast.Node) ast.Visitor { |
||||||
|
for _, decl := range w.fileAst.Decls { |
||||||
|
gd, ok := decl.(*ast.GenDecl) |
||||||
|
if !ok || gd.Tok != token.VAR { |
||||||
|
continue |
||||||
|
} |
||||||
|
for _, spec := range gd.Specs { |
||||||
|
spec := spec.(*ast.ValueSpec) |
||||||
|
if len(spec.Names) != 1 || len(spec.Values) != 1 { |
||||||
|
continue |
||||||
|
} |
||||||
|
ce, ok := spec.Values[0].(*ast.CallExpr) |
||||||
|
if !ok { |
||||||
|
continue |
||||||
|
} |
||||||
|
if !isPkgDot(ce.Fun, "errors", "New") && !isPkgDot(ce.Fun, "fmt", "Errorf") { |
||||||
|
continue |
||||||
|
} |
||||||
|
|
||||||
|
id := spec.Names[0] |
||||||
|
prefix := "err" |
||||||
|
if id.IsExported() { |
||||||
|
prefix = "Err" |
||||||
|
} |
||||||
|
if !strings.HasPrefix(id.Name, prefix) { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Node: id, |
||||||
|
Confidence: 0.9, |
||||||
|
Category: "naming", |
||||||
|
Failure: fmt.Sprintf("error var %s should have name of the form %sFoo", id.Name, prefix), |
||||||
|
}) |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
return nil |
||||||
|
} |
@ -0,0 +1,67 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// ErrorReturnRule lints given else constructs.
|
||||||
|
type ErrorReturnRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *ErrorReturnRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
fileAst := file.AST |
||||||
|
walker := lintErrorReturn{ |
||||||
|
file: file, |
||||||
|
fileAst: fileAst, |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(walker, fileAst) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *ErrorReturnRule) Name() string { |
||||||
|
return "error-return" |
||||||
|
} |
||||||
|
|
||||||
|
type lintErrorReturn struct { |
||||||
|
file *lint.File |
||||||
|
fileAst *ast.File |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintErrorReturn) Visit(n ast.Node) ast.Visitor { |
||||||
|
fn, ok := n.(*ast.FuncDecl) |
||||||
|
if !ok || fn.Type.Results == nil { |
||||||
|
return w |
||||||
|
} |
||||||
|
ret := fn.Type.Results.List |
||||||
|
if len(ret) <= 1 { |
||||||
|
return w |
||||||
|
} |
||||||
|
if isIdent(ret[len(ret)-1].Type, "error") { |
||||||
|
return nil |
||||||
|
} |
||||||
|
// An error return parameter should be the last parameter.
|
||||||
|
// Flag any error parameters found before the last.
|
||||||
|
for _, r := range ret[:len(ret)-1] { |
||||||
|
if isIdent(r.Type, "error") { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Category: "arg-order", |
||||||
|
Confidence: 0.9, |
||||||
|
Node: fn, |
||||||
|
Failure: "error should be the last type when returning multiple items", |
||||||
|
}) |
||||||
|
break // only flag one
|
||||||
|
} |
||||||
|
} |
||||||
|
return w |
||||||
|
} |
@ -0,0 +1,98 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"go/ast" |
||||||
|
"go/token" |
||||||
|
"strconv" |
||||||
|
"unicode" |
||||||
|
"unicode/utf8" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// ErrorStringsRule lints given else constructs.
|
||||||
|
type ErrorStringsRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *ErrorStringsRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
fileAst := file.AST |
||||||
|
walker := lintErrorStrings{ |
||||||
|
file: file, |
||||||
|
fileAst: fileAst, |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(walker, fileAst) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *ErrorStringsRule) Name() string { |
||||||
|
return "error-strings" |
||||||
|
} |
||||||
|
|
||||||
|
type lintErrorStrings struct { |
||||||
|
file *lint.File |
||||||
|
fileAst *ast.File |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintErrorStrings) Visit(n ast.Node) ast.Visitor { |
||||||
|
ce, ok := n.(*ast.CallExpr) |
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
if !isPkgDot(ce.Fun, "errors", "New") && !isPkgDot(ce.Fun, "fmt", "Errorf") { |
||||||
|
return w |
||||||
|
} |
||||||
|
if len(ce.Args) < 1 { |
||||||
|
return w |
||||||
|
} |
||||||
|
str, ok := ce.Args[0].(*ast.BasicLit) |
||||||
|
if !ok || str.Kind != token.STRING { |
||||||
|
return w |
||||||
|
} |
||||||
|
s, _ := strconv.Unquote(str.Value) // can assume well-formed Go
|
||||||
|
if s == "" { |
||||||
|
return w |
||||||
|
} |
||||||
|
clean, conf := lintErrorString(s) |
||||||
|
if clean { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Node: str, |
||||||
|
Confidence: conf, |
||||||
|
Category: "errors", |
||||||
|
Failure: "error strings should not be capitalized or end with punctuation or a newline", |
||||||
|
}) |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
func lintErrorString(s string) (isClean bool, conf float64) { |
||||||
|
const basicConfidence = 0.8 |
||||||
|
const capConfidence = basicConfidence - 0.2 |
||||||
|
first, firstN := utf8.DecodeRuneInString(s) |
||||||
|
last, _ := utf8.DecodeLastRuneInString(s) |
||||||
|
if last == '.' || last == ':' || last == '!' || last == '\n' { |
||||||
|
return false, basicConfidence |
||||||
|
} |
||||||
|
if unicode.IsUpper(first) { |
||||||
|
// People use proper nouns and exported Go identifiers in error strings,
|
||||||
|
// so decrease the confidence of warnings for capitalization.
|
||||||
|
if len(s) <= firstN { |
||||||
|
return false, capConfidence |
||||||
|
} |
||||||
|
// Flag strings starting with something that doesn't look like an initialism.
|
||||||
|
if second, _ := utf8.DecodeRuneInString(s[firstN:]); !unicode.IsUpper(second) { |
||||||
|
return false, capConfidence |
||||||
|
} |
||||||
|
} |
||||||
|
return true, 0 |
||||||
|
} |
@ -0,0 +1,93 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"go/ast" |
||||||
|
"regexp" |
||||||
|
"strings" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// ErrorfRule lints given else constructs.
|
||||||
|
type ErrorfRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *ErrorfRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
fileAst := file.AST |
||||||
|
walker := lintErrorf{ |
||||||
|
file: file, |
||||||
|
fileAst: fileAst, |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
file.Pkg.TypeCheck() |
||||||
|
ast.Walk(walker, fileAst) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *ErrorfRule) Name() string { |
||||||
|
return "errorf" |
||||||
|
} |
||||||
|
|
||||||
|
type lintErrorf struct { |
||||||
|
file *lint.File |
||||||
|
fileAst *ast.File |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintErrorf) Visit(n ast.Node) ast.Visitor { |
||||||
|
ce, ok := n.(*ast.CallExpr) |
||||||
|
if !ok || len(ce.Args) != 1 { |
||||||
|
return w |
||||||
|
} |
||||||
|
isErrorsNew := isPkgDot(ce.Fun, "errors", "New") |
||||||
|
var isTestingError bool |
||||||
|
se, ok := ce.Fun.(*ast.SelectorExpr) |
||||||
|
if ok && se.Sel.Name == "Error" { |
||||||
|
if typ := w.file.Pkg.TypeOf(se.X); typ != nil { |
||||||
|
isTestingError = typ.String() == "*testing.T" |
||||||
|
} |
||||||
|
} |
||||||
|
if !isErrorsNew && !isTestingError { |
||||||
|
return w |
||||||
|
} |
||||||
|
arg := ce.Args[0] |
||||||
|
ce, ok = arg.(*ast.CallExpr) |
||||||
|
if !ok || !isPkgDot(ce.Fun, "fmt", "Sprintf") { |
||||||
|
return w |
||||||
|
} |
||||||
|
errorfPrefix := "fmt" |
||||||
|
if isTestingError { |
||||||
|
errorfPrefix = w.file.Render(se.X) |
||||||
|
} |
||||||
|
|
||||||
|
failure := lint.Failure{ |
||||||
|
Category: "errors", |
||||||
|
Node: n, |
||||||
|
Confidence: 1, |
||||||
|
Failure: fmt.Sprintf("should replace %s(fmt.Sprintf(...)) with %s.Errorf(...)", w.file.Render(se), errorfPrefix), |
||||||
|
} |
||||||
|
|
||||||
|
m := srcLineWithMatch(w.file, ce, `^(.*)`+w.file.Render(se)+`\(fmt\.Sprintf\((.*)\)\)(.*)$`) |
||||||
|
if m != nil { |
||||||
|
failure.ReplacementLine = m[1] + errorfPrefix + ".Errorf(" + m[2] + ")" + m[3] |
||||||
|
} |
||||||
|
|
||||||
|
w.onFailure(failure) |
||||||
|
|
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
func srcLineWithMatch(file *lint.File, node ast.Node, pattern string) (m []string) { |
||||||
|
line := srcLine(file.Content(), file.ToPosition(node.Pos())) |
||||||
|
line = strings.TrimSuffix(line, "\n") |
||||||
|
rx := regexp.MustCompile(pattern) |
||||||
|
return rx.FindStringSubmatch(line) |
||||||
|
} |
@ -0,0 +1,272 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"go/ast" |
||||||
|
"go/token" |
||||||
|
"strings" |
||||||
|
"unicode" |
||||||
|
"unicode/utf8" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// ExportedRule lints given else constructs.
|
||||||
|
type ExportedRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *ExportedRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
if isTest(file) { |
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
fileAst := file.AST |
||||||
|
walker := lintExported{ |
||||||
|
file: file, |
||||||
|
fileAst: fileAst, |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
genDeclMissingComments: make(map[*ast.GenDecl]bool), |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(&walker, fileAst) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *ExportedRule) Name() string { |
||||||
|
return "exported" |
||||||
|
} |
||||||
|
|
||||||
|
type lintExported struct { |
||||||
|
file *lint.File |
||||||
|
fileAst *ast.File |
||||||
|
lastGen *ast.GenDecl |
||||||
|
genDeclMissingComments map[*ast.GenDecl]bool |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w *lintExported) lintFuncDoc(fn *ast.FuncDecl) { |
||||||
|
if !ast.IsExported(fn.Name.Name) { |
||||||
|
// func is unexported
|
||||||
|
return |
||||||
|
} |
||||||
|
kind := "function" |
||||||
|
name := fn.Name.Name |
||||||
|
if fn.Recv != nil && len(fn.Recv.List) > 0 { |
||||||
|
// method
|
||||||
|
kind = "method" |
||||||
|
recv := receiverType(fn) |
||||||
|
if !ast.IsExported(recv) { |
||||||
|
// receiver is unexported
|
||||||
|
return |
||||||
|
} |
||||||
|
if commonMethods[name] { |
||||||
|
return |
||||||
|
} |
||||||
|
switch name { |
||||||
|
case "Len", "Less", "Swap": |
||||||
|
if w.file.Pkg.Sortable[recv] { |
||||||
|
return |
||||||
|
} |
||||||
|
} |
||||||
|
name = recv + "." + name |
||||||
|
} |
||||||
|
if fn.Doc == nil { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Node: fn, |
||||||
|
Confidence: 1, |
||||||
|
Category: "comments", |
||||||
|
Failure: fmt.Sprintf("exported %s %s should have comment or be unexported", kind, name), |
||||||
|
}) |
||||||
|
return |
||||||
|
} |
||||||
|
s := normalizeText(fn.Doc.Text()) |
||||||
|
prefix := fn.Name.Name + " " |
||||||
|
if !strings.HasPrefix(s, prefix) { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Node: fn.Doc, |
||||||
|
Confidence: 0.8, |
||||||
|
Category: "comments", |
||||||
|
Failure: fmt.Sprintf(`comment on exported %s %s should be of the form "%s..."`, kind, name, prefix), |
||||||
|
}) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
func (w *lintExported) checkStutter(id *ast.Ident, thing string) { |
||||||
|
pkg, name := w.fileAst.Name.Name, id.Name |
||||||
|
if !ast.IsExported(name) { |
||||||
|
// unexported name
|
||||||
|
return |
||||||
|
} |
||||||
|
// A name stutters if the package name is a strict prefix
|
||||||
|
// and the next character of the name starts a new word.
|
||||||
|
if len(name) <= len(pkg) { |
||||||
|
// name is too short to stutter.
|
||||||
|
// This permits the name to be the same as the package name.
|
||||||
|
return |
||||||
|
} |
||||||
|
if !strings.EqualFold(pkg, name[:len(pkg)]) { |
||||||
|
return |
||||||
|
} |
||||||
|
// We can assume the name is well-formed UTF-8.
|
||||||
|
// If the next rune after the package name is uppercase or an underscore
|
||||||
|
// the it's starting a new word and thus this name stutters.
|
||||||
|
rem := name[len(pkg):] |
||||||
|
if next, _ := utf8.DecodeRuneInString(rem); next == '_' || unicode.IsUpper(next) { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Node: id, |
||||||
|
Confidence: 0.8, |
||||||
|
Category: "naming", |
||||||
|
Failure: fmt.Sprintf("%s name will be used as %s.%s by other packages, and that stutters; consider calling this %s", thing, pkg, name, rem), |
||||||
|
}) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
func (w *lintExported) lintTypeDoc(t *ast.TypeSpec, doc *ast.CommentGroup) { |
||||||
|
if !ast.IsExported(t.Name.Name) { |
||||||
|
return |
||||||
|
} |
||||||
|
if doc == nil { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Node: t, |
||||||
|
Confidence: 1, |
||||||
|
Category: "comments", |
||||||
|
Failure: fmt.Sprintf("exported type %v should have comment or be unexported", t.Name), |
||||||
|
}) |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
s := normalizeText(doc.Text()) |
||||||
|
articles := [...]string{"A", "An", "The", "This"} |
||||||
|
for _, a := range articles { |
||||||
|
if t.Name.Name == a { |
||||||
|
continue |
||||||
|
} |
||||||
|
if strings.HasPrefix(s, a+" ") { |
||||||
|
s = s[len(a)+1:] |
||||||
|
break |
||||||
|
} |
||||||
|
} |
||||||
|
if !strings.HasPrefix(s, t.Name.Name+" ") { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Node: doc, |
||||||
|
Confidence: 1, |
||||||
|
Category: "comments", |
||||||
|
Failure: fmt.Sprintf(`comment on exported type %v should be of the form "%v ..." (with optional leading article)`, t.Name, t.Name), |
||||||
|
}) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
func (w *lintExported) lintValueSpecDoc(vs *ast.ValueSpec, gd *ast.GenDecl, genDeclMissingComments map[*ast.GenDecl]bool) { |
||||||
|
kind := "var" |
||||||
|
if gd.Tok == token.CONST { |
||||||
|
kind = "const" |
||||||
|
} |
||||||
|
|
||||||
|
if len(vs.Names) > 1 { |
||||||
|
// Check that none are exported except for the first.
|
||||||
|
for _, n := range vs.Names[1:] { |
||||||
|
if ast.IsExported(n.Name) { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Category: "comments", |
||||||
|
Confidence: 1, |
||||||
|
Failure: fmt.Sprintf("exported %s %s should have its own declaration", kind, n.Name), |
||||||
|
Node: vs, |
||||||
|
}) |
||||||
|
return |
||||||
|
} |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// Only one name.
|
||||||
|
name := vs.Names[0].Name |
||||||
|
if !ast.IsExported(name) { |
||||||
|
return |
||||||
|
} |
||||||
|
|
||||||
|
if vs.Doc == nil && gd.Doc == nil { |
||||||
|
if genDeclMissingComments[gd] { |
||||||
|
return |
||||||
|
} |
||||||
|
block := "" |
||||||
|
if kind == "const" && gd.Lparen.IsValid() { |
||||||
|
block = " (or a comment on this block)" |
||||||
|
} |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: vs, |
||||||
|
Category: "comments", |
||||||
|
Failure: fmt.Sprintf("exported %s %s should have comment%s or be unexported", kind, name, block), |
||||||
|
}) |
||||||
|
genDeclMissingComments[gd] = true |
||||||
|
return |
||||||
|
} |
||||||
|
// If this GenDecl has parens and a comment, we don't check its comment form.
|
||||||
|
if gd.Lparen.IsValid() && gd.Doc != nil { |
||||||
|
return |
||||||
|
} |
||||||
|
// The relevant text to check will be on either vs.Doc or gd.Doc.
|
||||||
|
// Use vs.Doc preferentially.
|
||||||
|
doc := vs.Doc |
||||||
|
if doc == nil { |
||||||
|
doc = gd.Doc |
||||||
|
} |
||||||
|
prefix := name + " " |
||||||
|
s := normalizeText(doc.Text()) |
||||||
|
if !strings.HasPrefix(s, prefix) { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: doc, |
||||||
|
Category: "comments", |
||||||
|
Failure: fmt.Sprintf(`comment on exported %s %s should be of the form "%s..."`, kind, name, prefix), |
||||||
|
}) |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
// normalizeText is a helper function that normalizes comment strings by:
|
||||||
|
// * removing one leading space
|
||||||
|
//
|
||||||
|
// This function is needed because ast.CommentGroup.Text() does not handle //-style and /*-style comments uniformly
|
||||||
|
func normalizeText(t string) string { |
||||||
|
return strings.TrimPrefix(t, " ") |
||||||
|
} |
||||||
|
|
||||||
|
func (w *lintExported) Visit(n ast.Node) ast.Visitor { |
||||||
|
switch v := n.(type) { |
||||||
|
case *ast.GenDecl: |
||||||
|
if v.Tok == token.IMPORT { |
||||||
|
return nil |
||||||
|
} |
||||||
|
// token.CONST, token.TYPE or token.VAR
|
||||||
|
w.lastGen = v |
||||||
|
return w |
||||||
|
case *ast.FuncDecl: |
||||||
|
w.lintFuncDoc(v) |
||||||
|
if v.Recv == nil { |
||||||
|
// Only check for stutter on functions, not methods.
|
||||||
|
// Method names are not used package-qualified.
|
||||||
|
w.checkStutter(v.Name, "func") |
||||||
|
} |
||||||
|
// Don't proceed inside funcs.
|
||||||
|
return nil |
||||||
|
case *ast.TypeSpec: |
||||||
|
// inside a GenDecl, which usually has the doc
|
||||||
|
doc := v.Doc |
||||||
|
if doc == nil { |
||||||
|
doc = w.lastGen.Doc |
||||||
|
} |
||||||
|
w.lintTypeDoc(v, doc) |
||||||
|
w.checkStutter(v.Name, "type") |
||||||
|
// Don't proceed inside types.
|
||||||
|
return nil |
||||||
|
case *ast.ValueSpec: |
||||||
|
w.lintValueSpecDoc(v, w.lastGen, w.genDeclMissingComments) |
||||||
|
return nil |
||||||
|
} |
||||||
|
return w |
||||||
|
} |
@ -0,0 +1,69 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"regexp" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// FileHeaderRule lints given else constructs.
|
||||||
|
type FileHeaderRule struct{} |
||||||
|
|
||||||
|
var ( |
||||||
|
multiRegexp = regexp.MustCompile("^/\\*") |
||||||
|
singleRegexp = regexp.MustCompile("^//") |
||||||
|
) |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *FileHeaderRule) Apply(file *lint.File, arguments lint.Arguments) []lint.Failure { |
||||||
|
if len(arguments) != 1 { |
||||||
|
panic(`invalid configuration for "file-header" rule`) |
||||||
|
} |
||||||
|
|
||||||
|
header, ok := arguments[0].(string) |
||||||
|
if !ok { |
||||||
|
panic(`invalid argument for "file-header" rule: first argument should be a string`) |
||||||
|
} |
||||||
|
|
||||||
|
failure := []lint.Failure{ |
||||||
|
{ |
||||||
|
Node: file.AST, |
||||||
|
Confidence: 1, |
||||||
|
Failure: "the file doesn't have an appropriate header", |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
if len(file.AST.Comments) == 0 { |
||||||
|
return failure |
||||||
|
} |
||||||
|
|
||||||
|
g := file.AST.Comments[0] |
||||||
|
if g == nil { |
||||||
|
return failure |
||||||
|
} |
||||||
|
comment := "" |
||||||
|
for _, c := range g.List { |
||||||
|
text := c.Text |
||||||
|
if multiRegexp.Match([]byte(text)) { |
||||||
|
text = text[2 : len(text)-2] |
||||||
|
} else if singleRegexp.Match([]byte(text)) { |
||||||
|
text = text[2:] |
||||||
|
} |
||||||
|
comment += text |
||||||
|
} |
||||||
|
|
||||||
|
regex, err := regexp.Compile(header) |
||||||
|
if err != nil { |
||||||
|
panic(err.Error()) |
||||||
|
} |
||||||
|
|
||||||
|
if !regex.Match([]byte(comment)) { |
||||||
|
return failure |
||||||
|
} |
||||||
|
return nil |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *FileHeaderRule) Name() string { |
||||||
|
return "file-header" |
||||||
|
} |
@ -0,0 +1,104 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
"go/ast" |
||||||
|
) |
||||||
|
|
||||||
|
// FlagParamRule lints given else constructs.
|
||||||
|
type FlagParamRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *FlagParamRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
onFailure := func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
} |
||||||
|
|
||||||
|
w := lintFlagParamRule{onFailure: onFailure} |
||||||
|
ast.Walk(w, file.AST) |
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *FlagParamRule) Name() string { |
||||||
|
return "flag-parameter" |
||||||
|
} |
||||||
|
|
||||||
|
type lintFlagParamRule struct { |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintFlagParamRule) Visit(node ast.Node) ast.Visitor { |
||||||
|
fd, ok := node.(*ast.FuncDecl) |
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
if fd.Body == nil { |
||||||
|
return nil // skip whole function declaration
|
||||||
|
} |
||||||
|
|
||||||
|
for _, p := range fd.Type.Params.List { |
||||||
|
t := p.Type |
||||||
|
|
||||||
|
id, ok := t.(*ast.Ident) |
||||||
|
if !ok { |
||||||
|
continue |
||||||
|
} |
||||||
|
|
||||||
|
if id.Name != "bool" { |
||||||
|
continue |
||||||
|
} |
||||||
|
|
||||||
|
cv := conditionVisitor{p.Names, fd, w} |
||||||
|
ast.Walk(cv, fd.Body) |
||||||
|
} |
||||||
|
|
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
type conditionVisitor struct { |
||||||
|
ids []*ast.Ident |
||||||
|
fd *ast.FuncDecl |
||||||
|
linter lintFlagParamRule |
||||||
|
} |
||||||
|
|
||||||
|
func (w conditionVisitor) Visit(node ast.Node) ast.Visitor { |
||||||
|
ifStmt, ok := node.(*ast.IfStmt) |
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
fselect := func(n ast.Node) bool { |
||||||
|
ident, ok := n.(*ast.Ident) |
||||||
|
if !ok { |
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
for _, id := range w.ids { |
||||||
|
if ident.Name == id.Name { |
||||||
|
return true |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
uses := pick(ifStmt.Cond, fselect, nil) |
||||||
|
|
||||||
|
if len(uses) < 1 { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
w.linter.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Node: w.fd.Type.Params, |
||||||
|
Category: "bad practice", |
||||||
|
Failure: fmt.Sprintf("parameter '%s' seems to be a control flag, avoid control coupling", uses[0]), |
||||||
|
}) |
||||||
|
|
||||||
|
return nil |
||||||
|
} |
@ -0,0 +1,68 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"go/ast" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// FunctionResultsLimitRule lints given else constructs.
|
||||||
|
type FunctionResultsLimitRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *FunctionResultsLimitRule) Apply(file *lint.File, arguments lint.Arguments) []lint.Failure { |
||||||
|
if len(arguments) != 1 { |
||||||
|
panic(`invalid configuration for "function-result-limit"`) |
||||||
|
} |
||||||
|
|
||||||
|
max, ok := arguments[0].(int64) // Alt. non panicking version
|
||||||
|
if !ok { |
||||||
|
panic(fmt.Sprintf(`invalid value passed as return results number to the "function-result-limit" rule; need int64 but got %T`, arguments[0])) |
||||||
|
} |
||||||
|
if max < 0 { |
||||||
|
panic(`the value passed as return results number to the "function-result-limit" rule cannot be negative`) |
||||||
|
} |
||||||
|
|
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
walker := lintFunctionResultsNum{ |
||||||
|
max: int(max), |
||||||
|
onFailure: func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
}, |
||||||
|
} |
||||||
|
|
||||||
|
ast.Walk(walker, file.AST) |
||||||
|
|
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *FunctionResultsLimitRule) Name() string { |
||||||
|
return "function-result-limit" |
||||||
|
} |
||||||
|
|
||||||
|
type lintFunctionResultsNum struct { |
||||||
|
max int |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintFunctionResultsNum) Visit(n ast.Node) ast.Visitor { |
||||||
|
node, ok := n.(*ast.FuncDecl) |
||||||
|
if ok { |
||||||
|
num := 0 |
||||||
|
if node.Type.Results != nil { |
||||||
|
num = node.Type.Results.NumFields() |
||||||
|
} |
||||||
|
if num > w.max { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 1, |
||||||
|
Failure: fmt.Sprintf("maximum number of return results per function exceeded; max %d but got %d", w.max, num), |
||||||
|
Node: node.Type, |
||||||
|
}) |
||||||
|
return w |
||||||
|
} |
||||||
|
} |
||||||
|
return w |
||||||
|
} |
@ -0,0 +1,70 @@ |
|||||||
|
package rule |
||||||
|
|
||||||
|
import ( |
||||||
|
"fmt" |
||||||
|
"go/ast" |
||||||
|
"strings" |
||||||
|
|
||||||
|
"github.com/mgechev/revive/lint" |
||||||
|
) |
||||||
|
|
||||||
|
// GetReturnRule lints given else constructs.
|
||||||
|
type GetReturnRule struct{} |
||||||
|
|
||||||
|
// Apply applies the rule to given file.
|
||||||
|
func (r *GetReturnRule) Apply(file *lint.File, _ lint.Arguments) []lint.Failure { |
||||||
|
var failures []lint.Failure |
||||||
|
|
||||||
|
onFailure := func(failure lint.Failure) { |
||||||
|
failures = append(failures, failure) |
||||||
|
} |
||||||
|
|
||||||
|
w := lintReturnRule{onFailure} |
||||||
|
ast.Walk(w, file.AST) |
||||||
|
return failures |
||||||
|
} |
||||||
|
|
||||||
|
// Name returns the rule name.
|
||||||
|
func (r *GetReturnRule) Name() string { |
||||||
|
return "get-return" |
||||||
|
} |
||||||
|
|
||||||
|
type lintReturnRule struct { |
||||||
|
onFailure func(lint.Failure) |
||||||
|
} |
||||||
|
|
||||||
|
func isGetter(name string) bool { |
||||||
|
if strings.HasPrefix(strings.ToUpper(name), "GET") { |
||||||
|
if len(name) > 3 { |
||||||
|
c := name[3] |
||||||
|
return !(c >= 'a' && c <= 'z') |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
return false |
||||||
|
} |
||||||
|
|
||||||
|
func hasResults(rs *ast.FieldList) bool { |
||||||
|
return rs != nil && len(rs.List) > 0 |
||||||
|
} |
||||||
|
|
||||||
|
func (w lintReturnRule) Visit(node ast.Node) ast.Visitor { |
||||||
|
fd, ok := node.(*ast.FuncDecl) |
||||||
|
if !ok { |
||||||
|
return w |
||||||
|
} |
||||||
|
|
||||||
|
if !isGetter(fd.Name.Name) { |
||||||
|
return w |
||||||
|
} |
||||||
|
if !hasResults(fd.Type.Results) { |
||||||
|
w.onFailure(lint.Failure{ |
||||||
|
Confidence: 0.8, |
||||||
|
Node: fd, |
||||||
|
Category: "logic", |
||||||
|
Failure: fmt.Sprintf("function '%s' seems to be a getter but it does not return any result", fd.Name.Name), |
||||||
|
}) |
||||||
|
} |
||||||
|
|
||||||
|
return w |
||||||
|
} |
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in new issue