ref: 67df33f50069445d53d373790e8c636e47013a72
parent: 00f07c5374d859057c6e7c4731000b5978364819
author: Anthony Fok <[email protected]>
date: Wed Mar 11 07:34:57 EDT 2015
Correct initialisms as suggested by golint First step to use initialisms that golint suggests, for example: Line 116: func GetHtmlRenderer should be GetHTMLRenderer as see on http://goreportcard.com/report/spf13/hugo Thanks to @bep for the idea! Note that command-line flags (cobra and pflag) as well as struct fields like .BaseUrl and .Url that are used in Go HTML templates need more work to maintain backward-compatibility, and thus are NOT yet dealt with in this commit. First step in fixing #959.
--- a/commands/hugo.go
+++ b/commands/hugo.go
@@ -56,8 +56,8 @@
var hugoCmdV *cobra.Command
//Flags that are to be added to commands.
-var BuildWatch, IgnoreCache, Draft, Future, UglyUrls, Verbose, Logging, VerboseLog, DisableRSS, DisableSitemap, PluralizeListTitles, NoTimes bool
-var Source, CacheDir, Destination, Theme, BaseUrl, CfgFile, LogFile, Editor string
+var BuildWatch, IgnoreCache, Draft, Future, UglyURLs, Verbose, Logging, VerboseLog, DisableRSS, DisableSitemap, PluralizeListTitles, NoTimes bool
+var Source, CacheDir, Destination, Theme, BaseURL, CfgFile, LogFile, Editor string
//Execute adds all child commands to the root command HugoCmd and sets flags appropriately.
func Execute() {
@@ -89,8 +89,8 @@
HugoCmd.PersistentFlags().StringVarP(&Destination, "destination", "d", "", "filesystem path to write files to")
HugoCmd.PersistentFlags().StringVarP(&Theme, "theme", "t", "", "theme to use (located in /themes/THEMENAME/)")
HugoCmd.PersistentFlags().BoolVarP(&Verbose, "verbose", "v", false, "verbose output")
- HugoCmd.PersistentFlags().BoolVar(&UglyUrls, "uglyUrls", false, "if true, use /filename.html instead of /filename/")
- HugoCmd.PersistentFlags().StringVarP(&BaseUrl, "baseUrl", "b", "", "hostname (and path) to the root eg. http://spf13.com/")
+ HugoCmd.PersistentFlags().BoolVar(&UglyURLs, "uglyUrls", false, "if true, use /filename.html instead of /filename/")
+ HugoCmd.PersistentFlags().StringVarP(&BaseURL, "baseUrl", "b", "", "hostname (and path) to the root eg. http://spf13.com/")
HugoCmd.PersistentFlags().StringVar(&CfgFile, "config", "", "config file (default is path/config.yaml|json|toml)")
HugoCmd.PersistentFlags().StringVar(&Editor, "editor", "", "edit new content with this editor, if provided")
HugoCmd.PersistentFlags().BoolVar(&Logging, "log", false, "Enable Logging")
@@ -127,10 +127,10 @@
viper.SetDefault("DefaultLayout", "post")
viper.SetDefault("BuildDrafts", false)
viper.SetDefault("BuildFuture", false)
- viper.SetDefault("UglyUrls", false)
+ viper.SetDefault("UglyURLs", false)
viper.SetDefault("Verbose", false)
viper.SetDefault("IgnoreCache", false)
- viper.SetDefault("CanonifyUrls", false)
+ viper.SetDefault("CanonifyURLs", false)
viper.SetDefault("Taxonomies", map[string]string{"tag": "tags", "category": "categories"})
viper.SetDefault("Permalinks", make(hugolib.PermalinkOverrides, 0))
viper.SetDefault("Sitemap", hugolib.Sitemap{Priority: -1})
@@ -155,7 +155,7 @@
}
if hugoCmdV.PersistentFlags().Lookup("uglyUrls").Changed {
- viper.Set("UglyUrls", UglyUrls)
+ viper.Set("UglyURLs", UglyURLs)
}
if hugoCmdV.PersistentFlags().Lookup("disableRSS").Changed {
@@ -181,14 +181,14 @@
if hugoCmdV.PersistentFlags().Lookup("logFile").Changed {
viper.Set("LogFile", LogFile)
}
- if BaseUrl != "" {
- if !strings.HasSuffix(BaseUrl, "/") {
- BaseUrl = BaseUrl + "/"
+ if BaseURL != "" {
+ if !strings.HasSuffix(BaseURL, "/") {
+ BaseURL = BaseURL + "/"
}
- viper.Set("BaseUrl", BaseUrl)
+ viper.Set("BaseURL", BaseURL)
}
- if viper.GetString("BaseUrl") == "" {
+ if viper.GetString("BaseURL") == "" {
jww.ERROR.Println("No 'baseurl' set in configuration or as a flag. Features like page menus will not work without one.")
}
--- a/commands/server.go
+++ b/commands/server.go
@@ -84,11 +84,11 @@
viper.Set("port", serverPort)
- BaseUrl, err := fixUrl(BaseUrl)
+ BaseURL, err := fixURL(BaseURL)
if err != nil {
jww.ERROR.Fatal(err)
}
- viper.Set("BaseUrl", BaseUrl)
+ viper.Set("BaseURL", BaseURL)
if err := memStats(); err != nil {
jww.ERROR.Println("memstats error:", err)
@@ -114,9 +114,9 @@
httpFs := &afero.HttpFs{SourceFs: hugofs.DestinationFS}
fileserver := http.FileServer(httpFs.Dir(helpers.AbsPathify(viper.GetString("PublishDir"))))
- u, err := url.Parse(viper.GetString("BaseUrl"))
+ u, err := url.Parse(viper.GetString("BaseURL"))
if err != nil {
- jww.ERROR.Fatalf("Invalid BaseUrl: %s", err)
+ jww.ERROR.Fatalf("Invalid BaseURL: %s", err)
}
if u.Path == "" || u.Path == "/" {
http.Handle("/", fileserver)
@@ -137,10 +137,10 @@
// fixUrl massages the BaseUrl into a form needed for serving
// all pages correctly.
-func fixUrl(s string) (string, error) {
+func fixURL(s string) (string, error) {
useLocalhost := false
if s == "" {
- s = viper.GetString("BaseUrl")
+ s = viper.GetString("BaseURL")
useLocalhost = true
}
if !strings.HasPrefix(s, "http://") && !strings.HasPrefix(s, "https://") {
--- a/commands/server_test.go
+++ b/commands/server_test.go
@@ -6,11 +6,11 @@
"github.com/spf13/viper"
)
-func TestFixUrl(t *testing.T) {
+func TestFixURL(t *testing.T) {
type data struct {
TestName string
- CliBaseUrl string
- CfgBaseUrl string
+ CLIBaseURL string
+ CfgBaseURL string
AppendPort bool
Port int
Result string
@@ -28,11 +28,11 @@
}
for i, test := range tests {
- BaseUrl = test.CliBaseUrl
- viper.Set("BaseUrl", test.CfgBaseUrl)
+ BaseURL = test.CLIBaseURL
+ viper.Set("BaseURL", test.CfgBaseURL)
serverAppend = test.AppendPort
serverPort = test.Port
- result, err := fixUrl(BaseUrl)
+ result, err := fixURL(BaseURL)
if err != nil {
t.Errorf("Test #%d %s: unexpected error %s", i, test.TestName, err)
}
--- a/helpers/content.go
+++ b/helpers/content.go
@@ -41,7 +41,7 @@
type Blackfriday struct {
AngledQuotes bool
Fractions bool
- PlainIdAnchors bool
+ PlainIDAnchors bool
Extensions []string
}
@@ -50,7 +50,7 @@
return &Blackfriday{
AngledQuotes: false,
Fractions: true,
- PlainIdAnchors: false,
+ PlainIDAnchors: false,
}
}
@@ -113,17 +113,17 @@
}
// GetHtmlRenderer creates a new Renderer with the given configuration.
-func GetHtmlRenderer(defaultFlags int, ctx *RenderingContext) blackfriday.Renderer {
+func GetHTMLRenderer(defaultFlags int, ctx *RenderingContext) blackfriday.Renderer {
renderParameters := blackfriday.HtmlRendererParameters{
FootnoteAnchorPrefix: viper.GetString("FootnoteAnchorPrefix"),
FootnoteReturnLinkContents: viper.GetString("FootnoteReturnLinkContents"),
}
- b := len(ctx.DocumentId) != 0
+ b := len(ctx.DocumentID) != 0
- if b && !ctx.getConfig().PlainIdAnchors {
- renderParameters.FootnoteAnchorPrefix = ctx.DocumentId + ":" + renderParameters.FootnoteAnchorPrefix
- renderParameters.HeaderIDSuffix = ":" + ctx.DocumentId
+ if b && !ctx.getConfig().PlainIDAnchors {
+ renderParameters.FootnoteAnchorPrefix = ctx.DocumentID + ":" + renderParameters.FootnoteAnchorPrefix
+ renderParameters.HeaderIDSuffix = ":" + ctx.DocumentID
}
htmlFlags := defaultFlags
@@ -158,13 +158,13 @@
}
func markdownRender(ctx *RenderingContext) []byte {
- return blackfriday.Markdown(ctx.Content, GetHtmlRenderer(0, ctx),
+ return blackfriday.Markdown(ctx.Content, GetHTMLRenderer(0, ctx),
getMarkdownExtensions(ctx))
}
func markdownRenderWithTOC(ctx *RenderingContext) []byte {
return blackfriday.Markdown(ctx.Content,
- GetHtmlRenderer(blackfriday.HTML_TOC, ctx),
+ GetHTMLRenderer(blackfriday.HTML_TOC, ctx),
getMarkdownExtensions(ctx))
}
@@ -209,7 +209,7 @@
type RenderingContext struct {
Content []byte
PageFmt string
- DocumentId string
+ DocumentID string
Config *Blackfriday
configInit sync.Once
}
--- a/helpers/content_test.go
+++ b/helpers/content_test.go
@@ -7,7 +7,7 @@
"testing"
)
-const tstHtmlContent = "<!DOCTYPE html><html><head><script src=\"http://two/foobar.js\"></script></head><body><nav><ul><li hugo-nav=\"section_0\"></li><li hugo-nav=\"section_1\"></li></ul></nav><article>content <a href=\"http://two/foobar\">foobar</a>. Follow up</article><p>This is some text.<br>And some more.</p></body></html>"
+const tstHTMLContent = "<!DOCTYPE html><html><head><script src=\"http://two/foobar.js\"></script></head><body><nav><ul><li hugo-nav=\"section_0\"></li><li hugo-nav=\"section_1\"></li></ul></nav><article>content <a href=\"http://two/foobar\">foobar</a>. Follow up</article><p>This is some text.<br>And some more.</p></body></html>"
func TestStripHTML(t *testing.T) {
type test struct {
@@ -31,7 +31,7 @@
func BenchmarkStripHTML(b *testing.B) {
b.ResetTimer()
for i := 0; i < b.N; i++ {
- StripHTML(tstHtmlContent)
+ StripHTML(tstHTMLContent)
}
}
--- a/helpers/url.go
+++ b/helpers/url.go
@@ -52,7 +52,7 @@
var pathBridge PathBridge
-func sanitizeUrlWithFlags(in string, f purell.NormalizationFlags) string {
+func sanitizeURLWithFlags(in string, f purell.NormalizationFlags) string {
s, err := purell.NormalizeURLString(in, f)
if err != nil {
return in
@@ -88,13 +88,13 @@
}
// SanitizeUrl sanitizes the input URL string.
-func SanitizeUrl(in string) string {
- return sanitizeUrlWithFlags(in, purell.FlagsSafe|purell.FlagRemoveTrailingSlash|purell.FlagRemoveDotSegments|purell.FlagRemoveDuplicateSlashes|purell.FlagRemoveUnnecessaryHostDots|purell.FlagRemoveEmptyPortSeparator)
+func SanitizeURL(in string) string {
+ return sanitizeURLWithFlags(in, purell.FlagsSafe|purell.FlagRemoveTrailingSlash|purell.FlagRemoveDotSegments|purell.FlagRemoveDuplicateSlashes|purell.FlagRemoveUnnecessaryHostDots|purell.FlagRemoveEmptyPortSeparator)
}
// SanitizeUrlKeepTrailingSlash is the same as SanitizeUrl, but will keep any trailing slash.
-func SanitizeUrlKeepTrailingSlash(in string) string {
- return sanitizeUrlWithFlags(in, purell.FlagsSafe|purell.FlagRemoveDotSegments|purell.FlagRemoveDuplicateSlashes|purell.FlagRemoveUnnecessaryHostDots|purell.FlagRemoveEmptyPortSeparator)
+func SanitizeURLKeepTrailingSlash(in string) string {
+ return sanitizeURLWithFlags(in, purell.FlagsSafe|purell.FlagRemoveDotSegments|purell.FlagRemoveDuplicateSlashes|purell.FlagRemoveUnnecessaryHostDots|purell.FlagRemoveEmptyPortSeparator)
}
// Similar to MakePath, but with Unicode handling
@@ -101,7 +101,7 @@
// Example:
// uri: Vim (text editor)
// urlize: vim-text-editor
-func Urlize(uri string) string {
+func URLize(uri string) string {
sanitized := MakePathToLower(uri)
// escape unicode letters
@@ -148,9 +148,9 @@
// AddContextRoot adds the context root to an URL if it's not already set.
// For relative URL entries on sites with a base url with a context root set (i.e. http://example.com/mysite),
// relative URLs must not include the context root if canonifyUrls is enabled. But if it's disabled, it must be set.
-func AddContextRoot(baseUrl, relativePath string) string {
+func AddContextRoot(baseURL, relativePath string) string {
- url, err := url.Parse(baseUrl)
+ url, err := url.Parse(baseURL)
if err != nil {
panic(err)
}
@@ -164,16 +164,16 @@
return newPath
}
-func UrlizeAndPrep(in string) string {
- return UrlPrep(viper.GetBool("UglyUrls"), Urlize(in))
+func URLizeAndPrep(in string) string {
+ return URLPrep(viper.GetBool("UglyURLs"), URLize(in))
}
-func UrlPrep(ugly bool, in string) string {
+func URLPrep(ugly bool, in string) string {
if ugly {
- x := Uglify(SanitizeUrl(in))
+ x := Uglify(SanitizeURL(in))
return x
}
- x := PrettifyUrl(SanitizeUrl(in))
+ x := PrettifyURL(SanitizeURL(in))
if path.Ext(x) == ".xml" {
return x
}
@@ -186,8 +186,8 @@
}
// PrettifyUrl takes a URL string and returns a semantic, clean URL.
-func PrettifyUrl(in string) string {
- x := PrettifyUrlPath(in)
+func PrettifyURL(in string) string {
+ x := PrettifyURLPath(in)
if path.Base(x) == "index.html" {
return path.Dir(x)
@@ -205,7 +205,7 @@
// /section/name.html becomes /section/name/index.html
// /section/name/ becomes /section/name/index.html
// /section/name/index.html becomes /section/name/index.html
-func PrettifyUrlPath(in string) string {
+func PrettifyURLPath(in string) string {
return PrettiyPath(in, pathBridge)
}
--- a/helpers/url_test.go
+++ b/helpers/url_test.go
@@ -19,7 +19,7 @@
}
for _, test := range tests {
- output := Urlize(test.input)
+ output := URLize(test.input)
if output != test.expected {
t.Errorf("Expected %#v, got %#v\n", test.expected, output)
}
@@ -36,8 +36,8 @@
}
for _, test := range tests {
- o1 := SanitizeUrl(test.input)
- o2 := SanitizeUrlKeepTrailingSlash(test.input)
+ o1 := SanitizeURL(test.input)
+ o2 := SanitizeURLKeepTrailingSlash(test.input)
expected2 := test.expected
@@ -88,7 +88,7 @@
{true, "/section/name/index.html", "/section/name.html"},
}
for i, d := range data {
- output := UrlPrep(d.ugly, d.input)
+ output := URLPrep(d.ugly, d.input)
if d.output != output {
t.Errorf("Test #%d failed. Expected %q got %q", i, d.output, output)
}
@@ -98,7 +98,7 @@
func TestAddContextRoot(t *testing.T) {
tests := []struct {
- baseUrl string
+ baseURL string
url string
expected string
}{
@@ -114,7 +114,7 @@
}
for _, test := range tests {
- output := AddContextRoot(test.baseUrl, test.url)
+ output := AddContextRoot(test.baseURL, test.url)
if output != test.expected {
t.Errorf("Expected %#v, got %#v\n", test.expected, output)
}
@@ -122,22 +122,22 @@
}
func TestPretty(t *testing.T) {
- assert.Equal(t, PrettifyUrlPath("/section/name.html"), "/section/name/index.html")
- assert.Equal(t, PrettifyUrlPath("/section/sub/name.html"), "/section/sub/name/index.html")
- assert.Equal(t, PrettifyUrlPath("/section/name/"), "/section/name/index.html")
- assert.Equal(t, PrettifyUrlPath("/section/name/index.html"), "/section/name/index.html")
- assert.Equal(t, PrettifyUrlPath("/index.html"), "/index.html")
- assert.Equal(t, PrettifyUrlPath("/name.xml"), "/name/index.xml")
- assert.Equal(t, PrettifyUrlPath("/"), "/")
- assert.Equal(t, PrettifyUrlPath(""), "/")
- assert.Equal(t, PrettifyUrl("/section/name.html"), "/section/name")
- assert.Equal(t, PrettifyUrl("/section/sub/name.html"), "/section/sub/name")
- assert.Equal(t, PrettifyUrl("/section/name/"), "/section/name")
- assert.Equal(t, PrettifyUrl("/section/name/index.html"), "/section/name")
- assert.Equal(t, PrettifyUrl("/index.html"), "/")
- assert.Equal(t, PrettifyUrl("/name.xml"), "/name/index.xml")
- assert.Equal(t, PrettifyUrl("/"), "/")
- assert.Equal(t, PrettifyUrl(""), "/")
+ assert.Equal(t, PrettifyURLPath("/section/name.html"), "/section/name/index.html")
+ assert.Equal(t, PrettifyURLPath("/section/sub/name.html"), "/section/sub/name/index.html")
+ assert.Equal(t, PrettifyURLPath("/section/name/"), "/section/name/index.html")
+ assert.Equal(t, PrettifyURLPath("/section/name/index.html"), "/section/name/index.html")
+ assert.Equal(t, PrettifyURLPath("/index.html"), "/index.html")
+ assert.Equal(t, PrettifyURLPath("/name.xml"), "/name/index.xml")
+ assert.Equal(t, PrettifyURLPath("/"), "/")
+ assert.Equal(t, PrettifyURLPath(""), "/")
+ assert.Equal(t, PrettifyURL("/section/name.html"), "/section/name")
+ assert.Equal(t, PrettifyURL("/section/sub/name.html"), "/section/sub/name")
+ assert.Equal(t, PrettifyURL("/section/name/"), "/section/name")
+ assert.Equal(t, PrettifyURL("/section/name/index.html"), "/section/name")
+ assert.Equal(t, PrettifyURL("/index.html"), "/")
+ assert.Equal(t, PrettifyURL("/name.xml"), "/name/index.xml")
+ assert.Equal(t, PrettifyURL("/"), "/")
+ assert.Equal(t, PrettifyURL(""), "/")
}
func TestUgly(t *testing.T) {
--- a/hugolib/menu_test.go
+++ b/hugolib/menu_test.go
@@ -95,7 +95,7 @@
{"sect/doc3.md", MENU_PAGE_3},
}
-func tstCreateMenuPageWithNameToml(title, menu, name string) []byte {
+func tstCreateMenuPageWithNameTOML(title, menu, name string) []byte {
return []byte(fmt.Sprintf(`+++
title = "%s"
weight = 1
@@ -106,7 +106,7 @@
Front Matter with Menu with Name`, title, menu, name))
}
-func tstCreateMenuPageWithIdentifierToml(title, menu, identifier string) []byte {
+func tstCreateMenuPageWithIdentifierTOML(title, menu, identifier string) []byte {
return []byte(fmt.Sprintf(`+++
title = "%s"
weight = 1
@@ -118,7 +118,7 @@
Front Matter with Menu with Identifier`, title, menu, identifier))
}
-func tstCreateMenuPageWithNameYaml(title, menu, name string) []byte {
+func tstCreateMenuPageWithNameYAML(title, menu, name string) []byte {
return []byte(fmt.Sprintf(`---
title: "%s"
weight: 1
@@ -129,7 +129,7 @@
Front Matter with Menu with Name`, title, menu, name))
}
-func tstCreateMenuPageWithIdentifierYaml(title, menu, identifier string) []byte {
+func tstCreateMenuPageWithIdentifierYAML(title, menu, identifier string) []byte {
return []byte(fmt.Sprintf(`---
title: "%s"
weight: 1
@@ -144,7 +144,7 @@
type testMenuState struct {
site *Site
oldMenu interface{}
- oldBaseUrl interface{}
+ oldBaseURL interface{}
}
// Issue 817 - identifier should trump everything
@@ -151,15 +151,15 @@
func TestPageMenuWithIdentifier(t *testing.T) {
toml := []source.ByteSource{
- {"sect/doc1.md", tstCreateMenuPageWithIdentifierToml("t1", "m1", "i1")},
- {"sect/doc2.md", tstCreateMenuPageWithIdentifierToml("t1", "m1", "i2")},
- {"sect/doc3.md", tstCreateMenuPageWithIdentifierToml("t1", "m1", "i2")}, // duplicate
+ {"sect/doc1.md", tstCreateMenuPageWithIdentifierTOML("t1", "m1", "i1")},
+ {"sect/doc2.md", tstCreateMenuPageWithIdentifierTOML("t1", "m1", "i2")},
+ {"sect/doc3.md", tstCreateMenuPageWithIdentifierTOML("t1", "m1", "i2")}, // duplicate
}
yaml := []source.ByteSource{
- {"sect/doc1.md", tstCreateMenuPageWithIdentifierYaml("t1", "m1", "i1")},
- {"sect/doc2.md", tstCreateMenuPageWithIdentifierYaml("t1", "m1", "i2")},
- {"sect/doc3.md", tstCreateMenuPageWithIdentifierYaml("t1", "m1", "i2")}, // duplicate
+ {"sect/doc1.md", tstCreateMenuPageWithIdentifierYAML("t1", "m1", "i1")},
+ {"sect/doc2.md", tstCreateMenuPageWithIdentifierYAML("t1", "m1", "i2")},
+ {"sect/doc3.md", tstCreateMenuPageWithIdentifierYAML("t1", "m1", "i2")}, // duplicate
}
doTestPageMenuWithIdentifier(t, toml)
@@ -174,8 +174,8 @@
assert.Equal(t, 3, len(ts.site.Pages), "Not enough pages")
- me1 := ts.findTestMenuEntryById("m1", "i1")
- me2 := ts.findTestMenuEntryById("m1", "i2")
+ me1 := ts.findTestMenuEntryByID("m1", "i1")
+ me2 := ts.findTestMenuEntryByID("m1", "i2")
assert.NotNil(t, me1)
assert.NotNil(t, me2)
@@ -188,15 +188,15 @@
// Issue 817 contd - name should be second identifier in
func TestPageMenuWithDuplicateName(t *testing.T) {
toml := []source.ByteSource{
- {"sect/doc1.md", tstCreateMenuPageWithNameToml("t1", "m1", "n1")},
- {"sect/doc2.md", tstCreateMenuPageWithNameToml("t1", "m1", "n2")},
- {"sect/doc3.md", tstCreateMenuPageWithNameToml("t1", "m1", "n2")}, // duplicate
+ {"sect/doc1.md", tstCreateMenuPageWithNameTOML("t1", "m1", "n1")},
+ {"sect/doc2.md", tstCreateMenuPageWithNameTOML("t1", "m1", "n2")},
+ {"sect/doc3.md", tstCreateMenuPageWithNameTOML("t1", "m1", "n2")}, // duplicate
}
yaml := []source.ByteSource{
- {"sect/doc1.md", tstCreateMenuPageWithNameYaml("t1", "m1", "n1")},
- {"sect/doc2.md", tstCreateMenuPageWithNameYaml("t1", "m1", "n2")},
- {"sect/doc3.md", tstCreateMenuPageWithNameYaml("t1", "m1", "n2")}, // duplicate
+ {"sect/doc1.md", tstCreateMenuPageWithNameYAML("t1", "m1", "n1")},
+ {"sect/doc2.md", tstCreateMenuPageWithNameYAML("t1", "m1", "n2")},
+ {"sect/doc3.md", tstCreateMenuPageWithNameYAML("t1", "m1", "n2")}, // duplicate
}
doTestPageMenuWithDuplicateName(t, toml)
@@ -234,7 +234,7 @@
third := ts.site.Pages[2]
pOne := ts.findTestMenuEntryByName("p_one", "One")
- pTwo := ts.findTestMenuEntryById("p_two", "Two")
+ pTwo := ts.findTestMenuEntryByID("p_two", "Two")
for i, this := range []struct {
menu string
@@ -267,11 +267,11 @@
}
// issue #888
-func TestMenuWithHashInUrl(t *testing.T) {
+func TestMenuWithHashInURL(t *testing.T) {
ts := setupMenuTests(t, MENU_PAGE_SOURCES)
defer resetMenuTestState(ts)
- me := ts.findTestMenuEntryById("hash", "hash")
+ me := ts.findTestMenuEntryByID("hash", "hash")
assert.NotNil(t, me)
@@ -279,41 +279,41 @@
}
// issue #719
-func TestMenuWithUnicodeUrls(t *testing.T) {
- for _, uglyUrls := range []bool{true, false} {
- for _, canonifyUrls := range []bool{true, false} {
- doTestMenuWithUnicodeUrls(t, canonifyUrls, uglyUrls)
+func TestMenuWithUnicodeURLs(t *testing.T) {
+ for _, uglyURLs := range []bool{true, false} {
+ for _, canonifyURLs := range []bool{true, false} {
+ doTestMenuWithUnicodeURLs(t, canonifyURLs, uglyURLs)
}
}
}
-func doTestMenuWithUnicodeUrls(t *testing.T, canonifyUrls, uglyUrls bool) {
- viper.Set("CanonifyUrls", canonifyUrls)
- viper.Set("UglyUrls", uglyUrls)
+func doTestMenuWithUnicodeURLs(t *testing.T, canonifyURLs, uglyURLs bool) {
+ viper.Set("CanonifyURLs", canonifyURLs)
+ viper.Set("UglyURLs", uglyURLs)
ts := setupMenuTests(t, MENU_PAGE_SOURCES)
defer resetMenuTestState(ts)
- unicodeRussian := ts.findTestMenuEntryById("unicode", "unicode-russian")
+ unicodeRussian := ts.findTestMenuEntryByID("unicode", "unicode-russian")
expectedBase := "/%D0%BD%D0%BE%D0%B2%D0%BE%D1%81%D1%82%D0%B8-%D0%BF%D1%80%D0%BE%D0%B5%D0%BA%D1%82%D0%B0"
- if !canonifyUrls {
+ if !canonifyURLs {
expectedBase = "/Zoo" + expectedBase
}
var expected string
- if uglyUrls {
+ if uglyURLs {
expected = expectedBase + ".html"
} else {
expected = expectedBase + "/"
}
- assert.Equal(t, expected, unicodeRussian.Url, "uglyUrls[%t]", uglyUrls)
+ assert.Equal(t, expected, unicodeRussian.Url, "uglyURLs[%t]", uglyURLs)
}
func TestTaxonomyNodeMenu(t *testing.T) {
- viper.Set("CanonifyUrls", true)
+ viper.Set("CanonifyURLs", true)
ts := setupMenuTests(t, MENU_PAGE_SOURCES)
defer resetMenuTestState(ts)
@@ -325,9 +325,9 @@
hasMenuCurrent bool
}{
{"tax", taxRenderInfo{key: "key", singular: "one", plural: "two"},
- ts.findTestMenuEntryById("tax", "1"), true, false},
+ ts.findTestMenuEntryByID("tax", "1"), true, false},
{"tax", taxRenderInfo{key: "key", singular: "one", plural: "two"},
- ts.findTestMenuEntryById("tax", "2"), true, false},
+ ts.findTestMenuEntryByID("tax", "2"), true, false},
{"tax", taxRenderInfo{key: "key", singular: "one", plural: "two"},
&MenuEntry{Name: "Somewhere else", Url: "/somewhereelse"}, false, false},
} {
@@ -347,9 +347,9 @@
}
- menuEntryXml := ts.findTestMenuEntryById("tax", "xml")
+ menuEntryXML := ts.findTestMenuEntryByID("tax", "xml")
- if strings.HasSuffix(menuEntryXml.Url, "/") {
+ if strings.HasSuffix(menuEntryXML.Url, "/") {
t.Error("RSS menu item should not be padded with trailing slash")
}
}
@@ -370,9 +370,9 @@
{"main", homeMenuEntry, true, false},
{"doesnotexist", homeMenuEntry, false, false},
{"main", &MenuEntry{Name: "Somewhere else", Url: "/somewhereelse"}, false, false},
- {"grandparent", ts.findTestMenuEntryById("grandparent", "grandparentId"), false, false},
- {"grandparent", ts.findTestMenuEntryById("grandparent", "parentId"), false, true},
- {"grandparent", ts.findTestMenuEntryById("grandparent", "grandchildId"), true, false},
+ {"grandparent", ts.findTestMenuEntryByID("grandparent", "grandparentId"), false, false},
+ {"grandparent", ts.findTestMenuEntryByID("grandparent", "parentId"), false, true},
+ {"grandparent", ts.findTestMenuEntryByID("grandparent", "grandchildId"), true, false},
} {
isMenuCurrent := home.IsMenuCurrent(this.menu, this.menuItem)
@@ -391,7 +391,7 @@
var testMenuIdentityMatcher = func(me *MenuEntry, id string) bool { return me.Identifier == id }
var testMenuNameMatcher = func(me *MenuEntry, id string) bool { return me.Name == id }
-func (ts testMenuState) findTestMenuEntryById(mn string, id string) *MenuEntry {
+func (ts testMenuState) findTestMenuEntryByID(mn string, id string) *MenuEntry {
return ts.findTestMenuEntry(mn, id, testMenuIdentityMatcher)
}
func (ts testMenuState) findTestMenuEntryByName(mn string, id string) *MenuEntry {
@@ -447,7 +447,7 @@
}
func getTestMenuState(s *Site, t *testing.T) *testMenuState {
- menuState := &testMenuState{site: s, oldBaseUrl: viper.Get("baseurl"), oldMenu: viper.Get("menu")}
+ menuState := &testMenuState{site: s, oldBaseURL: viper.Get("baseurl"), oldMenu: viper.Get("menu")}
menus, err := tomlToMap(CONF_MENU1)
@@ -471,7 +471,7 @@
func resetMenuTestState(state *testMenuState) {
viper.Set("menu", state.oldMenu)
- viper.Set("baseurl", state.oldBaseUrl)
+ viper.Set("baseurl", state.oldBaseURL)
}
func createTestSite(pageSources []source.ByteSource) *Site {
--- a/hugolib/node.go
+++ b/hugolib/node.go
@@ -40,7 +40,7 @@
return time.Now()
}
-func (n *Node) HasMenuCurrent(menuId string, inme *MenuEntry) bool {
+func (n *Node) HasMenuCurrent(menuID string, inme *MenuEntry) bool {
if inme.HasChildren() {
me := MenuEntry{Name: n.Title, Url: n.Url}
@@ -54,7 +54,7 @@
return false
}
-func (n *Node) IsMenuCurrent(menuId string, inme *MenuEntry) bool {
+func (n *Node) IsMenuCurrent(menuID string, inme *MenuEntry) bool {
me := MenuEntry{Name: n.Title, Url: n.Url}
if !me.IsSameResource(inme) {
@@ -63,7 +63,7 @@
// this resource may be included in several menus
// search for it to make sure that it is in the menu with the given menuId
- if menu, ok := (*n.Site.Menus)[menuId]; ok {
+ if menu, ok := (*n.Site.Menus)[menuID]; ok {
for _, menuEntry := range *menu {
if menuEntry.IsSameResource(inme) {
return true
--- a/hugolib/page.go
+++ b/hugolib/page.go
@@ -149,8 +149,8 @@
return al
}
-func (p *Page) UniqueId() string {
- return p.Source.UniqueId()
+func (p *Page) UniqueID() string {
+ return p.Source.UniqueID()
}
func (p *Page) Ref(ref string) (string, error) {
@@ -200,12 +200,12 @@
func (p *Page) renderBytes(content []byte) []byte {
return helpers.RenderBytes(
&helpers.RenderingContext{Content: content, PageFmt: p.guessMarkupType(),
- DocumentId: p.UniqueId(), Config: p.getRenderingConfig()})
+ DocumentID: p.UniqueID(), Config: p.getRenderingConfig()})
}
func (p *Page) renderContent(content []byte) []byte {
return helpers.RenderBytesWithTOC(&helpers.RenderingContext{Content: content, PageFmt: p.guessMarkupType(),
- DocumentId: p.UniqueId(), Config: p.getRenderingConfig()})
+ DocumentID: p.UniqueID(), Config: p.getRenderingConfig()})
}
func (p *Page) getRenderingConfig() *helpers.Blackfriday {
@@ -341,15 +341,15 @@
}
func (p *Page) permalink() (*url.URL, error) {
- baseUrl := string(p.Site.BaseUrl)
+ baseURL := string(p.Site.BaseUrl)
dir := strings.TrimSpace(filepath.ToSlash(p.Source.Dir()))
pSlug := strings.TrimSpace(p.Slug)
- pUrl := strings.TrimSpace(p.Url)
+ pURL := strings.TrimSpace(p.Url)
var permalink string
var err error
- if len(pUrl) > 0 {
- return helpers.MakePermalink(baseUrl, pUrl), nil
+ if len(pURL) > 0 {
+ return helpers.MakePermalink(baseURL, pURL), nil
}
if override, ok := p.Site.Permalinks[p.Section()]; ok {
@@ -361,14 +361,14 @@
// fmt.Printf("have a section override for %q in section %s → %s\n", p.Title, p.Section, permalink)
} else {
if len(pSlug) > 0 {
- permalink = helpers.UrlPrep(viper.GetBool("UglyUrls"), path.Join(dir, p.Slug+"."+p.Extension()))
+ permalink = helpers.URLPrep(viper.GetBool("UglyURLs"), path.Join(dir, p.Slug+"."+p.Extension()))
} else {
_, t := filepath.Split(p.Source.LogicalName())
- permalink = helpers.UrlPrep(viper.GetBool("UglyUrls"), path.Join(dir, helpers.ReplaceExtension(strings.TrimSpace(t), p.Extension())))
+ permalink = helpers.URLPrep(viper.GetBool("UglyURLs"), path.Join(dir, helpers.ReplaceExtension(strings.TrimSpace(t), p.Extension())))
}
}
- return helpers.MakePermalink(baseUrl, permalink), nil
+ return helpers.MakePermalink(baseURL, permalink), nil
}
func (p *Page) Extension() string {
@@ -419,7 +419,7 @@
return "", err
}
- if viper.GetBool("CanonifyUrls") {
+ if viper.GetBool("CanonifyURLs") {
// replacements for relpermalink with baseUrl on the form http://myhost.com/sub/ will fail later on
// have to return the Url relative from baseUrl
relpath, err := helpers.GetRelativePath(link.String(), string(p.Site.BaseUrl))
@@ -452,12 +452,12 @@
case "description":
p.Description = cast.ToString(v)
case "slug":
- p.Slug = helpers.Urlize(cast.ToString(v))
+ p.Slug = helpers.URLize(cast.ToString(v))
case "url":
if url := cast.ToString(v); strings.HasPrefix(url, "http://") || strings.HasPrefix(url, "https://") {
return fmt.Errorf("Only relative urls are supported, %v provided", url)
}
- p.Url = helpers.Urlize(cast.ToString(v))
+ p.Url = helpers.URLize(cast.ToString(v))
case "type":
p.contentType = cast.ToString(v)
case "extension", "ext":
--- a/hugolib/page_permalink_test.go
+++ b/hugolib/page_permalink_test.go
@@ -16,8 +16,8 @@
base template.URL
slug string
url string
- uglyUrls bool
- canonifyUrls bool
+ uglyURLs bool
+ canonifyURLs bool
expectedAbs string
expectedRel string
}{
@@ -42,8 +42,8 @@
viper.Set("DefaultExtension", "html")
for i, test := range tests {
- viper.Set("uglyurls", test.uglyUrls)
- viper.Set("canonifyurls", test.canonifyUrls)
+ viper.Set("uglyurls", test.uglyURLs)
+ viper.Set("canonifyurls", test.canonifyURLs)
p := &Page{
Node: Node{
UrlPath: UrlPath{
--- a/hugolib/pagesPrevNext.go
+++ b/hugolib/pagesPrevNext.go
@@ -15,7 +15,7 @@
func (p Pages) Prev(cur *Page) *Page {
for x, c := range p {
- if c.UniqueId() == cur.UniqueId() {
+ if c.UniqueID() == cur.UniqueID() {
if x == 0 {
return p[len(p)-1]
}
@@ -27,7 +27,7 @@
func (p Pages) Next(cur *Page) *Page {
for x, c := range p {
- if c.UniqueId() == cur.UniqueId() {
+ if c.UniqueID() == cur.UniqueID() {
if x < len(p)-1 {
return p[x+1]
}
--- a/hugolib/pagination.go
+++ b/hugolib/pagination.go
@@ -35,12 +35,12 @@
type paginator struct {
paginatedPages []Pages
pagers
- paginationUrlFactory
+ paginationURLFactory
total int
size int
}
-type paginationUrlFactory func(int) string
+type paginationURLFactory func(int) string
// PageNumber returns the current page's number in the pager sequence.
func (p *pager) PageNumber() int {
@@ -49,7 +49,7 @@
// Url returns the url to the current page.
func (p *pager) Url() template.HTML {
- return template.HTML(p.paginationUrlFactory(p.PageNumber()))
+ return template.HTML(p.paginationURLFactory(p.PageNumber()))
}
// Pages returns the elements on this page.
@@ -225,7 +225,7 @@
return nil, errors.New(fmt.Sprintf("unsupported type in paginate, got %T", seq))
}
- urlFactory := newPaginationUrlFactory(section)
+ urlFactory := newPaginationURLFactory(section)
paginator, _ := newPaginator(pages, paginateSize, urlFactory)
pagers := paginator.Pagers()
@@ -232,7 +232,7 @@
return pagers, nil
}
-func newPaginator(pages Pages, size int, urlFactory paginationUrlFactory) (*paginator, error) {
+func newPaginator(pages Pages, size int, urlFactory paginationURLFactory) (*paginator, error) {
if size <= 0 {
return nil, errors.New("Paginator size must be positive")
@@ -240,7 +240,7 @@
split := splitPages(pages, size)
- p := &paginator{total: len(pages), paginatedPages: split, size: size, paginationUrlFactory: urlFactory}
+ p := &paginator{total: len(pages), paginatedPages: split, size: size, paginationURLFactory: urlFactory}
var ps pagers
@@ -259,7 +259,7 @@
return p, nil
}
-func newPaginationUrlFactory(pathElements ...string) paginationUrlFactory {
+func newPaginationURLFactory(pathElements ...string) paginationURLFactory {
paginatePath := viper.GetString("paginatePath")
return func(page int) string {
@@ -270,6 +270,6 @@
rel = fmt.Sprintf("/%s/%s/%d/", path.Join(pathElements...), paginatePath, page)
}
- return helpers.UrlizeAndPrep(rel)
+ return helpers.URLizeAndPrep(rel)
}
}
--- a/hugolib/pagination_test.go
+++ b/hugolib/pagination_test.go
@@ -99,8 +99,8 @@
func TestPaginationUrlFactory(t *testing.T) {
viper.Set("PaginatePath", "zoo")
- unicode := newPaginationUrlFactory("новости проекта")
- fooBar := newPaginationUrlFactory("foo", "bar")
+ unicode := newPaginationURLFactory("новости проекта")
+ fooBar := newPaginationURLFactory("foo", "bar")
assert.Equal(t, "/%D0%BD%D0%BE%D0%B2%D0%BE%D1%81%D1%82%D0%B8-%D0%BF%D1%80%D0%BE%D0%B5%D0%BA%D1%82%D0%B0/", unicode(1))
assert.Equal(t, "/foo/bar/", fooBar(1))
--- a/hugolib/path_separators_windows_test.go
+++ b/hugolib/path_separators_windows_test.go
@@ -11,7 +11,7 @@
)
func TestTemplatePathSeparator(t *testing.T) {
- tmpl := new(tpl.GoHtmlTemplate)
+ tmpl := new(tpl.GoHTMLTemplate)
if name := tmpl.GenerateTemplateNameFrom(win_base, win_path); name != "sub1/index.html" {
t.Fatalf("Template name incorrect. Expected: %s, Got: %s", "sub1/index.html", name)
}
--- a/hugolib/permalinks.go
+++ b/hugolib/permalinks.go
@@ -139,7 +139,7 @@
func pageToPermalinkTitle(p *Page, _ string) (string, error) {
// Page contains Node which has Title
// (also contains UrlPath which has Slug, sometimes)
- return helpers.Urlize(p.Title), nil
+ return helpers.URLize(p.Title), nil
}
// pageToPermalinkFilename returns the URL-safe form of the filename
@@ -146,7 +146,7 @@
func pageToPermalinkFilename(p *Page, _ string) (string, error) {
//var extension = p.Source.Ext
//var name = p.Source.Path()[0 : len(p.Source.Path())-len(extension)]
- return helpers.Urlize(p.Source.BaseFileName()), nil
+ return helpers.URLize(p.Source.BaseFileName()), nil
}
// if the page has a slug, return the slug, else return the title
--- a/hugolib/shortcode.go
+++ b/hugolib/shortcode.go
@@ -203,7 +203,7 @@
if sc.doMarkup {
newInner := helpers.RenderBytes(&helpers.RenderingContext{
Content: []byte(inner), PageFmt: p.guessMarkupType(),
- DocumentId: p.UniqueId(), Config: p.getRenderingConfig()})
+ DocumentID: p.UniqueID(), Config: p.getRenderingConfig()})
// If the type is “unknown” or “markdown”, we assume the markdown
// generation has been performed. Given the input: `a line`, markdown
--- a/hugolib/site.go
+++ b/hugolib/site.go
@@ -113,7 +113,7 @@
Permalinks PermalinkOverrides
Params map[string]interface{}
BuildDrafts bool
- canonifyUrls bool
+ canonifyURLs bool
paginationPageCount uint64
Data *map[string]interface{}
}
@@ -158,10 +158,10 @@
}
func (s *SiteInfo) refLink(ref string, page *Page, relative bool) (string, error) {
- var refUrl *url.URL
+ var refURL *url.URL
var err error
- refUrl, err = url.Parse(ref)
+ refURL, err = url.Parse(ref)
if err != nil {
return "", err
@@ -170,9 +170,9 @@
var target *Page
var link string
- if refUrl.Path != "" {
+ if refURL.Path != "" {
for _, page := range []*Page(*s.Pages) {
- if page.Source.Path() == refUrl.Path || page.Source.LogicalName() == refUrl.Path {
+ if page.Source.Path() == refURL.Path || page.Source.LogicalName() == refURL.Path {
target = page
break
}
@@ -179,7 +179,7 @@
}
if target == nil {
- return "", fmt.Errorf("No page found with path or logical name \"%s\".\n", refUrl.Path)
+ return "", fmt.Errorf("No page found with path or logical name \"%s\".\n", refURL.Path)
}
if relative {
@@ -193,13 +193,13 @@
}
}
- if refUrl.Fragment != "" {
- link = link + "#" + refUrl.Fragment
+ if refURL.Fragment != "" {
+ link = link + "#" + refURL.Fragment
- if refUrl.Path != "" && target != nil && !target.getRenderingConfig().PlainIdAnchors {
- link = link + ":" + target.UniqueId()
- } else if page != nil && !page.getRenderingConfig().PlainIdAnchors {
- link = link + ":" + page.UniqueId()
+ if refURL.Path != "" && target != nil && !target.getRenderingConfig().PlainIDAnchors {
+ link = link + ":" + target.UniqueID()
+ } else if page != nil && !page.getRenderingConfig().PlainIDAnchors {
+ link = link + ":" + page.UniqueID()
}
}
@@ -316,11 +316,11 @@
func readData(f *source.File) (interface{}, error) {
switch f.Extension() {
case "yaml", "yml":
- return parser.HandleYamlMetaData(f.Bytes())
+ return parser.HandleYAMLMetaData(f.Bytes())
case "json":
- return parser.HandleJsonMetaData(f.Bytes())
+ return parser.HandleJSONMetaData(f.Bytes())
case "toml":
- return parser.HandleTomlMetaData(f.Bytes())
+ return parser.HandleTOMLMetaData(f.Bytes())
default:
return nil, fmt.Errorf("Data not supported for extension '%s'", f.Extension())
}
@@ -444,7 +444,7 @@
}
s.Info = SiteInfo{
- BaseUrl: template.URL(helpers.SanitizeUrlKeepTrailingSlash(viper.GetString("BaseUrl"))),
+ BaseUrl: template.URL(helpers.SanitizeURLKeepTrailingSlash(viper.GetString("BaseURL"))),
Title: viper.GetString("Title"),
Author: viper.GetStringMap("author"),
LanguageCode: viper.GetString("languagecode"),
@@ -451,7 +451,7 @@
Copyright: viper.GetString("copyright"),
DisqusShortname: viper.GetString("DisqusShortname"),
BuildDrafts: viper.GetBool("BuildDrafts"),
- canonifyUrls: viper.GetBool("CanonifyUrls"),
+ canonifyURLs: viper.GetBool("CanonifyURLs"),
Pages: &s.Pages,
Recent: &s.Pages,
Menus: &s.Menus,
@@ -706,12 +706,12 @@
if strings.HasPrefix(menuEntry.Url, "/") {
// make it match the nodes
- menuEntryUrl := menuEntry.Url
- menuEntryUrl = helpers.UrlizeAndPrep(menuEntryUrl)
- if !s.Info.canonifyUrls {
- menuEntryUrl = helpers.AddContextRoot(string(s.Info.BaseUrl), menuEntryUrl)
+ menuEntryURL := menuEntry.Url
+ menuEntryURL = helpers.URLizeAndPrep(menuEntryURL)
+ if !s.Info.canonifyURLs {
+ menuEntryURL = helpers.AddContextRoot(string(s.Info.BaseUrl), menuEntryURL)
}
- menuEntry.Url = menuEntryUrl
+ menuEntry.Url = menuEntryURL
}
if ret[name] == nil {
@@ -1249,7 +1249,7 @@
}
}
- n.Url = helpers.Urlize("404.html")
+ n.Url = helpers.URLize("404.html")
n.Title = "404 Page not found"
n.Permalink = s.permalink("404.html")
@@ -1315,7 +1315,7 @@
}
func (s *Site) setUrls(n *Node, in string) {
- n.Url = helpers.UrlizeAndPrep(in)
+ n.Url = helpers.URLizeAndPrep(in)
n.Permalink = s.permalink(n.Url)
n.RSSLink = s.permalink(in + ".xml")
}
@@ -1325,7 +1325,7 @@
}
func (s *Site) permalinkStr(plink string) string {
- return helpers.MakePermalink(string(viper.GetString("BaseUrl")), helpers.UrlizeAndPrep(plink)).String()
+ return helpers.MakePermalink(string(viper.GetString("BaseURL")), helpers.URLizeAndPrep(plink)).String()
}
func (s *Site) NewNode() *Node {
@@ -1348,7 +1348,7 @@
err := s.render(name, d, renderBuffer, layouts...)
- absURLInXML, err := transform.AbsURLInXML(viper.GetString("BaseUrl"))
+ absURLInXML, err := transform.AbsURLInXML(viper.GetString("BaseURL"))
if err != nil {
return err
}
@@ -1377,8 +1377,8 @@
transformLinks := transform.NewEmptyTransforms()
- if viper.GetBool("CanonifyUrls") {
- absURL, err := transform.AbsURL(viper.GetString("BaseUrl"))
+ if viper.GetBool("CanonifyURLs") {
+ absURL, err := transform.AbsURL(viper.GetString("BaseURL"))
if err != nil {
return err
}
@@ -1460,7 +1460,7 @@
if s.Targets.Page == nil {
s.Targets.Page = &target.PagePub{
PublishDir: s.absPublishDir(),
- UglyUrls: viper.GetBool("UglyUrls"),
+ UglyURLs: viper.GetBool("UglyURLs"),
}
}
if s.Targets.File == nil {
--- a/hugolib/site_show_plan_test.go
+++ b/hugolib/site_show_plan_test.go
@@ -92,7 +92,7 @@
func TestPageTargetUgly(t *testing.T) {
s := &Site{
- Targets: targetList{Page: &target.PagePub{UglyUrls: true}},
+ Targets: targetList{Page: &target.PagePub{UglyURLs: true}},
Source: &source.InMemorySource{ByteSource: fakeSource},
}
s.AliasTarget()
--- a/hugolib/site_test.go
+++ b/hugolib/site_test.go
@@ -309,18 +309,18 @@
// Issue #939
func Test404ShouldAlwaysHaveUglyUrls(t *testing.T) {
- for _, uglyUrls := range []bool{true, false} {
- doTest404ShouldAlwaysHaveUglyUrls(t, uglyUrls)
+ for _, uglyURLs := range []bool{true, false} {
+ doTest404ShouldAlwaysHaveUglyUrls(t, uglyURLs)
}
}
-func doTest404ShouldAlwaysHaveUglyUrls(t *testing.T, uglyUrls bool) {
+func doTest404ShouldAlwaysHaveUglyUrls(t *testing.T, uglyURLs bool) {
viper.Set("verbose", true)
viper.Set("baseurl", "http://auth/bub")
viper.Set("DisableSitemap", false)
viper.Set("DisableRSS", false)
- viper.Set("UglyUrls", uglyUrls)
+ viper.Set("UglyURLs", uglyURLs)
sources := []source.ByteSource{
{filepath.FromSlash("sect/doc1.html"), []byte("---\nmarkup: markdown\n---\n# title\nsome *content*")},
@@ -328,7 +328,7 @@
s := &Site{
Source: &source.InMemorySource{ByteSource: sources},
- Targets: targetList{Page: &target.PagePub{UglyUrls: uglyUrls}},
+ Targets: targetList{Page: &target.PagePub{UglyURLs: uglyURLs}},
}
s.initializeSiteInfo()
@@ -347,7 +347,7 @@
s.RenderSitemap()
var expectedPagePath string
- if uglyUrls {
+ if uglyURLs {
expectedPagePath = "sect/doc1.html"
} else {
expectedPagePath = "sect/doc1/index.html"
@@ -392,11 +392,11 @@
}
viper.Set("verbose", true)
- viper.Set("CanonifyUrls", true)
+ viper.Set("CanonifyURLs", true)
viper.Set("baseurl", "http://auth/bub")
s := &Site{
Source: &source.InMemorySource{ByteSource: sources},
- Targets: targetList{Page: &target.PagePub{UglyUrls: true}},
+ Targets: targetList{Page: &target.PagePub{UglyURLs: true}},
}
s.initializeSiteInfo()
@@ -442,13 +442,13 @@
{filepath.FromSlash("content/blue/doc2.html"), []byte("---\nf: t\n---\n<!doctype html><html><body>more content</body></html>")},
}
for _, canonify := range []bool{true, false} {
- viper.Set("CanonifyUrls", canonify)
- viper.Set("BaseUrl", "http://auth/bub")
+ viper.Set("CanonifyURLs", canonify)
+ viper.Set("BaseURL", "http://auth/bub")
s := &Site{
Source: &source.InMemorySource{ByteSource: sources},
- Targets: targetList{Page: &target.PagePub{UglyUrls: true}},
+ Targets: targetList{Page: &target.PagePub{UglyURLs: true}},
}
- t.Logf("Rendering with BaseUrl %q and CanonifyUrls set %v", viper.GetString("baseUrl"), canonify)
+ t.Logf("Rendering with BaseURL %q and CanonifyURLs set %v", viper.GetString("baseURL"), canonify)
s.initializeSiteInfo()
templatePrep(s)
must(s.addTemplate("blue/single.html", TEMPLATE_WITH_URL_ABS))
@@ -823,13 +823,13 @@
}
}
-func TestDataDirJson(t *testing.T) {
+func TestDataDirJSON(t *testing.T) {
sources := []source.ByteSource{
{filepath.FromSlash("test/foo.json"), []byte(`{ "bar": "foofoo" }`)},
{filepath.FromSlash("test.json"), []byte(`{ "hello": [ { "world": "foo" } ] }`)},
}
- expected, err := parser.HandleJsonMetaData([]byte(`{ "test": { "hello": [{ "world": "foo" }] , "foo": { "bar":"foofoo" } } }`))
+ expected, err := parser.HandleJSONMetaData([]byte(`{ "test": { "hello": [{ "world": "foo" }] , "foo": { "bar":"foofoo" } } }`))
if err != nil {
t.Fatalf("Error %s", err)
@@ -843,7 +843,7 @@
{filepath.FromSlash("test/kung.toml"), []byte("[foo]\nbar = 1")},
}
- expected, err := parser.HandleTomlMetaData([]byte("[test]\n[test.kung]\n[test.kung.foo]\nbar = 1"))
+ expected, err := parser.HandleTOMLMetaData([]byte("[test]\n[test.kung]\n[test.kung.foo]\nbar = 1"))
if err != nil {
t.Fatalf("Error %s", err)
@@ -852,7 +852,7 @@
doTestDataDir(t, expected, []source.Input{&source.InMemorySource{ByteSource: sources}})
}
-func TestDataDirYamlWithOverridenValue(t *testing.T) {
+func TestDataDirYAMLWithOverridenValue(t *testing.T) {
sources := []source.ByteSource{
// filepath.Walk walks the files in lexical order, '/' comes before '.'. Simulate this:
{filepath.FromSlash("a.yaml"), []byte("a: 1")},
@@ -878,7 +878,7 @@
{filepath.FromSlash("test/second.toml"), []byte("tender = 2")},
}
- expected, _ := parser.HandleTomlMetaData([]byte("[test.first]\nbar = 1\n[test.second]\ntender=2"))
+ expected, _ := parser.HandleTOMLMetaData([]byte("[test.first]\nbar = 1\n[test.second]\ntender=2"))
doTestDataDir(t, expected, []source.Input{&source.InMemorySource{ByteSource: s1}, &source.InMemorySource{ByteSource: s2}})
--- a/hugolib/taxonomy.go
+++ b/hugolib/taxonomy.go
@@ -162,7 +162,7 @@
func (wp WeightedPages) Prev(cur *Page) *Page {
for x, c := range wp {
- if c.Page.UniqueId() == cur.UniqueId() {
+ if c.Page.UniqueID() == cur.UniqueID() {
if x == 0 {
return wp[len(wp)-1].Page
}
@@ -174,7 +174,7 @@
func (wp WeightedPages) Next(cur *Page) *Page {
for x, c := range wp {
- if c.Page.UniqueId() == cur.UniqueId() {
+ if c.Page.UniqueID() == cur.UniqueID() {
if x < len(wp)-1 {
return wp[x+1].Page
}
--- a/parser/frontmatter.go
+++ b/parser/frontmatter.go
@@ -151,19 +151,19 @@
func DetectFrontMatter(mark rune) (f *FrontmatterType) {
switch mark {
case '-':
- return &FrontmatterType{[]byte(YAML_DELIM), []byte(YAML_DELIM), HandleYamlMetaData, false}
+ return &FrontmatterType{[]byte(YAML_DELIM), []byte(YAML_DELIM), HandleYAMLMetaData, false}
case '+':
- return &FrontmatterType{[]byte(TOML_DELIM), []byte(TOML_DELIM), HandleTomlMetaData, false}
+ return &FrontmatterType{[]byte(TOML_DELIM), []byte(TOML_DELIM), HandleTOMLMetaData, false}
case '{':
- return &FrontmatterType{[]byte{'{'}, []byte{'}'}, HandleJsonMetaData, true}
+ return &FrontmatterType{[]byte{'{'}, []byte{'}'}, HandleJSONMetaData, true}
default:
return nil
}
}
-func HandleTomlMetaData(datum []byte) (interface{}, error) {
+func HandleTOMLMetaData(datum []byte) (interface{}, error) {
m := map[string]interface{}{}
- datum = removeTomlIdentifier(datum)
+ datum = removeTOMLIdentifier(datum)
if _, err := toml.Decode(string(datum), &m); err != nil {
return m, err
}
@@ -170,11 +170,11 @@
return m, nil
}
-func removeTomlIdentifier(datum []byte) []byte {
+func removeTOMLIdentifier(datum []byte) []byte {
return bytes.Replace(datum, []byte(TOML_DELIM), []byte(""), -1)
}
-func HandleYamlMetaData(datum []byte) (interface{}, error) {
+func HandleYAMLMetaData(datum []byte) (interface{}, error) {
m := map[string]interface{}{}
if err := yaml.Unmarshal(datum, &m); err != nil {
return m, err
@@ -182,7 +182,7 @@
return m, nil
}
-func HandleJsonMetaData(datum []byte) (interface{}, error) {
+func HandleJSONMetaData(datum []byte) (interface{}, error) {
var f interface{}
if err := json.Unmarshal(datum, &f); err != nil {
return f, err
--- a/source/file.go
+++ b/source/file.go
@@ -27,14 +27,14 @@
section string // The first directory
dir string // The full directory Path (minus file name)
ext string // Just the ext (eg txt)
- uniqueId string // MD5 of the filename
+ uniqueID string // MD5 of the filename
}
-func (f *File) UniqueId() string {
- if f.uniqueId == "" {
- f.uniqueId = helpers.Md5String(f.LogicalName())
+func (f *File) UniqueID() string {
+ if f.uniqueID == "" {
+ f.uniqueID = helpers.Md5String(f.LogicalName())
}
- return f.uniqueId
+ return f.uniqueID
}
func (f *File) String() string {
--- a/target/page.go
+++ b/target/page.go
@@ -16,7 +16,7 @@
}
type PagePub struct {
- UglyUrls bool
+ UglyURLs bool
DefaultExtension string
PublishDir string
}
@@ -46,7 +46,7 @@
dir = filepath.Join(pp.PublishDir, dir)
}
- if pp.UglyUrls || file == "index.html" || file == "404.html" {
+ if pp.UglyURLs || file == "index.html" || file == "404.html" {
return filepath.Join(dir, fmt.Sprintf("%s%s", name, ext)), nil
}
--- a/target/page_test.go
+++ b/target/page_test.go
@@ -74,7 +74,7 @@
}
for _, test := range tests {
- f := &PagePub{UglyUrls: true}
+ f := &PagePub{UglyURLs: true}
dest, err := f.Translate(filepath.FromSlash(test.content))
if err != nil {
t.Fatalf("Translate returned an unexpected err: %s", err)
--- a/tpl/template.go
+++ b/tpl/template.go
@@ -58,7 +58,7 @@
err error
}
-type GoHtmlTemplate struct {
+type GoHTMLTemplate struct {
template.Template
errors []*templateErr
}
@@ -81,7 +81,7 @@
// Return a new Hugo Template System
// With all the additional features, templates & functions
func New() Template {
- var templates = &GoHtmlTemplate{
+ var templates = &GoHTMLTemplate{
Template: *template.New(""),
errors: make([]*templateErr, 0),
}
@@ -934,21 +934,21 @@
return t.Format(layout), nil
}
-func SafeHtml(text string) template.HTML {
+func SafeHTML(text string) template.HTML {
return template.HTML(text)
}
// "safeHtmlAttr" is currently disabled, pending further discussion
// on its use case. 2015-01-19
-func SafeHtmlAttr(text string) template.HTMLAttr {
+func SafeHTMLAttr(text string) template.HTMLAttr {
return template.HTMLAttr(text)
}
-func SafeCss(text string) template.CSS {
+func SafeCSS(text string) template.CSS {
return template.CSS(text)
}
-func SafeUrl(text string) template.URL {
+func SafeURL(text string) template.URL {
return template.URL(text)
}
@@ -1151,12 +1151,12 @@
return template.HTML(b.String())
}
-func (t *GoHtmlTemplate) LoadEmbedded() {
+func (t *GoHTMLTemplate) LoadEmbedded() {
t.EmbedShortcodes()
t.EmbedTemplates()
}
-func (t *GoHtmlTemplate) AddInternalTemplate(prefix, name, tpl string) error {
+func (t *GoHTMLTemplate) AddInternalTemplate(prefix, name, tpl string) error {
if prefix != "" {
return t.AddTemplate("_internal/"+prefix+"/"+name, tpl)
} else {
@@ -1164,11 +1164,11 @@
}
}
-func (t *GoHtmlTemplate) AddInternalShortcode(name, content string) error {
+func (t *GoHTMLTemplate) AddInternalShortcode(name, content string) error {
return t.AddInternalTemplate("shortcodes", name, content)
}
-func (t *GoHtmlTemplate) AddTemplate(name, tpl string) error {
+func (t *GoHTMLTemplate) AddTemplate(name, tpl string) error {
_, err := t.New(name).Parse(tpl)
if err != nil {
t.errors = append(t.errors, &templateErr{name: name, err: err})
@@ -1176,7 +1176,7 @@
return err
}
-func (t *GoHtmlTemplate) AddTemplateFile(name, path string) error {
+func (t *GoHTMLTemplate) AddTemplateFile(name, path string) error {
// get the suffix and switch on that
ext := filepath.Ext(path)
switch ext {
@@ -1221,7 +1221,7 @@
}
-func (t *GoHtmlTemplate) GenerateTemplateNameFrom(base, path string) string {
+func (t *GoHTMLTemplate) GenerateTemplateNameFrom(base, path string) string {
name, _ := filepath.Rel(base, path)
return filepath.ToSlash(name)
}
@@ -1234,7 +1234,7 @@
return path[len(path)-1] == '~'
}
-func (t *GoHtmlTemplate) loadTemplates(absPath string, prefix string) {
+func (t *GoHTMLTemplate) loadTemplates(absPath string, prefix string) {
walker := func(path string, fi os.FileInfo, err error) error {
if err != nil {
return nil
@@ -1277,15 +1277,15 @@
filepath.Walk(absPath, walker)
}
-func (t *GoHtmlTemplate) LoadTemplatesWithPrefix(absPath string, prefix string) {
+func (t *GoHTMLTemplate) LoadTemplatesWithPrefix(absPath string, prefix string) {
t.loadTemplates(absPath, prefix)
}
-func (t *GoHtmlTemplate) LoadTemplates(absPath string) {
+func (t *GoHTMLTemplate) LoadTemplates(absPath string) {
t.loadTemplates(absPath, "")
}
-func (t *GoHtmlTemplate) PrintErrors() {
+func (t *GoHTMLTemplate) PrintErrors() {
for _, e := range t.errors {
jww.ERROR.Println(e.err)
}
@@ -1293,8 +1293,9 @@
func init() {
funcMap = template.FuncMap{
- "urlize": helpers.Urlize,
- "sanitizeurl": helpers.SanitizeUrl,
+ "urlize": helpers.URLize,
+ "sanitizeURL": helpers.SanitizeURL,
+ "sanitizeurl": helpers.SanitizeURL,
"eq": Eq,
"ne": Ne,
"gt": Gt,
@@ -1303,11 +1304,15 @@
"le": Le,
"in": In,
"intersect": Intersect,
+ "isSet": IsSet,
"isset": IsSet,
"echoParam": ReturnWhenSet,
- "safeHtml": SafeHtml,
- "safeCss": SafeCss,
- "safeUrl": SafeUrl,
+ "safeHTML": SafeHTML,
+ "safeHtml": SafeHTML,
+ "safeCSS": SafeCSS,
+ "safeCss": SafeCSS,
+ "safeURL": SafeURL,
+ "safeUrl": SafeURL,
"markdownify": Markdownify,
"first": First,
"where": Where,
@@ -1331,8 +1336,10 @@
"replace": Replace,
"trim": Trim,
"dateFormat": DateFormat,
- "getJson": GetJson,
- "getCsv": GetCsv,
+ "getJSON": GetJSON,
+ "getJson": GetJSON,
+ "getCSV": GetCSV,
+ "getCsv": GetCSV,
}
}
--- a/tpl/template_embedded.go
+++ b/tpl/template_embedded.go
@@ -18,7 +18,7 @@
Data string
}
-func (t *GoHtmlTemplate) EmbedShortcodes() {
+func (t *GoHTMLTemplate) EmbedShortcodes() {
t.AddInternalShortcode("ref.html", `{{ .Get 0 | ref .Page }}`)
t.AddInternalShortcode("relref.html", `{{ .Get 0 | relref .Page }}`)
t.AddInternalShortcode("highlight.html", `{{ .Get 0 | highlight .Inner }}`)
@@ -43,7 +43,7 @@
<!-- image -->`)
}
-func (t *GoHtmlTemplate) EmbedTemplates() {
+func (t *GoHTMLTemplate) EmbedTemplates() {
t.AddInternalTemplate("_default", "rss.xml", `<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
<channel>
--- a/tpl/template_resources.go
+++ b/tpl/template_resources.go
@@ -31,7 +31,7 @@
"github.com/spf13/viper"
)
-var remoteUrlLock = &remoteLock{m: make(map[string]*sync.Mutex)}
+var remoteURLLock = &remoteLock{m: make(map[string]*sync.Mutex)}
type remoteLock struct {
sync.RWMutex
@@ -39,7 +39,7 @@
}
// resLock locks an URL during download
-func (l *remoteLock) UrlLock(url string) {
+func (l *remoteLock) URLLock(url string) {
l.Lock()
if _, ok := l.m[url]; !ok {
l.m[url] = &sync.Mutex{}
@@ -49,7 +49,7 @@
}
// resUnlock unlocks an URL when the download has been finished. Use only in defer calls.
-func (l *remoteLock) UrlUnlock(url string) {
+func (l *remoteLock) URLUnlock(url string) {
l.RLock()
defer l.RUnlock()
if um, ok := l.m[url]; ok {
@@ -111,8 +111,8 @@
}
// avoid race condition with locks, block other goroutines if the current url is processing
- remoteUrlLock.UrlLock(url)
- defer func() { remoteUrlLock.UrlUnlock(url) }()
+ remoteURLLock.URLLock(url)
+ defer func() { remoteURLLock.URLUnlock(url) }()
// avoid multiple locks due to calling resGetCache twice
c, err = resGetCache(url, fs, viper.GetBool("IgnoreCache"))
@@ -176,7 +176,7 @@
// GetJson expects one or n-parts of a URL to a resource which can either be a local or a remote one.
// If you provide multiple parts they will be joined together to the final URL.
// GetJson returns nil or parsed JSON to use in a short code.
-func GetJson(urlParts ...string) interface{} {
+func GetJSON(urlParts ...string) interface{} {
url := strings.Join(urlParts, "")
c, err := resGetResource(url)
if err != nil {
@@ -194,7 +194,7 @@
}
// parseCsv parses bytes of csv data into a slice slice string or an error
-func parseCsv(c []byte, sep string) ([][]string, error) {
+func parseCSV(c []byte, sep string) ([][]string, error) {
if len(sep) != 1 {
return nil, errors.New("Incorrect length of csv separator: " + sep)
}
@@ -211,7 +211,7 @@
// The data separator can be a comma, semi-colon, pipe, etc, but only one character.
// If you provide multiple parts for the URL they will be joined together to the final URL.
// GetCsv returns nil or a slice slice to use in a short code.
-func GetCsv(sep string, urlParts ...string) [][]string {
+func GetCSV(sep string, urlParts ...string) [][]string {
url := strings.Join(urlParts, "")
c, err := resGetResource(url)
if err != nil {
@@ -218,7 +218,7 @@
jww.ERROR.Printf("Failed to get csv resource %s with error message %s", url, err)
return nil
}
- d, err := parseCsv(c, sep)
+ d, err := parseCSV(c, sep)
if err != nil {
jww.ERROR.Printf("Failed to read csv resource %s with error message %s", url, err)
return nil
--- a/tpl/template_resources_test.go
+++ b/tpl/template_resources_test.go
@@ -157,7 +157,7 @@
}
}
-func TestParseCsv(t *testing.T) {
+func TestParseCSV(t *testing.T) {
tests := []struct {
csv []byte
@@ -173,7 +173,7 @@
{[]byte("z|y|c\nd|e|f"), "|", "zycdef", false},
}
for _, test := range tests {
- csv, err := parseCsv(test.csv, test.sep)
+ csv, err := parseCSV(test.csv, test.sep)
if test.err && err == nil {
t.Error("Expecting an error")
}
--- a/tpl/template_test.go
+++ b/tpl/template_test.go
@@ -972,7 +972,7 @@
}
}
-func TestSafeHtml(t *testing.T) {
+func TestSafeHTML(t *testing.T) {
for i, this := range []struct {
str string
tmplStr string
@@ -997,7 +997,7 @@
}
buf.Reset()
- err = tmpl.Execute(buf, SafeHtml(this.str))
+ err = tmpl.Execute(buf, SafeHTML(this.str))
if err != nil {
t.Errorf("[%d] execute template with an escaped string value by SafeHtml returns unexpected error: %s", i, err)
}
@@ -1007,7 +1007,7 @@
}
}
-func TestSafeHtmlAttr(t *testing.T) {
+func TestSafeHTMLAttr(t *testing.T) {
for i, this := range []struct {
str string
tmplStr string
@@ -1032,7 +1032,7 @@
}
buf.Reset()
- err = tmpl.Execute(buf, SafeHtmlAttr(this.str))
+ err = tmpl.Execute(buf, SafeHTMLAttr(this.str))
if err != nil {
t.Errorf("[%d] execute template with an escaped string value by SafeHtmlAttr returns unexpected error: %s", i, err)
}
@@ -1042,7 +1042,7 @@
}
}
-func TestSafeCss(t *testing.T) {
+func TestSafeCSS(t *testing.T) {
for i, this := range []struct {
str string
tmplStr string
@@ -1067,7 +1067,7 @@
}
buf.Reset()
- err = tmpl.Execute(buf, SafeCss(this.str))
+ err = tmpl.Execute(buf, SafeCSS(this.str))
if err != nil {
t.Errorf("[%d] execute template with an escaped string value by SafeCss returns unexpected error: %s", i, err)
}
@@ -1077,7 +1077,7 @@
}
}
-func TestSafeUrl(t *testing.T) {
+func TestSafeURL(t *testing.T) {
for i, this := range []struct {
str string
tmplStr string
@@ -1102,7 +1102,7 @@
}
buf.Reset()
- err = tmpl.Execute(buf, SafeUrl(this.str))
+ err = tmpl.Execute(buf, SafeURL(this.str))
if err != nil {
t.Errorf("[%d] execute template with an escaped string value by SafeUrl returns unexpected error: %s", i, err)
}
--- a/transform/absurl.go
+++ b/transform/absurl.go
@@ -4,12 +4,12 @@
"sync"
)
-var absUrlInit sync.Once
-var ar *absurlReplacer
+var absURLInit sync.Once
+var ar *absURLReplacer
// for performance reasons, we reuse the first baseUrl given
func initAbsurlReplacer(baseURL string) {
- absUrlInit.Do(func() {
+ absURLInit.Do(func() {
ar = newAbsurlReplacer(baseURL)
})
}
@@ -18,7 +18,7 @@
initAbsurlReplacer(absURL)
trs = append(trs, func(content []byte) []byte {
- return ar.replaceInHtml(content)
+ return ar.replaceInHTML(content)
})
return
}
@@ -27,7 +27,7 @@
initAbsurlReplacer(absURL)
trs = append(trs, func(content []byte) []byte {
- return ar.replaceInXml(content)
+ return ar.replaceInXML(content)
})
return
}
--- a/transform/absurlreplacer.go
+++ b/transform/absurlreplacer.go
@@ -29,7 +29,7 @@
start int // item start position
width int // width of last element
- matchers []absurlMatcher
+ matchers []absURLMatcher
state stateFunc
prefixLookup *prefixes
@@ -101,13 +101,13 @@
var mainPrefixRunes = []prefixRunes{{'s', 'r', 'c', '='}, {'h', 'r', 'e', 'f', '='}}
-type absurlMatcher struct {
+type absURLMatcher struct {
prefix int
match []byte
replacement []byte
}
-func (a absurlMatcher) isSourceType() bool {
+func (a absURLMatcher) isSourceType() bool {
return a.prefix == matchPrefixSrc
}
@@ -177,7 +177,7 @@
}
}
-func doReplace(content []byte, matchers []absurlMatcher) []byte {
+func doReplace(content []byte, matchers []absURLMatcher) []byte {
b := bp.GetBuffer()
defer bp.PutBuffer(b)
@@ -191,48 +191,48 @@
return b.Bytes()
}
-type absurlReplacer struct {
- htmlMatchers []absurlMatcher
- xmlMatchers []absurlMatcher
+type absURLReplacer struct {
+ htmlMatchers []absURLMatcher
+ xmlMatchers []absURLMatcher
}
-func newAbsurlReplacer(baseUrl string) *absurlReplacer {
- u, _ := url.Parse(baseUrl)
+func newAbsurlReplacer(baseURL string) *absURLReplacer {
+ u, _ := url.Parse(baseURL)
base := strings.TrimRight(u.String(), "/")
// HTML
- dqHtmlMatch := []byte("\"/")
- sqHtmlMatch := []byte("'/")
+ dqHTMLMatch := []byte("\"/")
+ sqHTMLMatch := []byte("'/")
// XML
- dqXmlMatch := []byte(""/")
- sqXmlMatch := []byte("'/")
+ dqXMLMatch := []byte(""/")
+ sqXMLMatch := []byte("'/")
- dqHtml := []byte("\"" + base + "/")
- sqHtml := []byte("'" + base + "/")
+ dqHTML := []byte("\"" + base + "/")
+ sqHTML := []byte("'" + base + "/")
- dqXml := []byte(""" + base + "/")
- sqXml := []byte("'" + base + "/")
+ dqXML := []byte(""" + base + "/")
+ sqXML := []byte("'" + base + "/")
- return &absurlReplacer{
- htmlMatchers: []absurlMatcher{
- {matchPrefixSrc, dqHtmlMatch, dqHtml},
- {matchPrefixSrc, sqHtmlMatch, sqHtml},
- {matchPrefixHref, dqHtmlMatch, dqHtml},
- {matchPrefixHref, sqHtmlMatch, sqHtml}},
- xmlMatchers: []absurlMatcher{
- {matchPrefixSrc, dqXmlMatch, dqXml},
- {matchPrefixSrc, sqXmlMatch, sqXml},
- {matchPrefixHref, dqXmlMatch, dqXml},
- {matchPrefixHref, sqXmlMatch, sqXml},
+ return &absURLReplacer{
+ htmlMatchers: []absURLMatcher{
+ {matchPrefixSrc, dqHTMLMatch, dqHTML},
+ {matchPrefixSrc, sqHTMLMatch, sqHTML},
+ {matchPrefixHref, dqHTMLMatch, dqHTML},
+ {matchPrefixHref, sqHTMLMatch, sqHTML}},
+ xmlMatchers: []absURLMatcher{
+ {matchPrefixSrc, dqXMLMatch, dqXML},
+ {matchPrefixSrc, sqXMLMatch, sqXML},
+ {matchPrefixHref, dqXMLMatch, dqXML},
+ {matchPrefixHref, sqXMLMatch, sqXML},
}}
}
-func (au *absurlReplacer) replaceInHtml(content []byte) []byte {
+func (au *absURLReplacer) replaceInHTML(content []byte) []byte {
return doReplace(content, au.htmlMatchers)
}
-func (au *absurlReplacer) replaceInXml(content []byte) []byte {
+func (au *absURLReplacer) replaceInXML(content []byte) []byte {
return doReplace(content, au.xmlMatchers)
}
--- a/transform/chain_test.go
+++ b/transform/chain_test.go
@@ -54,7 +54,7 @@
}
}
-func BenchmarkAbsUrl(b *testing.B) {
+func BenchmarkAbsURL(b *testing.B) {
absURL, _ := AbsURL("http://base")
tr := NewChain(absURL...)
@@ -64,7 +64,7 @@
}
}
-func TestAbsUrl(t *testing.T) {
+func TestAbsURL(t *testing.T) {
absURL, _ := AbsURL("http://base")
tr := NewChain(absURL...)
@@ -72,7 +72,7 @@
}
-func BenchmarkXmlAbsUrl(b *testing.B) {
+func BenchmarkXMLAbsURL(b *testing.B) {
absURLInXML, _ := AbsURLInXML("http://base")
tr := NewChain(absURLInXML...)
@@ -82,7 +82,7 @@
}
}
-func TestXMLAbsUrl(t *testing.T) {
+func TestXMLAbsURL(t *testing.T) {
absURLInXML, _ := AbsURLInXML("http://base")
tr := NewChain(absURLInXML...)
apply(t.Errorf, tr, xml_abs_url_tests)