Welcome to mirror list, hosted at ThFree Co, Russian Federation.

github.com/gohugoio/hugo.git - Unnamed repository; edit this file 'description' to name the repository.
summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorBjørn Erik Pedersen <bjorn.erik.pedersen@gmail.com>2016-11-11 11:19:16 +0300
committerBjørn Erik Pedersen <bjorn.erik.pedersen@gmail.com>2016-11-22 11:57:03 +0300
commitc80308e6b3f3f4f6879ee6b7301575f162ac9209 (patch)
tree6647f3f6e8053bf20fcd468e1b2d4a3c233b9f02 /hugolib
parent063b78d2ece39b82d0eb32ac5559097c2780cef8 (diff)
node to page: Add Pages to Page
As an alias to .Data.Pages for home page etc. Also renamte NodeType to PageType and make it a string so it can be used in `where`. Updates #2297
Diffstat (limited to 'hugolib')
-rw-r--r--hugolib/hugo_sites.go36
-rw-r--r--hugolib/hugo_sites_test.go12
-rw-r--r--hugolib/menu_test.go8
-rw-r--r--hugolib/node.go92
-rw-r--r--hugolib/node_as_page_test.go9
-rw-r--r--hugolib/node_test.go40
-rw-r--r--hugolib/page.go122
-rw-r--r--hugolib/page_collections.go24
-rw-r--r--hugolib/page_permalink_test.go1
-rw-r--r--hugolib/page_test.go34
-rw-r--r--hugolib/pagination.go8
-rw-r--r--hugolib/site.go12
-rw-r--r--hugolib/site_render.go15
13 files changed, 204 insertions, 209 deletions
diff --git a/hugolib/hugo_sites.go b/hugolib/hugo_sites.go
index 72b91a123..ba743d4f1 100644
--- a/hugolib/hugo_sites.go
+++ b/hugolib/hugo_sites.go
@@ -196,8 +196,8 @@ func (h *HugoSites) renderCrossSitesArtifacts() error {
func (h *HugoSites) assignMissingTranslations() error {
// This looks heavy, but it should be a small number of nodes by now.
- allPages := h.findAllPagesByNodeTypeNotIn(NodePage)
- for _, nodeType := range []NodeType{NodeHome, NodeSection, NodeTaxonomy, NodeTaxonomyTerms} {
+ allPages := h.findAllPagesByNodeTypeNotIn(PagePage)
+ for _, nodeType := range []PageType{PageHome, PageSection, PageTaxonomy, PageTaxonomyTerm} {
nodes := h.findPagesByNodeTypeIn(nodeType, allPages)
// Assign translations
@@ -223,7 +223,7 @@ func (h *HugoSites) createMissingPages() error {
for _, s := range h.Sites {
// home pages
- home := s.findPagesByNodeType(NodeHome)
+ home := s.findPagesByNodeType(PageHome)
if len(home) > 1 {
panic("Too many homes")
}
@@ -236,8 +236,8 @@ func (h *HugoSites) createMissingPages() error {
// taxonomy list and terms pages
taxonomies := s.Language.GetStringMapString("taxonomies")
if len(taxonomies) > 0 {
- taxonomyPages := s.findPagesByNodeType(NodeTaxonomy)
- taxonomyTermsPages := s.findPagesByNodeType(NodeTaxonomyTerms)
+ taxonomyPages := s.findPagesByNodeType(PageTaxonomy)
+ taxonomyTermsPages := s.findPagesByNodeType(PageTaxonomyTerm)
for _, plural := range taxonomies {
tax := s.Taxonomies[plural]
foundTaxonomyPage := false
@@ -271,7 +271,7 @@ func (h *HugoSites) createMissingPages() error {
}
}
- sectionPages := s.findPagesByNodeType(NodeSection)
+ sectionPages := s.findPagesByNodeType(PageSection)
if len(sectionPages) < len(s.Sections) {
for name, section := range s.Sections {
// A section may be created for the root content folder if a
@@ -309,12 +309,12 @@ func (h *HugoSites) createMissingPages() error {
// TODO(bep) np move
// Move the new* methods after cleanup in site.go
-func (s *Site) newNodePage(typ NodeType) *Page {
+func (s *Site) newNodePage(typ PageType) *Page {
return &Page{
+ PageType: typ,
Node: Node{
Date: s.Info.LastChange,
Lastmod: s.Info.LastChange,
- NodeType: typ,
Data: make(map[string]interface{}),
Site: &s.Info,
language: s.Language,
@@ -322,9 +322,11 @@ func (s *Site) newNodePage(typ NodeType) *Page {
}
func (s *Site) newHomePage() *Page {
- p := s.newNodePage(NodeHome)
+ p := s.newNodePage(PageHome)
p.Title = s.Info.Title
- p.Data["Pages"] = Pages{}
+ pages := Pages{}
+ p.Data["Pages"] = pages
+ p.Pages = pages
s.setPageURLs(p, "/")
// TODO(bep) np check Data pages
// TODO(bep) np check setURLs
@@ -339,7 +341,7 @@ func (s *Site) setPageURLs(p *Page, in string) {
func (s *Site) newTaxonomyPage(plural, key string) *Page {
- p := s.newNodePage(NodeTaxonomy)
+ p := s.newNodePage(PageTaxonomy)
p.sections = []string{plural, key}
@@ -361,7 +363,7 @@ func (s *Site) newTaxonomyPage(plural, key string) *Page {
func (s *Site) newSectionPage(name string, section WeightedPages) *Page {
- p := s.newNodePage(NodeSection)
+ p := s.newNodePage(PageSection)
p.sections = []string{name}
sectionName := name
@@ -380,7 +382,7 @@ func (s *Site) newSectionPage(name string, section WeightedPages) *Page {
}
func (s *Site) newTaxonomyTermsPage(plural string) *Page {
- p := s.newNodePage(NodeTaxonomyTerms)
+ p := s.newNodePage(PageTaxonomyTerm)
p.sections = []string{plural}
p.Title = strings.Title(plural)
s.setPageURLs(p, plural)
@@ -571,19 +573,19 @@ func (s *Site) updateBuildStats(page *Page) {
}
// TODO(bep) np remove
-func (h *HugoSites) findAllPagesByNodeType(n NodeType) Pages {
+func (h *HugoSites) findAllPagesByNodeType(n PageType) Pages {
return h.Sites[0].findAllPagesByNodeType(n)
}
-func (h *HugoSites) findPagesByNodeTypeNotIn(n NodeType, inPages Pages) Pages {
+func (h *HugoSites) findPagesByNodeTypeNotIn(n PageType, inPages Pages) Pages {
return h.Sites[0].findPagesByNodeTypeNotIn(n, inPages)
}
-func (h *HugoSites) findPagesByNodeTypeIn(n NodeType, inPages Pages) Pages {
+func (h *HugoSites) findPagesByNodeTypeIn(n PageType, inPages Pages) Pages {
return h.Sites[0].findPagesByNodeTypeIn(n, inPages)
}
-func (h *HugoSites) findAllPagesByNodeTypeNotIn(n NodeType) Pages {
+func (h *HugoSites) findAllPagesByNodeTypeNotIn(n PageType) Pages {
return h.findPagesByNodeTypeNotIn(n, h.Sites[0].AllPages)
}
diff --git a/hugolib/hugo_sites_test.go b/hugolib/hugo_sites_test.go
index 1ed4744ca..a6118b5c3 100644
--- a/hugolib/hugo_sites_test.go
+++ b/hugolib/hugo_sites_test.go
@@ -297,7 +297,7 @@ func doTestMultiSitesBuild(t *testing.T, configTemplate, configSuffix string) {
assertFileContent(t, "public/en/sect/doc1-slug/index.html", true, "Single", "Shortcode: Hello")
// Check node translations
- homeEn := enSite.getPage(NodeHome)
+ homeEn := enSite.getPage(PageHome)
require.NotNil(t, homeEn)
require.Len(t, homeEn.Translations(), 3)
require.Equal(t, "fr", homeEn.Translations()[0].Lang())
@@ -307,7 +307,7 @@ func doTestMultiSitesBuild(t *testing.T, configTemplate, configSuffix string) {
require.Equal(t, "På bokmål", homeEn.Translations()[2].Title, configSuffix)
require.Equal(t, "Bokmål", homeEn.Translations()[2].Language().LanguageName, configSuffix)
- sectFr := frSite.getPage(NodeSection, "sect")
+ sectFr := frSite.getPage(PageSection, "sect")
require.NotNil(t, sectFr)
require.Equal(t, "fr", sectFr.Lang())
@@ -317,12 +317,12 @@ func doTestMultiSitesBuild(t *testing.T, configTemplate, configSuffix string) {
nnSite := sites.Sites[2]
require.Equal(t, "nn", nnSite.Language.Lang)
- taxNn := nnSite.getPage(NodeTaxonomyTerms, "lag")
+ taxNn := nnSite.getPage(PageTaxonomyTerm, "lag")
require.NotNil(t, taxNn)
require.Len(t, taxNn.Translations(), 1)
require.Equal(t, "nb", taxNn.Translations()[0].Lang())
- taxTermNn := nnSite.getPage(NodeTaxonomy, "lag", "sogndal")
+ taxTermNn := nnSite.getPage(PageTaxonomy, "lag", "sogndal")
require.NotNil(t, taxTermNn)
require.Len(t, taxTermNn.Translations(), 1)
require.Equal(t, "nb", taxTermNn.Translations()[0].Lang())
@@ -515,7 +515,7 @@ func TestMultiSitesRebuild(t *testing.T) {
docFr := readDestination(t, "public/fr/sect/doc1/index.html")
require.True(t, strings.Contains(docFr, "Salut"), "No Salut")
- homeEn := enSite.getPage(NodeHome)
+ homeEn := enSite.getPage(PageHome)
require.NotNil(t, homeEn)
require.Len(t, homeEn.Translations(), 3)
require.Equal(t, "fr", homeEn.Translations()[0].Lang())
@@ -622,7 +622,7 @@ title = "Svenska"
require.True(t, svSite.Language.Lang == "sv", svSite.Language.Lang)
require.True(t, frSite.Language.Lang == "fr", frSite.Language.Lang)
- homeEn := enSite.getPage(NodeHome)
+ homeEn := enSite.getPage(PageHome)
require.NotNil(t, homeEn)
require.Len(t, homeEn.Translations(), 4)
require.Equal(t, "sv", homeEn.Translations()[0].Lang())
diff --git a/hugolib/menu_test.go b/hugolib/menu_test.go
index bd186bd9f..0d5d81d29 100644
--- a/hugolib/menu_test.go
+++ b/hugolib/menu_test.go
@@ -415,11 +415,11 @@ func doTestSectionPagesMenu(canonifyURLs bool, t *testing.T) {
fishySectionPages := s.Sections["fish-and-chips"]
assert.Equal(t, 1, len(fishySectionPages))
- nodeFirst := s.getPage(NodeSection, "first")
+ nodeFirst := s.getPage(PageSection, "first")
require.NotNil(t, nodeFirst)
- nodeSecond := s.getPage(NodeSection, "second-section")
+ nodeSecond := s.getPage(PageSection, "second-section")
require.NotNil(t, nodeSecond)
- nodeFishy := s.getPage(NodeSection, "fish-and-chips")
+ nodeFishy := s.getPage(PageSection, "fish-and-chips")
require.Equal(t, "fish-and-chips", nodeFishy.sections[0])
firstSectionMenuEntry := findTestMenuEntryByID(s, "spm", "first")
@@ -555,7 +555,7 @@ func TestHomeNodeMenu(t *testing.T) {
s := setupMenuTests(t, menuPageSources)
- home := s.getPage(NodeHome)
+ home := s.getPage(PageHome)
homeMenuEntry := &MenuEntry{Name: home.Title, URL: home.URL()}
diff --git a/hugolib/node.go b/hugolib/node.go
index d77408545..ef16f9085 100644
--- a/hugolib/node.go
+++ b/hugolib/node.go
@@ -29,59 +29,7 @@ import (
// TODO(bep) np clean up node vs page
-type NodeType int
-
-const (
- NodePage NodeType = iota
-
- // Temporary state.
- NodeUnknown
-
- // The rest are node types; home page, sections etc.
- NodeHome
- NodeSection
- NodeTaxonomy
- NodeTaxonomyTerms
-
- // The following are (currently) temporary nodes,
- // i.e. nodes we create just to render in isolation.
- NodeSitemap
- NodeRobotsTXT
- Node404
-)
-
-func (p NodeType) String() string {
- switch p {
- case NodePage:
- return "page"
- case NodeHome:
- return "home page"
- case NodeSection:
- return "section list"
- case NodeTaxonomy:
- return "taxonomy list"
- case NodeTaxonomyTerms:
- return "taxonomy terms"
- case NodeSitemap:
- return "sitemap"
- case NodeRobotsTXT:
- return "robots.txt"
- case Node404:
- return "404 Not Found"
- case NodeUnknown:
- return "unknown"
- default:
- return "invalid value"
- }
-}
-
-func (p NodeType) IsNode() bool {
- return p >= NodeHome
-}
-
type Node struct {
- NodeType NodeType
-
RSSLink template.HTML
Site *SiteInfo `json:"-"`
// layout string
@@ -181,18 +129,6 @@ func (n *Node) RSSlink() template.HTML {
return n.RSSLink
}
-func (n *Node) IsNode() bool {
- return n.NodeType.IsNode()
-}
-
-func (n *Node) IsHome() bool {
- return n.NodeType == NodeHome
-}
-
-func (n *Node) IsPage() bool {
- return n.NodeType == NodePage
-}
-
func (n *Node) Ref(ref string) (string, error) {
return n.Site.Ref(ref, nil)
}
@@ -240,7 +176,7 @@ func (n *Node) Lang() string {
}
func (p *Page) isTranslation(candidate *Page) bool {
- if p == candidate || p.NodeType != candidate.NodeType {
+ if p == candidate || p.PageType != candidate.PageType {
return false
}
@@ -248,7 +184,7 @@ func (p *Page) isTranslation(candidate *Page) bool {
return false
}
- if p.NodeType == NodePage || p.NodeType == NodeUnknown {
+ if p.PageType == PagePage || p.PageType == pageUnknown {
panic("Node type not currently supported for this op")
}
@@ -355,41 +291,41 @@ func sectionsFromFilename(filename string) []string {
}
// TODO(bep) np node identificator
-func nodeTypeFromFilename(filename string) NodeType {
+func nodeTypeFromFilename(filename string) PageType {
if !strings.Contains(filename, "_index") {
- return NodePage
+ return PagePage
}
if strings.HasPrefix(filename, "_index") {
- return NodeHome
+ return PageHome
}
// We don't know enough yet to determine the type.
- return NodeUnknown
+ return pageUnknown
}
func (p *Page) setNodeTypeVars(s *Site) {
// TODO(bep) np taxonomies etc.
- if p.NodeType == NodeUnknown {
+ if p.PageType == pageUnknown {
// This is either a taxonomy list, taxonomy term or a section
nodeType := s.nodeTypeFromSections(p.sections)
- if nodeType == NodeUnknown {
+ if nodeType == pageUnknown {
panic(fmt.Sprintf("Unable to determine node type from %q", p.sections))
}
- p.NodeType = nodeType
+ p.PageType = nodeType
}
// TODO(bep) np node URL
// Set Node URL
- switch p.NodeType {
- case NodeHome:
+ switch p.PageType {
+ case PageHome:
p.URLPath.URL = ""
- case NodeSection:
+ case PageSection:
p.URLPath.URL = p.sections[0]
- case NodeTaxonomy:
+ case PageTaxonomy:
p.URLPath.URL = path.Join(p.sections...)
- case NodeTaxonomyTerms:
+ case PageTaxonomyTerm:
p.URLPath.URL = path.Join(p.sections...)
}
diff --git a/hugolib/node_as_page_test.go b/hugolib/node_as_page_test.go
index 04a3c30cf..78ccc6cef 100644
--- a/hugolib/node_as_page_test.go
+++ b/hugolib/node_as_page_test.go
@@ -84,7 +84,7 @@ Content Page %02d
assertFileContent(t, filepath.Join("public", "sect1", "regular1", "index.html"), false, "Single Title: Page 01", "Content Page 01")
h := s.owner
- nodes := h.findAllPagesByNodeType(NodeHome)
+ nodes := h.findAllPagesByNodeType(PageHome)
require.Len(t, nodes, 1)
home := nodes[0]
@@ -93,7 +93,7 @@ Content Page %02d
require.True(t, home.IsNode())
require.False(t, home.IsPage())
- pages := h.findAllPagesByNodeType(NodePage)
+ pages := h.findAllPagesByNodeType(PagePage)
require.Len(t, pages, 4)
first := pages[0]
@@ -115,7 +115,7 @@ Content Page %02d
assertFileContent(t, filepath.Join("public", "sect1", "page", "2", "index.html"), false,
"Pag: Page 02")
- sections := h.findAllPagesByNodeType(NodeSection)
+ sections := h.findAllPagesByNodeType(PageSection)
require.Len(t, sections, 2)
@@ -166,11 +166,12 @@ func TestNodesWithNoContentFile(t *testing.T) {
}
// Home page
- homePages := s.findIndexNodesByNodeType(NodeHome)
+ homePages := s.findIndexNodesByNodeType(PageHome)
require.Len(t, homePages, 1)
homePage := homePages[0]
require.Len(t, homePage.Data["Pages"], 9)
+ require.Len(t, homePage.Pages, 9) // Alias
assertFileContent(t, filepath.Join("public", "index.html"), false,
"Index Title: Hugo Rocks!")
diff --git a/hugolib/node_test.go b/hugolib/node_test.go
deleted file mode 100644
index 98c97cae6..000000000
--- a/hugolib/node_test.go
+++ /dev/null
@@ -1,40 +0,0 @@
-// Copyright 2016-present The Hugo Authors. All rights reserved.
-//
-// Licensed under the Apache License, Version 2.0 (the "License");
-// you may not use this file except in compliance with the License.
-// You may obtain a copy of the License at
-// http://www.apache.org/licenses/LICENSE-2.0
-//
-// Unless required by applicable law or agreed to in writing, software
-// distributed under the License is distributed on an "AS IS" BASIS,
-// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-// See the License for the specific language governing permissions and
-// limitations under the License.
-
-package hugolib
-
-import (
- "testing"
- "time"
-)
-
-func TestNodeSimpleMethods(t *testing.T) {
- for i, this := range []struct {
- assertFunc func(n *Node) bool
- }{
- {func(n *Node) bool { return n.IsNode() }},
- {func(n *Node) bool { return !n.IsPage() }},
- {func(n *Node) bool { return n.RSSlink() == "rssLink" }},
- {func(n *Node) bool { return n.Scratch() != nil }},
- {func(n *Node) bool { return n.Hugo() != nil }},
- {func(n *Node) bool { return n.Now().Unix() == time.Now().Unix() }},
- } {
-
- n := &Node{NodeType: NodeHome}
- n.RSSLink = "rssLink"
-
- if !this.assertFunc(n) {
- t.Errorf("[%d] Node method error", i)
- }
- }
-}
diff --git a/hugolib/page.go b/hugolib/page.go
index 6184aa0bc..28d4f7115 100644
--- a/hugolib/page.go
+++ b/hugolib/page.go
@@ -49,7 +49,46 @@ var (
cjk = regexp.MustCompile(`\p{Han}|\p{Hangul}|\p{Hiragana}|\p{Katakana}`)
)
+// PageType is the discriminator that identifies the different page types
+// in the different page collections. This can, as an example, be used
+// to to filter regular pages, find sections etc.
+// NOTE: THe exported constants below are used to filter pages from
+// templates in the wild, so do not change the values!
+type PageType string
+
+const (
+ PagePage PageType = "page"
+
+ // The rest are node types; home page, sections etc.
+ PageHome PageType = "home"
+ PageSection PageType = "section"
+ PageTaxonomy PageType = "taxonomy"
+ PageTaxonomyTerm PageType = "taxonomyTerm"
+
+ // Temporary state.
+ pageUnknown PageType = "unknown"
+
+ // The following are (currently) temporary nodes,
+ // i.e. nodes we create just to render in isolation.
+ pageSitemap PageType = "sitemap"
+ pageRobotsTXT PageType = "robotsTXT"
+ page404 PageType = "404"
+)
+
+func (p PageType) IsNode() bool {
+ return p != PagePage
+}
+
type Page struct {
+ PageType PageType
+
+ // Since Hugo 0.18 we got rid of the Node type. So now all pages are ...
+ // pages (regular pages, home page, sections etc.).
+ // Sections etc. will have child pages. These were earlier placed in .Data.Pages,
+ // but can now be more intuitively also be fetched directly from .Pages.
+ // This collection will be nil for regular pages.
+ Pages Pages
+
Params map[string]interface{}
Content template.HTML
Summary template.HTML
@@ -146,6 +185,18 @@ func (*PageMeta) ReadingTime() int {
return 0
}
+func (p *Page) IsNode() bool {
+ return p.PageType.IsNode()
+}
+
+func (p *Page) IsHome() bool {
+ return p.PageType == PageHome
+}
+
+func (p *Page) IsPage() bool {
+ return p.PageType == PagePage
+}
+
type Position struct {
Prev *Page
Next *Page
@@ -430,9 +481,11 @@ func (p *Page) getRenderingConfig() *helpers.Blackfriday {
}
func newPage(filename string) *Page {
- page := Page{contentType: "",
+ page := Page{
+ PageType: nodeTypeFromFilename(filename),
+ contentType: "",
Source: Source{File: *source.NewFile(filename)},
- Node: Node{NodeType: nodeTypeFromFilename(filename), Keywords: []string{}, Sitemap: Sitemap{Priority: -1}},
+ Node: Node{Keywords: []string{}, Sitemap: Sitemap{Priority: -1}},
Params: make(map[string]interface{}),
translations: make(Pages, 0),
sections: sectionsFromFilename(filename),
@@ -468,16 +521,16 @@ func (p *Page) layouts(l ...string) []string {
}
// TODO(bep) np taxonomy etc.
- switch p.NodeType {
- case NodeHome:
+ switch p.PageType {
+ case PageHome:
return []string{"index.html", "_default/list.html"}
- case NodeSection:
+ case PageSection:
section := p.sections[0]
return []string{"section/" + section + ".html", "_default/section.html", "_default/list.html", "indexes/" + section + ".html", "_default/indexes.html"}
- case NodeTaxonomy:
+ case PageTaxonomy:
singular := p.site.taxonomiesPluralSingular[p.sections[0]]
return []string{"taxonomy/" + singular + ".html", "indexes/" + singular + ".html", "_default/taxonomy.html", "_default/list.html"}
- case NodeTaxonomyTerms:
+ case PageTaxonomyTerm:
singular := p.site.taxonomiesPluralSingular[p.sections[0]]
return []string{"taxonomy/" + singular + ".terms.html", "_default/terms.html", "indexes/indexes.html"}
}
@@ -502,18 +555,18 @@ func (p *Page) layouts(l ...string) []string {
// rssLayouts returns RSS layouts to use for the RSS version of this page, nil
// if no RSS should be rendered.
func (p *Page) rssLayouts() []string {
- switch p.NodeType {
- case NodeHome:
+ switch p.PageType {
+ case PageHome:
return []string{"rss.xml", "_default/rss.xml", "_internal/_default/rss.xml"}
- case NodeSection:
+ case PageSection:
section := p.sections[0]
return []string{"section/" + section + ".rss.xml", "_default/rss.xml", "rss.xml", "_internal/_default/rss.xml"}
- case NodeTaxonomy:
+ case PageTaxonomy:
singular := p.site.taxonomiesPluralSingular[p.sections[0]]
return []string{"taxonomy/" + singular + ".rss.xml", "_default/rss.xml", "rss.xml", "_internal/_default/rss.xml"}
- case NodeTaxonomyTerms:
+ case PageTaxonomyTerm:
// No RSS for taxonomy terms
- case NodePage:
+ case PagePage:
// No RSS for regular pages
}
@@ -723,7 +776,7 @@ func (p *Page) IsExpired() bool {
func (p *Page) Permalink() (string, error) {
// TODO(bep) np permalink
- if p.NodeType.IsNode() {
+ if p.PageType.IsNode() {
return p.Node.Permalink(), nil
}
link, err := p.permalink()
@@ -961,7 +1014,7 @@ func (p *Page) getParam(key string, stringToLower bool) interface{} {
func (p *Page) HasMenuCurrent(menu string, me *MenuEntry) bool {
// TODO(bep) np menu
- if p.NodeType.IsNode() {
+ if p.PageType.IsNode() {
return p.Node.HasMenuCurrent(menu, me)
}
menus := p.Menus()
@@ -991,7 +1044,7 @@ func (p *Page) HasMenuCurrent(menu string, me *MenuEntry) bool {
func (p *Page) IsMenuCurrent(menu string, inme *MenuEntry) bool {
// TODO(bep) np menu
- if p.NodeType.IsNode() {
+ if p.PageType.IsNode() {
return p.Node.IsMenuCurrent(menu, inme)
}
menus := p.Menus()
@@ -1200,14 +1253,14 @@ func (p *Page) FullFilePath() string {
func (p *Page) TargetPath() (outfile string) {
// TODO(bep) np
- switch p.NodeType {
- case NodeHome:
+ switch p.PageType {
+ case PageHome:
return p.addLangFilepathPrefix(helpers.FilePathSeparator)
- case NodeSection:
+ case PageSection:
return p.addLangFilepathPrefix(p.sections[0])
- case NodeTaxonomy:
+ case PageTaxonomy:
return p.addLangFilepathPrefix(filepath.Join(p.sections...))
- case NodeTaxonomyTerms:
+ case PageTaxonomyTerm:
return p.addLangFilepathPrefix(filepath.Join(p.sections...))
}
@@ -1252,7 +1305,7 @@ func (p *Page) TargetPath() (outfile string) {
func (p *Page) prepareLayouts() error {
// TODO(bep): Check the IsRenderable logic.
- if p.NodeType == NodePage {
+ if p.PageType == PagePage {
var layouts []string
if !p.IsRenderable() {
self := "__" + p.TargetPath()
@@ -1273,18 +1326,20 @@ func (p *Page) prepareLayouts() error {
// TODO(bep) np naming, move some
func (p *Page) prepareData(s *Site) error {
+ var pages Pages
+
p.Data = make(map[string]interface{})
- switch p.NodeType {
- case NodePage:
- case NodeHome:
- p.Data["Pages"] = s.findPagesByNodeTypeNotIn(NodeHome, s.Pages)
- case NodeSection:
+ switch p.PageType {
+ case PagePage:
+ case PageHome:
+ pages = s.findPagesByNodeTypeNotIn(PageHome, s.Pages)
+ case PageSection:
sectionData, ok := s.Sections[p.sections[0]]
if !ok {
return fmt.Errorf("Data for section %s not found", p.Section())
}
- p.Data["Pages"] = sectionData.Pages()
- case NodeTaxonomy:
+ pages = sectionData.Pages()
+ case PageTaxonomy:
plural := p.sections[0]
term := p.sections[1]
@@ -1294,8 +1349,8 @@ func (p *Page) prepareData(s *Site) error {
p.Data[singular] = taxonomy
p.Data["Singular"] = singular
p.Data["Plural"] = plural
- p.Data["Pages"] = taxonomy.Pages()
- case NodeTaxonomyTerms:
+ pages = taxonomy.Pages()
+ case PageTaxonomyTerm:
plural := p.sections[0]
singular := s.taxonomiesPluralSingular[plural]
@@ -1307,6 +1362,9 @@ func (p *Page) prepareData(s *Site) error {
p.Data["Index"] = p.Data["Terms"]
}
+ p.Data["Pages"] = pages
+ p.Pages = pages
+
return nil
}
@@ -1315,7 +1373,7 @@ func (p *Page) prepareData(s *Site) error {
// the paginators etc., we do it manually here.
// TODO(bep) np do better
func (p *Page) copy() *Page {
- c := &Page{Node: Node{NodeType: p.NodeType, Site: p.Site}}
+ c := &Page{PageType: p.PageType, Node: Node{Site: p.Site}}
c.Title = p.Title
c.Data = p.Data
c.Date = p.Date
diff --git a/hugolib/page_collections.go b/hugolib/page_collections.go
index 71e40d964..e1ba36c18 100644
--- a/hugolib/page_collections.go
+++ b/hugolib/page_collections.go
@@ -40,12 +40,12 @@ type PageCollections struct {
}
func (c *PageCollections) refreshPageCaches() {
- c.indexPages = c.findPagesByNodeTypeNotIn(NodePage, c.Pages)
- c.regularPages = c.findPagesByNodeTypeIn(NodePage, c.Pages)
+ c.indexPages = c.findPagesByNodeTypeNotIn(PagePage, c.Pages)
+ c.regularPages = c.findPagesByNodeTypeIn(PagePage, c.Pages)
// TODO(bep) np remove eventually
for _, n := range c.Pages {
- if n.NodeType == NodeUnknown {
+ if n.PageType == pageUnknown {
panic(fmt.Sprintf("Got unknown type %s", n.Title))
}
}
@@ -61,11 +61,11 @@ func newPageCollectionsFromPages(pages Pages) *PageCollections {
// TODO(bep) np clean and remove finders
-func (c *PageCollections) findPagesByNodeType(n NodeType) Pages {
+func (c *PageCollections) findPagesByNodeType(n PageType) Pages {
return c.findPagesByNodeTypeIn(n, c.Pages)
}
-func (c *PageCollections) getPage(n NodeType, path ...string) *Page {
+func (c *PageCollections) getPage(n PageType, path ...string) *Page {
pages := c.findPagesByNodeTypeIn(n, c.Pages)
if len(pages) == 0 {
@@ -94,35 +94,35 @@ func (c *PageCollections) getPage(n NodeType, path ...string) *Page {
return nil
}
-func (c *PageCollections) findIndexNodesByNodeType(n NodeType) Pages {
+func (c *PageCollections) findIndexNodesByNodeType(n PageType) Pages {
return c.findPagesByNodeTypeIn(n, c.indexPages)
}
-func (*PageCollections) findPagesByNodeTypeIn(n NodeType, inPages Pages) Pages {
+func (*PageCollections) findPagesByNodeTypeIn(n PageType, inPages Pages) Pages {
var pages Pages
for _, p := range inPages {
- if p.NodeType == n {
+ if p.PageType == n {
pages = append(pages, p)
}
}
return pages
}
-func (*PageCollections) findPagesByNodeTypeNotIn(n NodeType, inPages Pages) Pages {
+func (*PageCollections) findPagesByNodeTypeNotIn(n PageType, inPages Pages) Pages {
var pages Pages
for _, p := range inPages {
- if p.NodeType != n {
+ if p.PageType != n {
pages = append(pages, p)
}
}
return pages
}
-func (c *PageCollections) findAllPagesByNodeType(n NodeType) Pages {
+func (c *PageCollections) findAllPagesByNodeType(n PageType) Pages {
return c.findPagesByNodeTypeIn(n, c.Pages)
}
-func (c *PageCollections) findRawAllPagesByNodeType(n NodeType) Pages {
+func (c *PageCollections) findRawAllPagesByNodeType(n PageType) Pages {
return c.findPagesByNodeTypeIn(n, c.rawAllPages)
}
diff --git a/hugolib/page_permalink_test.go b/hugolib/page_permalink_test.go
index fd8f11dda..2a24dc642 100644
--- a/hugolib/page_permalink_test.go
+++ b/hugolib/page_permalink_test.go
@@ -66,6 +66,7 @@ func TestPermalink(t *testing.T) {
info := newSiteInfo(siteBuilderCfg{baseURL: string(test.base), language: helpers.NewDefaultLanguage()})
p := &Page{
+ PageType: PagePage,
Node: Node{
URLPath: URLPath{
Section: "z",
diff --git a/hugolib/page_test.go b/hugolib/page_test.go
index a63ca1c9f..89979b8f2 100644
--- a/hugolib/page_test.go
+++ b/hugolib/page_test.go
@@ -1257,6 +1257,40 @@ func TestPageSimpleMethods(t *testing.T) {
}
}
+func TestIndexPageSimpleMethods(t *testing.T) {
+ for i, this := range []struct {
+ assertFunc func(n *Page) bool
+ }{
+ {func(n *Page) bool { return n.IsNode() }},
+ {func(n *Page) bool { return !n.IsPage() }},
+ {func(n *Page) bool { return n.RSSlink() == "rssLink" }},
+ {func(n *Page) bool { return n.Scratch() != nil }},
+ {func(n *Page) bool { return n.Hugo() != nil }},
+ {func(n *Page) bool { return n.Now().Unix() == time.Now().Unix() }},
+ } {
+
+ n := &Page{PageType: PageHome}
+ n.RSSLink = "rssLink"
+
+ if !this.assertFunc(n) {
+ t.Errorf("[%d] Node method error", i)
+ }
+ }
+}
+
+func TestPageType(t *testing.T) {
+
+ // Add tests for these constants to make sure they don't change
+ require.Equal(t, PageType("page"), PagePage)
+ require.Equal(t, PageType("home"), PageHome)
+ require.Equal(t, PageType("section"), PageSection)
+ require.Equal(t, PageType("taxonomy"), PageTaxonomy)
+ require.Equal(t, PageType("taxonomyTerm"), PageTaxonomyTerm)
+
+ require.False(t, PagePage.IsNode())
+ require.True(t, PageHome.IsNode())
+}
+
func TestChompBOM(t *testing.T) {
const utf8BOM = "\xef\xbb\xbf"
diff --git a/hugolib/pagination.go b/hugolib/pagination.go
index bb2142cb8..47bde38dc 100644
--- a/hugolib/pagination.go
+++ b/hugolib/pagination.go
@@ -260,8 +260,8 @@ func splitPageGroups(pageGroups PagesGroup, size int) []paginatedElement {
// Paginator gets this Node's paginator if it's already created.
// If it's not, one will be created with all pages in Data["Pages"].
func (n *Page) Paginator(options ...interface{}) (*Pager, error) {
- if !n.NodeType.IsNode() {
- return nil, fmt.Errorf("Paginators not supported for pages of type %q (%q)", n.NodeType, n.Title)
+ if !n.PageType.IsNode() {
+ return nil, fmt.Errorf("Paginators not supported for pages of type %q (%q)", n.PageType, n.Title)
}
pagerSize, err := resolvePagerSize(options...)
@@ -303,8 +303,8 @@ func (n *Page) Paginator(options ...interface{}) (*Pager, error) {
// If it's not, one will be created with the qiven sequence.
// Note that repeated calls will return the same result, even if the sequence is different.
func (n *Page) Paginate(seq interface{}, options ...interface{}) (*Pager, error) {
- if !n.NodeType.IsNode() {
- return nil, fmt.Errorf("Paginators not supported for pages of type %q (%q)", n.NodeType, n.Title)
+ if !n.PageType.IsNode() {
+ return nil, fmt.Errorf("Paginators not supported for pages of type %q (%q)", n.PageType, n.Title)
}
pagerSize, err := resolvePagerSize(options...)
diff --git a/hugolib/site.go b/hugolib/site.go
index d93fbf650..910f887e5 100644
--- a/hugolib/site.go
+++ b/hugolib/site.go
@@ -1533,8 +1533,8 @@ func (s *Site) assembleSections() {
s.Sections = make(Taxonomy)
s.Info.Sections = s.Sections
// TODO(bep) np check these vs the caches
- regularPages := s.findPagesByNodeType(NodePage)
- sectionPages := s.findPagesByNodeType(NodeSection)
+ regularPages := s.findPagesByNodeType(PagePage)
+ sectionPages := s.findPagesByNodeType(PageSection)
for i, p := range regularPages {
s.Sections.add(p.Section(), WeightedPage{regularPages[i].Weight, regularPages[i]}, s.Info.preserveTaxonomyNames)
@@ -1561,14 +1561,14 @@ func (s *Site) assembleSections() {
}
}
-func (s *Site) nodeTypeFromSections(sections []string) NodeType {
+func (s *Site) nodeTypeFromSections(sections []string) PageType {
if _, isTaxonomy := s.Taxonomies[sections[0]]; isTaxonomy {
if len(sections) == 1 {
- return NodeTaxonomyTerms
+ return PageTaxonomyTerm
}
- return NodeTaxonomy
+ return PageTaxonomy
}
- return NodeSection
+ return PageSection
}
func (s *Site) preparePages() error {
diff --git a/hugolib/site_render.go b/hugolib/site_render.go
index fe7db05f3..2cb921c86 100644
--- a/hugolib/site_render.go
+++ b/hugolib/site_render.go
@@ -67,14 +67,14 @@ func pageRenderer(s *Site, pages <-chan *Page, results chan<- error, wg *sync.Wa
for p := range pages {
targetPath := p.TargetPath()
layouts := p.layouts()
- jww.DEBUG.Printf("Render %s to %q with layouts %q", p.NodeType, targetPath, layouts)
+ jww.DEBUG.Printf("Render %s to %q with layouts %q", p.PageType, targetPath, layouts)
if err := s.renderAndWritePage("page "+p.FullFilePath(), targetPath, p, s.appendThemeTemplates(layouts)...); err != nil {
results <- err
}
// Taxonomy terms have no page set to paginate, so skip that for now.
- if p.NodeType.IsNode() && p.NodeType != NodeTaxonomyTerms {
+ if p.PageType.IsNode() && p.PageType != PageTaxonomyTerm {
if err := s.renderPaginator(p); err != nil {
results <- err
}
@@ -160,9 +160,10 @@ func (s *Site) render404() error {
return nil
}
- p := s.newNodePage(Node404)
+ p := s.newNodePage(page404)
p.Title = "404 Page not found"
p.Data["Pages"] = s.Pages
+ p.Pages = s.Pages
s.setPageURLs(p, "404.html")
nfLayouts := []string{"404.html"}
@@ -180,18 +181,19 @@ func (s *Site) renderSitemap() error {
sitemapDefault := parseSitemap(viper.GetStringMap("sitemap"))
- n := s.newNodePage(NodeSitemap)
+ n := s.newNodePage(pageSitemap)
// Include all pages (regular, home page, taxonomies etc.)
pages := s.Pages
- page := s.newNodePage(NodeSitemap)
+ page := s.newNodePage(pageSitemap)
page.URLPath.URL = ""
page.Sitemap.ChangeFreq = sitemapDefault.ChangeFreq
page.Sitemap.Priority = sitemapDefault.Priority
page.Sitemap.Filename = sitemapDefault.Filename
n.Data["Pages"] = pages
+ n.Pages = pages
// TODO(bep) this should be done somewhere else
for _, page := range pages {
@@ -222,8 +224,9 @@ func (s *Site) renderRobotsTXT() error {
return nil
}
- n := s.newNodePage(NodeRobotsTXT)
+ n := s.newNodePage(pageRobotsTXT)
n.Data["Pages"] = s.Pages
+ n.Pages = s.Pages
rLayouts := []string{"robots.txt", "_default/robots.txt", "_internal/_default/robots.txt"}
outBuffer := bp.GetBuffer()