Rewrite markdown rendering to blackfriday v2 and rewrite orgmode rendering to go-org (#8560)
* Rewrite markdown rendering to blackfriday v2.0 * Fix style * Fix go mod with golang 1.13 * Fix blackfriday v2 import * Inital orgmode renderer migration to go-org * Vendor go-org dependency * Ignore errors :/ * Update go-org to latest version * Update test * Fix go-org test * Remove unneeded code * Fix comments * Fix markdown test * Fix blackfriday regression rendering HTML blocktokarchuk/v1.17
parent
690a8ec502
commit
086a46994a
@ -1 +0,0 @@ |
||||
.DS_Store |
@ -1,12 +0,0 @@ |
||||
language: go |
||||
|
||||
go: |
||||
- 1.7 |
||||
|
||||
before_install: |
||||
- go get golang.org/x/tools/cmd/cover |
||||
- go get github.com/mattn/goveralls |
||||
|
||||
script: |
||||
- go test -v -covermode=count -coverprofile=coverage.out |
||||
- $HOME/gopath/bin/goveralls -coverprofile=coverage.out -service=travis-ci |
@ -1,66 +0,0 @@ |
||||
#+TITLE: chaseadamsio/goorgeous |
||||
|
||||
[[https://travis-ci.org/chaseadamsio/goorgeous.svg?branch=master]] |
||||
[[https://coveralls.io/repos/github/chaseadamsio/goorgeous/badge.svg?branch=master]] |
||||
|
||||
/goorgeous is a Go Org to HTML Parser./ |
||||
|
||||
[[file:gopher_small.gif]] |
||||
|
||||
*Pronounced: Go? Org? Yes!* |
||||
|
||||
#+BEGIN_QUOTE |
||||
"Org mode is for keeping notes, maintaining TODO lists, planning projects, and authoring documents with a fast and effective plain-text system." |
||||
|
||||
- [[orgmode.org]] |
||||
#+END_QUOTE |
||||
|
||||
The purpose of this package is to come as close as possible as parsing an =*.org= document into HTML, the same way one might publish [[http://orgmode.org/worg/org-tutorials/org-publish-html-tutorial.html][with org-publish-html from Emacs]]. |
||||
|
||||
* Installation |
||||
|
||||
#+BEGIN_SRC sh |
||||
go get -u github.com/chaseadamsio/goorgeous |
||||
#+END_SRC |
||||
|
||||
* Usage |
||||
|
||||
** Org Headers |
||||
|
||||
To retrieve the headers from a =[]byte=, call =OrgHeaders= and it will return a =map[string]interface{}=: |
||||
|
||||
#+BEGIN_SRC go |
||||
input := "#+title: goorgeous\n* Some Headline\n" |
||||
out := goorgeous.OrgHeaders(input) |
||||
#+END_SRC |
||||
|
||||
#+BEGIN_SRC go |
||||
map[string]interface{}{ |
||||
"title": "goorgeous" |
||||
} |
||||
#+END_SRC |
||||
|
||||
** Org Content |
||||
|
||||
After importing =github.com/chaseadamsio/goorgeous=, you can call =Org= with a =[]byte= and it will return an =html= version of the content as a =[]byte= |
||||
|
||||
#+BEGIN_SRC go |
||||
input := "#+TITLE: goorgeous\n* Some Headline\n" |
||||
out := goorgeous.Org(input) |
||||
#+END_SRC |
||||
|
||||
=out= will be: |
||||
|
||||
#+BEGIN_SRC html |
||||
<h1>Some Headline</h1>/n |
||||
#+END_SRC |
||||
|
||||
* Why? |
||||
|
||||
First off, I've become an unapologetic user of Emacs & ever since finding =org-mode= I use it for anything having to do with writing content, organizing my life and keeping documentation of my days/weeks/months. |
||||
|
||||
Although I like Emacs & =emacs-lisp=, I publish all of my html sites with [[https://gohugo.io][Hugo Static Site Generator]] and wanted to be able to write my content in =org-mode= in Emacs rather than markdown. |
||||
|
||||
Hugo's implementation of templating and speed are unmatched, so the only way I knew for sure I could continue to use Hugo and write in =org-mode= seamlessly was to write a golang parser for org content and submit a PR for Hugo to use it. |
||||
* Acknowledgements |
||||
I leaned heavily on russross' [[https://github.com/russross/blackfriday][blackfriday markdown renderer]] as both an example of how to write a parser (with some updates to leverage the go we know today) and reusing the blackfriday HTML Renderer so I didn't have to write my own! |
@ -1,803 +0,0 @@ |
||||
package goorgeous |
||||
|
||||
import ( |
||||
"bufio" |
||||
"bytes" |
||||
"regexp" |
||||
|
||||
"github.com/russross/blackfriday" |
||||
"github.com/shurcooL/sanitized_anchor_name" |
||||
) |
||||
|
||||
type inlineParser func(p *parser, out *bytes.Buffer, data []byte, offset int) int |
||||
|
||||
type footnotes struct { |
||||
id string |
||||
def string |
||||
} |
||||
|
||||
type parser struct { |
||||
r blackfriday.Renderer |
||||
inlineCallback [256]inlineParser |
||||
notes []footnotes |
||||
} |
||||
|
||||
// NewParser returns a new parser with the inlineCallbacks required for org content
|
||||
func NewParser(renderer blackfriday.Renderer) *parser { |
||||
p := new(parser) |
||||
p.r = renderer |
||||
|
||||
p.inlineCallback['='] = generateVerbatim |
||||
p.inlineCallback['~'] = generateCode |
||||
p.inlineCallback['/'] = generateEmphasis |
||||
p.inlineCallback['_'] = generateUnderline |
||||
p.inlineCallback['*'] = generateBold |
||||
p.inlineCallback['+'] = generateStrikethrough |
||||
p.inlineCallback['['] = generateLinkOrImg |
||||
|
||||
return p |
||||
} |
||||
|
||||
// OrgCommon is the easiest way to parse a byte slice of org content and makes assumptions
|
||||
// that the caller wants to use blackfriday's HTMLRenderer with XHTML
|
||||
func OrgCommon(input []byte) []byte { |
||||
renderer := blackfriday.HtmlRenderer(blackfriday.HTML_USE_XHTML, "", "") |
||||
return OrgOptions(input, renderer) |
||||
} |
||||
|
||||
// Org is a convenience name for OrgOptions
|
||||
func Org(input []byte, renderer blackfriday.Renderer) []byte { |
||||
return OrgOptions(input, renderer) |
||||
} |
||||
|
||||
// OrgOptions takes an org content byte slice and a renderer to use
|
||||
func OrgOptions(input []byte, renderer blackfriday.Renderer) []byte { |
||||
// in the case that we need to render something in isEmpty but there isn't a new line char
|
||||
input = append(input, '\n') |
||||
var output bytes.Buffer |
||||
|
||||
p := NewParser(renderer) |
||||
|
||||
scanner := bufio.NewScanner(bytes.NewReader(input)) |
||||
// used to capture code blocks
|
||||
marker := "" |
||||
syntax := "" |
||||
listType := "" |
||||
inParagraph := false |
||||
inList := false |
||||
inTable := false |
||||
inFixedWidthArea := false |
||||
var tmpBlock bytes.Buffer |
||||
|
||||
for scanner.Scan() { |
||||
data := scanner.Bytes() |
||||
|
||||
if !isEmpty(data) && isComment(data) || IsKeyword(data) { |
||||
switch { |
||||
case inList: |
||||
if tmpBlock.Len() > 0 { |
||||
p.generateList(&output, tmpBlock.Bytes(), listType) |
||||
} |
||||
inList = false |
||||
listType = "" |
||||
tmpBlock.Reset() |
||||
case inTable: |
||||
if tmpBlock.Len() > 0 { |
||||
p.generateTable(&output, tmpBlock.Bytes()) |
||||
} |
||||
inTable = false |
||||
tmpBlock.Reset() |
||||
case inParagraph: |
||||
if tmpBlock.Len() > 0 { |
||||
p.generateParagraph(&output, tmpBlock.Bytes()[:len(tmpBlock.Bytes())-1]) |
||||
} |
||||
inParagraph = false |
||||
tmpBlock.Reset() |
||||
case inFixedWidthArea: |
||||
if tmpBlock.Len() > 0 { |
||||
tmpBlock.WriteString("</pre>\n") |
||||
output.Write(tmpBlock.Bytes()) |
||||
} |
||||
inFixedWidthArea = false |
||||
tmpBlock.Reset() |
||||
} |
||||
|
||||
} |
||||
|
||||
switch { |
||||
case isEmpty(data): |
||||
switch { |
||||
case inList: |
||||
if tmpBlock.Len() > 0 { |
||||
p.generateList(&output, tmpBlock.Bytes(), listType) |
||||
} |
||||
inList = false |
||||
listType = "" |
||||
tmpBlock.Reset() |
||||
case inTable: |
||||
if tmpBlock.Len() > 0 { |
||||
p.generateTable(&output, tmpBlock.Bytes()) |
||||
} |
||||
inTable = false |
||||
tmpBlock.Reset() |
||||
case inParagraph: |
||||
if tmpBlock.Len() > 0 { |
||||
p.generateParagraph(&output, tmpBlock.Bytes()[:len(tmpBlock.Bytes())-1]) |
||||
} |
||||
inParagraph = false |
||||
tmpBlock.Reset() |
||||
case inFixedWidthArea: |
||||
if tmpBlock.Len() > 0 { |
||||
tmpBlock.WriteString("</pre>\n") |
||||
output.Write(tmpBlock.Bytes()) |
||||
} |
||||
inFixedWidthArea = false |
||||
tmpBlock.Reset() |
||||
case marker != "": |
||||
tmpBlock.WriteByte('\n') |
||||
default: |
||||
continue |
||||
} |
||||
case isPropertyDrawer(data) || marker == "PROPERTIES": |
||||
if marker == "" { |
||||
marker = "PROPERTIES" |
||||
} |
||||
if bytes.Equal(data, []byte(":END:")) { |
||||
marker = "" |
||||
} |
||||
continue |
||||
case isBlock(data) || marker != "": |
||||
matches := reBlock.FindSubmatch(data) |
||||
if len(matches) > 0 { |
||||
if string(matches[1]) == "END" { |
||||
switch marker { |
||||
case "QUOTE": |
||||
var tmpBuf bytes.Buffer |
||||
p.inline(&tmpBuf, tmpBlock.Bytes()) |
||||
p.r.BlockQuote(&output, tmpBuf.Bytes()) |
||||
case "CENTER": |
||||
var tmpBuf bytes.Buffer |
||||
output.WriteString("<center>\n") |
||||
p.inline(&tmpBuf, tmpBlock.Bytes()) |
||||
output.Write(tmpBuf.Bytes()) |
||||
output.WriteString("</center>\n") |
||||
default: |
||||
tmpBlock.WriteByte('\n') |
||||
p.r.BlockCode(&output, tmpBlock.Bytes(), syntax) |
||||
} |
||||
marker = "" |
||||
tmpBlock.Reset() |
||||
continue |
||||
} |
||||
|
||||
} |
||||
if marker != "" { |
||||
if marker != "SRC" && marker != "EXAMPLE" { |
||||
var tmpBuf bytes.Buffer |
||||
tmpBuf.Write([]byte("<p>\n")) |
||||
p.inline(&tmpBuf, data) |
||||
tmpBuf.WriteByte('\n') |
||||
tmpBuf.Write([]byte("</p>\n")) |
||||
tmpBlock.Write(tmpBuf.Bytes()) |
||||
|
||||
} else { |
||||
tmpBlock.WriteByte('\n') |
||||
tmpBlock.Write(data) |
||||
} |
||||
|
||||
} else { |
||||
marker = string(matches[2]) |
||||
syntax = string(matches[3]) |
||||
} |
||||
case isFootnoteDef(data): |
||||
matches := reFootnoteDef.FindSubmatch(data) |
||||
for i := range p.notes { |
||||
if p.notes[i].id == string(matches[1]) { |
||||
p.notes[i].def = string(matches[2]) |
||||
} |
||||
} |
||||
case isTable(data): |
||||
if inTable != true { |
||||
inTable = true |
||||
} |
||||
tmpBlock.Write(data) |
||||
tmpBlock.WriteByte('\n') |
||||
case IsKeyword(data): |
||||
continue |
||||
case isComment(data): |
||||
p.generateComment(&output, data) |
||||
case isHeadline(data): |
||||
p.generateHeadline(&output, data) |
||||
case isDefinitionList(data): |
||||
if inList != true { |
||||
listType = "dl" |
||||
inList = true |
||||
} |
||||
var work bytes.Buffer |
||||
flags := blackfriday.LIST_TYPE_DEFINITION |
||||
matches := reDefinitionList.FindSubmatch(data) |
||||
flags |= blackfriday.LIST_TYPE_TERM |
||||
p.inline(&work, matches[1]) |
||||
p.r.ListItem(&tmpBlock, work.Bytes(), flags) |
||||
work.Reset() |
||||
flags &= ^blackfriday.LIST_TYPE_TERM |
||||
p.inline(&work, matches[2]) |
||||
p.r.ListItem(&tmpBlock, work.Bytes(), flags) |
||||
case isUnorderedList(data): |
||||
if inList != true { |
||||
listType = "ul" |
||||
inList = true |
||||
} |
||||
matches := reUnorderedList.FindSubmatch(data) |
||||
var work bytes.Buffer |
||||
p.inline(&work, matches[2]) |
||||
p.r.ListItem(&tmpBlock, work.Bytes(), 0) |
||||
case isOrderedList(data): |
||||
if inList != true { |
||||
listType = "ol" |
||||
inList = true |
||||
} |
||||
matches := reOrderedList.FindSubmatch(data) |
||||
var work bytes.Buffer |
||||
tmpBlock.WriteString("<li") |
||||
if len(matches[2]) > 0 { |
||||
tmpBlock.WriteString(" value=\"") |
||||
tmpBlock.Write(matches[2]) |
||||
tmpBlock.WriteString("\"") |
||||
matches[3] = matches[3][1:] |
||||
} |
||||
p.inline(&work, matches[3]) |
||||
tmpBlock.WriteString(">") |
||||
tmpBlock.Write(work.Bytes()) |
||||
tmpBlock.WriteString("</li>\n") |
||||
case isHorizontalRule(data): |
||||
p.r.HRule(&output) |
||||
case isExampleLine(data): |
||||
if inParagraph == true { |
||||
if len(tmpBlock.Bytes()) > 0 { |
||||
p.generateParagraph(&output, tmpBlock.Bytes()[:len(tmpBlock.Bytes())-1]) |
||||
inParagraph = false |
||||
} |
||||
tmpBlock.Reset() |
||||
} |
||||
if inFixedWidthArea != true { |
||||
tmpBlock.WriteString("<pre class=\"example\">\n") |
||||
inFixedWidthArea = true |
||||
} |
||||
matches := reExampleLine.FindSubmatch(data) |
||||
tmpBlock.Write(matches[1]) |
||||
tmpBlock.WriteString("\n") |
||||
break |
||||
default: |
||||
if inParagraph == false { |
||||
inParagraph = true |
||||
if inFixedWidthArea == true { |
||||
if tmpBlock.Len() > 0 { |
||||
tmpBlock.WriteString("</pre>") |
||||
output.Write(tmpBlock.Bytes()) |
||||
} |
||||
inFixedWidthArea = false |
||||
tmpBlock.Reset() |
||||
} |
||||
} |
||||
tmpBlock.Write(data) |
||||
tmpBlock.WriteByte('\n') |
||||
} |
||||
} |
||||
|
||||
if len(tmpBlock.Bytes()) > 0 { |
||||
if inParagraph == true { |
||||
p.generateParagraph(&output, tmpBlock.Bytes()[:len(tmpBlock.Bytes())-1]) |
||||
} else if inFixedWidthArea == true { |
||||
tmpBlock.WriteString("</pre>\n") |
||||
output.Write(tmpBlock.Bytes()) |
||||
} |
||||
} |
||||
|
||||
// Writing footnote def. list
|
||||
if len(p.notes) > 0 { |
||||
flags := blackfriday.LIST_ITEM_BEGINNING_OF_LIST |
||||
p.r.Footnotes(&output, func() bool { |
||||
for i := range p.notes { |
||||
p.r.FootnoteItem(&output, []byte(p.notes[i].id), []byte(p.notes[i].def), flags) |
||||
} |
||||
return true |
||||
}) |
||||
} |
||||
|
||||
return output.Bytes() |
||||
} |
||||
|
||||
// Org Syntax has been broken up into 4 distinct sections based on
|
||||
// the org-syntax draft (http://orgmode.org/worg/dev/org-syntax.html):
|
||||
// - Headlines
|
||||
// - Greater Elements
|
||||
// - Elements
|
||||
// - Objects
|
||||
|
||||
// Headlines
|
||||
func isHeadline(data []byte) bool { |
||||
if !charMatches(data[0], '*') { |
||||
return false |
||||
} |
||||
level := 0 |
||||
for level < 6 && charMatches(data[level], '*') { |
||||
level++ |
||||
} |
||||
return charMatches(data[level], ' ') |
||||
} |
||||
|
||||
func (p *parser) generateHeadline(out *bytes.Buffer, data []byte) { |
||||
level := 1 |
||||
status := "" |
||||
priority := "" |
||||
|
||||
for level < 6 && data[level] == '*' { |
||||
level++ |
||||
} |
||||
|
||||
start := skipChar(data, level, ' ') |
||||
|
||||
data = data[start:] |
||||
i := 0 |
||||
|
||||
// Check if has a status so it can be rendered as a separate span that can be hidden or
|
||||
// modified with CSS classes
|
||||
if hasStatus(data[i:4]) { |
||||
status = string(data[i:4]) |
||||
i += 5 // one extra character for the next whitespace
|
||||
} |
||||
|
||||
// Check if the next byte is a priority marker
|
||||
if data[i] == '[' && hasPriority(data[i+1]) { |
||||
priority = string(data[i+1]) |
||||
i += 4 // for "[c]" + ' '
|
||||
} |
||||
|
||||
tags, tagsFound := findTags(data, i) |
||||
|
||||
headlineID := sanitized_anchor_name.Create(string(data[i:])) |
||||
|
||||
generate := func() bool { |
||||
dataEnd := len(data) |
||||
if tagsFound > 0 { |
||||
dataEnd = tagsFound |
||||
} |
||||
|
||||
headline := bytes.TrimRight(data[i:dataEnd], " \t") |
||||
|
||||
if status != "" { |
||||
out.WriteString("<span class=\"todo " + status + "\">" + status + "</span>") |
||||
out.WriteByte(' ') |
||||
} |
||||
|
||||
if priority != "" { |
||||
out.WriteString("<span class=\"priority " + priority + "\">[" + priority + "]</span>") |
||||
out.WriteByte(' ') |
||||
} |
||||
|
||||
p.inline(out, headline) |
||||
|
||||
if tagsFound > 0 { |
||||
for _, tag := range tags { |
||||
out.WriteByte(' ') |
||||
out.WriteString("<span class=\"tags " + tag + "\">" + tag + "</span>") |
||||
out.WriteByte(' ') |
||||
} |
||||
} |
||||
return true |
||||
} |
||||
|
||||
p.r.Header(out, generate, level, headlineID) |
||||
} |
||||
|
||||
func hasStatus(data []byte) bool { |
||||
return bytes.Contains(data, []byte("TODO")) || bytes.Contains(data, []byte("DONE")) |
||||
} |
||||
|
||||
func hasPriority(char byte) bool { |
||||
return (charMatches(char, 'A') || charMatches(char, 'B') || charMatches(char, 'C')) |
||||
} |
||||
|
||||
func findTags(data []byte, start int) ([]string, int) { |
||||
tags := []string{} |
||||
tagOpener := 0 |
||||
tagMarker := tagOpener |
||||
for tIdx := start; tIdx < len(data); tIdx++ { |
||||
if tagMarker > 0 && data[tIdx] == ':' { |
||||
tags = append(tags, string(data[tagMarker+1:tIdx])) |
||||
tagMarker = tIdx |
||||
} |
||||
if data[tIdx] == ':' && tagOpener == 0 && data[tIdx-1] == ' ' { |
||||
tagMarker = tIdx |
||||
tagOpener = tIdx |
||||
} |
||||
} |
||||
return tags, tagOpener |
||||
} |
||||
|
||||
// Greater Elements
|
||||
// ~~ Definition Lists
|
||||
var reDefinitionList = regexp.MustCompile(`^\s*-\s+(.+?)\s+::\s+(.*)`) |
||||
|
||||
func isDefinitionList(data []byte) bool { |
||||
return reDefinitionList.Match(data) |
||||
} |
||||
|
||||
// ~~ Example lines
|
||||
var reExampleLine = regexp.MustCompile(`^\s*:\s(\s*.*)|^\s*:$`) |
||||
|
||||
func isExampleLine(data []byte) bool { |
||||
return reExampleLine.Match(data) |
||||
} |
||||
|
||||
// ~~ Ordered Lists
|
||||
var reOrderedList = regexp.MustCompile(`^(\s*)\d+\.\s+\[?@?(\d*)\]?(.+)`) |
||||
|
||||
func isOrderedList(data []byte) bool { |
||||
return reOrderedList.Match(data) |
||||
} |
||||
|
||||
// ~~ Unordered Lists
|
||||
var reUnorderedList = regexp.MustCompile(`^(\s*)[-\+]\s+(.+)`) |
||||
|
||||
func isUnorderedList(data []byte) bool { |
||||
return reUnorderedList.Match(data) |
||||
} |
||||
|
||||
// ~~ Tables
|
||||
var reTableHeaders = regexp.MustCompile(`^[|+-]*$`) |
||||
|
||||
func isTable(data []byte) bool { |
||||
return charMatches(data[0], '|') |
||||
} |
||||
|
||||
func (p *parser) generateTable(output *bytes.Buffer, data []byte) { |
||||
var table bytes.Buffer |
||||
rows := bytes.Split(bytes.Trim(data, "\n"), []byte("\n")) |
||||
hasTableHeaders := len(rows) > 1 |
||||
if len(rows) > 1 { |
||||
hasTableHeaders = reTableHeaders.Match(rows[1]) |
||||
} |
||||
tbodySet := false |
||||
|
||||
for idx, row := range rows { |
||||
var rowBuff bytes.Buffer |
||||
if hasTableHeaders && idx == 0 { |
||||
table.WriteString("<thead>") |
||||
for _, cell := range bytes.Split(row[1:len(row)-1], []byte("|")) { |
||||
p.r.TableHeaderCell(&rowBuff, bytes.Trim(cell, " \t"), 0) |
||||
} |
||||
p.r.TableRow(&table, rowBuff.Bytes()) |
||||
table.WriteString("</thead>\n") |
||||
} else if hasTableHeaders && idx == 1 { |
||||
continue |
||||
} else { |
||||
if !tbodySet { |
||||
table.WriteString("<tbody>") |
||||
tbodySet = true |
||||
} |
||||
if !reTableHeaders.Match(row) { |
||||
for _, cell := range bytes.Split(row[1:len(row)-1], []byte("|")) { |
||||
var cellBuff bytes.Buffer |
||||
p.inline(&cellBuff, bytes.Trim(cell, " \t")) |
||||
p.r.TableCell(&rowBuff, cellBuff.Bytes(), 0) |
||||
} |
||||
p.r.TableRow(&table, rowBuff.Bytes()) |
||||
} |
||||
if tbodySet && idx == len(rows)-1 { |
||||
table.WriteString("</tbody>\n") |
||||
tbodySet = false |
||||
} |
||||
} |
||||
} |
||||
|
||||
output.WriteString("\n<table>\n") |
||||
output.Write(table.Bytes()) |
||||
output.WriteString("</table>\n") |
||||
} |
||||
|
||||
// ~~ Property Drawers
|
||||
|
||||
func isPropertyDrawer(data []byte) bool { |
||||
return bytes.Equal(data, []byte(":PROPERTIES:")) |
||||
} |
||||
|
||||
// ~~ Dynamic Blocks
|
||||
var reBlock = regexp.MustCompile(`^#\+(BEGIN|END)_(\w+)\s*([0-9A-Za-z_\-]*)?`) |
||||
|
||||
func isBlock(data []byte) bool { |
||||
return reBlock.Match(data) |
||||
} |
||||
|
||||
// ~~ Footnotes
|
||||
var reFootnoteDef = regexp.MustCompile(`^\[fn:([\w]+)\] +(.+)`) |
||||
|
||||
func isFootnoteDef(data []byte) bool { |
||||
return reFootnoteDef.Match(data) |
||||
} |
||||
|
||||
// Elements
|
||||
// ~~ Keywords
|
||||
func IsKeyword(data []byte) bool { |
||||
return len(data) > 2 && charMatches(data[0], '#') && charMatches(data[1], '+') && !charMatches(data[2], ' ') |
||||
} |
||||
|
||||
// ~~ Comments
|
||||
func isComment(data []byte) bool { |
||||
return charMatches(data[0], '#') && charMatches(data[1], ' ') |
||||
} |
||||
|
||||
func (p *parser) generateComment(out *bytes.Buffer, data []byte) { |
||||
var work bytes.Buffer |
||||
work.WriteString("<!-- ") |
||||
work.Write(data[2:]) |
||||
work.WriteString(" -->") |
||||
work.WriteByte('\n') |
||||
out.Write(work.Bytes()) |
||||
} |
||||
|
||||
// ~~ Horizontal Rules
|
||||
var reHorizontalRule = regexp.MustCompile(`^\s*?-----\s?$`) |
||||
|
||||
func isHorizontalRule(data []byte) bool { |
||||
return reHorizontalRule.Match(data) |
||||
} |
||||
|
||||
// ~~ Paragraphs
|
||||
func (p *parser) generateParagraph(out *bytes.Buffer, data []byte) { |
||||
generate := func() bool { |
||||
p.inline(out, bytes.Trim(data, " ")) |
||||
return true |
||||
} |
||||
p.r.Paragraph(out, generate) |
||||
} |
||||
|
||||
func (p *parser) generateList(output *bytes.Buffer, data []byte, listType string) { |
||||
generateList := func() bool { |
||||
output.WriteByte('\n') |
||||
p.inline(output, bytes.Trim(data, " ")) |
||||
return true |
||||
} |
||||
switch listType { |
||||
case "ul": |
||||
p.r.List(output, generateList, 0) |
||||
case "ol": |
||||
p.r.List(output, generateList, blackfriday.LIST_TYPE_ORDERED) |
||||
case "dl": |
||||
p.r.List(output, generateList, blackfriday.LIST_TYPE_DEFINITION) |
||||
} |
||||
} |
||||
|
||||
// Objects
|
||||
|
||||
func (p *parser) inline(out *bytes.Buffer, data []byte) { |
||||
i, end := 0, 0 |
||||
|
||||
for i < len(data) { |
||||
for end < len(data) && p.inlineCallback[data[end]] == nil { |
||||
end++ |
||||
} |
||||
|
||||
p.r.Entity(out, data[i:end]) |
||||
|
||||
if end >= len(data) { |
||||
break |
||||
} |
||||
i = end |
||||
|
||||
handler := p.inlineCallback[data[i]] |
||||
|
||||
if consumed := handler(p, out, data, i); consumed > 0 { |
||||
i += consumed |
||||
end = i |
||||
continue |
||||
} |
||||
|
||||
end = i + 1 |
||||
} |
||||
} |
||||
|
||||
func isAcceptablePreOpeningChar(dataIn, data []byte, offset int) bool { |
||||
if len(dataIn) == len(data) { |
||||
return true |
||||
} |
||||
|
||||
char := dataIn[offset-1] |
||||
return charMatches(char, ' ') || isPreChar(char) |
||||
} |
||||
|
||||
func isPreChar(char byte) bool { |
||||
return charMatches(char, '>') || charMatches(char, '(') || charMatches(char, '{') || charMatches(char, '[') |
||||
} |
||||
|
||||
func isAcceptablePostClosingChar(char byte) bool { |
||||
return charMatches(char, ' ') || isTerminatingChar(char) |
||||
} |
||||
|
||||
func isTerminatingChar(char byte) bool { |
||||
return charMatches(char, '.') || charMatches(char, ',') || charMatches(char, '?') || charMatches(char, '!') || charMatches(char, ')') || charMatches(char, '}') || charMatches(char, ']') |
||||
} |
||||
|
||||
func findLastCharInInline(data []byte, char byte) int { |
||||
timesFound := 0 |
||||
last := 0 |
||||
// Start from character after the inline indicator
|
||||
for i := 1; i < len(data); i++ { |
||||
if timesFound == 1 { |
||||
break |
||||
} |
||||
if data[i] == char { |
||||
if len(data) == i+1 || (len(data) > i+1 && isAcceptablePostClosingChar(data[i+1])) { |
||||
last = i |
||||
timesFound += 1 |
||||
} |
||||
} |
||||
} |
||||
return last |
||||
} |
||||
|
||||
func generator(p *parser, out *bytes.Buffer, dataIn []byte, offset int, char byte, doInline bool, renderer func(*bytes.Buffer, []byte)) int { |
||||
data := dataIn[offset:] |
||||
c := byte(char) |
||||
start := 1 |
||||
i := start |
||||
if len(data) <= 1 { |
||||
return 0 |
||||
} |
||||
|
||||
lastCharInside := findLastCharInInline(data, c) |
||||
|
||||
// Org mode spec says a non-whitespace character must immediately follow.
|
||||
// if the current char is the marker, then there's no text between, not a candidate
|
||||
if isSpace(data[i]) || lastCharInside == i || !isAcceptablePreOpeningChar(dataIn, data, offset) { |
||||
return 0 |
||||
} |
||||
|
||||
if lastCharInside > 0 { |
||||
var work bytes.Buffer |
||||
if doInline { |
||||
p.inline(&work, data[start:lastCharInside]) |
||||
renderer(out, work.Bytes()) |
||||
} else { |
||||
renderer(out, data[start:lastCharInside]) |
||||
} |
||||
next := lastCharInside + 1 |
||||
return next |
||||
} |
||||
|
||||
return 0 |
||||
} |
||||
|
||||
// ~~ Text Markup
|
||||
func generateVerbatim(p *parser, out *bytes.Buffer, data []byte, offset int) int { |
||||
return generator(p, out, data, offset, '=', false, p.r.CodeSpan) |
||||
} |
||||
|
||||
func generateCode(p *parser, out *bytes.Buffer, data []byte, offset int) int { |
||||
return generator(p, out, data, offset, '~', false, p.r.CodeSpan) |
||||
} |
||||
|
||||
func generateEmphasis(p *parser, out *bytes.Buffer, data []byte, offset int) int { |
||||
return generator(p, out, data, offset, '/', true, p.r.Emphasis) |
||||
} |
||||
|
||||
func generateUnderline(p *parser, out *bytes.Buffer, data []byte, offset int) int { |
||||
underline := func(out *bytes.Buffer, text []byte) { |
||||
out.WriteString("<span style=\"text-decoration: underline;\">") |
||||
out.Write(text) |
||||
out.WriteString("</span>") |
||||
} |
||||
|
||||
return generator(p, out, data, offset, '_', true, underline) |
||||
} |
||||
|
||||
func generateBold(p *parser, out *bytes.Buffer, data []byte, offset int) int { |
||||
return generator(p, out, data, offset, '*', true, p.r.DoubleEmphasis) |
||||
} |
||||
|
||||
func generateStrikethrough(p *parser, out *bytes.Buffer, data []byte, offset int) int { |
||||
return generator(p, out, data, offset, '+', true, p.r.StrikeThrough) |
||||
} |
||||
|
||||
// ~~ Images and Links (inc. Footnote)
|
||||
var reLinkOrImg = regexp.MustCompile(`\[\[(.+?)\]\[?(.*?)\]?\]`) |
||||
|
||||
func generateLinkOrImg(p *parser, out *bytes.Buffer, data []byte, offset int) int { |
||||
data = data[offset+1:] |
||||
start := 1 |
||||
i := start |
||||
var hyperlink []byte |
||||
isImage := false |
||||
isFootnote := false |
||||
closedLink := false |
||||
hasContent := false |
||||
|
||||
if bytes.Equal(data[0:3], []byte("fn:")) { |
||||
isFootnote = true |
||||
} else if data[0] != '[' { |
||||
return 0 |
||||
} |
||||
|
||||
if bytes.Equal(data[1:6], []byte("file:")) { |
||||
isImage = true |
||||
} |
||||
|
||||
for i < len(data) { |
||||
currChar := data[i] |
||||
switch { |
||||
case charMatches(currChar, ']') && closedLink == false: |
||||
if isImage { |
||||
hyperlink = data[start+5 : i] |
||||
} else if isFootnote { |
||||
refid := data[start+2 : i] |
||||
if bytes.Equal(refid, bytes.Trim(refid, " ")) { |
||||
p.notes = append(p.notes, footnotes{string(refid), "DEFINITION NOT FOUND"}) |
||||
p.r.FootnoteRef(out, refid, len(p.notes)) |
||||
return i + 2 |
||||
} else { |
||||
return 0 |
||||
} |
||||
} else if bytes.Equal(data[i-4:i], []byte(".org")) { |
||||
orgStart := start |
||||
if bytes.Equal(data[orgStart:orgStart+2], []byte("./")) { |
||||
orgStart = orgStart + 1 |
||||
} |
||||
hyperlink = data[orgStart : i-4] |
||||
} else { |
||||
hyperlink = data[start:i] |
||||
} |
||||
closedLink = true |
||||
case charMatches(currChar, '['): |
||||
start = i + 1 |
||||
hasContent = true |
||||
case charMatches(currChar, ']') && closedLink == true && hasContent == true && isImage == true: |
||||
p.r.Image(out, hyperlink, data[start:i], data[start:i]) |
||||
return i + 3 |
||||
case charMatches(currChar, ']') && closedLink == true && hasContent == true: |
||||
var tmpBuf bytes.Buffer |
||||
p.inline(&tmpBuf, data[start:i]) |
||||
p.r.Link(out, hyperlink, tmpBuf.Bytes(), tmpBuf.Bytes()) |
||||
return i + 3 |
||||
case charMatches(currChar, ']') && closedLink == true && hasContent == false && isImage == true: |
||||
p.r.Image(out, hyperlink, hyperlink, hyperlink) |
||||
return i + 2 |
||||
case charMatches(currChar, ']') && closedLink == true && hasContent == false: |
||||
p.r.Link(out, hyperlink, hyperlink, hyperlink) |
||||
return i + 2 |
||||
} |
||||
i++ |
||||
} |
||||
|
||||
return 0 |
||||
} |
||||
|
||||
// Helpers
|
||||
func skipChar(data []byte, start int, char byte) int { |
||||
i := start |
||||
for i < len(data) && charMatches(data[i], char) { |
||||
i++ |
||||
} |
||||
return i |
||||
} |
||||
|
||||
func isSpace(char byte) bool { |
||||
return charMatches(char, ' ') |
||||
} |
||||
|
||||
func isEmpty(data []byte) bool { |
||||
if len(data) == 0 { |
||||
return true |
||||
} |
||||
|
||||
for i := 0; i < len(data) && !charMatches(data[i], '\n'); i++ { |
||||
if !charMatches(data[i], ' ') && !charMatches(data[i], '\t') { |
||||
return false |
||||
} |
||||
} |
||||
return true |
||||
} |
||||
|
||||
func charMatches(a byte, b byte) bool { |
||||
return a == b |
||||
} |
Before Width: | Height: | Size: 15 KiB |
Before Width: | Height: | Size: 3.2 KiB |
@ -1,70 +0,0 @@ |
||||
package goorgeous |
||||
|
||||
import ( |
||||
"bufio" |
||||
"bytes" |
||||
"regexp" |
||||
"strings" |
||||
) |
||||
|
||||
// ExtractOrgHeaders finds and returns all of the headers
|
||||
// from a bufio.Reader and returns them as their own byte slice
|
||||
func ExtractOrgHeaders(r *bufio.Reader) (fm []byte, err error) { |
||||
var out bytes.Buffer |
||||
endOfHeaders := true |
||||
for endOfHeaders { |
||||
p, err := r.Peek(2) |
||||
if err != nil { |
||||
return nil, err |
||||
} |
||||
if !charMatches(p[0], '#') && !charMatches(p[1], '+') { |
||||
endOfHeaders = false |
||||
break |
||||
} |
||||
line, _, err := r.ReadLine() |
||||
if err != nil { |
||||
return nil, err |
||||
} |
||||
out.Write(line) |
||||
out.WriteByte('\n') |
||||
} |
||||
return out.Bytes(), nil |
||||
} |
||||
|
||||
var reHeader = regexp.MustCompile(`^#\+(\w+?): (.*)`) |
||||
|
||||
// OrgHeaders find all of the headers from a byte slice and returns
|
||||
// them as a map of string interface
|
||||
func OrgHeaders(input []byte) (map[string]interface{}, error) { |
||||
out := make(map[string]interface{}) |
||||
scanner := bufio.NewScanner(bytes.NewReader(input)) |
||||
|
||||
for scanner.Scan() { |
||||
data := scanner.Bytes() |
||||
if !charMatches(data[0], '#') && !charMatches(data[1], '+') { |
||||
return out, nil |
||||
} |
||||
matches := reHeader.FindSubmatch(data) |
||||
|
||||
if len(matches) < 3 { |
||||
continue |
||||
} |
||||
|
||||
key := string(matches[1]) |
||||
val := matches[2] |
||||
switch { |
||||
case strings.ToLower(key) == "tags" || strings.ToLower(key) == "categories" || strings.ToLower(key) == "aliases": |
||||
bTags := bytes.Split(val, []byte(" ")) |
||||
tags := make([]string, len(bTags)) |
||||
for idx, tag := range bTags { |
||||
tags[idx] = string(tag) |
||||
} |
||||
out[key] = tags |
||||
default: |
||||
out[key] = string(val) |
||||
} |
||||
|
||||
} |
||||
return out, nil |
||||
|
||||
} |
@ -1,6 +1,6 @@ |
||||
MIT License |
||||
|
||||
Copyright (c) 2017 Chase Adams <realchaseadams@gmail.com> |
||||
Copyright (c) 2018 Niklas Fasching |
||||
|
||||
Permission is hereby granted, free of charge, to any person obtaining a copy |
||||
of this software and associated documentation files (the "Software"), to deal |
@ -0,0 +1,84 @@ |
||||
package org |
||||
|
||||
import ( |
||||
"regexp" |
||||
"strings" |
||||
"unicode" |
||||
) |
||||
|
||||
type Block struct { |
||||
Name string |
||||
Parameters []string |
||||
Children []Node |
||||
} |
||||
|
||||
type Example struct { |
||||
Children []Node |
||||
} |
||||
|
||||
var exampleLineRegexp = regexp.MustCompile(`^(\s*):(\s(.*)|\s*$)`) |
||||
var beginBlockRegexp = regexp.MustCompile(`(?i)^(\s*)#\+BEGIN_(\w+)(.*)`) |
||||
var endBlockRegexp = regexp.MustCompile(`(?i)^(\s*)#\+END_(\w+)`) |
||||
|
||||
func lexBlock(line string) (token, bool) { |
||||
if m := beginBlockRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"beginBlock", len(m[1]), strings.ToUpper(m[2]), m}, true |
||||
} else if m := endBlockRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"endBlock", len(m[1]), strings.ToUpper(m[2]), m}, true |
||||
} |
||||
return nilToken, false |
||||
} |
||||
|
||||
func lexExample(line string) (token, bool) { |
||||
if m := exampleLineRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"example", len(m[1]), m[3], m}, true |
||||
} |
||||
return nilToken, false |
||||
} |
||||
|
||||
func isRawTextBlock(name string) bool { return name == "SRC" || name == "EXAMPLE" || name == "EXPORT" } |
||||
|
||||
func (d *Document) parseBlock(i int, parentStop stopFn) (int, Node) { |
||||
t, start := d.tokens[i], i |
||||
name, parameters := t.content, strings.Fields(t.matches[3]) |
||||
trim := trimIndentUpTo(d.tokens[i].lvl) |
||||
stop := func(d *Document, i int) bool { |
||||
return i >= len(d.tokens) || (d.tokens[i].kind == "endBlock" && d.tokens[i].content == name) |
||||
} |
||||
block, i := Block{name, parameters, nil}, i+1 |
||||
if isRawTextBlock(name) { |
||||
rawText := "" |
||||
for ; !stop(d, i); i++ { |
||||
rawText += trim(d.tokens[i].matches[0]) + "\n" |
||||
} |
||||
block.Children = d.parseRawInline(rawText) |
||||
} else { |
||||
consumed, nodes := d.parseMany(i, stop) |
||||
block.Children = nodes |
||||
i += consumed |
||||
} |
||||
if i < len(d.tokens) && d.tokens[i].kind == "endBlock" && d.tokens[i].content == name { |
||||
return i + 1 - start, block |
||||
} |
||||
return 0, nil |
||||
} |
||||
|
||||
func (d *Document) parseExample(i int, parentStop stopFn) (int, Node) { |
||||
example, start := Example{}, i |
||||
for ; !parentStop(d, i) && d.tokens[i].kind == "example"; i++ { |
||||
example.Children = append(example.Children, Text{d.tokens[i].content, true}) |
||||
} |
||||
return i - start, example |
||||
} |
||||
|
||||
func trimIndentUpTo(max int) func(string) string { |
||||
return func(line string) string { |
||||
i := 0 |
||||
for ; i < len(line) && i < max && unicode.IsSpace(rune(line[i])); i++ { |
||||
} |
||||
return line[i:] |
||||
} |
||||
} |
||||
|
||||
func (n Example) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n Block) String() string { return orgWriter.nodesAsString(n) } |
@ -0,0 +1,260 @@ |
||||
// Package org is an Org mode syntax processor.
|
||||
//
|
||||
// It parses plain text into an AST and can export it as HTML or pretty printed Org mode syntax.
|
||||
// Further export formats can be defined using the Writer interface.
|
||||
//
|
||||
// You probably want to start with something like this:
|
||||
// input := strings.NewReader("Your Org mode input")
|
||||
// html, err := org.New().Parse(input, "./").Write(org.NewHTMLWriter())
|
||||
// if err != nil {
|
||||
// log.Fatalf("Something went wrong: %s", err)
|
||||
// }
|
||||
// log.Print(html)
|
||||
package org |
||||
|
||||
import ( |
||||
"bufio" |
||||
"fmt" |
||||
"io" |
||||
"io/ioutil" |
||||
"log" |
||||
"os" |
||||
"strings" |
||||
) |
||||
|
||||
type Configuration struct { |
||||
MaxEmphasisNewLines int // Maximum number of newlines inside an emphasis. See org-emphasis-regexp-components newline.
|
||||
AutoLink bool // Try to convert text passages that look like hyperlinks into hyperlinks.
|
||||
DefaultSettings map[string]string // Default values for settings that are overriden by setting the same key in BufferSettings.
|
||||
Log *log.Logger // Log is used to print warnings during parsing.
|
||||
ReadFile func(filename string) ([]byte, error) // ReadFile is used to read e.g. #+INCLUDE files.
|
||||
} |
||||
|
||||
// Document contains the parsing results and a pointer to the Configuration.
|
||||
type Document struct { |
||||
*Configuration |
||||
Path string // Path of the file containing the parse input - used to resolve relative paths during parsing (e.g. INCLUDE).
|
||||
tokens []token |
||||
Nodes []Node |
||||
NamedNodes map[string]Node |
||||
Outline Outline // Outline is a Table Of Contents for the document and contains all sections (headline + content).
|
||||
BufferSettings map[string]string // Settings contains all settings that were parsed from keywords.
|
||||
Error error |
||||
} |
||||
|
||||
// Node represents a parsed node of the document.
|
||||
type Node interface { |
||||
String() string // String returns the pretty printed Org mode string for the node (see OrgWriter).
|
||||
} |
||||
|
||||
type lexFn = func(line string) (t token, ok bool) |
||||
type parseFn = func(*Document, int, stopFn) (int, Node) |
||||
type stopFn = func(*Document, int) bool |
||||
|
||||
type token struct { |
||||
kind string |
||||
lvl int |
||||
content string |
||||
matches []string |
||||
} |
||||
|
||||
var lexFns = []lexFn{ |
||||
lexHeadline, |
||||
lexDrawer, |
||||
lexBlock, |
||||
lexList, |
||||
lexTable, |
||||
lexHorizontalRule, |
||||
lexKeywordOrComment, |
||||
lexFootnoteDefinition, |
||||
lexExample, |
||||
lexText, |
||||
} |
||||
|
||||
var nilToken = token{"nil", -1, "", nil} |
||||
var orgWriter = NewOrgWriter() |
||||
|
||||
// New returns a new Configuration with (hopefully) sane defaults.
|
||||
func New() *Configuration { |
||||
return &Configuration{ |
||||
AutoLink: true, |
||||
MaxEmphasisNewLines: 1, |
||||
DefaultSettings: map[string]string{ |
||||
"TODO": "TODO | DONE", |
||||
"EXCLUDE_TAGS": "noexport", |
||||
"OPTIONS": "toc:t <:t e:t f:t pri:t todo:t tags:t", |
||||
}, |
||||
Log: log.New(os.Stderr, "go-org: ", 0), |
||||
ReadFile: ioutil.ReadFile, |
||||
} |
||||
} |
||||
|
||||
// String returns the pretty printed Org mode string for the given nodes (see OrgWriter).
|
||||
func String(nodes []Node) string { return orgWriter.nodesAsString(nodes...) } |
||||
|
||||
// Write is called after with an instance of the Writer interface to export a parsed Document into another format.
|
||||
func (d *Document) Write(w Writer) (out string, err error) { |
||||
defer func() { |
||||
if recovered := recover(); recovered != nil { |
||||
err = fmt.Errorf("could not write output: %s", recovered) |
||||
} |
||||
}() |
||||
if d.Error != nil { |
||||
return "", d.Error |
||||
} else if d.Nodes == nil { |
||||
return "", fmt.Errorf("could not write output: parse was not called") |
||||
} |
||||
w.Before(d) |
||||
WriteNodes(w, d.Nodes...) |
||||
w.After(d) |
||||
return w.String(), err |
||||
} |
||||
|
||||
// Parse parses the input into an AST (and some other helpful fields like Outline).
|
||||
// To allow method chaining, errors are stored in document.Error rather than being returned.
|
||||
func (c *Configuration) Parse(input io.Reader, path string) (d *Document) { |
||||
outlineSection := &Section{} |
||||
d = &Document{ |
||||
Configuration: c, |
||||
Outline: Outline{outlineSection, outlineSection, 0}, |
||||
BufferSettings: map[string]string{}, |
||||
NamedNodes: map[string]Node{}, |
||||
Path: path, |
||||
} |
||||
defer func() { |
||||
if recovered := recover(); recovered != nil { |
||||
d.Error = fmt.Errorf("could not parse input: %v", recovered) |
||||
} |
||||
}() |
||||
if d.tokens != nil { |
||||
d.Error = fmt.Errorf("parse was called multiple times") |
||||
} |
||||
d.tokenize(input) |
||||
_, nodes := d.parseMany(0, func(d *Document, i int) bool { return i >= len(d.tokens) }) |
||||
d.Nodes = nodes |
||||
return d |
||||
} |
||||
|
||||
// Silent disables all logging of warnings during parsing.
|
||||
func (c *Configuration) Silent() *Configuration { |
||||
c.Log = log.New(ioutil.Discard, "", 0) |
||||
return c |
||||
} |
||||
|
||||
func (d *Document) tokenize(input io.Reader) { |
||||
d.tokens = []token{} |
||||
scanner := bufio.NewScanner(input) |
||||
for scanner.Scan() { |
||||
d.tokens = append(d.tokens, tokenize(scanner.Text())) |
||||
} |
||||
if err := scanner.Err(); err != nil { |
||||
d.Error = fmt.Errorf("could not tokenize input: %s", err) |
||||
} |
||||
} |
||||
|
||||
// Get returns the value for key in BufferSettings or DefaultSettings if key does not exist in the former
|
||||
func (d *Document) Get(key string) string { |
||||
if v, ok := d.BufferSettings[key]; ok { |
||||
return v |
||||
} |
||||
if v, ok := d.DefaultSettings[key]; ok { |
||||
return v |
||||
} |
||||
return "" |
||||
} |
||||
|
||||
// GetOption returns the value associated to the export option key
|
||||
// Currently supported options:
|
||||
// - < (export timestamps)
|
||||
// - e (export org entities)
|
||||
// - f (export footnotes)
|
||||
// - toc (export table of content)
|
||||
// - todo (export headline todo status)
|
||||
// - pri (export headline priority)
|
||||
// - tags (export headline tags)
|
||||
// see https://orgmode.org/manual/Export-settings.html for more information
|
||||
func (d *Document) GetOption(key string) bool { |
||||
get := func(settings map[string]string) string { |
||||
for _, field := range strings.Fields(settings["OPTIONS"]) { |
||||
if strings.HasPrefix(field, key+":") { |
||||
return field[len(key)+1:] |
||||
} |
||||
} |
||||
return "" |
||||
} |
||||
value := get(d.BufferSettings) |
||||
if value == "" { |
||||
value = get(d.DefaultSettings) |
||||
} |
||||
switch value { |
||||
case "t": |
||||
return true |
||||
case "nil": |
||||
return false |
||||
default: |
||||
d.Log.Printf("Bad value for export option %s (%s)", key, value) |
||||
return false |
||||
} |
||||
} |
||||
|
||||
func (d *Document) parseOne(i int, stop stopFn) (consumed int, node Node) { |
||||
switch d.tokens[i].kind { |
||||
case "unorderedList", "orderedList": |
||||
consumed, node = d.parseList(i, stop) |
||||
case "tableRow", "tableSeparator": |
||||
consumed, node = d.parseTable(i, stop) |
||||
case "beginBlock": |
||||
consumed, node = d.parseBlock(i, stop) |
||||
case "beginDrawer": |
||||
consumed, node = d.parseDrawer(i, stop) |
||||
case "text": |
||||
consumed, node = d.parseParagraph(i, stop) |
||||
case "example": |
||||
consumed, node = d.parseExample(i, stop) |
||||
case "horizontalRule": |
||||
consumed, node = d.parseHorizontalRule(i, stop) |
||||
case "comment": |
||||
consumed, node = d.parseComment(i, stop) |
||||
case "keyword": |
||||
consumed, node = d.parseKeyword(i, stop) |
||||
case "headline": |
||||
consumed, node = d.parseHeadline(i, stop) |
||||
case "footnoteDefinition": |
||||
consumed, node = d.parseFootnoteDefinition(i, stop) |
||||
} |
||||
|
||||
if consumed != 0 { |
||||
return consumed, node |
||||
} |
||||
d.Log.Printf("Could not parse token %#v: Falling back to treating it as plain text.", d.tokens[i]) |
||||
m := plainTextRegexp.FindStringSubmatch(d.tokens[i].matches[0]) |
||||
d.tokens[i] = token{"text", len(m[1]), m[2], m} |
||||
return d.parseOne(i, stop) |
||||
} |
||||
|
||||
func (d *Document) parseMany(i int, stop stopFn) (int, []Node) { |
||||
start, nodes := i, []Node{} |
||||
for i < len(d.tokens) && !stop(d, i) { |
||||
consumed, node := d.parseOne(i, stop) |
||||
i += consumed |
||||
nodes = append(nodes, node) |
||||
} |
||||
return i - start, nodes |
||||
} |
||||
|
||||
func (d *Document) addHeadline(headline *Headline) int { |
||||
current := &Section{Headline: headline} |
||||
d.Outline.last.add(current) |
||||
d.Outline.count++ |
||||
d.Outline.last = current |
||||
return d.Outline.count |
||||
} |
||||
|
||||
func tokenize(line string) token { |
||||
for _, lexFn := range lexFns { |
||||
if token, ok := lexFn(line); ok { |
||||
return token |
||||
} |
||||
} |
||||
panic(fmt.Sprintf("could not lex line: %s", line)) |
||||
} |
@ -0,0 +1,97 @@ |
||||
package org |
||||
|
||||
import ( |
||||
"regexp" |
||||
"strings" |
||||
) |
||||
|
||||
type Drawer struct { |
||||
Name string |
||||
Children []Node |
||||
} |
||||
|
||||
type PropertyDrawer struct { |
||||
Properties [][]string |
||||
} |
||||
|
||||
var beginDrawerRegexp = regexp.MustCompile(`^(\s*):(\S+):\s*$`) |
||||
var endDrawerRegexp = regexp.MustCompile(`^(\s*):END:\s*$`) |
||||
var propertyRegexp = regexp.MustCompile(`^(\s*):(\S+):(\s+(.*)$|$)`) |
||||
|
||||
func lexDrawer(line string) (token, bool) { |
||||
if m := endDrawerRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"endDrawer", len(m[1]), "", m}, true |
||||
} else if m := beginDrawerRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"beginDrawer", len(m[1]), strings.ToUpper(m[2]), m}, true |
||||
} |
||||
return nilToken, false |
||||
} |
||||
|
||||
func (d *Document) parseDrawer(i int, parentStop stopFn) (int, Node) { |
||||
name := strings.ToUpper(d.tokens[i].content) |
||||
if name == "PROPERTIES" { |
||||
return d.parsePropertyDrawer(i, parentStop) |
||||
} |
||||
drawer, start := Drawer{Name: name}, i |
||||
i++ |
||||
stop := func(d *Document, i int) bool { |
||||
if parentStop(d, i) { |
||||
return true |
||||
} |
||||
kind := d.tokens[i].kind |
||||
return kind == "beginDrawer" || kind == "endDrawer" || kind == "headline" |
||||
} |
||||
for { |
||||
consumed, nodes := d.parseMany(i, stop) |
||||
i += consumed |
||||
drawer.Children = append(drawer.Children, nodes...) |
||||
if i < len(d.tokens) && d.tokens[i].kind == "beginDrawer" { |
||||
p := Paragraph{[]Node{Text{":" + d.tokens[i].content + ":", false}}} |
||||
drawer.Children = append(drawer.Children, p) |
||||
i++ |
||||
} else { |
||||
break |
||||
} |
||||
} |
||||
if i < len(d.tokens) && d.tokens[i].kind == "endDrawer" { |
||||
i++ |
||||
} |
||||
return i - start, drawer |
||||
} |
||||
|
||||
func (d *Document) parsePropertyDrawer(i int, parentStop stopFn) (int, Node) { |
||||
drawer, start := PropertyDrawer{}, i |
||||
i++ |
||||
stop := func(d *Document, i int) bool { |
||||
return parentStop(d, i) || (d.tokens[i].kind != "text" && d.tokens[i].kind != "beginDrawer") |
||||
} |
||||
for ; !stop(d, i); i++ { |
||||
m := propertyRegexp.FindStringSubmatch(d.tokens[i].matches[0]) |
||||
if m == nil { |
||||
return 0, nil |
||||
} |
||||
k, v := strings.ToUpper(m[2]), strings.TrimSpace(m[4]) |
||||
drawer.Properties = append(drawer.Properties, []string{k, v}) |
||||
} |
||||
if i < len(d.tokens) && d.tokens[i].kind == "endDrawer" { |
||||
i++ |
||||
} else { |
||||
return 0, nil |
||||
} |
||||
return i - start, drawer |
||||
} |
||||
|
||||
func (d *PropertyDrawer) Get(key string) (string, bool) { |
||||
if d == nil { |
||||
return "", false |
||||
} |
||||
for _, kvPair := range d.Properties { |
||||
if kvPair[0] == key { |
||||
return kvPair[1], true |
||||
} |
||||
} |
||||
return "", false |
||||
} |
||||
|
||||
func (n Drawer) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n PropertyDrawer) String() string { return orgWriter.nodesAsString(n) } |
@ -0,0 +1,35 @@ |
||||
package org |
||||
|
||||
import ( |
||||
"regexp" |
||||
) |
||||
|
||||
type FootnoteDefinition struct { |
||||
Name string |
||||
Children []Node |
||||
Inline bool |
||||
} |
||||
|
||||
var footnoteDefinitionRegexp = regexp.MustCompile(`^\[fn:([\w-]+)\](\s+(.+)|\s*$)`) |
||||
|
||||
func lexFootnoteDefinition(line string) (token, bool) { |
||||
if m := footnoteDefinitionRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"footnoteDefinition", 0, m[1], m}, true |
||||
} |
||||
return nilToken, false |
||||
} |
||||
|
||||
func (d *Document) parseFootnoteDefinition(i int, parentStop stopFn) (int, Node) { |
||||
start, name := i, d.tokens[i].content |
||||
d.tokens[i] = tokenize(d.tokens[i].matches[2]) |
||||
stop := func(d *Document, i int) bool { |
||||
return parentStop(d, i) || |
||||
(isSecondBlankLine(d, i) && i > start+1) || |
||||
d.tokens[i].kind == "headline" || d.tokens[i].kind == "footnoteDefinition" |
||||
} |
||||
consumed, nodes := d.parseMany(i, stop) |
||||
definition := FootnoteDefinition{name, nodes, false} |
||||
return consumed, definition |
||||
} |
||||
|
||||
func (n FootnoteDefinition) String() string { return orgWriter.nodesAsString(n) } |
@ -0,0 +1,27 @@ |
||||
// +build gofuzz
|
||||
|
||||
package org |
||||
|
||||
import ( |
||||
"bytes" |
||||
"strings" |
||||
) |
||||
|
||||
// Fuzz function to be used by https://github.com/dvyukov/go-fuzz
|
||||
func Fuzz(input []byte) int { |
||||
conf := New().Silent() |
||||
d := conf.Parse(bytes.NewReader(input), "") |
||||
orgOutput, err := d.Write(NewOrgWriter()) |
||||
if err != nil { |
||||
panic(err) |
||||
} |
||||
htmlOutputA, err := d.Write(NewHTMLWriter()) |
||||
if err != nil { |
||||
panic(err) |
||||
} |
||||
htmlOutputB, err := conf.Parse(strings.NewReader(orgOutput), "").Write(NewHTMLWriter()) |
||||
if htmlOutputA != htmlOutputB { |
||||
panic("rendered org results in different html than original input") |
||||
} |
||||
return 0 |
||||
} |
@ -0,0 +1,101 @@ |
||||
package org |
||||
|
||||
import ( |
||||
"fmt" |
||||
"regexp" |
||||
"strings" |
||||
"unicode" |
||||
) |
||||
|
||||
type Outline struct { |
||||
*Section |
||||
last *Section |
||||
count int |
||||
} |
||||
|
||||
type Section struct { |
||||
Headline *Headline |
||||
Parent *Section |
||||
Children []*Section |
||||
} |
||||
|
||||
type Headline struct { |
||||
Index int |
||||
Lvl int |
||||
Status string |
||||
Priority string |
||||
Properties *PropertyDrawer |
||||
Title []Node |
||||
Tags []string |
||||
Children []Node |
||||
} |
||||
|
||||
var headlineRegexp = regexp.MustCompile(`^([*]+)\s+(.*)`) |
||||
var tagRegexp = regexp.MustCompile(`(.*?)\s+(:[A-Za-z0-9_@#%:]+:\s*$)`) |
||||
|
||||
func lexHeadline(line string) (token, bool) { |
||||
if m := headlineRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"headline", len(m[1]), m[2], m}, true |
||||
} |
||||
return nilToken, false |
||||
} |
||||
|
||||
func (d *Document) parseHeadline(i int, parentStop stopFn) (int, Node) { |
||||
t, headline := d.tokens[i], Headline{} |
||||
headline.Lvl = t.lvl |
||||
|
||||
headline.Index = d.addHeadline(&headline) |
||||
|
||||
text := t.content |
||||
todoKeywords := strings.FieldsFunc(d.Get("TODO"), func(r rune) bool { return unicode.IsSpace(r) || r == '|' }) |
||||
for _, k := range todoKeywords { |
||||
if strings.HasPrefix(text, k) && len(text) > len(k) && unicode.IsSpace(rune(text[len(k)])) { |
||||
headline.Status = k |
||||
text = text[len(k)+1:] |
||||
break |
||||
} |
||||
} |
||||
|
||||
if len(text) >= 4 && text[0:2] == "[#" && strings.Contains("ABC", text[2:3]) && text[3] == ']' { |
||||
headline.Priority = text[2:3] |
||||
text = strings.TrimSpace(text[4:]) |
||||
} |
||||
|
||||
if m := tagRegexp.FindStringSubmatch(text); m != nil { |
||||
text = m[1] |
||||
headline.Tags = strings.FieldsFunc(m[2], func(r rune) bool { return r == ':' }) |
||||
} |
||||
|
||||
headline.Title = d.parseInline(text) |
||||
|
||||
stop := func(d *Document, i int) bool { |
||||
return parentStop(d, i) || d.tokens[i].kind == "headline" && d.tokens[i].lvl <= headline.Lvl |
||||
} |
||||
consumed, nodes := d.parseMany(i+1, stop) |
||||
if len(nodes) > 0 { |
||||
if d, ok := nodes[0].(PropertyDrawer); ok { |
||||
headline.Properties = &d |
||||
nodes = nodes[1:] |
||||
} |
||||
} |
||||
headline.Children = nodes |
||||
return consumed + 1, headline |
||||
} |
||||
|
||||
func (h Headline) ID() string { |
||||
if customID, ok := h.Properties.Get("CUSTOM_ID"); ok { |
||||
return customID |
||||
} |
||||
return fmt.Sprintf("headline-%d", h.Index) |
||||
} |
||||
|
||||
func (parent *Section) add(current *Section) { |
||||
if parent.Headline == nil || parent.Headline.Lvl < current.Headline.Lvl { |
||||
parent.Children = append(parent.Children, current) |
||||
current.Parent = parent |
||||
} else { |
||||
parent.Parent.add(current) |
||||
} |
||||
} |
||||
|
||||
func (n Headline) String() string { return orgWriter.nodesAsString(n) } |
@ -0,0 +1,504 @@ |
||||
package org |
||||
|
||||
import ( |
||||
"fmt" |
||||
"html" |
||||
"log" |
||||
"regexp" |
||||
"strings" |
||||
"unicode" |
||||
|
||||
h "golang.org/x/net/html" |
||||
"golang.org/x/net/html/atom" |
||||
) |
||||
|
||||
// HTMLWriter exports an org document into a html document.
|
||||
type HTMLWriter struct { |
||||
ExtendingWriter Writer |
||||
HighlightCodeBlock func(source, lang string) string |
||||
|
||||
strings.Builder |
||||
document *Document |
||||
htmlEscape bool |
||||
log *log.Logger |
||||
footnotes *footnotes |
||||
} |
||||
|
||||
type footnotes struct { |
||||
mapping map[string]int |
||||
list []*FootnoteDefinition |
||||
} |
||||
|
||||
var emphasisTags = map[string][]string{ |
||||
"/": []string{"<em>", "</em>"}, |
||||
"*": []string{"<strong>", "</strong>"}, |
||||
"+": []string{"<del>", "</del>"}, |
||||
"~": []string{"<code>", "</code>"}, |
||||
"=": []string{`<code class="verbatim">`, "</code>"}, |
||||
"_": []string{`<span style="text-decoration: underline;">`, "</span>"}, |
||||
"_{}": []string{"<sub>", "</sub>"}, |
||||
"^{}": []string{"<sup>", "</sup>"}, |
||||
} |
||||
|
||||
var listTags = map[string][]string{ |
||||
"unordered": []string{"<ul>", "</ul>"}, |
||||
"ordered": []string{"<ol>", "</ol>"}, |
||||
"descriptive": []string{"<dl>", "</dl>"}, |
||||
} |
||||
|
||||
var listItemStatuses = map[string]string{ |
||||
" ": "unchecked", |
||||
"-": "indeterminate", |
||||
"X": "checked", |
||||
} |
||||
|
||||
var cleanHeadlineTitleForHTMLAnchorRegexp = regexp.MustCompile(`</?a[^>]*>`) // nested a tags are not valid HTML
|
||||
|
||||
func NewHTMLWriter() *HTMLWriter { |
||||
defaultConfig := New() |
||||
return &HTMLWriter{ |
||||
document: &Document{Configuration: defaultConfig}, |
||||
log: defaultConfig.Log, |
||||
htmlEscape: true, |
||||
HighlightCodeBlock: func(source, lang string) string { |
||||
return fmt.Sprintf("<div class=\"highlight\">\n<pre>\n%s\n</pre>\n</div>", html.EscapeString(source)) |
||||
}, |
||||
footnotes: &footnotes{ |
||||
mapping: map[string]int{}, |
||||
}, |
||||
} |
||||
} |
||||
|
||||
func (w *HTMLWriter) emptyClone() *HTMLWriter { |
||||
wcopy := *w |
||||
wcopy.Builder = strings.Builder{} |
||||
return &wcopy |
||||
} |
||||
|
||||
func (w *HTMLWriter) nodesAsString(nodes ...Node) string { |
||||
tmp := w.emptyClone() |
||||
WriteNodes(tmp, nodes...) |
||||
return tmp.String() |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriterWithExtensions() Writer { |
||||
if w.ExtendingWriter != nil { |
||||
return w.ExtendingWriter |
||||
} |
||||
return w |
||||
} |
||||
|
||||
func (w *HTMLWriter) Before(d *Document) { |
||||
w.document = d |
||||
w.log = d.Log |
||||
w.WriteOutline(d) |
||||
} |
||||
|
||||
func (w *HTMLWriter) After(d *Document) { |
||||
w.WriteFootnotes(d) |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteComment(Comment) {} |
||||
func (w *HTMLWriter) WritePropertyDrawer(PropertyDrawer) {} |
||||
|
||||
func (w *HTMLWriter) WriteBlock(b Block) { |
||||
content := "" |
||||
if isRawTextBlock(b.Name) { |
||||
exportWriter := w.emptyClone() |
||||
exportWriter.htmlEscape = false |
||||
WriteNodes(exportWriter, b.Children...) |
||||
content = strings.TrimRightFunc(exportWriter.String(), unicode.IsSpace) |
||||
} else { |
||||
content = w.nodesAsString(b.Children...) |
||||
} |
||||
switch name := b.Name; { |
||||
case name == "SRC": |
||||
lang := "text" |
||||
if len(b.Parameters) >= 1 { |
||||
lang = strings.ToLower(b.Parameters[0]) |
||||
} |
||||
content = w.HighlightCodeBlock(content, lang) |
||||
w.WriteString(fmt.Sprintf("<div class=\"src src-%s\">\n%s\n</div>\n", lang, content)) |
||||
case name == "EXAMPLE": |
||||
w.WriteString(`<pre class="example">` + "\n" + content + "\n</pre>\n") |
||||
case name == "EXPORT" && len(b.Parameters) >= 1 && strings.ToLower(b.Parameters[0]) == "html": |
||||
w.WriteString(content + "\n") |
||||
case name == "QUOTE": |
||||
w.WriteString("<blockquote>\n" + content + "</blockquote>\n") |
||||
case name == "CENTER": |
||||
w.WriteString(`<div class="center-block" style="text-align: center; margin-left: auto; margin-right: auto;">` + "\n") |
||||
w.WriteString(content + "</div>\n") |
||||
default: |
||||
w.WriteString(fmt.Sprintf(`<div class="%s-block">`, strings.ToLower(b.Name)) + "\n") |
||||
w.WriteString(content + "</div>\n") |
||||
} |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteDrawer(d Drawer) { |
||||
WriteNodes(w, d.Children...) |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteKeyword(k Keyword) { |
||||
if k.Key == "HTML" { |
||||
w.WriteString(k.Value + "\n") |
||||
} |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteInclude(i Include) { |
||||
WriteNodes(w, i.Resolve()) |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteFootnoteDefinition(f FootnoteDefinition) { |
||||
w.footnotes.updateDefinition(f) |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteFootnotes(d *Document) { |
||||
if !w.document.GetOption("f") || len(w.footnotes.list) == 0 { |
||||
return |
||||
} |
||||
w.WriteString(`<div class="footnotes">` + "\n") |
||||
w.WriteString(`<hr class="footnotes-separatator">` + "\n") |
||||
w.WriteString(`<div class="footnote-definitions">` + "\n") |
||||
for i, definition := range w.footnotes.list { |
||||
id := i + 1 |
||||
if definition == nil { |
||||
name := "" |
||||
for k, v := range w.footnotes.mapping { |
||||
if v == i { |
||||
name = k |
||||
} |
||||
} |
||||
w.log.Printf("Missing footnote definition for [fn:%s] (#%d)", name, id) |
||||
continue |
||||
} |
||||
w.WriteString(`<div class="footnote-definition">` + "\n") |
||||
w.WriteString(fmt.Sprintf(`<sup id="footnote-%d"><a href="#footnote-reference-%d">%d</a></sup>`, id, id, id) + "\n") |
||||
w.WriteString(`<div class="footnote-body">` + "\n") |
||||
WriteNodes(w, definition.Children...) |
||||
w.WriteString("</div>\n</div>\n") |
||||
} |
||||
w.WriteString("</div>\n</div>\n") |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteOutline(d *Document) { |
||||
if w.document.GetOption("toc") && len(d.Outline.Children) != 0 { |
||||
w.WriteString("<nav>\n<ul>\n") |
||||
for _, section := range d.Outline.Children { |
||||
w.writeSection(section) |
||||
} |
||||
w.WriteString("</ul>\n</nav>\n") |
||||
} |
||||
} |
||||
|
||||
func (w *HTMLWriter) writeSection(section *Section) { |
||||
// NOTE: To satisfy hugo ExtractTOC() check we cannot use `<li>\n` here. Doesn't really matter, just a note.
|
||||
w.WriteString("<li>") |
||||
h := section.Headline |
||||
title := cleanHeadlineTitleForHTMLAnchorRegexp.ReplaceAllString(w.nodesAsString(h.Title...), "") |
||||
w.WriteString(fmt.Sprintf("<a href=\"#%s\">%s</a>\n", h.ID(), title)) |
||||
if len(section.Children) != 0 { |
||||
w.WriteString("<ul>\n") |
||||
for _, section := range section.Children { |
||||
w.writeSection(section) |
||||
} |
||||
w.WriteString("</ul>\n") |
||||
} |
||||
w.WriteString("</li>\n") |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteHeadline(h Headline) { |
||||
for _, excludeTag := range strings.Fields(w.document.Get("EXCLUDE_TAGS")) { |
||||
for _, tag := range h.Tags { |
||||
if excludeTag == tag { |
||||
return |
||||
} |
||||
} |
||||
} |
||||
|
||||
w.WriteString(fmt.Sprintf(`<h%d id="%s">`, h.Lvl, h.ID()) + "\n") |
||||
if w.document.GetOption("todo") && h.Status != "" { |
||||
w.WriteString(fmt.Sprintf(`<span class="todo">%s</span>`, h.Status) + "\n") |
||||
} |
||||
if w.document.GetOption("pri") && h.Priority != "" { |
||||
w.WriteString(fmt.Sprintf(`<span class="priority">[%s]</span>`, h.Priority) + "\n") |
||||
} |
||||
|
||||
WriteNodes(w, h.Title...) |
||||
if w.document.GetOption("tags") && len(h.Tags) != 0 { |
||||
tags := make([]string, len(h.Tags)) |
||||
for i, tag := range h.Tags { |
||||
tags[i] = fmt.Sprintf(`<span>%s</span>`, tag) |
||||
} |
||||
w.WriteString("   ") |
||||
w.WriteString(fmt.Sprintf(`<span class="tags">%s</span>`, strings.Join(tags, " "))) |
||||
} |
||||
w.WriteString(fmt.Sprintf("\n</h%d>\n", h.Lvl)) |
||||
WriteNodes(w, h.Children...) |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteText(t Text) { |
||||
if !w.htmlEscape { |
||||
w.WriteString(t.Content) |
||||
} else if !w.document.GetOption("e") || t.IsRaw { |
||||
w.WriteString(html.EscapeString(t.Content)) |
||||
} else { |
||||
w.WriteString(html.EscapeString(htmlEntityReplacer.Replace(t.Content))) |
||||
} |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteEmphasis(e Emphasis) { |
||||
tags, ok := emphasisTags[e.Kind] |
||||
if !ok { |
||||
panic(fmt.Sprintf("bad emphasis %#v", e)) |
||||
} |
||||
w.WriteString(tags[0]) |
||||
WriteNodes(w, e.Content...) |
||||
w.WriteString(tags[1]) |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteLatexFragment(l LatexFragment) { |
||||
w.WriteString(l.OpeningPair) |
||||
WriteNodes(w, l.Content...) |
||||
w.WriteString(l.ClosingPair) |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteStatisticToken(s StatisticToken) { |
||||
w.WriteString(fmt.Sprintf(`<code class="statistic">[%s]</code>`, s.Content)) |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteLineBreak(l LineBreak) { |
||||
w.WriteString(strings.Repeat("\n", l.Count)) |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteExplicitLineBreak(l ExplicitLineBreak) { |
||||
w.WriteString("<br>\n") |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteFootnoteLink(l FootnoteLink) { |
||||
if !w.document.GetOption("f") { |
||||
return |
||||
} |
||||
i := w.footnotes.add(l) |
||||
id := i + 1 |
||||
w.WriteString(fmt.Sprintf(`<sup class="footnote-reference"><a id="footnote-reference-%d" href="#footnote-%d">%d</a></sup>`, id, id, id)) |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteTimestamp(t Timestamp) { |
||||
if !w.document.GetOption("<") { |
||||
return |
||||
} |
||||
w.WriteString(`<span class="timestamp"><`) |
||||
if t.IsDate { |
||||
w.WriteString(t.Time.Format(datestampFormat)) |
||||
} else { |
||||
w.WriteString(t.Time.Format(timestampFormat)) |
||||
} |
||||
if t.Interval != "" { |
||||
w.WriteString(" " + t.Interval) |
||||
} |
||||
w.WriteString(`></span>`) |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteRegularLink(l RegularLink) { |
||||
url := html.EscapeString(l.URL) |
||||
if l.Protocol == "file" { |
||||
url = url[len("file:"):] |
||||
} |
||||
description := url |
||||
if l.Description != nil { |
||||
description = w.nodesAsString(l.Description...) |
||||
} |
||||
switch l.Kind() { |
||||
case "image": |
||||
w.WriteString(fmt.Sprintf(`<img src="%s" alt="%s" title="%s" />`, url, description, description)) |
||||
case "video": |
||||
w.WriteString(fmt.Sprintf(`<video src="%s" title="%s">%s</video>`, url, description, description)) |
||||
default: |
||||
w.WriteString(fmt.Sprintf(`<a href="%s">%s</a>`, url, description)) |
||||
} |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteList(l List) { |
||||
tags, ok := listTags[l.Kind] |
||||
if !ok { |
||||
panic(fmt.Sprintf("bad list kind %#v", l)) |
||||
} |
||||
w.WriteString(tags[0] + "\n") |
||||
WriteNodes(w, l.Items...) |
||||
w.WriteString(tags[1] + "\n") |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteListItem(li ListItem) { |
||||
if li.Status != "" { |
||||
w.WriteString(fmt.Sprintf("<li class=\"%s\">\n", listItemStatuses[li.Status])) |
||||
} else { |
||||
w.WriteString("<li>\n") |
||||
} |
||||
WriteNodes(w, li.Children...) |
||||
w.WriteString("</li>\n") |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteDescriptiveListItem(di DescriptiveListItem) { |
||||
if di.Status != "" { |
||||
w.WriteString(fmt.Sprintf("<dt class=\"%s\">\n", listItemStatuses[di.Status])) |
||||
} else { |
||||
w.WriteString("<dt>\n") |
||||
} |
||||
|
||||
if len(di.Term) != 0 { |
||||
WriteNodes(w, di.Term...) |
||||
} else { |
||||
w.WriteString("?") |
||||
} |
||||
w.WriteString("\n</dt>\n") |
||||
w.WriteString("<dd>\n") |
||||
WriteNodes(w, di.Details...) |
||||
w.WriteString("</dd>\n") |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteParagraph(p Paragraph) { |
||||
if len(p.Children) == 0 { |
||||
return |
||||
} |
||||
w.WriteString("<p>") |
||||
if _, ok := p.Children[0].(LineBreak); !ok { |
||||
w.WriteString("\n") |
||||
} |
||||
WriteNodes(w, p.Children...) |
||||
w.WriteString("\n</p>\n") |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteExample(e Example) { |
||||
w.WriteString(`<pre class="example">` + "\n") |
||||
if len(e.Children) != 0 { |
||||
for _, n := range e.Children { |
||||
WriteNodes(w, n) |
||||
w.WriteString("\n") |
||||
} |
||||
} |
||||
w.WriteString("</pre>\n") |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteHorizontalRule(h HorizontalRule) { |
||||
w.WriteString("<hr>\n") |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteNodeWithMeta(n NodeWithMeta) { |
||||
out := w.nodesAsString(n.Node) |
||||
if p, ok := n.Node.(Paragraph); ok { |
||||
if len(p.Children) == 1 && isImageOrVideoLink(p.Children[0]) { |
||||
out = w.nodesAsString(p.Children[0]) |
||||
} |
||||
} |
||||
for _, attributes := range n.Meta.HTMLAttributes { |
||||
out = w.withHTMLAttributes(out, attributes...) + "\n" |
||||
} |
||||
if len(n.Meta.Caption) != 0 { |
||||
caption := "" |
||||
for i, ns := range n.Meta.Caption { |
||||
if i != 0 { |
||||
caption += " " |
||||
} |
||||
caption += w.nodesAsString(ns...) |
||||
} |
||||
out = fmt.Sprintf("<figure>\n%s<figcaption>\n%s\n</figcaption>\n</figure>\n", out, caption) |
||||
} |
||||
w.WriteString(out) |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteNodeWithName(n NodeWithName) { |
||||
WriteNodes(w, n.Node) |
||||
} |
||||
|
||||
func (w *HTMLWriter) WriteTable(t Table) { |
||||
w.WriteString("<table>\n") |
||||
beforeFirstContentRow := true |
||||
for i, row := range t.Rows { |
||||
if row.IsSpecial || len(row.Columns) == 0 { |
||||
continue |
||||
} |
||||
if beforeFirstContentRow { |
||||
beforeFirstContentRow = false |
||||
if i+1 < len(t.Rows) && len(t.Rows[i+1].Columns) == 0 { |
||||
w.WriteString("<thead>\n") |
||||
w.writeTableColumns(row.Columns, "th") |
||||
w.WriteString("</thead>\n<tbody>\n") |
||||
continue |
||||
} else { |
||||
w.WriteString("<tbody>\n") |
||||
} |
||||
} |
||||
w.writeTableColumns(row.Columns, "td") |
||||
} |
||||
w.WriteString("</tbody>\n</table>\n") |
||||
} |
||||
|
||||
func (w *HTMLWriter) writeTableColumns(columns []Column, tag string) { |
||||
w.WriteString("<tr>\n") |
||||
for _, column := range columns { |
||||
if column.Align == "" { |
||||
w.WriteString(fmt.Sprintf("<%s>", tag)) |
||||
} else { |
||||
w.WriteString(fmt.Sprintf(`<%s class="align-%s">`, tag, column.Align)) |
||||
} |
||||
WriteNodes(w, column.Children...) |
||||
w.WriteString(fmt.Sprintf("</%s>\n", tag)) |
||||
} |
||||
w.WriteString("</tr>\n") |
||||
} |
||||
|
||||
func (w *HTMLWriter) withHTMLAttributes(input string, kvs ...string) string { |
||||
if len(kvs)%2 != 0 { |
||||
w.log.Printf("withHTMLAttributes: Len of kvs must be even: %#v", kvs) |
||||
return input |
||||
} |
||||
context := &h.Node{Type: h.ElementNode, Data: "body", DataAtom: atom.Body} |
||||
nodes, err := h.ParseFragment(strings.NewReader(strings.TrimSpace(input)), context) |
||||
if err != nil || len(nodes) != 1 { |
||||
w.log.Printf("withHTMLAttributes: Could not extend attributes of %s: %v (%s)", input, nodes, err) |
||||
return input |
||||
} |
||||
out, node := strings.Builder{}, nodes[0] |
||||
for i := 0; i < len(kvs)-1; i += 2 { |
||||
node.Attr = setHTMLAttribute(node.Attr, strings.TrimPrefix(kvs[i], ":"), kvs[i+1]) |
||||
} |
||||
err = h.Render(&out, nodes[0]) |
||||
if err != nil { |
||||
w.log.Printf("withHTMLAttributes: Could not extend attributes of %s: %v (%s)", input, node, err) |
||||
return input |
||||
} |
||||
return out.String() |
||||
} |
||||
|
||||
func setHTMLAttribute(attributes []h.Attribute, k, v string) []h.Attribute { |
||||
for i, a := range attributes { |
||||
if strings.ToLower(a.Key) == strings.ToLower(k) { |
||||
switch strings.ToLower(k) { |
||||
case "class", "style": |
||||
attributes[i].Val += " " + v |
||||
default: |
||||
attributes[i].Val = v |
||||
} |
||||
return attributes |
||||
} |
||||
} |
||||
return append(attributes, h.Attribute{Namespace: "", Key: k, Val: v}) |
||||
} |
||||
|
||||
func (fs *footnotes) add(f FootnoteLink) int { |
||||
if i, ok := fs.mapping[f.Name]; ok && f.Name != "" { |
||||
return i |
||||
} |
||||
fs.list = append(fs.list, f.Definition) |
||||
i := len(fs.list) - 1 |
||||
if f.Name != "" { |
||||
fs.mapping[f.Name] = i |
||||
} |
||||
return i |
||||
} |
||||
|
||||
func (fs *footnotes) updateDefinition(f FootnoteDefinition) { |
||||
if i, ok := fs.mapping[f.Name]; ok { |
||||
fs.list[i] = &f |
||||
} |
||||
} |
@ -0,0 +1,357 @@ |
||||
package org |
||||
|
||||
import ( |
||||
"fmt" |
||||
"path" |
||||
"regexp" |
||||
"strings" |
||||
"time" |
||||
"unicode" |
||||
) |
||||
|
||||
type Text struct { |
||||
Content string |
||||
IsRaw bool |
||||
} |
||||
|
||||
type LineBreak struct{ Count int } |
||||
type ExplicitLineBreak struct{} |
||||
|
||||
type StatisticToken struct{ Content string } |
||||
|
||||
type Timestamp struct { |
||||
Time time.Time |
||||
IsDate bool |
||||
Interval string |
||||
} |
||||
|
||||
type Emphasis struct { |
||||
Kind string |
||||
Content []Node |
||||
} |
||||
|
||||
type LatexFragment struct { |
||||
OpeningPair string |
||||
ClosingPair string |
||||
Content []Node |
||||
} |
||||
|
||||
type FootnoteLink struct { |
||||
Name string |
||||
Definition *FootnoteDefinition |
||||
} |
||||
|
||||
type RegularLink struct { |
||||
Protocol string |
||||
Description []Node |
||||
URL string |
||||
AutoLink bool |
||||
} |
||||
|
||||
var validURLCharacters = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789-._~:/?#[]@!$&'()*+,;=" |
||||
var autolinkProtocols = regexp.MustCompile(`^(https?|ftp|file)$`) |
||||
var imageExtensionRegexp = regexp.MustCompile(`^[.](png|gif|jpe?g|svg|tiff?)$`) |
||||
var videoExtensionRegexp = regexp.MustCompile(`^[.](webm|mp4)$`) |
||||
|
||||
var subScriptSuperScriptRegexp = regexp.MustCompile(`^([_^]){([^{}]+?)}`) |
||||
var timestampRegexp = regexp.MustCompile(`^<(\d{4}-\d{2}-\d{2})( [A-Za-z]+)?( \d{2}:\d{2})?( \+\d+[dwmy])?>`) |
||||
var footnoteRegexp = regexp.MustCompile(`^\[fn:([\w-]*?)(:(.*?))?\]`) |
||||
var statisticsTokenRegexp = regexp.MustCompile(`^\[(\d+/\d+|\d+%)\]`) |
||||
var latexFragmentRegexp = regexp.MustCompile(`(?s)^\\begin{(\w+)}(.*)\\end{(\w+)}`) |
||||
|
||||
var timestampFormat = "2006-01-02 Mon 15:04" |
||||
var datestampFormat = "2006-01-02 Mon" |
||||
|
||||
var latexFragmentPairs = map[string]string{ |
||||
`\(`: `\)`, |
||||
`\[`: `\]`, |
||||
`$$`: `$$`, |
||||
} |
||||
|
||||
func (d *Document) parseInline(input string) (nodes []Node) { |
||||
previous, current := 0, 0 |
||||
for current < len(input) { |
||||
rewind, consumed, node := 0, 0, (Node)(nil) |
||||
switch input[current] { |
||||
case '^': |
||||
consumed, node = d.parseSubOrSuperScript(input, current) |
||||
case '_': |
||||
consumed, node = d.parseSubScriptOrEmphasis(input, current) |
||||
case '*', '/', '+': |
||||
consumed, node = d.parseEmphasis(input, current, false) |
||||
case '=', '~': |
||||
consumed, node = d.parseEmphasis(input, current, true) |
||||
case '[': |
||||
consumed, node = d.parseOpeningBracket(input, current) |
||||
case '<': |
||||
consumed, node = d.parseTimestamp(input, current) |
||||
case '\\': |
||||
consumed, node = d.parseExplicitLineBreakOrLatexFragment(input, current) |
||||
case '$': |
||||
consumed, node = d.parseLatexFragment(input, current) |
||||
case '\n': |
||||
consumed, node = d.parseLineBreak(input, current) |
||||
case ':': |
||||
rewind, consumed, node = d.parseAutoLink(input, current) |
||||
current -= rewind |
||||
} |
||||
if consumed != 0 { |
||||
if current > previous { |
||||
nodes = append(nodes, Text{input[previous:current], false}) |
||||
} |
||||
if node != nil { |
||||
nodes = append(nodes, node) |
||||
} |
||||
current += consumed |
||||
previous = current |
||||
} else { |
||||
current++ |
||||
} |
||||
} |
||||
|
||||
if previous < len(input) { |
||||
nodes = append(nodes, Text{input[previous:], false}) |
||||
} |
||||
return nodes |
||||
} |
||||
|
||||
func (d *Document) parseRawInline(input string) (nodes []Node) { |
||||
previous, current := 0, 0 |
||||
for current < len(input) { |
||||
if input[current] == '\n' { |
||||
consumed, node := d.parseLineBreak(input, current) |
||||
if current > previous { |
||||
nodes = append(nodes, Text{input[previous:current], true}) |
||||
} |
||||
nodes = append(nodes, node) |
||||
current += consumed |
||||
previous = current |
||||
} else { |
||||
current++ |
||||
} |
||||
} |
||||
if previous < len(input) { |
||||
nodes = append(nodes, Text{input[previous:], true}) |
||||
} |
||||
return nodes |
||||
} |
||||
|
||||
func (d *Document) parseLineBreak(input string, start int) (int, Node) { |
||||
i := start |
||||
for ; i < len(input) && input[i] == '\n'; i++ { |
||||
} |
||||
return i - start, LineBreak{i - start} |
||||
} |
||||
|
||||
func (d *Document) parseExplicitLineBreakOrLatexFragment(input string, start int) (int, Node) { |
||||
switch { |
||||
case start+2 >= len(input): |
||||
case input[start+1] == '\\' && start != 0 && input[start-1] != '\n': |
||||
for i := start + 2; unicode.IsSpace(rune(input[i])); i++ { |
||||
if i >= len(input) || input[i] == '\n' { |
||||
return i + 1 - start, ExplicitLineBreak{} |
||||
} |
||||
} |
||||
case input[start+1] == '(' || input[start+1] == '[': |
||||
return d.parseLatexFragment(input, start) |
||||
case strings.Index(input[start:], `\begin{`) == 0: |
||||
if m := latexFragmentRegexp.FindStringSubmatch(input[start:]); m != nil { |
||||
if open, content, close := m[1], m[2], m[3]; open == close { |
||||
openingPair, closingPair := `\begin{`+open+`}`, `\end{`+close+`}` |
||||
i := strings.Index(input[start:], closingPair) |
||||
return i + len(closingPair), LatexFragment{openingPair, closingPair, d.parseRawInline(content)} |
||||
} |
||||
} |
||||
} |
||||
return 0, nil |
||||
} |
||||
|
||||
func (d *Document) parseLatexFragment(input string, start int) (int, Node) { |
||||
if start+2 >= len(input) { |
||||
return 0, nil |
||||
} |
||||
openingPair := input[start : start+2] |
||||
closingPair := latexFragmentPairs[openingPair] |
||||
if i := strings.Index(input[start+2:], closingPair); i != -1 { |
||||
content := d.parseRawInline(input[start+2 : start+2+i]) |
||||
return i + 2 + 2, LatexFragment{openingPair, closingPair, content} |
||||
} |
||||
return 0, nil |
||||
} |
||||
|
||||
func (d *Document) parseSubOrSuperScript(input string, start int) (int, Node) { |
||||
if m := subScriptSuperScriptRegexp.FindStringSubmatch(input[start:]); m != nil { |
||||
return len(m[2]) + 3, Emphasis{m[1] + "{}", []Node{Text{m[2], false}}} |
||||
} |
||||
return 0, nil |
||||
} |
||||
|
||||
func (d *Document) parseSubScriptOrEmphasis(input string, start int) (int, Node) { |
||||
if consumed, node := d.parseSubOrSuperScript(input, start); consumed != 0 { |
||||
return consumed, node |
||||
} |
||||
return d.parseEmphasis(input, start, false) |
||||
} |
||||
|
||||
func (d *Document) parseOpeningBracket(input string, start int) (int, Node) { |
||||
if len(input[start:]) >= 2 && input[start] == '[' && input[start+1] == '[' { |
||||
return d.parseRegularLink(input, start) |
||||
} else if footnoteRegexp.MatchString(input[start:]) { |
||||
return d.parseFootnoteReference(input, start) |
||||
} else if statisticsTokenRegexp.MatchString(input[start:]) { |
||||
return d.parseStatisticToken(input, start) |
||||
} |
||||
return 0, nil |
||||
} |
||||
|
||||
func (d *Document) parseFootnoteReference(input string, start int) (int, Node) { |
||||
if m := footnoteRegexp.FindStringSubmatch(input[start:]); m != nil { |
||||
name, definition := m[1], m[3] |
||||
if name == "" && definition == "" { |
||||
return 0, nil |
||||
} |
||||
link := FootnoteLink{name, nil} |
||||
if definition != "" { |
||||
link.Definition = &FootnoteDefinition{name, []Node{Paragraph{d.parseInline(definition)}}, true} |
||||
} |
||||
return len(m[0]), link |
||||
} |
||||
return 0, nil |
||||
} |
||||
|
||||
func (d *Document) parseStatisticToken(input string, start int) (int, Node) { |
||||
if m := statisticsTokenRegexp.FindStringSubmatch(input[start:]); m != nil { |
||||
return len(m[1]) + 2, StatisticToken{m[1]} |
||||
} |
||||
return 0, nil |
||||
} |
||||
|
||||
func (d *Document) parseAutoLink(input string, start int) (int, int, Node) { |
||||
if !d.AutoLink || start == 0 || len(input[start:]) < 3 || input[start:start+3] != "://" { |
||||
return 0, 0, nil |
||||
} |
||||
protocolStart, protocol := start-1, "" |
||||
for ; protocolStart > 0; protocolStart-- { |
||||
if !unicode.IsLetter(rune(input[protocolStart])) { |
||||
protocolStart++ |
||||
break |
||||
} |
||||
} |
||||
if m := autolinkProtocols.FindStringSubmatch(input[protocolStart:start]); m != nil { |
||||
protocol = m[1] |
||||
} else { |
||||
return 0, 0, nil |
||||
} |
||||
end := start |
||||
for ; end < len(input) && strings.ContainsRune(validURLCharacters, rune(input[end])); end++ { |
||||
} |
||||
path := input[start:end] |
||||
if path == "://" { |
||||
return 0, 0, nil |
||||
} |
||||
return len(protocol), len(path + protocol), RegularLink{protocol, nil, protocol + path, true} |
||||
} |
||||
|
||||
func (d *Document) parseRegularLink(input string, start int) (int, Node) { |
||||
input = input[start:] |
||||
if len(input) < 3 || input[:2] != "[[" || input[2] == '[' { |
||||
return 0, nil |
||||
} |
||||
end := strings.Index(input, "]]") |
||||
if end == -1 { |
||||
return 0, nil |
||||
} |
||||
rawLinkParts := strings.Split(input[2:end], "][") |
||||
description, link := ([]Node)(nil), rawLinkParts[0] |
||||
if len(rawLinkParts) == 2 { |
||||
link, description = rawLinkParts[0], d.parseInline(rawLinkParts[1]) |
||||
} |
||||
if strings.ContainsRune(link, '\n') { |
||||
return 0, nil |
||||
} |
||||
consumed := end + 2 |
||||
protocol, linkParts := "", strings.SplitN(link, ":", 2) |
||||
if len(linkParts) == 2 { |
||||
protocol = linkParts[0] |
||||
} |
||||
return consumed, RegularLink{protocol, description, link, false} |
||||
} |
||||
|
||||
func (d *Document) parseTimestamp(input string, start int) (int, Node) { |
||||
if m := timestampRegexp.FindStringSubmatch(input[start:]); m != nil { |
||||
ddmmyy, hhmm, interval, isDate := m[1], m[3], strings.TrimSpace(m[4]), false |
||||
if hhmm == "" { |
||||
hhmm, isDate = "00:00", true |
||||
} |
||||
t, err := time.Parse(timestampFormat, fmt.Sprintf("%s Mon %s", ddmmyy, hhmm)) |
||||
if err != nil { |
||||
return 0, nil |
||||
} |
||||
timestamp := Timestamp{t, isDate, interval} |
||||
return len(m[0]), timestamp |
||||
} |
||||
return 0, nil |
||||
} |
||||
|
||||
func (d *Document) parseEmphasis(input string, start int, isRaw bool) (int, Node) { |
||||
marker, i := input[start], start |
||||
if !hasValidPreAndBorderChars(input, i) { |
||||
return 0, nil |
||||
} |
||||
for i, consumedNewLines := i+1, 0; i < len(input) && consumedNewLines <= d.MaxEmphasisNewLines; i++ { |
||||
if input[i] == '\n' { |
||||
consumedNewLines++ |
||||
} |
||||
|
||||
if input[i] == marker && i != start+1 && hasValidPostAndBorderChars(input, i) { |
||||
if isRaw { |
||||
return i + 1 - start, Emphasis{input[start : start+1], d.parseRawInline(input[start+1 : i])} |
||||
} |
||||
return i + 1 - start, Emphasis{input[start : start+1], d.parseInline(input[start+1 : i])} |
||||
} |
||||
} |
||||
return 0, nil |
||||
} |
||||
|
||||
// see org-emphasis-regexp-components (emacs elisp variable)
|
||||
|
||||
func hasValidPreAndBorderChars(input string, i int) bool { |
||||
return (i+1 >= len(input) || isValidBorderChar(rune(input[i+1]))) && (i == 0 || isValidPreChar(rune(input[i-1]))) |
||||
} |
||||
|
||||
func hasValidPostAndBorderChars(input string, i int) bool { |
||||
return (i == 0 || isValidBorderChar(rune(input[i-1]))) && (i+1 >= len(input) || isValidPostChar(rune(input[i+1]))) |
||||
} |
||||
|
||||
func isValidPreChar(r rune) bool { |
||||
return unicode.IsSpace(r) || strings.ContainsRune(`-({'"`, r) |
||||
} |
||||
|
||||
func isValidPostChar(r rune) bool { |
||||
return unicode.IsSpace(r) || strings.ContainsRune(`-.,:!?;'")}[`, r) |
||||
} |
||||
|
||||
func isValidBorderChar(r rune) bool { return !unicode.IsSpace(r) } |
||||
|
||||
func (l RegularLink) Kind() string { |
||||
if p := l.Protocol; l.Description != nil || (p != "" && p != "file" && p != "http" && p != "https") { |
||||
return "regular" |
||||
} |
||||
if imageExtensionRegexp.MatchString(path.Ext(l.URL)) { |
||||
return "image" |
||||
} |
||||
if videoExtensionRegexp.MatchString(path.Ext(l.URL)) { |
||||
return "video" |
||||
} |
||||
return "regular" |
||||
} |
||||
|
||||
func (n Text) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n LineBreak) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n ExplicitLineBreak) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n StatisticToken) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n Emphasis) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n LatexFragment) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n FootnoteLink) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n RegularLink) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n Timestamp) String() string { return orgWriter.nodesAsString(n) } |
@ -0,0 +1,184 @@ |
||||
package org |
||||
|
||||
import ( |
||||
"bytes" |
||||
"path/filepath" |
||||
"regexp" |
||||
"strings" |
||||
) |
||||
|
||||
type Comment struct{ Content string } |
||||
|
||||
type Keyword struct { |
||||
Key string |
||||
Value string |
||||
} |
||||
|
||||
type NodeWithName struct { |
||||
Name string |
||||
Node Node |
||||
} |
||||
|
||||
type NodeWithMeta struct { |
||||
Node Node |
||||
Meta Metadata |
||||
} |
||||
|
||||
type Metadata struct { |
||||
Caption [][]Node |
||||
HTMLAttributes [][]string |
||||
} |
||||
|
||||
type Include struct { |
||||
Keyword |
||||
Resolve func() Node |
||||
} |
||||
|
||||
var keywordRegexp = regexp.MustCompile(`^(\s*)#\+([^:]+):(\s+(.*)|$)`) |
||||
var commentRegexp = regexp.MustCompile(`^(\s*)#(.*)`) |
||||
|
||||
var includeFileRegexp = regexp.MustCompile(`(?i)^"([^"]+)" (src|example|export) (\w+)$`) |
||||
var attributeRegexp = regexp.MustCompile(`(?:^|\s+)(:[-\w]+)\s+(.*)$`) |
||||
|
||||
func lexKeywordOrComment(line string) (token, bool) { |
||||
if m := keywordRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"keyword", len(m[1]), m[2], m}, true |
||||
} else if m := commentRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"comment", len(m[1]), m[2], m}, true |
||||
} |
||||
return nilToken, false |
||||
} |
||||
|
||||
func (d *Document) parseComment(i int, stop stopFn) (int, Node) { |
||||
return 1, Comment{d.tokens[i].content} |
||||
} |
||||
|
||||
func (d *Document) parseKeyword(i int, stop stopFn) (int, Node) { |
||||
k := parseKeyword(d.tokens[i]) |
||||
switch k.Key { |
||||
case "NAME": |
||||
return d.parseNodeWithName(k, i, stop) |
||||
case "SETUPFILE": |
||||
return d.loadSetupFile(k) |
||||
case "INCLUDE": |
||||
return d.parseInclude(k) |
||||
case "CAPTION", "ATTR_HTML": |
||||
consumed, node := d.parseAffiliated(i, stop) |
||||
if consumed != 0 { |
||||
return consumed, node |
||||
} |
||||
fallthrough |
||||
default: |
||||
if _, ok := d.BufferSettings[k.Key]; ok { |
||||
d.BufferSettings[k.Key] = strings.Join([]string{d.BufferSettings[k.Key], k.Value}, "\n") |
||||
} else { |
||||
d.BufferSettings[k.Key] = k.Value |
||||
} |
||||
return 1, k |
||||
} |
||||
} |
||||
|
||||
func (d *Document) parseNodeWithName(k Keyword, i int, stop stopFn) (int, Node) { |
||||
if stop(d, i+1) { |
||||
return 0, nil |
||||
} |
||||
consumed, node := d.parseOne(i+1, stop) |
||||
if consumed == 0 || node == nil { |
||||
return 0, nil |
||||
} |
||||
d.NamedNodes[k.Value] = node |
||||
return consumed + 1, NodeWithName{k.Value, node} |
||||
} |
||||
|
||||
func (d *Document) parseAffiliated(i int, stop stopFn) (int, Node) { |
||||
start, meta := i, Metadata{} |
||||
for ; !stop(d, i) && d.tokens[i].kind == "keyword"; i++ { |
||||
switch k := parseKeyword(d.tokens[i]); k.Key { |
||||
case "CAPTION": |
||||
meta.Caption = append(meta.Caption, d.parseInline(k.Value)) |
||||
case "ATTR_HTML": |
||||
attributes, rest := []string{}, k.Value |
||||
for { |
||||
if k, m := "", attributeRegexp.FindStringSubmatch(rest); m != nil { |
||||
k, rest = m[1], m[2] |
||||
attributes = append(attributes, k) |
||||
if v, m := "", attributeRegexp.FindStringSubmatchIndex(rest); m != nil { |
||||
v, rest = rest[:m[0]], rest[m[0]:] |
||||
attributes = append(attributes, v) |
||||
} else { |
||||
attributes = append(attributes, strings.TrimSpace(rest)) |
||||
break |
||||
} |
||||
} else { |
||||
break |
||||
} |
||||
} |
||||
meta.HTMLAttributes = append(meta.HTMLAttributes, attributes) |
||||
default: |
||||
return 0, nil |
||||
} |
||||
} |
||||
if stop(d, i) { |
||||
return 0, nil |
||||
} |
||||
consumed, node := d.parseOne(i, stop) |
||||
if consumed == 0 || node == nil { |
||||
return 0, nil |
||||
} |
||||
i += consumed |
||||
return i - start, NodeWithMeta{node, meta} |
||||
} |
||||
|
||||
func parseKeyword(t token) Keyword { |
||||
k, v := t.matches[2], t.matches[4] |
||||
return Keyword{strings.ToUpper(k), strings.TrimSpace(v)} |
||||
} |
||||
|
||||
func (d *Document) parseInclude(k Keyword) (int, Node) { |
||||
resolve := func() Node { |
||||
d.Log.Printf("Bad include %#v", k) |
||||
return k |
||||
} |
||||
if m := includeFileRegexp.FindStringSubmatch(k.Value); m != nil { |
||||
path, kind, lang := m[1], m[2], m[3] |
||||
if !filepath.IsAbs(path) { |
||||
path = filepath.Join(filepath.Dir(d.Path), path) |
||||
} |
||||
resolve = func() Node { |
||||
bs, err := d.ReadFile(path) |
||||
if err != nil { |
||||
d.Log.Printf("Bad include %#v: %s", k, err) |
||||
return k |
||||
} |
||||
return Block{strings.ToUpper(kind), []string{lang}, d.parseRawInline(string(bs))} |
||||
} |
||||
} |
||||
return 1, Include{k, resolve} |
||||
} |
||||
|
||||
func (d *Document) loadSetupFile(k Keyword) (int, Node) { |
||||
path := k.Value |
||||
if !filepath.IsAbs(path) { |
||||
path = filepath.Join(filepath.Dir(d.Path), path) |
||||
} |
||||
bs, err := d.ReadFile(path) |
||||
if err != nil { |
||||
d.Log.Printf("Bad setup file: %#v: %s", k, err) |
||||
return 1, k |
||||
} |
||||
setupDocument := d.Configuration.Parse(bytes.NewReader(bs), path) |
||||
if err := setupDocument.Error; err != nil { |
||||
d.Log.Printf("Bad setup file: %#v: %s", k, err) |
||||
return 1, k |
||||
} |
||||
for k, v := range setupDocument.BufferSettings { |
||||
d.BufferSettings[k] = v |
||||
} |
||||
return 1, k |
||||
} |
||||
|
||||
func (n Comment) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n Keyword) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n NodeWithMeta) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n NodeWithName) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n Include) String() string { return orgWriter.nodesAsString(n) } |
@ -0,0 +1,114 @@ |
||||
package org |
||||
|
||||
import ( |
||||
"fmt" |
||||
"regexp" |
||||
"strings" |
||||
"unicode" |
||||
) |
||||
|
||||
type List struct { |
||||
Kind string |
||||
Items []Node |
||||
} |
||||
|
||||
type ListItem struct { |
||||
Bullet string |
||||
Status string |
||||
Children []Node |
||||
} |
||||
|
||||
type DescriptiveListItem struct { |
||||
Bullet string |
||||
Status string |
||||
Term []Node |
||||
Details []Node |
||||
} |
||||
|
||||
var unorderedListRegexp = regexp.MustCompile(`^(\s*)([+*-])(\s+(.*)|$)`) |
||||
var orderedListRegexp = regexp.MustCompile(`^(\s*)(([0-9]+|[a-zA-Z])[.)])(\s+(.*)|$)`) |
||||
var descriptiveListItemRegexp = regexp.MustCompile(`\s::(\s|$)`) |
||||
var listItemStatusRegexp = regexp.MustCompile(`\[( |X|-)\]\s`) |
||||
|
||||
func lexList(line string) (token, bool) { |
||||
if m := unorderedListRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"unorderedList", len(m[1]), m[4], m}, true |
||||
} else if m := orderedListRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"orderedList", len(m[1]), m[5], m}, true |
||||
} |
||||
return nilToken, false |
||||
} |
||||
|
||||
func isListToken(t token) bool { |
||||
return t.kind == "unorderedList" || t.kind == "orderedList" |
||||
} |
||||
|
||||
func listKind(t token) (string, string) { |
||||
kind := "" |
||||
switch bullet := t.matches[2]; { |
||||
case bullet == "*" || bullet == "+" || bullet == "-": |
||||
kind = "unordered" |
||||
case unicode.IsLetter(rune(bullet[0])), unicode.IsDigit(rune(bullet[0])): |
||||
kind = "ordered" |
||||
default: |
||||
panic(fmt.Sprintf("bad list bullet '%s': %#v", bullet, t)) |
||||
} |
||||
if descriptiveListItemRegexp.MatchString(t.content) { |
||||
return kind, "descriptive" |
||||
} |
||||
return kind, kind |
||||
} |
||||
|
||||
func (d *Document) parseList(i int, parentStop stopFn) (int, Node) { |
||||
start, lvl := i, d.tokens[i].lvl |
||||
listMainKind, kind := listKind(d.tokens[i]) |
||||
list := List{Kind: kind} |
||||
stop := func(*Document, int) bool { |
||||
if parentStop(d, i) || d.tokens[i].lvl != lvl || !isListToken(d.tokens[i]) { |
||||
return true |
||||
} |
||||
itemMainKind, _ := listKind(d.tokens[i]) |
||||
return itemMainKind != listMainKind |
||||
} |
||||
for !stop(d, i) { |
||||
consumed, node := d.parseListItem(list, i, parentStop) |
||||
i += consumed |
||||
list.Items = append(list.Items, node) |
||||
} |
||||
return i - start, list |
||||
} |
||||
|
||||
func (d *Document) parseListItem(l List, i int, parentStop stopFn) (int, Node) { |
||||
start, nodes, bullet := i, []Node{}, d.tokens[i].matches[2] |
||||
minIndent, dterm, content, status := d.tokens[i].lvl+len(bullet), "", d.tokens[i].content, "" |
||||
if m := listItemStatusRegexp.FindStringSubmatch(content); m != nil { |
||||
status, content = m[1], content[len("[ ] "):] |
||||
} |
||||
if l.Kind == "descriptive" { |
||||
if m := descriptiveListItemRegexp.FindStringIndex(content); m != nil { |
||||
dterm, content = content[:m[0]], content[m[1]:] |
||||
} |
||||
} |
||||
|
||||
d.tokens[i] = tokenize(strings.Repeat(" ", minIndent) + content) |
||||
stop := func(d *Document, i int) bool { |
||||
if parentStop(d, i) { |
||||
return true |
||||
} |
||||
t := d.tokens[i] |
||||
return t.lvl < minIndent && !(t.kind == "text" && t.content == "") |
||||
} |
||||
for !stop(d, i) && (i <= start+1 || !isSecondBlankLine(d, i)) { |
||||
consumed, node := d.parseOne(i, stop) |
||||
i += consumed |
||||
nodes = append(nodes, node) |
||||
} |
||||
if l.Kind == "descriptive" { |
||||
return i - start, DescriptiveListItem{bullet, status, d.parseInline(dterm), nodes} |
||||
} |
||||
return i - start, ListItem{bullet, status, nodes} |
||||
} |
||||
|
||||
func (n List) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n ListItem) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n DescriptiveListItem) String() string { return orgWriter.nodesAsString(n) } |
@ -0,0 +1,334 @@ |
||||
package org |
||||
|
||||
import ( |
||||
"fmt" |
||||
"strings" |
||||
"unicode" |
||||
"unicode/utf8" |
||||
) |
||||
|
||||
// OrgWriter export an org document into pretty printed org document.
|
||||
type OrgWriter struct { |
||||
ExtendingWriter Writer |
||||
TagsColumn int |
||||
|
||||
strings.Builder |
||||
indent string |
||||
} |
||||
|
||||
var emphasisOrgBorders = map[string][]string{ |
||||
"_": []string{"_", "_"}, |
||||
"*": []string{"*", "*"}, |
||||
"/": []string{"/", "/"}, |
||||
"+": []string{"+", "+"}, |
||||
"~": []string{"~", "~"}, |
||||
"=": []string{"=", "="}, |
||||
"_{}": []string{"_{", "}"}, |
||||
"^{}": []string{"^{", "}"}, |
||||
} |
||||
|
||||
func NewOrgWriter() *OrgWriter { |
||||
return &OrgWriter{ |
||||
TagsColumn: 77, |
||||
} |
||||
} |
||||
|
||||
func (w *OrgWriter) WriterWithExtensions() Writer { |
||||
if w.ExtendingWriter != nil { |
||||
return w.ExtendingWriter |
||||
} |
||||
return w |
||||
} |
||||
|
||||
func (w *OrgWriter) Before(d *Document) {} |
||||
func (w *OrgWriter) After(d *Document) {} |
||||
|
||||
func (w *OrgWriter) emptyClone() *OrgWriter { |
||||
wcopy := *w |
||||
wcopy.Builder = strings.Builder{} |
||||
return &wcopy |
||||
} |
||||
|
||||
func (w *OrgWriter) nodesAsString(nodes ...Node) string { |
||||
tmp := w.emptyClone() |
||||
WriteNodes(tmp, nodes...) |
||||
return tmp.String() |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteHeadline(h Headline) { |
||||
tmp := w.emptyClone() |
||||
tmp.WriteString(strings.Repeat("*", h.Lvl)) |
||||
if h.Status != "" { |
||||
tmp.WriteString(" " + h.Status) |
||||
} |
||||
if h.Priority != "" { |
||||
tmp.WriteString(" [#" + h.Priority + "]") |
||||
} |
||||
tmp.WriteString(" ") |
||||
WriteNodes(tmp, h.Title...) |
||||
hString := tmp.String() |
||||
if len(h.Tags) != 0 { |
||||
tString := ":" + strings.Join(h.Tags, ":") + ":" |
||||
if n := w.TagsColumn - len(tString) - len(hString); n > 0 { |
||||
w.WriteString(hString + strings.Repeat(" ", n) + tString) |
||||
} else { |
||||
w.WriteString(hString + " " + tString) |
||||
} |
||||
} else { |
||||
w.WriteString(hString) |
||||
} |
||||
w.WriteString("\n") |
||||
if len(h.Children) != 0 { |
||||
w.WriteString(w.indent) |
||||
} |
||||
if h.Properties != nil { |
||||
WriteNodes(w, *h.Properties) |
||||
} |
||||
WriteNodes(w, h.Children...) |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteBlock(b Block) { |
||||
w.WriteString(w.indent + "#+BEGIN_" + b.Name) |
||||
if len(b.Parameters) != 0 { |
||||
w.WriteString(" " + strings.Join(b.Parameters, " ")) |
||||
} |
||||
w.WriteString("\n") |
||||
if isRawTextBlock(b.Name) { |
||||
w.WriteString(w.indent) |
||||
} |
||||
WriteNodes(w, b.Children...) |
||||
if !isRawTextBlock(b.Name) { |
||||
w.WriteString(w.indent) |
||||
} |
||||
w.WriteString("#+END_" + b.Name + "\n") |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteDrawer(d Drawer) { |
||||
w.WriteString(w.indent + ":" + d.Name + ":\n") |
||||
WriteNodes(w, d.Children...) |
||||
w.WriteString(w.indent + ":END:\n") |
||||
} |
||||
|
||||
func (w *OrgWriter) WritePropertyDrawer(d PropertyDrawer) { |
||||
w.WriteString(":PROPERTIES:\n") |
||||
for _, kvPair := range d.Properties { |
||||
k, v := kvPair[0], kvPair[1] |
||||
if v != "" { |
||||
v = " " + v |
||||
} |
||||
w.WriteString(fmt.Sprintf(":%s:%s\n", k, v)) |
||||
} |
||||
w.WriteString(":END:\n") |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteFootnoteDefinition(f FootnoteDefinition) { |
||||
w.WriteString(fmt.Sprintf("[fn:%s]", f.Name)) |
||||
content := w.nodesAsString(f.Children...) |
||||
if content != "" && !unicode.IsSpace(rune(content[0])) { |
||||
w.WriteString(" ") |
||||
} |
||||
w.WriteString(content) |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteParagraph(p Paragraph) { |
||||
content := w.nodesAsString(p.Children...) |
||||
if len(content) > 0 && content[0] != '\n' { |
||||
w.WriteString(w.indent) |
||||
} |
||||
w.WriteString(content + "\n") |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteExample(e Example) { |
||||
for _, n := range e.Children { |
||||
w.WriteString(w.indent + ":") |
||||
if content := w.nodesAsString(n); content != "" { |
||||
w.WriteString(" " + content) |
||||
} |
||||
w.WriteString("\n") |
||||
} |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteKeyword(k Keyword) { |
||||
w.WriteString(w.indent + "#+" + k.Key + ":") |
||||
if k.Value != "" { |
||||
w.WriteString(" " + k.Value) |
||||
} |
||||
w.WriteString("\n") |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteInclude(i Include) { |
||||
w.WriteKeyword(i.Keyword) |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteNodeWithMeta(n NodeWithMeta) { |
||||
for _, ns := range n.Meta.Caption { |
||||
w.WriteString("#+CAPTION: ") |
||||
WriteNodes(w, ns...) |
||||
w.WriteString("\n") |
||||
} |
||||
for _, attributes := range n.Meta.HTMLAttributes { |
||||
w.WriteString("#+ATTR_HTML: ") |
||||
w.WriteString(strings.Join(attributes, " ") + "\n") |
||||
} |
||||
WriteNodes(w, n.Node) |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteNodeWithName(n NodeWithName) { |
||||
w.WriteString(fmt.Sprintf("#+NAME: %s\n", n.Name)) |
||||
WriteNodes(w, n.Node) |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteComment(c Comment) { |
||||
w.WriteString(w.indent + "#" + c.Content + "\n") |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteList(l List) { WriteNodes(w, l.Items...) } |
||||
|
||||
func (w *OrgWriter) WriteListItem(li ListItem) { |
||||
liWriter := w.emptyClone() |
||||
liWriter.indent = w.indent + strings.Repeat(" ", len(li.Bullet)+1) |
||||
WriteNodes(liWriter, li.Children...) |
||||
content := strings.TrimPrefix(liWriter.String(), liWriter.indent) |
||||
w.WriteString(w.indent + li.Bullet) |
||||
if li.Status != "" { |
||||
w.WriteString(fmt.Sprintf(" [%s]", li.Status)) |
||||
} |
||||
if len(content) > 0 && content[0] == '\n' { |
||||
w.WriteString(content) |
||||
} else { |
||||
w.WriteString(" " + content) |
||||
} |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteDescriptiveListItem(di DescriptiveListItem) { |
||||
w.WriteString(w.indent + di.Bullet) |
||||
if di.Status != "" { |
||||
w.WriteString(fmt.Sprintf(" [%s]", di.Status)) |
||||
} |
||||
indent := w.indent + strings.Repeat(" ", len(di.Bullet)+1) |
||||
if len(di.Term) != 0 { |
||||
term := w.nodesAsString(di.Term...) |
||||
w.WriteString(" " + term + " ::") |
||||
indent = indent + strings.Repeat(" ", len(term)+4) |
||||
} |
||||
diWriter := w.emptyClone() |
||||
diWriter.indent = indent |
||||
WriteNodes(diWriter, di.Details...) |
||||
details := strings.TrimPrefix(diWriter.String(), diWriter.indent) |
||||
if len(details) > 0 && details[0] == '\n' { |
||||
w.WriteString(details) |
||||
} else { |
||||
w.WriteString(" " + details) |
||||
} |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteTable(t Table) { |
||||
for _, row := range t.Rows { |
||||
w.WriteString(w.indent) |
||||
if len(row.Columns) == 0 { |
||||
w.WriteString(`|`) |
||||
for i := 0; i < len(t.ColumnInfos); i++ { |
||||
w.WriteString(strings.Repeat("-", t.ColumnInfos[i].Len+2)) |
||||
if i < len(t.ColumnInfos)-1 { |
||||
w.WriteString("+") |
||||
} |
||||
} |
||||
w.WriteString(`|`) |
||||
|
||||
} else { |
||||
w.WriteString(`|`) |
||||
for _, column := range row.Columns { |
||||
w.WriteString(` `) |
||||
content := w.nodesAsString(column.Children...) |
||||
if content == "" { |
||||
content = " " |
||||
} |
||||
n := column.Len - utf8.RuneCountInString(content) |
||||
if n < 0 { |
||||
n = 0 |
||||
} |
||||
if column.Align == "center" { |
||||
if n%2 != 0 { |
||||
w.WriteString(" ") |
||||
} |
||||
w.WriteString(strings.Repeat(" ", n/2) + content + strings.Repeat(" ", n/2)) |
||||
} else if column.Align == "right" { |
||||
w.WriteString(strings.Repeat(" ", n) + content) |
||||
} else { |
||||
w.WriteString(content + strings.Repeat(" ", n)) |
||||
} |
||||
w.WriteString(` |`) |
||||
} |
||||
} |
||||
w.WriteString("\n") |
||||
} |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteHorizontalRule(hr HorizontalRule) { |
||||
w.WriteString(w.indent + "-----\n") |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteText(t Text) { w.WriteString(t.Content) } |
||||
|
||||
func (w *OrgWriter) WriteEmphasis(e Emphasis) { |
||||
borders, ok := emphasisOrgBorders[e.Kind] |
||||
if !ok { |
||||
panic(fmt.Sprintf("bad emphasis %#v", e)) |
||||
} |
||||
w.WriteString(borders[0]) |
||||
WriteNodes(w, e.Content...) |
||||
w.WriteString(borders[1]) |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteLatexFragment(l LatexFragment) { |
||||
w.WriteString(l.OpeningPair) |
||||
WriteNodes(w, l.Content...) |
||||
w.WriteString(l.ClosingPair) |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteStatisticToken(s StatisticToken) { |
||||
w.WriteString(fmt.Sprintf("[%s]", s.Content)) |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteLineBreak(l LineBreak) { |
||||
w.WriteString(strings.Repeat("\n"+w.indent, l.Count)) |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteExplicitLineBreak(l ExplicitLineBreak) { |
||||
w.WriteString(`\\` + "\n" + w.indent) |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteTimestamp(t Timestamp) { |
||||
w.WriteString("<") |
||||
if t.IsDate { |
||||
w.WriteString(t.Time.Format(datestampFormat)) |
||||
} else { |
||||
w.WriteString(t.Time.Format(timestampFormat)) |
||||
} |
||||
if t.Interval != "" { |
||||
w.WriteString(" " + t.Interval) |
||||
} |
||||
w.WriteString(">") |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteFootnoteLink(l FootnoteLink) { |
||||
w.WriteString("[fn:" + l.Name) |
||||
if l.Definition != nil { |
||||
w.WriteString(":") |
||||
WriteNodes(w, l.Definition.Children[0].(Paragraph).Children...) |
||||
} |
||||
w.WriteString("]") |
||||
} |
||||
|
||||
func (w *OrgWriter) WriteRegularLink(l RegularLink) { |
||||
if l.AutoLink { |
||||
w.WriteString(l.URL) |
||||
} else if l.Description == nil { |
||||
w.WriteString(fmt.Sprintf("[[%s]]", l.URL)) |
||||
} else { |
||||
descriptionWriter := w.emptyClone() |
||||
WriteNodes(descriptionWriter, l.Description...) |
||||
description := descriptionWriter.String() |
||||
w.WriteString(fmt.Sprintf("[[%s][%s]]", l.URL, description)) |
||||
} |
||||
} |
@ -0,0 +1,46 @@ |
||||
package org |
||||
|
||||
import ( |
||||
"regexp" |
||||
"strings" |
||||
) |
||||
|
||||
type Paragraph struct{ Children []Node } |
||||
type HorizontalRule struct{} |
||||
|
||||
var horizontalRuleRegexp = regexp.MustCompile(`^(\s*)-{5,}\s*$`) |
||||
var plainTextRegexp = regexp.MustCompile(`^(\s*)(.*)`) |
||||
|
||||
func lexText(line string) (token, bool) { |
||||
if m := plainTextRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"text", len(m[1]), m[2], m}, true |
||||
} |
||||
return nilToken, false |
||||
} |
||||
|
||||
func lexHorizontalRule(line string) (token, bool) { |
||||
if m := horizontalRuleRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"horizontalRule", len(m[1]), "", m}, true |
||||
} |
||||
return nilToken, false |
||||
} |
||||
|
||||
func (d *Document) parseParagraph(i int, parentStop stopFn) (int, Node) { |
||||
lines, start := []string{d.tokens[i].content}, i |
||||
i++ |
||||
stop := func(d *Document, i int) bool { |
||||
return parentStop(d, i) || d.tokens[i].kind != "text" || d.tokens[i].content == "" |
||||
} |
||||
for ; !stop(d, i); i++ { |
||||
lines = append(lines, d.tokens[i].content) |
||||
} |
||||
consumed := i - start |
||||
return consumed, Paragraph{d.parseInline(strings.Join(lines, "\n"))} |
||||
} |
||||
|
||||
func (d *Document) parseHorizontalRule(i int, parentStop stopFn) (int, Node) { |
||||
return 1, HorizontalRule{} |
||||
} |
||||
|
||||
func (n Paragraph) String() string { return orgWriter.nodesAsString(n) } |
||||
func (n HorizontalRule) String() string { return orgWriter.nodesAsString(n) } |
@ -0,0 +1,130 @@ |
||||
package org |
||||
|
||||
import ( |
||||
"regexp" |
||||
"strconv" |
||||
"strings" |
||||
"unicode/utf8" |
||||
) |
||||
|
||||
type Table struct { |
||||
Rows []Row |
||||
ColumnInfos []ColumnInfo |
||||
} |
||||
|
||||
type Row struct { |
||||
Columns []Column |
||||
IsSpecial bool |
||||
} |
||||
|
||||
type Column struct { |
||||
Children []Node |
||||
*ColumnInfo |
||||
} |
||||
|
||||
type ColumnInfo struct { |
||||
Align string |
||||
Len int |
||||
} |
||||
|
||||
var tableSeparatorRegexp = regexp.MustCompile(`^(\s*)(\|[+-|]*)\s*$`) |
||||
var tableRowRegexp = regexp.MustCompile(`^(\s*)(\|.*)`) |
||||
|
||||
var columnAlignRegexp = regexp.MustCompile(`^<(l|c|r)>$`) |
||||
|
||||
func lexTable(line string) (token, bool) { |
||||
if m := tableSeparatorRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"tableSeparator", len(m[1]), m[2], m}, true |
||||
} else if m := tableRowRegexp.FindStringSubmatch(line); m != nil { |
||||
return token{"tableRow", len(m[1]), m[2], m}, true |
||||
} |
||||
return nilToken, false |
||||
} |
||||
|
||||
func (d *Document) parseTable(i int, parentStop stopFn) (int, Node) { |
||||
rawRows, start := [][]string{}, i |
||||
for ; !parentStop(d, i); i++ { |
||||
if t := d.tokens[i]; t.kind == "tableRow" { |
||||
rawRow := strings.FieldsFunc(d.tokens[i].content, func(r rune) bool { return r == '|' }) |
||||
for i := range rawRow { |
||||
rawRow[i] = strings.TrimSpace(rawRow[i]) |
||||
} |
||||
rawRows = append(rawRows, rawRow) |
||||
} else if t.kind == "tableSeparator" { |
||||
rawRows = append(rawRows, nil) |
||||
} else { |
||||
break |
||||
} |
||||
} |
||||
|
||||
table := Table{nil, getColumnInfos(rawRows)} |
||||
for _, rawColumns := range rawRows { |
||||
row := Row{nil, isSpecialRow(rawColumns)} |
||||
if len(rawColumns) != 0 { |
||||
for i := range table.ColumnInfos { |
||||
column := Column{nil, &table.ColumnInfos[i]} |
||||
if i < len(rawColumns) { |
||||
column.Children = d.parseInline(rawColumns[i]) |
||||
} |
||||
row.Columns = append(row.Columns, column) |
||||
} |
||||
} |
||||
table.Rows = append(table.Rows, row) |
||||
} |
||||
return i - start, table |
||||
} |
||||
|
||||
func getColumnInfos(rows [][]string) []ColumnInfo { |
||||
columnCount := 0 |
||||
for _, columns := range rows { |
||||
if n := len(columns); n > columnCount { |
||||
columnCount = n |
||||
} |
||||
} |
||||
|
||||
columnInfos := make([]ColumnInfo, columnCount) |
||||
for i := 0; i < columnCount; i++ { |
||||
countNumeric, countNonNumeric := 0, 0 |
||||
for _, columns := range rows { |
||||
if i >= len(columns) { |
||||
continue |
||||
} |
||||
|
||||
if n := utf8.RuneCountInString(columns[i]); n > columnInfos[i].Len { |
||||
columnInfos[i].Len = n |
||||
} |
||||
|
||||
if m := columnAlignRegexp.FindStringSubmatch(columns[i]); m != nil && isSpecialRow(columns) { |
||||
switch m[1] { |
||||
case "l": |
||||
columnInfos[i].Align = "left" |
||||
case "c": |
||||
columnInfos[i].Align = "center" |
||||
case "r": |
||||
columnInfos[i].Align = "right" |
||||
} |
||||
} else if _, err := strconv.ParseFloat(columns[i], 32); err == nil { |
||||
countNumeric++ |
||||
} else if strings.TrimSpace(columns[i]) != "" { |
||||
countNonNumeric++ |
||||
} |
||||
} |
||||
|
||||
if columnInfos[i].Align == "" && countNumeric >= countNonNumeric { |
||||
columnInfos[i].Align = "right" |
||||
} |
||||
} |
||||
return columnInfos |
||||
} |
||||
|
||||
func isSpecialRow(rawColumns []string) bool { |
||||
isAlignRow := true |
||||
for _, rawColumn := range rawColumns { |
||||
if !columnAlignRegexp.MatchString(rawColumn) && rawColumn != "" { |
||||
isAlignRow = false |
||||
} |
||||
} |
||||
return isAlignRow |
||||
} |
||||
|
||||
func (n Table) String() string { return orgWriter.nodesAsString(n) } |
@ -0,0 +1,19 @@ |
||||
package org |
||||
|
||||
func isSecondBlankLine(d *Document, i int) bool { |
||||
if i-1 <= 0 { |
||||
return false |
||||
} |
||||
t1, t2 := d.tokens[i-1], d.tokens[i] |
||||
if t1.kind == "text" && t2.kind == "text" && t1.content == "" && t2.content == "" { |
||||
return true |
||||
} |
||||
return false |
||||
} |
||||
|
||||
func isImageOrVideoLink(n Node) bool { |
||||
if l, ok := n.(RegularLink); ok && l.Kind() == "video" || l.Kind() == "image" { |
||||
return true |
||||
} |
||||
return false |
||||
} |
@ -0,0 +1,103 @@ |
||||
package org |
||||
|
||||
import "fmt" |
||||
|
||||
// Writer is the interface that is used to export a parsed document into a new format. See Document.Write().
|
||||
type Writer interface { |
||||
Before(*Document) // Before is called before any nodes are passed to the writer.
|
||||
After(*Document) // After is called after all nodes have been passed to the writer.
|
||||
String() string // String is called at the very end to retrieve the final output.
|
||||
|
||||
WriterWithExtensions() Writer |
||||
|
||||
WriteKeyword(Keyword) |
||||
WriteInclude(Include) |
||||
WriteComment(Comment) |
||||
WriteNodeWithMeta(NodeWithMeta) |
||||
WriteNodeWithName(NodeWithName) |
||||
WriteHeadline(Headline) |
||||
WriteBlock(Block) |
||||
WriteExample(Example) |
||||
WriteDrawer(Drawer) |
||||
WritePropertyDrawer(PropertyDrawer) |
||||
WriteList(List) |
||||
WriteListItem(ListItem) |
||||
WriteDescriptiveListItem(DescriptiveListItem) |
||||
WriteTable(Table) |
||||
WriteHorizontalRule(HorizontalRule) |
||||
WriteParagraph(Paragraph) |
||||
WriteText(Text) |
||||
WriteEmphasis(Emphasis) |
||||
WriteLatexFragment(LatexFragment) |
||||
WriteStatisticToken(StatisticToken) |
||||
WriteExplicitLineBreak(ExplicitLineBreak) |
||||
WriteLineBreak(LineBreak) |
||||
WriteRegularLink(RegularLink) |
||||
WriteTimestamp(Timestamp) |
||||
WriteFootnoteLink(FootnoteLink) |
||||
WriteFootnoteDefinition(FootnoteDefinition) |
||||
} |
||||
|
||||
func WriteNodes(w Writer, nodes ...Node) { |
||||
w = w.WriterWithExtensions() |
||||
for _, n := range nodes { |
||||
switch n := n.(type) { |
||||
case Keyword: |
||||
w.WriteKeyword(n) |
||||
case Include: |
||||
w.WriteInclude(n) |
||||
case Comment: |
||||
w.WriteComment(n) |
||||
case NodeWithMeta: |
||||
w.WriteNodeWithMeta(n) |
||||
case NodeWithName: |
||||
w.WriteNodeWithName(n) |
||||
case Headline: |
||||
w.WriteHeadline(n) |
||||
case Block: |
||||
w.WriteBlock(n) |
||||
case Example: |
||||
w.WriteExample(n) |
||||
case Drawer: |
||||
w.WriteDrawer(n) |
||||
case PropertyDrawer: |
||||
w.WritePropertyDrawer(n) |
||||
case List: |
||||
w.WriteList(n) |
||||
case ListItem: |
||||
w.WriteListItem(n) |
||||
case DescriptiveListItem: |
||||
w.WriteDescriptiveListItem(n) |
||||
case Table: |
||||
w.WriteTable(n) |
||||
case HorizontalRule: |
||||
w.WriteHorizontalRule(n) |
||||
case Paragraph: |
||||
w.WriteParagraph(n) |
||||
case Text: |
||||
w.WriteText(n) |
||||
case Emphasis: |
||||
w.WriteEmphasis(n) |
||||
case LatexFragment: |
||||
w.WriteLatexFragment(n) |
||||
case StatisticToken: |
||||
w.WriteStatisticToken(n) |
||||
case ExplicitLineBreak: |
||||
w.WriteExplicitLineBreak(n) |
||||
case LineBreak: |
||||
w.WriteLineBreak(n) |
||||
case RegularLink: |
||||
w.WriteRegularLink(n) |
||||
case Timestamp: |
||||
w.WriteTimestamp(n) |
||||
case FootnoteLink: |
||||
w.WriteFootnoteLink(n) |
||||
case FootnoteDefinition: |
||||
w.WriteFootnoteDefinition(n) |
||||
default: |
||||
if n != nil { |
||||
panic(fmt.Sprintf("bad node %T %#v", n, n)) |
||||
} |
||||
} |
||||
} |
||||
} |
@ -1,32 +0,0 @@ |
||||
// Package blackfriday is a Markdown processor.
|
||||
//
|
||||
// It translates plain text with simple formatting rules into HTML or LaTeX.
|
||||
//
|
||||
// Sanitized Anchor Names
|
||||
//
|
||||
// Blackfriday includes an algorithm for creating sanitized anchor names
|
||||
// corresponding to a given input text. This algorithm is used to create
|
||||
// anchors for headings when EXTENSION_AUTO_HEADER_IDS is enabled. The
|
||||
// algorithm is specified below, so that other packages can create
|
||||
// compatible anchor names and links to those anchors.
|
||||
//
|
||||
// The algorithm iterates over the input text, interpreted as UTF-8,
|
||||
// one Unicode code point (rune) at a time. All runes that are letters (category L)
|
||||
// or numbers (category N) are considered valid characters. They are mapped to
|
||||
// lower case, and included in the output. All other runes are considered
|
||||
// invalid characters. Invalid characters that preceed the first valid character,
|
||||
// as well as invalid character that follow the last valid character
|
||||
// are dropped completely. All other sequences of invalid characters
|
||||
// between two valid characters are replaced with a single dash character '-'.
|
||||
//
|
||||
// SanitizedAnchorName exposes this functionality, and can be used to
|
||||
// create compatible links to the anchor names generated by blackfriday.
|
||||
// This algorithm is also implemented in a small standalone package at
|
||||
// github.com/shurcooL/sanitized_anchor_name. It can be useful for clients
|
||||
// that want a small package and don't need full functionality of blackfriday.
|
||||
package blackfriday |
||||
|
||||
// NOTE: Keep Sanitized Anchor Name algorithm in sync with package
|
||||
// github.com/shurcooL/sanitized_anchor_name.
|
||||
// Otherwise, users of sanitized_anchor_name will get anchor names
|
||||
// that are incompatible with those generated by blackfriday.
|
@ -1,938 +0,0 @@ |
||||
//
|
||||
// Blackfriday Markdown Processor
|
||||
// Available at http://github.com/russross/blackfriday
|
||||
//
|
||||
// Copyright © 2011 Russ Ross <russ@russross.com>.
|
||||
// Distributed under the Simplified BSD License.
|
||||
// See README.md for details.
|
||||
//
|
||||
|
||||
//
|
||||
//
|
||||
// HTML rendering backend
|
||||
//
|
||||
//
|
||||
|
||||
package blackfriday |
||||
|
||||
import ( |
||||
"bytes" |
||||
"fmt" |
||||
"regexp" |
||||
"strconv" |
||||
"strings" |
||||
) |
||||
|
||||
// Html renderer configuration options.
|
||||
const ( |
||||
HTML_SKIP_HTML = 1 << iota // skip preformatted HTML blocks
|
||||
HTML_SKIP_STYLE // skip embedded <style> elements
|
||||
HTML_SKIP_IMAGES // skip embedded images
|
||||
HTML_SKIP_LINKS // skip all links
|
||||
HTML_SAFELINK // only link to trusted protocols
|
||||
HTML_NOFOLLOW_LINKS // only link with rel="nofollow"
|
||||
HTML_NOREFERRER_LINKS // only link with rel="noreferrer"
|
||||
HTML_HREF_TARGET_BLANK // add a blank target
|
||||
HTML_TOC // generate a table of contents
|
||||
HTML_OMIT_CONTENTS // skip the main contents (for a standalone table of contents)
|
||||
HTML_COMPLETE_PAGE // generate a complete HTML page
|
||||
HTML_USE_XHTML // generate XHTML output instead of HTML
|
||||
HTML_USE_SMARTYPANTS // enable smart punctuation substitutions
|
||||
HTML_SMARTYPANTS_FRACTIONS // enable smart fractions (with HTML_USE_SMARTYPANTS)
|
||||
HTML_SMARTYPANTS_DASHES // enable smart dashes (with HTML_USE_SMARTYPANTS)
|
||||
HTML_SMARTYPANTS_LATEX_DASHES // enable LaTeX-style dashes (with HTML_USE_SMARTYPANTS and HTML_SMARTYPANTS_DASHES)
|
||||
HTML_SMARTYPANTS_ANGLED_QUOTES // enable angled double quotes (with HTML_USE_SMARTYPANTS) for double quotes rendering
|
||||
HTML_SMARTYPANTS_QUOTES_NBSP // enable "French guillemets" (with HTML_USE_SMARTYPANTS)
|
||||
HTML_FOOTNOTE_RETURN_LINKS // generate a link at the end of a footnote to return to the source
|
||||
) |
||||
|
||||
var ( |
||||
alignments = []string{ |
||||
"left", |
||||
"right", |
||||
"center", |
||||
} |
||||
|
||||
// TODO: improve this regexp to catch all possible entities:
|
||||
htmlEntity = regexp.MustCompile(`&[a-z]{2,5};`) |
||||
) |
||||
|
||||
type HtmlRendererParameters struct { |
||||
// Prepend this text to each relative URL.
|
||||
AbsolutePrefix string |
||||
// Add this text to each footnote anchor, to ensure uniqueness.
|
||||
FootnoteAnchorPrefix string |
||||
// Show this text inside the <a> tag for a footnote return link, if the
|
||||
// HTML_FOOTNOTE_RETURN_LINKS flag is enabled. If blank, the string
|
||||
// <sup>[return]</sup> is used.
|
||||
FootnoteReturnLinkContents string |
||||
// If set, add this text to the front of each Header ID, to ensure
|
||||
// uniqueness.
|
||||
HeaderIDPrefix string |
||||
// If set, add this text to the back of each Header ID, to ensure uniqueness.
|
||||
HeaderIDSuffix string |
||||
} |
||||
|
||||
// Html is a type that implements the Renderer interface for HTML output.
|
||||
//
|
||||
// Do not create this directly, instead use the HtmlRenderer function.
|
||||
type Html struct { |
||||
flags int // HTML_* options
|
||||
closeTag string // how to end singleton tags: either " />" or ">"
|
||||
title string // document title
|
||||
css string // optional css file url (used with HTML_COMPLETE_PAGE)
|
||||
|
||||
parameters HtmlRendererParameters |
||||
|
||||
// table of contents data
|
||||
tocMarker int |
||||
headerCount int |
||||
currentLevel int |
||||
toc *bytes.Buffer |
||||
|
||||
// Track header IDs to prevent ID collision in a single generation.
|
||||
headerIDs map[string]int |
||||
|
||||
smartypants *smartypantsRenderer |
||||
} |
||||
|
||||
const ( |
||||
xhtmlClose = " />" |
||||
htmlClose = ">" |
||||
) |
||||
|
||||
// HtmlRenderer creates and configures an Html object, which
|
||||
// satisfies the Renderer interface.
|
||||
//
|
||||
// flags is a set of HTML_* options ORed together.
|
||||
// title is the title of the document, and css is a URL for the document's
|
||||
// stylesheet.
|
||||
// title and css are only used when HTML_COMPLETE_PAGE is selected.
|
||||
func HtmlRenderer(flags int, title string, css string) Renderer { |
||||
return HtmlRendererWithParameters(flags, title, css, HtmlRendererParameters{}) |
||||
} |
||||
|
||||
func HtmlRendererWithParameters(flags int, title string, |
||||
css string, renderParameters HtmlRendererParameters) Renderer { |
||||
// configure the rendering engine
|
||||
closeTag := htmlClose |
||||
if flags&HTML_USE_XHTML != 0 { |
||||
closeTag = xhtmlClose |
||||
} |
||||
|
||||
if renderParameters.FootnoteReturnLinkContents == "" { |
||||
renderParameters.FootnoteReturnLinkContents = `<sup>[return]</sup>` |
||||
} |
||||
|
||||
return &Html{ |
||||
flags: flags, |
||||
closeTag: closeTag, |
||||
title: title, |
||||
css: css, |
||||
parameters: renderParameters, |
||||
|
||||
headerCount: 0, |
||||
currentLevel: 0, |
||||
toc: new(bytes.Buffer), |
||||
|
||||
headerIDs: make(map[string]int), |
||||
|
||||
smartypants: smartypants(flags), |
||||
} |
||||
} |
||||
|
||||
// Using if statements is a bit faster than a switch statement. As the compiler
|
||||
// improves, this should be unnecessary this is only worthwhile because
|
||||
// attrEscape is the single largest CPU user in normal use.
|
||||
// Also tried using map, but that gave a ~3x slowdown.
|
||||
func escapeSingleChar(char byte) (string, bool) { |
||||
if char == '"' { |
||||
return """, true |
||||
} |
||||
if char == '&' { |
||||
return "&", true |
||||
} |
||||
if char == '<' { |
||||
return "<", true |
||||
} |
||||
if char == '>' { |
||||
return ">", true |
||||
} |
||||
return "", false |
||||
} |
||||
|
||||
func attrEscape(out *bytes.Buffer, src []byte) { |
||||
org := 0 |
||||
for i, ch := range src { |
||||
if entity, ok := escapeSingleChar(ch); ok { |
||||
if i > org { |
||||
// copy all the normal characters since the last escape
|
||||
out.Write(src[org:i]) |
||||
} |
||||
org = i + 1 |
||||
out.WriteString(entity) |
||||
} |
||||
} |
||||
if org < len(src) { |
||||
out.Write(src[org:]) |
||||
} |
||||
} |
||||
|
||||
func entityEscapeWithSkip(out *bytes.Buffer, src []byte, skipRanges [][]int) { |
||||
end := 0 |
||||
for _, rang := range skipRanges { |
||||
attrEscape(out, src[end:rang[0]]) |
||||
out.Write(src[rang[0]:rang[1]]) |
||||
end = rang[1] |
||||
} |
||||
attrEscape(out, src[end:]) |
||||
} |
||||
|
||||
func (options *Html) GetFlags() int { |
||||
return options.flags |
||||
} |
||||
|
||||
func (options *Html) TitleBlock(out *bytes.Buffer, text []byte) { |
||||
text = bytes.TrimPrefix(text, []byte("% ")) |
||||
text = bytes.Replace(text, []byte("\n% "), []byte("\n"), -1) |
||||
out.WriteString("<h1 class=\"title\">") |
||||
out.Write(text) |
||||
out.WriteString("\n</h1>") |
||||
} |
||||
|
||||
func (options *Html) Header(out *bytes.Buffer, text func() bool, level int, id string) { |
||||
marker := out.Len() |
||||
doubleSpace(out) |
||||
|
||||
if id == "" && options.flags&HTML_TOC != 0 { |
||||
id = fmt.Sprintf("toc_%d", options.headerCount) |
||||
} |
||||
|
||||
if id != "" { |
||||
id = options.ensureUniqueHeaderID(id) |
||||
|
||||
if options.parameters.HeaderIDPrefix != "" { |
||||
id = options.parameters.HeaderIDPrefix + id |
||||
} |
||||
|
||||
if options.parameters.HeaderIDSuffix != "" { |
||||
id = id + options.parameters.HeaderIDSuffix |
||||
} |
||||
|
||||
out.WriteString(fmt.Sprintf("<h%d id=\"%s\">", level, id)) |
||||
} else { |
||||
out.WriteString(fmt.Sprintf("<h%d>", level)) |
||||
} |
||||
|
||||
tocMarker := out.Len() |
||||
if !text() { |
||||
out.Truncate(marker) |
||||
return |
||||
} |
||||
|
||||
// are we building a table of contents?
|
||||
if options.flags&HTML_TOC != 0 { |
||||
options.TocHeaderWithAnchor(out.Bytes()[tocMarker:], level, id) |
||||
} |
||||
|
||||
out.WriteString(fmt.Sprintf("</h%d>\n", level)) |
||||
} |
||||
|
||||
func (options *Html) BlockHtml(out *bytes.Buffer, text []byte) { |
||||
if options.flags&HTML_SKIP_HTML != 0 { |
||||
return |
||||
} |
||||
|
||||
doubleSpace(out) |
||||
out.Write(text) |
||||
out.WriteByte('\n') |
||||
} |
||||
|
||||
func (options *Html) HRule(out *bytes.Buffer) { |
||||
doubleSpace(out) |
||||
out.WriteString("<hr") |
||||
out.WriteString(options.closeTag) |
||||
out.WriteByte('\n') |
||||
} |
||||
|
||||
func (options *Html) BlockCode(out *bytes.Buffer, text []byte, info string) { |
||||
doubleSpace(out) |
||||
|
||||
endOfLang := strings.IndexAny(info, "\t ") |
||||
if endOfLang < 0 { |
||||
endOfLang = len(info) |
||||
} |
||||
lang := info[:endOfLang] |
||||
if len(lang) == 0 || lang == "." { |
||||
out.WriteString("<pre><code>") |
||||
} else { |
||||
out.WriteString("<pre><code class=\"language-") |
||||
attrEscape(out, []byte(lang)) |
||||
out.WriteString("\">") |
||||
} |
||||
attrEscape(out, text) |
||||
out.WriteString("</code></pre>\n") |
||||
} |
||||
|
||||
func (options *Html) BlockQuote(out *bytes.Buffer, text []byte) { |
||||
doubleSpace(out) |
||||
out.WriteString("<blockquote>\n") |
||||
out.Write(text) |
||||
out.WriteString("</blockquote>\n") |
||||
} |
||||
|
||||
func (options *Html) Table(out *bytes.Buffer, header []byte, body []byte, columnData []int) { |
||||
doubleSpace(out) |
||||
out.WriteString("<table>\n<thead>\n") |
||||
out.Write(header) |
||||
out.WriteString("</thead>\n\n<tbody>\n") |
||||
out.Write(body) |
||||
out.WriteString("</tbody>\n</table>\n") |
||||
} |
||||
|
||||
func (options *Html) TableRow(out *bytes.Buffer, text []byte) { |
||||
doubleSpace(out) |
||||
out.WriteString("<tr>\n") |
||||
out.Write(text) |
||||
out.WriteString("\n</tr>\n") |
||||
} |
||||
|
||||
func (options *Html) TableHeaderCell(out *bytes.Buffer, text []byte, align int) { |
||||
doubleSpace(out) |
||||
switch align { |
||||
case TABLE_ALIGNMENT_LEFT: |
||||
out.WriteString("<th align=\"left\">") |
||||
case TABLE_ALIGNMENT_RIGHT: |
||||
out.WriteString("<th align=\"right\">") |
||||
case TABLE_ALIGNMENT_CENTER: |
||||
out.WriteString("<th align=\"center\">") |
||||
default: |
||||
out.WriteString("<th>") |
||||
} |
||||
|
||||
out.Write(text) |
||||
out.WriteString("</th>") |
||||
} |
||||
|
||||
func (options *Html) TableCell(out *bytes.Buffer, text []byte, align int) { |
||||
doubleSpace(out) |
||||
switch align { |
||||
case TABLE_ALIGNMENT_LEFT: |
||||
out.WriteString("<td align=\"left\">") |
||||
case TABLE_ALIGNMENT_RIGHT: |
||||
out.WriteString("<td align=\"right\">") |
||||
case TABLE_ALIGNMENT_CENTER: |
||||
out.WriteString("<td align=\"center\">") |
||||
default: |
||||
out.WriteString("<td>") |
||||
} |
||||
|
||||
out.Write(text) |
||||
out.WriteString("</td>") |
||||
} |
||||
|
||||
func (options *Html) Footnotes(out *bytes.Buffer, text func() bool) { |
||||
out.WriteString("<div class=\"footnotes\">\n") |
||||
options.HRule(out) |
||||
options.List(out, text, LIST_TYPE_ORDERED) |
||||
out.WriteString("</div>\n") |
||||
} |
||||
|
||||
func (options *Html) FootnoteItem(out *bytes.Buffer, name, text []byte, flags int) { |
||||
if flags&LIST_ITEM_CONTAINS_BLOCK != 0 || flags&LIST_ITEM_BEGINNING_OF_LIST != 0 { |
||||
doubleSpace(out) |
||||
} |
||||
slug := slugify(name) |
||||
out.WriteString(`<li id="`) |
||||
out.WriteString(`fn:`) |
||||
out.WriteString(options.parameters.FootnoteAnchorPrefix) |
||||
out.Write(slug) |
||||
out.WriteString(`">`) |
||||
out.Write(text) |
||||
if options.flags&HTML_FOOTNOTE_RETURN_LINKS != 0 { |
||||
out.WriteString(` <a class="footnote-return" href="#`) |
||||
out.WriteString(`fnref:`) |
||||
out.WriteString(options.parameters.FootnoteAnchorPrefix) |
||||
out.Write(slug) |
||||
out.WriteString(`">`) |
||||
out.WriteString(options.parameters.FootnoteReturnLinkContents) |
||||
out.WriteString(`</a>`) |
||||
} |
||||
out.WriteString("</li>\n") |
||||
} |
||||
|
||||
func (options *Html) List(out *bytes.Buffer, text func() bool, flags int) { |
||||
marker := out.Len() |
||||
doubleSpace(out) |
||||
|
||||
if flags&LIST_TYPE_DEFINITION != 0 { |
||||
out.WriteString("<dl>") |
||||
} else if flags&LIST_TYPE_ORDERED != 0 { |
||||
out.WriteString("<ol>") |
||||
} else { |
||||
out.WriteString("<ul>") |
||||
} |
||||
if !text() { |
||||
out.Truncate(marker) |
||||
return |
||||
} |
||||
if flags&LIST_TYPE_DEFINITION != 0 { |
||||
out.WriteString("</dl>\n") |
||||
} else if flags&LIST_TYPE_ORDERED != 0 { |
||||
out.WriteString("</ol>\n") |
||||
} else { |
||||
out.WriteString("</ul>\n") |
||||
} |
||||
} |
||||
|
||||
func (options *Html) ListItem(out *bytes.Buffer, text []byte, flags int) { |
||||
if (flags&LIST_ITEM_CONTAINS_BLOCK != 0 && flags&LIST_TYPE_DEFINITION == 0) || |
||||
flags&LIST_ITEM_BEGINNING_OF_LIST != 0 { |
||||
doubleSpace(out) |
||||
} |
||||
if flags&LIST_TYPE_TERM != 0 { |
||||
out.WriteString("<dt>") |
||||
} else if flags&LIST_TYPE_DEFINITION != 0 { |
||||
out.WriteString("<dd>") |
||||
} else { |
||||
out.WriteString("<li>") |
||||
} |
||||
out.Write(text) |
||||
if flags&LIST_TYPE_TERM != 0 { |
||||
out.WriteString("</dt>\n") |
||||
} else if flags&LIST_TYPE_DEFINITION != 0 { |
||||
out.WriteString("</dd>\n") |
||||
} else { |
||||
out.WriteString("</li>\n") |
||||
} |
||||
} |
||||
|
||||
func (options *Html) Paragraph(out *bytes.Buffer, text func() bool) { |
||||
marker := out.Len() |
||||
doubleSpace(out) |
||||
|
||||
out.WriteString("<p>") |
||||
if !text() { |
||||
out.Truncate(marker) |
||||
return |
||||
} |
||||
out.WriteString("</p>\n") |
||||
} |
||||
|
||||
func (options *Html) AutoLink(out *bytes.Buffer, link []byte, kind int) { |
||||
skipRanges := htmlEntity.FindAllIndex(link, -1) |
||||
if options.flags&HTML_SAFELINK != 0 && !isSafeLink(link) && kind != LINK_TYPE_EMAIL { |
||||
// mark it but don't link it if it is not a safe link: no smartypants
|
||||
out.WriteString("<tt>") |
||||
entityEscapeWithSkip(out, link, skipRanges) |
||||
out.WriteString("</tt>") |
||||
return |
||||
} |
||||
|
||||
out.WriteString("<a href=\"") |
||||
if kind == LINK_TYPE_EMAIL { |
||||
out.WriteString("mailto:") |
||||
} else { |
||||
options.maybeWriteAbsolutePrefix(out, link) |
||||
} |
||||
|
||||
entityEscapeWithSkip(out, link, skipRanges) |
||||
|
||||
var relAttrs []string |
||||
if options.flags&HTML_NOFOLLOW_LINKS != 0 && !isRelativeLink(link) { |
||||
relAttrs = append(relAttrs, "nofollow") |
||||
} |
||||
if options.flags&HTML_NOREFERRER_LINKS != 0 && !isRelativeLink(link) { |
||||
relAttrs = append(relAttrs, "noreferrer") |
||||
} |
||||
if len(relAttrs) > 0 { |
||||
out.WriteString(fmt.Sprintf("\" rel=\"%s", strings.Join(relAttrs, " "))) |
||||
} |
||||
|
||||
// blank target only add to external link
|
||||
if options.flags&HTML_HREF_TARGET_BLANK != 0 && !isRelativeLink(link) { |
||||
out.WriteString("\" target=\"_blank") |
||||
} |
||||
|
||||
out.WriteString("\">") |
||||
|
||||
// Pretty print: if we get an email address as
|
||||
// an actual URI, e.g. `mailto:foo@bar.com`, we don't
|
||||
// want to print the `mailto:` prefix
|
||||
switch { |
||||
case bytes.HasPrefix(link, []byte("mailto://")): |
||||
attrEscape(out, link[len("mailto://"):]) |
||||
case bytes.HasPrefix(link, []byte("mailto:")): |
||||
attrEscape(out, link[len("mailto:"):]) |
||||
default: |
||||
entityEscapeWithSkip(out, link, skipRanges) |
||||
} |
||||
|
||||
out.WriteString("</a>") |
||||
} |
||||
|
||||
func (options *Html) CodeSpan(out *bytes.Buffer, text []byte) { |
||||
out.WriteString("<code>") |
||||
attrEscape(out, text) |
||||
out.WriteString("</code>") |
||||
} |
||||
|
||||
func (options *Html) DoubleEmphasis(out *bytes.Buffer, text []byte) { |
||||
out.WriteString("<strong>") |
||||
out.Write(text) |
||||
out.WriteString("</strong>") |
||||
} |
||||
|
||||
func (options *Html) Emphasis(out *bytes.Buffer, text []byte) { |
||||
if len(text) == 0 { |
||||
return |
||||
} |
||||
out.WriteString("<em>") |
||||
out.Write(text) |
||||
out.WriteString("</em>") |
||||
} |
||||
|
||||
func (options *Html) maybeWriteAbsolutePrefix(out *bytes.Buffer, link []byte) { |
||||
if options.parameters.AbsolutePrefix != "" && isRelativeLink(link) && link[0] != '.' { |
||||
out.WriteString(options.parameters.AbsolutePrefix) |
||||
if link[0] != '/' { |
||||
out.WriteByte('/') |
||||
} |
||||
} |
||||
} |
||||
|
||||
func (options *Html) Image(out *bytes.Buffer, link []byte, title []byte, alt []byte) { |
||||
if options.flags&HTML_SKIP_IMAGES != 0 { |
||||
return |
||||
} |
||||
|
||||
out.WriteString("<img src=\"") |
||||
options.maybeWriteAbsolutePrefix(out, link) |
||||
attrEscape(out, link) |
||||
out.WriteString("\" alt=\"") |
||||
if len(alt) > 0 { |
||||
attrEscape(out, alt) |
||||
} |
||||
if len(title) > 0 { |
||||
out.WriteString("\" title=\"") |
||||
attrEscape(out, title) |
||||
} |
||||
|
||||
out.WriteByte('"') |
||||
out.WriteString(options.closeTag) |
||||
} |
||||
|
||||
func (options *Html) LineBreak(out *bytes.Buffer) { |
||||
out.WriteString("<br") |
||||
out.WriteString(options.closeTag) |
||||
out.WriteByte('\n') |
||||
} |
||||
|
||||
func (options *Html) Link(out *bytes.Buffer, link []byte, title []byte, content []byte) { |
||||
if options.flags&HTML_SKIP_LINKS != 0 { |
||||
// write the link text out but don't link it, just mark it with typewriter font
|
||||
out.WriteString("<tt>") |
||||
attrEscape(out, content) |
||||
out.WriteString("</tt>") |
||||
return |
||||
} |
||||
|
||||
if options.flags&HTML_SAFELINK != 0 && !isSafeLink(link) { |
||||
// write the link text out but don't link it, just mark it with typewriter font
|
||||
out.WriteString("<tt>") |
||||
attrEscape(out, content) |
||||
out.WriteString("</tt>") |
||||
return |
||||
} |
||||
|
||||
out.WriteString("<a href=\"") |
||||
options.maybeWriteAbsolutePrefix(out, link) |
||||
attrEscape(out, link) |
||||
if len(title) > 0 { |
||||
out.WriteString("\" title=\"") |
||||
attrEscape(out, title) |
||||
} |
||||
var relAttrs []string |
||||
if options.flags&HTML_NOFOLLOW_LINKS != 0 && !isRelativeLink(link) { |
||||
relAttrs = append(relAttrs, "nofollow") |
||||
} |
||||
if options.flags&HTML_NOREFERRER_LINKS != 0 && !isRelativeLink(link) { |
||||
relAttrs = append(relAttrs, "noreferrer") |
||||
} |
||||
if len(relAttrs) > 0 { |
||||
out.WriteString(fmt.Sprintf("\" rel=\"%s", strings.Join(relAttrs, " "))) |
||||
} |
||||
|
||||
// blank target only add to external link
|
||||
if options.flags&HTML_HREF_TARGET_BLANK != 0 && !isRelativeLink(link) { |
||||
out.WriteString("\" target=\"_blank") |
||||
} |
||||
|
||||
out.WriteString("\">") |
||||
out.Write(content) |
||||
out.WriteString("</a>") |
||||
return |
||||
} |
||||
|
||||
func (options *Html) RawHtmlTag(out *bytes.Buffer, text []byte) { |
||||
if options.flags&HTML_SKIP_HTML != 0 { |
||||
return |
||||
} |
||||
if options.flags&HTML_SKIP_STYLE != 0 && isHtmlTag(text, "style") { |
||||
return |
||||
} |
||||
if options.flags&HTML_SKIP_LINKS != 0 && isHtmlTag(text, "a") { |
||||
return |
||||
} |
||||
if options.flags&HTML_SKIP_IMAGES != 0 && isHtmlTag(text, "img") { |
||||
return |
||||
} |
||||
out.Write(text) |
||||
} |
||||
|
||||
func (options *Html) TripleEmphasis(out *bytes.Buffer, text []byte) { |
||||
out.WriteString("<strong><em>") |
||||
out.Write(text) |
||||
out.WriteString("</em></strong>") |
||||
} |
||||
|
||||
func (options *Html) StrikeThrough(out *bytes.Buffer, text []byte) { |
||||
out.WriteString("<del>") |
||||
out.Write(text) |
||||
out.WriteString("</del>") |
||||
} |
||||
|
||||
func (options *Html) FootnoteRef(out *bytes.Buffer, ref []byte, id int) { |
||||
slug := slugify(ref) |
||||
out.WriteString(`<sup class="footnote-ref" id="`) |
||||
out.WriteString(`fnref:`) |
||||
out.WriteString(options.parameters.FootnoteAnchorPrefix) |
||||
out.Write(slug) |
||||
out.WriteString(`"><a href="#`) |
||||
out.WriteString(`fn:`) |
||||
out.WriteString(options.parameters.FootnoteAnchorPrefix) |
||||
out.Write(slug) |
||||
out.WriteString(`">`) |
||||
out.WriteString(strconv.Itoa(id)) |
||||
out.WriteString(`</a></sup>`) |
||||
} |
||||
|
||||
func (options *Html) Entity(out *bytes.Buffer, entity []byte) { |
||||
out.Write(entity) |
||||
} |
||||
|
||||
func (options *Html) NormalText(out *bytes.Buffer, text []byte) { |
||||
if options.flags&HTML_USE_SMARTYPANTS != 0 { |
||||
options.Smartypants(out, text) |
||||
} else { |
||||
attrEscape(out, text) |
||||
} |
||||
} |
||||
|
||||
func (options *Html) Smartypants(out *bytes.Buffer, text []byte) { |
||||
smrt := smartypantsData{false, false} |
||||
|
||||
// first do normal entity escaping
|
||||
var escaped bytes.Buffer |
||||
attrEscape(&escaped, text) |
||||
text = escaped.Bytes() |
||||
|
||||
mark := 0 |
||||
for i := 0; i < len(text); i++ { |
||||
if action := options.smartypants[text[i]]; action != nil { |
||||
if i > mark { |
||||
out.Write(text[mark:i]) |
||||
} |
||||
|
||||
previousChar := byte(0) |
||||
if i > 0 { |
||||
previousChar = text[i-1] |
||||
} |
||||
i += action(out, &smrt, previousChar, text[i:]) |
||||
mark = i + 1 |
||||
} |
||||
} |
||||
|
||||
if mark < len(text) { |
||||
out.Write(text[mark:]) |
||||
} |
||||
} |
||||
|
||||
func (options *Html) DocumentHeader(out *bytes.Buffer) { |
||||
if options.flags&HTML_COMPLETE_PAGE == 0 { |
||||
return |
||||
} |
||||
|
||||
ending := "" |
||||
if options.flags&HTML_USE_XHTML != 0 { |
||||
out.WriteString("<!DOCTYPE html PUBLIC \"-//W3C//DTD XHTML 1.0 Transitional//EN\" ") |
||||
out.WriteString("\"http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd\">\n") |
||||
out.WriteString("<html xmlns=\"http://www.w3.org/1999/xhtml\">\n") |
||||
ending = " /" |
||||
} else { |
||||
out.WriteString("<!DOCTYPE html>\n") |
||||
out.WriteString("<html>\n") |
||||
} |
||||
out.WriteString("<head>\n") |
||||
out.WriteString(" <title>") |
||||
options.NormalText(out, []byte(options.title)) |
||||
out.WriteString("</title>\n") |
||||
out.WriteString(" <meta name=\"GENERATOR\" content=\"Blackfriday Markdown Processor v") |
||||
out.WriteString(VERSION) |
||||
out.WriteString("\"") |
||||
out.WriteString(ending) |
||||
out.WriteString(">\n") |
||||
out.WriteString(" <meta charset=\"utf-8\"") |
||||
out.WriteString(ending) |
||||
out.WriteString(">\n") |
||||
if options.css != "" { |
||||
out.WriteString(" <link rel=\"stylesheet\" type=\"text/css\" href=\"") |
||||
attrEscape(out, []byte(options.css)) |
||||
out.WriteString("\"") |
||||
out.WriteString(ending) |
||||
out.WriteString(">\n") |
||||
} |
||||
out.WriteString("</head>\n") |
||||
out.WriteString("<body>\n") |
||||
|
||||
options.tocMarker = out.Len() |
||||
} |
||||
|
||||
func (options *Html) DocumentFooter(out *bytes.Buffer) { |
||||
// finalize and insert the table of contents
|
||||
if options.flags&HTML_TOC != 0 { |
||||
options.TocFinalize() |
||||
|
||||
// now we have to insert the table of contents into the document
|
||||
var temp bytes.Buffer |
||||
|
||||
// start by making a copy of everything after the document header
|
||||
temp.Write(out.Bytes()[options.tocMarker:]) |
||||
|
||||
// now clear the copied material from the main output buffer
|
||||
out.Truncate(options.tocMarker) |
||||
|
||||
// corner case spacing issue
|
||||
if options.flags&HTML_COMPLETE_PAGE != 0 { |
||||
out.WriteByte('\n') |
||||
} |
||||
|
||||
// insert the table of contents
|
||||
out.WriteString("<nav>\n") |
||||
out.Write(options.toc.Bytes()) |
||||
out.WriteString("</nav>\n") |
||||
|
||||
// corner case spacing issue
|
||||
if options.flags&HTML_COMPLETE_PAGE == 0 && options.flags&HTML_OMIT_CONTENTS == 0 { |
||||
out.WriteByte('\n') |
||||
} |
||||
|
||||
// write out everything that came after it
|
||||
if options.flags&HTML_OMIT_CONTENTS == 0 { |
||||
out.Write(temp.Bytes()) |
||||
} |
||||
} |
||||
|
||||
if options.flags&HTML_COMPLETE_PAGE != 0 { |
||||
out.WriteString("\n</body>\n") |
||||
out.WriteString("</html>\n") |
||||
} |
||||
|
||||
} |
||||
|
||||
func (options *Html) TocHeaderWithAnchor(text []byte, level int, anchor string) { |
||||
for level > options.currentLevel { |
||||
switch { |
||||
case bytes.HasSuffix(options.toc.Bytes(), []byte("</li>\n")): |
||||
// this sublist can nest underneath a header
|
||||
size := options.toc.Len() |
||||
options.toc.Truncate(size - len("</li>\n")) |
||||
|
||||
case options.currentLevel > 0: |
||||
options.toc.WriteString("<li>") |
||||
} |
||||
if options.toc.Len() > 0 { |
||||
options.toc.WriteByte('\n') |
||||
} |
||||
options.toc.WriteString("<ul>\n") |
||||
options.currentLevel++ |
||||
} |
||||
|
||||
for level < options.currentLevel { |
||||
options.toc.WriteString("</ul>") |
||||
if options.currentLevel > 1 { |
||||
options.toc.WriteString("</li>\n") |
||||
} |
||||
options.currentLevel-- |
||||
} |
||||
|
||||
options.toc.WriteString("<li><a href=\"#") |
||||
if anchor != "" { |
||||
options.toc.WriteString(anchor) |
||||
} else { |
||||
options.toc.WriteString("toc_") |
||||
options.toc.WriteString(strconv.Itoa(options.headerCount)) |
||||
} |
||||
options.toc.WriteString("\">") |
||||
options.headerCount++ |
||||
|
||||
options.toc.Write(text) |
||||
|
||||
options.toc.WriteString("</a></li>\n") |
||||
} |
||||
|
||||
func (options *Html) TocHeader(text []byte, level int) { |
||||
options.TocHeaderWithAnchor(text, level, "") |
||||
} |
||||
|
||||
func (options *Html) TocFinalize() { |
||||
for options.currentLevel > 1 { |
||||
options.toc.WriteString("</ul></li>\n") |
||||
options.currentLevel-- |
||||
} |
||||
|
||||
if options.currentLevel > 0 { |
||||
options.toc.WriteString("</ul>\n") |
||||
} |
||||
} |
||||
|
||||
func isHtmlTag(tag []byte, tagname string) bool { |
||||
found, _ := findHtmlTagPos(tag, tagname) |
||||
return found |
||||
} |
||||
|
||||
// Look for a character, but ignore it when it's in any kind of quotes, it
|
||||
// might be JavaScript
|
||||
func skipUntilCharIgnoreQuotes(html []byte, start int, char byte) int { |
||||
inSingleQuote := false |
||||
inDoubleQuote := false |
||||
inGraveQuote := false |
||||
i := start |
||||
for i < len(html) { |
||||
switch { |
||||
case html[i] == char && !inSingleQuote && !inDoubleQuote && !inGraveQuote: |
||||
return i |
||||
case html[i] == '\'': |
||||
inSingleQuote = !inSingleQuote |
||||
case html[i] == '"': |
||||
inDoubleQuote = !inDoubleQuote |
||||
case html[i] == '`': |
||||
inGraveQuote = !inGraveQuote |
||||
} |
||||
i++ |
||||
} |
||||
return start |
||||
} |
||||
|
||||
func findHtmlTagPos(tag []byte, tagname string) (bool, int) { |
||||
i := 0 |
||||
if i < len(tag) && tag[0] != '<' { |
||||
return false, -1 |
||||
} |
||||
i++ |
||||
i = skipSpace(tag, i) |
||||
|
||||
if i < len(tag) && tag[i] == '/' { |
||||
i++ |
||||
} |
||||
|
||||
i = skipSpace(tag, i) |
||||
j := 0 |
||||
for ; i < len(tag); i, j = i+1, j+1 { |
||||
if j >= len(tagname) { |
||||
break |
||||
} |
||||
|
||||
if strings.ToLower(string(tag[i]))[0] != tagname[j] { |
||||
return false, -1 |
||||
} |
||||
} |
||||
|
||||
if i == len(tag) { |
||||
return false, -1 |
||||
} |
||||
|
||||
rightAngle := skipUntilCharIgnoreQuotes(tag, i, '>') |
||||
if rightAngle > i { |
||||
return true, rightAngle |
||||
} |
||||
|
||||
return false, -1 |
||||
} |
||||
|
||||
func skipUntilChar(text []byte, start int, char byte) int { |
||||
i := start |
||||
for i < len(text) && text[i] != char { |
||||
i++ |
||||
} |
||||
return i |
||||
} |
||||
|
||||
func skipSpace(tag []byte, i int) int { |
||||
for i < len(tag) && isspace(tag[i]) { |
||||
i++ |
||||
} |
||||
return i |
||||
} |
||||
|
||||
func skipChar(data []byte, start int, char byte) int { |
||||
i := start |
||||
for i < len(data) && data[i] == char { |
||||
i++ |
||||
} |
||||
return i |
||||
} |
||||
|
||||
func doubleSpace(out *bytes.Buffer) { |
||||
if out.Len() > 0 { |
||||
out.WriteByte('\n') |
||||
} |
||||
} |
||||
|
||||
func isRelativeLink(link []byte) (yes bool) { |
||||
// a tag begin with '#'
|
||||
if link[0] == '#' { |
||||
return true |
||||
} |
||||
|
||||
// link begin with '/' but not '//', the second maybe a protocol relative link
|
||||
if len(link) >= 2 && link[0] == '/' && link[1] != '/' { |
||||
return true |
||||
} |
||||
|
||||
// only the root '/'
|
||||
if len(link) == 1 && link[0] == '/' { |
||||
return true |
||||
} |
||||
|
||||
// current directory : begin with "./"
|
||||
if bytes.HasPrefix(link, []byte("./")) { |
||||
return true |
||||
} |
||||
|
||||
// parent directory : begin with "../"
|
||||
if bytes.HasPrefix(link, []byte("../")) { |
||||
return true |
||||
} |
||||
|
||||
return false |
||||
} |
||||
|
||||
func (options *Html) ensureUniqueHeaderID(id string) string { |
||||
for count, found := options.headerIDs[id]; found; count, found = options.headerIDs[id] { |
||||
tmp := fmt.Sprintf("%s-%d", id, count+1) |
||||
|
||||
if _, tmpFound := options.headerIDs[tmp]; !tmpFound { |
||||
options.headerIDs[id] = count + 1 |
||||
id = tmp |
||||
} else { |
||||
id = id + "-1" |
||||
} |
||||
} |
||||
|
||||
if _, found := options.headerIDs[id]; !found { |
||||
options.headerIDs[id] = 0 |
||||
} |
||||
|
||||
return id |
||||
} |
@ -1,334 +0,0 @@ |
||||
//
|
||||
// Blackfriday Markdown Processor
|
||||
// Available at http://github.com/russross/blackfriday
|
||||
//
|
||||
// Copyright © 2011 Russ Ross <russ@russross.com>.
|
||||
// Distributed under the Simplified BSD License.
|
||||
// See README.md for details.
|
||||
//
|
||||
|
||||
//
|
||||
//
|
||||
// LaTeX rendering backend
|
||||
//
|
||||
//
|
||||
|
||||
package blackfriday |
||||
|
||||
import ( |
||||
"bytes" |
||||
"strings" |
||||
) |
||||
|
||||
// Latex is a type that implements the Renderer interface for LaTeX output.
|
||||
//
|
||||
// Do not create this directly, instead use the LatexRenderer function.
|
||||
type Latex struct { |
||||
} |
||||
|
||||
// LatexRenderer creates and configures a Latex object, which
|
||||
// satisfies the Renderer interface.
|
||||
//
|
||||
// flags is a set of LATEX_* options ORed together (currently no such options
|
||||
// are defined).
|
||||
func LatexRenderer(flags int) Renderer { |
||||
return &Latex{} |
||||
} |
||||
|
||||
func (options *Latex) GetFlags() int { |
||||
return 0 |
||||
} |
||||
|
||||
// render code chunks using verbatim, or listings if we have a language
|
||||
func (options *Latex) BlockCode(out *bytes.Buffer, text []byte, info string) { |
||||
if info == "" { |
||||
out.WriteString("\n\\begin{verbatim}\n") |
||||
} else { |
||||
lang := strings.Fields(info)[0] |
||||
out.WriteString("\n\\begin{lstlisting}[language=") |
||||
out.WriteString(lang) |
||||
out.WriteString("]\n") |
||||
} |
||||
out.Write(text) |
||||
if info == "" { |
||||
out.WriteString("\n\\end{verbatim}\n") |
||||
} else { |
||||
out.WriteString("\n\\end{lstlisting}\n") |
||||
} |
||||
} |
||||
|
||||
func (options *Latex) TitleBlock(out *bytes.Buffer, text []byte) { |
||||
|
||||
} |
||||
|
||||
func (options *Latex) BlockQuote(out *bytes.Buffer, text []byte) { |
||||
out.WriteString("\n\\begin{quotation}\n") |
||||
out.Write(text) |
||||
out.WriteString("\n\\end{quotation}\n") |
||||
} |
||||
|
||||
func (options *Latex) BlockHtml(out *bytes.Buffer, text []byte) { |
||||
// a pretty lame thing to do...
|
||||
out.WriteString("\n\\begin{verbatim}\n") |
||||
out.Write(text) |
||||
out.WriteString("\n\\end{verbatim}\n") |
||||
} |
||||
|
||||
func (options *Latex) Header(out *bytes.Buffer, text func() bool, level int, id string) { |
||||
marker := out.Len() |
||||
|
||||
switch level { |
||||
case 1: |
||||
out.WriteString("\n\\section{") |
||||
case 2: |
||||
out.WriteString("\n\\subsection{") |
||||
case 3: |
||||
out.WriteString("\n\\subsubsection{") |
||||
case 4: |
||||
out.WriteString("\n\\paragraph{") |
||||
case 5: |
||||
out.WriteString("\n\\subparagraph{") |
||||
case 6: |
||||
out.WriteString("\n\\textbf{") |
||||
} |
||||
if !text() { |
||||
out.Truncate(marker) |
||||
return |
||||
} |
||||
out.WriteString("}\n") |
||||
} |
||||
|
||||
func (options *Latex) HRule(out *bytes.Buffer) { |
||||
out.WriteString("\n\\HRule\n") |
||||
} |
||||
|
||||
func (options *Latex) List(out *bytes.Buffer, text func() bool, flags int) { |
||||
marker := out.Len() |
||||
if flags&LIST_TYPE_ORDERED != 0 { |
||||
out.WriteString("\n\\begin{enumerate}\n") |
||||
} else { |
||||
out.WriteString("\n\\begin{itemize}\n") |
||||
} |
||||
if !text() { |
||||
out.Truncate(marker) |
||||
return |
||||
} |
||||
if flags&LIST_TYPE_ORDERED != 0 { |
||||
out.WriteString("\n\\end{enumerate}\n") |
||||
} else { |
||||
out.WriteString("\n\\end{itemize}\n") |
||||
} |
||||
} |
||||
|
||||
func (options *Latex) ListItem(out *bytes.Buffer, text []byte, flags int) { |
||||
out.WriteString("\n\\item ") |
||||
out.Write(text) |
||||
} |
||||
|
||||
func (options *Latex) Paragraph(out *bytes.Buffer, text func() bool) { |
||||
marker := out.Len() |
||||
out.WriteString("\n") |
||||
if !text() { |
||||
out.Truncate(marker) |
||||
return |
||||
} |
||||
out.WriteString("\n") |
||||
} |
||||
|
||||
func (options *Latex) Table(out *bytes.Buffer, header []byte, body []byte, columnData []int) { |
||||
out.WriteString("\n\\begin{tabular}{") |
||||
for _, elt := range columnData { |
||||
switch elt { |
||||
case TABLE_ALIGNMENT_LEFT: |
||||
out.WriteByte('l') |
||||
case TABLE_ALIGNMENT_RIGHT: |
||||
out.WriteByte('r') |
||||
default: |
||||
out.WriteByte('c') |
||||
} |
||||
} |
||||
out.WriteString("}\n") |
||||
out.Write(header) |
||||
out.WriteString(" \\\\\n\\hline\n") |
||||
out.Write(body) |
||||
out.WriteString("\n\\end{tabular}\n") |
||||
} |
||||
|
||||
func (options *Latex) TableRow(out *bytes.Buffer, text []byte) { |
||||
if out.Len() > 0 { |
||||
out.WriteString(" \\\\\n") |
||||
} |
||||
out.Write(text) |
||||
} |
||||
|
||||
func (options *Latex) TableHeaderCell(out *bytes.Buffer, text []byte, align int) { |
||||
if out.Len() > 0 { |
||||
out.WriteString(" & ") |
||||
} |
||||
out.Write(text) |
||||
} |
||||
|
||||
func (options *Latex) TableCell(out *bytes.Buffer, text []byte, align int) { |
||||
if out.Len() > 0 { |
||||
out.WriteString(" & ") |
||||
} |
||||
out.Write(text) |
||||
} |
||||
|
||||
// TODO: this
|
||||
func (options *Latex) Footnotes(out *bytes.Buffer, text func() bool) { |
||||
|
||||
} |
||||
|
||||
func (options *Latex) FootnoteItem(out *bytes.Buffer, name, text []byte, flags int) { |
||||
|
||||
} |
||||
|
||||
func (options *Latex) AutoLink(out *bytes.Buffer, link []byte, kind int) { |
||||
out.WriteString("\\href{") |
||||
if kind == LINK_TYPE_EMAIL { |
||||
out.WriteString("mailto:") |
||||
} |
||||
out.Write(link) |
||||
out.WriteString("}{") |
||||
out.Write(link) |
||||
out.WriteString("}") |
||||
} |
||||
|
||||
func (options *Latex) CodeSpan(out *bytes.Buffer, text []byte) { |
||||
out.WriteString("\\texttt{") |
||||
escapeSpecialChars(out, text) |
||||
out.WriteString("}") |
||||
} |
||||
|
||||
func (options *Latex) DoubleEmphasis(out *bytes.Buffer, text []byte) { |
||||
out.WriteString("\\textbf{") |
||||
out.Write(text) |
||||
out.WriteString("}") |
||||
} |
||||
|
||||
func (options *Latex) Emphasis(out *bytes.Buffer, text []byte) { |
||||
out.WriteString("\\textit{") |
||||
out.Write(text) |
||||
out.WriteString("}") |
||||
} |
||||
|
||||
func (options *Latex) Image(out *bytes.Buffer, link []byte, title []byte, alt []byte) { |
||||
if bytes.HasPrefix(link, []byte("http://")) || bytes.HasPrefix(link, []byte("https://")) { |
||||
// treat it like a link
|
||||
out.WriteString("\\href{") |
||||
out.Write(link) |
||||
out.WriteString("}{") |
||||
out.Write(alt) |
||||
out.WriteString("}") |
||||
} else { |
||||
out.WriteString("\\includegraphics{") |
||||
out.Write(link) |
||||
out.WriteString("}") |
||||
} |
||||
} |
||||
|
||||
func (options *Latex) LineBreak(out *bytes.Buffer) { |
||||
out.WriteString(" \\\\\n") |
||||
} |
||||
|
||||
func (options *Latex) Link(out *bytes.Buffer, link []byte, title []byte, content []byte) { |
||||
out.WriteString("\\href{") |
||||
out.Write(link) |
||||
out.WriteString("}{") |
||||
out.Write(content) |
||||
out.WriteString("}") |
||||
} |
||||
|
||||
func (options *Latex) RawHtmlTag(out *bytes.Buffer, tag []byte) { |
||||
} |
||||
|
||||
func (options *Latex) TripleEmphasis(out *bytes.Buffer, text []byte) { |
||||
out.WriteString("\\textbf{\\textit{") |
||||
out.Write(text) |
||||
out.WriteString("}}") |
||||
} |
||||
|
||||
func (options *Latex) StrikeThrough(out *bytes.Buffer, text []byte) { |
||||
out.WriteString("\\sout{") |
||||
out.Write(text) |
||||
out.WriteString("}") |
||||
} |
||||
|
||||
// TODO: this
|
||||
func (options *Latex) FootnoteRef(out *bytes.Buffer, ref []byte, id int) { |
||||
|
||||
} |
||||
|
||||
func needsBackslash(c byte) bool { |
||||
for _, r := range []byte("_{}%$&\\~#") { |
||||
if c == r { |
||||
return true |
||||
} |
||||
} |
||||
return false |
||||
} |
||||
|
||||
func escapeSpecialChars(out *bytes.Buffer, text []byte) { |
||||
for i := 0; i < len(text); i++ { |
||||
// directly copy normal characters
|
||||
org := i |
||||
|
||||
for i < len(text) && !needsBackslash(text[i]) { |
||||
i++ |
||||
} |
||||
if i > org { |
||||
out.Write(text[org:i]) |
||||
} |
||||
|
||||
// escape a character
|
||||
if i >= len(text) { |
||||
break |
||||
} |
||||
out.WriteByte('\\') |
||||
out.WriteByte(text[i]) |
||||
} |
||||
} |
||||
|
||||
func (options *Latex) Entity(out *bytes.Buffer, entity []byte) { |
||||
// TODO: convert this into a unicode character or something
|
||||
out.Write(entity) |
||||
} |
||||
|
||||
func (options *Latex) NormalText(out *bytes.Buffer, text []byte) { |
||||
escapeSpecialChars(out, text) |
||||
} |
||||
|
||||
// header and footer
|
||||
func (options *Latex) DocumentHeader(out *bytes.Buffer) { |
||||
out.WriteString("\\documentclass{article}\n") |
||||
out.WriteString("\n") |
||||
out.WriteString("\\usepackage{graphicx}\n") |
||||
out.WriteString("\\usepackage{listings}\n") |
||||
out.WriteString("\\usepackage[margin=1in]{geometry}\n") |
||||
out.WriteString("\\usepackage[utf8]{inputenc}\n") |
||||
out.WriteString("\\usepackage{verbatim}\n") |
||||
out.WriteString("\\usepackage[normalem]{ulem}\n") |
||||
out.WriteString("\\usepackage{hyperref}\n") |
||||
out.WriteString("\n") |
||||
out.WriteString("\\hypersetup{colorlinks,%\n") |
||||
out.WriteString(" citecolor=black,%\n") |
||||
out.WriteString(" filecolor=black,%\n") |
||||
out.WriteString(" linkcolor=black,%\n") |
||||
out.WriteString(" urlcolor=black,%\n") |
||||
out.WriteString(" pdfstartview=FitH,%\n") |
||||
out.WriteString(" breaklinks=true,%\n") |
||||
out.WriteString(" pdfauthor={Blackfriday Markdown Processor v") |
||||
out.WriteString(VERSION) |
||||
out.WriteString("}}\n") |
||||
out.WriteString("\n") |
||||
out.WriteString("\\newcommand{\\HRule}{\\rule{\\linewidth}{0.5mm}}\n") |
||||
out.WriteString("\\addtolength{\\parskip}{0.5\\baselineskip}\n") |
||||
out.WriteString("\\parindent=0pt\n") |
||||
out.WriteString("\n") |
||||
out.WriteString("\\begin{document}\n") |
||||
} |
||||
|
||||
func (options *Latex) DocumentFooter(out *bytes.Buffer) { |
||||
out.WriteString("\n\\end{document}\n") |
||||
} |
@ -1,30 +1,17 @@ |
||||
sudo: false |
||||
language: go |
||||
go: |
||||
- 1.5.4 |
||||
- 1.6.2 |
||||
- "1.10.x" |
||||
- "1.11.x" |
||||
- tip |
||||
matrix: |
||||
include: |
||||
- go: 1.2.2 |
||||
script: |
||||
- go get -t -v ./... |
||||
- go test -v -race ./... |
||||
- go: 1.3.3 |
||||
script: |
||||
- go get -t -v ./... |
||||
- go test -v -race ./... |
||||
- go: 1.4.3 |
||||
script: |
||||
- go get -t -v ./... |
||||
- go test -v -race ./... |
||||
fast_finish: true |
||||
allow_failures: |
||||
- go: tip |
||||
fast_finish: true |
||||
install: |
||||
- # Do nothing. This is needed to prevent default install action "go get -t -v ./..." from happening here (we want it to happen inside script step). |
||||
script: |
||||
- go get -t -v ./... |
||||
- diff -u <(echo -n) <(gofmt -d -s .) |
||||
- go tool vet . |
||||
- go test -v -race ./... |
||||
- go test -v ./... |
File diff suppressed because it is too large
Load Diff
@ -0,0 +1,18 @@ |
||||
// Package blackfriday is a markdown processor.
|
||||
//
|
||||
// It translates plain text with simple formatting rules into an AST, which can
|
||||
// then be further processed to HTML (provided by Blackfriday itself) or other
|
||||
// formats (provided by the community).
|
||||
//
|
||||
// The simplest way to invoke Blackfriday is to call the Run function. It will
|
||||
// take a text input and produce a text output in HTML (or other format).
|
||||
//
|
||||
// A slightly more sophisticated way to use Blackfriday is to create a Markdown
|
||||
// processor and to call Parse, which returns a syntax tree for the input
|
||||
// document. You can leverage Blackfriday's parsing for content extraction from
|
||||
// markdown documents. You can assign a custom renderer and set various options
|
||||
// to the Markdown processor.
|
||||
//
|
||||
// If you're interested in calling Blackfriday from command line, see
|
||||
// https://github.com/russross/blackfriday-tool.
|
||||
package blackfriday |
@ -0,0 +1,34 @@ |
||||
package blackfriday |
||||
|
||||
import ( |
||||
"html" |
||||
"io" |
||||
) |
||||
|
||||
var htmlEscaper = [256][]byte{ |
||||
'&': []byte("&"), |
||||
'<': []byte("<"), |
||||
'>': []byte(">"), |
||||
'"': []byte("""), |
||||
} |
||||
|
||||
func escapeHTML(w io.Writer, s []byte) { |
||||
var start, end int |
||||
for end < len(s) { |
||||
escSeq := htmlEscaper[s[end]] |
||||
if escSeq != nil { |
||||
w.Write(s[start:end]) |
||||
w.Write(escSeq) |
||||
start = end + 1 |
||||
} |
||||
end++ |
||||
} |
||||
if start < len(s) && end <= len(s) { |
||||
w.Write(s[start:end]) |
||||
} |
||||
} |
||||
|
||||
func escLink(w io.Writer, text []byte) { |
||||
unesc := html.UnescapeString(string(text)) |
||||
escapeHTML(w, []byte(unesc)) |
||||
} |
@ -0,0 +1 @@ |
||||
module github.com/russross/blackfriday/v2 |
@ -0,0 +1,949 @@ |
||||
//
|
||||
// Blackfriday Markdown Processor
|
||||
// Available at http://github.com/russross/blackfriday
|
||||
//
|
||||
// Copyright © 2011 Russ Ross <russ@russross.com>.
|
||||
// Distributed under the Simplified BSD License.
|
||||
// See README.md for details.
|
||||
//
|
||||
|
||||
//
|
||||
//
|
||||
// HTML rendering backend
|
||||
//
|
||||
//
|
||||
|
||||
package blackfriday |
||||
|
||||
import ( |
||||
"bytes" |
||||
"fmt" |
||||
"io" |
||||
"regexp" |
||||
"strings" |
||||
) |
||||
|
||||
// HTMLFlags control optional behavior of HTML renderer.
|
||||
type HTMLFlags int |
||||
|
||||
// HTML renderer configuration options.
|
||||
const ( |
||||
HTMLFlagsNone HTMLFlags = 0 |
||||
SkipHTML HTMLFlags = 1 << iota // Skip preformatted HTML blocks
|
||||
SkipImages // Skip embedded images
|
||||
SkipLinks // Skip all links
|
||||
Safelink // Only link to trusted protocols
|
||||
NofollowLinks // Only link with rel="nofollow"
|
||||
NoreferrerLinks // Only link with rel="noreferrer"
|
||||
NoopenerLinks // Only link with rel="noopener"
|
||||
HrefTargetBlank // Add a blank target
|
||||
CompletePage // Generate a complete HTML page
|
||||
UseXHTML // Generate XHTML output instead of HTML
|
||||
FootnoteReturnLinks // Generate a link at the end of a footnote to return to the source
|
||||
Smartypants // Enable smart punctuation substitutions
|
||||
SmartypantsFractions // Enable smart fractions (with Smartypants)
|
||||
SmartypantsDashes // Enable smart dashes (with Smartypants)
|
||||
SmartypantsLatexDashes // Enable LaTeX-style dashes (with Smartypants)
|
||||
SmartypantsAngledQuotes // Enable angled double quotes (with Smartypants) for double quotes rendering
|
||||
SmartypantsQuotesNBSP // Enable « French guillemets » (with Smartypants)
|
||||
TOC // Generate a table of contents
|
||||
) |
||||
|
||||
var ( |
||||
htmlTagRe = regexp.MustCompile("(?i)^" + htmlTag) |
||||
) |
||||
|
||||
const ( |
||||
htmlTag = "(?:" + openTag + "|" + closeTag + "|" + htmlComment + "|" + |
||||
processingInstruction + "|" + declaration + "|" + cdata + ")" |
||||
closeTag = "</" + tagName + "\\s*[>]" |
||||
openTag = "<" + tagName + attribute + "*" + "\\s*/?>" |
||||
attribute = "(?:" + "\\s+" + attributeName + attributeValueSpec + "?)" |
||||
attributeValue = "(?:" + unquotedValue + "|" + singleQuotedValue + "|" + doubleQuotedValue + ")" |
||||
attributeValueSpec = "(?:" + "\\s*=" + "\\s*" + attributeValue + ")" |
||||
attributeName = "[a-zA-Z_:][a-zA-Z0-9:._-]*" |
||||
cdata = "<!\\[CDATA\\[[\\s\\S]*?\\]\\]>" |
||||
declaration = "<![A-Z]+" + "\\s+[^>]*>" |
||||
doubleQuotedValue = "\"[^\"]*\"" |
||||
htmlComment = "<!---->|<!--(?:-?[^>-])(?:-?[^-])*-->" |
||||
processingInstruction = "[<][?].*?[?][>]" |
||||
singleQuotedValue = "'[^']*'" |
||||
tagName = "[A-Za-z][A-Za-z0-9-]*" |
||||
unquotedValue = "[^\"'=<>`\\x00-\\x20]+" |
||||
) |
||||
|
||||
// HTMLRendererParameters is a collection of supplementary parameters tweaking
|
||||
// the behavior of various parts of HTML renderer.
|
||||
type HTMLRendererParameters struct { |
||||
// Prepend this text to each relative URL.
|
||||
AbsolutePrefix string |
||||
// Add this text to each footnote anchor, to ensure uniqueness.
|
||||
FootnoteAnchorPrefix string |
||||
// Show this text inside the <a> tag for a footnote return link, if the
|
||||
// HTML_FOOTNOTE_RETURN_LINKS flag is enabled. If blank, the string
|
||||
// <sup>[return]</sup> is used.
|
||||
FootnoteReturnLinkContents string |
||||
// If set, add this text to the front of each Heading ID, to ensure
|
||||
// uniqueness.
|
||||
HeadingIDPrefix string |
||||
// If set, add this text to the back of each Heading ID, to ensure uniqueness.
|
||||
HeadingIDSuffix string |
||||
// Increase heading levels: if the offset is 1, <h1> becomes <h2> etc.
|
||||
// Negative offset is also valid.
|
||||
// Resulting levels are clipped between 1 and 6.
|
||||
HeadingLevelOffset int |
||||
|
||||
Title string // Document title (used if CompletePage is set)
|
||||
CSS string // Optional CSS file URL (used if CompletePage is set)
|
||||
Icon string // Optional icon file URL (used if CompletePage is set)
|
||||
|
||||
Flags HTMLFlags // Flags allow customizing this renderer's behavior
|
||||
} |
||||
|
||||
// HTMLRenderer is a type that implements the Renderer interface for HTML output.
|
||||
//
|
||||
// Do not create this directly, instead use the NewHTMLRenderer function.
|
||||
type HTMLRenderer struct { |
||||
HTMLRendererParameters |
||||
|
||||
closeTag string // how to end singleton tags: either " />" or ">"
|
||||
|
||||
// Track heading IDs to prevent ID collision in a single generation.
|
||||
headingIDs map[string]int |
||||
|
||||
lastOutputLen int |
||||
disableTags int |
||||
|
||||
sr *SPRenderer |
||||
} |
||||
|
||||
const ( |
||||
xhtmlClose = " />" |
||||
htmlClose = ">" |
||||
) |
||||
|
||||
// NewHTMLRenderer creates and configures an HTMLRenderer object, which
|
||||
// satisfies the Renderer interface.
|
||||
func NewHTMLRenderer(params HTMLRendererParameters) *HTMLRenderer { |
||||
// configure the rendering engine
|
||||
closeTag := htmlClose |
||||
if params.Flags&UseXHTML != 0 { |
||||
closeTag = xhtmlClose |
||||
} |
||||
|
||||
if params.FootnoteReturnLinkContents == "" { |
||||
params.FootnoteReturnLinkContents = `<sup>[return]</sup>` |
||||
} |
||||
|
||||
return &HTMLRenderer{ |
||||
HTMLRendererParameters: params, |
||||
|
||||
closeTag: closeTag, |
||||
headingIDs: make(map[string]int), |
||||
|
||||
sr: NewSmartypantsRenderer(params.Flags), |
||||
} |
||||
} |
||||
|
||||
func isHTMLTag(tag []byte, tagname string) bool { |
||||
found, _ := findHTMLTagPos(tag, tagname) |
||||
return found |
||||
} |
||||
|
||||
// Look for a character, but ignore it when it's in any kind of quotes, it
|
||||
// might be JavaScript
|
||||
func skipUntilCharIgnoreQuotes(html []byte, start int, char byte) int { |
||||
inSingleQuote := false |
||||
inDoubleQuote := false |
||||
inGraveQuote := false |
||||
i := start |
||||
for i < len(html) { |
||||
switch { |
||||
case html[i] == char && !inSingleQuote && !inDoubleQuote && !inGraveQuote: |
||||
return i |
||||
case html[i] == '\'': |
||||
inSingleQuote = !inSingleQuote |
||||
case html[i] == '"': |
||||
inDoubleQuote = !inDoubleQuote |
||||
case html[i] == '`': |
||||
inGraveQuote = !inGraveQuote |
||||
} |
||||
i++ |
||||
} |
||||
return start |
||||
} |
||||
|
||||
func findHTMLTagPos(tag []byte, tagname string) (bool, int) { |
||||
i := 0 |
||||
if i < len(tag) && tag[0] != '<' { |
||||
return false, -1 |
||||
} |
||||
i++ |
||||
i = skipSpace(tag, i) |
||||
|
||||
if i < len(tag) && tag[i] == '/' { |
||||
i++ |
||||
} |
||||
|
||||
i = skipSpace(tag, i) |
||||
j := 0 |
||||
for ; i < len(tag); i, j = i+1, j+1 { |
||||
if j >= len(tagname) { |
||||
break |
||||
} |
||||
|
||||
if strings.ToLower(string(tag[i]))[0] != tagname[j] { |
||||
return false, -1 |
||||
} |
||||
} |
||||
|
||||
if i == len(tag) { |
||||
return false, -1 |
||||
} |
||||
|
||||
rightAngle := skipUntilCharIgnoreQuotes(tag, i, '>') |
||||
if rightAngle >= i { |
||||
return true, rightAngle |
||||
} |
||||
|
||||
return false, -1 |
||||
} |
||||
|
||||
func skipSpace(tag []byte, i int) int { |
||||
for i < len(tag) && isspace(tag[i]) { |
||||
i++ |
||||
} |
||||
return i |
||||
} |
||||
|
||||
func isRelativeLink(link []byte) (yes bool) { |
||||
// a tag begin with '#'
|
||||
if link[0] == '#' { |
||||
return true |
||||
} |
||||
|
||||
// link begin with '/' but not '//', the second maybe a protocol relative link
|
||||
if len(link) >= 2 && link[0] == '/' && link[1] != '/' { |
||||
return true |
||||
} |
||||
|
||||
// only the root '/'
|
||||
if len(link) == 1 && link[0] == '/' { |
||||
return true |
||||
} |
||||
|
||||
// current directory : begin with "./"
|
||||
if bytes.HasPrefix(link, []byte("./")) { |
||||
return true |
||||
} |
||||
|
||||
// parent directory : begin with "../"
|
||||
if bytes.HasPrefix(link, []byte("../")) { |
||||
return true |
||||
} |
||||
|
||||
return false |
||||
} |
||||
|
||||
func (r *HTMLRenderer) ensureUniqueHeadingID(id string) string { |
||||
for count, found := r.headingIDs[id]; found; count, found = r.headingIDs[id] { |
||||
tmp := fmt.Sprintf("%s-%d", id, count+1) |
||||
|
||||
if _, tmpFound := r.headingIDs[tmp]; !tmpFound { |
||||
r.headingIDs[id] = count + 1 |
||||
id = tmp |
||||
} else { |
||||
id = id + "-1" |
||||
} |
||||
} |
||||
|
||||
if _, found := r.headingIDs[id]; !found { |
||||
r.headingIDs[id] = 0 |
||||
} |
||||
|
||||
return id |
||||
} |
||||
|
||||
func (r *HTMLRenderer) addAbsPrefix(link []byte) []byte { |
||||
if r.AbsolutePrefix != "" && isRelativeLink(link) && link[0] != '.' { |
||||
newDest := r.AbsolutePrefix |
||||
if link[0] != '/' { |
||||
newDest += "/" |
||||
} |
||||
newDest += string(link) |
||||
return []byte(newDest) |
||||
} |
||||
return link |
||||
} |
||||
|
||||
func appendLinkAttrs(attrs []string, flags HTMLFlags, link []byte) []string { |
||||
if isRelativeLink(link) { |
||||
return attrs |
||||
} |
||||
val := []string{} |
||||
if flags&NofollowLinks != 0 { |
||||
val = append(val, "nofollow") |
||||
} |
||||
if flags&NoreferrerLinks != 0 { |
||||
val = append(val, "noreferrer") |
||||
} |
||||
if flags&NoopenerLinks != 0 { |
||||
val = append(val, "noopener") |
||||
} |
||||
if flags&HrefTargetBlank != 0 { |
||||
attrs = append(attrs, "target=\"_blank\"") |
||||
} |
||||
if len(val) == 0 { |
||||
return attrs |
||||
} |
||||
attr := fmt.Sprintf("rel=%q", strings.Join(val, " ")) |
||||
return append(attrs, attr) |
||||
} |
||||
|
||||
func isMailto(link []byte) bool { |
||||
return bytes.HasPrefix(link, []byte("mailto:")) |
||||
} |
||||
|
||||
func needSkipLink(flags HTMLFlags, dest []byte) bool { |
||||
if flags&SkipLinks != 0 { |
||||
return true |
||||
} |
||||
return flags&Safelink != 0 && !isSafeLink(dest) && !isMailto(dest) |
||||
} |
||||
|
||||
func isSmartypantable(node *Node) bool { |
||||
pt := node.Parent.Type |
||||
return pt != Link && pt != CodeBlock && pt != Code |
||||
} |
||||
|
||||
func appendLanguageAttr(attrs []string, info []byte) []string { |
||||
if len(info) == 0 { |
||||
return attrs |
||||
} |
||||
endOfLang := bytes.IndexAny(info, "\t ") |
||||
if endOfLang < 0 { |
||||
endOfLang = len(info) |
||||
} |
||||
return append(attrs, fmt.Sprintf("class=\"language-%s\"", info[:endOfLang])) |
||||
} |
||||
|
||||
func (r *HTMLRenderer) tag(w io.Writer, name []byte, attrs []string) { |
||||
w.Write(name) |
||||
if len(attrs) > 0 { |
||||
w.Write(spaceBytes) |
||||
w.Write([]byte(strings.Join(attrs, " "))) |
||||
} |
||||
w.Write(gtBytes) |
||||
r.lastOutputLen = 1 |
||||
} |
||||
|
||||
func footnoteRef(prefix string, node *Node) []byte { |
||||
urlFrag := prefix + string(slugify(node.Destination)) |
||||
anchor := fmt.Sprintf(`<a href="#fn:%s">%d</a>`, urlFrag, node.NoteID) |
||||
return []byte(fmt.Sprintf(`<sup class="footnote-ref" id="fnref:%s">%s</sup>`, urlFrag, anchor)) |
||||
} |
||||
|
||||
func footnoteItem(prefix string, slug []byte) []byte { |
||||
return []byte(fmt.Sprintf(`<li id="fn:%s%s">`, prefix, slug)) |
||||
} |
||||
|
||||
func footnoteReturnLink(prefix, returnLink string, slug []byte) []byte { |
||||
const format = ` <a class="footnote-return" href="#fnref:%s%s">%s</a>` |
||||
return []byte(fmt.Sprintf(format, prefix, slug, returnLink)) |
||||
} |
||||
|
||||
func itemOpenCR(node *Node) bool { |
||||
if node.Prev == nil { |
||||
return false |
||||
} |
||||
ld := node.Parent.ListData |
||||
return !ld.Tight && ld.ListFlags&ListTypeDefinition == 0 |
||||
} |
||||
|
||||
func skipParagraphTags(node *Node) bool { |
||||
grandparent := node.Parent.Parent |
||||
if grandparent == nil || grandparent.Type != List { |
||||
return false |
||||
} |
||||
tightOrTerm := grandparent.Tight || node.Parent.ListFlags&ListTypeTerm != 0 |
||||
return grandparent.Type == List && tightOrTerm |
||||
} |
||||
|
||||
func cellAlignment(align CellAlignFlags) string { |
||||
switch align { |
||||
case TableAlignmentLeft: |
||||
return "left" |
||||
case TableAlignmentRight: |
||||
return "right" |
||||
case TableAlignmentCenter: |
||||
return "center" |
||||
default: |
||||
return "" |
||||
} |
||||
} |
||||
|
||||
func (r *HTMLRenderer) out(w io.Writer, text []byte) { |
||||
if r.disableTags > 0 { |
||||
w.Write(htmlTagRe.ReplaceAll(text, []byte{})) |
||||
} else { |
||||
w.Write(text) |
||||
} |
||||
r.lastOutputLen = len(text) |
||||
} |
||||
|
||||
func (r *HTMLRenderer) cr(w io.Writer) { |
||||
if r.lastOutputLen > 0 { |
||||
r.out(w, nlBytes) |
||||
} |
||||
} |
||||
|
||||
var ( |
||||
nlBytes = []byte{'\n'} |
||||
gtBytes = []byte{'>'} |
||||
spaceBytes = []byte{' '} |
||||
) |
||||
|
||||
var ( |
||||
brTag = []byte("<br>") |
||||
brXHTMLTag = []byte("<br />") |
||||
emTag = []byte("<em>") |
||||
emCloseTag = []byte("</em>") |
||||
strongTag = []byte("<strong>") |
||||
strongCloseTag = []byte("</strong>") |
||||
delTag = []byte("<del>") |
||||
delCloseTag = []byte("</del>") |
||||
ttTag = []byte("<tt>") |
||||
ttCloseTag = []byte("</tt>") |
||||
aTag = []byte("<a") |
||||
aCloseTag = []byte("</a>") |
||||
preTag = []byte("<pre>") |
||||
preCloseTag = []byte("</pre>") |
||||
codeTag = []byte("<code>") |
||||
codeCloseTag = []byte("</code>") |
||||
pTag = []byte("<p>") |
||||
pCloseTag = []byte("</p>") |
||||
blockquoteTag = []byte("<blockquote>") |
||||
blockquoteCloseTag = []byte("</blockquote>") |
||||
hrTag = []byte("<hr>") |
||||
hrXHTMLTag = []byte("<hr />") |
||||
ulTag = []byte("<ul>") |
||||
ulCloseTag = []byte("</ul>") |
||||
olTag = []byte("<ol>") |
||||
olCloseTag = []byte("</ol>") |
||||
dlTag = []byte("<dl>") |
||||
dlCloseTag = []byte("</dl>") |
||||
liTag = []byte("<li>") |
||||
liCloseTag = []byte("</li>") |
||||
ddTag = []byte("<dd>") |
||||
ddCloseTag = []byte("</dd>") |
||||
dtTag = []byte("<dt>") |
||||
dtCloseTag = []byte("</dt>") |
||||
tableTag = []byte("<table>") |
||||
tableCloseTag = []byte("</table>") |
||||
tdTag = []byte("<td") |
||||
tdCloseTag = []byte("</td>") |
||||
thTag = []byte("<th") |
||||
thCloseTag = []byte("</th>") |
||||
theadTag = []byte("<thead>") |
||||
theadCloseTag = []byte("</thead>") |
||||
tbodyTag = []byte("<tbody>") |
||||
tbodyCloseTag = []byte("</tbody>") |
||||
trTag = []byte("<tr>") |
||||
trCloseTag = []byte("</tr>") |
||||
h1Tag = []byte("<h1") |
||||
h1CloseTag = []byte("</h1>") |
||||
h2Tag = []byte("<h2") |
||||
h2CloseTag = []byte("</h2>") |
||||
h3Tag = []byte("<h3") |
||||
h3CloseTag = []byte("</h3>") |
||||
h4Tag = []byte("<h4") |
||||
h4CloseTag = []byte("</h4>") |
||||
h5Tag = []byte("<h5") |
||||
h5CloseTag = []byte("</h5>") |
||||
h6Tag = []byte("<h6") |
||||
h6CloseTag = []byte("</h6>") |
||||
|
||||
footnotesDivBytes = []byte("\n<div class=\"footnotes\">\n\n") |
||||
footnotesCloseDivBytes = []byte("\n</div>\n") |
||||
) |
||||
|
||||
func headingTagsFromLevel(level int) ([]byte, []byte) { |
||||
if level <= 1 { |
||||
return h1Tag, h1CloseTag |
||||
} |
||||
switch level { |
||||
case 2: |
||||
return h2Tag, h2CloseTag |
||||
case 3: |
||||
return h3Tag, h3CloseTag |
||||
case 4: |
||||
return h4Tag, h4CloseTag |
||||
case 5: |
||||
return h5Tag, h5CloseTag |
||||
} |
||||
return h6Tag, h6CloseTag |
||||
} |
||||
|
||||
func (r *HTMLRenderer) outHRTag(w io.Writer) { |
||||
if r.Flags&UseXHTML == 0 { |
||||
r.out(w, hrTag) |
||||
} else { |
||||
r.out(w, hrXHTMLTag) |
||||
} |
||||
} |
||||
|
||||
// RenderNode is a default renderer of a single node of a syntax tree. For
|
||||
// block nodes it will be called twice: first time with entering=true, second
|
||||
// time with entering=false, so that it could know when it's working on an open
|
||||
// tag and when on close. It writes the result to w.
|
||||
//
|
||||
// The return value is a way to tell the calling walker to adjust its walk
|
||||
// pattern: e.g. it can terminate the traversal by returning Terminate. Or it
|
||||
// can ask the walker to skip a subtree of this node by returning SkipChildren.
|
||||
// The typical behavior is to return GoToNext, which asks for the usual
|
||||
// traversal to the next node.
|
||||
func (r *HTMLRenderer) RenderNode(w io.Writer, node *Node, entering bool) WalkStatus { |
||||
attrs := []string{} |
||||
switch node.Type { |
||||
case Text: |
||||
if r.Flags&Smartypants != 0 { |
||||
var tmp bytes.Buffer |
||||
escapeHTML(&tmp, node.Literal) |
||||
r.sr.Process(w, tmp.Bytes()) |
||||
} else { |
||||
if node.Parent.Type == Link { |
||||
escLink(w, node.Literal) |
||||
} else { |
||||
escapeHTML(w, node.Literal) |
||||
} |
||||
} |
||||
case Softbreak: |
||||
r.cr(w) |
||||
// TODO: make it configurable via out(renderer.softbreak)
|
||||
case Hardbreak: |
||||
if r.Flags&UseXHTML == 0 { |
||||
r.out(w, brTag) |
||||
} else { |
||||
r.out(w, brXHTMLTag) |
||||
} |
||||
r.cr(w) |
||||
case Emph: |
||||
if entering { |
||||
r.out(w, emTag) |
||||
} else { |
||||
r.out(w, emCloseTag) |
||||
} |
||||
case Strong: |
||||
if entering { |
||||
r.out(w, strongTag) |
||||
} else { |
||||
r.out(w, strongCloseTag) |
||||
} |
||||
case Del: |
||||
if entering { |
||||
r.out(w, delTag) |
||||
} else { |
||||
r.out(w, delCloseTag) |
||||
} |
||||
case HTMLSpan: |
||||
if r.Flags&SkipHTML != 0 { |
||||
break |
||||
} |
||||
r.out(w, node.Literal) |
||||
case Link: |
||||
// mark it but don't link it if it is not a safe link: no smartypants
|
||||
dest := node.LinkData.Destination |
||||
if needSkipLink(r.Flags, dest) { |
||||
if entering { |
||||
r.out(w, ttTag) |
||||
} else { |
||||
r.out(w, ttCloseTag) |
||||
} |
||||
} else { |
||||
if entering { |
||||
dest = r.addAbsPrefix(dest) |
||||
var hrefBuf bytes.Buffer |
||||
hrefBuf.WriteString("href=\"") |
||||
escLink(&hrefBuf, dest) |
||||
hrefBuf.WriteByte('"') |
||||
attrs = append(attrs, hrefBuf.String()) |
||||
if node.NoteID != 0 { |
||||
r.out(w, footnoteRef(r.FootnoteAnchorPrefix, node)) |
||||
break |
||||
} |
||||
attrs = appendLinkAttrs(attrs, r.Flags, dest) |
||||
if len(node.LinkData.Title) > 0 { |
||||
var titleBuff bytes.Buffer |
||||
titleBuff.WriteString("title=\"") |
||||
escapeHTML(&titleBuff, node.LinkData.Title) |
||||
titleBuff.WriteByte('"') |
||||
attrs = append(attrs, titleBuff.String()) |
||||
} |
||||
r.tag(w, aTag, attrs) |
||||
} else { |
||||
if node.NoteID != 0 { |
||||
break |
||||
} |
||||
r.out(w, aCloseTag) |
||||
} |
||||
} |
||||
case Image: |
||||
if r.Flags&SkipImages != 0 { |
||||
return SkipChildren |
||||
} |
||||
if entering { |
||||
dest := node.LinkData.Destination |
||||
dest = r.addAbsPrefix(dest) |
||||
if r.disableTags == 0 { |
||||
//if options.safe && potentiallyUnsafe(dest) {
|
||||
//out(w, `<img src="" alt="`)
|
||||
//} else {
|
||||
r.out(w, []byte(`<img src="`)) |
||||
escLink(w, dest) |
||||
r.out(w, []byte(`" alt="`)) |
||||
//}
|
||||
} |
||||
r.disableTags++ |
||||
} else { |
||||
r.disableTags-- |
||||
if r.disableTags == 0 { |
||||
if node.LinkData.Title != nil { |
||||
r.out(w, []byte(`" title="`)) |
||||
escapeHTML(w, node.LinkData.Title) |
||||
} |
||||
r.out(w, []byte(`" />`)) |
||||
} |
||||
} |
||||
case Code: |
||||
r.out(w, codeTag) |
||||
escapeHTML(w, node.Literal) |
||||
r.out(w, codeCloseTag) |
||||
case Document: |
||||
break |
||||
case Paragraph: |
||||
if skipParagraphTags(node) { |
||||
break |
||||
} |
||||
if entering { |
||||
// TODO: untangle this clusterfuck about when the newlines need
|
||||
// to be added and when not.
|
||||
if node.Prev != nil { |
||||
switch node.Prev.Type { |
||||
case HTMLBlock, List, Paragraph, Heading, CodeBlock, BlockQuote, HorizontalRule: |
||||
r.cr(w) |
||||
} |
||||
} |
||||
if node.Parent.Type == BlockQuote && node.Prev == nil { |
||||
r.cr(w) |
||||
} |
||||
r.out(w, pTag) |
||||
} else { |
||||
r.out(w, pCloseTag) |
||||
if !(node.Parent.Type == Item && node.Next == nil) { |
||||
r.cr(w) |
||||
} |
||||
} |
||||
case BlockQuote: |
||||
if entering { |
||||
r.cr(w) |
||||
r.out(w, blockquoteTag) |
||||
} else { |
||||
r.out(w, blockquoteCloseTag) |
||||
r.cr(w) |
||||
} |
||||
case HTMLBlock: |
||||
if r.Flags&SkipHTML != 0 { |
||||
break |
||||
} |
||||
r.cr(w) |
||||
r.out(w, node.Literal) |
||||
r.cr(w) |
||||
case Heading: |
||||
headingLevel := r.HTMLRendererParameters.HeadingLevelOffset + node.Level |
||||
openTag, closeTag := headingTagsFromLevel(headingLevel) |
||||
if entering { |
||||
if node.IsTitleblock { |
||||
attrs = append(attrs, `class="title"`) |
||||
} |
||||
if node.HeadingID != "" { |
||||
id := r.ensureUniqueHeadingID(node.HeadingID) |
||||
if r.HeadingIDPrefix != "" { |
||||
id = r.HeadingIDPrefix + id |
||||
} |
||||
if r.HeadingIDSuffix != "" { |
||||
id = id + r.HeadingIDSuffix |
||||
} |
||||
attrs = append(attrs, fmt.Sprintf(`id="%s"`, id)) |
||||
} |
||||
r.cr(w) |
||||
r.tag(w, openTag, attrs) |
||||
} else { |
||||
r.out(w, closeTag) |
||||
if !(node.Parent.Type == Item && node.Next == nil) { |
||||
r.cr(w) |
||||
} |
||||
} |
||||
case HorizontalRule: |
||||
r.cr(w) |
||||
r.outHRTag(w) |
||||
r.cr(w) |
||||
case List: |
||||
openTag := ulTag |
||||
closeTag := ulCloseTag |
||||
if node.ListFlags&ListTypeOrdered != 0 { |
||||
openTag = olTag |
||||
closeTag = olCloseTag |
||||
} |
||||
if node.ListFlags&ListTypeDefinition != 0 { |
||||
openTag = dlTag |
||||
closeTag = dlCloseTag |
||||
} |
||||
if entering { |
||||
if node.IsFootnotesList { |
||||
r.out(w, footnotesDivBytes) |
||||
r.outHRTag(w) |
||||
r.cr(w) |
||||
} |
||||
r.cr(w) |
||||
if node.Parent.Type == Item && node.Parent.Parent.Tight { |
||||
r.cr(w) |
||||
} |
||||
r.tag(w, openTag[:len(openTag)-1], attrs) |
||||
r.cr(w) |
||||
} else { |
||||
r.out(w, closeTag) |
||||
//cr(w)
|
||||
//if node.parent.Type != Item {
|
||||
// cr(w)
|
||||
//}
|
||||
if node.Parent.Type == Item && node.Next != nil { |
||||
r.cr(w) |
||||
} |
||||
if node.Parent.Type == Document || node.Parent.Type == BlockQuote { |
||||
r.cr(w) |
||||
} |
||||
if node.IsFootnotesList { |
||||
r.out(w, footnotesCloseDivBytes) |
||||
} |
||||
} |
||||
case Item: |
||||
openTag := liTag |
||||
closeTag := liCloseTag |
||||
if node.ListFlags&ListTypeDefinition != 0 { |
||||
openTag = ddTag |
||||
closeTag = ddCloseTag |
||||
} |
||||
if node.ListFlags&ListTypeTerm != 0 { |
||||
openTag = dtTag |
||||
closeTag = dtCloseTag |
||||
} |
||||
if entering { |
||||
if itemOpenCR(node) { |
||||
r.cr(w) |
||||
} |
||||
if node.ListData.RefLink != nil { |
||||
slug := slugify(node.ListData.RefLink) |
||||
r.out(w, footnoteItem(r.FootnoteAnchorPrefix, slug)) |
||||
break |
||||
} |
||||
r.out(w, openTag) |
||||
} else { |
||||
if node.ListData.RefLink != nil { |
||||
slug := slugify(node.ListData.RefLink) |
||||
if r.Flags&FootnoteReturnLinks != 0 { |
||||
r.out(w, footnoteReturnLink(r.FootnoteAnchorPrefix, r.FootnoteReturnLinkContents, slug)) |
||||
} |
||||
} |
||||
r.out(w, closeTag) |
||||
r.cr(w) |
||||
} |
||||
case CodeBlock: |
||||
attrs = appendLanguageAttr(attrs, node.Info) |
||||
r.cr(w) |
||||
r.out(w, preTag) |
||||
r.tag(w, codeTag[:len(codeTag)-1], attrs) |
||||
escapeHTML(w, node.Literal) |
||||
r.out(w, codeCloseTag) |
||||
r.out(w, preCloseTag) |
||||
if node.Parent.Type != Item { |
||||
r.cr(w) |
||||
} |
||||
case Table: |
||||
if entering { |
||||
r.cr(w) |
||||
r.out(w, tableTag) |
||||
} else { |
||||
r.out(w, tableCloseTag) |
||||
r.cr(w) |
||||
} |
||||
case TableCell: |
||||
openTag := tdTag |
||||
closeTag := tdCloseTag |
||||
if node.IsHeader { |
||||
openTag = thTag |
||||
closeTag = thCloseTag |
||||
} |
||||
if entering { |
||||
align := cellAlignment(node.Align) |
||||
if align != "" { |
||||
attrs = append(attrs, fmt.Sprintf(`align="%s"`, align)) |
||||
} |
||||
if node.Prev == nil { |
||||
r.cr(w) |
||||
} |
||||
r.tag(w, openTag, attrs) |
||||
} else { |
||||
r.out(w, closeTag) |
||||
r.cr(w) |
||||
} |
||||
case TableHead: |
||||
if entering { |
||||
r.cr(w) |
||||
r.out(w, theadTag) |
||||
} else { |
||||
r.out(w, theadCloseTag) |
||||
r.cr(w) |
||||
} |
||||
case TableBody: |
||||
if entering { |
||||
r.cr(w) |
||||
r.out(w, tbodyTag) |
||||
// XXX: this is to adhere to a rather silly test. Should fix test.
|
||||
if node.FirstChild == nil { |
||||
r.cr(w) |
||||
} |
||||
} else { |
||||
r.out(w, tbodyCloseTag) |
||||
r.cr(w) |
||||
} |
||||
case TableRow: |
||||
if entering { |
||||
r.cr(w) |
||||
r.out(w, trTag) |
||||
} else { |
||||
r.out(w, trCloseTag) |
||||
r.cr(w) |
||||
} |
||||
default: |
||||
panic("Unknown node type " + node.Type.String()) |
||||
} |
||||
return GoToNext |
||||
} |
||||
|
||||
// RenderHeader writes HTML document preamble and TOC if requested.
|
||||
func (r *HTMLRenderer) RenderHeader(w io.Writer, ast *Node) { |
||||
r.writeDocumentHeader(w) |
||||
if r.Flags&TOC != 0 { |
||||
r.writeTOC(w, ast) |
||||
} |
||||
} |
||||
|
||||
// RenderFooter writes HTML document footer.
|
||||
func (r *HTMLRenderer) RenderFooter(w io.Writer, ast *Node) { |
||||
if r.Flags&CompletePage == 0 { |
||||
return |
||||
} |
||||
io.WriteString(w, "\n</body>\n</html>\n") |
||||
} |
||||
|
||||
func (r *HTMLRenderer) writeDocumentHeader(w io.Writer) { |
||||
if r.Flags&CompletePage == 0 { |
||||
return |
||||
} |
||||
ending := "" |
||||
if r.Flags&UseXHTML != 0 { |
||||
io.WriteString(w, "<!DOCTYPE html PUBLIC \"-//W3C//DTD XHTML 1.0 Transitional//EN\" ") |
||||
io.WriteString(w, "\"http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd\">\n") |
||||
io.WriteString(w, "<html xmlns=\"http://www.w3.org/1999/xhtml\">\n") |
||||
ending = " /" |
||||
} else { |
||||
io.WriteString(w, "<!DOCTYPE html>\n") |
||||
io.WriteString(w, "<html>\n") |
||||
} |
||||
io.WriteString(w, "<head>\n") |
||||
io.WriteString(w, " <title>") |
||||
if r.Flags&Smartypants != 0 { |
||||
r.sr.Process(w, []byte(r.Title)) |
||||
} else { |
||||
escapeHTML(w, []byte(r.Title)) |
||||
} |
||||
io.WriteString(w, "</title>\n") |
||||
io.WriteString(w, " <meta name=\"GENERATOR\" content=\"Blackfriday Markdown Processor v") |
||||
io.WriteString(w, Version) |
||||
io.WriteString(w, "\"") |
||||
io.WriteString(w, ending) |
||||
io.WriteString(w, ">\n") |
||||
io.WriteString(w, " <meta charset=\"utf-8\"") |
||||
io.WriteString(w, ending) |
||||
io.WriteString(w, ">\n") |
||||
if r.CSS != "" { |
||||
io.WriteString(w, " <link rel=\"stylesheet\" type=\"text/css\" href=\"") |
||||
escapeHTML(w, []byte(r.CSS)) |
||||
io.WriteString(w, "\"") |
||||
io.WriteString(w, ending) |
||||
io.WriteString(w, ">\n") |
||||
} |
||||
if r.Icon != "" { |
||||
io.WriteString(w, " <link rel=\"icon\" type=\"image/x-icon\" href=\"") |
||||
escapeHTML(w, []byte(r.Icon)) |
||||
io.WriteString(w, "\"") |
||||
io.WriteString(w, ending) |
||||
io.WriteString(w, ">\n") |
||||
} |
||||
io.WriteString(w, "</head>\n") |
||||
io.WriteString(w, "<body>\n\n") |
||||
} |
||||
|
||||
func (r *HTMLRenderer) writeTOC(w io.Writer, ast *Node) { |
||||
buf := bytes.Buffer{} |
||||
|
||||
inHeading := false |
||||
tocLevel := 0 |
||||
headingCount := 0 |
||||
|
||||
ast.Walk(func(node *Node, entering bool) WalkStatus { |
||||
if node.Type == Heading && !node.HeadingData.IsTitleblock { |
||||
inHeading = entering |
||||
if entering { |
||||
node.HeadingID = fmt.Sprintf("toc_%d", headingCount) |
||||
if node.Level == tocLevel { |
||||
buf.WriteString("</li>\n\n<li>") |
||||
} else if node.Level < tocLevel { |
||||
for node.Level < tocLevel { |
||||
tocLevel-- |
||||
buf.WriteString("</li>\n</ul>") |
||||
} |
||||
buf.WriteString("</li>\n\n<li>") |
||||
} else { |
||||
for node.Level > tocLevel { |
||||
tocLevel++ |
||||
buf.WriteString("\n<ul>\n<li>") |
||||
} |
||||
} |
||||
|
||||
fmt.Fprintf(&buf, `<a href="#toc_%d">`, headingCount) |
||||
headingCount++ |
||||
} else { |
||||
buf.WriteString("</a>") |
||||
} |
||||
return GoToNext |
||||
} |
||||
|
||||
if inHeading { |
||||
return r.RenderNode(&buf, node, entering) |
||||
} |
||||
|
||||
return GoToNext |
||||
}) |
||||
|
||||
for ; tocLevel > 0; tocLevel-- { |
||||
buf.WriteString("</li>\n</ul>") |
||||
} |
||||
|
||||
if buf.Len() > 0 { |
||||
io.WriteString(w, "<nav>\n") |
||||
w.Write(buf.Bytes()) |
||||
io.WriteString(w, "\n\n</nav>\n") |
||||
} |
||||
r.lastOutputLen = buf.Len() |
||||
} |
File diff suppressed because it is too large
Load Diff
@ -0,0 +1,354 @@ |
||||
package blackfriday |
||||
|
||||
import ( |
||||
"bytes" |
||||
"fmt" |
||||
) |
||||
|
||||
// NodeType specifies a type of a single node of a syntax tree. Usually one
|
||||
// node (and its type) corresponds to a single markdown feature, e.g. emphasis
|
||||
// or code block.
|
||||
type NodeType int |
||||
|
||||
// Constants for identifying different types of nodes. See NodeType.
|
||||
const ( |
||||
Document NodeType = iota |
||||
BlockQuote |
||||
List |
||||
Item |
||||
Paragraph |
||||
Heading |
||||
HorizontalRule |
||||
Emph |
||||
Strong |
||||
Del |
||||
Link |
||||
Image |
||||
Text |
||||
HTMLBlock |
||||
CodeBlock |
||||
Softbreak |
||||
Hardbreak |
||||
Code |
||||
HTMLSpan |
||||
Table |
||||
TableCell |
||||
TableHead |
||||
TableBody |
||||
TableRow |
||||
) |
||||
|
||||
var nodeTypeNames = []string{ |
||||
Document: "Document", |
||||
BlockQuote: "BlockQuote", |
||||
List: "List", |
||||
Item: "Item", |
||||
Paragraph: "Paragraph", |
||||
Heading: "Heading", |
||||
HorizontalRule: "HorizontalRule", |
||||
Emph: "Emph", |
||||
Strong: "Strong", |
||||
Del: "Del", |
||||
Link: "Link", |
||||
Image: "Image", |
||||
Text: "Text", |
||||
HTMLBlock: "HTMLBlock", |
||||
CodeBlock: "CodeBlock", |
||||
Softbreak: "Softbreak", |
||||
Hardbreak: "Hardbreak", |
||||
Code: "Code", |
||||
HTMLSpan: "HTMLSpan", |
||||
Table: "Table", |
||||
TableCell: "TableCell", |
||||
TableHead: "TableHead", |
||||
TableBody: "TableBody", |
||||
TableRow: "TableRow", |
||||
} |
||||
|
||||
func (t NodeType) String() string { |
||||
return nodeTypeNames[t] |
||||
} |
||||
|
||||
// ListData contains fields relevant to a List and Item node type.
|
||||
type ListData struct { |
||||
ListFlags ListType |
||||
Tight bool // Skip <p>s around list item data if true
|
||||
BulletChar byte // '*', '+' or '-' in bullet lists
|
||||
Delimiter byte // '.' or ')' after the number in ordered lists
|
||||
RefLink []byte // If not nil, turns this list item into a footnote item and triggers different rendering
|
||||
IsFootnotesList bool // This is a list of footnotes
|
||||
} |
||||
|
||||
// LinkData contains fields relevant to a Link node type.
|
||||
type LinkData struct { |
||||
Destination []byte // Destination is what goes into a href
|
||||
Title []byte // Title is the tooltip thing that goes in a title attribute
|
||||
NoteID int // NoteID contains a serial number of a footnote, zero if it's not a footnote
|
||||
Footnote *Node // If it's a footnote, this is a direct link to the footnote Node. Otherwise nil.
|
||||
} |
||||
|
||||
// CodeBlockData contains fields relevant to a CodeBlock node type.
|
||||
type CodeBlockData struct { |
||||
IsFenced bool // Specifies whether it's a fenced code block or an indented one
|
||||
Info []byte // This holds the info string
|
||||
FenceChar byte |
||||
FenceLength int |
||||
FenceOffset int |
||||
} |
||||
|
||||
// TableCellData contains fields relevant to a TableCell node type.
|
||||
type TableCellData struct { |
||||
IsHeader bool // This tells if it's under the header row
|
||||
Align CellAlignFlags // This holds the value for align attribute
|
||||
} |
||||
|
||||
// HeadingData contains fields relevant to a Heading node type.
|
||||
type HeadingData struct { |
||||
Level int // This holds the heading level number
|
||||
HeadingID string // This might hold heading ID, if present
|
||||
IsTitleblock bool // Specifies whether it's a title block
|
||||
} |
||||
|
||||
// Node is a single element in the abstract syntax tree of the parsed document.
|
||||
// It holds connections to the structurally neighboring nodes and, for certain
|
||||
// types of nodes, additional information that might be needed when rendering.
|
||||
type Node struct { |
||||
Type NodeType // Determines the type of the node
|
||||
Parent *Node // Points to the parent
|
||||
FirstChild *Node // Points to the first child, if any
|
||||
LastChild *Node // Points to the last child, if any
|
||||
Prev *Node // Previous sibling; nil if it's the first child
|
||||
Next *Node // Next sibling; nil if it's the last child
|
||||
|
||||
Literal []byte // Text contents of the leaf nodes
|
||||
|
||||
HeadingData // Populated if Type is Heading
|
||||
ListData // Populated if Type is List
|
||||
CodeBlockData // Populated if Type is CodeBlock
|
||||
LinkData // Populated if Type is Link
|
||||
TableCellData // Populated if Type is TableCell
|
||||
|
||||
content []byte // Markdown content of the block nodes
|
||||
open bool // Specifies an open block node that has not been finished to process yet
|
||||
} |
||||
|
||||
// NewNode allocates a node of a specified type.
|
||||
func NewNode(typ NodeType) *Node { |
||||
return &Node{ |
||||
Type: typ, |
||||
open: true, |
||||
} |
||||
} |
||||
|
||||
func (n *Node) String() string { |
||||
ellipsis := "" |
||||
snippet := n.Literal |
||||
if len(snippet) > 16 { |
||||
snippet = snippet[:16] |
||||
ellipsis = "..." |
||||
} |
||||
return fmt.Sprintf("%s: '%s%s'", n.Type, snippet, ellipsis) |
||||
} |
||||
|
||||
// Unlink removes node 'n' from the tree.
|
||||
// It panics if the node is nil.
|
||||
func (n *Node) Unlink() { |
||||
if n.Prev != nil { |
||||
n.Prev.Next = n.Next |
||||
} else if n.Parent != nil { |
||||
n.Parent.FirstChild = n.Next |
||||
} |
||||
if n.Next != nil { |
||||
n.Next.Prev = n.Prev |
||||
} else if n.Parent != nil { |
||||
n.Parent.LastChild = n.Prev |
||||
} |
||||
n.Parent = nil |
||||
n.Next = nil |
||||
n.Prev = nil |
||||
} |
||||
|
||||
// AppendChild adds a node 'child' as a child of 'n'.
|
||||
// It panics if either node is nil.
|
||||
func (n *Node) AppendChild(child *Node) { |
||||
child.Unlink() |
||||
child.Parent = n |
||||
if n.LastChild != nil { |
||||
n.LastChild.Next = child |
||||
child.Prev = n.LastChild |
||||
n.LastChild = child |
||||
} else { |
||||
n.FirstChild = child |
||||
n.LastChild = child |
||||
} |
||||
} |
||||
|
||||
// InsertBefore inserts 'sibling' immediately before 'n'.
|
||||
// It panics if either node is nil.
|
||||
func (n *Node) InsertBefore(sibling *Node) { |
||||
sibling.Unlink() |
||||
sibling.Prev = n.Prev |
||||
if sibling.Prev != nil { |
||||
sibling.Prev.Next = sibling |
||||
} |
||||
sibling.Next = n |
||||
n.Prev = sibling |
||||
sibling.Parent = n.Parent |
||||
if sibling.Prev == nil { |
||||
sibling.Parent.FirstChild = sibling |
||||
} |
||||
} |
||||
|
||||
func (n *Node) isContainer() bool { |
||||
switch n.Type { |
||||
case Document: |
||||
fallthrough |
||||
case BlockQuote: |
||||
fallthrough |
||||
case List: |
||||
fallthrough |
||||
case Item: |
||||
fallthrough |
||||
case Paragraph: |
||||
fallthrough |
||||
case Heading: |
||||
fallthrough |
||||
case Emph: |
||||
fallthrough |
||||
case Strong: |
||||
fallthrough |
||||
case Del: |
||||
fallthrough |
||||
case Link: |
||||
fallthrough |
||||
case Image: |
||||
fallthrough |
||||
case Table: |
||||
fallthrough |
||||
case TableHead: |
||||
fallthrough |
||||
case TableBody: |
||||
fallthrough |
||||
case TableRow: |
||||
fallthrough |
||||
case TableCell: |
||||
return true |
||||
default: |
||||
return false |
||||
} |
||||
} |
||||
|
||||
func (n *Node) canContain(t NodeType) bool { |
||||
if n.Type == List { |
||||
return t == Item |
||||
} |
||||
if n.Type == Document || n.Type == BlockQuote || n.Type == Item { |
||||
return t != Item |
||||
} |
||||
if n.Type == Table { |
||||
return t == TableHead || t == TableBody |
||||
} |
||||
if n.Type == TableHead || n.Type == TableBody { |
||||
return t == TableRow |
||||
} |
||||
if n.Type == TableRow { |
||||
return t == TableCell |
||||
} |
||||
return false |
||||
} |
||||
|
||||
// WalkStatus allows NodeVisitor to have some control over the tree traversal.
|
||||
// It is returned from NodeVisitor and different values allow Node.Walk to
|
||||
// decide which node to go to next.
|
||||
type WalkStatus int |
||||
|
||||
const ( |
||||
// GoToNext is the default traversal of every node.
|
||||
GoToNext WalkStatus = iota |
||||
// SkipChildren tells walker to skip all children of current node.
|
||||
SkipChildren |
||||
// Terminate tells walker to terminate the traversal.
|
||||
Terminate |
||||
) |
||||
|
||||
// NodeVisitor is a callback to be called when traversing the syntax tree.
|
||||
// Called twice for every node: once with entering=true when the branch is
|
||||
// first visited, then with entering=false after all the children are done.
|
||||
type NodeVisitor func(node *Node, entering bool) WalkStatus |
||||
|
||||
// Walk is a convenience method that instantiates a walker and starts a
|
||||
// traversal of subtree rooted at n.
|
||||
func (n *Node) Walk(visitor NodeVisitor) { |
||||
w := newNodeWalker(n) |
||||
for w.current != nil { |
||||
status := visitor(w.current, w.entering) |
||||
switch status { |
||||
case GoToNext: |
||||
w.next() |
||||
case SkipChildren: |
||||
w.entering = false |
||||
w.next() |
||||
case Terminate: |
||||
return |
||||
} |
||||
} |
||||
} |
||||
|
||||
type nodeWalker struct { |
||||
current *Node |
||||
root *Node |
||||
entering bool |
||||
} |
||||
|
||||
func newNodeWalker(root *Node) *nodeWalker { |
||||
return &nodeWalker{ |
||||
current: root, |
||||
root: root, |
||||
entering: true, |
||||
} |
||||
} |
||||
|
||||
func (nw *nodeWalker) next() { |
||||
if (!nw.current.isContainer() || !nw.entering) && nw.current == nw.root { |
||||
nw.current = nil |
||||
return |
||||
} |
||||
if nw.entering && nw.current.isContainer() { |
||||
if nw.current.FirstChild != nil { |
||||
nw.current = nw.current.FirstChild |
||||
nw.entering = true |
||||
} else { |
||||
nw.entering = false |
||||
} |
||||
} else if nw.current.Next == nil { |
||||
nw.current = nw.current.Parent |
||||
nw.entering = false |
||||
} else { |
||||
nw.current = nw.current.Next |
||||
nw.entering = true |
||||
} |
||||
} |
||||
|
||||
func dump(ast *Node) { |
||||
fmt.Println(dumpString(ast)) |
||||
} |
||||
|
||||
func dumpR(ast *Node, depth int) string { |
||||
if ast == nil { |
||||
return "" |
||||
} |
||||
indent := bytes.Repeat([]byte("\t"), depth) |
||||
content := ast.Literal |
||||
if content == nil { |
||||
content = ast.content |
||||
} |
||||
result := fmt.Sprintf("%s%s(%q)\n", indent, ast.Type, content) |
||||
for n := ast.FirstChild; n != nil; n = n.Next { |
||||
result += dumpR(n, depth+1) |
||||
} |
||||
return result |
||||
} |
||||
|
||||
func dumpString(ast *Node) string { |
||||
return dumpR(ast, 0) |
||||
} |
@ -0,0 +1 @@ |
||||
module github.com/shurcooL/sanitized_anchor_name |
Loading…
Reference in new issue