Since writers are normally only used synchronously (i.e. to write one document at a time), we don't guard modifications to their internal state (e.g. temporarily replacing the string.Builder in WriteNodesAsString) against race conditions. The package global `orgWriter` and corresponding use cases of it (`org.String`, `$node.String`) break that pattern - the writer is potentially used from multiple go routines at the same time. This results in race conditions that manifest as error messages like e.g. could not write output: runtime error: invalid memory address or nil pointer dereference. Using unrendered content. Additionally, since we catch panics in `Document.Write`, the corresponding stack trace is lost and dependents of go-org never know what hit them. As using a writer across simultaneously across go routines is not a standard pattern, we'll sync the use of the global `orgWriter` instead of trying to make the actual writer threadsafe; less code noise for the common use case.
123 lines
3.6 KiB
Go
123 lines
3.6 KiB
Go
package org
|
|
|
|
import (
|
|
"fmt"
|
|
"regexp"
|
|
"strings"
|
|
"unicode"
|
|
)
|
|
|
|
type List struct {
|
|
Kind string
|
|
Items []Node
|
|
}
|
|
|
|
type ListItem struct {
|
|
Bullet string
|
|
Status string
|
|
Value string
|
|
Children []Node
|
|
}
|
|
|
|
type DescriptiveListItem struct {
|
|
Bullet string
|
|
Status string
|
|
Term []Node
|
|
Details []Node
|
|
}
|
|
|
|
var unorderedListRegexp = regexp.MustCompile(`^(\s*)([+*-])(\s+(.*)|$)`)
|
|
var orderedListRegexp = regexp.MustCompile(`^(\s*)(([0-9]+|[a-zA-Z])[.)])(\s+(.*)|$)`)
|
|
var descriptiveListItemRegexp = regexp.MustCompile(`\s::(\s|$)`)
|
|
var listItemValueRegexp = regexp.MustCompile(`\[@(\d+)\]\s`)
|
|
var listItemStatusRegexp = regexp.MustCompile(`\[( |X|-)\]\s`)
|
|
|
|
func lexList(line string) (token, bool) {
|
|
if m := unorderedListRegexp.FindStringSubmatch(line); m != nil {
|
|
return token{"unorderedList", len(m[1]), m[4], m}, true
|
|
} else if m := orderedListRegexp.FindStringSubmatch(line); m != nil {
|
|
return token{"orderedList", len(m[1]), m[5], m}, true
|
|
}
|
|
return nilToken, false
|
|
}
|
|
|
|
func isListToken(t token) bool {
|
|
return t.kind == "unorderedList" || t.kind == "orderedList"
|
|
}
|
|
|
|
func listKind(t token) (string, string) {
|
|
kind := ""
|
|
switch bullet := t.matches[2]; {
|
|
case bullet == "*" || bullet == "+" || bullet == "-":
|
|
kind = "unordered"
|
|
case unicode.IsLetter(rune(bullet[0])), unicode.IsDigit(rune(bullet[0])):
|
|
kind = "ordered"
|
|
default:
|
|
panic(fmt.Sprintf("bad list bullet '%s': %#v", bullet, t))
|
|
}
|
|
if descriptiveListItemRegexp.MatchString(t.content) {
|
|
return kind, "descriptive"
|
|
}
|
|
return kind, kind
|
|
}
|
|
|
|
func (d *Document) parseList(i int, parentStop stopFn) (int, Node) {
|
|
start, lvl := i, d.tokens[i].lvl
|
|
listMainKind, kind := listKind(d.tokens[i])
|
|
list := List{Kind: kind}
|
|
stop := func(*Document, int) bool {
|
|
if parentStop(d, i) || d.tokens[i].lvl != lvl || !isListToken(d.tokens[i]) {
|
|
return true
|
|
}
|
|
itemMainKind, _ := listKind(d.tokens[i])
|
|
return itemMainKind != listMainKind
|
|
}
|
|
for !stop(d, i) {
|
|
consumed, node := d.parseListItem(list, i, parentStop)
|
|
i += consumed
|
|
list.Items = append(list.Items, node)
|
|
}
|
|
return i - start, list
|
|
}
|
|
|
|
func (d *Document) parseListItem(l List, i int, parentStop stopFn) (int, Node) {
|
|
start, nodes, bullet := i, []Node{}, d.tokens[i].matches[2]
|
|
minIndent, dterm, content, status, value := d.tokens[i].lvl+len(bullet), "", d.tokens[i].content, "", ""
|
|
originalBaseLvl := d.baseLvl
|
|
d.baseLvl = minIndent + 1
|
|
if m := listItemValueRegexp.FindStringSubmatch(content); m != nil && l.Kind == "ordered" {
|
|
value, content = m[1], content[len("[@] ")+len(m[1]):]
|
|
}
|
|
if m := listItemStatusRegexp.FindStringSubmatch(content); m != nil {
|
|
status, content = m[1], content[len("[ ] "):]
|
|
}
|
|
if l.Kind == "descriptive" {
|
|
if m := descriptiveListItemRegexp.FindStringIndex(content); m != nil {
|
|
dterm, content = content[:m[0]], content[m[1]:]
|
|
d.baseLvl = strings.Index(d.tokens[i].matches[0], " ::") + 4
|
|
}
|
|
}
|
|
|
|
d.tokens[i] = tokenize(strings.Repeat(" ", minIndent) + content)
|
|
stop := func(d *Document, i int) bool {
|
|
if parentStop(d, i) {
|
|
return true
|
|
}
|
|
t := d.tokens[i]
|
|
return t.lvl < minIndent && !(t.kind == "text" && t.content == "")
|
|
}
|
|
for !stop(d, i) && (i <= start+1 || !isSecondBlankLine(d, i)) {
|
|
consumed, node := d.parseOne(i, stop)
|
|
i += consumed
|
|
nodes = append(nodes, node)
|
|
}
|
|
d.baseLvl = originalBaseLvl
|
|
if l.Kind == "descriptive" {
|
|
return i - start, DescriptiveListItem{bullet, status, d.parseInline(dterm), nodes}
|
|
}
|
|
return i - start, ListItem{bullet, status, value, nodes}
|
|
}
|
|
|
|
func (n List) String() string { return String(n) }
|
|
func (n ListItem) String() string { return String(n) }
|
|
func (n DescriptiveListItem) String() string { return String(n) }
|