diff options
author | Andrew Boyarshin <andrew.boyarshin@gmail.com> | 2017-02-14 08:13:59 +0700 |
---|---|---|
committer | Lunny Xiao <xiaolunwen@gmail.com> | 2017-02-14 09:13:59 +0800 |
commit | dc8248f8a49e4801e119008a32b28cd2ad6e1a57 (patch) | |
tree | f3749cdfb1b4766e6f9d7d601f2de7654b747087 /modules/markdown/markdown.go | |
parent | 5cc275b1defc56d54bec23d1a5740c3fadcff2b0 (diff) | |
download | gitea-dc8248f8a49e4801e119008a32b28cd2ad6e1a57.tar.gz gitea-dc8248f8a49e4801e119008a32b28cd2ad6e1a57.zip |
Markdown rendering overhaul (#186)
* Markdown rendering overhaul
Cleaned up and squashed commits into single one.
Signed-off-by: Andrew Boyarshin <boyarshinand@gmail.com>
* Fix markdown API, add markdown module and API tests, improve code coverage
Signed-off-by: Andrew Boyarshin <boyarshinand@gmail.com>
Diffstat (limited to 'modules/markdown/markdown.go')
-rw-r--r-- | modules/markdown/markdown.go | 458 |
1 files changed, 357 insertions, 101 deletions
diff --git a/modules/markdown/markdown.go b/modules/markdown/markdown.go index c0b7c91f83..5218af8f27 100644 --- a/modules/markdown/markdown.go +++ b/modules/markdown/markdown.go @@ -8,6 +8,7 @@ import ( "bytes" "fmt" "io" + "net/url" "path" "path/filepath" "regexp" @@ -35,23 +36,17 @@ var Sanitizer = bluemonday.UGCPolicy() // This function should only be called once during entire application lifecycle. func BuildSanitizer() { // Normal markdown-stuff - Sanitizer.AllowAttrs("class").Matching(regexp.MustCompile(`[\p{L}\p{N}\s\-_',:\[\]!\./\\\(\)&]*`)).OnElements("code") + Sanitizer.AllowAttrs("class").Matching(regexp.MustCompile(`[\p{L}\p{N}\s\-_',:\[\]!\./\\\(\)&]*`)).OnElements("code", "div", "ul", "ol", "dl") // Checkboxes Sanitizer.AllowAttrs("type").Matching(regexp.MustCompile(`^checkbox$`)).OnElements("input") Sanitizer.AllowAttrs("checked", "disabled").OnElements("input") + Sanitizer.AllowNoAttrs().OnElements("label") // Custom URL-Schemes Sanitizer.AllowURLSchemes(setting.Markdown.CustomURLSchemes...) } -var validLinksPattern = regexp.MustCompile(`^[a-z][\w-]+://`) - -// isLink reports whether link fits valid format. -func isLink(link []byte) bool { - return validLinksPattern.Match(link) -} - // IsMarkdownFile reports whether name looks like a Markdown file // based on its extension. func IsMarkdownFile(name string) bool { @@ -65,7 +60,7 @@ func IsMarkdownFile(name string) bool { } // IsReadmeFile reports whether name looks like a README file -// based on its extension. +// based on its name. func IsReadmeFile(name string) bool { name = strings.ToLower(name) if len(name) < 6 { @@ -80,13 +75,6 @@ var ( // MentionPattern matches string that mentions someone, e.g. @Unknwon MentionPattern = regexp.MustCompile(`(\s|^|\W)@[0-9a-zA-Z-_\.]+`) - // CommitPattern matches link to certain commit with or without trailing hash, - // e.g. https://try.gogs.io/gogs/gogs/commit/d8a994ef243349f321568f9e36d5c3f444b99cae#diff-2 - CommitPattern = regexp.MustCompile(`(\s|^)https?.*commit/[0-9a-zA-Z]+(#+[0-9a-zA-Z-]*)?`) - - // IssueFullPattern matches link to an issue with or without trailing hash, - // e.g. https://try.gogs.io/gogs/gogs/issues/4#issue-685 - IssueFullPattern = regexp.MustCompile(`(\s|^)https?.*issues/[0-9]+(#+[0-9a-zA-Z-]*)?`) // IssueNumericPattern matches string that references to a numeric issue, e.g. #1287 IssueNumericPattern = regexp.MustCompile(`( |^|\()#[0-9]+\b`) // IssueAlphanumericPattern matches string that references to an alphanumeric issue, e.g. ABC-1234 @@ -96,11 +84,27 @@ var ( CrossReferenceIssueNumericPattern = regexp.MustCompile(`( |^)[0-9a-zA-Z]+/[0-9a-zA-Z]+#[0-9]+\b`) // Sha1CurrentPattern matches string that represents a commit SHA, e.g. d8a994ef243349f321568f9e36d5c3f444b99cae - // FIXME: this pattern matches pure numbers as well, right now we do a hack to check in RenderSha1CurrentPattern + // FIXME: this pattern matches pure numbers as well, right now we do a hack to check in renderSha1CurrentPattern // by converting string to a number. - Sha1CurrentPattern = regexp.MustCompile(`\b[0-9a-f]{40}\b`) + Sha1CurrentPattern = regexp.MustCompile(`(?:^|\s|\()[0-9a-f]{40}\b`) + + // ShortLinkPattern matches short but difficult to parse [[name|link|arg=test]] syntax + ShortLinkPattern = regexp.MustCompile(`(\[\[.*\]\]\w*)`) + + // AnySHA1Pattern allows to split url containing SHA into parts + AnySHA1Pattern = regexp.MustCompile(`http\S+//(\S+)/(\S+)/(\S+)/(\S+)/([0-9a-f]{40})(?:/?([^#\s]+)?(?:#(\S+))?)?`) + + // IssueFullPattern allows to split issue (and pull) URLs into parts + IssueFullPattern = regexp.MustCompile(`(?:^|\s|\()http\S+//((?:[^\s/]+/)+)((?:\w{1,10}-)?[1-9][0-9]*)([\?|#]\S+.(\S+)?)?\b`) + + validLinksPattern = regexp.MustCompile(`^[a-z][\w-]+://`) ) +// isLink reports whether link fits valid format. +func isLink(link []byte) bool { + return validLinksPattern.Match(link) +} + // FindAllMentions matches mention patterns in given content // and returns a list of found user names without @ prefix. func FindAllMentions(content string) []string { @@ -114,79 +118,67 @@ func FindAllMentions(content string) []string { // Renderer is a extended version of underlying render object. type Renderer struct { blackfriday.Renderer - urlPrefix string + urlPrefix string + isWikiMarkdown bool } // Link defines how formal links should be processed to produce corresponding HTML elements. func (r *Renderer) Link(out *bytes.Buffer, link []byte, title []byte, content []byte) { if len(link) > 0 && !isLink(link) { if link[0] != '#' { - link = []byte(path.Join(r.urlPrefix, string(link))) + mLink := URLJoin(r.urlPrefix, string(link)) + if r.isWikiMarkdown { + mLink = URLJoin(r.urlPrefix, "wiki", string(link)) + } + link = []byte(mLink) } } r.Renderer.Link(out, link, title, content) } -// AutoLink defines how auto-detected links should be processed to produce corresponding HTML elements. -// Reference for kind: https://github.com/russross/blackfriday/blob/master/markdown.go#L69-L76 -func (r *Renderer) AutoLink(out *bytes.Buffer, link []byte, kind int) { - if kind != blackfriday.LINK_TYPE_NORMAL { - r.Renderer.AutoLink(out, link, kind) - return +// List renders markdown bullet or digit lists to HTML +func (r *Renderer) List(out *bytes.Buffer, text func() bool, flags int) { + marker := out.Len() + if out.Len() > 0 { + out.WriteByte('\n') } - // Since this method could only possibly serve one link at a time, - // we do not need to find all. - if bytes.HasPrefix(link, []byte(setting.AppURL)) { - m := CommitPattern.Find(link) - if m != nil { - m = bytes.TrimSpace(m) - i := strings.Index(string(m), "commit/") - j := strings.Index(string(m), "#") - if j == -1 { - j = len(m) - } - out.WriteString(fmt.Sprintf(` <code><a href="%s">%s</a></code>`, m, base.ShortSha(string(m[i+7:j])))) - return - } - - m = IssueFullPattern.Find(link) - if m != nil { - m = bytes.TrimSpace(m) - i := strings.Index(string(m), "issues/") - j := strings.Index(string(m), "#") - if j == -1 { - j = len(m) - } - - issue := string(m[i+7 : j]) - fullRepoURL := setting.AppURL + strings.TrimPrefix(r.urlPrefix, "/") - var link string - if strings.HasPrefix(string(m), fullRepoURL) { - // Use a short issue reference if the URL refers to this repository - link = fmt.Sprintf(`<a href="%s">#%s</a>`, m, issue) - } else { - // Use a cross-repository issue reference if the URL refers to a different repository - repo := string(m[len(setting.AppURL) : i-1]) - link = fmt.Sprintf(`<a href="%s">%s#%s</a>`, m, repo, issue) - } - out.WriteString(link) - return - } + if flags&blackfriday.LIST_TYPE_DEFINITION != 0 { + out.WriteString("<dl>") + } else if flags&blackfriday.LIST_TYPE_ORDERED != 0 { + out.WriteString("<ol class='ui list'>") + } else { + out.WriteString("<ul class='ui list'>") + } + if !text() { + out.Truncate(marker) + return + } + if flags&blackfriday.LIST_TYPE_DEFINITION != 0 { + out.WriteString("</dl>\n") + } else if flags&blackfriday.LIST_TYPE_ORDERED != 0 { + out.WriteString("</ol>\n") + } else { + out.WriteString("</ul>\n") } - - r.Renderer.AutoLink(out, link, kind) } // ListItem defines how list items should be processed to produce corresponding HTML elements. func (r *Renderer) ListItem(out *bytes.Buffer, text []byte, flags int) { // Detect procedures to draw checkboxes. + prefix := "" + if bytes.HasPrefix(text, []byte("<p>")) { + prefix = "<p>" + } switch { - case bytes.HasPrefix(text, []byte("[ ] ")): - text = append([]byte(`<input type="checkbox" disabled="" />`), text[3:]...) - case bytes.HasPrefix(text, []byte("[x] ")): - text = append([]byte(`<input type="checkbox" disabled="" checked="" />`), text[3:]...) + case bytes.HasPrefix(text, []byte(prefix+"[ ] ")): + text = append([]byte(`<div class="ui fitted disabled checkbox"><input type="checkbox" disabled="disabled" /><label /></div>`), text[3+len(prefix):]...) + case bytes.HasPrefix(text, []byte(prefix+"[x] ")): + text = append([]byte(`<div class="ui checked fitted disabled checkbox"><input type="checkbox" checked="" disabled="disabled" /><label /></div>`), text[3+len(prefix):]...) + } + if prefix != "" { + text = bytes.Replace(text, []byte("</p>"), []byte{}, 1) } r.Renderer.ListItem(out, text, flags) } @@ -196,15 +188,15 @@ func (r *Renderer) ListItem(out *bytes.Buffer, text []byte, flags int) { var ( svgSuffix = []byte(".svg") svgSuffixWithMark = []byte(".svg?") - spaceBytes = []byte(" ") - spaceEncodedBytes = []byte("%20") - space = " " - spaceEncoded = "%20" ) // Image defines how images should be processed to produce corresponding HTML elements. func (r *Renderer) Image(out *bytes.Buffer, link []byte, title []byte, alt []byte) { - prefix := strings.Replace(r.urlPrefix, "/src/", "/raw/", 1) + prefix := r.urlPrefix + if r.isWikiMarkdown { + prefix = URLJoin(prefix, "wiki", "src") + } + prefix = strings.Replace(prefix, "/src/", "/raw/", 1) if len(link) > 0 { if isLink(link) { // External link with .svg suffix usually means CI status. @@ -215,10 +207,11 @@ func (r *Renderer) Image(out *bytes.Buffer, link []byte, title []byte, alt []byt } } else { if link[0] != '/' { - prefix += "/" + if !strings.HasSuffix(prefix, "/") { + prefix += "/" + } } - link = bytes.Replace([]byte((prefix + string(link))), spaceBytes, spaceEncodedBytes, -1) - fmt.Println(333, string(link)) + link = []byte(url.QueryEscape(prefix + string(link))) } } @@ -247,6 +240,19 @@ func cutoutVerbosePrefix(prefix string) string { return prefix } +// URLJoin joins url components, like path.Join, but preserving contents +func URLJoin(elem ...string) string { + res := "" + last := len(elem) - 1 + for i, item := range elem { + res += item + if !strings.HasSuffix(res, "/") && i != last { + res += "/" + } + } + return res +} + // RenderIssueIndexPattern renders issue indexes to corresponding links. func RenderIssueIndexPattern(rawBytes []byte, urlPrefix string, metas map[string]string) []byte { urlPrefix = cutoutVerbosePrefix(urlPrefix) @@ -263,7 +269,7 @@ func RenderIssueIndexPattern(rawBytes []byte, urlPrefix string, metas map[string } var link string if metas == nil { - link = fmt.Sprintf(`<a href="%s/issues/%s">%s</a>`, urlPrefix, m[1:], m) + link = fmt.Sprintf(`<a href="%s">%s</a>`, URLJoin(urlPrefix, "issues", string(m[1:])), m) } else { // Support for external issue tracker if metas["style"] == IssueNameStyleAlphanumeric { @@ -278,6 +284,238 @@ func RenderIssueIndexPattern(rawBytes []byte, urlPrefix string, metas map[string return rawBytes } +// IsSameDomain checks if given url string has the same hostname as current Gitea instance +func IsSameDomain(s string) bool { + if uapp, err := url.Parse(setting.AppURL); err == nil { + if u, err := url.Parse(s); err == nil { + return u.Host == uapp.Host + } + return false + } + return false +} + +// renderFullSha1Pattern renders SHA containing URLs +func renderFullSha1Pattern(rawBytes []byte, urlPrefix string) []byte { + ms := AnySHA1Pattern.FindAllSubmatch(rawBytes, -1) + for _, m := range ms { + all := m[0] + paths := string(m[1]) + var path = "//" + paths + author := string(m[2]) + repoName := string(m[3]) + path = URLJoin(path, author, repoName) + ltype := "src" + itemType := m[4] + if IsSameDomain(paths) { + ltype = string(itemType) + } else if string(itemType) == "commit" { + ltype = "commit" + } + sha := m[5] + var subtree string + if len(m) > 6 && len(m[6]) > 0 { + subtree = string(m[6]) + } + var line []byte + if len(m) > 7 && len(m[7]) > 0 { + line = m[7] + } + urlSuffix := "" + text := base.ShortSha(string(sha)) + if subtree != "" { + urlSuffix = "/" + subtree + text += urlSuffix + } + if line != nil { + value := string(line) + urlSuffix += "#" + urlSuffix += value + text += " (" + text += value + text += ")" + } + rawBytes = bytes.Replace(rawBytes, all, []byte(fmt.Sprintf( + `<a href="%s">%s</a>`, URLJoin(path, ltype, string(sha))+urlSuffix, text)), -1) + } + return rawBytes +} + +// renderFullIssuePattern renders issues-like URLs +func renderFullIssuePattern(rawBytes []byte, urlPrefix string) []byte { + ms := IssueFullPattern.FindAllSubmatch(rawBytes, -1) + for _, m := range ms { + all := m[0] + paths := bytes.Split(m[1], []byte("/")) + paths = paths[:len(paths)-1] + if bytes.HasPrefix(paths[0], []byte("gist.")) { + continue + } + var path string + if len(paths) > 3 { + // Internal one + path = URLJoin(urlPrefix, "issues") + } else { + path = "//" + string(m[1]) + } + id := string(m[2]) + path = URLJoin(path, id) + var comment []byte + if len(m) > 3 { + comment = m[3] + } + urlSuffix := "" + text := "#" + id + if comment != nil { + urlSuffix += string(comment) + text += " <i class='comment icon'></i>" + } + rawBytes = bytes.Replace(rawBytes, all, []byte(fmt.Sprintf( + `<a href="%s%s">%s</a>`, path, urlSuffix, text)), -1) + } + return rawBytes +} + +func firstIndexOfByte(sl []byte, target byte) int { + for i := 0; i < len(sl); i++ { + if sl[i] == target { + return i + } + } + return -1 +} + +func lastIndexOfByte(sl []byte, target byte) int { + for i := len(sl) - 1; i >= 0; i-- { + if sl[i] == target { + return i + } + } + return -1 +} + +// renderShortLinks processes [[syntax]] +func renderShortLinks(rawBytes []byte, urlPrefix string, noLink bool) []byte { + ms := ShortLinkPattern.FindAll(rawBytes, -1) + for _, m := range ms { + orig := bytes.TrimSpace(m) + m = orig[2:] + tailPos := lastIndexOfByte(m, ']') + 1 + tail := []byte{} + if tailPos < len(m) { + tail = m[tailPos:] + m = m[:tailPos-1] + } + m = m[:len(m)-2] + props := map[string]string{} + + // MediaWiki uses [[link|text]], while GitHub uses [[text|link]] + // It makes page handling terrible, but we prefer GitHub syntax + // And fall back to MediaWiki only when it is obvious from the look + // Of text and link contents + sl := bytes.Split(m, []byte("|")) + for _, v := range sl { + switch bytes.Count(v, []byte("=")) { + + // Piped args without = sign, these are mandatory arguments + case 0: + { + sv := string(v) + if props["name"] == "" { + if isLink(v) { + // If we clearly see it is a link, we save it so + + // But first we need to ensure, that if both mandatory args provided + // look like links, we stick to GitHub syntax + if props["link"] != "" { + props["name"] = props["link"] + } + + props["link"] = strings.TrimSpace(sv) + } else { + props["name"] = sv + } + } else { + props["link"] = strings.TrimSpace(sv) + } + } + + // Piped args with = sign, these are optional arguments + case 1: + { + sep := firstIndexOfByte(v, '=') + key, val := string(v[:sep]), html.UnescapeString(string(v[sep+1:])) + lastCharIndex := len(val) - 1 + if (val[0] == '"' || val[0] == '\'') && (val[lastCharIndex] == '"' || val[lastCharIndex] == '\'') { + val = val[1:lastCharIndex] + } + props[key] = val + } + } + } + + var name string + var link string + if props["link"] != "" { + link = props["link"] + } else if props["name"] != "" { + link = props["name"] + } + if props["title"] != "" { + name = props["title"] + } else if props["name"] != "" { + name = props["name"] + } else { + name = link + } + + name += string(tail) + image := false + ext := filepath.Ext(string(link)) + if ext != "" { + switch ext { + case ".jpg", ".jpeg", ".png", ".tif", ".tiff", ".webp", ".gif", ".bmp", ".ico", ".svg": + { + image = true + } + } + } + absoluteLink := isLink([]byte(link)) + if !absoluteLink { + link = url.QueryEscape(link) + } + if image { + if !absoluteLink { + link = URLJoin(urlPrefix, "wiki", "raw", link) + } + title := props["title"] + if title == "" { + title = props["alt"] + } + if title == "" { + title = path.Base(string(name)) + } + alt := props["alt"] + if alt == "" { + alt = name + } + if alt != "" { + alt = `alt="` + alt + `"` + } + name = fmt.Sprintf(`<img src="%s" %s title="%s" />`, link, alt, title) + } else if !absoluteLink { + link = URLJoin(urlPrefix, "wiki", link) + } + if noLink { + rawBytes = bytes.Replace(rawBytes, orig, []byte(name), -1) + } else { + rawBytes = bytes.Replace(rawBytes, orig, + []byte(fmt.Sprintf(`<a href="%s">%s</a>`, link, name)), -1) + } + } + return rawBytes +} + // RenderCrossReferenceIssueIndexPattern renders issue indexes from other repositories to corresponding links. func RenderCrossReferenceIssueIndexPattern(rawBytes []byte, urlPrefix string, metas map[string]string) []byte { ms := CrossReferenceIssueNumericPattern.FindAll(rawBytes, -1) @@ -289,20 +527,24 @@ func RenderCrossReferenceIssueIndexPattern(rawBytes []byte, urlPrefix string, me repo := string(bytes.Split(m, []byte("#"))[0]) issue := string(bytes.Split(m, []byte("#"))[1]) - link := fmt.Sprintf(`<a href="%s%s/issues/%s">%s</a>`, setting.AppURL, repo, issue, m) + link := fmt.Sprintf(`<a href="%s">%s</a>`, URLJoin(urlPrefix, repo, "issues", issue), m) rawBytes = bytes.Replace(rawBytes, m, []byte(link), 1) } return rawBytes } -// RenderSha1CurrentPattern renders SHA1 strings to corresponding links that assumes in the same repository. -func RenderSha1CurrentPattern(rawBytes []byte, urlPrefix string) []byte { - return []byte(Sha1CurrentPattern.ReplaceAllStringFunc(string(rawBytes[:]), func(m string) string { - if com.StrTo(m).MustInt() > 0 { - return m +// renderSha1CurrentPattern renders SHA1 strings to corresponding links that assumes in the same repository. +func renderSha1CurrentPattern(rawBytes []byte, urlPrefix string) []byte { + ms := Sha1CurrentPattern.FindAllSubmatch(rawBytes, -1) + for _, m := range ms { + all := m[0] + if com.StrTo(all).MustInt() > 0 { + continue } - return fmt.Sprintf(`<a href="%s/commit/%s"><code>%s</code></a>`, urlPrefix, m, base.ShortSha(m)) - })) + rawBytes = bytes.Replace(rawBytes, all, []byte(fmt.Sprintf( + `<a href="%s">%s</a>`, URLJoin(urlPrefix, "commit", string(all)), base.ShortSha(string(all)))), -1) + } + return rawBytes } // RenderSpecialLink renders mentions, indexes and SHA1 strings to corresponding links. @@ -311,23 +553,27 @@ func RenderSpecialLink(rawBytes []byte, urlPrefix string, metas map[string]strin for _, m := range ms { m = m[bytes.Index(m, []byte("@")):] rawBytes = bytes.Replace(rawBytes, m, - []byte(fmt.Sprintf(`<a href="%s/%s">%s</a>`, setting.AppSubURL, m[1:], m)), -1) + []byte(fmt.Sprintf(`<a href="%s">%s</a>`, URLJoin(setting.AppURL, string(m[1:])), m)), -1) } + rawBytes = renderShortLinks(rawBytes, urlPrefix, false) rawBytes = RenderIssueIndexPattern(rawBytes, urlPrefix, metas) rawBytes = RenderCrossReferenceIssueIndexPattern(rawBytes, urlPrefix, metas) - rawBytes = RenderSha1CurrentPattern(rawBytes, urlPrefix) + rawBytes = renderFullSha1Pattern(rawBytes, urlPrefix) + rawBytes = renderSha1CurrentPattern(rawBytes, urlPrefix) + rawBytes = renderFullIssuePattern(rawBytes, urlPrefix) return rawBytes } // RenderRaw renders Markdown to HTML without handling special links. -func RenderRaw(body []byte, urlPrefix string) []byte { +func RenderRaw(body []byte, urlPrefix string, wikiMarkdown bool) []byte { htmlFlags := 0 htmlFlags |= blackfriday.HTML_SKIP_STYLE htmlFlags |= blackfriday.HTML_OMIT_CONTENTS renderer := &Renderer{ - Renderer: blackfriday.HtmlRenderer(htmlFlags, "", ""), - urlPrefix: urlPrefix, + Renderer: blackfriday.HtmlRenderer(htmlFlags, "", ""), + urlPrefix: urlPrefix, + isWikiMarkdown: wikiMarkdown, } // set up the parser @@ -335,9 +581,7 @@ func RenderRaw(body []byte, urlPrefix string) []byte { extensions |= blackfriday.EXTENSION_NO_INTRA_EMPHASIS extensions |= blackfriday.EXTENSION_TABLES extensions |= blackfriday.EXTENSION_FENCED_CODE - extensions |= blackfriday.EXTENSION_AUTOLINK extensions |= blackfriday.EXTENSION_STRIKETHROUGH - extensions |= blackfriday.EXTENSION_SPACE_HEADERS extensions |= blackfriday.EXTENSION_NO_EMPTY_LINE_BEFORE_BLOCK if setting.Markdown.EnableHardLineBreak { @@ -379,10 +623,12 @@ OUTER_LOOP: token = tokenizer.Token() // Copy the token to the output verbatim - buf.WriteString(token.String()) + buf.Write(renderShortLinks([]byte(token.String()), urlPrefix, true)) if token.Type == html.StartTagToken { - stackNum++ + if !com.IsSliceContainsStr(noEndTags, token.Data) { + stackNum++ + } } // If this is the close tag to the outer-most, we are done @@ -425,16 +671,26 @@ OUTER_LOOP: return rawHTML } -// Render renders Markdown to HTML with special links. -func Render(rawBytes []byte, urlPrefix string, metas map[string]string) []byte { - urlPrefix = strings.Replace(urlPrefix, space, spaceEncoded, -1) - result := RenderRaw(rawBytes, urlPrefix) +// Render renders Markdown to HTML with all specific handling stuff. +func render(rawBytes []byte, urlPrefix string, metas map[string]string, isWikiMarkdown bool) []byte { + urlPrefix = strings.Replace(urlPrefix, " ", "%20", -1) + result := RenderRaw(rawBytes, urlPrefix, isWikiMarkdown) result = PostProcess(result, urlPrefix, metas) result = Sanitizer.SanitizeBytes(result) return result } +// Render renders Markdown to HTML with all specific handling stuff. +func Render(rawBytes []byte, urlPrefix string, metas map[string]string) []byte { + return render(rawBytes, urlPrefix, metas, false) +} + // RenderString renders Markdown to HTML with special links and returns string type. func RenderString(raw, urlPrefix string, metas map[string]string) string { - return string(Render([]byte(raw), urlPrefix, metas)) + return string(render([]byte(raw), urlPrefix, metas, false)) +} + +// RenderWiki renders markdown wiki page to HTML and return HTML string +func RenderWiki(rawBytes []byte, urlPrefix string, metas map[string]string) string { + return string(render(rawBytes, urlPrefix, metas, true)) } |