diff options
| -rw-r--r-- | .gitignore | 4 | ||||
| -rw-r--r-- | makefile | 6 | ||||
| -rw-r--r-- | misc/util/d/tools/markup_conversion/README | 1 | ||||
| -rwxr-xr-x | misc/util/d/tools/markup_conversion/endnotes_inline_from_binary.d (renamed from misc/util/d/tools/endnotes_inline_from_binary.d) | 16 | ||||
| -rw-r--r-- | misc/util/d/tools/markup_conversion/markup_changes.d | 136 | ||||
| -rwxr-xr-x | misc/util/d/tools/markup_conversion/markup_changes_header_and_content.d | 244 | ||||
| -rwxr-xr-x | misc/util/d/tools/markup_conversion/markup_conversion_from_sisu_ruby_to_sisu_spine.d | 354 | ||||
| -rw-r--r-- | org/spine_build_scaffold.org | 10 | ||||
| -rw-r--r-- | org/util_spine_markup_conversion_from_sisu.org | 951 | ||||
| -rw-r--r-- | org/util_spine_syntax_highlighting_emacs.org (renamed from org/spine_syntax_highlighting_emacs.org) | 0 | ||||
| -rw-r--r-- | org/util_spine_syntax_highlighting_vim.org (renamed from org/spine_syntax_highlighting_vim.org) | 0 | 
11 files changed, 1702 insertions, 20 deletions
| @@ -21,10 +21,6 @@  !org  !misc  !misc/** -!util -!util/** -!editor-syntax-etc -!editor-syntax-etc/**  !ext_lib  !ext_lib/**  !src @@ -162,7 +162,8 @@ skel:  	mkdir -p build; \  	mkdir -p views; \  	mkdir -p data; \ -	mkdir -p misc/util; \ +	mkdir -p misc/util/d/cgi/search/cgi-bin/src; \ +	mkdir -p misc/util/d/tools/markup_conversion; \  	mkdir -p misc/editor-syntax-etc/emacs; \  	mkdir -p misc/editor-syntax-etc/vim/syntax; \  	mkdir -p misc/editor-syntax-etc/vim/colors; \ @@ -196,7 +197,8 @@ expunge:  distclean: expunge  distclean_and_init: expunge  	mkdir -p views; \ -	mkdir -p misc/util; \ +	mkdir -p misc/util/d/cgi/search/cgi-bin/src; \ +	mkdir -p misc/util/d/tools/markup_conversion; \  	mkdir -p misc/editor-syntax-etc/emacs; \  	mkdir -p misc/editor-syntax-etc/vim/syntax; \  	mkdir -p misc/editor-syntax-etc/vim/colors; \ diff --git a/misc/util/d/tools/markup_conversion/README b/misc/util/d/tools/markup_conversion/README new file mode 100644 index 0000000..8b13789 --- /dev/null +++ b/misc/util/d/tools/markup_conversion/README @@ -0,0 +1 @@ + diff --git a/misc/util/d/tools/endnotes_inline_from_binary.d b/misc/util/d/tools/markup_conversion/endnotes_inline_from_binary.d index 4d9ceb0..abd4e45 100755 --- a/misc/util/d/tools/endnotes_inline_from_binary.d +++ b/misc/util/d/tools/markup_conversion/endnotes_inline_from_binary.d @@ -29,6 +29,13 @@ import std.regex;  import std.format;  import std.conv;  void main(string[] args) { +  static comment                 = ctRegex!(`^%+ `); +  static block_tic_code_open     = ctRegex!("^`{3} code(?:[.](?P<syntax>[a-z][0-9a-z#+_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?"); +  static block_tic_close         = ctRegex!("^(`{3})$","m"); +  static block_curly_code_open   = ctRegex!(`^(?:code(?:[.](?P<syntax>[a-z][0-9a-z_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?[{][ ]*$)`); +  static block_curly_code_close  = ctRegex!(`^([}]code)`); +  auto rgx_endnote_ref           = ctRegex!(`([~]\^)(?P<tail>[)\]]? |$)`, "gm"); +  auto rgx_endnote               = ctRegex!(`^\^~\s+(.+|\n)`, "gm");    foreach(arg; args[1..$]) {      if (        !(arg.match(regex(r"--\w+"))) @@ -45,13 +52,6 @@ void main(string[] args) {            "curly_code"                 : 0,            "tic_code"                   : 0,          ]; -        static comment                 = ctRegex!(`^%+ `); -        static block_tic_code_open     = ctRegex!("^`{3} code(?:[.](?P<syntax>[a-z][0-9a-z#+_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?"); -        static block_tic_close         = ctRegex!("^(`{3})$","m"); -        static block_curly_code_open   = ctRegex!(`^(?:code(?:[.](?P<syntax>[a-z][0-9a-z_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?[{][ ]*$)`); -        static block_curly_code_close  = ctRegex!(`^([}]code)`); -        auto rgx_endnote_ref           = ctRegex!(`([~]\^)(?P<tail>[)\]]? |$)`, "gm"); -        auto rgx_endnote               = ctRegex!(`^\^~\s+(.+|\n)`, "gm");          foreach (paragraph; paragraphs) { /+ loop to gather binary endnotes +/            if ( !( type["curly_code"] == 1 || type["tic_code"] == 1)              && paragraph.match(rgx_endnote) @@ -68,7 +68,7 @@ void main(string[] args) {              ) {                type["tic_code"] = 0;              } else if ( type["curly_code"] == 1 || type["tic_code"] == 1) { -              // prevent search for endnotes +              // skip, prevent search for endnotes              } else if (paragraph.matchFirst(block_curly_code_open)) {                type["curly_code"] = 1;              } else if (paragraph.matchFirst(block_tic_code_open)) { diff --git a/misc/util/d/tools/markup_conversion/markup_changes.d b/misc/util/d/tools/markup_conversion/markup_changes.d new file mode 100644 index 0000000..4274f78 --- /dev/null +++ b/misc/util/d/tools/markup_conversion/markup_changes.d @@ -0,0 +1,136 @@ +#!/usr/bin/env rdmd +/+ +  - read in file .sst .ssi .ssm +  - loop twice +    - first +      - check for and skip code blocks +      - use unique code marker for endnote markers in text and give an endnote +        number ★1, increment +      - extract all endnotes in array +    - second +      - check that the footnote marker number count matches the number of notes +        in the array +        - if they match either: +            - substitute each endnote marker with the array footnote[number-1] +            - substitute each endnote marker with footnote +              as inlined footnote markup (footnote number not needed) +        - if they do not match exit +  - check whether changes have been made +    - if so write file with inline footnotes in sub-directory converted_output_/ +      using the same name as the original file +    - else, exit ++/ +import std.stdio; +import std.file; +import std.array : split; +import std.exception; +import core.stdc.errno; +import std.regex; +import std.format; +import std.conv; +void main(string[] args) { +  static comment                 = ctRegex!(`^%+ `); +  static block_tic_code_open     = ctRegex!("^`{3} code(?:[.](?P<syntax>[a-z][0-9a-z#+_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?"); +  static block_tic_close         = ctRegex!("^(`{3})$","m"); +  static block_curly_code_open   = ctRegex!(`^(?:code(?:[.](?P<syntax>[a-z][0-9a-z_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?[{][ ]*$)`); +  static block_curly_code_close  = ctRegex!(`^([}]code)`); +  auto rgx_endnote_ref           = ctRegex!(`([~]\^)(?P<tail>[)\]]? |$)`, "gm"); +  auto rgx_endnote               = ctRegex!(`^\^~\s+(.+|\n)`, "gm"); +  foreach(arg; args[1..$]) { +    if ( +      !(arg.match(regex(r"--\w+"))) +      && arg.match(regex(r"\w+?\.ss[itm]")) +    ) { +      writeln(arg); +      string filename                  = arg; +      try { +        string[] contents, endnotes, endnote_refs; +        string text                    = filename.readText; +        string[] paragraphs            = text.split("\n\n"); +        int endnote_ref_count          = 0; +        int[string] type = [ +          "curly_code"                 : 0, +          "tic_code"                   : 0, +        ]; +        foreach (paragraph; paragraphs) { /+ loop to gather binary endnotes +/ +          if ( !( type["curly_code"] == 1 || type["tic_code"] == 1) +            && paragraph.match(rgx_endnote) +          ) { +            endnotes ~= replaceAll!(m => m[1]) +              (paragraph, rgx_endnote); +          } else { +            if ( type["curly_code"] == 1 || type["tic_code"] == 1 +              || paragraph.matchFirst(block_curly_code_open) +              || paragraph.matchFirst(block_tic_code_open) +            ) { /+ code blocks identified, no munging +/ +              if ( type["curly_code"] == 1 +                && paragraph.matchFirst(block_curly_code_close) +              ) { +                type["curly_code"] = 0; +              } else if (type["tic_code"] == 1 +                && paragraph.matchFirst(block_tic_close) +              ) { +                type["tic_code"] = 0; +              } else if (paragraph.matchFirst(block_curly_code_open)) { +                type["curly_code"] = 1; +              } else if (paragraph.matchFirst(block_tic_code_open)) { +                type["tic_code"] = 1; +              } +              contents ~= paragraph; +            } else { /+ regular content, not a code block +/ +              if (auto m = paragraph.matchAll(rgx_endnote_ref)) { +                foreach (n; m) { +                  endnote_ref_count++; // endnote_refs ~= (n.captures[1]); +                } +              } +              paragraph = replaceAll!(m => " \\\\ " ) +                (paragraph, regex(r"\s*<(?:/\s*|:)?br>\s*")); // (paragraph, regex(r"(<br>)")); +              contents ~= paragraph; +            } +          } +        } +        { +          import std.outbuffer; +          auto buffer = new OutBuffer(); +          if (endnotes.length == endnote_ref_count) { +            // writeln("endnote ref count:         ", endnote_ref_count); +            // writeln("number of binary endnotes: ", endnotes.length); +            int endnote_count = -1; +            foreach (content; contents) { /+ loop to inline endnotes +/ +              content = replaceAll!(m => "~{ " ~ endnotes[++endnote_count] ~ " }~" ~ m["tail"] ) +                (content, rgx_endnote_ref); // endnote_ref cannot occur in a code block or else fail +              buffer.write(content ~ "\n\n"); +            } +            if (buffer) { +              try { +                string dir_out = "converted_output_"; +                string path_and_file_out = dir_out ~ "/" ~ filename; +                dir_out.mkdirRecurse; +                auto f = File(path_and_file_out, "w"); +                f.write(buffer); +                writeln("wrote: ", path_and_file_out); +              } catch (FileException ex) { +                writeln("did not write file"); +                // Handle errors +              } +            } +          } else { +            foreach (content; contents) { /+ loop to inline endnotes +/ +              buffer.write(content ~ "\n\n"); +            } +          } +        } +      } catch (ErrnoException ex) { +        switch(ex.errno) { +          case EPERM: +          case EACCES: // Permission denied +            break; +          case ENOENT: // File does not exist +            break; +          default:     // Handle other errors +            break; +        } +      } +    } +  } +} diff --git a/misc/util/d/tools/markup_conversion/markup_changes_header_and_content.d b/misc/util/d/tools/markup_conversion/markup_changes_header_and_content.d new file mode 100755 index 0000000..86792ff --- /dev/null +++ b/misc/util/d/tools/markup_conversion/markup_changes_header_and_content.d @@ -0,0 +1,244 @@ +#!/usr/bin/env rdmd +/+ +  - read in file .sst .ssi .ssm +  - loop twice +    - first +      - check for and skip code blocks +      - use unique code marker for endnote markers in text and give an endnote +        number ★1, increment +      - extract all endnotes in array +    - second +      - check that the footnote marker number count matches the number of notes +        in the array +        - if they match either: +            - substitute each endnote marker with the array footnote[number-1] +            - substitute each endnote marker with footnote +              as inlined footnote markup (footnote number not needed) +        - if they do not match exit +  - check whether changes have been made +    - if so write file with inline footnotes in sub-directory converted_output_/ +      using the same name as the original file +    - else, exit ++/ +import std.stdio; +import std.file; +import std.array : split; +import std.exception; +// import std.range; +import core.stdc.errno; +import std.regex; +import std.format; +import std.conv; +void main(string[] args) { +  static heading_a               = ctRegex!(`^:?[A][~] `, "m"); +  static comment                 = ctRegex!(`^%+ `); +  static block_tic_code_open     = ctRegex!("^`{3} code(?:[.](?P<syntax>[a-z][0-9a-z#+_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?"); +  static block_tic_close         = ctRegex!("^(`{3})$","m"); +  static block_curly_code_open   = ctRegex!(`^(?:code(?:[.](?P<syntax>[a-z][0-9a-z_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?[{][ ]*$)`); +  static block_curly_code_close  = ctRegex!(`^([}]code)`); +  auto rgx_endnote_ref           = ctRegex!(`([~]\^)(?P<tail>[)\]]? |$)`, "gm"); +  auto rgx_endnote               = ctRegex!(`^\^~\s+(.+|\n)`, "gm"); +  char[][] header0Content1(in string src_text) { // cast(char[]) +    /+ split string on _first_ match of "^:?A~\s" into [header, content] array/tuple +/ +    char[][] header_and_content; +    auto m = (cast(char[]) src_text).matchFirst(heading_a); +    header_and_content ~= m.pre; +    header_and_content ~= m.hit ~ m.post; +    assert(header_and_content.length == 2, +      "document markup is broken, header body split == " +      ~ header_and_content.length.to!string +      ~ "; (header / body array split should == 2 (split is on level A~))" +    ); +    return header_and_content; +  } +  foreach(arg; args[1..$]) { +    if ( +      !(arg.match(regex(r"--\w+"))) +      && arg.match(regex(r"\w+?\.ss[itm]")) +    ) { +      writeln(arg); +      string filename                  = arg; +      try { +        string[] munged_header, munged_contents, munged_endnotes, endnote_refs; +        string text                    = filename.readText; +        char[][] hc                    = header0Content1(text); +        char[] src_header              = hc[0]; +        string[] headers               = src_header.to!string.split("\n\n"); +        char[] src_txt                 = hc[1]; +        string[] paragraphs            = src_txt.to!string.split("\n\n"); +        int endnote_ref_count          = 0; +        int[string] type = [ +          "curly_code"                 : 0, +          "tic_code"                   : 0, +        ]; +        string _tmp_header; +        foreach (h_; headers) {                                                                                          /+ loop to inline endnotes +/ +          _tmp_header = ""; +          if (h_.match(regex(r"^[@\[]?title[:\]]?"))) {                                                                  // title +            if (auto m = h_.match(regex(r"^@title:(?:\s+(?P<c>.+)|$)"))) {                                               // sisu bespoke markup +              if (m.captures["c"].length == 0) { +                _tmp_header ~= "title:"; +              } else { +                _tmp_header ~= "title:\n  main: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } +            } else if (auto m = h_.match(regex(r"^title\s*=\s*(?P<c>.+)"))) {                                            // toml? +              if (m.captures["c"].length == 0) { +                _tmp_header ~= "title:"; +              } else { +                _tmp_header ~= "title:\n  main: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } +            } else if (auto m = h_.match(regex(r"^\[title\]"))) {                                                        // toml markup +              _tmp_header ~= "title:"; +            } else if (auto m = h_.match(regex(r"^title(?:\s+(?P<c>.+)|\s+\\$)"))) {                                     // sdlang markup +              if (m.captures["c"].length == 0) { +                _tmp_header ~= "title:"; +              } else { +                _tmp_header ~= "title:\n  main: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } +            } +            if (h_.match(regex(r"^\s*[:]?(?:main)[:= ]?", "m"))) { +              if (auto m = h_.match(regex(r"^\s+(?P<h>:main):(?:\s+(?P<c>.+)|$)", "m"))) {                                // sisu bespoke markup +                _tmp_header ~= "  main: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } else if (auto m = h_.match(regex(r"^\s*(?P<h>main)\s*=\s*(?P<c>.+)", "m"))) {                            // toml? +                _tmp_header ~= "  main: " ~ m.captures["c"]; +              } else if (auto m = h_.match(regex(r"^\s+(?P<h>main)(?:\s*\s*(?P<c>.+)|$)", "m"))) {                       // toml markup +                _tmp_header ~= "  main: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } else if (auto m = h_.match(regex(r"^\s+(?P<h>main)(?:\s+(?P<c>.+)|\s+\\$)", "m"))) {                     // sdlang markup +                _tmp_header ~= "  main: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } +            } +            if (h_.match(regex(r"^\s*[:]?(?:sub(title)?)[:= ]?", "m"))) { +              if (auto m = h_.match(regex(r"^\s+:sub(?:title)?:(?:\s+(?P<c>.+)|$)", "m"))) {                             // sisu bespoke markup +                _tmp_header ~= "  subtitle: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } else if (auto m = h_.match(regex(r"^\s*sub(?:title)?\s*=\s*(?P<c>.+)$", "m"))) {                         // toml? +                _tmp_header ~= "  subtitle: " ~ m.captures["c"]; +              } else if (auto m = h_.match(regex(r"^\s+(?:title)?(?:\s*\s*(?P<c>.+)|$)", "m"))) {                        // toml markup +                _tmp_header ~= "  subtitle: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } else if (auto m = h_.match(regex(r"^\s+(?:title)?(?:\s+(?P<c>.+)|\s+\\$)", "m"))) {                      // sdlang markup +                _tmp_header ~= "  subtitle: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } +            } +          } +          if (h_.match(regex(r"^[@\[]?rights[:\]]?"))) {                                                                 // rights +            if (auto m = h_.match(regex(r"^@rights:[ ]+(?P<c>.+)$"))) {                                                  // sisu bespoke markup +              _tmp_header ~= "rights: \n  copyright: \"" ~ m.captures["c"] ~ "\""; +            } else if (auto m = h_.match(regex(r"^@rights:"))) {                                                         // sisu bespoke markup +              _tmp_header ~= "rights:"; +            } else if (auto m = h_.match(regex(r"^\[rights\]", "m"))) {                                                  // toml markup +              _tmp_header ~= "rights:"; +            } else if (auto m = h_.match(regex(r"^rights:"))) {                                                          // sdlang markup +              _tmp_header ~= "rights:"; +            } +            if (h_.match(regex(r"^\s*[:]?copyright[:= ]?", "m"))) { +              if (auto m = h_.match(regex(r"^\s+:copyright:(?:\s+(?P<c>.+)|$)", "m"))) {                                 // sisu bespoke markup +                _tmp_header ~= "  copyright: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } else if (auto m = h_.match(regex(r"^\s*copyright\s*=\s*(?P<c>.+)", "m"))) {                              // toml? +                _tmp_header ~= "  copyright: " ~ m.captures["c"]; +              } else if (auto m = h_.match(regex(r"^\s+<h>copyright(?:\s*\s*(?P<c>.+)|$)", "m"))) {                      // toml markup +                _tmp_header ~= "  copyright: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } else if (auto m = h_.match(regex(r"^\s+copyright(?:\s+(?P<c>.+)|\s+\\$)", "m"))) {                       // sdlang markup +                _tmp_header ~= "  copyright: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } +            } +            if (h_.match(regex(r"^\s*[:]?licen[cs]e[:= ]?", "m"))) { +              if (auto m = h_.match(regex(r"^\s+:licen[cs]e:(?:\s+(?P<c>.+)|$)", "m"))) {                                // sisu bespoke markup +                _tmp_header ~= "  license: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } else if (auto m = h_.match(regex(r"^\s*licen[cs]e\s*=\s*(?P<c>.+)$", "m"))) {                            // toml? +                _tmp_header ~= "  license: " ~ m.captures["c"]; +              } else if (auto m = h_.match(regex(r"^\s+licen[cs]e(?:\s*\s*(?P<c>.+)|$)", "m"))) {                        // toml markup +                _tmp_header ~= "  license: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } else if (auto m = h_.match(regex(r"^\s+licen[cs]e(?:\s+(?P<c>.+)|\s+\\$)", "m"))) {                      // sdlang markup +                _tmp_header ~= "  license: " ~ "\"" ~ m.captures["c"] ~ "\""; +              } +            } +          } +          if (_tmp_header.length > 0) { +            munged_header ~= _tmp_header; +          } else { +            munged_header ~= h_; +          } +        } +        writeln(munged_header); +        foreach (paragraph; paragraphs) {                                                                                  /+ loop to gather binary endnotes +/ +          if ( !( type["curly_code"] == 1 || type["tic_code"] == 1) +            && paragraph.match(rgx_endnote) +          ) { +            munged_endnotes ~= replaceAll!(m => m[1]) +              (paragraph, rgx_endnote); +          } else { +            if ( type["curly_code"] == 1 || type["tic_code"] == 1 +              || paragraph.matchFirst(block_curly_code_open) +              || paragraph.matchFirst(block_tic_code_open) +            ) { /+ code blocks identified, no munging +/ +              if ( type["curly_code"] == 1 +                && paragraph.matchFirst(block_curly_code_close) +              ) { +                type["curly_code"] = 0; +              } else if (type["tic_code"] == 1 +                && paragraph.matchFirst(block_tic_close) +              ) { +                type["tic_code"] = 0; +              } else if (paragraph.matchFirst(block_curly_code_open)) { +                type["curly_code"] = 1; +              } else if (paragraph.matchFirst(block_tic_code_open)) { +                type["tic_code"] = 1; +              } +              munged_contents ~= paragraph; +            } else { /+ regular content, not a code block +/ +              if (auto m = paragraph.matchAll(rgx_endnote_ref)) { +                foreach (n; m) { +                  endnote_ref_count++; // endnote_refs ~= (n.captures[1]); +                } +              } +              paragraph = replaceAll!(m => " \\\\ " ) +                (paragraph, regex(r"\s*<(?:/\s*|:)?br>\s*")); // (paragraph, regex(r"(<br>)")); +              munged_contents ~= paragraph; +            } +          } +        } +        { +          import std.outbuffer; +          auto buffer = new OutBuffer(); +          foreach (header; munged_header) { /+ loop to inline endnotes +/ +            buffer.write(header ~ "\n\n"); +          } +          if (munged_endnotes.length == endnote_ref_count) { +            int endnote_count = -1; +            foreach (content; munged_contents) { /+ loop to inline endnotes +/ +              content = replaceAll!(m => "~{ " ~ munged_endnotes[++endnote_count] ~ " }~" ~ m["tail"] ) +                (content, rgx_endnote_ref); // endnote_ref cannot occur in a code block or else fail +              buffer.write(content ~ "\n\n"); +            } +            if (buffer) { +              try { +                string dir_out = "converted_output_"; +                string path_and_file_out = dir_out ~ "/" ~ filename; +                dir_out.mkdirRecurse; +                auto f = File(path_and_file_out, "w"); +                f.write(buffer); +                // writeln("wrote: ", path_and_file_out); +              } catch (FileException ex) { +                writeln("did not write file"); +                // Handle errors +              } +            } +          } else { +            foreach (content; munged_contents) { /+ loop to inline endnotes +/ +              buffer.write(content ~ "\n\n"); +            } +          } +        } +      } catch (ErrnoException ex) { +        switch(ex.errno) { +          case EPERM: +          case EACCES: // Permission denied +            break; +          case ENOENT: // File does not exist +            break; +          default:     // Handle other errors +            break; +        } +      } +    } +  } +} diff --git a/misc/util/d/tools/markup_conversion/markup_conversion_from_sisu_ruby_to_sisu_spine.d b/misc/util/d/tools/markup_conversion/markup_conversion_from_sisu_ruby_to_sisu_spine.d new file mode 100755 index 0000000..94e8718 --- /dev/null +++ b/misc/util/d/tools/markup_conversion/markup_conversion_from_sisu_ruby_to_sisu_spine.d @@ -0,0 +1,354 @@ +#!/usr/bin/env rdmd +/+ +  - read in file .sst .ssi .ssm +  - loop twice +    - first +      - check for and skip code blocks +      - use unique code marker for endnote markers in text and give an endnote +        number ★1, increment +      - extract all endnotes in array +    - second +      - check that the footnote marker number count matches the number of notes +        in the array +        - if they match either: +            - substitute each endnote marker with the array footnote[number-1] +            - substitute each endnote marker with footnote +              as inlined footnote markup (footnote number not needed) +        - if they do not match exit +  - check whether changes have been made +    - if so write file with inline footnotes in sub-directory converted_output_/ +      using the same name as the original file +    - else, exit ++/ +import std.stdio; +import std.file; +import std.array : split, join; +import std.exception; +// import std.range; +import core.stdc.errno; +import std.regex; +import std.format; +import std.conv; +void main(string[] args) { +  static heading_a               = ctRegex!(`^:?[A][~] `, "m"); +  static comment                 = ctRegex!(`^%+ `); +  static block_tic_code_open     = ctRegex!("^`{3} code(?:[.](?P<syntax>[a-z][0-9a-z#+_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?"); +  static block_tic_close         = ctRegex!("^(`{3})$","m"); +  static block_curly_code_open   = ctRegex!(`^(?:code(?:[.](?P<syntax>[a-z][0-9a-z_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?[{][ ]*$)`); +  static block_curly_code_close  = ctRegex!(`^([}]code)`); +  auto rgx_endnote_ref           = ctRegex!(`([~]\^)(?P<tail>[)\]]? |$)`, "gm"); +  auto rgx_endnote               = ctRegex!(`^\^~\s+(.+|\n)`, "gm"); +  char[][] header0Content1(in string src_text) { // cast(char[]) +    /+ split string on _first_ match of "^:?A~\s" into [header, content] array/tuple +/ +    char[][] header_and_content; +    auto m = (cast(char[]) src_text).matchFirst(heading_a); +    header_and_content ~= m.pre; +    header_and_content ~= m.hit ~ m.post; +    assert(header_and_content.length == 2, +      "document markup is broken, header body split == " +      ~ header_and_content.length.to!string +      ~ "; (header / body array split should == 2 (split is on level A~))" +    ); +    return header_and_content; +  } +  string format_body_string(string s) { +    string o; +    o = s +     .replaceAll(regex("^<(?:/[ ]*)?br>[ ]*"), " \\\\ ") +     .replaceAll(regex("[ ]*<(?:/[ ]*)?br>$"), " \\\\") +     .replaceAll(regex("[ ]*<(?:/[ ]*)?br>[ ]*"), " \\\\ "); +    return o; +  } +  string format_header_string(string s) { +    string o; +    o = s +     .replaceAll(regex("\""), "\\\"") +     .replaceAll(regex("[ ]*<(?:/[ ]*)?br>$"), " \\\\") +     .replaceAll(regex("[ ]*<(?:/[ ]*)?br>[ ]*"), " \\\\ "); +    return o; +  } +  string format_main_header(string hm, string hs = "", string c = "") { +    string o; +    if (c.length == 0) { +      o ~= hm ~ ":\n"; +    } else { +      o ~= hm ~ ":\n" +        ~ "  " ~ hs ~ ": " +        ~ "\"" ~ format_header_string(c) ~ "\"\n"; +    } +    return o; +  } +  string format_sub_header(string hs, string c) { +    string o; +    o ~= "  " ~ hs ~ ": " +      ~ "\"" ~ format_header_string(c) ~ "\"\n"; +    return o; +  } +  foreach(arg; args[1..$]) { +    if ( +      !(arg.match(regex(r"--\w+"))) +      && arg.match(regex(r"\w+?\.ss[itm]")) +    ) { +      writeln(arg); +      string filename                  = arg; +      try { +        string[] munged_header, munged_contents, munged_endnotes, endnote_refs; +        string text                    = filename.readText; +        char[][] hc                    = header0Content1(text); +        char[] src_header              = hc[0]; +        string[] headers               = src_header.to!string.split("\n\n"); +        char[] src_txt                 = hc[1]; +        string[] paragraphs            = src_txt.to!string.split("\n\n"); +        int endnote_ref_count          = 0; +        int[string] type = [ +          "curly_code"                 : 0, +          "tic_code"                   : 0, +        ]; +        string _tmp_header; +        headers[0] = headers[0].replaceFirst(regex(r"^%\s+SiSU.+", "i"), "# SiSU 8.0 spine (auto-conversion)"); +        foreach (h_; headers) { +          _tmp_header = ""; +          if (auto m = h_.match(regex(r"^%\s*", "m"))) { +            h_ = h_.replaceAll(regex(r"^%\s*", "m"), "# ") ~ "\n"; +          } +          if (h_.match(regex(r"^@title:|@subtitle"))) { +            if (auto m = h_.match(regex(r"^@(?P<h>title):(?:[ ]+(?P<c>.+)|\n)"))) { +              _tmp_header ~= format_main_header(m.captures["h"], "main", m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^@(?P<h>subtitle):(?:[ ]+(?P<c>.+)|$)"))) { +              if (m.captures["c"].length == 0) { +              } else { +                _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +              } +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>main):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:sub(?:title)?:(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header("subtitle", m.captures["c"]); +            } +          } else if (h_.match(regex(r"^@creator:|@author:"))) { +            if (auto m = h_.match(regex(r"^(?:@creator:|@author:)(?:[ ]+(?P<c>.+)|\n)"))) { +              _tmp_header ~= format_main_header("creator", "author", m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>author):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +          } else if (h_.match(regex(r"^@rights:"))) { +            if (auto m = h_.match(regex(r"^@(?P<h>rights):(?:[ ]+(?P<c>.+)|\n)"))) { +              _tmp_header ~= format_main_header(m.captures["h"], "copyright", m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>copyright):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:licen[cs]e:(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header("license", m.captures["c"]); +            } +          } else if (h_.match(regex(r"^@date:|@date\."))) { +            if (auto m = h_.match(regex(r"^@(?P<h>date):(?:[ ]+(?P<c>.+)|\n)"))) { +              _tmp_header ~= format_main_header(m.captures["h"], "published", m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>published):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>available):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>modified):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>created):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>issued):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>valid):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^@date\.(?P<h>available):[ ]+(?P<c>.+)$"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^@date\.(?P<h>modified):[ ]+(?P<c>.+)$"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^@date\.(?P<h>created):[ ]+(?P<c>.+)$"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^@date\.(?P<h>issued):[ ]+(?P<c>.+)$"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^@date\.(?P<h>valid):[ ]+(?P<c>.+)$"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +          } else if (h_.match(regex(r"^@classify:"))) { +            if (auto m = h_.match(regex(r"^@classify:"))) { +              _tmp_header ~= "classify:\n"; +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>topic_register):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:type:(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= "#  type: " ~ "\"" ~ m.captures["c"] ~ "\"\n"; +            } +          } else if (h_.match(regex(r"^(?:@identifier:|@identify:)"))) { +            if (auto m = h_.match(regex(r"^(?:@identifier:|@idenfify)"))) { +              _tmp_header ~= "identify:\n"; +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>oclc):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>isbn):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>dewey):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +          } else if (h_.match(regex(r"^@publisher:"))) { +            if (auto m = h_.match(regex(r"^@publisher:[ ]+(?P<c>.+)$"))) { +              _tmp_header ~= "publisher: " ~  "\"" ~ m.captures["c"] ~ "\"\n"; +            } +          } else if (h_.match(regex(r"^@make:"))) { +            // writeln(h_); +            if (auto m = h_.match(regex(r"^@make:"))) { +              _tmp_header ~= "make:\n"; +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>breaks):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>num_top):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>headings):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>italics):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>bold):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>emphasis):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>texpdf_font):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>home_button_text):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>home_button_image):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>cover_image):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            if (auto m = h_.match(regex(r"^\s+:(?P<h>footer):(?:[ ]+(?P<c>.+)|$)", "m"))) { +              _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +            } +            // writeln(_tmp_header); +          } else if (h_.match(regex(r"^@\w+:"))) { +            _tmp_header ~= "# " ~ h_.split("\n").join("\n# ") ~ "\n"; +          } else if (h_.match(regex(r"^\s+:\w+:", "m"))) { +            if (auto m = h_.match(regex(r"^(?P<g>\s+:\w+:.*)"))) { +              _tmp_header ~= "# " ~ m.captures["g"] ~ "\n"; +            } +          } +          if (h_.match(regex(r"^#", "m"))) { +            if (auto m = h_.match(regex(r"^(?P<g>#.*)", "m"))) { +              _tmp_header ~= m.captures["g"] ~ "\n"; +            } +           } +          if (_tmp_header.length > 0) { +            munged_header ~= _tmp_header.split("\n\n"); +          } else if (h_.length > 0) { +            writeln("munging required: ", h_); +            h_ = h_.replaceAll((regex(r"\n\n\n+", "m")), "\n\n"); +            munged_header ~= h_; +          } +        } +        // writeln(munged_header.join("\n")); +        foreach (paragraph; paragraphs) {                                                                                  /+ loop to gather binary endnotes +/ +          if ( !( type["curly_code"] == 1 || type["tic_code"] == 1) +            && paragraph.match(rgx_endnote) +          ) { +            munged_endnotes ~= replaceAll!(m => m[1]) +              (paragraph, rgx_endnote); +          } else { +            if ( type["curly_code"] == 1 || type["tic_code"] == 1 +              || paragraph.matchFirst(block_curly_code_open) +              || paragraph.matchFirst(block_tic_code_open) +            ) { /+ code blocks identified, no munging +/ +              if ( type["curly_code"] == 1 +                && paragraph.matchFirst(block_curly_code_close) +              ) { +                type["curly_code"] = 0; +              } else if (type["tic_code"] == 1 +                && paragraph.matchFirst(block_tic_close) +              ) { +                type["tic_code"] = 0; +              } else if (paragraph.matchFirst(block_curly_code_open)) { +                type["curly_code"] = 1; +              } else if (paragraph.matchFirst(block_tic_code_open)) { +                type["tic_code"] = 1; +              } +              munged_contents ~= paragraph; +            } else { /+ regular content, not a code block +/ +              if (auto m = paragraph.matchAll(rgx_endnote_ref)) { +                foreach (n; m) { +                  endnote_ref_count++; // endnote_refs ~= (n.captures[1]); +                } +              } +              paragraph = format_body_string(paragraph); +              // paragraph = replaceAll!(m => " \\\\ " ) +              //   (paragraph, regex(r"\s*<(?:/\s*|:)?br>\s*")); // (paragraph, regex(r"(<br>)")); +              munged_contents ~= paragraph; +            } +          } +        } +        { +          import std.outbuffer; +          auto buffer = new OutBuffer(); +          foreach (header; munged_header) { /+ loop to inline endnotes +/ +            buffer.write(header ~ "\n"); +          } +          if (munged_endnotes.length == endnote_ref_count) { +            int endnote_count = -1; +            foreach (content; munged_contents) { /+ loop to inline endnotes +/ +              content = replaceAll!(m => "~{ " ~ munged_endnotes[++endnote_count] ~ " }~" ~ m["tail"] ) +                (content, rgx_endnote_ref); // endnote_ref cannot occur in a code block or else fail +              buffer.write(content ~ "\n\n"); +            } +            if (buffer) { +              try { +                string dir_out = "converted_output_"; +                string path_and_file_out = dir_out ~ "/" ~ filename; +                dir_out.mkdirRecurse; +                auto f = File(path_and_file_out, "w"); +                f.write(buffer); +                // writeln("wrote: ", path_and_file_out); +              } catch (FileException ex) { +                writeln("did not write file"); +                // Handle errors +              } +            } +          } else { +            foreach (content; munged_contents) { /+ loop to inline endnotes +/ +              buffer.write(content ~ "\n\n"); +            } +          } +        } +      } catch (ErrnoException ex) { +        switch(ex.errno) { +          case EPERM: +          case EACCES: // Permission denied +            break; +          case ENOENT: // File does not exist +            break; +          default:     // Handle other errors +            break; +        } +      } +    } +  } +} diff --git a/org/spine_build_scaffold.org b/org/spine_build_scaffold.org index 144a82c..8842f52 100644 --- a/org/spine_build_scaffold.org +++ b/org/spine_build_scaffold.org @@ -371,7 +371,8 @@ skel:  	mkdir -p build; \  	mkdir -p views; \  	mkdir -p data; \ -	mkdir -p misc/util; \ +	mkdir -p misc/util/d/cgi/search/cgi-bin/src; \ +	mkdir -p misc/util/d/tools/markup_conversion; \  	mkdir -p misc/editor-syntax-etc/emacs; \  	mkdir -p misc/editor-syntax-etc/vim/syntax; \  	mkdir -p misc/editor-syntax-etc/vim/colors; \ @@ -405,7 +406,8 @@ expunge:  distclean: expunge  distclean_and_init: expunge  	mkdir -p views; \ -	mkdir -p misc/util; \ +	mkdir -p misc/util/d/cgi/search/cgi-bin/src; \ +	mkdir -p misc/util/d/tools/markup_conversion; \  	mkdir -p misc/editor-syntax-etc/emacs; \  	mkdir -p misc/editor-syntax-etc/vim/syntax; \  	mkdir -p misc/editor-syntax-etc/vim/colors; \ @@ -1151,10 +1153,6 @@ spine_exe = executable('spine',  !org  !misc  !misc/** -!util -!util/** -!editor-syntax-etc -!editor-syntax-etc/**  !ext_lib  !ext_lib/**  !src diff --git a/org/util_spine_markup_conversion_from_sisu.org b/org/util_spine_markup_conversion_from_sisu.org new file mode 100644 index 0000000..8053bcf --- /dev/null +++ b/org/util_spine_markup_conversion_from_sisu.org @@ -0,0 +1,951 @@ +-*- mode: org -*- +#+TITLE:       spine (doc_reform) information files +#+DESCRIPTION: documents - structuring, various output representations & search +#+FILETAGS:    :spine:info: +#+AUTHOR:      Ralph Amissah +#+EMAIL:       [[mailto:ralph.amissah@gmail.com][ralph.amissah@gmail.com]] +#+COPYRIGHT:   Copyright (C) 2015 - 2020 Ralph Amissah +#+LANGUAGE:    en +#+STARTUP:     content hideblocks hidestars noindent entitiespretty +#+PROPERTY:    header-args  :exports code +#+PROPERTY:    header-args+ :noweb yes +#+PROPERTY:    header-args+ :eval no +#+PROPERTY:    header-args+ :results no +#+PROPERTY:    header-args+ :cache no +#+PROPERTY:    header-args+ :padline no + +* Markup conversion tools + +** README + +#+BEGIN_SRC text  :tangle "../misc/util/d/tools/markup_conversion/README" +#+END_SRC + +** endnotes, inline from binary +*** tangle + +#+BEGIN_SRC d  :tangle "../misc/util/d/tools/markup_conversion/endnotes_inline_from_binary.d" :tangle-mode (identity #o755) :shebang #!/usr/bin/env rdmd +<<inline_notes_head>> +<<inline_notes_imports>> +void main(string[] args) { +  <<inline_notes_init>> +  foreach(arg; args[1..$]) { +    if ( +      !(arg.match(regex(r"--\w+"))) +      && arg.match(regex(r"\w+?\.ss[itm]")) +    ) { +      <<inline_notes_pre_loops>> +        <<inline_notes_loop_doc_body>> +        <<inline_notes_loop_adjust_and_output>> +      <<inline_notes_exceptions>> +    } +  } +} +#+END_SRC + +*** head + +#+NAME: inline_notes_head +#+BEGIN_SRC d +/+ +  - read in file .sst .ssi .ssm +  - loop twice +    - first +      - check for and skip code blocks +      - use unique code marker for endnote markers in text and give an endnote +        number ★1, increment +      - extract all endnotes in array +    - second +      - check that the footnote marker number count matches the number of notes +        in the array +        - if they match either: +            - substitute each endnote marker with the array footnote[number-1] +            - substitute each endnote marker with footnote +              as inlined footnote markup (footnote number not needed) +        - if they do not match exit +  - check whether changes have been made +    - if so write file with inline footnotes in sub-directory converted_output_/ +      using the same name as the original file +    - else, exit ++/ +#+END_SRC + +*** import + +#+NAME: inline_notes_imports +#+BEGIN_SRC d +import std.stdio; +import std.file; +import std.array : split; +import std.exception; +import core.stdc.errno; +import std.regex; +import std.format; +import std.conv; +#+END_SRC + +*** init + +#+NAME: inline_notes_init +#+BEGIN_SRC d +static comment                 = ctRegex!(`^%+ `); +static block_tic_code_open     = ctRegex!("^`{3} code(?:[.](?P<syntax>[a-z][0-9a-z#+_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?"); +static block_tic_close         = ctRegex!("^(`{3})$","m"); +static block_curly_code_open   = ctRegex!(`^(?:code(?:[.](?P<syntax>[a-z][0-9a-z_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?[{][ ]*$)`); +static block_curly_code_close  = ctRegex!(`^([}]code)`); +auto rgx_endnote_ref           = ctRegex!(`([~]\^)(?P<tail>[)\]]? |$)`, "gm"); +auto rgx_endnote               = ctRegex!(`^\^~\s+(.+|\n)`, "gm"); +#+END_SRC + +*** pre-loops + +#+NAME: inline_notes_pre_loops +#+BEGIN_SRC d +writeln(arg); +string filename                  = arg; +try { +  string[] contents, endnotes, endnote_refs; +  string text                    = filename.readText; +  string[] paragraphs            = text.split("\n\n"); +  int endnote_ref_count          = 0; +  int[string] type = [ +    "curly_code"                 : 0, +    "tic_code"                   : 0, +  ]; +#+END_SRC + +*** loop doc body + +#+NAME: inline_notes_loop_doc_body +#+BEGIN_SRC d +foreach (paragraph; paragraphs) { /+ loop to gather binary endnotes +/ +  if ( !( type["curly_code"] == 1 || type["tic_code"] == 1) +    && paragraph.match(rgx_endnote) +  ) { +    endnotes ~= replaceAll!(m => m[1]) +      (paragraph, rgx_endnote); +  } else { +    if ( type["curly_code"] == 1 +      && paragraph.matchFirst(block_curly_code_close) +    ) { +      type["curly_code"] = 0; +    } else if (type["tic_code"] == 1 +      && paragraph.matchFirst(block_tic_close) +    ) { +      type["tic_code"] = 0; +    } else if ( type["curly_code"] == 1 || type["tic_code"] == 1) { +      // skip, prevent search for endnotes +    } else if (paragraph.matchFirst(block_curly_code_open)) { +      type["curly_code"] = 1; +    } else if (paragraph.matchFirst(block_tic_code_open)) { +      type["tic_code"] = 1; +    } else if (auto m = paragraph.matchAll(rgx_endnote_ref)) { +      foreach (n; m) { +        endnote_ref_count++; // endnote_refs ~= (n.captures[1]); +      } +    } +    contents ~= paragraph; +  } +} +#+END_SRC + +*** (loop to) adjustment & output + +#+NAME: inline_notes_loop_adjust_and_output +#+BEGIN_SRC d +if (endnotes.length == endnote_ref_count) { +  import std.outbuffer; +  writeln("endnote ref count:         ", endnote_ref_count); +  writeln("number of binary endnotes: ", endnotes.length); +  int endnote_count = -1; +  auto buffer = new OutBuffer(); +  foreach (content; contents) { /+ loop to inline endnotes +/ +    content = replaceAll!(m => "~{ " ~ endnotes[++endnote_count] ~ " }~" ~ m["tail"] ) +      (content, rgx_endnote_ref); +    buffer.write(content ~ "\n\n"); +  } +  if (buffer) { +    try { +      string dir_out = "converted_output_"; +      string path_and_file_out = dir_out ~ "/" ~ filename; +      dir_out.mkdirRecurse; +      auto f = File(path_and_file_out, "w"); +      f.write(buffer); +      writeln("wrote: ", path_and_file_out); +    } catch (FileException ex) { +      writeln("did not write file"); +      // Handle errors +    } +  } +} else { +  writeln("ERROR binary endnote mismatch, check markup,\nmisatch in the number of endnotes & endnote references!"); +  writeln("  number of endnotes:     ", endnotes.length); +  writeln("  number of endnote refs: ", endnote_ref_count); // endnote_refs.length, +} +// assert(endnotes.length == endnote_ref_count); +#+END_SRC + +*** exceptions + +#+NAME: inline_notes_exceptions +#+BEGIN_SRC d +} catch (ErrnoException ex) { +  switch(ex.errno) { +    case EPERM: +    case EACCES: // Permission denied +      break; +    case ENOENT: // File does not exist +      break; +    default:     // Handle other errors +      break; +  } +} +#+END_SRC + +** conversion from sisu (sisu bespoke headers) any binary to inline notes TODO +*** tangle + +#+BEGIN_SRC d  :tangle "../misc/util/d/tools/markup_conversion/markup_conversion_from_sisu_ruby_to_sisu_spine.d" :tangle-mode (identity #o755) :shebang #!/usr/bin/env rdmd +<<from_sisu_rb_head>> +<<from_sisu_rb_imports>> +void main(string[] args) { +  <<from_sisu_rb_init>> +  <<from_sisu_rb_body_format>> +  <<from_sisu_rb_header_format>> +  foreach(arg; args[1..$]) { +    if ( +      !(arg.match(regex(r"--\w+"))) +      && arg.match(regex(r"\w+?\.ss[itm]")) +    ) { +      <<from_sisu_rb_pre_loops>> +        <<from_sisu_rb_loop_doc_header>> +        <<from_sisu_rb_loop_doc_body>> +        <<from_sisu_rb_loop_adjust_and_output>> +      <<from_sisu_rb_exceptions>> +    } +  } +} +#+END_SRC + +*** head + +#+NAME: from_sisu_rb_head +#+BEGIN_SRC d +/+ +  - read in file .sst .ssi .ssm +  - loop twice +    - first +      - check for and skip code blocks +      - use unique code marker for endnote markers in text and give an endnote +        number ★1, increment +      - extract all endnotes in array +    - second +      - check that the footnote marker number count matches the number of notes +        in the array +        - if they match either: +            - substitute each endnote marker with the array footnote[number-1] +            - substitute each endnote marker with footnote +              as inlined footnote markup (footnote number not needed) +        - if they do not match exit +  - check whether changes have been made +    - if so write file with inline footnotes in sub-directory converted_output_/ +      using the same name as the original file +    - else, exit ++/ +#+END_SRC + +*** import + +#+NAME: from_sisu_rb_imports +#+BEGIN_SRC d +import std.stdio; +import std.file; +import std.array : split, join; +import std.exception; +// import std.range; +import core.stdc.errno; +import std.regex; +import std.format; +import std.conv; +#+END_SRC + +*** init + +#+NAME: from_sisu_rb_init +#+BEGIN_SRC d +static heading_a               = ctRegex!(`^:?[A][~] `, "m"); +static comment                 = ctRegex!(`^%+ `); +static block_tic_code_open     = ctRegex!("^`{3} code(?:[.](?P<syntax>[a-z][0-9a-z#+_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?"); +static block_tic_close         = ctRegex!("^(`{3})$","m"); +static block_curly_code_open   = ctRegex!(`^(?:code(?:[.](?P<syntax>[a-z][0-9a-z_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?[{][ ]*$)`); +static block_curly_code_close  = ctRegex!(`^([}]code)`); +auto rgx_endnote_ref           = ctRegex!(`([~]\^)(?P<tail>[)\]]? |$)`, "gm"); +auto rgx_endnote               = ctRegex!(`^\^~\s+(.+|\n)`, "gm"); +char[][] header0Content1(in string src_text) { // cast(char[]) +  /+ split string on _first_ match of "^:?A~\s" into [header, content] array/tuple +/ +  char[][] header_and_content; +  auto m = (cast(char[]) src_text).matchFirst(heading_a); +  header_and_content ~= m.pre; +  header_and_content ~= m.hit ~ m.post; +  assert(header_and_content.length == 2, +    "document markup is broken, header body split == " +    ~ header_and_content.length.to!string +    ~ "; (header / body array split should == 2 (split is on level A~))" +  ); +  return header_and_content; +} +#+END_SRC + +*** body format +**** format body string + +#+NAME: from_sisu_rb_body_format +#+BEGIN_SRC d +string format_body_string(string s) { +  string o; +  o = s +   .replaceAll(regex("^<(?:/[ ]*)?br>[ ]*"), " \\\\ ") +   .replaceAll(regex("[ ]*<(?:/[ ]*)?br>$"), " \\\\") +   .replaceAll(regex("[ ]*<(?:/[ ]*)?br>[ ]*"), " \\\\ "); +  return o; +} +#+END_SRC + +*** header format +**** format header string + +#+NAME: from_sisu_rb_header_format +#+BEGIN_SRC d +string format_header_string(string s) { +  string o; +  o = s +   .replaceAll(regex("\""), "\\\"") +   .replaceAll(regex("[ ]*<(?:/[ ]*)?br>$"), " \\\\") +   .replaceAll(regex("[ ]*<(?:/[ ]*)?br>[ ]*"), " \\\\ "); +  return o; +} +#+END_SRC + +**** format main header + +#+NAME: from_sisu_rb_header_format +#+BEGIN_SRC d +string format_main_header(string hm, string hs = "", string c = "") { +  string o; +  if (c.length == 0) { +    o ~= hm ~ ":\n"; +  } else { +    o ~= hm ~ ":\n" +      ~ "  " ~ hs ~ ": " +      ~ "\"" ~ format_header_string(c) ~ "\"\n"; +  } +  return o; +} +#+END_SRC + +**** format sub header + +#+NAME: from_sisu_rb_header_format +#+BEGIN_SRC d +string format_sub_header(string hs, string c) { +  string o; +  o ~= "  " ~ hs ~ ": " +    ~ "\"" ~ format_header_string(c) ~ "\"\n"; +  return o; +} + #+END_SRC + +*** pre-loops + +#+NAME: from_sisu_rb_pre_loops +#+BEGIN_SRC d +writeln(arg); +string filename                  = arg; +try { +  string[] munged_header, munged_contents, munged_endnotes, endnote_refs; +  string text                    = filename.readText; +  char[][] hc                    = header0Content1(text); +  char[] src_header              = hc[0]; +  string[] headers               = src_header.to!string.split("\n\n"); +  char[] src_txt                 = hc[1]; +  string[] paragraphs            = src_txt.to!string.split("\n\n"); +  int endnote_ref_count          = 0; +  int[string] type = [ +    "curly_code"                 : 0, +    "tic_code"                   : 0, +  ]; +  string _tmp_header; +#+END_SRC + +*** loop doc header + +#+NAME: from_sisu_rb_loop_doc_header +#+BEGIN_SRC d +headers[0] = headers[0].replaceFirst(regex(r"^%\s+SiSU.+", "i"), "# SiSU 8.0 spine (auto-conversion)"); +foreach (h_; headers) { +  _tmp_header = ""; +  if (auto m = h_.match(regex(r"^%\s*", "m"))) { +    h_ = h_.replaceAll(regex(r"^%\s*", "m"), "# ") ~ "\n"; +  } +  if (h_.match(regex(r"^@title:|@subtitle"))) { +    if (auto m = h_.match(regex(r"^@(?P<h>title):(?:[ ]+(?P<c>.+)|\n)"))) { +      _tmp_header ~= format_main_header(m.captures["h"], "main", m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^@(?P<h>subtitle):(?:[ ]+(?P<c>.+)|$)"))) { +      if (m.captures["c"].length == 0) { +      } else { +        _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +      } +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>main):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:sub(?:title)?:(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header("subtitle", m.captures["c"]); +    } +  } else if (h_.match(regex(r"^@creator:|@author:"))) { +    if (auto m = h_.match(regex(r"^(?:@creator:|@author:)(?:[ ]+(?P<c>.+)|\n)"))) { +      _tmp_header ~= format_main_header("creator", "author", m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>author):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +  } else if (h_.match(regex(r"^@rights:"))) { +    if (auto m = h_.match(regex(r"^@(?P<h>rights):(?:[ ]+(?P<c>.+)|\n)"))) { +      _tmp_header ~= format_main_header(m.captures["h"], "copyright", m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>copyright):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:licen[cs]e:(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header("license", m.captures["c"]); +    } +  } else if (h_.match(regex(r"^@date:|@date\."))) { +    if (auto m = h_.match(regex(r"^@(?P<h>date):(?:[ ]+(?P<c>.+)|\n)"))) { +      _tmp_header ~= format_main_header(m.captures["h"], "published", m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>published):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>available):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>modified):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>created):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>issued):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>valid):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^@date\.(?P<h>available):[ ]+(?P<c>.+)$"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^@date\.(?P<h>modified):[ ]+(?P<c>.+)$"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^@date\.(?P<h>created):[ ]+(?P<c>.+)$"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^@date\.(?P<h>issued):[ ]+(?P<c>.+)$"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^@date\.(?P<h>valid):[ ]+(?P<c>.+)$"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +  } else if (h_.match(regex(r"^@classify:"))) { +    if (auto m = h_.match(regex(r"^@classify:"))) { +      _tmp_header ~= "classify:\n"; +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>topic_register):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:type:(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= "#  type: " ~ "\"" ~ m.captures["c"] ~ "\"\n"; +    } +  } else if (h_.match(regex(r"^(?:@identifier:|@identify:)"))) { +    if (auto m = h_.match(regex(r"^(?:@identifier:|@idenfify)"))) { +      _tmp_header ~= "identify:\n"; +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>oclc):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>isbn):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>dewey):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +  } else if (h_.match(regex(r"^@publisher:"))) { +    if (auto m = h_.match(regex(r"^@publisher:[ ]+(?P<c>.+)$"))) { +      _tmp_header ~= "publisher: " ~  "\"" ~ m.captures["c"] ~ "\"\n"; +    } +  } else if (h_.match(regex(r"^@make:"))) { +    // writeln(h_); +    if (auto m = h_.match(regex(r"^@make:"))) { +      _tmp_header ~= "make:\n"; +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>breaks):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>num_top):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>headings):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>italics):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>bold):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>emphasis):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>texpdf_font):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>home_button_text):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>home_button_image):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>cover_image):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    if (auto m = h_.match(regex(r"^\s+:(?P<h>footer):(?:[ ]+(?P<c>.+)|$)", "m"))) { +      _tmp_header ~= format_sub_header(m.captures["h"], m.captures["c"]); +    } +    // writeln(_tmp_header); +  } else if (h_.match(regex(r"^@\w+:"))) { +    _tmp_header ~= "# " ~ h_.split("\n").join("\n# ") ~ "\n"; +  } else if (h_.match(regex(r"^\s+:\w+:", "m"))) { +    if (auto m = h_.match(regex(r"^(?P<g>\s+:\w+:.*)"))) { +      _tmp_header ~= "# " ~ m.captures["g"] ~ "\n"; +    } +  } +  if (h_.match(regex(r"^#", "m"))) { +    if (auto m = h_.match(regex(r"^(?P<g>#.*)", "m"))) { +      _tmp_header ~= m.captures["g"] ~ "\n"; +    } +   } +  if (_tmp_header.length > 0) { +    munged_header ~= _tmp_header.split("\n\n"); +  } else if (h_.length > 0) { +    writeln("munging required: ", h_); +    h_ = h_.replaceAll((regex(r"\n\n\n+", "m")), "\n\n"); +    munged_header ~= h_; +  } +} +// writeln(munged_header.join("\n")); +#+END_SRC + +*** loop doc body (identify & ignore code blocks) + +#+NAME: from_sisu_rb_loop_doc_body +#+BEGIN_SRC d +foreach (paragraph; paragraphs) {                                                                                  /+ loop to gather binary endnotes +/ +  if ( !( type["curly_code"] == 1 || type["tic_code"] == 1) +    && paragraph.match(rgx_endnote) +  ) { +    munged_endnotes ~= replaceAll!(m => m[1]) +      (paragraph, rgx_endnote); +  } else { +    if ( type["curly_code"] == 1 || type["tic_code"] == 1 +      || paragraph.matchFirst(block_curly_code_open) +      || paragraph.matchFirst(block_tic_code_open) +    ) { /+ code blocks identified, no munging +/ +      if ( type["curly_code"] == 1 +        && paragraph.matchFirst(block_curly_code_close) +      ) { +        type["curly_code"] = 0; +      } else if (type["tic_code"] == 1 +        && paragraph.matchFirst(block_tic_close) +      ) { +        type["tic_code"] = 0; +      } else if (paragraph.matchFirst(block_curly_code_open)) { +        type["curly_code"] = 1; +      } else if (paragraph.matchFirst(block_tic_code_open)) { +        type["tic_code"] = 1; +      } +      munged_contents ~= paragraph; +    } else { /+ regular content, not a code block +/ +      if (auto m = paragraph.matchAll(rgx_endnote_ref)) { +        foreach (n; m) { +          endnote_ref_count++; // endnote_refs ~= (n.captures[1]); +        } +      } +      paragraph = format_body_string(paragraph); +      // paragraph = replaceAll!(m => " \\\\ " ) +      //   (paragraph, regex(r"\s*<(?:/\s*|:)?br>\s*")); // (paragraph, regex(r"(<br>)")); +      munged_contents ~= paragraph; +    } +  } +} +#+END_SRC + +*** (loop to) adjustment & output + +#+NAME: from_sisu_rb_loop_adjust_and_output +#+BEGIN_SRC d +{ +  import std.outbuffer; +  auto buffer = new OutBuffer(); +  foreach (header; munged_header) { /+ loop to inline endnotes +/ +    buffer.write(header ~ "\n"); +  } +  if (munged_endnotes.length == endnote_ref_count) { +    int endnote_count = -1; +    foreach (content; munged_contents) { /+ loop to inline endnotes +/ +      content = replaceAll!(m => "~{ " ~ munged_endnotes[++endnote_count] ~ " }~" ~ m["tail"] ) +        (content, rgx_endnote_ref); // endnote_ref cannot occur in a code block or else fail +      buffer.write(content ~ "\n\n"); +    } +    if (buffer) { +      try { +        string dir_out = "converted_output_"; +        string path_and_file_out = dir_out ~ "/" ~ filename; +        dir_out.mkdirRecurse; +        auto f = File(path_and_file_out, "w"); +        f.write(buffer); +        // writeln("wrote: ", path_and_file_out); +      } catch (FileException ex) { +        writeln("did not write file"); +        // Handle errors +      } +    } +  } else { +    foreach (content; munged_contents) { /+ loop to inline endnotes +/ +      buffer.write(content ~ "\n\n"); +    } +  } +} +#+END_SRC + +*** exceptions + +#+NAME: from_sisu_rb_exceptions +#+BEGIN_SRC d +} catch (ErrnoException ex) { +  switch(ex.errno) { +    case EPERM: +    case EACCES: // Permission denied +      break; +    case ENOENT: // File does not exist +      break; +    default:     // Handle other errors +      break; +  } +} +#+END_SRC + +** conversion from sisu and multiple headers (sisu bespoke, sdlang, toml) incomplete +*** tangle + +#+BEGIN_SRC d  :tangle "../misc/util/d/tools/markup_conversion/markup_changes_header_and_content.d" :tangle-mode (identity #o755) :shebang #!/usr/bin/env rdmd +<<from_previous_markups_head>> +<<from_previous_markups_imports>> +void main(string[] args) { +  <<from_previous_markups_init>> +  foreach(arg; args[1..$]) { +    if ( +      !(arg.match(regex(r"--\w+"))) +      && arg.match(regex(r"\w+?\.ss[itm]")) +    ) { +      <<from_previous_markups_pre_loops>> +        <<from_previous_markups_loop_doc_header>> +        <<from_previous_markups_loop_doc_body>> +        <<from_previous_markups_loop_adjust_and_output>> +      <<from_previous_markups_exceptions>> +    } +  } +} +#+END_SRC + +*** head + +#+NAME: from_previous_markups_head +#+BEGIN_SRC d +/+ +  - read in file .sst .ssi .ssm +  - loop twice +    - first +      - check for and skip code blocks +      - use unique code marker for endnote markers in text and give an endnote +        number ★1, increment +      - extract all endnotes in array +    - second +      - check that the footnote marker number count matches the number of notes +        in the array +        - if they match either: +            - substitute each endnote marker with the array footnote[number-1] +            - substitute each endnote marker with footnote +              as inlined footnote markup (footnote number not needed) +        - if they do not match exit +  - check whether changes have been made +    - if so write file with inline footnotes in sub-directory converted_output_/ +      using the same name as the original file +    - else, exit ++/ +#+END_SRC + +*** imports + +#+NAME: from_previous_markups_imports +#+BEGIN_SRC d +import std.stdio; +import std.file; +import std.array : split; +import std.exception; +// import std.range; +import core.stdc.errno; +import std.regex; +import std.format; +import std.conv; +#+END_SRC + +*** init + +#+NAME: from_previous_markups_init +#+BEGIN_SRC d +static heading_a               = ctRegex!(`^:?[A][~] `, "m"); +static comment                 = ctRegex!(`^%+ `); +static block_tic_code_open     = ctRegex!("^`{3} code(?:[.](?P<syntax>[a-z][0-9a-z#+_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?"); +static block_tic_close         = ctRegex!("^(`{3})$","m"); +static block_curly_code_open   = ctRegex!(`^(?:code(?:[.](?P<syntax>[a-z][0-9a-z_]+))?(?:[(](?P<attrib>[ a-zA-Z0-9;:,]*)[)])?[{][ ]*$)`); +static block_curly_code_close  = ctRegex!(`^([}]code)`); +auto rgx_endnote_ref           = ctRegex!(`([~]\^)(?P<tail>[)\]]? |$)`, "gm"); +auto rgx_endnote               = ctRegex!(`^\^~\s+(.+|\n)`, "gm"); +char[][] header0Content1(in string src_text) { // cast(char[]) +  /+ split string on _first_ match of "^:?A~\s" into [header, content] array/tuple +/ +  char[][] header_and_content; +  auto m = (cast(char[]) src_text).matchFirst(heading_a); +  header_and_content ~= m.pre; +  header_and_content ~= m.hit ~ m.post; +  assert(header_and_content.length == 2, +    "document markup is broken, header body split == " +    ~ header_and_content.length.to!string +    ~ "; (header / body array split should == 2 (split is on level A~))" +  ); +  return header_and_content; +} +#+END_SRC + +*** pre-loops + +#+NAME: from_previous_markups_pre_loops +#+BEGIN_SRC d +writeln(arg); +string filename                  = arg; +try { +  string[] munged_header, munged_contents, munged_endnotes, endnote_refs; +  string text                    = filename.readText; +  char[][] hc                    = header0Content1(text); +  char[] src_header              = hc[0]; +  string[] headers               = src_header.to!string.split("\n\n"); +  char[] src_txt                 = hc[1]; +  string[] paragraphs            = src_txt.to!string.split("\n\n"); +  int endnote_ref_count          = 0; +  int[string] type = [ +    "curly_code"                 : 0, +    "tic_code"                   : 0, +  ]; +  string _tmp_header; +#+END_SRC + +*** loop doc header + +#+NAME: from_previous_markups_loop_doc_header +#+BEGIN_SRC d +foreach (h_; headers) {                                                                                          /+ loop to inline endnotes +/ +  _tmp_header = ""; +  if (h_.match(regex(r"^[@\[]?title[:\]]?"))) {                                                                  // title +    if (auto m = h_.match(regex(r"^@title:(?:\s+(?P<c>.+)|$)"))) {                                               // sisu bespoke markup +      if (m.captures["c"].length == 0) { +        _tmp_header ~= "title:"; +      } else { +        _tmp_header ~= "title:\n  main: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } +    } else if (auto m = h_.match(regex(r"^title\s*=\s*(?P<c>.+)"))) {                                            // toml? +      if (m.captures["c"].length == 0) { +        _tmp_header ~= "title:"; +      } else { +        _tmp_header ~= "title:\n  main: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } +    } else if (auto m = h_.match(regex(r"^\[title\]"))) {                                                        // toml markup +      _tmp_header ~= "title:"; +    } else if (auto m = h_.match(regex(r"^title(?:\s+(?P<c>.+)|\s+\\$)"))) {                                     // sdlang markup +      if (m.captures["c"].length == 0) { +        _tmp_header ~= "title:"; +      } else { +        _tmp_header ~= "title:\n  main: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } +    } +    if (h_.match(regex(r"^\s*[:]?(?:main)[:= ]?", "m"))) { +      if (auto m = h_.match(regex(r"^\s+(?P<h>:main):(?:\s+(?P<c>.+)|$)", "m"))) {                                // sisu bespoke markup +        _tmp_header ~= "  main: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } else if (auto m = h_.match(regex(r"^\s*(?P<h>main)\s*=\s*(?P<c>.+)", "m"))) {                            // toml? +        _tmp_header ~= "  main: " ~ m.captures["c"]; +      } else if (auto m = h_.match(regex(r"^\s+(?P<h>main)(?:\s*\s*(?P<c>.+)|$)", "m"))) {                       // toml markup +        _tmp_header ~= "  main: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } else if (auto m = h_.match(regex(r"^\s+(?P<h>main)(?:\s+(?P<c>.+)|\s+\\$)", "m"))) {                     // sdlang markup +        _tmp_header ~= "  main: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } +    } +    if (h_.match(regex(r"^\s*[:]?(?:sub(title)?)[:= ]?", "m"))) { +      if (auto m = h_.match(regex(r"^\s+:sub(?:title)?:(?:\s+(?P<c>.+)|$)", "m"))) {                             // sisu bespoke markup +        _tmp_header ~= "  subtitle: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } else if (auto m = h_.match(regex(r"^\s*sub(?:title)?\s*=\s*(?P<c>.+)$", "m"))) {                         // toml? +        _tmp_header ~= "  subtitle: " ~ m.captures["c"]; +      } else if (auto m = h_.match(regex(r"^\s+(?:title)?(?:\s*\s*(?P<c>.+)|$)", "m"))) {                        // toml markup +        _tmp_header ~= "  subtitle: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } else if (auto m = h_.match(regex(r"^\s+(?:title)?(?:\s+(?P<c>.+)|\s+\\$)", "m"))) {                      // sdlang markup +        _tmp_header ~= "  subtitle: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } +    } +  } +  if (h_.match(regex(r"^[@\[]?rights[:\]]?"))) {                                                                 // rights +    if (auto m = h_.match(regex(r"^@rights:[ ]+(?P<c>.+)$"))) {                                                  // sisu bespoke markup +      _tmp_header ~= "rights: \n  copyright: \"" ~ m.captures["c"] ~ "\""; +    } else if (auto m = h_.match(regex(r"^@rights:"))) {                                                         // sisu bespoke markup +      _tmp_header ~= "rights:"; +    } else if (auto m = h_.match(regex(r"^\[rights\]", "m"))) {                                                  // toml markup +      _tmp_header ~= "rights:"; +    } else if (auto m = h_.match(regex(r"^rights:"))) {                                                          // sdlang markup +      _tmp_header ~= "rights:"; +    } +    if (h_.match(regex(r"^\s*[:]?copyright[:= ]?", "m"))) { +      if (auto m = h_.match(regex(r"^\s+:copyright:(?:\s+(?P<c>.+)|$)", "m"))) {                                 // sisu bespoke markup +        _tmp_header ~= "  copyright: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } else if (auto m = h_.match(regex(r"^\s*copyright\s*=\s*(?P<c>.+)", "m"))) {                              // toml? +        _tmp_header ~= "  copyright: " ~ m.captures["c"]; +      } else if (auto m = h_.match(regex(r"^\s+<h>copyright(?:\s*\s*(?P<c>.+)|$)", "m"))) {                      // toml markup +        _tmp_header ~= "  copyright: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } else if (auto m = h_.match(regex(r"^\s+copyright(?:\s+(?P<c>.+)|\s+\\$)", "m"))) {                       // sdlang markup +        _tmp_header ~= "  copyright: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } +    } +    if (h_.match(regex(r"^\s*[:]?licen[cs]e[:= ]?", "m"))) { +      if (auto m = h_.match(regex(r"^\s+:licen[cs]e:(?:\s+(?P<c>.+)|$)", "m"))) {                                // sisu bespoke markup +        _tmp_header ~= "  license: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } else if (auto m = h_.match(regex(r"^\s*licen[cs]e\s*=\s*(?P<c>.+)$", "m"))) {                            // toml? +        _tmp_header ~= "  license: " ~ m.captures["c"]; +      } else if (auto m = h_.match(regex(r"^\s+licen[cs]e(?:\s*\s*(?P<c>.+)|$)", "m"))) {                        // toml markup +        _tmp_header ~= "  license: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } else if (auto m = h_.match(regex(r"^\s+licen[cs]e(?:\s+(?P<c>.+)|\s+\\$)", "m"))) {                      // sdlang markup +        _tmp_header ~= "  license: " ~ "\"" ~ m.captures["c"] ~ "\""; +      } +    } +  } +  if (_tmp_header.length > 0) { +    munged_header ~= _tmp_header; +  } else { +    munged_header ~= h_; +  } +} +writeln(munged_header); +#+END_SRC + +*** loop doc body + +#+NAME: from_previous_markups_loop_doc_body +#+BEGIN_SRC d +foreach (paragraph; paragraphs) {                                                                                  /+ loop to gather binary endnotes +/ +  if ( !( type["curly_code"] == 1 || type["tic_code"] == 1) +    && paragraph.match(rgx_endnote) +  ) { +    munged_endnotes ~= replaceAll!(m => m[1]) +      (paragraph, rgx_endnote); +  } else { +    if ( type["curly_code"] == 1 || type["tic_code"] == 1 +      || paragraph.matchFirst(block_curly_code_open) +      || paragraph.matchFirst(block_tic_code_open) +    ) { /+ code blocks identified, no munging +/ +      if ( type["curly_code"] == 1 +        && paragraph.matchFirst(block_curly_code_close) +      ) { +        type["curly_code"] = 0; +      } else if (type["tic_code"] == 1 +        && paragraph.matchFirst(block_tic_close) +      ) { +        type["tic_code"] = 0; +      } else if (paragraph.matchFirst(block_curly_code_open)) { +        type["curly_code"] = 1; +      } else if (paragraph.matchFirst(block_tic_code_open)) { +        type["tic_code"] = 1; +      } +      munged_contents ~= paragraph; +    } else { /+ regular content, not a code block +/ +      if (auto m = paragraph.matchAll(rgx_endnote_ref)) { +        foreach (n; m) { +          endnote_ref_count++; // endnote_refs ~= (n.captures[1]); +        } +      } +      paragraph = replaceAll!(m => " \\\\ " ) +        (paragraph, regex(r"\s*<(?:/\s*|:)?br>\s*")); // (paragraph, regex(r"(<br>)")); +      munged_contents ~= paragraph; +    } +  } +} +#+END_SRC + +*** (loop to) adjust & output + +#+NAME: from_previous_markups_loop_adjust_and_output +#+BEGIN_SRC d +{ +  import std.outbuffer; +  auto buffer = new OutBuffer(); +  foreach (header; munged_header) { /+ loop to inline endnotes +/ +    buffer.write(header ~ "\n\n"); +  } +  if (munged_endnotes.length == endnote_ref_count) { +    int endnote_count = -1; +    foreach (content; munged_contents) { /+ loop to inline endnotes +/ +      content = replaceAll!(m => "~{ " ~ munged_endnotes[++endnote_count] ~ " }~" ~ m["tail"] ) +        (content, rgx_endnote_ref); // endnote_ref cannot occur in a code block or else fail +      buffer.write(content ~ "\n\n"); +    } +    if (buffer) { +      try { +        string dir_out = "converted_output_"; +        string path_and_file_out = dir_out ~ "/" ~ filename; +        dir_out.mkdirRecurse; +        auto f = File(path_and_file_out, "w"); +        f.write(buffer); +        // writeln("wrote: ", path_and_file_out); +      } catch (FileException ex) { +        writeln("did not write file"); +        // Handle errors +      } +    } +  } else { +    foreach (content; munged_contents) { /+ loop to inline endnotes +/ +      buffer.write(content ~ "\n\n"); +    } +  } +} +#+END_SRC + +*** exceptions + +#+NAME: from_previous_markups_exceptions +#+BEGIN_SRC d +} catch (ErrnoException ex) { +  switch(ex.errno) { +    case EPERM: +    case EACCES: // Permission denied +      break; +    case ENOENT: // File does not exist +      break; +    default:     // Handle other errors +      break; +  } +} +#+END_SRC diff --git a/org/spine_syntax_highlighting_emacs.org b/org/util_spine_syntax_highlighting_emacs.org index 0007e48..0007e48 100644 --- a/org/spine_syntax_highlighting_emacs.org +++ b/org/util_spine_syntax_highlighting_emacs.org diff --git a/org/spine_syntax_highlighting_vim.org b/org/util_spine_syntax_highlighting_vim.org index 137d734..137d734 100644 --- a/org/spine_syntax_highlighting_vim.org +++ b/org/util_spine_syntax_highlighting_vim.org | 
