require "isodoc" require_relative "metadata" module IsoDoc module Mpfd # A {Converter} implementation that generates Word output, and a document # schema encapsulation of the document for validation class WordConvert < IsoDoc::WordConvert def rsd_html_path(file) File.join(File.dirname(__FILE__), File.join("html", file)) end def initialize(options) super @wordstylesheet = generate_css(rsd_html_path("wordstyle.scss"), false, default_fonts(options)) @standardstylesheet = generate_css(rsd_html_path("rsd.scss"), false, default_fonts(options)) @header = rsd_html_path("header.html") @wordcoverpage = rsd_html_path("word_rsd_titlepage.html") @wordintropage = rsd_html_path("word_rsd_intro.html") @ulstyle = "l3" @olstyle = "l2" system "cp #{rsd_html_path('logo.jpg')} logo.jpg" system "cp #{rsd_html_path('mpfa-logo-no-text@4x.png')} mpfa-logo-no-text@4x.png" @files_to_delete << "logo.jpg" @files_to_delete << "mpfa-logo-no-text@4x.png" end def default_fonts(options) b = options[:bodyfont] || (options[:script] == "Hans" ? '"SimSun",serif' : '"Arial",sans-serif') h = options[:headerfont] || (options[:script] == "Hans" ? '"SimHei",sans-serif' : '"Arial",sans-serif') m = options[:monospacefont] || '"Courier New",monospace' "$bodyfont: #{b};\n$headerfont: #{h};\n$monospacefont: #{m};\n" end def metadata_init(lang, script, labels) @meta = Metadata.new(lang, script, labels) end def make_body(xml, docxml) body_attr = { lang: "EN-US", link: "blue", vlink: "#954F72" } xml.body **body_attr do |body| make_body1(body, docxml) make_body2(body, docxml) make_body3(body, docxml) end end def make_body2(body, docxml) body.div **{ class: "WordSection2" } do |div2| info docxml, div2 div2.p { |p| p << " " } # placeholder end #body.br **{ clear: "all", style: "page-break-before:auto;mso-break-type:section-break;" } section_break(body) end def title(isoxml, _out) main = isoxml&.at(ns("//title[@language='en']"))&.text set_metadata(:doctitle, main) end def generate_header(filename, dir) return unless @header template = Liquid::Template.parse(File.read(@header, encoding: "UTF-8")) meta = @meta.get meta[:filename] = filename params = meta.map { |k, v| [k.to_s, v] }.to_h File.open("header.html", "w") { |f| f.write(template.render(params)) } @files_to_delete << "header.html" "header.html" end def header_strip(h) h = h.to_s.gsub(%r{
}, " ").sub(/<\/?h[12][^>]*>/, "") h1 = to_xhtml_fragment(h.dup) h1.traverse do |x| x.replace(" ") if x.name == "span" && /mso-tab-count/.match(x["style"]) x.remove if x.name == "span" && x["class"] == "MsoCommentReference" x.remove if x.name == "a" && x["epub:type"] == "footnote" x.replace(x.children) if x.name == "a" end from_xhtml(h1) end def info(isoxml, out) @meta.security isoxml, out super end def annex_name(annex, name, div) div.h1 **{ class: "Annex" } do |t| t << "#{get_anchors[annex['id']][:label]} " t << "#{name.text}" end end def annex_name_lbl(clause, num) obl = l10n("(#{@inform_annex_lbl})") obl = l10n("(#{@norm_annex_lbl})") if clause["obligation"] == "normative" l10n("#{@annex_lbl} #{num} #{obl}") end def pre_parse(node, out) out.pre node.text # content.gsub(//, ">") end def term_defs_boilerplate(div, source, term, preface) if source.empty? && term.nil? div << @no_terms_boilerplate else div << term_defs_boilerplate_cont(source, term) end end def i18n_init(lang, script) super @annex_lbl = "Appendix" end def error_parse(node, out) # catch elements not defined in ISO case node.name when "pre" pre_parse(node, out) when "keyword" out.span node.text, **{ class: "keyword" } else super end end def fileloc(loc) File.join(File.dirname(__FILE__), loc) end def i18n_init(lang, script) super y = if lang == "en" YAML.load_file(File.join(File.dirname(__FILE__), "i18n-en.yaml")) elsif lang == "zh" && script == "Hans" YAML.load_file(File.join(File.dirname(__FILE__), "i18n-zh-Hans.yaml")) else YAML.load_file(File.join(File.dirname(__FILE__), "i18n-zh-Hans.yaml")) end @labels = @labels.merge(y) @clause_lbl = y["clause"] end def terms_defs_title(f) return f&.at(ns("./title"))&.content end TERM_CLAUSE = "//preface/terms | "\ "//preface/clause[descendant::terms]".freeze def terms_defs(isoxml, out, num) f = isoxml.at(ns(TERM_CLAUSE)) or return num out.div **attr_code(id: f["id"]) do |div| clause_name(nil, terms_defs_title(f), div, nil) f.elements.each do |e| parse(e, div) unless %w{title source}.include? e.name end end num end FRONT_CLAUSE = "//*[parent::preface]".freeze #FRONT_CLAUSE = "//clause[parent::preface] | //terms[parent::preface]".freeze def preface(isoxml, out) isoxml.xpath(ns(FRONT_CLAUSE)).each do |c| if c.name == "terms" then terms_defs isoxml, out, 0 else out.div **attr_code(id: c["id"]) do |s| clause_name(get_anchors[c['id']][:label], c&.at(ns("./title"))&.content, s, nil) c.elements.reject { |c1| c1.name == "title" }.each do |c1| parse(c1, s) end end end end end def initial_anchor_names(d) d.xpath(ns(FRONT_CLAUSE)).each do |c| preface_names(c) sequential_asset_names(c) end middle_section_asset_names(d) clause_names(d, 0) termnote_anchor_names(d) termexample_anchor_names(d) end def middle(isoxml, out) middle_title(out) clause isoxml, out annex isoxml, out bibliography isoxml, out end def make_body2(body, docxml) body.div **{ class: "WordSection2" } do |div2| info docxml, div2 foreword docxml, div2 introduction docxml, div2 terms_defs docxml, div2, 0 div2.p { |p| p << " " } # placeholder end section_break(body) end def middle(isoxml, out) middle_title(out) clause isoxml, out annex isoxml, out bibliography isoxml, out end def termdef_parse(node, out) set_termdomain("") node.children.each { |n| parse(n, out) } end def annex_name_lbl(clause, num) l10n("#{@annex_lbl} #{num}") end def clause_names(docxml, sect_num) q = "//clause[parent::sections]" @topnum = nil lvl = 0 docxml.xpath(ns(q)).each do |c| container_names(c, 0) sect_num, lvl = sect_names(c, nil, sect_num, 0, lvl) end end def container_names(clause, lvl) if clause["container"] @anchors[clause["id"]] = { label: nil, xref: clause.at(ns("./title"))&.text, level: lvl+1 } end clause.xpath(ns("./clause | ./term | ./terms | "\ "./definitions")).each do |c| container_names(c, clause["container"] ? lvl+1 : lvl) end end def sect_names(clause, num, i, lvl, prev_lvl) return i if clause.nil? curr = i if clause["container"] retlvl = lvl+1 else retlvl = lvl i+=1 curr = i name = num.nil? ? i.to_s : "#{num}.#{i}" @anchors[clause["id"]] = { label: name, xref: l10n("#{@clause_lbl} #{name}"), level: lvl+1 } end prev = lvl j = 0 clause.xpath(ns("./clause | ./term | ./terms | "\ "./definitions")).each do |c| if clause["container"] i, lvl = sect_names(c, num, i, lvl, lvl) else j, prev = sect_names(c, name, j, lvl+1, prev) end end i = j if j >0 i = curr if lvl < prev [i, prev] end def annex_naming(c, num, lvl, i) if c["guidance"] then annex_names1(c, "#{num}E", lvl + 1) else i+= 1 annex_names1(c, "#{num}.#{i}", lvl + 1) end i end def annex_names(clause, num) @anchors[clause["id"]] = { label: annex_name_lbl(clause, num), xref: "#{@annex_lbl} #{num}", level: 1 } i = 0 clause.xpath(ns("./clause")).each do |c| i = annex_naming(c, num, 1, i) end hierarchical_asset_names(clause, num) end def annex_names1(clause, num, level) @anchors[clause["id"]] = { label: num, xref: "#{@annex_lbl} #{num}", level: level } i = 0 clause.xpath(ns("./clause")).each do |c| i = annex_naming(c, num, level, i) end end def clause(isoxml, out) isoxml.xpath(ns(MIDDLE_CLAUSE)).each do |c| out.div **attr_code(id: c["id"]) do |s| clause_name(get_anchors[c['id']][:label], c&.at(ns("./title"))&.content, s, class: c["container"] ? "containerhdr" : nil ) c.elements.reject { |c1| c1.name == "title" }.each do |c1| parse(c1, s) end end end end def clause_parse_title(node, div, c1, out) if node["inline-header"] == "true" inline_header_title(out, node, c1) else attrs = { class: node["container"] ? "containerhdr" : nil } div.send "h#{get_anchors[node['id']][:level]}", **attr_code(attrs) do |h| lbl = get_anchors[node['id']][:label] h << "#{lbl}. " if lbl c1&.children&.each { |c2| parse(c2, h) } end end end def ol_depth(node) ol_style(node["type"]) end end end end