require "isodoc"
require_relative "metadata"
require "fileutils"
module IsoDoc
module Mpfd
# A {Converter} implementation that generates Word output, and a document
# schema encapsulation of the document for validation
class WordConvert < IsoDoc::WordConvert
def rsd_html_path(file)
File.join(File.dirname(__FILE__), File.join("html", file))
end
def initialize(options)
@libdir = File.dirname(__FILE__)
super
FileUtils.cp html_doc_path("logo.jpg"), "logo.jpg"
FileUtils.cp html_doc_path('mpfa-logo-no-text@4x.png'), "mpfa-logo-no-text@4x.png"
@files_to_delete << "logo.jpg"
@files_to_delete << "mpfa-logo-no-text@4x.png"
end
def default_fonts(options)
{
bodyfont: (options[:script] == "Hans" ? '"SimSun",serif' : '"Arial",sans-serif'),
headerfont: (options[:script] == "Hans" ? '"SimHei",sans-serif' : '"Arial",sans-serif'),
monospacefont: '"Courier New",monospace'
}
end
def default_file_locations(options)
{
htmlstylesheet: html_doc_path("htmlstyle.scss"),
htmlcoverpage: html_doc_path("html_rsd_titlepage.html"),
htmlintropage: html_doc_path("html_rsd_intro.html"),
scripts: html_doc_path("scripts.html"),
wordstylesheet: html_doc_path("wordstyle.scss"),
standardstylesheet: html_doc_path("rsd.scss"),
header: html_doc_path("header.html"),
wordcoverpage: html_doc_path("word_rsd_titlepage.html"),
wordintropage: html_doc_path("word_rsd_intro.html"),
ulstyle: "l3",
olstyle: "l2",
}
end
def metadata_init(lang, script, labels)
@meta = Metadata.new(lang, script, labels)
end
def make_body(xml, docxml)
body_attr = { lang: "EN-US", link: "blue", vlink: "#954F72" }
xml.body **body_attr do |body|
make_body1(body, docxml)
make_body2(body, docxml)
make_body3(body, docxml)
end
end
def make_body2(body, docxml)
body.div **{ class: "WordSection2" } do |div2|
info docxml, div2
div2.p { |p| p << " " } # placeholder
end
#body.br **{ clear: "all", style: "page-break-before:auto;mso-break-type:section-break;" }
section_break(body)
end
def title(isoxml, _out)
main = isoxml&.at(ns("//title[@language='en']"))&.text
set_metadata(:doctitle, main)
end
def generate_header(filename, dir)
return unless @header
template = Liquid::Template.parse(File.read(@header, encoding: "UTF-8"))
meta = @meta.get
meta[:filename] = filename
params = meta.map { |k, v| [k.to_s, v] }.to_h
File.open("header.html", "w") { |f| f.write(template.render(params)) }
@files_to_delete << "header.html"
"header.html"
end
def header_strip(h)
h = h.to_s.gsub(%r{
}, " ").sub(/<\/?h[12][^>]*>/, "")
h1 = to_xhtml_fragment(h.dup)
h1.traverse do |x|
x.replace(" ") if x.name == "span" &&
/mso-tab-count/.match(x["style"])
x.remove if x.name == "span" && x["class"] == "MsoCommentReference"
x.remove if x.name == "a" && x["epub:type"] == "footnote"
x.replace(x.children) if x.name == "a"
end
from_xhtml(h1)
end
def info(isoxml, out)
@meta.security isoxml, out
super
end
def annex_name(annex, name, div)
div.h1 **{ class: "Annex" } do |t|
t << "#{get_anchors[annex['id']][:label]} "
t << "#{name.text}"
end
end
def annex_name_lbl(clause, num)
obl = l10n("(#{@inform_annex_lbl})")
obl = l10n("(#{@norm_annex_lbl})") if clause["obligation"] == "normative"
l10n("#{@annex_lbl} #{num} #{obl}")
end
def pre_parse(node, out)
out.pre node.text # content.gsub(/, "<").gsub(/>/, ">")
end
def term_defs_boilerplate(div, source, term, preface)
if source.empty? && term.nil?
div << @no_terms_boilerplate
else
div << term_defs_boilerplate_cont(source, term)
end
end
def i18n_init(lang, script)
super
@annex_lbl = "Appendix"
end
def error_parse(node, out)
# catch elements not defined in ISO
case node.name
when "pre"
pre_parse(node, out)
when "keyword"
out.span node.text, **{ class: "keyword" }
else
super
end
end
def fileloc(loc)
File.join(File.dirname(__FILE__), loc)
end
def i18n_init(lang, script)
super
y = if lang == "en"
YAML.load_file(File.join(File.dirname(__FILE__), "i18n-en.yaml"))
elsif lang == "zh" && script == "Hans"
YAML.load_file(File.join(File.dirname(__FILE__),
"i18n-zh-Hans.yaml"))
else
YAML.load_file(File.join(File.dirname(__FILE__), "i18n-zh-Hans.yaml"))
end
@labels = @labels.merge(y)
@clause_lbl = y["clause"]
end
def terms_defs_title(f)
return f&.at(ns("./title"))&.content
end
TERM_CLAUSE = "//preface/terms | "\
"//preface/clause[descendant::terms]".freeze
def terms_defs(isoxml, out, num)
f = isoxml.at(ns(TERM_CLAUSE)) or return num
out.div **attr_code(id: f["id"]) do |div|
clause_name(nil, terms_defs_title(f), div, nil)
f.elements.each do |e|
parse(e, div) unless %w{title source}.include? e.name
end
end
num
end
FRONT_CLAUSE = "//*[parent::preface]".freeze
#FRONT_CLAUSE = "//clause[parent::preface] | //terms[parent::preface]".freeze
def preface(isoxml, out)
isoxml.xpath(ns(FRONT_CLAUSE)).each do |c|
if c.name == "terms" then terms_defs isoxml, out, 0
else
out.div **attr_code(id: c["id"]) do |s|
clause_name(get_anchors[c['id']][:label],
c&.at(ns("./title"))&.content, s, nil)
c.elements.reject { |c1| c1.name == "title" }.each do |c1|
parse(c1, s)
end
end
end
end
end
def initial_anchor_names(d)
d.xpath(ns(FRONT_CLAUSE)).each do |c|
preface_names(c)
sequential_asset_names(c)
end
middle_section_asset_names(d)
clause_names(d, 0)
termnote_anchor_names(d)
termexample_anchor_names(d)
end
def middle(isoxml, out)
middle_title(out)
clause isoxml, out
annex isoxml, out
bibliography isoxml, out
end
def make_body2(body, docxml)
body.div **{ class: "WordSection2" } do |div2|
info docxml, div2
foreword docxml, div2
introduction docxml, div2
terms_defs docxml, div2, 0
div2.p { |p| p << " " } # placeholder
end
section_break(body)
end
def middle(isoxml, out)
middle_title(out)
clause isoxml, out
annex isoxml, out
bibliography isoxml, out
end
def termdef_parse(node, out)
set_termdomain("")
node.children.each { |n| parse(n, out) }
end
def annex_name_lbl(clause, num)
l10n("#{@annex_lbl} #{num}")
end
def clause_names(docxml, sect_num)
q = "//clause[parent::sections]"
@topnum = nil
lvl = 0
docxml.xpath(ns(q)).each do |c|
container_names(c, 0)
sect_num, lvl = sect_names(c, nil, sect_num, 0, lvl)
end
end
def container_names(clause, lvl)
if clause["container"]
@anchors[clause["id"]] =
{ label: nil, xref: clause.at(ns("./title"))&.text, level: lvl+1 }
end
clause.xpath(ns("./clause | ./term | ./terms | "\
"./definitions")).each do |c|
container_names(c, clause["container"] ? lvl+1 : lvl)
end
end
def sect_names(clause, num, i, lvl, prev_lvl)
return i if clause.nil?
curr = i
if clause["container"]
retlvl = lvl+1
else
retlvl = lvl
i+=1
curr = i
name = num.nil? ? i.to_s : "#{num}.#{i}"
@anchors[clause["id"]] = { label: name, xref: l10n("#{@clause_lbl} #{name}"), level: lvl+1 }
end
prev = lvl
j = 0
clause.xpath(ns("./clause | ./term | ./terms | "\
"./definitions")).each do |c|
if clause["container"]
i, lvl = sect_names(c, num, i, lvl, lvl)
else
j, prev = sect_names(c, name, j, lvl+1, prev)
end
end
i = j if j >0
i = curr if lvl < prev
[i, prev]
end
def annex_naming(c, num, lvl, i)
if c["guidance"] then annex_names1(c, "#{num}E", lvl + 1)
else
i+= 1
annex_names1(c, "#{num}.#{i}", lvl + 1)
end
i
end
def annex_names(clause, num)
@anchors[clause["id"]] = { label: annex_name_lbl(clause, num),
xref: "#{@annex_lbl} #{num}", level: 1 }
i = 0
clause.xpath(ns("./clause")).each do |c|
i = annex_naming(c, num, 1, i)
end
hierarchical_asset_names(clause, num)
end
def annex_names1(clause, num, level)
@anchors[clause["id"]] = { label: num, xref: "#{@annex_lbl} #{num}",
level: level }
i = 0
clause.xpath(ns("./clause")).each do |c|
i = annex_naming(c, num, level, i)
end
end
def clause(isoxml, out)
isoxml.xpath(ns(MIDDLE_CLAUSE)).each do |c|
out.div **attr_code(id: c["id"]) do |s|
clause_name(get_anchors[c['id']][:label],
c&.at(ns("./title"))&.content, s, class: c["container"] ? "containerhdr" : nil )
c.elements.reject { |c1| c1.name == "title" }.each do |c1|
parse(c1, s)
end
end
end
end
def clause_parse_title(node, div, c1, out)
if node["inline-header"] == "true"
inline_header_title(out, node, c1)
else
attrs = { class: node["container"] ? "containerhdr" : nil }
div.send "h#{get_anchors[node['id']][:level]}", **attr_code(attrs) do |h|
lbl = get_anchors[node['id']][:label]
h << "#{lbl}. " if lbl
c1&.children&.each { |c2| parse(c2, h) }
end
end
end
def ol_depth(node)
ol_style(node["type"])
end
end
end
end