]>
pere.pagekite.me Git - text-mekanikerord.git/blob - make-glossary
6 from lxml
.etree
import tostring
10 filemakerxml
= 'meksme-utf8.xml'
12 tree
= etree
.parse(filemakerxml
)
16 #print(tostring(tree))
19 'topic', 'se', 'desc-se', 'desc-nb', 'nb', 'sv', 'fi', 'en', 'is',
39 'fuolahas': 'bearbeiding',
43 resultset
= root
.find("{http://www.filemaker.com/fmpxmlresult}RESULTSET")
46 for row
in resultset
.getchildren():
49 for col
in row
.findall("{http://www.filemaker.com/fmpxmlresult}COL"):
50 t
= col
.getchildren()[0].text
53 t
= re
.sub(r
'\s+', ' ', t
)
58 def langsort(lang
, e
):
60 return locale
.strxfrm(e
[lang
])
62 return locale
.strxfrm(e
['se'])
64 def make_glossary_docbook(lang
, desccodes
, langcodes
, output
='glossary.xml'):
66 E
= lxml
.builder
.ElementMaker(
68 # 'xi': "http://www.w3.org/2001/XInclude",
72 def indexit(entry
, wlist
, lang
=None):
73 for w
in wlist
.split(","):
75 if lang
and '[' not in w
:
77 entry
.append(E
.indexterm(E
.primary(w
)))
79 glossary
= E
.glossary()
80 for e
in sorted(words
, key
=lambda x
: langsort(lang
, x
)):
81 ldesc
= 'desc-%s' % lang
82 if 'topic' in e
and lang
in topicmap
:
83 e
['topic'] = topicmap
[lang
][e
['topic']]
85 w
= e
[lang
].split(',')
98 print("warning: term %s missing primary language %s description" % (e
[lang
], lang
))
99 entry
= E
.glossentry(id=id)
100 if list_topic
and 'topic' in e
:
101 entry
.append(E
.glossterm('%s [%s]' % (e
[lang
], e
['topic'])))
103 entry
.append(E
.glossterm(e
[lang
]))
104 indexit(entry
, e
[lang
])
107 if l
!= lang
and l
in e
:
108 lstr
+= "%s (%s) " % (e
[l
], l
)
109 # Add foreign words to index, split on comma
110 indexit(entry
, e
[l
], l
)
112 entry
.append(E
.glossdef(E
.para(lstr
)))
114 # only single word witout translations, skip it
116 for desccode
in desccodes
:
117 codestr
= 'desc-%s' % desccode
119 entry
.append(E
.glossdef(E
.para("(%s): %s" % (desccode
,
121 glossary
.append(entry
)
123 # Add See also entries pointing to main entry
126 t
= t
.strip().lstrip()
127 #print("info: Adding see also entry for %s" % t)
128 seeentry
= E
.glossentry()
129 seeentry
.append(E
.glossterm(t
))
130 seeentry
.append(E
.glosssee(otherterm
=id))
131 glossary
.append(seeentry
)
133 def glosstermlocale(x
):
134 # Look up glossterm (FIXME figure out more robust way)
135 t
= x
.getchildren()[0].text
137 return locale
.strxfrm(t
)
140 # Sort list to mix seealso entries into their correct location.
141 glossary
[:] = sorted(glossary
, key
=glosstermlocale
)
144 print("info: dictionary contain %d entries" % l
)
146 content
= lxml
.etree
.tostring(glossary
,
148 xml_declaration
=True,
151 with open(output
, 'wb') as f
:
155 parser
= argparse
.ArgumentParser()
156 parser
.add_argument("langcode", help="language code to generate glossary for")
157 parser
.add_argument("--output", help="where to store the glossary")
158 args
= parser
.parse_args()
160 locale
.setlocale(locale
.LC_ALL
, '')
162 if 'nb' == args
.langcode
:
163 print("Norsk/bokmål")
165 make_glossary_docbook(lang
='nb', desccodes
=('nb',),
166 langcodes
=('en', 'se', 'sv', 'da', 'fi', 'is',),
168 elif 'se' == args
.langcode
:
171 make_glossary_docbook(lang
='se', desccodes
=('se', 'nb'),
172 langcodes
=('nb', 'en', 'sv', 'da', 'fi', 'is',),
174 elif 'en' == args
.langcode
:
177 make_glossary_docbook(lang
='en', desccodes
=('en'),
178 langcodes
=('en', 'nb', 'se', 'sv', 'da', 'fi', 'is',),
181 print("error: Unknown language code %s" % args
.langcode
)