1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
|
#!/usr/bin/python3
# -*- coding: utf-8 -*-
#
# Generate a translit_combining file from a UnicodeData file.
# Copyright (C) 2015-2019 Free Software Foundation, Inc.
# This file is part of the GNU C Library.
#
# The GNU C Library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# The GNU C Library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with the GNU C Library; if not, see
# <https://www.gnu.org/licenses/>.
'''
Generate a translit_combining file from UnicodeData.txt
To see how this script is used, call it with the “-h” option:
$ ./gen_translit_combining -h
… prints usage message …
'''
import argparse
import time
import unicode_utils
def read_input_file(filename):
'''Reads the original glibc translit_combining file to get the
original head and tail.
We want to replace only the part of the file between
“translit_start” and “translit_end”
'''
head = tail = ''
with open(filename, mode='r') as translit_file:
for line in translit_file:
head = head + line
if line.startswith('translit_start'):
break
for line in translit_file:
if line.startswith('translit_end'):
tail = line
break
for line in translit_file:
tail = tail + line
return (head, tail)
def output_head(translit_file, unicode_version, head=''):
'''Write the header of the output file, i.e. the part of the file
before the “translit_start” line.
'''
if ARGS.input_file and head:
translit_file.write(head)
else:
translit_file.write('escape_char /\n')
translit_file.write('comment_char %\n')
translit_file.write(unicode_utils.COMMENT_HEADER)
translit_file.write('\n')
translit_file.write('% Transliterations that remove all ')
translit_file.write('combining characters (accents,\n')
translit_file.write('% pronounciation marks, etc.).\n')
translit_file.write('% Generated automatically from UnicodeData.txt '
+ 'by gen_translit_combining.py '
+ 'on {:s} '.format(time.strftime('%Y-%m-%d'))
+ 'for Unicode {:s}.\n'.format(unicode_version))
translit_file.write('\n')
translit_file.write('LC_CTYPE\n')
translit_file.write('\n')
translit_file.write('translit_start\n')
def output_tail(translit_file, tail=''):
'''Write the tail of the output file'''
if ARGS.input_file and tail:
translit_file.write(tail)
else:
translit_file.write('translit_end\n')
translit_file.write('\n')
translit_file.write('END LC_CTYPE\n')
def is_combining_remove(code_point):
'''Check whether this is a combining character which should be listed
in the section of the translit_combining file where combining
characters are replaced by empty strings.
We ignore combining characters from many scripts here because
the original translit_combining file didn’t do this for the
combining characters from these scripts either and I am not
sure yet whether this would be useful to do for all combining
characters or not. For the moment I think it is better to keep
close to the spirit of the original file.
'''
if not unicode_utils.is_combining(code_point):
return False
name = unicode_utils.UNICODE_ATTRIBUTES[code_point]['name']
for substring in ('DEVANAGARI',
'BENGALI',
'CYRILLIC',
'SYRIAC',
'THAANA',
'NKO',
'GURMUKHI',
'TAMIL',
'GUJARATI',
'ORIYA',
'TELUGU',
'KANNADA',
'MALAYALAM',
'SINHALA',
'THAI',
'LAO',
'TIBETAN',
'MYANMAR',
'ETHIOPIC',
'TAGALOG',
'HANUNOO',
'BUHID',
'TAGBANWA',
'KHMER',
'MONGOLIAN',
'LIMBU',
'NEW TAI LUE',
'BUGINESE',
'BALINESE',
'SUNDANESE',
'LEPCHA',
'IDEOGRAPHIC',
'HANGUL',
'SYLOTI',
'SAURASHTRA',
'KAYAH',
'REJANG',
'CHAM',
'VARIATION SELECTOR',
'KHAROSHTHI',
'MUSICAL SYMBOL',
'SAMARITAN',
'MANDAIC',
'TAI THAM',
'BATAK',
'VEDIC',
'COPTIC',
'TIFINAGH',
'BAMUM',
'JAVANESE',
'TAI VIET',
'MEETEI',
'MANICHAEAN',
'BRAHMI',
'KAITHI',
'CHAKMA',
'MAHAJANI',
'SHARADA',
'KHOJKI',
'KHUDAWADI',
'GRANTHA',
'TIRHUTA',
'SIDDHAM',
'MODI VOWEL',
'MODI SIGN',
'TAKRI',
'BASSA VAH',
'PAHAWH HMONG',
'MIAO',
'DUPLOYAN',
'MENDE KIKAKUI',
'AHOM',
'SIGNWRITING'
):
if substring in name:
return False
return True
def canonical_decompose(code_point):
'''http://www.unicode.org/reports/tr44/#Character_Decomposition_Mappings
In some instances a canonical mapping or a compatibility mapping
may consist of a single character. For a canonical mapping, this
indicates that the character is a canonical equivalent of another
single character. For a compatibility mapping, this indicates that
the character is a compatibility equivalent of another single
character.
A canonical mapping may also consist of a pair of characters, but
is never longer than two characters. When a canonical mapping
consists of a pair of characters, the first character may itself
be a character with a decomposition mapping, but the second
character never has a decomposition mapping.
We ignore the canonical decomposition for code points
matching certain substrings because the original translit_combining
file didn’t include these types of characters either. I am unsure
about the usefulness of including them and want to keep close
to the spirit of the original file for the moment.
'''
name = unicode_utils.UNICODE_ATTRIBUTES[code_point]['name']
for substring in ('MUSICAL SYMBOL',
'CJK COMPATIBILITY IDEOGRAPH',
'BALINESE',
'KAITHI LETTER',
'CHAKMA VOWEL',
'GRANTHA VOWEL',
'TIRHUTA VOWEL',
'SIDDHAM VOWEL'):
if substring in name:
return []
decomposition = unicode_utils.UNICODE_ATTRIBUTES[
code_point]['decomposition']
if decomposition and not decomposition.startswith('<'):
decomposed_code_points = [int(x, 16) for x in decomposition.split(' ')]
if decomposed_code_points:
cd0 = canonical_decompose(decomposed_code_points[0])
if cd0:
decomposed_code_points = cd0 + decomposed_code_points[1:]
return decomposed_code_points
else:
return []
def special_decompose(code_point_list):
'''
Decompositions which are not canonical or which are not in
UnicodeData.txt at all but some of these were used in the original
translit_combining file in glibc and they seemed to make sense.
I want to keep the update of translit_combining close to the
spirit of the original file, therefore I added these special
decomposition rules here.
'''
special_decompose_dict = {
# Ø U+00D8 is already handled in translit_neutral. But
# translit_combining is usually included after translit_neutral
# and Ǿ U+01FE LATIN CAPITAL LETTER O WITH STROKE AND ACUTE
# has a canonical decomposition to Ø U+00D8 and we want to
# further decompose this to U+004F.
(0x00D8,): [0x004F], # Ø → O
# ø U+00F8 is already handled in translit_neutral. But
# translit_combining is usually included after translit_neutral
# and ǿ U+01FF LATIN SMALL LETTER O WITH STROKE AND ACUTE
# has a canonical decomposition to ø U+00F8 and we want to
# further decompose this to U+006F.
(0x00F8,): [0x006F], # ø → o
# æ U+00E6 is already in translit_compat because ligatures
# are handled in translit_compat. But ǣ U+01E3 has a
# canonical decomposition to U+00E6, U+0304 and we want to
# further decompose this to “ae”.
(0x00E6,): [0x0061, 0x0065], # æ → ae
# Æ U+00C6 is already in translit_compat because ligatures
# are handled in translit_compat. But Ǣ U+01E2 has a
# canonical decomposition to U+00C6, U+0304 and we want to
# further decompose this to “AE”
(0x00C6,): [0x0041, 0x0045], # Æ → AE
# U+05F2 HEBREW LIGATURE YIDDISH DOUBLE YOD is already in
# translit_compat because ligatures are handled in translit_compat.
# But U+FB1F has a canonical decomposition to U+05F2 and
# we want to further decompose this to U+05D9, U+05D9.
(0x05F2,): [0x05D9, 0x05D9], # ײ → יי
# 0x2002 has a <compat> decomposition to 0x0020 in UnicodeData.txt
# But U+2000 EN QUAD has a canonical decomposition U+2002
# and we want to further decompose this to U+0020.
(0x2002,): [0x0020], # EN SPACE → SPACE
# 0x2003 has a <compat> decomposition to 0x0020 in UnicodeData.txt
# But U+2001 EM QUAD has a canonical decomposition to U+2003
# and we want to further decompose this to U+0020.
(0x2003,): [0x0020], # EM SPACE → SPACE
# U+2260 ≠ has the canonical decomposition U+003D U+0338
# (= followed by ̸). After stripping the combining characters,
# the result is only = which reverses the meaning.
# Therefore, we add a special rules here for such mathematical
# negations:
(0x21AE,): [0x0021, 0x003C, 0x002D, 0x003E], # ↮ → !<->
(0x21CD,): [0x0021, 0x003C, 0x003D], # ⇍ → !<=
(0x21CE,): [0x0021, 0x003C, 0x003D, 0x003E], # ⇎ → !<=>
(0x21CF,): [0x0021, 0x003D, 0x003E], # ⇏ → !=>
(0x2204,): [0x0021, 0x2203], # ∄ → !∃
(0x2209,): [0x0021, 0x2208], # ∉ → !∈
(0x220C,): [0x0021, 0x220B], # ∌ → !∋
(0x2224,): [0x0021, 0x2223], # ∤ → !∣
(0x2226,): [0x0021, 0x2225], # ∦ → !∥
(0x2241,): [0x0021, 0x007E], # ≁ → !~
(0x2244,): [0x0021, 0x007E, 0x002D], # ≄ → !~-
(0x2247,): [0x0021, 0x007E, 0x003D], # ≇ → !~=
(0x2249,): [0x0021, 0x007E, 0x007E], # ≉ → !~~
(0x2260,): [0x0021, 0x003D], # ≠ → !=
(0x2262,): [0x0021, 0x003D, 0x003D], # ≢ → !==
(0x226D,): [0x0021, 0x224D], # ≭ → !≍
(0x226E,): [0x0021, 0x003C], # ≮ → !<
(0x226F,): [0x0021, 0x003E], # ≯ → !>
(0x2270,): [0x0021, 0x003C, 0x003D], # ≰ → !<=
(0x2271,): [0x0021, 0x003E, 0x003D], # ≱ → !>=
(0x2274,): [0x0021, 0x003C, 0x007E], # ≴ → !<~
(0x2275,): [0x0021, 0x003E, 0x007E], # ≵ → !>~
(0x2278,): [0x0021, 0x003C, 0x003E], # ≸ → !<>
(0x2279,): [0x0021, 0x003E, 0x003C], # ≹ → !><
(0x2280,): [0x0021, 0x227A], # ⊀ → !≺
(0x2281,): [0x0021, 0x227B], # ⊁ → !≻
(0x2284,): [0x0021, 0x2282], # ⊄ → !⊂
(0x2285,): [0x0021, 0x2283], # ⊅ → !⊃
(0x2288,): [0x0021, 0x2282, 0x003D], # ⊈ → !⊂=
(0x2289,): [0x0021, 0x2283, 0x003D], # ⊉ → !⊃=
(0x22AC,): [0x0021, 0x22A2], # ⊬ → !⊢
(0x22AD,): [0x0021, 0x22A8], # ⊭ → !⊨
(0x22AE,): [0x0021, 0x22A9], # ⊮ → !⊩
(0x22AF,): [0x0021, 0x22AB], # ⊯ → !⊫
(0x22E0,): [0x0021, 0x227C], # ⋠ → !≼
(0x22E1,): [0x0021, 0x227D], # ⋡ → !≽
(0x22E2,): [0x0021, 0x2291], # ⋢ → !⊑
(0x22E3,): [0x0021, 0x2292], # ⋣ → !⊒
(0x22EA,): [0x0021, 0x22B2], # ⋪ → !⊲
(0x22EB,): [0x0021, 0x22B3], # ⋫ → !⊳
(0x22EC,): [0x0021, 0x22B4], # ⋬ → !⊴
(0x22ED,): [0x0021, 0x22B5], # ⋭ → !⊵
(0x2ADC,): [0x0021, 0x2ADD], # ⫝̸ → !⫝
# Special rule for 〈 U+3008 is added
# because 〉 U+2329 has the canonical decomposition U+3008
# and we want to further decompose this to > U+003C.
(0x3008,): [0x003C], # 〈 → <
# Special rule for 〉 U+3009 is added
# because 〉 U+232A has the canonical decomposition U+3009
# and we want to further decompose this to < U+003E.
(0x3009,): [0x003E], # 〉→ >
}
if tuple(code_point_list) in special_decompose_dict:
return special_decompose_dict[tuple(code_point_list)]
else:
return code_point_list
def output_combining_remove(translit_file):
'''Write the section of the translit_combining file where combining
characters are replaced by empty strings.
'''
translit_file.write('\n')
for code_point in sorted(unicode_utils.UNICODE_ATTRIBUTES):
name = unicode_utils.UNICODE_ATTRIBUTES[code_point]['name']
if is_combining_remove(code_point):
translit_file.write('% {:s}\n'.format(name))
translit_file.write('{:s} ""\n'.format(
unicode_utils.ucs_symbol(code_point)))
translit_file.write('\n')
def output_decompositions(translit_file):
'''Write the section of the translit_combining file where characters
characters are decomposed and combining characters stripped from
the decompositions.
'''
for code_point in sorted(unicode_utils.UNICODE_ATTRIBUTES):
if special_decompose([code_point]) != [code_point]:
decomposed_code_points = [special_decompose([code_point])]
else:
decomposed_code_points = [canonical_decompose(code_point)]
if decomposed_code_points[0]:
while True:
special_decomposed_code_points = special_decompose(
decomposed_code_points[-1])
if (special_decomposed_code_points
!= decomposed_code_points[-1]):
decomposed_code_points.append(
special_decomposed_code_points)
continue
special_decomposed_code_points = []
for decomposed_code_point in decomposed_code_points[-1]:
special_decomposed_code_points += special_decompose(
[decomposed_code_point])
if (special_decomposed_code_points
== decomposed_code_points[-1]):
break
decomposed_code_points.append(
special_decomposed_code_points)
for index in range(0, len(decomposed_code_points)):
decomposed_code_points[index] = [
x for x in decomposed_code_points[index]
if not is_combining_remove(x)]
if decomposed_code_points[0]:
translit_file.write('% {:s}\n'.format(
unicode_utils.UNICODE_ATTRIBUTES[code_point]['name']))
translit_file.write('{:s} '.format(
unicode_utils.ucs_symbol(code_point)))
for index in range(0, len(decomposed_code_points)):
if index > 0:
translit_file.write(';')
if len(decomposed_code_points[index]) > 1:
translit_file.write('"')
for decomposed_code_point in decomposed_code_points[index]:
translit_file.write('{:s}'.format(
unicode_utils.ucs_symbol(decomposed_code_point)))
if len(decomposed_code_points[index]) > 1:
translit_file.write('"')
translit_file.write('\n')
translit_file.write('\n')
def output_transliteration(translit_file):
'''Write the new transliteration to the output file'''
output_combining_remove(translit_file)
output_decompositions(translit_file)
if __name__ == "__main__":
PARSER = argparse.ArgumentParser(
description='''
Generate a translit_combining file from UnicodeData.txt.
''')
PARSER.add_argument(
'-u', '--unicode_data_file',
nargs='?',
type=str,
default='UnicodeData.txt',
help=('The UnicodeData.txt file to read, '
+ 'default: %(default)s'))
PARSER.add_argument(
'-i', '--input_file',
nargs='?',
type=str,
help=''' The original glibc/localedata/locales/translit_combining
file.''')
PARSER.add_argument(
'-o', '--output_file',
nargs='?',
type=str,
default='translit_combining.new',
help='''The new translit_combining file, default: %(default)s. If the
original glibc/localedata/locales/translit_combining file has
been given as an option, the header up to the
“translit_start” line and the tail from the “translit_end”
line to the end of the file will be copied unchanged into the
output file. ''')
PARSER.add_argument(
'--unicode_version',
nargs='?',
required=True,
type=str,
help='The Unicode version of the input files used.')
ARGS = PARSER.parse_args()
unicode_utils.fill_attributes(ARGS.unicode_data_file)
HEAD = TAIL = ''
if ARGS.input_file:
(HEAD, TAIL) = read_input_file(ARGS.input_file)
with open(ARGS.output_file, mode='w') as TRANSLIT_FILE:
output_head(TRANSLIT_FILE, ARGS.unicode_version, head=HEAD)
output_transliteration(TRANSLIT_FILE)
output_tail(TRANSLIT_FILE, tail=TAIL)
|