HarfBuzz text shaping engine
http://harfbuzz.github.io/
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
259 lines
6.2 KiB
259 lines
6.2 KiB
#!/usr/bin/env python |
|
|
|
from __future__ import print_function, division, absolute_import |
|
|
|
import io, sys |
|
|
|
if len (sys.argv) != 4: |
|
print ("usage: ./gen-indic-table.py IndicSyllabicCategory.txt IndicPositionalCategory.txt Blocks.txt", file=sys.stderr) |
|
sys.exit (1) |
|
|
|
ALLOWED_SINGLES = [0x00A0, 0x25CC] |
|
ALLOWED_BLOCKS = [ |
|
'Basic Latin', |
|
'Latin-1 Supplement', |
|
'Devanagari', |
|
'Bengali', |
|
'Gurmukhi', |
|
'Gujarati', |
|
'Oriya', |
|
'Tamil', |
|
'Telugu', |
|
'Kannada', |
|
'Malayalam', |
|
'Sinhala', |
|
'Myanmar', |
|
'Khmer', |
|
'Vedic Extensions', |
|
'General Punctuation', |
|
'Superscripts and Subscripts', |
|
'Devanagari Extended', |
|
'Myanmar Extended-B', |
|
'Myanmar Extended-A', |
|
] |
|
|
|
files = [io.open (x, encoding='utf-8') for x in sys.argv[1:]] |
|
|
|
headers = [[f.readline () for i in range (2)] for f in files] |
|
|
|
data = [{} for f in files] |
|
values = [{} for f in files] |
|
for i, f in enumerate (files): |
|
for line in f: |
|
|
|
j = line.find ('#') |
|
if j >= 0: |
|
line = line[:j] |
|
|
|
fields = [x.strip () for x in line.split (';')] |
|
if len (fields) == 1: |
|
continue |
|
|
|
uu = fields[0].split ('..') |
|
start = int (uu[0], 16) |
|
if len (uu) == 1: |
|
end = start |
|
else: |
|
end = int (uu[1], 16) |
|
|
|
t = fields[1] |
|
|
|
for u in range (start, end + 1): |
|
data[i][u] = t |
|
values[i][t] = values[i].get (t, 0) + end - start + 1 |
|
|
|
# Merge data into one dict: |
|
defaults = ('Other', 'Not_Applicable', 'No_Block') |
|
for i,v in enumerate (defaults): |
|
values[i][v] = values[i].get (v, 0) + 1 |
|
combined = {} |
|
for i,d in enumerate (data): |
|
for u,v in d.items (): |
|
if i == 2 and not u in combined: |
|
continue |
|
if not u in combined: |
|
combined[u] = list (defaults) |
|
combined[u][i] = v |
|
combined = {k:v for k,v in combined.items() if k in ALLOWED_SINGLES or v[2] in ALLOWED_BLOCKS} |
|
data = combined |
|
del combined |
|
num = len (data) |
|
|
|
for u in [0x17CD, 0x17CE, 0x17CF, 0x17D0, 0x17D3]: |
|
if data[u][0] == 'Other': |
|
data[u][0] = "Vowel_Dependent" |
|
|
|
# Move the outliers NO-BREAK SPACE and DOTTED CIRCLE out |
|
singles = {} |
|
for u in ALLOWED_SINGLES: |
|
singles[u] = data[u] |
|
del data[u] |
|
|
|
print ("/* == Start of generated table == */") |
|
print ("/*") |
|
print (" * The following table is generated by running:") |
|
print (" *") |
|
print (" * ./gen-indic-table.py IndicSyllabicCategory.txt IndicPositionalCategory.txt Blocks.txt") |
|
print (" *") |
|
print (" * on files with these headers:") |
|
print (" *") |
|
for h in headers: |
|
for l in h: |
|
print (" * %s" % (l.strip())) |
|
print (" */") |
|
print () |
|
print ('#include "hb-ot-shape-complex-indic-private.hh"') |
|
print () |
|
|
|
# Shorten values |
|
short = [{ |
|
"Bindu": 'Bi', |
|
"Cantillation_Mark": 'Ca', |
|
"Joiner": 'ZWJ', |
|
"Non_Joiner": 'ZWNJ', |
|
"Number": 'Nd', |
|
"Visarga": 'Vs', |
|
"Vowel": 'Vo', |
|
"Vowel_Dependent": 'M', |
|
"Consonant_Prefixed": 'CPrf', |
|
"Other": 'x', |
|
},{ |
|
"Not_Applicable": 'x', |
|
}] |
|
all_shorts = [{},{}] |
|
|
|
# Add some of the values, to make them more readable, and to avoid duplicates |
|
|
|
|
|
for i in range (2): |
|
for v,s in short[i].items (): |
|
all_shorts[i][s] = v |
|
|
|
what = ["INDIC_SYLLABIC_CATEGORY", "INDIC_MATRA_CATEGORY"] |
|
what_short = ["ISC", "IMC"] |
|
for i in range (2): |
|
print () |
|
vv = sorted (values[i].keys ()) |
|
for v in vv: |
|
v_no_and = v.replace ('_And_', '_') |
|
if v in short[i]: |
|
s = short[i][v] |
|
else: |
|
s = ''.join ([c for c in v_no_and if ord ('A') <= ord (c) <= ord ('Z')]) |
|
if s in all_shorts[i]: |
|
raise Exception ("Duplicate short value alias", v, all_shorts[i][s]) |
|
all_shorts[i][s] = v |
|
short[i][v] = s |
|
print ("#define %s_%s %s_%s %s/* %3d chars; %s */" % |
|
(what_short[i], s, what[i], v.upper (), |
|
' '* ((48-1 - len (what[i]) - 1 - len (v)) // 8), |
|
values[i][v], v)) |
|
print () |
|
print ("#define _(S,M) INDIC_COMBINE_CATEGORIES (ISC_##S, IMC_##M)") |
|
print () |
|
print () |
|
|
|
total = 0 |
|
used = 0 |
|
last_block = None |
|
def print_block (block, start, end, data): |
|
global total, used, last_block |
|
if block and block != last_block: |
|
print () |
|
print () |
|
print (" /* %s */" % block) |
|
num = 0 |
|
assert start % 8 == 0 |
|
assert (end+1) % 8 == 0 |
|
for u in range (start, end+1): |
|
if u % 8 == 0: |
|
print () |
|
print (" /* %04X */" % u, end="") |
|
if u in data: |
|
num += 1 |
|
d = data.get (u, defaults) |
|
print ("%9s" % ("_(%s,%s)," % (short[0][d[0]], short[1][d[1]])), end="") |
|
|
|
total += end - start + 1 |
|
used += num |
|
if block: |
|
last_block = block |
|
|
|
uu = sorted (data.keys ()) |
|
|
|
last = -100000 |
|
num = 0 |
|
offset = 0 |
|
starts = [] |
|
ends = [] |
|
print ("static const INDIC_TABLE_ELEMENT_TYPE indic_table[] = {") |
|
for u in uu: |
|
if u <= last: |
|
continue |
|
block = data[u][2] |
|
|
|
start = u//8*8 |
|
end = start+1 |
|
while end in uu and block == data[end][2]: |
|
end += 1 |
|
end = (end-1)//8*8 + 7 |
|
|
|
if start != last + 1: |
|
if start - last <= 1+16*3: |
|
print_block (None, last+1, start-1, data) |
|
last = start-1 |
|
else: |
|
if last >= 0: |
|
ends.append (last + 1) |
|
offset += ends[-1] - starts[-1] |
|
print () |
|
print () |
|
print ("#define indic_offset_0x%04xu %d" % (start, offset)) |
|
starts.append (start) |
|
|
|
print_block (block, start, end, data) |
|
last = end |
|
ends.append (last + 1) |
|
offset += ends[-1] - starts[-1] |
|
print () |
|
print () |
|
occupancy = used * 100. / total |
|
page_bits = 12 |
|
print ("}; /* Table items: %d; occupancy: %d%% */" % (offset, occupancy)) |
|
print () |
|
print ("INDIC_TABLE_ELEMENT_TYPE") |
|
print ("hb_indic_get_categories (hb_codepoint_t u)") |
|
print ("{") |
|
print (" switch (u >> %d)" % page_bits) |
|
print (" {") |
|
pages = set ([u>>page_bits for u in starts+ends+list (singles.keys ())]) |
|
for p in sorted(pages): |
|
print (" case 0x%0Xu:" % p) |
|
for u,d in singles.items (): |
|
if p != u>>page_bits: continue |
|
print (" if (unlikely (u == 0x%04Xu)) return _(%s,%s);" % (u, short[0][d[0]], short[1][d[1]])) |
|
for (start,end) in zip (starts, ends): |
|
if p not in [start>>page_bits, end>>page_bits]: continue |
|
offset = "indic_offset_0x%04xu" % start |
|
print (" if (hb_in_range<hb_codepoint_t> (u, 0x%04Xu, 0x%04Xu)) return indic_table[u - 0x%04Xu + %s];" % (start, end-1, start, offset)) |
|
print (" break;") |
|
print ("") |
|
print (" default:") |
|
print (" break;") |
|
print (" }") |
|
print (" return _(x,x);") |
|
print ("}") |
|
print () |
|
print ("#undef _") |
|
for i in range (2): |
|
print |
|
vv = sorted (values[i].keys ()) |
|
for v in vv: |
|
print ("#undef %s_%s" % |
|
(what_short[i], short[i][v])) |
|
print () |
|
print ("/* == End of generated table == */") |
|
|
|
# Maintain at least 30% occupancy in the table */ |
|
if occupancy < 30: |
|
raise Exception ("Table too sparse, please investigate: ", occupancy)
|
|
|