mirror of
https://github.com/python/cpython.git
synced 2026-02-22 15:10:47 +00:00
Since the code for rule NR2 is already here, to support CJK unified ideographs and Tangut ideographs, it can also be used for other names derived by rule NR2.
1324 lines
49 KiB
Python
1324 lines
49 KiB
Python
#
|
|
# (re)generate unicode property and type databases
|
|
#
|
|
# This script converts Unicode database files to Modules/unicodedata_db.h,
|
|
# Modules/unicodename_db.h, and Objects/unicodetype_db.h
|
|
#
|
|
# history:
|
|
# 2000-09-24 fl created (based on bits and pieces from unidb)
|
|
# 2000-09-25 fl merged tim's splitbin fixes, separate decomposition table
|
|
# 2000-09-25 fl added character type table
|
|
# 2000-09-26 fl added LINEBREAK, DECIMAL, and DIGIT flags/fields (2.0)
|
|
# 2000-11-03 fl expand first/last ranges
|
|
# 2001-01-19 fl added character name tables (2.1)
|
|
# 2001-01-21 fl added decomp compression; dynamic phrasebook threshold
|
|
# 2002-09-11 wd use string methods
|
|
# 2002-10-18 mvl update to Unicode 3.2
|
|
# 2002-10-22 mvl generate NFC tables
|
|
# 2002-11-24 mvl expand all ranges, sort names version-independently
|
|
# 2002-11-25 mvl add UNIDATA_VERSION
|
|
# 2004-05-29 perky add east asian width information
|
|
# 2006-03-10 mvl update to Unicode 4.1; add UCD 3.2 delta
|
|
# 2008-06-11 gb add PRINTABLE_MASK for Atsuo Ishimoto's ascii() patch
|
|
# 2011-10-21 ezio add support for name aliases and named sequences
|
|
# 2012-01 benjamin add full case mappings
|
|
#
|
|
# written by Fredrik Lundh (fredrik@pythonware.com)
|
|
#
|
|
|
|
import dataclasses
|
|
import os
|
|
import sys
|
|
import re
|
|
import zipfile
|
|
|
|
from functools import partial
|
|
from textwrap import dedent
|
|
from typing import Iterator, List, Optional, Set, Tuple
|
|
|
|
SCRIPT = os.path.normpath(sys.argv[0])
|
|
VERSION = "3.3"
|
|
|
|
# The Unicode Database
|
|
# --------------------
|
|
# When changing UCD version please update
|
|
# * Doc/library/stdtypes.rst, and
|
|
# * Doc/library/unicodedata.rst
|
|
# * Doc/reference/lexical_analysis.rst (three occurrences)
|
|
UNIDATA_VERSION = "17.0.0"
|
|
UNICODE_DATA = "UnicodeData%s.txt"
|
|
COMPOSITION_EXCLUSIONS = "CompositionExclusions%s.txt"
|
|
EASTASIAN_WIDTH = "EastAsianWidth%s.txt"
|
|
UNIHAN = "Unihan%s.zip"
|
|
DERIVED_BIDI_CLASS = "extracted/DerivedBidiClass%s.txt"
|
|
DERIVED_CORE_PROPERTIES = "DerivedCoreProperties%s.txt"
|
|
DERIVEDNORMALIZATION_PROPS = "DerivedNormalizationProps%s.txt"
|
|
LINE_BREAK = "LineBreak%s.txt"
|
|
NAME_ALIASES = "NameAliases%s.txt"
|
|
NAMED_SEQUENCES = "NamedSequences%s.txt"
|
|
SPECIAL_CASING = "SpecialCasing%s.txt"
|
|
CASE_FOLDING = "CaseFolding%s.txt"
|
|
GRAPHEME_CLUSTER_BREAK = "auxiliary/GraphemeBreakProperty%s.txt"
|
|
EMOJI_DATA = "emoji/emoji-data%s.txt"
|
|
|
|
# Private Use Areas -- in planes 1, 15, 16
|
|
PUA_1 = range(0xE000, 0xF900)
|
|
PUA_15 = range(0xF0000, 0xFFFFE)
|
|
PUA_16 = range(0x100000, 0x10FFFE)
|
|
|
|
# we use this ranges of PUA_15 to store name aliases and named sequences
|
|
NAME_ALIASES_START = 0xF0000
|
|
NAMED_SEQUENCES_START = 0xF0200
|
|
|
|
old_versions = ["3.2.0"]
|
|
|
|
CATEGORY_NAMES = [ "Cn", "Lu", "Ll", "Lt", "Mn", "Mc", "Me", "Nd",
|
|
"Nl", "No", "Zs", "Zl", "Zp", "Cc", "Cf", "Cs", "Co", "Cn", "Lm",
|
|
"Lo", "Pc", "Pd", "Ps", "Pe", "Pi", "Pf", "Po", "Sm", "Sc", "Sk",
|
|
"So" ]
|
|
|
|
BIDIRECTIONAL_NAMES = [ "", "L", "LRE", "LRO", "R", "AL", "RLE", "RLO",
|
|
"PDF", "EN", "ES", "ET", "AN", "CS", "NSM", "BN", "B", "S", "WS",
|
|
"ON", "LRI", "RLI", "FSI", "PDI" ]
|
|
|
|
# https://www.unicode.org/reports/tr44/#BC_Values_Table
|
|
BIDI_LONG_NAMES = {
|
|
'Left_To_Right': 'L',
|
|
'Right_To_Left': 'R',
|
|
'Arabic_Letter': 'AL',
|
|
'European_Number': 'EN',
|
|
'European_Separator': 'ES',
|
|
'European_Terminator': 'ET',
|
|
'Arabic_Number': 'AN',
|
|
'Common_Separator': 'CS',
|
|
'Nonspacing_Mark': 'NSM',
|
|
'Boundary_Neutral': 'BN',
|
|
'Paragraph_Separator': 'B',
|
|
'Segment_Separator': 'S',
|
|
'White_Space': 'WS',
|
|
'Other_Neutral': 'ON',
|
|
'Left_To_Right_Embedding': 'LRE',
|
|
'Left_To_Right_Override': 'LRO',
|
|
'Right_To_Left_Embedding': 'RLE',
|
|
'Right_To_Left_Override': 'RLO',
|
|
'Pop_Directional_Format': 'PDF',
|
|
'Left_To_Right_Isolate': 'LRI',
|
|
'Right_To_Left_Isolate': 'RLI',
|
|
'First_Strong_Isolate': 'FSI',
|
|
'Pop_Directional_Isolate': 'PDI',
|
|
}
|
|
|
|
# "Other" needs to be the first entry, see the comment in makeunicodedata
|
|
GRAPHEME_CLUSTER_NAMES = [ 'Other', 'Prepend', 'CR', 'LF', 'Control',
|
|
'Extend', 'Regional_Indicator', 'SpacingMark', 'L', 'V', 'T', 'LV', 'LVT',
|
|
'ZWJ' ]
|
|
|
|
# "None" needs to be the first entry, see the comment in makeunicodedata
|
|
INDIC_CONJUNCT_BREAK_NAMES = [ 'None', 'Linker', 'Consonant', 'Extend' ]
|
|
|
|
# "N" needs to be the first entry, see the comment in makeunicodedata
|
|
EASTASIANWIDTH_NAMES = [ "N", "H", "W", "Na", "A", "F" ]
|
|
|
|
MANDATORY_LINE_BREAKS = [ "BK", "CR", "LF", "NL" ]
|
|
|
|
# note: should match definitions in Objects/unicodectype.c
|
|
ALPHA_MASK = 0x01
|
|
DECIMAL_MASK = 0x02
|
|
DIGIT_MASK = 0x04
|
|
LOWER_MASK = 0x08
|
|
LINEBREAK_MASK = 0x10
|
|
SPACE_MASK = 0x20
|
|
TITLE_MASK = 0x40
|
|
UPPER_MASK = 0x80
|
|
XID_START_MASK = 0x100
|
|
XID_CONTINUE_MASK = 0x200
|
|
PRINTABLE_MASK = 0x400
|
|
NUMERIC_MASK = 0x800
|
|
CASE_IGNORABLE_MASK = 0x1000
|
|
CASED_MASK = 0x2000
|
|
EXTENDED_CASE_MASK = 0x4000
|
|
|
|
# Maps the range names in UnicodeData.txt to prefixes for
|
|
# derived names specified by rule NR2.
|
|
# Hangul should always be at index 0, since it uses special format.
|
|
derived_name_range_names = [
|
|
("Hangul Syllable", "HANGUL SYLLABLE "),
|
|
("CJK Ideograph", "CJK UNIFIED IDEOGRAPH-"),
|
|
("Tangut Ideograph", "TANGUT IDEOGRAPH-"),
|
|
]
|
|
|
|
|
|
def maketables(trace=0):
|
|
|
|
print("--- Reading", UNICODE_DATA % "", "...")
|
|
|
|
unicode = UnicodeData(UNIDATA_VERSION)
|
|
|
|
print(len(list(filter(None, unicode.table))), "characters")
|
|
|
|
for version in old_versions:
|
|
print("--- Reading", UNICODE_DATA % ("-"+version), "...")
|
|
old_unicode = UnicodeData(version, ideograph_check=False)
|
|
print(len(list(filter(None, old_unicode.table))), "characters")
|
|
merge_old_version(version, unicode, old_unicode)
|
|
|
|
makeunicodename(unicode, trace)
|
|
makeunicodedata(unicode, trace)
|
|
makeunicodetype(unicode, trace)
|
|
|
|
|
|
# --------------------------------------------------------------------
|
|
# unicode character properties
|
|
|
|
def makeunicodedata(unicode, trace):
|
|
|
|
# the default value of east_asian_width is "N", for unassigned code points
|
|
# not mentioned in EastAsianWidth.txt
|
|
# in addition there are some reserved but unassigned code points in CJK
|
|
# ranges that are classified as "W". code points in private use areas
|
|
# have a width of "A". both of these have entries in
|
|
# EastAsianWidth.txt
|
|
# see https://unicode.org/reports/tr11/#Unassigned
|
|
assert EASTASIANWIDTH_NAMES[0] == "N"
|
|
assert GRAPHEME_CLUSTER_NAMES[0] == "Other"
|
|
assert INDIC_CONJUNCT_BREAK_NAMES[0] == "None"
|
|
dummy = (0, 0, 0, 0, 0, 0, 0, 0, 0)
|
|
table = [dummy]
|
|
cache = {0: dummy}
|
|
index = [0] * len(unicode.chars)
|
|
|
|
FILE = "Modules/unicodedata_db.h"
|
|
|
|
print("--- Preparing", FILE, "...")
|
|
|
|
# 1) database properties
|
|
|
|
for char in unicode.chars:
|
|
record = unicode.table[char]
|
|
eastasianwidth = EASTASIANWIDTH_NAMES.index(unicode.widths[char] or 'N')
|
|
graphemebreak = GRAPHEME_CLUSTER_NAMES.index(unicode.grapheme_breaks[char] or 'Other')
|
|
extpict = unicode.ext_picts[char]
|
|
bidirectional = BIDIRECTIONAL_NAMES.index(unicode.bidi_classes[char])
|
|
if record:
|
|
# extract database properties
|
|
category = CATEGORY_NAMES.index(record.general_category)
|
|
combining = int(record.canonical_combining_class)
|
|
mirrored = record.bidi_mirrored == "Y"
|
|
normalizationquickcheck = record.quick_check
|
|
incb = INDIC_CONJUNCT_BREAK_NAMES.index(record.incb)
|
|
item = (
|
|
category, combining, bidirectional, mirrored, eastasianwidth,
|
|
normalizationquickcheck, graphemebreak, incb, extpict,
|
|
)
|
|
else:
|
|
if eastasianwidth or graphemebreak or extpict or bidirectional:
|
|
item = (0, 0, bidirectional, 0, eastasianwidth,
|
|
0, graphemebreak, 0, extpict)
|
|
else:
|
|
continue
|
|
|
|
# add entry to index and item tables
|
|
i = cache.get(item)
|
|
if i is None:
|
|
cache[item] = i = len(table)
|
|
table.append(item)
|
|
index[char] = i
|
|
|
|
# 2) decomposition data
|
|
|
|
decomp_data_cache = {}
|
|
decomp_data = [0]
|
|
decomp_prefix = [""]
|
|
decomp_index = [0] * len(unicode.chars)
|
|
decomp_size = 0
|
|
|
|
comp_pairs = []
|
|
comp_first = [None] * len(unicode.chars)
|
|
comp_last = [None] * len(unicode.chars)
|
|
|
|
for char in unicode.chars:
|
|
record = unicode.table[char]
|
|
if record:
|
|
if record.decomposition_type:
|
|
decomp = record.decomposition_type.split()
|
|
if len(decomp) > 19:
|
|
raise Exception("character %x has a decomposition too large for nfd_nfkd" % char)
|
|
# prefix
|
|
if decomp[0][0] == "<":
|
|
prefix = decomp.pop(0)
|
|
else:
|
|
prefix = ""
|
|
try:
|
|
i = decomp_prefix.index(prefix)
|
|
except ValueError:
|
|
i = len(decomp_prefix)
|
|
decomp_prefix.append(prefix)
|
|
prefix = i
|
|
assert prefix < 256
|
|
# content
|
|
decomp = [prefix + (len(decomp)<<8)] + [int(s, 16) for s in decomp]
|
|
# Collect NFC pairs
|
|
if not prefix and len(decomp) == 3 and \
|
|
char not in unicode.exclusions and \
|
|
unicode.table[decomp[1]].canonical_combining_class == "0":
|
|
p, l, r = decomp
|
|
comp_first[l] = 1
|
|
comp_last[r] = 1
|
|
comp_pairs.append((l,r,char))
|
|
key = tuple(decomp)
|
|
i = decomp_data_cache.get(key, -1)
|
|
if i == -1:
|
|
i = len(decomp_data)
|
|
decomp_data.extend(decomp)
|
|
decomp_size = decomp_size + len(decomp) * 2
|
|
decomp_data_cache[key] = i
|
|
else:
|
|
assert decomp_data[i:i+len(decomp)] == decomp
|
|
else:
|
|
i = 0
|
|
decomp_index[char] = i
|
|
|
|
f = l = 0
|
|
comp_first_ranges = []
|
|
comp_last_ranges = []
|
|
prev_f = prev_l = None
|
|
for i in unicode.chars:
|
|
if comp_first[i] is not None:
|
|
comp_first[i] = f
|
|
f += 1
|
|
if prev_f is None:
|
|
prev_f = (i,i)
|
|
elif prev_f[1]+1 == i:
|
|
prev_f = prev_f[0],i
|
|
else:
|
|
comp_first_ranges.append(prev_f)
|
|
prev_f = (i,i)
|
|
if comp_last[i] is not None:
|
|
comp_last[i] = l
|
|
l += 1
|
|
if prev_l is None:
|
|
prev_l = (i,i)
|
|
elif prev_l[1]+1 == i:
|
|
prev_l = prev_l[0],i
|
|
else:
|
|
comp_last_ranges.append(prev_l)
|
|
prev_l = (i,i)
|
|
comp_first_ranges.append(prev_f)
|
|
comp_last_ranges.append(prev_l)
|
|
total_first = f
|
|
total_last = l
|
|
|
|
comp_data = [0]*(total_first*total_last)
|
|
for f,l,char in comp_pairs:
|
|
f = comp_first[f]
|
|
l = comp_last[l]
|
|
comp_data[f*total_last+l] = char
|
|
|
|
print(len(table), "unique properties")
|
|
print(len(decomp_prefix), "unique decomposition prefixes")
|
|
print(len(decomp_data), "unique decomposition entries:", end=' ')
|
|
print(decomp_size, "bytes")
|
|
print(total_first, "first characters in NFC")
|
|
print(total_last, "last characters in NFC")
|
|
print(len(comp_pairs), "NFC pairs")
|
|
|
|
print("--- Writing", FILE, "...")
|
|
|
|
with open(FILE, "w") as fp:
|
|
fprint = partial(print, file=fp)
|
|
|
|
fprint("/* this file was generated by %s %s */" % (SCRIPT, VERSION))
|
|
fprint()
|
|
fprint('#define UNIDATA_VERSION "%s"' % UNIDATA_VERSION)
|
|
fprint("/* a list of unique database records */")
|
|
fprint("const _PyUnicode_DatabaseRecord _PyUnicode_Database_Records[] = {")
|
|
for item in table:
|
|
fprint(" {%d, %d, %d, %d, %d, %d, %d, %d, %d}," % item)
|
|
fprint("};")
|
|
fprint()
|
|
|
|
fprint("/* Reindexing of NFC first characters. */")
|
|
fprint("#define TOTAL_FIRST",total_first)
|
|
fprint("#define TOTAL_LAST",total_last)
|
|
fprint("struct reindex{int start;short count,index;};")
|
|
fprint("static struct reindex nfc_first[] = {")
|
|
for start,end in comp_first_ranges:
|
|
fprint(" { %d, %d, %d}," % (start,end-start,comp_first[start]))
|
|
fprint(" {0,0,0}")
|
|
fprint("};\n")
|
|
fprint("static struct reindex nfc_last[] = {")
|
|
for start,end in comp_last_ranges:
|
|
fprint(" { %d, %d, %d}," % (start,end-start,comp_last[start]))
|
|
fprint(" {0,0,0}")
|
|
fprint("};\n")
|
|
|
|
# FIXME: <fl> the following tables could be made static, and
|
|
# the support code moved into unicodedatabase.c
|
|
|
|
fprint("/* string literals */")
|
|
fprint("const char *_PyUnicode_CategoryNames[] = {")
|
|
for name in CATEGORY_NAMES:
|
|
fprint(" \"%s\"," % name)
|
|
fprint(" NULL")
|
|
fprint("};")
|
|
|
|
fprint("const char *_PyUnicode_BidirectionalNames[] = {")
|
|
for name in BIDIRECTIONAL_NAMES:
|
|
fprint(" \"%s\"," % name)
|
|
fprint(" NULL")
|
|
fprint("};")
|
|
|
|
fprint("const char *_PyUnicode_EastAsianWidthNames[] = {")
|
|
for name in EASTASIANWIDTH_NAMES:
|
|
fprint(" \"%s\"," % name)
|
|
fprint(" NULL")
|
|
fprint("};")
|
|
|
|
for i, name in enumerate(GRAPHEME_CLUSTER_NAMES):
|
|
fprint("#define GCB_%s %d" % (name, i))
|
|
|
|
fprint("const char * const _PyUnicode_GraphemeBreakNames[] = {")
|
|
for name in GRAPHEME_CLUSTER_NAMES:
|
|
fprint(' "%s",' % name)
|
|
fprint(" NULL")
|
|
fprint("};")
|
|
|
|
for i, name in enumerate(INDIC_CONJUNCT_BREAK_NAMES):
|
|
fprint("#define InCB_%s %d" % (name, i))
|
|
|
|
fprint("const char * const _PyUnicode_IndicConjunctBreakNames[] = {")
|
|
for name in INDIC_CONJUNCT_BREAK_NAMES:
|
|
fprint(' "%s",' % name)
|
|
fprint(" NULL")
|
|
fprint("};")
|
|
|
|
fprint("static const char *decomp_prefix[] = {")
|
|
for name in decomp_prefix:
|
|
fprint(" \"%s\"," % name)
|
|
fprint(" NULL")
|
|
fprint("};")
|
|
|
|
# split record index table
|
|
index1, index2, shift = splitbins(index, trace)
|
|
|
|
fprint("/* index tables for the database records */")
|
|
fprint("#define SHIFT", shift)
|
|
Array("index1", index1).dump(fp, trace)
|
|
Array("index2", index2).dump(fp, trace)
|
|
|
|
# split decomposition index table
|
|
index1, index2, shift = splitbins(decomp_index, trace)
|
|
|
|
fprint("/* decomposition data */")
|
|
Array("decomp_data", decomp_data).dump(fp, trace)
|
|
|
|
fprint("/* index tables for the decomposition data */")
|
|
fprint("#define DECOMP_SHIFT", shift)
|
|
Array("decomp_index1", index1).dump(fp, trace)
|
|
Array("decomp_index2", index2).dump(fp, trace)
|
|
|
|
index, index2, shift = splitbins(comp_data, trace)
|
|
fprint("/* NFC pairs */")
|
|
fprint("#define COMP_SHIFT", shift)
|
|
Array("comp_index", index).dump(fp, trace)
|
|
Array("comp_data", index2).dump(fp, trace)
|
|
|
|
# Generate delta tables for old versions
|
|
for version, table, normalization in unicode.changed:
|
|
cversion = version.replace(".","_")
|
|
records = [table[0]]
|
|
cache = {table[0]:0}
|
|
index = [0] * len(table)
|
|
for i, record in enumerate(table):
|
|
try:
|
|
index[i] = cache[record]
|
|
except KeyError:
|
|
index[i] = cache[record] = len(records)
|
|
records.append(record)
|
|
index1, index2, shift = splitbins(index, trace)
|
|
fprint("static const change_record change_records_%s[] = {" % cversion)
|
|
for record in records:
|
|
fprint(" { %s }," % ", ".join(map(str,record)))
|
|
fprint("};")
|
|
Array("changes_%s_index" % cversion, index1).dump(fp, trace)
|
|
Array("changes_%s_data" % cversion, index2).dump(fp, trace)
|
|
fprint("static const change_record* get_change_%s(Py_UCS4 n)" % cversion)
|
|
fprint("{")
|
|
fprint(" int index;")
|
|
fprint(" if (n >= 0x110000) index = 0;")
|
|
fprint(" else {")
|
|
fprint(" index = changes_%s_index[n>>%d];" % (cversion, shift))
|
|
fprint(" index = changes_%s_data[(index<<%d)+(n & %d)];" % \
|
|
(cversion, shift, ((1<<shift)-1)))
|
|
fprint(" }")
|
|
fprint(" return change_records_%s+index;" % cversion)
|
|
fprint("}\n")
|
|
fprint("static Py_UCS4 normalization_%s(Py_UCS4 n)" % cversion)
|
|
fprint("{")
|
|
fprint(" switch(n) {")
|
|
for k, v in normalization:
|
|
fprint(" case %s: return 0x%s;" % (hex(k), v))
|
|
fprint(" default: return 0;")
|
|
fprint(" }\n}\n")
|
|
|
|
|
|
# --------------------------------------------------------------------
|
|
# unicode character type tables
|
|
|
|
def makeunicodetype(unicode, trace):
|
|
|
|
FILE = "Objects/unicodetype_db.h"
|
|
|
|
print("--- Preparing", FILE, "...")
|
|
|
|
# extract unicode types
|
|
dummy = (0, 0, 0, 0, 0, 0)
|
|
table = [dummy]
|
|
cache = {dummy: 0}
|
|
index = [0] * len(unicode.chars)
|
|
numeric = {}
|
|
spaces = []
|
|
linebreaks = []
|
|
extra_casing = []
|
|
|
|
for char in unicode.chars:
|
|
record = unicode.table[char]
|
|
if record:
|
|
# extract database properties
|
|
category = record.general_category
|
|
bidirectional = unicode.bidi_classes[char]
|
|
properties = record.binary_properties
|
|
flags = 0
|
|
if category in ["Lm", "Lt", "Lu", "Ll", "Lo"]:
|
|
flags |= ALPHA_MASK
|
|
if "Lowercase" in properties:
|
|
flags |= LOWER_MASK
|
|
if 'Line_Break' in properties or bidirectional == "B":
|
|
flags |= LINEBREAK_MASK
|
|
linebreaks.append(char)
|
|
if category == "Zs" or bidirectional in ("WS", "B", "S"):
|
|
flags |= SPACE_MASK
|
|
spaces.append(char)
|
|
if category == "Lt":
|
|
flags |= TITLE_MASK
|
|
if "Uppercase" in properties:
|
|
flags |= UPPER_MASK
|
|
if char == ord(" ") or category[0] not in ("C", "Z"):
|
|
flags |= PRINTABLE_MASK
|
|
if "XID_Start" in properties:
|
|
flags |= XID_START_MASK
|
|
if "XID_Continue" in properties:
|
|
flags |= XID_CONTINUE_MASK
|
|
if "Cased" in properties:
|
|
flags |= CASED_MASK
|
|
if "Case_Ignorable" in properties:
|
|
flags |= CASE_IGNORABLE_MASK
|
|
sc = unicode.special_casing.get(char)
|
|
cf = unicode.case_folding.get(char, [char])
|
|
if record.simple_uppercase_mapping:
|
|
upper = int(record.simple_uppercase_mapping, 16)
|
|
else:
|
|
upper = char
|
|
if record.simple_lowercase_mapping:
|
|
lower = int(record.simple_lowercase_mapping, 16)
|
|
else:
|
|
lower = char
|
|
if record.simple_titlecase_mapping:
|
|
title = int(record.simple_titlecase_mapping, 16)
|
|
else:
|
|
title = upper
|
|
if sc is None and cf != [lower]:
|
|
sc = ([lower], [title], [upper])
|
|
if sc is None:
|
|
if upper == lower == title:
|
|
upper = lower = title = 0
|
|
else:
|
|
upper = upper - char
|
|
lower = lower - char
|
|
title = title - char
|
|
assert (abs(upper) <= 2147483647 and
|
|
abs(lower) <= 2147483647 and
|
|
abs(title) <= 2147483647)
|
|
else:
|
|
# This happens either when some character maps to more than one
|
|
# character in uppercase, lowercase, or titlecase or the
|
|
# casefolded version of the character is different from the
|
|
# lowercase. The extra characters are stored in a different
|
|
# array.
|
|
flags |= EXTENDED_CASE_MASK
|
|
lower = len(extra_casing) | (len(sc[0]) << 24)
|
|
extra_casing.extend(sc[0])
|
|
if cf != sc[0]:
|
|
lower |= len(cf) << 20
|
|
extra_casing.extend(cf)
|
|
upper = len(extra_casing) | (len(sc[2]) << 24)
|
|
extra_casing.extend(sc[2])
|
|
# Title is probably equal to upper.
|
|
if sc[1] == sc[2]:
|
|
title = upper
|
|
else:
|
|
title = len(extra_casing) | (len(sc[1]) << 24)
|
|
extra_casing.extend(sc[1])
|
|
# decimal digit, integer digit
|
|
decimal = 0
|
|
if record.decomposition_mapping:
|
|
flags |= DECIMAL_MASK
|
|
decimal = int(record.decomposition_mapping)
|
|
digit = 0
|
|
if record.numeric_type:
|
|
flags |= DIGIT_MASK
|
|
digit = int(record.numeric_type)
|
|
if record.numeric_value:
|
|
flags |= NUMERIC_MASK
|
|
numeric.setdefault(record.numeric_value, []).append(char)
|
|
item = (
|
|
upper, lower, title, decimal, digit, flags
|
|
)
|
|
# add entry to index and item tables
|
|
i = cache.get(item)
|
|
if i is None:
|
|
cache[item] = i = len(table)
|
|
table.append(item)
|
|
index[char] = i
|
|
|
|
print(len(table), "unique character type entries")
|
|
print(sum(map(len, numeric.values())), "numeric code points")
|
|
print(len(spaces), "whitespace code points")
|
|
print(len(linebreaks), "linebreak code points")
|
|
print(len(extra_casing), "extended case array")
|
|
|
|
print("--- Writing", FILE, "...")
|
|
|
|
with open(FILE, "w") as fp:
|
|
fprint = partial(print, file=fp)
|
|
|
|
fprint("/* this file was generated by %s %s */" % (SCRIPT, VERSION))
|
|
fprint()
|
|
fprint("/* a list of unique character type descriptors */")
|
|
fprint("const _PyUnicode_TypeRecord _PyUnicode_TypeRecords[] = {")
|
|
for item in table:
|
|
fprint(" {%d, %d, %d, %d, %d, %d}," % item)
|
|
fprint("};")
|
|
fprint()
|
|
|
|
fprint("/* extended case mappings */")
|
|
fprint()
|
|
fprint("const Py_UCS4 _PyUnicode_ExtendedCase[] = {")
|
|
for c in extra_casing:
|
|
fprint(" %d," % c)
|
|
fprint("};")
|
|
fprint()
|
|
|
|
# split decomposition index table
|
|
index1, index2, shift = splitbins(index, trace)
|
|
|
|
fprint("/* type indexes */")
|
|
fprint("#define SHIFT", shift)
|
|
Array("index1", index1).dump(fp, trace)
|
|
Array("index2", index2).dump(fp, trace)
|
|
|
|
# Generate code for _PyUnicode_ToNumeric()
|
|
numeric_items = sorted(numeric.items())
|
|
fprint('/* Returns the numeric value as double for Unicode characters')
|
|
fprint(' * having this property, -1.0 otherwise.')
|
|
fprint(' */')
|
|
fprint('double _PyUnicode_ToNumeric(Py_UCS4 ch)')
|
|
fprint('{')
|
|
fprint(' switch (ch) {')
|
|
for value, codepoints in numeric_items:
|
|
# Turn text into float literals
|
|
parts = value.split('/')
|
|
parts = [repr(float(part)) for part in parts]
|
|
value = '/'.join(parts)
|
|
|
|
codepoints.sort()
|
|
for codepoint in codepoints:
|
|
fprint(' case 0x%04X:' % (codepoint,))
|
|
fprint(' return (double) %s;' % (value,))
|
|
fprint(' }')
|
|
fprint(' return -1.0;')
|
|
fprint('}')
|
|
fprint()
|
|
|
|
# Generate code for _PyUnicode_IsWhitespace()
|
|
fprint("/* Returns 1 for Unicode characters having the bidirectional")
|
|
fprint(" * type 'WS', 'B' or 'S' or the category 'Zs', 0 otherwise.")
|
|
fprint(" */")
|
|
fprint('int _PyUnicode_IsWhitespace(const Py_UCS4 ch)')
|
|
fprint('{')
|
|
fprint(' switch (ch) {')
|
|
|
|
for codepoint in sorted(spaces):
|
|
fprint(' case 0x%04X:' % (codepoint,))
|
|
fprint(' return 1;')
|
|
|
|
fprint(' }')
|
|
fprint(' return 0;')
|
|
fprint('}')
|
|
fprint()
|
|
|
|
# Generate code for _PyUnicode_IsLinebreak()
|
|
fprint("/* Returns 1 for Unicode characters having the line break")
|
|
fprint(" * property 'BK', 'CR', 'LF' or 'NL' or having bidirectional")
|
|
fprint(" * type 'B', 0 otherwise.")
|
|
fprint(" */")
|
|
fprint('int _PyUnicode_IsLinebreak(const Py_UCS4 ch)')
|
|
fprint('{')
|
|
fprint(' switch (ch) {')
|
|
for codepoint in sorted(linebreaks):
|
|
fprint(' case 0x%04X:' % (codepoint,))
|
|
fprint(' return 1;')
|
|
|
|
fprint(' }')
|
|
fprint(' return 0;')
|
|
fprint('}')
|
|
fprint()
|
|
|
|
|
|
# --------------------------------------------------------------------
|
|
# unicode name database
|
|
|
|
def makeunicodename(unicode, trace):
|
|
from dawg import build_compression_dawg
|
|
|
|
FILE = "Modules/unicodename_db.h"
|
|
|
|
print("--- Preparing", FILE, "...")
|
|
|
|
# unicode name hash table
|
|
|
|
# extract names
|
|
data = []
|
|
for char in unicode.chars:
|
|
record = unicode.table[char]
|
|
if record:
|
|
name = record.name.strip()
|
|
if name and name[0] != "<":
|
|
data.append((name, char))
|
|
|
|
print("--- Writing", FILE, "...")
|
|
|
|
with open(FILE, "w") as fp:
|
|
fprint = partial(print, file=fp)
|
|
|
|
fprint("/* this file was generated by %s %s */" % (SCRIPT, VERSION))
|
|
fprint()
|
|
fprint("#define NAME_MAXLEN", 256)
|
|
assert max(len(x) for x in data) < 256
|
|
fprint()
|
|
|
|
fprint("/* name->code dictionary */")
|
|
packed_dawg, pos_to_codepoint = build_compression_dawg(data)
|
|
notfound = len(pos_to_codepoint)
|
|
inverse_list = [notfound] * len(unicode.chars)
|
|
for pos, codepoint in enumerate(pos_to_codepoint):
|
|
inverse_list[codepoint] = pos
|
|
Array("packed_name_dawg", list(packed_dawg)).dump(fp, trace)
|
|
Array("dawg_pos_to_codepoint", pos_to_codepoint).dump(fp, trace)
|
|
index1, index2, shift = splitbins(inverse_list, trace)
|
|
fprint("#define DAWG_CODEPOINT_TO_POS_SHIFT", shift)
|
|
fprint("#define DAWG_CODEPOINT_TO_POS_NOTFOUND", notfound)
|
|
Array("dawg_codepoint_to_pos_index1", index1).dump(fp, trace)
|
|
Array("dawg_codepoint_to_pos_index2", index2).dump(fp, trace)
|
|
|
|
fprint()
|
|
fprint('static const unsigned int aliases_start = %#x;' %
|
|
NAME_ALIASES_START)
|
|
fprint('static const unsigned int aliases_end = %#x;' %
|
|
(NAME_ALIASES_START + len(unicode.aliases)))
|
|
|
|
fprint('static const unsigned int name_aliases[] = {')
|
|
for name, codepoint in unicode.aliases:
|
|
fprint(' 0x%04X,' % codepoint)
|
|
fprint('};')
|
|
|
|
# In Unicode 6.0.0, the sequences contain at most 4 BMP chars,
|
|
# so we are using Py_UCS2 seq[4]. This needs to be updated if longer
|
|
# sequences or sequences with non-BMP chars are added.
|
|
# unicodedata_lookup should be adapted too.
|
|
fprint(dedent("""
|
|
typedef struct NamedSequence {
|
|
int seqlen;
|
|
Py_UCS2 seq[4];
|
|
} named_sequence;
|
|
"""))
|
|
|
|
fprint('static const unsigned int named_sequences_start = %#x;' %
|
|
NAMED_SEQUENCES_START)
|
|
fprint('static const unsigned int named_sequences_end = %#x;' %
|
|
(NAMED_SEQUENCES_START + len(unicode.named_sequences)))
|
|
|
|
fprint('static const named_sequence named_sequences[] = {')
|
|
for name, sequence in unicode.named_sequences:
|
|
seq_str = ', '.join('0x%04X' % cp for cp in sequence)
|
|
fprint(' {%d, {%s}},' % (len(sequence), seq_str))
|
|
fprint('};')
|
|
|
|
fprint(dedent("""
|
|
typedef struct {
|
|
Py_UCS4 first;
|
|
Py_UCS4 last;
|
|
int prefixid;
|
|
} derived_name_range;
|
|
"""))
|
|
|
|
fprint('static const derived_name_range derived_name_ranges[] = {')
|
|
for name_range in unicode.derived_name_ranges:
|
|
fprint(' {0x%s, 0x%s, %d},' % tuple(name_range))
|
|
fprint('};')
|
|
|
|
fprint('static const char * const derived_name_prefixes[] = {')
|
|
for prefix in unicode.derived_name_prefixes:
|
|
fprint(' "%s",' % prefix)
|
|
fprint('};')
|
|
|
|
def merge_old_version(version, new, old):
|
|
# Changes to exclusion file not implemented yet
|
|
if old.exclusions != new.exclusions:
|
|
raise NotImplementedError("exclusions differ")
|
|
|
|
# In these change records, 0xFF means "no change"
|
|
bidir_changes = [0xFF]*0x110000
|
|
category_changes = [0xFF]*0x110000
|
|
decimal_changes = [0xFF]*0x110000
|
|
mirrored_changes = [0xFF]*0x110000
|
|
east_asian_width_changes = [0xFF]*0x110000
|
|
# In numeric data, 0 means "no change",
|
|
# -1 means "did not have a numeric value
|
|
numeric_changes = [0] * 0x110000
|
|
# normalization_changes is a list of key-value pairs
|
|
normalization_changes = []
|
|
for i in range(0x110000):
|
|
if new.table[i] is None:
|
|
# Characters unassigned in the new version ought to
|
|
# be unassigned in the old one
|
|
assert old.table[i] is None
|
|
continue
|
|
# check characters unassigned in the old version
|
|
if old.table[i] is None:
|
|
# category 0 is "unassigned"
|
|
category_changes[i] = 0
|
|
continue
|
|
if old.bidi_classes[i] != new.bidi_classes[i]:
|
|
bidir_changes[i] = BIDIRECTIONAL_NAMES.index(old.bidi_classes[i])
|
|
# check characters that differ
|
|
if old.table[i] != new.table[i]:
|
|
for k, field in enumerate(dataclasses.fields(UcdRecord)):
|
|
value = getattr(old.table[i], field.name)
|
|
new_value = getattr(new.table[i], field.name)
|
|
if value != new_value:
|
|
if k == 1 and i in PUA_15:
|
|
# the name is not set in the old.table, but in the
|
|
# new.table we are using it for aliases and named seq
|
|
assert value == ''
|
|
elif k == 2:
|
|
category_changes[i] = CATEGORY_NAMES.index(value)
|
|
elif k == 4:
|
|
# bidi_class changes handled via bidi_classes
|
|
pass
|
|
elif k == 5:
|
|
# We assume that all normalization changes are in 1:1 mappings
|
|
assert " " not in value
|
|
normalization_changes.append((i, value))
|
|
elif k == 6:
|
|
# we only support changes where the old value is a single digit
|
|
assert value in "0123456789"
|
|
decimal_changes[i] = int(value)
|
|
elif k == 8:
|
|
# Since 0 encodes "no change", the old value is better not 0
|
|
if not value:
|
|
numeric_changes[i] = -1
|
|
else:
|
|
numeric_changes[i] = float(value)
|
|
assert numeric_changes[i] not in (0, -1)
|
|
elif k == 9:
|
|
if value == 'Y':
|
|
mirrored_changes[i] = '1'
|
|
else:
|
|
mirrored_changes[i] = '0'
|
|
elif k == 11:
|
|
# change to ISO comment, ignore
|
|
pass
|
|
elif k == 12:
|
|
# change to simple uppercase mapping; ignore
|
|
pass
|
|
elif k == 13:
|
|
# change to simple lowercase mapping; ignore
|
|
pass
|
|
elif k == 14:
|
|
# change to simple titlecase mapping; ignore
|
|
pass
|
|
elif k == 15:
|
|
# change to east asian width
|
|
east_asian_width_changes[i] = EASTASIANWIDTH_NAMES.index(value)
|
|
elif k == 16:
|
|
# derived property changes; not yet
|
|
pass
|
|
elif k == 17:
|
|
# normalization quickchecks are not performed
|
|
# for older versions
|
|
pass
|
|
elif k == 18:
|
|
# The Indic_Conjunct_Break property did not exist for
|
|
# older versions
|
|
pass
|
|
else:
|
|
class Difference(Exception):pass
|
|
raise Difference(hex(i), k, old.table[i], new.table[i])
|
|
new.changed.append((version, list(zip(bidir_changes, category_changes,
|
|
decimal_changes, mirrored_changes,
|
|
east_asian_width_changes,
|
|
numeric_changes)),
|
|
normalization_changes))
|
|
|
|
|
|
DATA_DIR = os.path.join('Tools', 'unicode', 'data')
|
|
|
|
def open_data(template, version):
|
|
local = os.path.join(DATA_DIR, template % ('-'+version,))
|
|
if not os.path.exists(local):
|
|
import urllib.request
|
|
if version == '3.2.0':
|
|
# irregular url structure
|
|
url = ('https://www.unicode.org/Public/3.2-Update/'+template) % ('-'+version,)
|
|
else:
|
|
url = ('https://www.unicode.org/Public/%s/ucd/'+template) % (version, '')
|
|
os.makedirs(os.path.dirname(local), exist_ok=True)
|
|
urllib.request.urlretrieve(url, filename=local)
|
|
if local.endswith('.txt'):
|
|
return open(local, encoding='utf-8')
|
|
else:
|
|
# Unihan.zip
|
|
return open(local, 'rb')
|
|
|
|
|
|
def expand_range(char_range: str) -> Iterator[int]:
|
|
'''
|
|
Parses ranges of code points, as described in UAX #44:
|
|
https://www.unicode.org/reports/tr44/#Code_Point_Ranges
|
|
'''
|
|
if '..' in char_range:
|
|
first, last = [int(c, 16) for c in char_range.split('..')]
|
|
else:
|
|
first = last = int(char_range, 16)
|
|
for char in range(first, last+1):
|
|
yield char
|
|
|
|
|
|
class UcdFile:
|
|
'''
|
|
A file in the standard format of the UCD.
|
|
|
|
See: https://www.unicode.org/reports/tr44/#Format_Conventions
|
|
|
|
Note that, as described there, the Unihan data files have their
|
|
own separate format.
|
|
'''
|
|
|
|
def __init__(self, template: str, version: str) -> None:
|
|
self.template = template
|
|
self.version = version
|
|
|
|
def records(self) -> Iterator[List[str]]:
|
|
with open_data(self.template, self.version) as file:
|
|
for line in file:
|
|
line = line.split('#', 1)[0].strip()
|
|
if not line:
|
|
continue
|
|
yield [field.strip() for field in line.split(';')]
|
|
|
|
def __iter__(self) -> Iterator[List[str]]:
|
|
return self.records()
|
|
|
|
def expanded(self) -> Iterator[Tuple[int, List[str]]]:
|
|
for record in self.records():
|
|
char_range, rest = record[0], record[1:]
|
|
for char in expand_range(char_range):
|
|
yield char, rest
|
|
|
|
|
|
@dataclasses.dataclass
|
|
class UcdRecord:
|
|
# 15 fields from UnicodeData.txt . See:
|
|
# https://www.unicode.org/reports/tr44/#UnicodeData.txt
|
|
codepoint: str
|
|
name: str
|
|
general_category: str
|
|
canonical_combining_class: str
|
|
bidi_class: str
|
|
decomposition_type: str
|
|
decomposition_mapping: str
|
|
numeric_type: str
|
|
numeric_value: str
|
|
bidi_mirrored: str
|
|
unicode_1_name: str # obsolete
|
|
iso_comment: str # obsolete
|
|
simple_uppercase_mapping: str
|
|
simple_lowercase_mapping: str
|
|
simple_titlecase_mapping: str
|
|
|
|
# https://www.unicode.org/reports/tr44/#EastAsianWidth.txt
|
|
east_asian_width: Optional[str]
|
|
|
|
# Binary properties, as a set of those that are true.
|
|
# Taken from multiple files:
|
|
# https://www.unicode.org/reports/tr44/#DerivedCoreProperties.txt
|
|
# https://www.unicode.org/reports/tr44/#LineBreak.txt
|
|
binary_properties: Set[str]
|
|
|
|
# The Quick_Check properties related to normalization:
|
|
# https://www.unicode.org/reports/tr44/#Decompositions_and_Normalization
|
|
# We store them as a bitmask.
|
|
quick_check: int
|
|
|
|
# The Indic_Conjunct_Break property from DerivedCoreProperties.txt. See:
|
|
# https://www.unicode.org/reports/tr44/#DerivedCoreProperties.txt
|
|
incb: str
|
|
|
|
|
|
def from_row(row: List[str]) -> UcdRecord:
|
|
return UcdRecord(*row, None, set(), 0, "None")
|
|
|
|
|
|
# --------------------------------------------------------------------
|
|
# the following support code is taken from the unidb utilities
|
|
# Copyright (c) 1999-2000 by Secret Labs AB
|
|
|
|
# load a unicode-data file from disk
|
|
|
|
class UnicodeData:
|
|
# table: List[Optional[UcdRecord]] # index is codepoint; None means unassigned
|
|
|
|
def __init__(self, version, ideograph_check=True):
|
|
self.changed = []
|
|
table = [None] * 0x110000
|
|
for s in UcdFile(UNICODE_DATA, version):
|
|
char = int(s[0], 16)
|
|
table[char] = from_row(s)
|
|
|
|
self.derived_name_ranges = []
|
|
self.derived_name_prefixes = {
|
|
prefix: i
|
|
for i, (_, prefix) in enumerate(derived_name_range_names)
|
|
}
|
|
|
|
# expand first-last ranges
|
|
field = None
|
|
for i in range(0, 0x110000):
|
|
# The file UnicodeData.txt has its own distinct way of
|
|
# expressing ranges. See:
|
|
# https://www.unicode.org/reports/tr44/#Code_Point_Ranges
|
|
s = table[i]
|
|
if s:
|
|
if s.name[-6:] == "First>":
|
|
s.name = ""
|
|
field = dataclasses.astuple(s)[:15]
|
|
elif s.name[-5:] == "Last>":
|
|
for j, (rangename, _) in enumerate(derived_name_range_names):
|
|
if s.name.startswith("<" + rangename):
|
|
self.derived_name_ranges.append(
|
|
(field[0], s.codepoint, j))
|
|
break
|
|
s.name = ""
|
|
field = None
|
|
else:
|
|
codepoint = s.codepoint
|
|
if s.name.endswith(codepoint):
|
|
prefix = s.name[:-len(codepoint)]
|
|
j = self.derived_name_prefixes.get(prefix)
|
|
if j is None:
|
|
j = len(self.derived_name_prefixes)
|
|
self.derived_name_prefixes[prefix] = j
|
|
if (self.derived_name_ranges
|
|
and self.derived_name_ranges[-1][2] == j
|
|
and int(self.derived_name_ranges[-1][1], 16) == i - 1):
|
|
self.derived_name_ranges[-1][1] = codepoint
|
|
else:
|
|
self.derived_name_ranges.append(
|
|
[codepoint, codepoint, j])
|
|
s.name = ""
|
|
elif field:
|
|
table[i] = from_row(('%04X' % i,) + field[1:])
|
|
|
|
# public attributes
|
|
self.filename = UNICODE_DATA % ''
|
|
self.table = table
|
|
self.chars = list(range(0x110000)) # unicode 3.2
|
|
|
|
# check for name aliases and named sequences, see #12753
|
|
# aliases and named sequences are not in 3.2.0
|
|
if version != '3.2.0':
|
|
self.aliases = []
|
|
# store aliases in the Private Use Area 15, in range U+F0000..U+F00FF,
|
|
# in order to take advantage of the compression and lookup
|
|
# algorithms used for the other characters
|
|
pua_index = NAME_ALIASES_START
|
|
for char, name, abbrev in UcdFile(NAME_ALIASES, version):
|
|
char = int(char, 16)
|
|
self.aliases.append((name, char))
|
|
# also store the name in the PUA 1
|
|
self.table[pua_index].name = name
|
|
pua_index += 1
|
|
assert pua_index - NAME_ALIASES_START == len(self.aliases)
|
|
|
|
self.named_sequences = []
|
|
# store named sequences in the PUA 1, in range U+F0100..,
|
|
# in order to take advantage of the compression and lookup
|
|
# algorithms used for the other characters.
|
|
|
|
assert pua_index < NAMED_SEQUENCES_START
|
|
pua_index = NAMED_SEQUENCES_START
|
|
for name, chars in UcdFile(NAMED_SEQUENCES, version):
|
|
chars = tuple(int(char, 16) for char in chars.split())
|
|
# check that the structure defined in makeunicodename is OK
|
|
assert 2 <= len(chars) <= 4, "change the Py_UCS2 array size"
|
|
assert all(c <= 0xFFFF for c in chars), ("use Py_UCS4 in "
|
|
"the NamedSequence struct and in unicodedata_lookup")
|
|
self.named_sequences.append((name, chars))
|
|
# also store these in the PUA 1
|
|
self.table[pua_index].name = name
|
|
pua_index += 1
|
|
assert pua_index - NAMED_SEQUENCES_START == len(self.named_sequences)
|
|
|
|
self.exclusions = {}
|
|
for char, in UcdFile(COMPOSITION_EXCLUSIONS, version):
|
|
char = int(char, 16)
|
|
self.exclusions[char] = 1
|
|
|
|
widths = [None] * 0x110000
|
|
for char, (width,) in UcdFile(EASTASIAN_WIDTH, version).expanded():
|
|
widths[char] = width
|
|
|
|
for i in range(0, 0x110000):
|
|
if table[i] is not None:
|
|
table[i].east_asian_width = widths[i]
|
|
self.widths = widths
|
|
|
|
# Read DerivedBidiClass.txt for bidi classes
|
|
# see https://www.unicode.org/reports/tr44/#Missing_Conventions
|
|
bidi_classes = [None] * 0x110000
|
|
for i in range(0, 0x110000):
|
|
if table[i] is not None:
|
|
bidi_classes[i] = table[i].bidi_class
|
|
if version != '3.2.0':
|
|
missing_re = re.compile(
|
|
r'# @missing: ([\dA-F]+\.\.[\dA-F]+); (\w+)'
|
|
)
|
|
with open_data(DERIVED_BIDI_CLASS, version) as f:
|
|
for l in f:
|
|
m = missing_re.match(l)
|
|
if not m:
|
|
continue
|
|
name = BIDI_LONG_NAMES[m[2]]
|
|
for i in expand_range(m[1]):
|
|
bidi_classes[i] = name
|
|
for char, (bidi,) in UcdFile(DERIVED_BIDI_CLASS, version).expanded():
|
|
bidi_classes[char] = bidi
|
|
self.bidi_classes = bidi_classes
|
|
|
|
for char, (propname, *propinfo) in UcdFile(DERIVED_CORE_PROPERTIES, version).expanded():
|
|
if not propinfo:
|
|
# binary property
|
|
if table[char]:
|
|
# Some properties (e.g. Default_Ignorable_Code_Point)
|
|
# apply to unassigned code points; ignore them
|
|
table[char].binary_properties.add(propname)
|
|
elif propname == 'InCB': # Indic_Conjunct_Break
|
|
table[char].incb, = propinfo
|
|
|
|
for char_range, value in UcdFile(LINE_BREAK, version):
|
|
if value not in MANDATORY_LINE_BREAKS:
|
|
continue
|
|
for char in expand_range(char_range):
|
|
table[char].binary_properties.add('Line_Break')
|
|
|
|
# We only want the quickcheck properties
|
|
# Format: NF?_QC; Y(es)/N(o)/M(aybe)
|
|
# Yes is the default, hence only N and M occur
|
|
# In 3.2.0, the format was different (NF?_NO)
|
|
# The parsing will incorrectly determine these as
|
|
# "yes", however, unicodedata.c will not perform quickchecks
|
|
# for older versions, and no delta records will be created.
|
|
quickchecks = [0] * 0x110000
|
|
qc_order = 'NFD_QC NFKD_QC NFC_QC NFKC_QC'.split()
|
|
for s in UcdFile(DERIVEDNORMALIZATION_PROPS, version):
|
|
if len(s) < 2 or s[1] not in qc_order:
|
|
continue
|
|
quickcheck = 'MN'.index(s[2]) + 1 # Maybe or No
|
|
quickcheck_shift = qc_order.index(s[1])*2
|
|
quickcheck <<= quickcheck_shift
|
|
for char in expand_range(s[0]):
|
|
assert not (quickchecks[char]>>quickcheck_shift)&3
|
|
quickchecks[char] |= quickcheck
|
|
for i in range(0, 0x110000):
|
|
if table[i] is not None:
|
|
table[i].quick_check = quickchecks[i]
|
|
|
|
with open_data(UNIHAN, version) as file:
|
|
zip = zipfile.ZipFile(file)
|
|
if version == '3.2.0':
|
|
data = zip.open('Unihan-3.2.0.txt').read()
|
|
else:
|
|
data = zip.open('Unihan_NumericValues.txt').read()
|
|
for line in data.decode("utf-8").splitlines():
|
|
if not line.startswith('U+'):
|
|
continue
|
|
code, tag, value = line.split(None, 3)[:3]
|
|
if tag not in ('kAccountingNumeric', 'kPrimaryNumeric',
|
|
'kOtherNumeric'):
|
|
continue
|
|
value = value.strip().replace(',', '')
|
|
i = int(code[2:], 16)
|
|
# Patch the numeric field
|
|
if table[i] is not None:
|
|
table[i].numeric_value = value
|
|
|
|
sc = self.special_casing = {}
|
|
for data in UcdFile(SPECIAL_CASING, version):
|
|
if data[4]:
|
|
# We ignore all conditionals (since they depend on
|
|
# languages) except for one, which is hardcoded. See
|
|
# handle_capital_sigma in unicodeobject.c.
|
|
continue
|
|
c = int(data[0], 16)
|
|
lower = [int(char, 16) for char in data[1].split()]
|
|
title = [int(char, 16) for char in data[2].split()]
|
|
upper = [int(char, 16) for char in data[3].split()]
|
|
sc[c] = (lower, title, upper)
|
|
|
|
cf = self.case_folding = {}
|
|
if version != '3.2.0':
|
|
for data in UcdFile(CASE_FOLDING, version):
|
|
if data[1] in "CF":
|
|
c = int(data[0], 16)
|
|
cf[c] = [int(char, 16) for char in data[2].split()]
|
|
|
|
if version != "3.2.0":
|
|
grapheme_breaks = [None] * 0x110000
|
|
for char, (prop,) in UcdFile(GRAPHEME_CLUSTER_BREAK, version).expanded():
|
|
grapheme_breaks[char] = prop
|
|
self.grapheme_breaks = grapheme_breaks
|
|
|
|
ext_picts = [False] * 0x110000
|
|
for char, (prop,) in UcdFile(EMOJI_DATA, version).expanded():
|
|
if prop == 'Extended_Pictographic':
|
|
ext_picts[char] = True
|
|
self.ext_picts = ext_picts
|
|
|
|
|
|
def uselatin1(self):
|
|
# restrict character range to ISO Latin 1
|
|
self.chars = list(range(256))
|
|
|
|
|
|
|
|
# stuff to deal with arrays of unsigned integers
|
|
|
|
class Array:
|
|
|
|
def __init__(self, name, data):
|
|
self.name = name
|
|
self.data = data
|
|
|
|
def dump(self, file, trace=0):
|
|
# write data to file, as a C array
|
|
size = getsize(self.data)
|
|
if trace:
|
|
print(self.name+":", size*len(self.data), "bytes", file=sys.stderr)
|
|
file.write("static const ")
|
|
if size == 1:
|
|
file.write("unsigned char")
|
|
elif size == 2:
|
|
file.write("unsigned short")
|
|
else:
|
|
file.write("unsigned int")
|
|
file.write(" " + self.name + "[] = {\n")
|
|
if self.data:
|
|
s = " "
|
|
for item in self.data:
|
|
i = str(item) + ", "
|
|
if len(s) + len(i) > 78:
|
|
file.write(s.rstrip() + "\n")
|
|
s = " " + i
|
|
else:
|
|
s = s + i
|
|
if s.strip():
|
|
file.write(s.rstrip() + "\n")
|
|
file.write("};\n\n")
|
|
|
|
|
|
def getsize(data):
|
|
# return smallest possible integer size for the given array
|
|
maxdata = max(data)
|
|
if maxdata < 256:
|
|
return 1
|
|
elif maxdata < 65536:
|
|
return 2
|
|
else:
|
|
return 4
|
|
|
|
|
|
def splitbins(t, trace=0):
|
|
"""t, trace=0 -> (t1, t2, shift). Split a table to save space.
|
|
|
|
t is a sequence of ints. This function can be useful to save space if
|
|
many of the ints are the same. t1 and t2 are lists of ints, and shift
|
|
is an int, chosen to minimize the combined size of t1 and t2 (in C
|
|
code), and where for each i in range(len(t)),
|
|
t[i] == t2[(t1[i >> shift] << shift) + (i & mask)]
|
|
where mask is a bitmask isolating the last "shift" bits.
|
|
|
|
If optional arg trace is non-zero (default zero), progress info
|
|
is printed to sys.stderr. The higher the value, the more info
|
|
you'll get.
|
|
"""
|
|
|
|
if trace:
|
|
def dump(t1, t2, shift, bytes):
|
|
print("%d+%d bins at shift %d; %d bytes" % (
|
|
len(t1), len(t2), shift, bytes), file=sys.stderr)
|
|
print("Size of original table:", len(t)*getsize(t), "bytes",
|
|
file=sys.stderr)
|
|
n = len(t)-1 # last valid index
|
|
maxshift = 0 # the most we can shift n and still have something left
|
|
if n > 0:
|
|
while n >> 1:
|
|
n >>= 1
|
|
maxshift += 1
|
|
del n
|
|
bytes = sys.maxsize # smallest total size so far
|
|
t = tuple(t) # so slices can be dict keys
|
|
for shift in range(maxshift + 1):
|
|
t1 = []
|
|
t2 = []
|
|
size = 2**shift
|
|
bincache = {}
|
|
for i in range(0, len(t), size):
|
|
bin = t[i:i+size]
|
|
index = bincache.get(bin)
|
|
if index is None:
|
|
index = len(t2)
|
|
bincache[bin] = index
|
|
t2.extend(bin)
|
|
t1.append(index >> shift)
|
|
# determine memory size
|
|
b = len(t1)*getsize(t1) + len(t2)*getsize(t2)
|
|
if trace > 1:
|
|
dump(t1, t2, shift, b)
|
|
if b < bytes:
|
|
best = t1, t2, shift
|
|
bytes = b
|
|
t1, t2, shift = best
|
|
if trace:
|
|
print("Best:", end=' ', file=sys.stderr)
|
|
dump(t1, t2, shift, bytes)
|
|
if __debug__:
|
|
# exhaustively verify that the decomposition is correct
|
|
mask = ~((~0) << shift) # i.e., low-bit mask of shift bits
|
|
for i in range(len(t)):
|
|
assert t[i] == t2[(t1[i >> shift] << shift) + (i & mask)]
|
|
return best
|
|
|
|
|
|
if __name__ == "__main__":
|
|
maketables(1)
|