| """cffLib: read/write Adobe CFF fonts |
| |
| OpenType fonts with PostScript outlines contain a completely independent |
| font file, Adobe's *Compact Font Format*. So dealing with OpenType fonts |
| requires also dealing with CFF. This module allows you to read and write |
| fonts written in the CFF format. |
| |
| In 2016, OpenType 1.8 introduced the `CFF2 <https://docs.microsoft.com/en-us/typography/opentype/spec/cff2>`_ |
| format which, along with other changes, extended the CFF format to deal with |
| the demands of variable fonts. This module parses both original CFF and CFF2. |
| |
| """ |
| |
| from fontTools.misc import sstruct |
| from fontTools.misc import psCharStrings |
| from fontTools.misc.arrayTools import unionRect, intRect |
| from fontTools.misc.textTools import bytechr, byteord, bytesjoin, tobytes, tostr, safeEval |
| from fontTools.ttLib import TTFont |
| from fontTools.ttLib.tables.otBase import OTTableWriter |
| from fontTools.ttLib.tables.otBase import OTTableReader |
| from fontTools.ttLib.tables import otTables as ot |
| from io import BytesIO |
| import struct |
| import logging |
| import re |
| |
| # mute cffLib debug messages when running ttx in verbose mode |
| DEBUG = logging.DEBUG - 1 |
| log = logging.getLogger(__name__) |
| |
| cffHeaderFormat = """ |
| major: B |
| minor: B |
| hdrSize: B |
| """ |
| |
| maxStackLimit = 513 |
| # maxstack operator has been deprecated. max stack is now always 513. |
| |
| |
| class StopHintCountEvent(Exception): |
| pass |
| |
| |
| class _DesubroutinizingT2Decompiler(psCharStrings.SimpleT2Decompiler): |
| stop_hintcount_ops = ("op_hintmask", "op_cntrmask", "op_rmoveto", "op_hmoveto", |
| "op_vmoveto") |
| |
| def __init__(self, localSubrs, globalSubrs, private=None): |
| psCharStrings.SimpleT2Decompiler.__init__(self, localSubrs, globalSubrs, |
| private) |
| |
| def execute(self, charString): |
| self.need_hintcount = True # until proven otherwise |
| for op_name in self.stop_hintcount_ops: |
| setattr(self, op_name, self.stop_hint_count) |
| |
| if hasattr(charString, '_desubroutinized'): |
| # If a charstring has already been desubroutinized, we will still |
| # need to execute it if we need to count hints in order to |
| # compute the byte length for mask arguments, and haven't finished |
| # counting hints pairs. |
| if self.need_hintcount and self.callingStack: |
| try: |
| psCharStrings.SimpleT2Decompiler.execute(self, charString) |
| except StopHintCountEvent: |
| del self.callingStack[-1] |
| return |
| |
| charString._patches = [] |
| psCharStrings.SimpleT2Decompiler.execute(self, charString) |
| desubroutinized = charString.program[:] |
| for idx, expansion in reversed(charString._patches): |
| assert idx >= 2 |
| assert desubroutinized[idx - 1] in ['callsubr', 'callgsubr'], desubroutinized[idx - 1] |
| assert type(desubroutinized[idx - 2]) == int |
| if expansion[-1] == 'return': |
| expansion = expansion[:-1] |
| desubroutinized[idx-2:idx] = expansion |
| if not self.private.in_cff2: |
| if 'endchar' in desubroutinized: |
| # Cut off after first endchar |
| desubroutinized = desubroutinized[:desubroutinized.index('endchar') + 1] |
| else: |
| if not len(desubroutinized) or desubroutinized[-1] != 'return': |
| desubroutinized.append('return') |
| |
| charString._desubroutinized = desubroutinized |
| del charString._patches |
| |
| def op_callsubr(self, index): |
| subr = self.localSubrs[self.operandStack[-1]+self.localBias] |
| psCharStrings.SimpleT2Decompiler.op_callsubr(self, index) |
| self.processSubr(index, subr) |
| |
| def op_callgsubr(self, index): |
| subr = self.globalSubrs[self.operandStack[-1]+self.globalBias] |
| psCharStrings.SimpleT2Decompiler.op_callgsubr(self, index) |
| self.processSubr(index, subr) |
| |
| def stop_hint_count(self, *args): |
| self.need_hintcount = False |
| for op_name in self.stop_hintcount_ops: |
| setattr(self, op_name, None) |
| cs = self.callingStack[-1] |
| if hasattr(cs, '_desubroutinized'): |
| raise StopHintCountEvent() |
| |
| def op_hintmask(self, index): |
| psCharStrings.SimpleT2Decompiler.op_hintmask(self, index) |
| if self.need_hintcount: |
| self.stop_hint_count() |
| |
| def processSubr(self, index, subr): |
| cs = self.callingStack[-1] |
| if not hasattr(cs, '_desubroutinized'): |
| cs._patches.append((index, subr._desubroutinized)) |
| |
| |
| class CFFFontSet(object): |
| """A CFF font "file" can contain more than one font, although this is |
| extremely rare (and not allowed within OpenType fonts). |
| |
| This class is the entry point for parsing a CFF table. To actually |
| manipulate the data inside the CFF font, you will want to access the |
| ``CFFFontSet``'s :class:`TopDict` object. To do this, a ``CFFFontSet`` |
| object can either be treated as a dictionary (with appropriate |
| ``keys()`` and ``values()`` methods) mapping font names to :class:`TopDict` |
| objects, or as a list. |
| |
| .. code:: python |
| |
| from fontTools import ttLib |
| tt = ttLib.TTFont("Tests/cffLib/data/LinLibertine_RBI.otf") |
| tt["CFF "].cff |
| # <fontTools.cffLib.CFFFontSet object at 0x101e24c90> |
| tt["CFF "].cff[0] # Here's your actual font data |
| # <fontTools.cffLib.TopDict object at 0x1020f1fd0> |
| |
| """ |
| |
| def decompile(self, file, otFont, isCFF2=None): |
| """Parse a binary CFF file into an internal representation. ``file`` |
| should be a file handle object. ``otFont`` is the top-level |
| :py:class:`fontTools.ttLib.ttFont.TTFont` object containing this CFF file. |
| |
| If ``isCFF2`` is passed and set to ``True`` or ``False``, then the |
| library makes an assertion that the CFF header is of the appropriate |
| version. |
| """ |
| |
| self.otFont = otFont |
| sstruct.unpack(cffHeaderFormat, file.read(3), self) |
| if isCFF2 is not None: |
| # called from ttLib: assert 'major' as read from file matches the |
| # expected version |
| expected_major = (2 if isCFF2 else 1) |
| if self.major != expected_major: |
| raise ValueError( |
| "Invalid CFF 'major' version: expected %d, found %d" % |
| (expected_major, self.major)) |
| else: |
| # use 'major' version from file to determine if isCFF2 |
| assert self.major in (1, 2), "Unknown CFF format" |
| isCFF2 = self.major == 2 |
| if not isCFF2: |
| self.offSize = struct.unpack("B", file.read(1))[0] |
| file.seek(self.hdrSize) |
| self.fontNames = list(tostr(s) for s in Index(file, isCFF2=isCFF2)) |
| self.topDictIndex = TopDictIndex(file, isCFF2=isCFF2) |
| self.strings = IndexedStrings(file) |
| else: # isCFF2 |
| self.topDictSize = struct.unpack(">H", file.read(2))[0] |
| file.seek(self.hdrSize) |
| self.fontNames = ["CFF2Font"] |
| cff2GetGlyphOrder = otFont.getGlyphOrder |
| # in CFF2, offsetSize is the size of the TopDict data. |
| self.topDictIndex = TopDictIndex( |
| file, cff2GetGlyphOrder, self.topDictSize, isCFF2=isCFF2) |
| self.strings = None |
| self.GlobalSubrs = GlobalSubrsIndex(file, isCFF2=isCFF2) |
| self.topDictIndex.strings = self.strings |
| self.topDictIndex.GlobalSubrs = self.GlobalSubrs |
| |
| def __len__(self): |
| return len(self.fontNames) |
| |
| def keys(self): |
| return list(self.fontNames) |
| |
| def values(self): |
| return self.topDictIndex |
| |
| def __getitem__(self, nameOrIndex): |
| """ Return TopDict instance identified by name (str) or index (int |
| or any object that implements `__index__`). |
| """ |
| if hasattr(nameOrIndex, "__index__"): |
| index = nameOrIndex.__index__() |
| elif isinstance(nameOrIndex, str): |
| name = nameOrIndex |
| try: |
| index = self.fontNames.index(name) |
| except ValueError: |
| raise KeyError(nameOrIndex) |
| else: |
| raise TypeError(nameOrIndex) |
| return self.topDictIndex[index] |
| |
| def compile(self, file, otFont, isCFF2=None): |
| """Write the object back into binary representation onto the given file. |
| ``file`` should be a file handle object. ``otFont`` is the top-level |
| :py:class:`fontTools.ttLib.ttFont.TTFont` object containing this CFF file. |
| |
| If ``isCFF2`` is passed and set to ``True`` or ``False``, then the |
| library makes an assertion that the CFF header is of the appropriate |
| version. |
| """ |
| self.otFont = otFont |
| if isCFF2 is not None: |
| # called from ttLib: assert 'major' value matches expected version |
| expected_major = (2 if isCFF2 else 1) |
| if self.major != expected_major: |
| raise ValueError( |
| "Invalid CFF 'major' version: expected %d, found %d" % |
| (expected_major, self.major)) |
| else: |
| # use current 'major' value to determine output format |
| assert self.major in (1, 2), "Unknown CFF format" |
| isCFF2 = self.major == 2 |
| |
| if otFont.recalcBBoxes and not isCFF2: |
| for topDict in self.topDictIndex: |
| topDict.recalcFontBBox() |
| |
| if not isCFF2: |
| strings = IndexedStrings() |
| else: |
| strings = None |
| writer = CFFWriter(isCFF2) |
| topCompiler = self.topDictIndex.getCompiler(strings, self, isCFF2=isCFF2) |
| if isCFF2: |
| self.hdrSize = 5 |
| writer.add(sstruct.pack(cffHeaderFormat, self)) |
| # Note: topDictSize will most likely change in CFFWriter.toFile(). |
| self.topDictSize = topCompiler.getDataLength() |
| writer.add(struct.pack(">H", self.topDictSize)) |
| else: |
| self.hdrSize = 4 |
| self.offSize = 4 # will most likely change in CFFWriter.toFile(). |
| writer.add(sstruct.pack(cffHeaderFormat, self)) |
| writer.add(struct.pack("B", self.offSize)) |
| if not isCFF2: |
| fontNames = Index() |
| for name in self.fontNames: |
| fontNames.append(name) |
| writer.add(fontNames.getCompiler(strings, self, isCFF2=isCFF2)) |
| writer.add(topCompiler) |
| if not isCFF2: |
| writer.add(strings.getCompiler()) |
| writer.add(self.GlobalSubrs.getCompiler(strings, self, isCFF2=isCFF2)) |
| |
| for topDict in self.topDictIndex: |
| if not hasattr(topDict, "charset") or topDict.charset is None: |
| charset = otFont.getGlyphOrder() |
| topDict.charset = charset |
| children = topCompiler.getChildren(strings) |
| for child in children: |
| writer.add(child) |
| |
| writer.toFile(file) |
| |
| def toXML(self, xmlWriter): |
| """Write the object into XML representation onto the given |
| :class:`fontTools.misc.xmlWriter.XMLWriter`. |
| |
| .. code:: python |
| |
| writer = xmlWriter.XMLWriter(sys.stdout) |
| tt["CFF "].cff.toXML(writer) |
| |
| """ |
| |
| xmlWriter.simpletag("major", value=self.major) |
| xmlWriter.newline() |
| xmlWriter.simpletag("minor", value=self.minor) |
| xmlWriter.newline() |
| for fontName in self.fontNames: |
| xmlWriter.begintag("CFFFont", name=tostr(fontName)) |
| xmlWriter.newline() |
| font = self[fontName] |
| font.toXML(xmlWriter) |
| xmlWriter.endtag("CFFFont") |
| xmlWriter.newline() |
| xmlWriter.newline() |
| xmlWriter.begintag("GlobalSubrs") |
| xmlWriter.newline() |
| self.GlobalSubrs.toXML(xmlWriter) |
| xmlWriter.endtag("GlobalSubrs") |
| xmlWriter.newline() |
| |
| def fromXML(self, name, attrs, content, otFont=None): |
| """Reads data from the XML element into the ``CFFFontSet`` object.""" |
| self.otFont = otFont |
| |
| # set defaults. These will be replaced if there are entries for them |
| # in the XML file. |
| if not hasattr(self, "major"): |
| self.major = 1 |
| if not hasattr(self, "minor"): |
| self.minor = 0 |
| |
| if name == "CFFFont": |
| if self.major == 1: |
| if not hasattr(self, "offSize"): |
| # this will be recalculated when the cff is compiled. |
| self.offSize = 4 |
| if not hasattr(self, "hdrSize"): |
| self.hdrSize = 4 |
| if not hasattr(self, "GlobalSubrs"): |
| self.GlobalSubrs = GlobalSubrsIndex() |
| if not hasattr(self, "fontNames"): |
| self.fontNames = [] |
| self.topDictIndex = TopDictIndex() |
| fontName = attrs["name"] |
| self.fontNames.append(fontName) |
| topDict = TopDict(GlobalSubrs=self.GlobalSubrs) |
| topDict.charset = None # gets filled in later |
| elif self.major == 2: |
| if not hasattr(self, "hdrSize"): |
| self.hdrSize = 5 |
| if not hasattr(self, "GlobalSubrs"): |
| self.GlobalSubrs = GlobalSubrsIndex() |
| if not hasattr(self, "fontNames"): |
| self.fontNames = ["CFF2Font"] |
| cff2GetGlyphOrder = self.otFont.getGlyphOrder |
| topDict = TopDict( |
| GlobalSubrs=self.GlobalSubrs, |
| cff2GetGlyphOrder=cff2GetGlyphOrder) |
| self.topDictIndex = TopDictIndex(None, cff2GetGlyphOrder, None) |
| self.topDictIndex.append(topDict) |
| for element in content: |
| if isinstance(element, str): |
| continue |
| name, attrs, content = element |
| topDict.fromXML(name, attrs, content) |
| |
| if hasattr(topDict, "VarStore") and topDict.FDArray[0].vstore is None: |
| fdArray = topDict.FDArray |
| for fontDict in fdArray: |
| if hasattr(fontDict, "Private"): |
| fontDict.Private.vstore = topDict.VarStore |
| |
| elif name == "GlobalSubrs": |
| subrCharStringClass = psCharStrings.T2CharString |
| if not hasattr(self, "GlobalSubrs"): |
| self.GlobalSubrs = GlobalSubrsIndex() |
| for element in content: |
| if isinstance(element, str): |
| continue |
| name, attrs, content = element |
| subr = subrCharStringClass() |
| subr.fromXML(name, attrs, content) |
| self.GlobalSubrs.append(subr) |
| elif name == "major": |
| self.major = int(attrs['value']) |
| elif name == "minor": |
| self.minor = int(attrs['value']) |
| |
| def convertCFFToCFF2(self, otFont): |
| """Converts this object from CFF format to CFF2 format. This conversion |
| is done 'in-place'. The conversion cannot be reversed. |
| |
| This assumes a decompiled CFF table. (i.e. that the object has been |
| filled via :meth:`decompile`.)""" |
| self.major = 2 |
| cff2GetGlyphOrder = self.otFont.getGlyphOrder |
| topDictData = TopDictIndex(None, cff2GetGlyphOrder, None) |
| topDictData.items = self.topDictIndex.items |
| self.topDictIndex = topDictData |
| topDict = topDictData[0] |
| if hasattr(topDict, 'Private'): |
| privateDict = topDict.Private |
| else: |
| privateDict = None |
| opOrder = buildOrder(topDictOperators2) |
| topDict.order = opOrder |
| topDict.cff2GetGlyphOrder = cff2GetGlyphOrder |
| for entry in topDictOperators: |
| key = entry[1] |
| if key not in opOrder: |
| if key in topDict.rawDict: |
| del topDict.rawDict[key] |
| if hasattr(topDict, key): |
| delattr(topDict, key) |
| |
| if not hasattr(topDict, "FDArray"): |
| fdArray = topDict.FDArray = FDArrayIndex() |
| fdArray.strings = None |
| fdArray.GlobalSubrs = topDict.GlobalSubrs |
| topDict.GlobalSubrs.fdArray = fdArray |
| charStrings = topDict.CharStrings |
| if charStrings.charStringsAreIndexed: |
| charStrings.charStringsIndex.fdArray = fdArray |
| else: |
| charStrings.fdArray = fdArray |
| fontDict = FontDict() |
| fontDict.setCFF2(True) |
| fdArray.append(fontDict) |
| fontDict.Private = privateDict |
| privateOpOrder = buildOrder(privateDictOperators2) |
| for entry in privateDictOperators: |
| key = entry[1] |
| if key not in privateOpOrder: |
| if key in privateDict.rawDict: |
| # print "Removing private dict", key |
| del privateDict.rawDict[key] |
| if hasattr(privateDict, key): |
| delattr(privateDict, key) |
| # print "Removing privateDict attr", key |
| else: |
| # clean up the PrivateDicts in the fdArray |
| fdArray = topDict.FDArray |
| privateOpOrder = buildOrder(privateDictOperators2) |
| for fontDict in fdArray: |
| fontDict.setCFF2(True) |
| for key in fontDict.rawDict.keys(): |
| if key not in fontDict.order: |
| del fontDict.rawDict[key] |
| if hasattr(fontDict, key): |
| delattr(fontDict, key) |
| |
| privateDict = fontDict.Private |
| for entry in privateDictOperators: |
| key = entry[1] |
| if key not in privateOpOrder: |
| if key in privateDict.rawDict: |
| # print "Removing private dict", key |
| del privateDict.rawDict[key] |
| if hasattr(privateDict, key): |
| delattr(privateDict, key) |
| # print "Removing privateDict attr", key |
| # At this point, the Subrs and Charstrings are all still T2Charstring class |
| # easiest to fix this by compiling, then decompiling again |
| file = BytesIO() |
| self.compile(file, otFont, isCFF2=True) |
| file.seek(0) |
| self.decompile(file, otFont, isCFF2=True) |
| |
| def desubroutinize(self): |
| for fontName in self.fontNames: |
| font = self[fontName] |
| cs = font.CharStrings |
| for g in font.charset: |
| c, _ = cs.getItemAndSelector(g) |
| c.decompile() |
| subrs = getattr(c.private, "Subrs", []) |
| decompiler = _DesubroutinizingT2Decompiler(subrs, c.globalSubrs, c.private) |
| decompiler.execute(c) |
| c.program = c._desubroutinized |
| del c._desubroutinized |
| # Delete all the local subrs |
| if hasattr(font, 'FDArray'): |
| for fd in font.FDArray: |
| pd = fd.Private |
| if hasattr(pd, 'Subrs'): |
| del pd.Subrs |
| if 'Subrs' in pd.rawDict: |
| del pd.rawDict['Subrs'] |
| else: |
| pd = font.Private |
| if hasattr(pd, 'Subrs'): |
| del pd.Subrs |
| if 'Subrs' in pd.rawDict: |
| del pd.rawDict['Subrs'] |
| # as well as the global subrs |
| self.GlobalSubrs.clear() |
| |
| |
| class CFFWriter(object): |
| """Helper class for serializing CFF data to binary. Used by |
| :meth:`CFFFontSet.compile`.""" |
| def __init__(self, isCFF2): |
| self.data = [] |
| self.isCFF2 = isCFF2 |
| |
| def add(self, table): |
| self.data.append(table) |
| |
| def toFile(self, file): |
| lastPosList = None |
| count = 1 |
| while True: |
| log.log(DEBUG, "CFFWriter.toFile() iteration: %d", count) |
| count = count + 1 |
| pos = 0 |
| posList = [pos] |
| for item in self.data: |
| if hasattr(item, "getDataLength"): |
| endPos = pos + item.getDataLength() |
| if isinstance(item, TopDictIndexCompiler) and item.isCFF2: |
| self.topDictSize = item.getDataLength() |
| else: |
| endPos = pos + len(item) |
| if hasattr(item, "setPos"): |
| item.setPos(pos, endPos) |
| pos = endPos |
| posList.append(pos) |
| if posList == lastPosList: |
| break |
| lastPosList = posList |
| log.log(DEBUG, "CFFWriter.toFile() writing to file.") |
| begin = file.tell() |
| if self.isCFF2: |
| self.data[1] = struct.pack(">H", self.topDictSize) |
| else: |
| self.offSize = calcOffSize(lastPosList[-1]) |
| self.data[1] = struct.pack("B", self.offSize) |
| posList = [0] |
| for item in self.data: |
| if hasattr(item, "toFile"): |
| item.toFile(file) |
| else: |
| file.write(item) |
| posList.append(file.tell() - begin) |
| assert posList == lastPosList |
| |
| |
| def calcOffSize(largestOffset): |
| if largestOffset < 0x100: |
| offSize = 1 |
| elif largestOffset < 0x10000: |
| offSize = 2 |
| elif largestOffset < 0x1000000: |
| offSize = 3 |
| else: |
| offSize = 4 |
| return offSize |
| |
| |
| class IndexCompiler(object): |
| """Base class for writing CFF `INDEX data <https://docs.microsoft.com/en-us/typography/opentype/spec/cff2#5-index-data>`_ |
| to binary.""" |
| |
| def __init__(self, items, strings, parent, isCFF2=None): |
| if isCFF2 is None and hasattr(parent, "isCFF2"): |
| isCFF2 = parent.isCFF2 |
| assert isCFF2 is not None |
| self.isCFF2 = isCFF2 |
| self.items = self.getItems(items, strings) |
| self.parent = parent |
| |
| def getItems(self, items, strings): |
| return items |
| |
| def getOffsets(self): |
| # An empty INDEX contains only the count field. |
| if self.items: |
| pos = 1 |
| offsets = [pos] |
| for item in self.items: |
| if hasattr(item, "getDataLength"): |
| pos = pos + item.getDataLength() |
| else: |
| pos = pos + len(item) |
| offsets.append(pos) |
| else: |
| offsets = [] |
| return offsets |
| |
| def getDataLength(self): |
| if self.isCFF2: |
| countSize = 4 |
| else: |
| countSize = 2 |
| |
| if self.items: |
| lastOffset = self.getOffsets()[-1] |
| offSize = calcOffSize(lastOffset) |
| dataLength = ( |
| countSize + # count |
| 1 + # offSize |
| (len(self.items) + 1) * offSize + # the offsets |
| lastOffset - 1 # size of object data |
| ) |
| else: |
| # count. For empty INDEX tables, this is the only entry. |
| dataLength = countSize |
| |
| return dataLength |
| |
| def toFile(self, file): |
| offsets = self.getOffsets() |
| if self.isCFF2: |
| writeCard32(file, len(self.items)) |
| else: |
| writeCard16(file, len(self.items)) |
| # An empty INDEX contains only the count field. |
| if self.items: |
| offSize = calcOffSize(offsets[-1]) |
| writeCard8(file, offSize) |
| offSize = -offSize |
| pack = struct.pack |
| for offset in offsets: |
| binOffset = pack(">l", offset)[offSize:] |
| assert len(binOffset) == -offSize |
| file.write(binOffset) |
| for item in self.items: |
| if hasattr(item, "toFile"): |
| item.toFile(file) |
| else: |
| data = tobytes(item, encoding="latin1") |
| file.write(data) |
| |
| |
| class IndexedStringsCompiler(IndexCompiler): |
| |
| def getItems(self, items, strings): |
| return items.strings |
| |
| |
| class TopDictIndexCompiler(IndexCompiler): |
| """Helper class for writing the TopDict to binary.""" |
| |
| def getItems(self, items, strings): |
| out = [] |
| for item in items: |
| out.append(item.getCompiler(strings, self)) |
| return out |
| |
| def getChildren(self, strings): |
| children = [] |
| for topDict in self.items: |
| children.extend(topDict.getChildren(strings)) |
| return children |
| |
| def getOffsets(self): |
| if self.isCFF2: |
| offsets = [0, self.items[0].getDataLength()] |
| return offsets |
| else: |
| return super(TopDictIndexCompiler, self).getOffsets() |
| |
| def getDataLength(self): |
| if self.isCFF2: |
| dataLength = self.items[0].getDataLength() |
| return dataLength |
| else: |
| return super(TopDictIndexCompiler, self).getDataLength() |
| |
| def toFile(self, file): |
| if self.isCFF2: |
| self.items[0].toFile(file) |
| else: |
| super(TopDictIndexCompiler, self).toFile(file) |
| |
| |
| class FDArrayIndexCompiler(IndexCompiler): |
| """Helper class for writing the |
| `Font DICT INDEX <https://docs.microsoft.com/en-us/typography/opentype/spec/cff2#10-font-dict-index-font-dicts-and-fdselect>`_ |
| to binary.""" |
| |
| def getItems(self, items, strings): |
| out = [] |
| for item in items: |
| out.append(item.getCompiler(strings, self)) |
| return out |
| |
| def getChildren(self, strings): |
| children = [] |
| for fontDict in self.items: |
| children.extend(fontDict.getChildren(strings)) |
| return children |
| |
| def toFile(self, file): |
| offsets = self.getOffsets() |
| if self.isCFF2: |
| writeCard32(file, len(self.items)) |
| else: |
| writeCard16(file, len(self.items)) |
| offSize = calcOffSize(offsets[-1]) |
| writeCard8(file, offSize) |
| offSize = -offSize |
| pack = struct.pack |
| for offset in offsets: |
| binOffset = pack(">l", offset)[offSize:] |
| assert len(binOffset) == -offSize |
| file.write(binOffset) |
| for item in self.items: |
| if hasattr(item, "toFile"): |
| item.toFile(file) |
| else: |
| file.write(item) |
| |
| def setPos(self, pos, endPos): |
| self.parent.rawDict["FDArray"] = pos |
| |
| |
| class GlobalSubrsCompiler(IndexCompiler): |
| """Helper class for writing the `global subroutine INDEX <https://docs.microsoft.com/en-us/typography/opentype/spec/cff2#9-local-and-global-subr-indexes>`_ |
| to binary.""" |
| |
| def getItems(self, items, strings): |
| out = [] |
| for cs in items: |
| cs.compile(self.isCFF2) |
| out.append(cs.bytecode) |
| return out |
| |
| |
| class SubrsCompiler(GlobalSubrsCompiler): |
| """Helper class for writing the `local subroutine INDEX <https://docs.microsoft.com/en-us/typography/opentype/spec/cff2#9-local-and-global-subr-indexes>`_ |
| to binary.""" |
| |
| def setPos(self, pos, endPos): |
| offset = pos - self.parent.pos |
| self.parent.rawDict["Subrs"] = offset |
| |
| |
| class CharStringsCompiler(GlobalSubrsCompiler): |
| """Helper class for writing the `CharStrings INDEX <https://docs.microsoft.com/en-us/typography/opentype/spec/cff2#9-local-and-global-subr-indexes>`_ |
| to binary.""" |
| def getItems(self, items, strings): |
| out = [] |
| for cs in items: |
| cs.compile(self.isCFF2) |
| out.append(cs.bytecode) |
| return out |
| |
| def setPos(self, pos, endPos): |
| self.parent.rawDict["CharStrings"] = pos |
| |
| |
| class Index(object): |
| """This class represents what the CFF spec calls an INDEX (an array of |
| variable-sized objects). `Index` items can be addressed and set using |
| Python list indexing.""" |
| |
| compilerClass = IndexCompiler |
| |
| def __init__(self, file=None, isCFF2=None): |
| assert (isCFF2 is None) == (file is None) |
| self.items = [] |
| name = self.__class__.__name__ |
| if file is None: |
| return |
| self._isCFF2 = isCFF2 |
| log.log(DEBUG, "loading %s at %s", name, file.tell()) |
| self.file = file |
| if isCFF2: |
| count = readCard32(file) |
| else: |
| count = readCard16(file) |
| if count == 0: |
| return |
| self.items = [None] * count |
| offSize = readCard8(file) |
| log.log(DEBUG, " index count: %s offSize: %s", count, offSize) |
| assert offSize <= 4, "offSize too large: %s" % offSize |
| self.offsets = offsets = [] |
| pad = b'\0' * (4 - offSize) |
| for index in range(count + 1): |
| chunk = file.read(offSize) |
| chunk = pad + chunk |
| offset, = struct.unpack(">L", chunk) |
| offsets.append(int(offset)) |
| self.offsetBase = file.tell() - 1 |
| file.seek(self.offsetBase + offsets[-1]) # pretend we've read the whole lot |
| log.log(DEBUG, " end of %s at %s", name, file.tell()) |
| |
| def __len__(self): |
| return len(self.items) |
| |
| def __getitem__(self, index): |
| item = self.items[index] |
| if item is not None: |
| return item |
| offset = self.offsets[index] + self.offsetBase |
| size = self.offsets[index + 1] - self.offsets[index] |
| file = self.file |
| file.seek(offset) |
| data = file.read(size) |
| assert len(data) == size |
| item = self.produceItem(index, data, file, offset) |
| self.items[index] = item |
| return item |
| |
| def __setitem__(self, index, item): |
| self.items[index] = item |
| |
| def produceItem(self, index, data, file, offset): |
| return data |
| |
| def append(self, item): |
| """Add an item to an INDEX.""" |
| self.items.append(item) |
| |
| def getCompiler(self, strings, parent, isCFF2=None): |
| return self.compilerClass(self, strings, parent, isCFF2=isCFF2) |
| |
| def clear(self): |
| """Empty the INDEX.""" |
| del self.items[:] |
| |
| |
| class GlobalSubrsIndex(Index): |
| """This index contains all the global subroutines in the font. A global |
| subroutine is a set of ``CharString`` data which is accessible to any |
| glyph in the font, and are used to store repeated instructions - for |
| example, components may be encoded as global subroutines, but so could |
| hinting instructions. |
| |
| Remember that when interpreting a ``callgsubr`` instruction (or indeed |
| a ``callsubr`` instruction) that you will need to add the "subroutine |
| number bias" to number given: |
| |
| .. code:: python |
| |
| tt = ttLib.TTFont("Almendra-Bold.otf") |
| u = tt["CFF "].cff[0].CharStrings["udieresis"] |
| u.decompile() |
| |
| u.toXML(XMLWriter(sys.stdout)) |
| # <some stuff> |
| # -64 callgsubr <-- Subroutine which implements the dieresis mark |
| # <other stuff> |
| |
| tt["CFF "].cff[0].GlobalSubrs[-64] # <-- WRONG |
| # <T2CharString (bytecode) at 103451d10> |
| |
| tt["CFF "].cff[0].GlobalSubrs[-64 + 107] # <-- RIGHT |
| # <T2CharString (source) at 103451390> |
| |
| ("The bias applied depends on the number of subrs (gsubrs). If the number of |
| subrs (gsubrs) is less than 1240, the bias is 107. Otherwise if it is less |
| than 33900, it is 1131; otherwise it is 32768.", |
| `Subroutine Operators <https://docs.microsoft.com/en-us/typography/opentype/otspec180/cff2charstr#section4.4>`) |
| """ |
| |
| compilerClass = GlobalSubrsCompiler |
| subrClass = psCharStrings.T2CharString |
| charStringClass = psCharStrings.T2CharString |
| |
| def __init__(self, file=None, globalSubrs=None, private=None, |
| fdSelect=None, fdArray=None, isCFF2=None): |
| super(GlobalSubrsIndex, self).__init__(file, isCFF2=isCFF2) |
| self.globalSubrs = globalSubrs |
| self.private = private |
| if fdSelect: |
| self.fdSelect = fdSelect |
| if fdArray: |
| self.fdArray = fdArray |
| |
| def produceItem(self, index, data, file, offset): |
| if self.private is not None: |
| private = self.private |
| elif hasattr(self, 'fdArray') and self.fdArray is not None: |
| if hasattr(self, 'fdSelect') and self.fdSelect is not None: |
| fdIndex = self.fdSelect[index] |
| else: |
| fdIndex = 0 |
| private = self.fdArray[fdIndex].Private |
| else: |
| private = None |
| return self.subrClass(data, private=private, globalSubrs=self.globalSubrs) |
| |
| def toXML(self, xmlWriter): |
| """Write the subroutines index into XML representation onto the given |
| :class:`fontTools.misc.xmlWriter.XMLWriter`. |
| |
| .. code:: python |
| |
| writer = xmlWriter.XMLWriter(sys.stdout) |
| tt["CFF "].cff[0].GlobalSubrs.toXML(writer) |
| |
| """ |
| xmlWriter.comment( |
| "The 'index' attribute is only for humans; " |
| "it is ignored when parsed.") |
| xmlWriter.newline() |
| for i in range(len(self)): |
| subr = self[i] |
| if subr.needsDecompilation(): |
| xmlWriter.begintag("CharString", index=i, raw=1) |
| else: |
| xmlWriter.begintag("CharString", index=i) |
| xmlWriter.newline() |
| subr.toXML(xmlWriter) |
| xmlWriter.endtag("CharString") |
| xmlWriter.newline() |
| |
| def fromXML(self, name, attrs, content): |
| if name != "CharString": |
| return |
| subr = self.subrClass() |
| subr.fromXML(name, attrs, content) |
| self.append(subr) |
| |
| def getItemAndSelector(self, index): |
| sel = None |
| if hasattr(self, 'fdSelect'): |
| sel = self.fdSelect[index] |
| return self[index], sel |
| |
| |
| class SubrsIndex(GlobalSubrsIndex): |
| """This index contains a glyph's local subroutines. A local subroutine is a |
| private set of ``CharString`` data which is accessible only to the glyph to |
| which the index is attached.""" |
| |
| compilerClass = SubrsCompiler |
| |
| |
| class TopDictIndex(Index): |
| """This index represents the array of ``TopDict`` structures in the font |
| (again, usually only one entry is present). Hence the following calls are |
| equivalent: |
| |
| .. code:: python |
| |
| tt["CFF "].cff[0] |
| # <fontTools.cffLib.TopDict object at 0x102ed6e50> |
| tt["CFF "].cff.topDictIndex[0] |
| # <fontTools.cffLib.TopDict object at 0x102ed6e50> |
| |
| """ |
| |
| compilerClass = TopDictIndexCompiler |
| |
| def __init__(self, file=None, cff2GetGlyphOrder=None, topSize=0, |
| isCFF2=None): |
| assert (isCFF2 is None) == (file is None) |
| self.cff2GetGlyphOrder = cff2GetGlyphOrder |
| if file is not None and isCFF2: |
| self._isCFF2 = isCFF2 |
| self.items = [] |
| name = self.__class__.__name__ |
| log.log(DEBUG, "loading %s at %s", name, file.tell()) |
| self.file = file |
| count = 1 |
| self.items = [None] * count |
| self.offsets = [0, topSize] |
| self.offsetBase = file.tell() |
| # pretend we've read the whole lot |
| file.seek(self.offsetBase + topSize) |
| log.log(DEBUG, " end of %s at %s", name, file.tell()) |
| else: |
| super(TopDictIndex, self).__init__(file, isCFF2=isCFF2) |
| |
| def produceItem(self, index, data, file, offset): |
| top = TopDict( |
| self.strings, file, offset, self.GlobalSubrs, |
| self.cff2GetGlyphOrder, isCFF2=self._isCFF2) |
| top.decompile(data) |
| return top |
| |
| def toXML(self, xmlWriter): |
| for i in range(len(self)): |
| xmlWriter.begintag("FontDict", index=i) |
| xmlWriter.newline() |
| self[i].toXML(xmlWriter) |
| xmlWriter.endtag("FontDict") |
| xmlWriter.newline() |
| |
| |
| class FDArrayIndex(Index): |
| |
| compilerClass = FDArrayIndexCompiler |
| |
| def toXML(self, xmlWriter): |
| for i in range(len(self)): |
| xmlWriter.begintag("FontDict", index=i) |
| xmlWriter.newline() |
| self[i].toXML(xmlWriter) |
| xmlWriter.endtag("FontDict") |
| xmlWriter.newline() |
| |
| def produceItem(self, index, data, file, offset): |
| fontDict = FontDict( |
| self.strings, file, offset, self.GlobalSubrs, isCFF2=self._isCFF2, |
| vstore=self.vstore) |
| fontDict.decompile(data) |
| return fontDict |
| |
| def fromXML(self, name, attrs, content): |
| if name != "FontDict": |
| return |
| fontDict = FontDict() |
| for element in content: |
| if isinstance(element, str): |
| continue |
| name, attrs, content = element |
| fontDict.fromXML(name, attrs, content) |
| self.append(fontDict) |
| |
| |
| class VarStoreData(object): |
| |
| def __init__(self, file=None, otVarStore=None): |
| self.file = file |
| self.data = None |
| self.otVarStore = otVarStore |
| self.font = TTFont() # dummy font for the decompile function. |
| |
| def decompile(self): |
| if self.file: |
| class GlobalState(object): |
| def __init__(self, tableType, cachingStats): |
| self.tableType = tableType |
| self.cachingStats = cachingStats |
| globalState = GlobalState(tableType="VarStore", cachingStats={}) |
| # read data in from file. Assume position is correct. |
| length = readCard16(self.file) |
| self.data = self.file.read(length) |
| globalState = {} |
| reader = OTTableReader(self.data, globalState) |
| self.otVarStore = ot.VarStore() |
| self.otVarStore.decompile(reader, self.font) |
| return self |
| |
| def compile(self): |
| writer = OTTableWriter() |
| self.otVarStore.compile(writer, self.font) |
| # Note that this omits the initial Card16 length from the CFF2 |
| # VarStore data block |
| self.data = writer.getAllData() |
| |
| def writeXML(self, xmlWriter, name): |
| self.otVarStore.toXML(xmlWriter, self.font) |
| |
| def xmlRead(self, name, attrs, content, parent): |
| self.otVarStore = ot.VarStore() |
| for element in content: |
| if isinstance(element, tuple): |
| name, attrs, content = element |
| self.otVarStore.fromXML(name, attrs, content, self.font) |
| else: |
| pass |
| return None |
| |
| def __len__(self): |
| return len(self.data) |
| |
| def getNumRegions(self, vsIndex): |
| varData = self.otVarStore.VarData[vsIndex] |
| numRegions = varData.VarRegionCount |
| return numRegions |
| |
| |
| class FDSelect(object): |
| |
| def __init__(self, file=None, numGlyphs=None, format=None): |
| if file: |
| # read data in from file |
| self.format = readCard8(file) |
| if self.format == 0: |
| from array import array |
| self.gidArray = array("B", file.read(numGlyphs)).tolist() |
| elif self.format == 3: |
| gidArray = [None] * numGlyphs |
| nRanges = readCard16(file) |
| fd = None |
| prev = None |
| for i in range(nRanges): |
| first = readCard16(file) |
| if prev is not None: |
| for glyphID in range(prev, first): |
| gidArray[glyphID] = fd |
| prev = first |
| fd = readCard8(file) |
| if prev is not None: |
| first = readCard16(file) |
| for glyphID in range(prev, first): |
| gidArray[glyphID] = fd |
| self.gidArray = gidArray |
| elif self.format == 4: |
| gidArray = [None] * numGlyphs |
| nRanges = readCard32(file) |
| fd = None |
| prev = None |
| for i in range(nRanges): |
| first = readCard32(file) |
| if prev is not None: |
| for glyphID in range(prev, first): |
| gidArray[glyphID] = fd |
| prev = first |
| fd = readCard16(file) |
| if prev is not None: |
| first = readCard32(file) |
| for glyphID in range(prev, first): |
| gidArray[glyphID] = fd |
| self.gidArray = gidArray |
| else: |
| assert False, "unsupported FDSelect format: %s" % format |
| else: |
| # reading from XML. Make empty gidArray, and leave format as passed in. |
| # format is None will result in the smallest representation being used. |
| self.format = format |
| self.gidArray = [] |
| |
| def __len__(self): |
| return len(self.gidArray) |
| |
| def __getitem__(self, index): |
| return self.gidArray[index] |
| |
| def __setitem__(self, index, fdSelectValue): |
| self.gidArray[index] = fdSelectValue |
| |
| def append(self, fdSelectValue): |
| self.gidArray.append(fdSelectValue) |
| |
| |
| class CharStrings(object): |
| """The ``CharStrings`` in the font represent the instructions for drawing |
| each glyph. This object presents a dictionary interface to the font's |
| CharStrings, indexed by glyph name: |
| |
| .. code:: python |
| |
| tt["CFF "].cff[0].CharStrings["a"] |
| # <T2CharString (bytecode) at 103451e90> |
| |
| See :class:`fontTools.misc.psCharStrings.T1CharString` and |
| :class:`fontTools.misc.psCharStrings.T2CharString` for how to decompile, |
| compile and interpret the glyph drawing instructions in the returned objects. |
| |
| """ |
| |
| def __init__(self, file, charset, globalSubrs, private, fdSelect, fdArray, |
| isCFF2=None): |
| self.globalSubrs = globalSubrs |
| if file is not None: |
| self.charStringsIndex = SubrsIndex( |
| file, globalSubrs, private, fdSelect, fdArray, isCFF2=isCFF2) |
| self.charStrings = charStrings = {} |
| for i in range(len(charset)): |
| charStrings[charset[i]] = i |
| # read from OTF file: charStrings.values() are indices into |
| # charStringsIndex. |
| self.charStringsAreIndexed = 1 |
| else: |
| self.charStrings = {} |
| # read from ttx file: charStrings.values() are actual charstrings |
| self.charStringsAreIndexed = 0 |
| self.private = private |
| if fdSelect is not None: |
| self.fdSelect = fdSelect |
| if fdArray is not None: |
| self.fdArray = fdArray |
| |
| def keys(self): |
| return list(self.charStrings.keys()) |
| |
| def values(self): |
| if self.charStringsAreIndexed: |
| return self.charStringsIndex |
| else: |
| return list(self.charStrings.values()) |
| |
| def has_key(self, name): |
| return name in self.charStrings |
| |
| __contains__ = has_key |
| |
| def __len__(self): |
| return len(self.charStrings) |
| |
| def __getitem__(self, name): |
| charString = self.charStrings[name] |
| if self.charStringsAreIndexed: |
| charString = self.charStringsIndex[charString] |
| return charString |
| |
| def __setitem__(self, name, charString): |
| if self.charStringsAreIndexed: |
| index = self.charStrings[name] |
| self.charStringsIndex[index] = charString |
| else: |
| self.charStrings[name] = charString |
| |
| def getItemAndSelector(self, name): |
| if self.charStringsAreIndexed: |
| index = self.charStrings[name] |
| return self.charStringsIndex.getItemAndSelector(index) |
| else: |
| if hasattr(self, 'fdArray'): |
| if hasattr(self, 'fdSelect'): |
| sel = self.charStrings[name].fdSelectIndex |
| else: |
| sel = 0 |
| else: |
| sel = None |
| return self.charStrings[name], sel |
| |
| def toXML(self, xmlWriter): |
| names = sorted(self.keys()) |
| for name in names: |
| charStr, fdSelectIndex = self.getItemAndSelector(name) |
| if charStr.needsDecompilation(): |
| raw = [("raw", 1)] |
| else: |
| raw = [] |
| if fdSelectIndex is None: |
| xmlWriter.begintag("CharString", [('name', name)] + raw) |
| else: |
| xmlWriter.begintag( |
| "CharString", |
| [('name', name), ('fdSelectIndex', fdSelectIndex)] + raw) |
| xmlWriter.newline() |
| charStr.toXML(xmlWriter) |
| xmlWriter.endtag("CharString") |
| xmlWriter.newline() |
| |
| def fromXML(self, name, attrs, content): |
| for element in content: |
| if isinstance(element, str): |
| continue |
| name, attrs, content = element |
| if name != "CharString": |
| continue |
| fdID = -1 |
| if hasattr(self, "fdArray"): |
| try: |
| fdID = safeEval(attrs["fdSelectIndex"]) |
| except KeyError: |
| fdID = 0 |
| private = self.fdArray[fdID].Private |
| else: |
| private = self.private |
| |
| glyphName = attrs["name"] |
| charStringClass = psCharStrings.T2CharString |
| charString = charStringClass( |
| private=private, |
| globalSubrs=self.globalSubrs) |
| charString.fromXML(name, attrs, content) |
| if fdID >= 0: |
| charString.fdSelectIndex = fdID |
| self[glyphName] = charString |
| |
| |
| def readCard8(file): |
| return byteord(file.read(1)) |
| |
| |
| def readCard16(file): |
| value, = struct.unpack(">H", file.read(2)) |
| return value |
| |
| |
| def readCard32(file): |
| value, = struct.unpack(">L", file.read(4)) |
| return value |
| |
| |
| def writeCard8(file, value): |
| file.write(bytechr(value)) |
| |
| |
| def writeCard16(file, value): |
| file.write(struct.pack(">H", value)) |
| |
| |
| def writeCard32(file, value): |
| file.write(struct.pack(">L", value)) |
| |
| |
| def packCard8(value): |
| return bytechr(value) |
| |
| |
| def packCard16(value): |
| return struct.pack(">H", value) |
| |
| |
| def packCard32(value): |
| return struct.pack(">L", value) |
| |
| |
| def buildOperatorDict(table): |
| d = {} |
| for op, name, arg, default, conv in table: |
| d[op] = (name, arg) |
| return d |
| |
| |
| def buildOpcodeDict(table): |
| d = {} |
| for op, name, arg, default, conv in table: |
| if isinstance(op, tuple): |
| op = bytechr(op[0]) + bytechr(op[1]) |
| else: |
| op = bytechr(op) |
| d[name] = (op, arg) |
| return d |
| |
| |
| def buildOrder(table): |
| l = [] |
| for op, name, arg, default, conv in table: |
| l.append(name) |
| return l |
| |
| |
| def buildDefaults(table): |
| d = {} |
| for op, name, arg, default, conv in table: |
| if default is not None: |
| d[name] = default |
| return d |
| |
| |
| def buildConverters(table): |
| d = {} |
| for op, name, arg, default, conv in table: |
| d[name] = conv |
| return d |
| |
| |
| class SimpleConverter(object): |
| |
| def read(self, parent, value): |
| if not hasattr(parent, "file"): |
| return self._read(parent, value) |
| file = parent.file |
| pos = file.tell() |
| try: |
| return self._read(parent, value) |
| finally: |
| file.seek(pos) |
| |
| def _read(self, parent, value): |
| return value |
| |
| def write(self, parent, value): |
| return value |
| |
| def xmlWrite(self, xmlWriter, name, value): |
| xmlWriter.simpletag(name, value=value) |
| xmlWriter.newline() |
| |
| def xmlRead(self, name, attrs, content, parent): |
| return attrs["value"] |
| |
| |
| class ASCIIConverter(SimpleConverter): |
| |
| def _read(self, parent, value): |
| return tostr(value, encoding='ascii') |
| |
| def write(self, parent, value): |
| return tobytes(value, encoding='ascii') |
| |
| def xmlWrite(self, xmlWriter, name, value): |
| xmlWriter.simpletag(name, value=tostr(value, encoding="ascii")) |
| xmlWriter.newline() |
| |
| def xmlRead(self, name, attrs, content, parent): |
| return tobytes(attrs["value"], encoding=("ascii")) |
| |
| |
| class Latin1Converter(SimpleConverter): |
| |
| def _read(self, parent, value): |
| return tostr(value, encoding='latin1') |
| |
| def write(self, parent, value): |
| return tobytes(value, encoding='latin1') |
| |
| def xmlWrite(self, xmlWriter, name, value): |
| value = tostr(value, encoding="latin1") |
| if name in ['Notice', 'Copyright']: |
| value = re.sub(r"[\r\n]\s+", " ", value) |
| xmlWriter.simpletag(name, value=value) |
| xmlWriter.newline() |
| |
| def xmlRead(self, name, attrs, content, parent): |
| return tobytes(attrs["value"], encoding=("latin1")) |
| |
| |
| def parseNum(s): |
| try: |
| value = int(s) |
| except: |
| value = float(s) |
| return value |
| |
| |
| def parseBlendList(s): |
| valueList = [] |
| for element in s: |
| if isinstance(element, str): |
| continue |
| name, attrs, content = element |
| blendList = attrs["value"].split() |
| blendList = [eval(val) for val in blendList] |
| valueList.append(blendList) |
| if len(valueList) == 1: |
| valueList = valueList[0] |
| return valueList |
| |
| |
| class NumberConverter(SimpleConverter): |
| def xmlWrite(self, xmlWriter, name, value): |
| if isinstance(value, list): |
| xmlWriter.begintag(name) |
| xmlWriter.newline() |
| xmlWriter.indent() |
| blendValue = " ".join([str(val) for val in value]) |
| xmlWriter.simpletag(kBlendDictOpName, value=blendValue) |
| xmlWriter.newline() |
| xmlWriter.dedent() |
| xmlWriter.endtag(name) |
| xmlWriter.newline() |
| else: |
| xmlWriter.simpletag(name, value=value) |
| xmlWriter.newline() |
| |
| def xmlRead(self, name, attrs, content, parent): |
| valueString = attrs.get("value", None) |
| if valueString is None: |
| value = parseBlendList(content) |
| else: |
| value = parseNum(attrs["value"]) |
| return value |
| |
| |
| class ArrayConverter(SimpleConverter): |
| def xmlWrite(self, xmlWriter, name, value): |
| if value and isinstance(value[0], list): |
| xmlWriter.begintag(name) |
| xmlWriter.newline() |
| xmlWriter.indent() |
| for valueList in value: |
| blendValue = " ".join([str(val) for val in valueList]) |
| xmlWriter.simpletag(kBlendDictOpName, value=blendValue) |
| xmlWriter.newline() |
| xmlWriter.dedent() |
| xmlWriter.endtag(name) |
| xmlWriter.newline() |
| else: |
| value = " ".join([str(val) for val in value]) |
| xmlWriter.simpletag(name, value=value) |
| xmlWriter.newline() |
| |
| def xmlRead(self, name, attrs, content, parent): |
| valueString = attrs.get("value", None) |
| if valueString is None: |
| valueList = parseBlendList(content) |
| else: |
| values = valueString.split() |
| valueList = [parseNum(value) for value in values] |
| return valueList |
| |
| |
| class TableConverter(SimpleConverter): |
| |
| def xmlWrite(self, xmlWriter, name, value): |
| xmlWriter.begintag(name) |
| xmlWriter.newline() |
| value.toXML(xmlWriter) |
| xmlWriter.endtag(name) |
| xmlWriter.newline() |
| |
| def xmlRead(self, name, attrs, content, parent): |
| ob = self.getClass()() |
| for element in content: |
| if isinstance(element, str): |
| continue |
| name, attrs, content = element |
| ob.fromXML(name, attrs, content) |
| return ob |
| |
| |
| class PrivateDictConverter(TableConverter): |
| |
| def getClass(self): |
| return PrivateDict |
| |
| def _read(self, parent, value): |
| size, offset = value |
| file = parent.file |
| isCFF2 = parent._isCFF2 |
| try: |
| vstore = parent.vstore |
| except AttributeError: |
| vstore = None |
| priv = PrivateDict( |
| parent.strings, file, offset, isCFF2=isCFF2, vstore=vstore) |
| file.seek(offset) |
| data = file.read(size) |
| assert len(data) == size |
| priv.decompile(data) |
| return priv |
| |
| def write(self, parent, value): |
| return (0, 0) # dummy value |
| |
| |
| class SubrsConverter(TableConverter): |
| |
| def getClass(self): |
| return SubrsIndex |
| |
| def _read(self, parent, value): |
| file = parent.file |
| isCFF2 = parent._isCFF2 |
| file.seek(parent.offset + value) # Offset(self) |
| return SubrsIndex(file, isCFF2=isCFF2) |
| |
| def write(self, parent, value): |
| return 0 # dummy value |
| |
| |
| class CharStringsConverter(TableConverter): |
| |
| def _read(self, parent, value): |
| file = parent.file |
| isCFF2 = parent._isCFF2 |
| charset = parent.charset |
| globalSubrs = parent.GlobalSubrs |
| if hasattr(parent, "FDArray"): |
| fdArray = parent.FDArray |
| if hasattr(parent, "FDSelect"): |
| fdSelect = parent.FDSelect |
| else: |
| fdSelect = None |
| private = None |
| else: |
| fdSelect, fdArray = None, None |
| private = parent.Private |
| file.seek(value) # Offset(0) |
| charStrings = CharStrings( |
| file, charset, globalSubrs, private, fdSelect, fdArray, isCFF2=isCFF2) |
| return charStrings |
| |
| def write(self, parent, value): |
| return 0 # dummy value |
| |
| def xmlRead(self, name, attrs, content, parent): |
| if hasattr(parent, "FDArray"): |
| # if it is a CID-keyed font, then the private Dict is extracted from the |
| # parent.FDArray |
| fdArray = parent.FDArray |
| if hasattr(parent, "FDSelect"): |
| fdSelect = parent.FDSelect |
| else: |
| fdSelect = None |
| private = None |
| else: |
| # if it is a name-keyed font, then the private dict is in the top dict, |
| # and |
| # there is no fdArray. |
| private, fdSelect, fdArray = parent.Private, None, None |
| charStrings = CharStrings( |
| None, None, parent.GlobalSubrs, private, fdSelect, fdArray) |
| charStrings.fromXML(name, attrs, content) |
| return charStrings |
| |
| |
| class CharsetConverter(SimpleConverter): |
| def _read(self, parent, value): |
| isCID = hasattr(parent, "ROS") |
| if value > 2: |
| numGlyphs = parent.numGlyphs |
| file = parent.file |
| file.seek(value) |
| log.log(DEBUG, "loading charset at %s", value) |
| format = readCard8(file) |
| if format == 0: |
| charset = parseCharset0(numGlyphs, file, parent.strings, isCID) |
| elif format == 1 or format == 2: |
| charset = parseCharset(numGlyphs, file, parent.strings, isCID, format) |
| else: |
| raise NotImplementedError |
| assert len(charset) == numGlyphs |
| log.log(DEBUG, " charset end at %s", file.tell()) |
| # make sure glyph names are unique |
| allNames = {} |
| newCharset = [] |
| for glyphName in charset: |
| if glyphName in allNames: |
| # make up a new glyphName that's unique |
| n = allNames[glyphName] |
| while (glyphName + "#" + str(n)) in allNames: |
| n += 1 |
| allNames[glyphName] = n + 1 |
| glyphName = glyphName + "#" + str(n) |
| allNames[glyphName] = 1 |
| newCharset.append(glyphName) |
| charset = newCharset |
| else: # offset == 0 -> no charset data. |
| if isCID or "CharStrings" not in parent.rawDict: |
| # We get here only when processing fontDicts from the FDArray of |
| # CFF-CID fonts. Only the real topDict references the chrset. |
| assert value == 0 |
| charset = None |
| elif value == 0: |
| charset = cffISOAdobeStrings |
| elif value == 1: |
| charset = cffIExpertStrings |
| elif value == 2: |
| charset = cffExpertSubsetStrings |
| if charset and (len(charset) != parent.numGlyphs): |
| charset = charset[:parent.numGlyphs] |
| return charset |
| |
| def write(self, parent, value): |
| return 0 # dummy value |
| |
| def xmlWrite(self, xmlWriter, name, value): |
| # XXX only write charset when not in OT/TTX context, where we |
| # dump charset as a separate "GlyphOrder" table. |
| # # xmlWriter.simpletag("charset") |
| xmlWriter.comment("charset is dumped separately as the 'GlyphOrder' element") |
| xmlWriter.newline() |
| |
| def xmlRead(self, name, attrs, content, parent): |
| pass |
| |
| |
| class CharsetCompiler(object): |
| |
| def __init__(self, strings, charset, parent): |
| assert charset[0] == '.notdef' |
| isCID = hasattr(parent.dictObj, "ROS") |
| data0 = packCharset0(charset, isCID, strings) |
| data = packCharset(charset, isCID, strings) |
| if len(data) < len(data0): |
| self.data = data |
| else: |
| self.data = data0 |
| self.parent = parent |
| |
| def setPos(self, pos, endPos): |
| self.parent.rawDict["charset"] = pos |
| |
| def getDataLength(self): |
| return len(self.data) |
| |
| def toFile(self, file): |
| file.write(self.data) |
| |
| |
| def getStdCharSet(charset): |
| # check to see if we can use a predefined charset value. |
| predefinedCharSetVal = None |
| predefinedCharSets = [ |
| (cffISOAdobeStringCount, cffISOAdobeStrings, 0), |
| (cffExpertStringCount, cffIExpertStrings, 1), |
| (cffExpertSubsetStringCount, cffExpertSubsetStrings, 2)] |
| lcs = len(charset) |
| for cnt, pcs, csv in predefinedCharSets: |
| if predefinedCharSetVal is not None: |
| break |
| if lcs > cnt: |
| continue |
| predefinedCharSetVal = csv |
| for i in range(lcs): |
| if charset[i] != pcs[i]: |
| predefinedCharSetVal = None |
| break |
| return predefinedCharSetVal |
| |
| |
| def getCIDfromName(name, strings): |
| return int(name[3:]) |
| |
| |
| def getSIDfromName(name, strings): |
| return strings.getSID(name) |
| |
| |
| def packCharset0(charset, isCID, strings): |
| fmt = 0 |
| data = [packCard8(fmt)] |
| if isCID: |
| getNameID = getCIDfromName |
| else: |
| getNameID = getSIDfromName |
| |
| for name in charset[1:]: |
| data.append(packCard16(getNameID(name, strings))) |
| return bytesjoin(data) |
| |
| |
| def packCharset(charset, isCID, strings): |
| fmt = 1 |
| ranges = [] |
| first = None |
| end = 0 |
| if isCID: |
| getNameID = getCIDfromName |
| else: |
| getNameID = getSIDfromName |
| |
| for name in charset[1:]: |
| SID = getNameID(name, strings) |
| if first is None: |
| first = SID |
| elif end + 1 != SID: |
| nLeft = end - first |
| if nLeft > 255: |
| fmt = 2 |
| ranges.append((first, nLeft)) |
| first = SID |
| end = SID |
| if end: |
| nLeft = end - first |
| if nLeft > 255: |
| fmt = 2 |
| ranges.append((first, nLeft)) |
| |
| data = [packCard8(fmt)] |
| if fmt == 1: |
| nLeftFunc = packCard8 |
| else: |
| nLeftFunc = packCard16 |
| for first, nLeft in ranges: |
| data.append(packCard16(first) + nLeftFunc(nLeft)) |
| return bytesjoin(data) |
| |
| |
| def parseCharset0(numGlyphs, file, strings, isCID): |
| charset = [".notdef"] |
| if isCID: |
| for i in range(numGlyphs - 1): |
| CID = readCard16(file) |
| charset.append("cid" + str(CID).zfill(5)) |
| else: |
| for i in range(numGlyphs - 1): |
| SID = readCard16(file) |
| charset.append(strings[SID]) |
| return charset |
| |
| |
| def parseCharset(numGlyphs, file, strings, isCID, fmt): |
| charset = ['.notdef'] |
| count = 1 |
| if fmt == 1: |
| nLeftFunc = readCard8 |
| else: |
| nLeftFunc = readCard16 |
| while count < numGlyphs: |
| first = readCard16(file) |
| nLeft = nLeftFunc(file) |
| if isCID: |
| for CID in range(first, first + nLeft + 1): |
| charset.append("cid" + str(CID).zfill(5)) |
| else: |
| for SID in range(first, first + nLeft + 1): |
| charset.append(strings[SID]) |
| count = count + nLeft + 1 |
| return charset |
| |
| |
| class EncodingCompiler(object): |
| |
| def __init__(self, strings, encoding, parent): |
| assert not isinstance(encoding, str) |
| data0 = packEncoding0(parent.dictObj.charset, encoding, parent.strings) |
| data1 = packEncoding1(parent.dictObj.charset, encoding, parent.strings) |
| if len(data0) < len(data1): |
| self.data = data0 |
| else: |
| self.data = data1 |
| self.parent = parent |
| |
| def setPos(self, pos, endPos): |
| self.parent.rawDict["Encoding"] = pos |
| |
| def getDataLength(self): |
| return len(self.data) |
| |
| def toFile(self, file): |
| file.write(self.data) |
| |
| |
| class EncodingConverter(SimpleConverter): |
| |
| def _read(self, parent, value): |
| if value == 0: |
| return "StandardEncoding" |
| elif value == 1: |
| return "ExpertEncoding" |
| else: |
| assert value > 1 |
| file = parent.file |
| file.seek(value) |
| log.log(DEBUG, "loading Encoding at %s", value) |
| fmt = readCard8(file) |
| haveSupplement = fmt & 0x80 |
| if haveSupplement: |
| raise NotImplementedError("Encoding supplements are not yet supported") |
| fmt = fmt & 0x7f |
| if fmt == 0: |
| encoding = parseEncoding0(parent.charset, file, haveSupplement, |
| parent.strings) |
| elif fmt == 1: |
| encoding = parseEncoding1(parent.charset, file, haveSupplement, |
| parent.strings) |
| return encoding |
| |
| def write(self, parent, value): |
| if value == "StandardEncoding": |
| return 0 |
| elif value == "ExpertEncoding": |
| return 1 |
| return 0 # dummy value |
| |
| def xmlWrite(self, xmlWriter, name, value): |
| if value in ("StandardEncoding", "ExpertEncoding"): |
| xmlWriter.simpletag(name, name=value) |
| xmlWriter.newline() |
| return |
| xmlWriter.begintag(name) |
| xmlWriter.newline() |
| for code in range(len(value)): |
| glyphName = value[code] |
| if glyphName != ".notdef": |
| xmlWriter.simpletag("map", code=hex(code), name=glyphName) |
| xmlWriter.newline() |
| xmlWriter.endtag(name) |
| xmlWriter.newline() |
| |
| def xmlRead(self, name, attrs, content, parent): |
| if "name" in attrs: |
| return attrs["name"] |
| encoding = [".notdef"] * 256 |
| for element in content: |
| if isinstance(element, str): |
| continue |
| name, attrs, content = element |
| code = safeEval(attrs["code"]) |
| glyphName = attrs["name"] |
| encoding[code] = glyphName |
| return encoding |
| |
| |
| def parseEncoding0(charset, file, haveSupplement, strings): |
| nCodes = readCard8(file) |
| encoding = [".notdef"] * 256 |
| for glyphID in range(1, nCodes + 1): |
| code = readCard8(file) |
| if code != 0: |
| encoding[code] = charset[glyphID] |
| return encoding |
| |
| |
| def parseEncoding1(charset, file, haveSupplement, strings): |
| nRanges = readCard8(file) |
| encoding = [".notdef"] * 256 |
| glyphID = 1 |
| for i in range(nRanges): |
| code = readCard8(file) |
| nLeft = readCard8(file) |
| for glyphID in range(glyphID, glyphID + nLeft + 1): |
| encoding[code] = charset[glyphID] |
| code = code + 1 |
| glyphID = glyphID + 1 |
| return encoding |
| |
| |
| def packEncoding0(charset, encoding, strings): |
| fmt = 0 |
| m = {} |
| for code in range(len(encoding)): |
| name = encoding[code] |
| if name != ".notdef": |
| m[name] = code |
| codes = [] |
| for name in charset[1:]: |
| code = m.get(name) |
| codes.append(code) |
| |
| while codes and codes[-1] is None: |
| codes.pop() |
| |
| data = [packCard8(fmt), packCard8(len(codes))] |
| for code in codes: |
| if code is None: |
| code = 0 |
| data.append(packCard8(code)) |
| return bytesjoin(data) |
| |
| |
| def packEncoding1(charset, encoding, strings): |
| fmt = 1 |
| m = {} |
| for code in range(len(encoding)): |
| name = encoding[code] |
| if name != ".notdef": |
| m[name] = code |
| ranges = [] |
| first = None |
| end = 0 |
| for name in charset[1:]: |
| code = m.get(name, -1) |
| if first is None: |
| first = code |
| elif end + 1 != code: |
| nLeft = end - first |
| ranges.append((first, nLeft)) |
| first = code |
| end = code |
| nLeft = end - first |
| ranges.append((first, nLeft)) |
| |
| # remove unencoded glyphs at the end. |
| while ranges and ranges[-1][0] == -1: |
| ranges.pop() |
| |
| data = [packCard8(fmt), packCard8(len(ranges))] |
| for first, nLeft in ranges: |
| if first == -1: # unencoded |
| first = 0 |
| data.append(packCard8(first) + packCard8(nLeft)) |
| return bytesjoin(data) |
| |
| |
| class FDArrayConverter(TableConverter): |
| |
| def _read(self, parent, value): |
| try: |
| vstore = parent.VarStore |
| except AttributeError: |
| vstore = None |
| file = parent.file |
| isCFF2 = parent._isCFF2 |
| file.seek(value) |
| fdArray = FDArrayIndex(file, isCFF2=isCFF2) |
| fdArray.vstore = vstore |
| fdArray.strings = parent.strings |
| fdArray.GlobalSubrs = parent.GlobalSubrs |
| return fdArray |
| |
| def write(self, parent, value): |
| return 0 # dummy value |
| |
| def xmlRead(self, name, attrs, content, parent): |
| fdArray = FDArrayIndex() |
| for element in content: |
| if isinstance(element, str): |
| continue |
| name, attrs, content = element |
| fdArray.fromXML(name, attrs, content) |
| return fdArray |
| |
| |
| class FDSelectConverter(SimpleConverter): |
| |
| def _read(self, parent, value): |
| file = parent.file |
| file.seek(value) |
| fdSelect = FDSelect(file, parent.numGlyphs) |
| return fdSelect |
| |
| def write(self, parent, value): |
| return 0 # dummy value |
| |
| # The FDSelect glyph data is written out to XML in the charstring keys, |
| # so we write out only the format selector |
| def xmlWrite(self, xmlWriter, name, value): |
| xmlWriter.simpletag(name, [('format', value.format)]) |
| xmlWriter.newline() |
| |
| def xmlRead(self, name, attrs, content, parent): |
| fmt = safeEval(attrs["format"]) |
| file = None |
| numGlyphs = None |
| fdSelect = FDSelect(file, numGlyphs, fmt) |
| return fdSelect |
| |
| |
| class VarStoreConverter(SimpleConverter): |
| |
| def _read(self, parent, value): |
| file = parent.file |
| file.seek(value) |
| varStore = VarStoreData(file) |
| varStore.decompile() |
| return varStore |
| |
| def write(self, parent, value): |
| return 0 # dummy value |
| |
| def xmlWrite(self, xmlWriter, name, value): |
| value.writeXML(xmlWriter, name) |
| |
| def xmlRead(self, name, attrs, content, parent): |
| varStore = VarStoreData() |
| varStore.xmlRead(name, attrs, content, parent) |
| return varStore |
| |
| |
| def packFDSelect0(fdSelectArray): |
| fmt = 0 |
| data = [packCard8(fmt)] |
| for index in fdSelectArray: |
| data.append(packCard8(index)) |
| return bytesjoin(data) |
| |
| |
| def packFDSelect3(fdSelectArray): |
| fmt = 3 |
| fdRanges = [] |
| lenArray = len(fdSelectArray) |
| lastFDIndex = -1 |
| for i in range(lenArray): |
| fdIndex = fdSelectArray[i] |
| if lastFDIndex != fdIndex: |
| fdRanges.append([i, fdIndex]) |
| lastFDIndex = fdIndex |
| sentinelGID = i + 1 |
| |
| data = [packCard8(fmt)] |
| data.append(packCard16(len(fdRanges))) |
| for fdRange in fdRanges: |
| data.append(packCard16(fdRange[0])) |
| data.append(packCard8(fdRange[1])) |
| data.append(packCard16(sentinelGID)) |
| return bytesjoin(data) |
| |
| |
| def packFDSelect4(fdSelectArray): |
| fmt = 4 |
| fdRanges = [] |
| lenArray = len(fdSelectArray) |
| lastFDIndex = -1 |
| for i in range(lenArray): |
| fdIndex = fdSelectArray[i] |
| if lastFDIndex != fdIndex: |
| fdRanges.append([i, fdIndex]) |
| lastFDIndex = fdIndex |
| sentinelGID = i + 1 |
| |
| data = [packCard8(fmt)] |
| data.append(packCard32(len(fdRanges))) |
| for fdRange in fdRanges: |
| data.append(packCard32(fdRange[0])) |
| data.append(packCard16(fdRange[1])) |
| data.append(packCard32(sentinelGID)) |
| return bytesjoin(data) |
| |
| |
| class FDSelectCompiler(object): |
| |
| def __init__(self, fdSelect, parent): |
| fmt = fdSelect.format |
| fdSelectArray = fdSelect.gidArray |
| if fmt == 0: |
| self.data = packFDSelect0(fdSelectArray) |
| elif fmt == 3: |
| self.data = packFDSelect3(fdSelectArray) |
| elif fmt == 4: |
| self.data = packFDSelect4(fdSelectArray) |
| else: |
| # choose smaller of the two formats |
| data0 = packFDSelect0(fdSelectArray) |
| data3 = packFDSelect3(fdSelectArray) |
| if len(data0) < len(data3): |
| self.data = data0 |
| fdSelect.format = 0 |
| else: |
| self.data = data3 |
| fdSelect.format = 3 |
| |
| self.parent = parent |
| |
| def setPos(self, pos, endPos): |
| self.parent.rawDict["FDSelect"] = pos |
| |
| def getDataLength(self): |
| return len(self.data) |
| |
| def toFile(self, file): |
| file.write(self.data) |
| |
| |
| class VarStoreCompiler(object): |
| |
| def __init__(self, varStoreData, parent): |
| self.parent = parent |
| if not varStoreData.data: |
| varStoreData.compile() |
| data = [ |
| packCard16(len(varStoreData.data)), |
| varStoreData.data |
| ] |
| self.data = bytesjoin(data) |
| |
| def setPos(self, pos, endPos): |
| self.parent.rawDict["VarStore"] = pos |
| |
| def getDataLength(self): |
| return len(self.data) |
| |
| def toFile(self, file): |
| file.write(self.data) |
| |
| |
| class ROSConverter(SimpleConverter): |
| |
| def xmlWrite(self, xmlWriter, name, value): |
| registry, order, supplement = value |
| xmlWriter.simpletag( |
| name, |
| [ |
| ('Registry', tostr(registry)), |
| ('Order', tostr(order)), |
| ('Supplement', supplement) |
| ]) |
| xmlWriter.newline() |
| |
| def xmlRead(self, name, attrs, content, parent): |
| return (attrs['Registry'], attrs['Order'], safeEval(attrs['Supplement'])) |
| |
| topDictOperators = [ |
| # opcode name argument type default converter |
| (25, 'maxstack', 'number', None, None), |
| ((12, 30), 'ROS', ('SID', 'SID', 'number'), None, ROSConverter()), |
| ((12, 20), 'SyntheticBase', 'number', None, None), |
| (0, 'version', 'SID', None, None), |
| (1, 'Notice', 'SID', None, Latin1Converter()), |
| ((12, 0), 'Copyright', 'SID', None, Latin1Converter()), |
| (2, 'FullName', 'SID', None, None), |
| ((12, 38), 'FontName', 'SID', None, None), |
| (3, 'FamilyName', 'SID', None, None), |
| (4, 'Weight', 'SID', None, None), |
| ((12, 1), 'isFixedPitch', 'number', 0, None), |
| ((12, 2), 'ItalicAngle', 'number', 0, None), |
| ((12, 3), 'UnderlinePosition', 'number', -100, None), |
| ((12, 4), 'UnderlineThickness', 'number', 50, None), |
| ((12, 5), 'PaintType', 'number', 0, None), |
| ((12, 6), 'CharstringType', 'number', 2, None), |
| ((12, 7), 'FontMatrix', 'array', [0.001, 0, 0, 0.001, 0, 0], None), |
| (13, 'UniqueID', 'number', None, None), |
| (5, 'FontBBox', 'array', [0, 0, 0, 0], None), |
| ((12, 8), 'StrokeWidth', 'number', 0, None), |
| (14, 'XUID', 'array', None, None), |
| ((12, 21), 'PostScript', 'SID', None, None), |
| ((12, 22), 'BaseFontName', 'SID', None, None), |
| ((12, 23), 'BaseFontBlend', 'delta', None, None), |
| ((12, 31), 'CIDFontVersion', 'number', 0, None), |
| ((12, 32), 'CIDFontRevision', 'number', 0, None), |
| ((12, 33), 'CIDFontType', 'number', 0, None), |
| ((12, 34), 'CIDCount', 'number', 8720, None), |
| (15, 'charset', 'number', None, CharsetConverter()), |
| ((12, 35), 'UIDBase', 'number', None, None), |
| (16, 'Encoding', 'number', 0, EncodingConverter()), |
| (18, 'Private', ('number', 'number'), None, PrivateDictConverter()), |
| ((12, 37), 'FDSelect', 'number', None, FDSelectConverter()), |
| ((12, 36), 'FDArray', 'number', None, FDArrayConverter()), |
| (17, 'CharStrings', 'number', None, CharStringsConverter()), |
| (24, 'VarStore', 'number', None, VarStoreConverter()), |
| ] |
| |
| topDictOperators2 = [ |
| # opcode name argument type default converter |
| (25, 'maxstack', 'number', None, None), |
| ((12, 7), 'FontMatrix', 'array', [0.001, 0, 0, 0.001, 0, 0], None), |
| ((12, 37), 'FDSelect', 'number', None, FDSelectConverter()), |
| ((12, 36), 'FDArray', 'number', None, FDArrayConverter()), |
| (17, 'CharStrings', 'number', None, CharStringsConverter()), |
| (24, 'VarStore', 'number', None, VarStoreConverter()), |
| ] |
| |
| # Note! FDSelect and FDArray must both preceed CharStrings in the output XML build order, |
| # in order for the font to compile back from xml. |
| |
| kBlendDictOpName = "blend" |
| blendOp = 23 |
| |
| privateDictOperators = [ |
| # opcode name argument type default converter |
| (22, "vsindex", 'number', None, None), |
| (blendOp, kBlendDictOpName, 'blendList', None, None), # This is for reading to/from XML: it not written to CFF. |
| (6, 'BlueValues', 'delta', None, None), |
| (7, 'OtherBlues', 'delta', None, None), |
| (8, 'FamilyBlues', 'delta', None, None), |
| (9, 'FamilyOtherBlues', 'delta', None, None), |
| ((12, 9), 'BlueScale', 'number', 0.039625, None), |
| ((12, 10), 'BlueShift', 'number', 7, None), |
| ((12, 11), 'BlueFuzz', 'number', 1, None), |
| (10, 'StdHW', 'number', None, None), |
| (11, 'StdVW', 'number', None, None), |
| ((12, 12), 'StemSnapH', 'delta', None, None), |
| ((12, 13), 'StemSnapV', 'delta', None, None), |
| ((12, 14), 'ForceBold', 'number', 0, None), |
| ((12, 15), 'ForceBoldThreshold', 'number', None, None), # deprecated |
| ((12, 16), 'lenIV', 'number', None, None), # deprecated |
| ((12, 17), 'LanguageGroup', 'number', 0, None), |
| ((12, 18), 'ExpansionFactor', 'number', 0.06, None), |
| ((12, 19), 'initialRandomSeed', 'number', 0, None), |
| (20, 'defaultWidthX', 'number', 0, None), |
| (21, 'nominalWidthX', 'number', 0, None), |
| (19, 'Subrs', 'number', None, SubrsConverter()), |
| ] |
| |
| privateDictOperators2 = [ |
| # opcode name argument type default converter |
| (22, "vsindex", 'number', None, None), |
| (blendOp, kBlendDictOpName, 'blendList', None, None), # This is for reading to/from XML: it not written to CFF. |
| (6, 'BlueValues', 'delta', None, None), |
| (7, 'OtherBlues', 'delta', None, None), |
| (8, 'FamilyBlues', 'delta', None, None), |
| (9, 'FamilyOtherBlues', 'delta', None, None), |
| ((12, 9), 'BlueScale', 'number', 0.039625, None), |
| ((12, 10), 'BlueShift', 'number', 7, None), |
| ((12, 11), 'BlueFuzz', 'number', 1, None), |
| (10, 'StdHW', 'number', None, None), |
| (11, 'StdVW', 'number', None, None), |
| ((12, 12), 'StemSnapH', 'delta', None, None), |
| ((12, 13), 'StemSnapV', 'delta', None, None), |
| ((12, 17), 'LanguageGroup', 'number', 0, None), |
| ((12, 18), 'ExpansionFactor', 'number', 0.06, None), |
| (19, 'Subrs', 'number', None, SubrsConverter()), |
| ] |
| |
| |
| def addConverters(table): |
| for i in range(len(table)): |
| op, name, arg, default, conv = table[i] |
| if conv is not None: |
| continue |
| if arg in ("delta", "array"): |
| conv = ArrayConverter() |
| elif arg == "number": |
| conv = NumberConverter() |
| elif arg == "SID": |
| conv = ASCIIConverter() |
| elif arg == 'blendList': |
| conv = None |
| else: |
| assert False |
| table[i] = op, name, arg, default, conv |
| |
| |
| addConverters(privateDictOperators) |
| addConverters(topDictOperators) |
| |
| |
| class TopDictDecompiler(psCharStrings.DictDecompiler): |
| operators = buildOperatorDict(topDictOperators) |
| |
| |
| class PrivateDictDecompiler(psCharStrings.DictDecompiler): |
| operators = buildOperatorDict(privateDictOperators) |
| |
| |
| class DictCompiler(object): |
| maxBlendStack = 0 |
| |
| def __init__(self, dictObj, strings, parent, isCFF2=None): |
| if strings: |
| assert isinstance(strings, IndexedStrings) |
| if isCFF2 is None and hasattr(parent, "isCFF2"): |
| isCFF2 = parent.isCFF2 |
| assert isCFF2 is not None |
| self.isCFF2 = isCFF2 |
| self.dictObj = dictObj |
| self.strings = strings |
| self.parent = parent |
| rawDict = {} |
| for name in dictObj.order: |
| value = getattr(dictObj, name, None) |
| if value is None: |
| continue |
| conv = dictObj.converters[name] |
| value = conv.write(dictObj, value) |
| if value == dictObj.defaults.get(name): |
| continue |
| rawDict[name] = value |
| self.rawDict = rawDict |
| |
| def setPos(self, pos, endPos): |
| pass |
| |
| def getDataLength(self): |
| return len(self.compile("getDataLength")) |
| |
| def compile(self, reason): |
| log.log(DEBUG, "-- compiling %s for %s", self.__class__.__name__, reason) |
| rawDict = self.rawDict |
| data = [] |
| for name in self.dictObj.order: |
| value = rawDict.get(name) |
| if value is None: |
| continue |
| op, argType = self.opcodes[name] |
| if isinstance(argType, tuple): |
| l = len(argType) |
| assert len(value) == l, "value doesn't match arg type" |
| for i in range(l): |
| arg = argType[i] |
| v = value[i] |
| arghandler = getattr(self, "arg_" + arg) |
| data.append(arghandler(v)) |
| else: |
| arghandler = getattr(self, "arg_" + argType) |
| data.append(arghandler(value)) |
| data.append(op) |
| data = bytesjoin(data) |
| return data |
| |
| def toFile(self, file): |
| data = self.compile("toFile") |
| file.write(data) |
| |
| def arg_number(self, num): |
| if isinstance(num, list): |
| data = [encodeNumber(val) for val in num] |
| data.append(encodeNumber(1)) |
| data.append(bytechr(blendOp)) |
| datum = bytesjoin(data) |
| else: |
| datum = encodeNumber(num) |
| return datum |
| |
| def arg_SID(self, s): |
| return psCharStrings.encodeIntCFF(self.strings.getSID(s)) |
| |
| def arg_array(self, value): |
| data = [] |
| for num in value: |
| data.append(self.arg_number(num)) |
| return bytesjoin(data) |
| |
| def arg_delta(self, value): |
| if not value: |
| return b"" |
| val0 = value[0] |
| if isinstance(val0, list): |
| data = self.arg_delta_blend(value) |
| else: |
| out = [] |
| last = 0 |
| for v in value: |
| out.append(v - last) |
| last = v |
| data = [] |
| for num in out: |
| data.append(encodeNumber(num)) |
| return bytesjoin(data) |
| |
| |
| def arg_delta_blend(self, value): |
| """A delta list with blend lists has to be *all* blend lists. |
| |
| The value is a list is arranged as follows:: |
| |
| [ |
| [V0, d0..dn] |
| [V1, d0..dn] |
| ... |
| [Vm, d0..dn] |
| ] |
| |
| ``V`` is the absolute coordinate value from the default font, and ``d0-dn`` |
| are the delta values from the *n* regions. Each ``V`` is an absolute |
| coordinate from the default font. |
| |
| We want to return a list:: |
| |
| [ |
| [v0, v1..vm] |
| [d0..dn] |
| ... |
| [d0..dn] |
| numBlends |
| blendOp |
| ] |
| |
| where each ``v`` is relative to the previous default font value. |
| """ |
| numMasters = len(value[0]) |
| numBlends = len(value) |
| numStack = (numBlends * numMasters) + 1 |
| if numStack > self.maxBlendStack: |
| # Figure out the max number of value we can blend |
| # and divide this list up into chunks of that size. |
| |
| numBlendValues = int((self.maxBlendStack - 1) / numMasters) |
| out = [] |
| while True: |
| numVal = min(len(value), numBlendValues) |
| if numVal == 0: |
| break |
| valList = value[0:numVal] |
| out1 = self.arg_delta_blend(valList) |
| out.extend(out1) |
| value = value[numVal:] |
| else: |
| firstList = [0] * numBlends |
| deltaList = [None] * numBlends |
| i = 0 |
| prevVal = 0 |
| while i < numBlends: |
| # For PrivateDict BlueValues, the default font |
| # values are absolute, not relative. |
| # Must convert these back to relative coordinates |
| # befor writing to CFF2. |
| defaultValue = value[i][0] |
| firstList[i] = defaultValue - prevVal |
| prevVal = defaultValue |
| deltaList[i] = value[i][1:] |
| i += 1 |
| |
| relValueList = firstList |
| for blendList in deltaList: |
| relValueList.extend(blendList) |
| out = [encodeNumber(val) for val in relValueList] |
| out.append(encodeNumber(numBlends)) |
| out.append(bytechr(blendOp)) |
| return out |
| |
| |
| def encodeNumber(num): |
| if isinstance(num, float): |
| return psCharStrings.encodeFloat(num) |
| else: |
| return psCharStrings.encodeIntCFF(num) |
| |
| |
| class TopDictCompiler(DictCompiler): |
| |
| opcodes = buildOpcodeDict(topDictOperators) |
| |
| def getChildren(self, strings): |
| isCFF2 = self.isCFF2 |
| children = [] |
| if self.dictObj.cff2GetGlyphOrder is None: |
| if hasattr(self.dictObj, "charset") and self.dictObj.charset: |
| if hasattr(self.dictObj, "ROS"): # aka isCID |
| charsetCode = None |
| else: |
| charsetCode = getStdCharSet(self.dictObj.charset) |
| if charsetCode is None: |
| children.append(CharsetCompiler(strings, self.dictObj.charset, self)) |
| else: |
| self.rawDict["charset"] = charsetCode |
| if hasattr(self.dictObj, "Encoding") and self.dictObj.Encoding: |
| encoding = self.dictObj.Encoding |
| if not isinstance(encoding, str): |
| children.append(EncodingCompiler(strings, encoding, self)) |
| else: |
| if hasattr(self.dictObj, "VarStore"): |
| varStoreData = self.dictObj.VarStore |
| varStoreComp = VarStoreCompiler(varStoreData, self) |
| children.append(varStoreComp) |
| if hasattr(self.dictObj, "FDSelect"): |
| # I have not yet supported merging a ttx CFF-CID font, as there are |
| # interesting issues about merging the FDArrays. Here I assume that |
| # either the font was read from XML, and the FDSelect indices are all |
| # in the charstring data, or the FDSelect array is already fully defined. |
| fdSelect = self.dictObj.FDSelect |
| # probably read in from XML; assume fdIndex in CharString data |
| if len(fdSelect) == 0: |
| charStrings = self.dictObj.CharStrings |
| for name in self.dictObj.charset: |
| fdSelect.append(charStrings[name].fdSelectIndex) |
| fdSelectComp = FDSelectCompiler(fdSelect, self) |
| children.append(fdSelectComp) |
| if hasattr(self.dictObj, "CharStrings"): |
| items = [] |
| charStrings = self.dictObj.CharStrings |
| for name in self.dictObj.charset: |
| items.append(charStrings[name]) |
| charStringsComp = CharStringsCompiler( |
| items, strings, self, isCFF2=isCFF2) |
| children.append(charStringsComp) |
| if hasattr(self.dictObj, "FDArray"): |
| # I have not yet supported merging a ttx CFF-CID font, as there are |
| # interesting issues about merging the FDArrays. Here I assume that the |
| # FDArray info is correct and complete. |
| fdArrayIndexComp = self.dictObj.FDArray.getCompiler(strings, self) |
| children.append(fdArrayIndexComp) |
| children.extend(fdArrayIndexComp.getChildren(strings)) |
| if hasattr(self.dictObj, "Private"): |
| privComp = self.dictObj.Private.getCompiler(strings, self) |
| children.append(privComp) |
| children.extend(privComp.getChildren(strings)) |
| return children |
| |
| |
| class FontDictCompiler(DictCompiler): |
| opcodes = buildOpcodeDict(topDictOperators) |
| |
| def __init__(self, dictObj, strings, parent, isCFF2=None): |
| super(FontDictCompiler, self).__init__(dictObj, strings, parent, isCFF2=isCFF2) |
| # |
| # We now take some effort to detect if there were any key/value pairs |
| # supplied that were ignored in the FontDict context, and issue a warning |
| # for those cases. |
| # |
| ignoredNames = [] |
| dictObj = self.dictObj |
| for name in sorted(set(dictObj.converters) - set(dictObj.order)): |
| if name in dictObj.rawDict: |
| # The font was directly read from binary. In this |
| # case, we want to report *all* "useless" key/value |
| # pairs that are in the font, not just the ones that |
| # are different from the default. |
| ignoredNames.append(name) |
| else: |
| # The font was probably read from a TTX file. We only |
| # warn about keys whos value is not the default. The |
| # ones that have the default value will not be written |
| # to binary anyway. |
| default = dictObj.defaults.get(name) |
| if default is not None: |
| conv = dictObj.converters[name] |
| default = conv.read(dictObj, default) |
| if getattr(dictObj, name, None) != default: |
| ignoredNames.append(name) |
| if ignoredNames: |
| log.warning( |
| "Some CFF FDArray/FontDict keys were ignored upon compile: " + |
| " ".join(sorted(ignoredNames))) |
| |
| def getChildren(self, strings): |
| children = [] |
| if hasattr(self.dictObj, "Private"): |
| privComp = self.dictObj.Private.getCompiler(strings, self) |
| children.append(privComp) |
| children.extend(privComp.getChildren(strings)) |
| return children |
| |
| |
| class PrivateDictCompiler(DictCompiler): |
| |
| maxBlendStack = maxStackLimit |
| opcodes = buildOpcodeDict(privateDictOperators) |
| |
| def setPos(self, pos, endPos): |
| size = endPos - pos |
| self.parent.rawDict["Private"] = size, pos |
| self.pos = pos |
| |
| def getChildren(self, strings): |
| children = [] |
| if hasattr(self.dictObj, "Subrs"): |
| children.append(self.dictObj.Subrs.getCompiler(strings, self)) |
| return children |
| |
| |
| class BaseDict(object): |
| |
| def __init__(self, strings=None, file=None, offset=None, isCFF2=None): |
| assert (isCFF2 is None) == (file is None) |
| self.rawDict = {} |
| self.skipNames = [] |
| self.strings = strings |
| if file is None: |
| return |
| self._isCFF2 = isCFF2 |
| self.file = file |
| if offset is not None: |
| log.log(DEBUG, "loading %s at %s", self.__class__.__name__, offset) |
| self.offset = offset |
| |
| def decompile(self, data): |
| log.log(DEBUG, " length %s is %d", self.__class__.__name__, len(data)) |
| dec = self.decompilerClass(self.strings, self) |
| dec.decompile(data) |
| self.rawDict = dec.getDict() |
| self.postDecompile() |
| |
| def postDecompile(self): |
| pass |
| |
| def getCompiler(self, strings, parent, isCFF2=None): |
| return self.compilerClass(self, strings, parent, isCFF2=isCFF2) |
| |
| def __getattr__(self, name): |
| if name[:2] == name[-2:] == "__": |
| # to make deepcopy() and pickle.load() work, we need to signal with |
| # AttributeError that dunder methods like '__deepcopy__' or '__getstate__' |
| # aren't implemented. For more details, see: |
| # https://github.com/fonttools/fonttools/pull/1488 |
| raise AttributeError(name) |
| value = self.rawDict.get(name, None) |
| if value is None: |
| value = self.defaults.get(name) |
| if value is None: |
| raise AttributeError(name) |
| conv = self.converters[name] |
| value = conv.read(self, value) |
| setattr(self, name, value) |
| return value |
| |
| def toXML(self, xmlWriter): |
| for name in self.order: |
| if name in self.skipNames: |
| continue |
| value = getattr(self, name, None) |
| # XXX For "charset" we never skip calling xmlWrite even if the |
| # value is None, so we always write the following XML comment: |
| # |
| # <!-- charset is dumped separately as the 'GlyphOrder' element --> |
| # |
| # Charset is None when 'CFF ' table is imported from XML into an |
| # empty TTFont(). By writing this comment all the time, we obtain |
| # the same XML output whether roundtripping XML-to-XML or |
| # dumping binary-to-XML |
| if value is None and name != "charset": |
| continue |
| conv = self.converters[name] |
| conv.xmlWrite(xmlWriter, name, value) |
| ignoredNames = set(self.rawDict) - set(self.order) |
| if ignoredNames: |
| xmlWriter.comment( |
| "some keys were ignored: %s" % " ".join(sorted(ignoredNames))) |
| xmlWriter.newline() |
| |
| def fromXML(self, name, attrs, content): |
| conv = self.converters[name] |
| value = conv.xmlRead(name, attrs, content, self) |
| setattr(self, name, value) |
| |
| |
| class TopDict(BaseDict): |
| """The ``TopDict`` represents the top-level dictionary holding font |
| information. CFF2 tables contain a restricted set of top-level entries |
| as described `here <https://docs.microsoft.com/en-us/typography/opentype/spec/cff2#7-top-dict-data>`_, |
| but CFF tables may contain a wider range of information. This information |
| can be accessed through attributes or through the dictionary returned |
| through the ``rawDict`` property: |
| |
| .. code:: python |
| |
| font = tt["CFF "].cff[0] |
| font.FamilyName |
| # 'Linux Libertine O' |
| font.rawDict["FamilyName"] |
| # 'Linux Libertine O' |
| |
| More information is available in the CFF file's private dictionary, accessed |
| via the ``Private`` property: |
| |
| .. code:: python |
| |
| tt["CFF "].cff[0].Private.BlueValues |
| # [-15, 0, 515, 515, 666, 666] |
| |
| """ |
| |
| defaults = buildDefaults(topDictOperators) |
| converters = buildConverters(topDictOperators) |
| compilerClass = TopDictCompiler |
| order = buildOrder(topDictOperators) |
| decompilerClass = TopDictDecompiler |
| |
| def __init__(self, strings=None, file=None, offset=None, |
| GlobalSubrs=None, cff2GetGlyphOrder=None, isCFF2=None): |
| super(TopDict, self).__init__(strings, file, offset, isCFF2=isCFF2) |
| self.cff2GetGlyphOrder = cff2GetGlyphOrder |
| self.GlobalSubrs = GlobalSubrs |
| if isCFF2: |
| self.defaults = buildDefaults(topDictOperators2) |
| self.charset = cff2GetGlyphOrder() |
| self.order = buildOrder(topDictOperators2) |
| else: |
| self.defaults = buildDefaults(topDictOperators) |
| self.order = buildOrder(topDictOperators) |
| |
| def getGlyphOrder(self): |
| """Returns a list of glyph names in the CFF font.""" |
| return self.charset |
| |
| def postDecompile(self): |
| offset = self.rawDict.get("CharStrings") |
| if offset is None: |
| return |
| # get the number of glyphs beforehand. |
| self.file.seek(offset) |
| if self._isCFF2: |
| self.numGlyphs = readCard32(self.file) |
| else: |
| self.numGlyphs = readCard16(self.file) |
| |
| def toXML(self, xmlWriter): |
| if hasattr(self, "CharStrings"): |
| self.decompileAllCharStrings() |
| if hasattr(self, "ROS"): |
| self.skipNames = ['Encoding'] |
| if not hasattr(self, "ROS") or not hasattr(self, "CharStrings"): |
| # these values have default values, but I only want them to show up |
| # in CID fonts. |
| self.skipNames = [ |
| 'CIDFontVersion', 'CIDFontRevision', 'CIDFontType', 'CIDCount'] |
| BaseDict.toXML(self, xmlWriter) |
| |
| def decompileAllCharStrings(self): |
| # Make sure that all the Private Dicts have been instantiated. |
| for i, charString in enumerate(self.CharStrings.values()): |
| try: |
| charString.decompile() |
| except: |
| log.error("Error in charstring %s", i) |
| raise |
| |
| def recalcFontBBox(self): |
| fontBBox = None |
| for charString in self.CharStrings.values(): |
| bounds = charString.calcBounds(self.CharStrings) |
| if bounds is not None: |
| if fontBBox is not None: |
| fontBBox = unionRect(fontBBox, bounds) |
| else: |
| fontBBox = bounds |
| |
| if fontBBox is None: |
| self.FontBBox = self.defaults['FontBBox'][:] |
| else: |
| self.FontBBox = list(intRect(fontBBox)) |
| |
| |
| class FontDict(BaseDict): |
| # |
| # Since fonttools used to pass a lot of fields that are not relevant in the FDArray |
| # FontDict, there are 'ttx' files in the wild that contain all these. These got in |
| # the ttx files because fonttools writes explicit values for all the TopDict default |
| # values. These are not actually illegal in the context of an FDArray FontDict - you |
| # can legally, per spec, put any arbitrary key/value pair in a FontDict - but are |
| # useless since current major company CFF interpreters ignore anything but the set |
| # listed in this file. So, we just silently skip them. An exception is Weight: this |
| # is not used by any interpreter, but some foundries have asked that this be |
| # supported in FDArray FontDicts just to preserve information about the design when |
| # the font is being inspected. |
| # |
| # On top of that, there are fonts out there that contain such useless FontDict values. |
| # |
| # By subclassing TopDict, we *allow* all key/values from TopDict, both when reading |
| # from binary or when reading from XML, but by overriding `order` with a limited |
| # list of names, we ensure that only the useful names ever get exported to XML and |
| # ever get compiled into the binary font. |
| # |
| # We override compilerClass so we can warn about "useless" key/value pairs, either |
| # from the original binary font or from TTX input. |
| # |
| # See: |
| # - https://github.com/fonttools/fonttools/issues/740 |
| # - https://github.com/fonttools/fonttools/issues/601 |
| # - https://github.com/adobe-type-tools/afdko/issues/137 |
| # |
| defaults = {} |
| converters = buildConverters(topDictOperators) |
| compilerClass = FontDictCompiler |
| orderCFF = ['FontName', 'FontMatrix', 'Weight', 'Private'] |
| orderCFF2 = ['Private'] |
| decompilerClass = TopDictDecompiler |
| |
| def __init__(self, strings=None, file=None, offset=None, |
| GlobalSubrs=None, isCFF2=None, vstore=None): |
| super(FontDict, self).__init__(strings, file, offset, isCFF2=isCFF2) |
| self.vstore = vstore |
| self.setCFF2(isCFF2) |
| |
| def setCFF2(self, isCFF2): |
| # isCFF2 may be None. |
| if isCFF2: |
| self.order = self.orderCFF2 |
| self._isCFF2 = True |
| else: |
| self.order = self.orderCFF |
| self._isCFF2 = False |
| |
| |
| class PrivateDict(BaseDict): |
| defaults = buildDefaults(privateDictOperators) |
| converters = buildConverters(privateDictOperators) |
| order = buildOrder(privateDictOperators) |
| decompilerClass = PrivateDictDecompiler |
| compilerClass = PrivateDictCompiler |
| |
| def __init__(self, strings=None, file=None, offset=None, isCFF2=None, |
| vstore=None): |
| super(PrivateDict, self).__init__(strings, file, offset, isCFF2=isCFF2) |
| self.vstore = vstore |
| if isCFF2: |
| self.defaults = buildDefaults(privateDictOperators2) |
| self.order = buildOrder(privateDictOperators2) |
| # Provide dummy values. This avoids needing to provide |
| # an isCFF2 state in a lot of places. |
| self.nominalWidthX = self.defaultWidthX = None |
| else: |
| self.defaults = buildDefaults(privateDictOperators) |
| self.order = buildOrder(privateDictOperators) |
| |
| @property |
| def in_cff2(self): |
| return self._isCFF2 |
| |
| def getNumRegions(self, vi=None): # called from misc/psCharStrings.py |
| # if getNumRegions is being called, we can assume that VarStore exists. |
| if vi is None: |
| if hasattr(self, 'vsindex'): |
| vi = self.vsindex |
| else: |
| vi = 0 |
| numRegions = self.vstore.getNumRegions(vi) |
| return numRegions |
| |
| |
| class IndexedStrings(object): |
| |
| """SID -> string mapping.""" |
| |
| def __init__(self, file=None): |
| if file is None: |
| strings = [] |
| else: |
| strings = [ |
| tostr(s, encoding="latin1") |
| for s in Index(file, isCFF2=False) |
| ] |
| self.strings = strings |
| |
| def getCompiler(self): |
| return IndexedStringsCompiler(self, None, self, isCFF2=False) |
| |
| def __len__(self): |
| return len(self.strings) |
| |
| def __getitem__(self, SID): |
| if SID < cffStandardStringCount: |
| return cffStandardStrings[SID] |
| else: |
| return self.strings[SID - cffStandardStringCount] |
| |
| def getSID(self, s): |
| if not hasattr(self, "stringMapping"): |
| self.buildStringMapping() |
| s = tostr(s, encoding="latin1") |
| if s in cffStandardStringMapping: |
| SID = cffStandardStringMapping[s] |
| elif s in self.stringMapping: |
| SID = self.stringMapping[s] |
| else: |
| SID = len(self.strings) + cffStandardStringCount |
| self.strings.append(s) |
| self.stringMapping[s] = SID |
| return SID |
| |
| def getStrings(self): |
| return self.strings |
| |
| def buildStringMapping(self): |
| self.stringMapping = {} |
| for index in range(len(self.strings)): |
| self.stringMapping[self.strings[index]] = index + cffStandardStringCount |
| |
| |
| # The 391 Standard Strings as used in the CFF format. |
| # from Adobe Technical None #5176, version 1.0, 18 March 1998 |
| |
| cffStandardStrings = ['.notdef', 'space', 'exclam', 'quotedbl', 'numbersign', |
| 'dollar', 'percent', 'ampersand', 'quoteright', 'parenleft', 'parenright', |
| 'asterisk', 'plus', 'comma', 'hyphen', 'period', 'slash', 'zero', 'one', |
| 'two', 'three', 'four', 'five', 'six', 'seven', 'eight', 'nine', 'colon', |
| 'semicolon', 'less', 'equal', 'greater', 'question', 'at', 'A', 'B', 'C', |
| 'D', 'E', 'F', 'G', 'H', 'I', 'J', 'K', 'L', 'M', 'N', 'O', 'P', 'Q', 'R', |
| 'S', 'T', 'U', 'V', 'W', 'X', 'Y', 'Z', 'bracketleft', 'backslash', |
| 'bracketright', 'asciicircum', 'underscore', 'quoteleft', 'a', 'b', 'c', |
| 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k', 'l', 'm', 'n', 'o', 'p', 'q', 'r', |
| 's', 't', 'u', 'v', 'w', 'x', 'y', 'z', 'braceleft', 'bar', 'braceright', |
| 'asciitilde', 'exclamdown', 'cent', 'sterling', 'fraction', 'yen', 'florin', |
| 'section', 'currency', 'quotesingle', 'quotedblleft', 'guillemotleft', |
| 'guilsinglleft', 'guilsinglright', 'fi', 'fl', 'endash', 'dagger', |
| 'daggerdbl', 'periodcentered', 'paragraph', 'bullet', 'quotesinglbase', |
| 'quotedblbase', 'quotedblright', 'guillemotright', 'ellipsis', 'perthousand', |
| 'questiondown', 'grave', 'acute', 'circumflex', 'tilde', 'macron', 'breve', |
| 'dotaccent', 'dieresis', 'ring', 'cedilla', 'hungarumlaut', 'ogonek', 'caron', |
| 'emdash', 'AE', 'ordfeminine', 'Lslash', 'Oslash', 'OE', 'ordmasculine', 'ae', |
| 'dotlessi', 'lslash', 'oslash', 'oe', 'germandbls', 'onesuperior', |
| 'logicalnot', 'mu', 'trademark', 'Eth', 'onehalf', 'plusminus', 'Thorn', |
| 'onequarter', 'divide', 'brokenbar', 'degree', 'thorn', 'threequarters', |
| 'twosuperior', 'registered', 'minus', 'eth', 'multiply', 'threesuperior', |
| 'copyright', 'Aacute', 'Acircumflex', 'Adieresis', 'Agrave', 'Aring', |
| 'Atilde', 'Ccedilla', 'Eacute', 'Ecircumflex', 'Edieresis', 'Egrave', |
| 'Iacute', 'Icircumflex', 'Idieresis', 'Igrave', 'Ntilde', 'Oacute', |
| 'Ocircumflex', 'Odieresis', 'Ograve', 'Otilde', 'Scaron', 'Uacute', |
| 'Ucircumflex', 'Udieresis', 'Ugrave', 'Yacute', 'Ydieresis', 'Zcaron', |
| 'aacute', 'acircumflex', 'adieresis', 'agrave', 'aring', 'atilde', 'ccedilla', |
| 'eacute', 'ecircumflex', 'edieresis', 'egrave', 'iacute', 'icircumflex', |
| 'idieresis', 'igrave', 'ntilde', 'oacute', 'ocircumflex', 'odieresis', |
| 'ograve', 'otilde', 'scaron', 'uacute', 'ucircumflex', 'udieresis', 'ugrave', |
| 'yacute', 'ydieresis', 'zcaron', 'exclamsmall', 'Hungarumlautsmall', |
| 'dollaroldstyle', 'dollarsuperior', 'ampersandsmall', 'Acutesmall', |
| 'parenleftsuperior', 'parenrightsuperior', 'twodotenleader', 'onedotenleader', |
| 'zerooldstyle', 'oneoldstyle', 'twooldstyle', 'threeoldstyle', 'fouroldstyle', |
| 'fiveoldstyle', 'sixoldstyle', 'sevenoldstyle', 'eightoldstyle', |
| 'nineoldstyle', 'commasuperior', 'threequartersemdash', 'periodsuperior', |
| 'questionsmall', 'asuperior', 'bsuperior', 'centsuperior', 'dsuperior', |
| 'esuperior', 'isuperior', 'lsuperior', 'msuperior', 'nsuperior', 'osuperior', |
| 'rsuperior', 'ssuperior', 'tsuperior', 'ff', 'ffi', 'ffl', 'parenleftinferior', |
| 'parenrightinferior', 'Circumflexsmall', 'hyphensuperior', 'Gravesmall', |
| 'Asmall', 'Bsmall', 'Csmall', 'Dsmall', 'Esmall', 'Fsmall', 'Gsmall', 'Hsmall', |
| 'Ismall', 'Jsmall', 'Ksmall', 'Lsmall', 'Msmall', 'Nsmall', 'Osmall', 'Psmall', |
| 'Qsmall', 'Rsmall', 'Ssmall', 'Tsmall', 'Usmall', 'Vsmall', 'Wsmall', 'Xsmall', |
| 'Ysmall', 'Zsmall', 'colonmonetary', 'onefitted', 'rupiah', 'Tildesmall', |
| 'exclamdownsmall', 'centoldstyle', 'Lslashsmall', 'Scaronsmall', 'Zcaronsmall', |
| 'Dieresissmall', 'Brevesmall', 'Caronsmall', 'Dotaccentsmall', 'Macronsmall', |
| 'figuredash', 'hypheninferior', 'Ogoneksmall', 'Ringsmall', 'Cedillasmall', |
| 'questiondownsmall', 'oneeighth', 'threeeighths', 'fiveeighths', 'seveneighths', |
| 'onethird', 'twothirds', 'zerosuperior', 'foursuperior', 'fivesuperior', |
| 'sixsuperior', 'sevensuperior', 'eightsuperior', 'ninesuperior', 'zeroinferior', |
| 'oneinferior', 'twoinferior', 'threeinferior', 'fourinferior', 'fiveinferior', |
| 'sixinferior', 'seveninferior', 'eightinferior', 'nineinferior', 'centinferior', |
| 'dollarinferior', 'periodinferior', 'commainferior', 'Agravesmall', |
| 'Aacutesmall', 'Acircumflexsmall', 'Atildesmall', 'Adieresissmall', 'Aringsmall', |
| 'AEsmall', 'Ccedillasmall', 'Egravesmall', 'Eacutesmall', 'Ecircumflexsmall', |
| 'Edieresissmall', 'Igravesmall', 'Iacutesmall', 'Icircumflexsmall', |
| 'Idieresissmall', 'Ethsmall', 'Ntildesmall', 'Ogravesmall', 'Oacutesmall', |
| 'Ocircumflexsmall', 'Otildesmall', 'Odieresissmall', 'OEsmall', 'Oslashsmall', |
| 'Ugravesmall', 'Uacutesmall', 'Ucircumflexsmall', 'Udieresissmall', |
| 'Yacutesmall', 'Thornsmall', 'Ydieresissmall', '001.000', '001.001', '001.002', |
| '001.003', 'Black', 'Bold', 'Book', 'Light', 'Medium', 'Regular', 'Roman', |
| 'Semibold' |
| ] |
| |
| cffStandardStringCount = 391 |
| assert len(cffStandardStrings) == cffStandardStringCount |
| # build reverse mapping |
| cffStandardStringMapping = {} |
| for _i in range(cffStandardStringCount): |
| cffStandardStringMapping[cffStandardStrings[_i]] = _i |
| |
| cffISOAdobeStrings = [".notdef", "space", "exclam", "quotedbl", "numbersign", |
| "dollar", "percent", "ampersand", "quoteright", "parenleft", "parenright", |
| "asterisk", "plus", "comma", "hyphen", "period", "slash", "zero", "one", "two", |
| "three", "four", "five", "six", "seven", "eight", "nine", "colon", "semicolon", |
| "less", "equal", "greater", "question", "at", "A", "B", "C", "D", "E", "F", "G", |
| "H", "I", "J", "K", "L", "M", "N", "O", "P", "Q", "R", "S", "T", "U", "V", "W", |
| "X", "Y", "Z", "bracketleft", "backslash", "bracketright", "asciicircum", |
| "underscore", "quoteleft", "a", "b", "c", "d", "e", "f", "g", "h", "i", "j", |
| "k", "l", "m", "n", "o", "p", "q", "r", "s", "t", "u", "v", "w", "x", "y", "z", |
| "braceleft", "bar", "braceright", "asciitilde", "exclamdown", "cent", |
| "sterling", "fraction", "yen", "florin", "section", "currency", "quotesingle", |
| "quotedblleft", "guillemotleft", "guilsinglleft", "guilsinglright", "fi", "fl", |
| "endash", "dagger", "daggerdbl", "periodcentered", "paragraph", "bullet", |
| "quotesinglbase", "quotedblbase", "quotedblright", "guillemotright", "ellipsis", |
| "perthousand", "questiondown", "grave", "acute", "circumflex", "tilde", |
| "macron", "breve", "dotaccent", "dieresis", "ring", "cedilla", "hungarumlaut", |
| "ogonek", "caron", "emdash", "AE", "ordfeminine", "Lslash", "Oslash", "OE", |
| "ordmasculine", "ae", "dotlessi", "lslash", "oslash", "oe", "germandbls", |
| "onesuperior", "logicalnot", "mu", "trademark", "Eth", "onehalf", "plusminus", |
| "Thorn", "onequarter", "divide", "brokenbar", "degree", "thorn", |
| "threequarters", "twosuperior", "registered", "minus", "eth", "multiply", |
| "threesuperior", "copyright", "Aacute", "Acircumflex", "Adieresis", "Agrave", |
| "Aring", "Atilde", "Ccedilla", "Eacute", "Ecircumflex", "Edieresis", "Egrave", |
| "Iacute", "Icircumflex", "Idieresis", "Igrave", "Ntilde", "Oacute", |
| "Ocircumflex", "Odieresis", "Ograve", "Otilde", "Scaron", "Uacute", |
| "Ucircumflex", "Udieresis", "Ugrave", "Yacute", "Ydieresis", "Zcaron", "aacute", |
| "acircumflex", "adieresis", "agrave", "aring", "atilde", "ccedilla", "eacute", |
| "ecircumflex", "edieresis", "egrave", "iacute", "icircumflex", "idieresis", |
| "igrave", "ntilde", "oacute", "ocircumflex", "odieresis", "ograve", "otilde", |
| "scaron", "uacute", "ucircumflex", "udieresis", "ugrave", "yacute", "ydieresis", |
| "zcaron"] |
| |
| cffISOAdobeStringCount = 229 |
| assert len(cffISOAdobeStrings) == cffISOAdobeStringCount |
| |
| cffIExpertStrings = [".notdef", "space", "exclamsmall", "Hungarumlautsmall", |
| "dollaroldstyle", "dollarsuperior", "ampersandsmall", "Acutesmall", |
| "parenleftsuperior", "parenrightsuperior", "twodotenleader", "onedotenleader", |
| "comma", "hyphen", "period", "fraction", "zerooldstyle", "oneoldstyle", |
| "twooldstyle", "threeoldstyle", "fouroldstyle", "fiveoldstyle", "sixoldstyle", |
| "sevenoldstyle", "eightoldstyle", "nineoldstyle", "colon", "semicolon", |
| "commasuperior", "threequartersemdash", "periodsuperior", "questionsmall", |
| "asuperior", "bsuperior", "centsuperior", "dsuperior", "esuperior", "isuperior", |
| "lsuperior", "msuperior", "nsuperior", "osuperior", "rsuperior", "ssuperior", |
| "tsuperior", "ff", "fi", "fl", "ffi", "ffl", "parenleftinferior", |
| "parenrightinferior", "Circumflexsmall", "hyphensuperior", "Gravesmall", |
| "Asmall", "Bsmall", "Csmall", "Dsmall", "Esmall", "Fsmall", "Gsmall", "Hsmall", |
| "Ismall", "Jsmall", "Ksmall", "Lsmall", "Msmall", "Nsmall", "Osmall", "Psmall", |
| "Qsmall", "Rsmall", "Ssmall", "Tsmall", "Usmall", "Vsmall", "Wsmall", "Xsmall", |
| "Ysmall", "Zsmall", "colonmonetary", "onefitted", "rupiah", "Tildesmall", |
| "exclamdownsmall", "centoldstyle", "Lslashsmall", "Scaronsmall", "Zcaronsmall", |
| "Dieresissmall", "Brevesmall", "Caronsmall", "Dotaccentsmall", "Macronsmall", |
| "figuredash", "hypheninferior", "Ogoneksmall", "Ringsmall", "Cedillasmall", |
| "onequarter", "onehalf", "threequarters", "questiondownsmall", "oneeighth", |
| "threeeighths", "fiveeighths", "seveneighths", "onethird", "twothirds", |
| "zerosuperior", "onesuperior", "twosuperior", "threesuperior", "foursuperior", |
| "fivesuperior", "sixsuperior", "sevensuperior", "eightsuperior", "ninesuperior", |
| "zeroinferior", "oneinferior", "twoinferior", "threeinferior", "fourinferior", |
| "fiveinferior", "sixinferior", "seveninferior", "eightinferior", "nineinferior", |
| "centinferior", "dollarinferior", "periodinferior", "commainferior", |
| "Agravesmall", "Aacutesmall", "Acircumflexsmall", "Atildesmall", |
| "Adieresissmall", "Aringsmall", "AEsmall", "Ccedillasmall", "Egravesmall", |
| "Eacutesmall", "Ecircumflexsmall", "Edieresissmall", "Igravesmall", |
| "Iacutesmall", "Icircumflexsmall", "Idieresissmall", "Ethsmall", "Ntildesmall", |
| "Ogravesmall", "Oacutesmall", "Ocircumflexsmall", "Otildesmall", |
| "Odieresissmall", "OEsmall", "Oslashsmall", "Ugravesmall", "Uacutesmall", |
| "Ucircumflexsmall", "Udieresissmall", "Yacutesmall", "Thornsmall", |
| "Ydieresissmall"] |
| |
| cffExpertStringCount = 166 |
| assert len(cffIExpertStrings) == cffExpertStringCount |
| |
| cffExpertSubsetStrings = [".notdef", "space", "dollaroldstyle", |
| "dollarsuperior", "parenleftsuperior", "parenrightsuperior", "twodotenleader", |
| "onedotenleader", "comma", "hyphen", "period", "fraction", "zerooldstyle", |
| "oneoldstyle", "twooldstyle", "threeoldstyle", "fouroldstyle", "fiveoldstyle", |
| "sixoldstyle", "sevenoldstyle", "eightoldstyle", "nineoldstyle", "colon", |
| "semicolon", "commasuperior", "threequartersemdash", "periodsuperior", |
| "asuperior", "bsuperior", "centsuperior", "dsuperior", "esuperior", "isuperior", |
| "lsuperior", "msuperior", "nsuperior", "osuperior", "rsuperior", "ssuperior", |
| "tsuperior", "ff", "fi", "fl", "ffi", "ffl", "parenleftinferior", |
| "parenrightinferior", "hyphensuperior", "colonmonetary", "onefitted", "rupiah", |
| "centoldstyle", "figuredash", "hypheninferior", "onequarter", "onehalf", |
| "threequarters", "oneeighth", "threeeighths", "fiveeighths", "seveneighths", |
| "onethird", "twothirds", "zerosuperior", "onesuperior", "twosuperior", |
| "threesuperior", "foursuperior", "fivesuperior", "sixsuperior", "sevensuperior", |
| "eightsuperior", "ninesuperior", "zeroinferior", "oneinferior", "twoinferior", |
| "threeinferior", "fourinferior", "fiveinferior", "sixinferior", "seveninferior", |
| "eightinferior", "nineinferior", "centinferior", "dollarinferior", |
| "periodinferior", "commainferior"] |
| |
| cffExpertSubsetStringCount = 87 |
| assert len(cffExpertSubsetStrings) == cffExpertSubsetStringCount |