| | """ |
| | A PDF Matplotlib backend. |
| | |
| | Author: Jouni K Seppänen <jks@iki.fi> and others. |
| | """ |
| |
|
| | import codecs |
| | from datetime import timezone |
| | from datetime import datetime |
| | from enum import Enum |
| | from functools import total_ordering |
| | from io import BytesIO |
| | import itertools |
| | import logging |
| | import math |
| | import os |
| | import string |
| | import struct |
| | import sys |
| | import time |
| | import types |
| | import warnings |
| | import zlib |
| |
|
| | import numpy as np |
| | from PIL import Image |
| |
|
| | import matplotlib as mpl |
| | from matplotlib import _api, _text_helpers, _type1font, cbook, dviread |
| | from matplotlib._pylab_helpers import Gcf |
| | from matplotlib.backend_bases import ( |
| | _Backend, FigureCanvasBase, FigureManagerBase, GraphicsContextBase, |
| | RendererBase) |
| | from matplotlib.backends.backend_mixed import MixedModeRenderer |
| | from matplotlib.figure import Figure |
| | from matplotlib.font_manager import get_font, fontManager as _fontManager |
| | from matplotlib._afm import AFM |
| | from matplotlib.ft2font import (FIXED_WIDTH, ITALIC, LOAD_NO_SCALE, |
| | LOAD_NO_HINTING, KERNING_UNFITTED, FT2Font) |
| | from matplotlib.transforms import Affine2D, BboxBase |
| | from matplotlib.path import Path |
| | from matplotlib.dates import UTC |
| | from matplotlib import _path |
| | from . import _backend_pdf_ps |
| |
|
| | _log = logging.getLogger(__name__) |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| |
|
| | @_api.deprecated("3.6", alternative="a vendored copy of _fill") |
| | def fill(strings, linelen=75): |
| | return _fill(strings, linelen=linelen) |
| |
|
| |
|
| | def _fill(strings, linelen=75): |
| | """ |
| | Make one string from sequence of strings, with whitespace in between. |
| | |
| | The whitespace is chosen to form lines of at most *linelen* characters, |
| | if possible. |
| | """ |
| | currpos = 0 |
| | lasti = 0 |
| | result = [] |
| | for i, s in enumerate(strings): |
| | length = len(s) |
| | if currpos + length < linelen: |
| | currpos += length + 1 |
| | else: |
| | result.append(b' '.join(strings[lasti:i])) |
| | lasti = i |
| | currpos = length |
| | result.append(b' '.join(strings[lasti:])) |
| | return b'\n'.join(result) |
| |
|
| |
|
| | def _create_pdf_info_dict(backend, metadata): |
| | """ |
| | Create a PDF infoDict based on user-supplied metadata. |
| | |
| | A default ``Creator``, ``Producer``, and ``CreationDate`` are added, though |
| | the user metadata may override it. The date may be the current time, or a |
| | time set by the ``SOURCE_DATE_EPOCH`` environment variable. |
| | |
| | Metadata is verified to have the correct keys and their expected types. Any |
| | unknown keys/types will raise a warning. |
| | |
| | Parameters |
| | ---------- |
| | backend : str |
| | The name of the backend to use in the Producer value. |
| | |
| | metadata : dict[str, Union[str, datetime, Name]] |
| | A dictionary of metadata supplied by the user with information |
| | following the PDF specification, also defined in |
| | `~.backend_pdf.PdfPages` below. |
| | |
| | If any value is *None*, then the key will be removed. This can be used |
| | to remove any pre-defined values. |
| | |
| | Returns |
| | ------- |
| | dict[str, Union[str, datetime, Name]] |
| | A validated dictionary of metadata. |
| | """ |
| |
|
| | |
| | |
| | source_date_epoch = os.getenv("SOURCE_DATE_EPOCH") |
| | if source_date_epoch: |
| | source_date = datetime.fromtimestamp(int(source_date_epoch), timezone.utc) |
| | source_date = source_date.replace(tzinfo=UTC) |
| | else: |
| | source_date = datetime.today() |
| |
|
| | info = { |
| | 'Creator': f'Matplotlib v{mpl.__version__}, https://matplotlib.org', |
| | 'Producer': f'Matplotlib {backend} backend v{mpl.__version__}', |
| | 'CreationDate': source_date, |
| | **metadata |
| | } |
| | info = {k: v for (k, v) in info.items() if v is not None} |
| |
|
| | def is_string_like(x): |
| | return isinstance(x, str) |
| | is_string_like.text_for_warning = "an instance of str" |
| |
|
| | def is_date(x): |
| | return isinstance(x, datetime) |
| | is_date.text_for_warning = "an instance of datetime.datetime" |
| |
|
| | def check_trapped(x): |
| | if isinstance(x, Name): |
| | return x.name in (b'True', b'False', b'Unknown') |
| | else: |
| | return x in ('True', 'False', 'Unknown') |
| | check_trapped.text_for_warning = 'one of {"True", "False", "Unknown"}' |
| |
|
| | keywords = { |
| | 'Title': is_string_like, |
| | 'Author': is_string_like, |
| | 'Subject': is_string_like, |
| | 'Keywords': is_string_like, |
| | 'Creator': is_string_like, |
| | 'Producer': is_string_like, |
| | 'CreationDate': is_date, |
| | 'ModDate': is_date, |
| | 'Trapped': check_trapped, |
| | } |
| | for k in info: |
| | if k not in keywords: |
| | _api.warn_external(f'Unknown infodict keyword: {k!r}. ' |
| | f'Must be one of {set(keywords)!r}.') |
| | elif not keywords[k](info[k]): |
| | _api.warn_external(f'Bad value for infodict keyword {k}. ' |
| | f'Got {info[k]!r} which is not ' |
| | f'{keywords[k].text_for_warning}.') |
| | if 'Trapped' in info: |
| | info['Trapped'] = Name(info['Trapped']) |
| |
|
| | return info |
| |
|
| |
|
| | def _datetime_to_pdf(d): |
| | """ |
| | Convert a datetime to a PDF string representing it. |
| | |
| | Used for PDF and PGF. |
| | """ |
| | r = d.strftime('D:%Y%m%d%H%M%S') |
| | z = d.utcoffset() |
| | if z is not None: |
| | z = z.seconds |
| | else: |
| | if time.daylight: |
| | z = time.altzone |
| | else: |
| | z = time.timezone |
| | if z == 0: |
| | r += 'Z' |
| | elif z < 0: |
| | r += "+%02d'%02d'" % ((-z) // 3600, (-z) % 3600) |
| | else: |
| | r += "-%02d'%02d'" % (z // 3600, z % 3600) |
| | return r |
| |
|
| |
|
| | def _calculate_quad_point_coordinates(x, y, width, height, angle=0): |
| | """ |
| | Calculate the coordinates of rectangle when rotated by angle around x, y |
| | """ |
| |
|
| | angle = math.radians(-angle) |
| | sin_angle = math.sin(angle) |
| | cos_angle = math.cos(angle) |
| | a = x + height * sin_angle |
| | b = y + height * cos_angle |
| | c = x + width * cos_angle + height * sin_angle |
| | d = y - width * sin_angle + height * cos_angle |
| | e = x + width * cos_angle |
| | f = y - width * sin_angle |
| | return ((x, y), (e, f), (c, d), (a, b)) |
| |
|
| |
|
| | def _get_coordinates_of_block(x, y, width, height, angle=0): |
| | """ |
| | Get the coordinates of rotated rectangle and rectangle that covers the |
| | rotated rectangle. |
| | """ |
| |
|
| | vertices = _calculate_quad_point_coordinates(x, y, width, |
| | height, angle) |
| |
|
| | |
| | |
| | |
| | |
| | |
| |
|
| | pad = 0.00001 if angle % 90 else 0 |
| | min_x = min(v[0] for v in vertices) - pad |
| | min_y = min(v[1] for v in vertices) - pad |
| | max_x = max(v[0] for v in vertices) + pad |
| | max_y = max(v[1] for v in vertices) + pad |
| | return (tuple(itertools.chain.from_iterable(vertices)), |
| | (min_x, min_y, max_x, max_y)) |
| |
|
| |
|
| | def _get_link_annotation(gc, x, y, width, height, angle=0): |
| | """ |
| | Create a link annotation object for embedding URLs. |
| | """ |
| | quadpoints, rect = _get_coordinates_of_block(x, y, width, height, angle) |
| | link_annotation = { |
| | 'Type': Name('Annot'), |
| | 'Subtype': Name('Link'), |
| | 'Rect': rect, |
| | 'Border': [0, 0, 0], |
| | 'A': { |
| | 'S': Name('URI'), |
| | 'URI': gc.get_url(), |
| | }, |
| | } |
| | if angle % 90: |
| | |
| | link_annotation['QuadPoints'] = quadpoints |
| | return link_annotation |
| |
|
| |
|
| | |
| | |
| | |
| | |
| | |
| | _str_escapes = str.maketrans({ |
| | '\\': '\\\\', '(': '\\(', ')': '\\)', '\n': '\\n', '\r': '\\r'}) |
| |
|
| |
|
| | def pdfRepr(obj): |
| | """Map Python objects to PDF syntax.""" |
| |
|
| | |
| | if hasattr(obj, 'pdfRepr'): |
| | return obj.pdfRepr() |
| |
|
| | |
| | |
| | |
| | elif isinstance(obj, (float, np.floating)): |
| | if not np.isfinite(obj): |
| | raise ValueError("Can only output finite numbers in PDF") |
| | r = b"%.10f" % obj |
| | return r.rstrip(b'0').rstrip(b'.') |
| |
|
| | |
| | |
| | elif isinstance(obj, bool): |
| | return [b'false', b'true'][obj] |
| |
|
| | |
| | elif isinstance(obj, (int, np.integer)): |
| | return b"%d" % obj |
| |
|
| | |
| | elif isinstance(obj, str): |
| | return pdfRepr(obj.encode('ascii') if obj.isascii() |
| | else codecs.BOM_UTF16_BE + obj.encode('UTF-16BE')) |
| |
|
| | |
| | |
| | |
| | |
| | |
| | elif isinstance(obj, bytes): |
| | return ( |
| | b'(' + |
| | obj.decode('latin-1').translate(_str_escapes).encode('latin-1') |
| | + b')') |
| |
|
| | |
| | |
| | |
| | |
| | elif isinstance(obj, dict): |
| | return _fill([ |
| | b"<<", |
| | *[Name(k).pdfRepr() + b" " + pdfRepr(v) for k, v in obj.items()], |
| | b">>", |
| | ]) |
| |
|
| | |
| | elif isinstance(obj, (list, tuple)): |
| | return _fill([b"[", *[pdfRepr(val) for val in obj], b"]"]) |
| |
|
| | |
| | elif obj is None: |
| | return b'null' |
| |
|
| | |
| | elif isinstance(obj, datetime): |
| | return pdfRepr(_datetime_to_pdf(obj)) |
| |
|
| | |
| | elif isinstance(obj, BboxBase): |
| | return _fill([pdfRepr(val) for val in obj.bounds]) |
| |
|
| | else: |
| | raise TypeError("Don't know a PDF representation for {} objects" |
| | .format(type(obj))) |
| |
|
| |
|
| | def _font_supports_glyph(fonttype, glyph): |
| | """ |
| | Returns True if the font is able to provide codepoint *glyph* in a PDF. |
| | |
| | For a Type 3 font, this method returns True only for single-byte |
| | characters. For Type 42 fonts this method return True if the character is |
| | from the Basic Multilingual Plane. |
| | """ |
| | if fonttype == 3: |
| | return glyph <= 255 |
| | if fonttype == 42: |
| | return glyph <= 65535 |
| | raise NotImplementedError() |
| |
|
| |
|
| | class Reference: |
| | """ |
| | PDF reference object. |
| | |
| | Use PdfFile.reserveObject() to create References. |
| | """ |
| |
|
| | def __init__(self, id): |
| | self.id = id |
| |
|
| | def __repr__(self): |
| | return "<Reference %d>" % self.id |
| |
|
| | def pdfRepr(self): |
| | return b"%d 0 R" % self.id |
| |
|
| | def write(self, contents, file): |
| | write = file.write |
| | write(b"%d 0 obj\n" % self.id) |
| | write(pdfRepr(contents)) |
| | write(b"\nendobj\n") |
| |
|
| |
|
| | @total_ordering |
| | class Name: |
| | """PDF name object.""" |
| | __slots__ = ('name',) |
| | _hexify = {c: '#%02x' % c |
| | for c in {*range(256)} - {*range(ord('!'), ord('~') + 1)}} |
| |
|
| | def __init__(self, name): |
| | if isinstance(name, Name): |
| | self.name = name.name |
| | else: |
| | if isinstance(name, bytes): |
| | name = name.decode('ascii') |
| | self.name = name.translate(self._hexify).encode('ascii') |
| |
|
| | def __repr__(self): |
| | return "<Name %s>" % self.name |
| |
|
| | def __str__(self): |
| | return '/' + self.name.decode('ascii') |
| |
|
| | def __eq__(self, other): |
| | return isinstance(other, Name) and self.name == other.name |
| |
|
| | def __lt__(self, other): |
| | return isinstance(other, Name) and self.name < other.name |
| |
|
| | def __hash__(self): |
| | return hash(self.name) |
| |
|
| | @staticmethod |
| | @_api.deprecated("3.6") |
| | def hexify(match): |
| | return '#%02x' % ord(match.group()) |
| |
|
| | def pdfRepr(self): |
| | return b'/' + self.name |
| |
|
| |
|
| | @_api.deprecated("3.6") |
| | class Operator: |
| | __slots__ = ('op',) |
| |
|
| | def __init__(self, op): |
| | self.op = op |
| |
|
| | def __repr__(self): |
| | return '<Operator %s>' % self.op |
| |
|
| | def pdfRepr(self): |
| | return self.op |
| |
|
| |
|
| | class Verbatim: |
| | """Store verbatim PDF command content for later inclusion in the stream.""" |
| | def __init__(self, x): |
| | self._x = x |
| |
|
| | def pdfRepr(self): |
| | return self._x |
| |
|
| |
|
| | class Op(Enum): |
| | """PDF operators (not an exhaustive list).""" |
| |
|
| | close_fill_stroke = b'b' |
| | fill_stroke = b'B' |
| | fill = b'f' |
| | closepath = b'h' |
| | close_stroke = b's' |
| | stroke = b'S' |
| | endpath = b'n' |
| | begin_text = b'BT' |
| | end_text = b'ET' |
| | curveto = b'c' |
| | rectangle = b're' |
| | lineto = b'l' |
| | moveto = b'm' |
| | concat_matrix = b'cm' |
| | use_xobject = b'Do' |
| | setgray_stroke = b'G' |
| | setgray_nonstroke = b'g' |
| | setrgb_stroke = b'RG' |
| | setrgb_nonstroke = b'rg' |
| | setcolorspace_stroke = b'CS' |
| | setcolorspace_nonstroke = b'cs' |
| | setcolor_stroke = b'SCN' |
| | setcolor_nonstroke = b'scn' |
| | setdash = b'd' |
| | setlinejoin = b'j' |
| | setlinecap = b'J' |
| | setgstate = b'gs' |
| | gsave = b'q' |
| | grestore = b'Q' |
| | textpos = b'Td' |
| | selectfont = b'Tf' |
| | textmatrix = b'Tm' |
| | show = b'Tj' |
| | showkern = b'TJ' |
| | setlinewidth = b'w' |
| | clip = b'W' |
| | shading = b'sh' |
| |
|
| | op = _api.deprecated('3.6')(property(lambda self: self.value)) |
| |
|
| | def pdfRepr(self): |
| | return self.value |
| |
|
| | @classmethod |
| | def paint_path(cls, fill, stroke): |
| | """ |
| | Return the PDF operator to paint a path. |
| | |
| | Parameters |
| | ---------- |
| | fill : bool |
| | Fill the path with the fill color. |
| | stroke : bool |
| | Stroke the outline of the path with the line color. |
| | """ |
| | if stroke: |
| | if fill: |
| | return cls.fill_stroke |
| | else: |
| | return cls.stroke |
| | else: |
| | if fill: |
| | return cls.fill |
| | else: |
| | return cls.endpath |
| |
|
| |
|
| | class Stream: |
| | """ |
| | PDF stream object. |
| | |
| | This has no pdfRepr method. Instead, call begin(), then output the |
| | contents of the stream by calling write(), and finally call end(). |
| | """ |
| | __slots__ = ('id', 'len', 'pdfFile', 'file', 'compressobj', 'extra', 'pos') |
| |
|
| | def __init__(self, id, len, file, extra=None, png=None): |
| | """ |
| | Parameters |
| | ---------- |
| | id : int |
| | Object id of the stream. |
| | len : Reference or None |
| | An unused Reference object for the length of the stream; |
| | None means to use a memory buffer so the length can be inlined. |
| | file : PdfFile |
| | The underlying object to write the stream to. |
| | extra : dict from Name to anything, or None |
| | Extra key-value pairs to include in the stream header. |
| | png : dict or None |
| | If the data is already png encoded, the decode parameters. |
| | """ |
| | self.id = id |
| | self.len = len |
| | self.pdfFile = file |
| | self.file = file.fh |
| | self.compressobj = None |
| | if extra is None: |
| | self.extra = dict() |
| | else: |
| | self.extra = extra.copy() |
| | if png is not None: |
| | self.extra.update({'Filter': Name('FlateDecode'), |
| | 'DecodeParms': png}) |
| |
|
| | self.pdfFile.recordXref(self.id) |
| | if mpl.rcParams['pdf.compression'] and not png: |
| | self.compressobj = zlib.compressobj( |
| | mpl.rcParams['pdf.compression']) |
| | if self.len is None: |
| | self.file = BytesIO() |
| | else: |
| | self._writeHeader() |
| | self.pos = self.file.tell() |
| |
|
| | def _writeHeader(self): |
| | write = self.file.write |
| | write(b"%d 0 obj\n" % self.id) |
| | dict = self.extra |
| | dict['Length'] = self.len |
| | if mpl.rcParams['pdf.compression']: |
| | dict['Filter'] = Name('FlateDecode') |
| |
|
| | write(pdfRepr(dict)) |
| | write(b"\nstream\n") |
| |
|
| | def end(self): |
| | """Finalize stream.""" |
| |
|
| | self._flush() |
| | if self.len is None: |
| | contents = self.file.getvalue() |
| | self.len = len(contents) |
| | self.file = self.pdfFile.fh |
| | self._writeHeader() |
| | self.file.write(contents) |
| | self.file.write(b"\nendstream\nendobj\n") |
| | else: |
| | length = self.file.tell() - self.pos |
| | self.file.write(b"\nendstream\nendobj\n") |
| | self.pdfFile.writeObject(self.len, length) |
| |
|
| | def write(self, data): |
| | """Write some data on the stream.""" |
| |
|
| | if self.compressobj is None: |
| | self.file.write(data) |
| | else: |
| | compressed = self.compressobj.compress(data) |
| | self.file.write(compressed) |
| |
|
| | def _flush(self): |
| | """Flush the compression object.""" |
| |
|
| | if self.compressobj is not None: |
| | compressed = self.compressobj.flush() |
| | self.file.write(compressed) |
| | self.compressobj = None |
| |
|
| |
|
| | def _get_pdf_charprocs(font_path, glyph_ids): |
| | font = get_font(font_path, hinting_factor=1) |
| | conv = 1000 / font.units_per_EM |
| | procs = {} |
| | for glyph_id in glyph_ids: |
| | g = font.load_glyph(glyph_id, LOAD_NO_SCALE) |
| | |
| | |
| | |
| | d1 = (np.array([g.horiAdvance, 0, *g.bbox]) * conv + .5).astype(int) |
| | v, c = font.get_path() |
| | v = (v * 64).astype(int) |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | quads, = np.nonzero(c == 3) |
| | quads_on = quads[1::2] |
| | quads_mid_on = np.array( |
| | sorted({*quads_on} & {*(quads - 1)} & {*(quads + 1)}), int) |
| | implicit = quads_mid_on[ |
| | (v[quads_mid_on] |
| | == ((v[quads_mid_on - 1] + v[quads_mid_on + 1]) / 2).astype(int)) |
| | .all(axis=1)] |
| | if (font.postscript_name, glyph_id) in [ |
| | ("DejaVuSerif-Italic", 77), |
| | ("DejaVuSerif-Italic", 135), |
| | ]: |
| | v[:, 0] -= 1 |
| | v = (v * conv + .5).astype(int) |
| | v[implicit] = (( |
| | (v[implicit - 1] + v[implicit + 1]) / 2).astype(int)) |
| | procs[font.get_glyph_name(glyph_id)] = ( |
| | " ".join(map(str, d1)).encode("ascii") + b" d1\n" |
| | + _path.convert_to_string( |
| | Path(v, c), None, None, False, None, -1, |
| | |
| | [b"m", b"l", b"", b"c", b"h"], True) |
| | + b"f") |
| | return procs |
| |
|
| |
|
| | class PdfFile: |
| | """PDF file object.""" |
| |
|
| | def __init__(self, filename, metadata=None): |
| | """ |
| | Parameters |
| | ---------- |
| | filename : str or path-like or file-like |
| | Output target; if a string, a file will be opened for writing. |
| | |
| | metadata : dict from strings to strings and dates |
| | Information dictionary object (see PDF reference section 10.2.1 |
| | 'Document Information Dictionary'), e.g.: |
| | ``{'Creator': 'My software', 'Author': 'Me', 'Title': 'Awesome'}``. |
| | |
| | The standard keys are 'Title', 'Author', 'Subject', 'Keywords', |
| | 'Creator', 'Producer', 'CreationDate', 'ModDate', and |
| | 'Trapped'. Values have been predefined for 'Creator', 'Producer' |
| | and 'CreationDate'. They can be removed by setting them to `None`. |
| | """ |
| | super().__init__() |
| |
|
| | self._object_seq = itertools.count(1) |
| | self.xrefTable = [[0, 65535, 'the zero object']] |
| | self.passed_in_file_object = False |
| | self.original_file_like = None |
| | self.tell_base = 0 |
| | fh, opened = cbook.to_filehandle(filename, "wb", return_opened=True) |
| | if not opened: |
| | try: |
| | self.tell_base = filename.tell() |
| | except IOError: |
| | fh = BytesIO() |
| | self.original_file_like = filename |
| | else: |
| | fh = filename |
| | self.passed_in_file_object = True |
| |
|
| | self.fh = fh |
| | self.currentstream = None |
| | fh.write(b"%PDF-1.4\n") |
| | |
| | |
| | |
| | fh.write(b"%\254\334 \253\272\n") |
| |
|
| | self.rootObject = self.reserveObject('root') |
| | self.pagesObject = self.reserveObject('pages') |
| | self.pageList = [] |
| | self.fontObject = self.reserveObject('fonts') |
| | self._extGStateObject = self.reserveObject('extended graphics states') |
| | self.hatchObject = self.reserveObject('tiling patterns') |
| | self.gouraudObject = self.reserveObject('Gouraud triangles') |
| | self.XObjectObject = self.reserveObject('external objects') |
| | self.resourceObject = self.reserveObject('resources') |
| |
|
| | root = {'Type': Name('Catalog'), |
| | 'Pages': self.pagesObject} |
| | self.writeObject(self.rootObject, root) |
| |
|
| | self.infoDict = _create_pdf_info_dict('pdf', metadata or {}) |
| |
|
| | self.fontNames = {} |
| | self._internal_font_seq = (Name(f'F{i}') for i in itertools.count(1)) |
| | self.dviFontInfo = {} |
| | |
| | self.type1Descriptors = {} |
| | self._character_tracker = _backend_pdf_ps.CharacterTracker() |
| |
|
| | self.alphaStates = {} |
| | self._alpha_state_seq = (Name(f'A{i}') for i in itertools.count(1)) |
| | self._soft_mask_states = {} |
| | self._soft_mask_seq = (Name(f'SM{i}') for i in itertools.count(1)) |
| | self._soft_mask_groups = [] |
| | self.hatchPatterns = {} |
| | self._hatch_pattern_seq = (Name(f'H{i}') for i in itertools.count(1)) |
| | self.gouraudTriangles = [] |
| |
|
| | self._images = {} |
| | self._image_seq = (Name(f'I{i}') for i in itertools.count(1)) |
| |
|
| | self.markers = {} |
| | self.multi_byte_charprocs = {} |
| |
|
| | self.paths = [] |
| |
|
| | |
| | |
| | |
| | self._annotations = [] |
| | |
| | |
| | self.pageAnnotations = [] |
| |
|
| | |
| | procsets = [Name(x) for x in "PDF Text ImageB ImageC ImageI".split()] |
| |
|
| | |
| | |
| | |
| | resources = {'Font': self.fontObject, |
| | 'XObject': self.XObjectObject, |
| | 'ExtGState': self._extGStateObject, |
| | 'Pattern': self.hatchObject, |
| | 'Shading': self.gouraudObject, |
| | 'ProcSet': procsets} |
| | self.writeObject(self.resourceObject, resources) |
| |
|
| | def newPage(self, width, height): |
| | self.endStream() |
| |
|
| | self.width, self.height = width, height |
| | contentObject = self.reserveObject('page contents') |
| | annotsObject = self.reserveObject('annotations') |
| | thePage = {'Type': Name('Page'), |
| | 'Parent': self.pagesObject, |
| | 'Resources': self.resourceObject, |
| | 'MediaBox': [0, 0, 72 * width, 72 * height], |
| | 'Contents': contentObject, |
| | 'Annots': annotsObject, |
| | } |
| | pageObject = self.reserveObject('page') |
| | self.writeObject(pageObject, thePage) |
| | self.pageList.append(pageObject) |
| | self._annotations.append((annotsObject, self.pageAnnotations)) |
| |
|
| | self.beginStream(contentObject.id, |
| | self.reserveObject('length of content stream')) |
| | |
| | |
| | self.output(Name('DeviceRGB'), Op.setcolorspace_stroke) |
| | self.output(Name('DeviceRGB'), Op.setcolorspace_nonstroke) |
| | self.output(GraphicsContextPdf.joinstyles['round'], Op.setlinejoin) |
| |
|
| | |
| | self.pageAnnotations = [] |
| |
|
| | def newTextnote(self, text, positionRect=[-100, -100, 0, 0]): |
| | |
| | theNote = {'Type': Name('Annot'), |
| | 'Subtype': Name('Text'), |
| | 'Contents': text, |
| | 'Rect': positionRect, |
| | } |
| | self.pageAnnotations.append(theNote) |
| |
|
| | def _get_subsetted_psname(self, ps_name, charmap): |
| | def toStr(n, base): |
| | if n < base: |
| | return string.ascii_uppercase[n] |
| | else: |
| | return ( |
| | toStr(n // base, base) + string.ascii_uppercase[n % base] |
| | ) |
| |
|
| | |
| | hashed = hash(frozenset(charmap.keys())) % ((sys.maxsize + 1) * 2) |
| | prefix = toStr(hashed, 26) |
| |
|
| | |
| | return prefix[:6] + "+" + ps_name |
| |
|
| | def finalize(self): |
| | """Write out the various deferred objects and the pdf end matter.""" |
| |
|
| | self.endStream() |
| | self._write_annotations() |
| | self.writeFonts() |
| | self.writeExtGSTates() |
| | self._write_soft_mask_groups() |
| | self.writeHatches() |
| | self.writeGouraudTriangles() |
| | xobjects = { |
| | name: ob for image, name, ob in self._images.values()} |
| | for tup in self.markers.values(): |
| | xobjects[tup[0]] = tup[1] |
| | for name, value in self.multi_byte_charprocs.items(): |
| | xobjects[name] = value |
| | for name, path, trans, ob, join, cap, padding, filled, stroked \ |
| | in self.paths: |
| | xobjects[name] = ob |
| | self.writeObject(self.XObjectObject, xobjects) |
| | self.writeImages() |
| | self.writeMarkers() |
| | self.writePathCollectionTemplates() |
| | self.writeObject(self.pagesObject, |
| | {'Type': Name('Pages'), |
| | 'Kids': self.pageList, |
| | 'Count': len(self.pageList)}) |
| | self.writeInfoDict() |
| |
|
| | |
| | self.writeXref() |
| | self.writeTrailer() |
| |
|
| | def close(self): |
| | """Flush all buffers and free all resources.""" |
| |
|
| | self.endStream() |
| | if self.passed_in_file_object: |
| | self.fh.flush() |
| | else: |
| | if self.original_file_like is not None: |
| | self.original_file_like.write(self.fh.getvalue()) |
| | self.fh.close() |
| |
|
| | def write(self, data): |
| | if self.currentstream is None: |
| | self.fh.write(data) |
| | else: |
| | self.currentstream.write(data) |
| |
|
| | def output(self, *data): |
| | self.write(_fill([pdfRepr(x) for x in data])) |
| | self.write(b'\n') |
| |
|
| | def beginStream(self, id, len, extra=None, png=None): |
| | assert self.currentstream is None |
| | self.currentstream = Stream(id, len, self, extra, png) |
| |
|
| | def endStream(self): |
| | if self.currentstream is not None: |
| | self.currentstream.end() |
| | self.currentstream = None |
| |
|
| | def outputStream(self, ref, data, *, extra=None): |
| | self.beginStream(ref.id, None, extra) |
| | self.currentstream.write(data) |
| | self.endStream() |
| |
|
| | def _write_annotations(self): |
| | for annotsObject, annotations in self._annotations: |
| | self.writeObject(annotsObject, annotations) |
| |
|
| | def fontName(self, fontprop): |
| | """ |
| | Select a font based on fontprop and return a name suitable for |
| | Op.selectfont. If fontprop is a string, it will be interpreted |
| | as the filename of the font. |
| | """ |
| |
|
| | if isinstance(fontprop, str): |
| | filenames = [fontprop] |
| | elif mpl.rcParams['pdf.use14corefonts']: |
| | filenames = _fontManager._find_fonts_by_props( |
| | fontprop, fontext='afm', directory=RendererPdf._afm_font_dir |
| | ) |
| | else: |
| | filenames = _fontManager._find_fonts_by_props(fontprop) |
| | first_Fx = None |
| | for fname in filenames: |
| | Fx = self.fontNames.get(fname) |
| | if not first_Fx: |
| | first_Fx = Fx |
| | if Fx is None: |
| | Fx = next(self._internal_font_seq) |
| | self.fontNames[fname] = Fx |
| | _log.debug('Assigning font %s = %r', Fx, fname) |
| | if not first_Fx: |
| | first_Fx = Fx |
| |
|
| | |
| | |
| | return first_Fx |
| |
|
| | def dviFontName(self, dvifont): |
| | """ |
| | Given a dvi font object, return a name suitable for Op.selectfont. |
| | This registers the font information in ``self.dviFontInfo`` if not yet |
| | registered. |
| | """ |
| |
|
| | dvi_info = self.dviFontInfo.get(dvifont.texname) |
| | if dvi_info is not None: |
| | return dvi_info.pdfname |
| |
|
| | tex_font_map = dviread.PsfontsMap(dviread._find_tex_file('pdftex.map')) |
| | psfont = tex_font_map[dvifont.texname] |
| | if psfont.filename is None: |
| | raise ValueError( |
| | "No usable font file found for {} (TeX: {}); " |
| | "the font may lack a Type-1 version" |
| | .format(psfont.psname, dvifont.texname)) |
| |
|
| | pdfname = next(self._internal_font_seq) |
| | _log.debug('Assigning font %s = %s (dvi)', pdfname, dvifont.texname) |
| | self.dviFontInfo[dvifont.texname] = types.SimpleNamespace( |
| | dvifont=dvifont, |
| | pdfname=pdfname, |
| | fontfile=psfont.filename, |
| | basefont=psfont.psname, |
| | encodingfile=psfont.encoding, |
| | effects=psfont.effects) |
| | return pdfname |
| |
|
| | def writeFonts(self): |
| | fonts = {} |
| | for dviname, info in sorted(self.dviFontInfo.items()): |
| | Fx = info.pdfname |
| | _log.debug('Embedding Type-1 font %s from dvi.', dviname) |
| | fonts[Fx] = self._embedTeXFont(info) |
| | for filename in sorted(self.fontNames): |
| | Fx = self.fontNames[filename] |
| | _log.debug('Embedding font %s.', filename) |
| | if filename.endswith('.afm'): |
| | |
| | _log.debug('Writing AFM font.') |
| | fonts[Fx] = self._write_afm_font(filename) |
| | else: |
| | |
| | _log.debug('Writing TrueType font.') |
| | chars = self._character_tracker.used.get(filename) |
| | if chars: |
| | fonts[Fx] = self.embedTTF(filename, chars) |
| | self.writeObject(self.fontObject, fonts) |
| |
|
| | def _write_afm_font(self, filename): |
| | with open(filename, 'rb') as fh: |
| | font = AFM(fh) |
| | fontname = font.get_fontname() |
| | fontdict = {'Type': Name('Font'), |
| | 'Subtype': Name('Type1'), |
| | 'BaseFont': Name(fontname), |
| | 'Encoding': Name('WinAnsiEncoding')} |
| | fontdictObject = self.reserveObject('font dictionary') |
| | self.writeObject(fontdictObject, fontdict) |
| | return fontdictObject |
| |
|
| | def _embedTeXFont(self, fontinfo): |
| | _log.debug('Embedding TeX font %s - fontinfo=%s', |
| | fontinfo.dvifont.texname, fontinfo.__dict__) |
| |
|
| | |
| | widthsObject = self.reserveObject('font widths') |
| | self.writeObject(widthsObject, fontinfo.dvifont.widths) |
| |
|
| | |
| | fontdictObject = self.reserveObject('font dictionary') |
| | fontdict = { |
| | 'Type': Name('Font'), |
| | 'Subtype': Name('Type1'), |
| | 'FirstChar': 0, |
| | 'LastChar': len(fontinfo.dvifont.widths) - 1, |
| | 'Widths': widthsObject, |
| | } |
| |
|
| | |
| | if fontinfo.encodingfile is not None: |
| | fontdict['Encoding'] = { |
| | 'Type': Name('Encoding'), |
| | 'Differences': [ |
| | 0, *map(Name, dviread._parse_enc(fontinfo.encodingfile))], |
| | } |
| |
|
| | |
| | if fontinfo.fontfile is None: |
| | _log.warning( |
| | "Because of TeX configuration (pdftex.map, see updmap option " |
| | "pdftexDownloadBase14) the font %s is not embedded. This is " |
| | "deprecated as of PDF 1.5 and it may cause the consumer " |
| | "application to show something that was not intended.", |
| | fontinfo.basefont) |
| | fontdict['BaseFont'] = Name(fontinfo.basefont) |
| | self.writeObject(fontdictObject, fontdict) |
| | return fontdictObject |
| |
|
| | |
| | t1font = _type1font.Type1Font(fontinfo.fontfile) |
| | if fontinfo.effects: |
| | t1font = t1font.transform(fontinfo.effects) |
| | fontdict['BaseFont'] = Name(t1font.prop['FontName']) |
| |
|
| | |
| | |
| | |
| | effects = (fontinfo.effects.get('slant', 0.0), |
| | fontinfo.effects.get('extend', 1.0)) |
| | fontdesc = self.type1Descriptors.get((fontinfo.fontfile, effects)) |
| | if fontdesc is None: |
| | fontdesc = self.createType1Descriptor(t1font, fontinfo.fontfile) |
| | self.type1Descriptors[(fontinfo.fontfile, effects)] = fontdesc |
| | fontdict['FontDescriptor'] = fontdesc |
| |
|
| | self.writeObject(fontdictObject, fontdict) |
| | return fontdictObject |
| |
|
| | def createType1Descriptor(self, t1font, fontfile): |
| | |
| | |
| | fontdescObject = self.reserveObject('font descriptor') |
| | fontfileObject = self.reserveObject('font file') |
| |
|
| | italic_angle = t1font.prop['ItalicAngle'] |
| | fixed_pitch = t1font.prop['isFixedPitch'] |
| |
|
| | flags = 0 |
| | |
| | if fixed_pitch: |
| | flags |= 1 << 0 |
| | |
| | if 0: |
| | flags |= 1 << 1 |
| | |
| | if 1: |
| | flags |= 1 << 2 |
| | |
| | else: |
| | flags |= 1 << 5 |
| | |
| | if italic_angle: |
| | flags |= 1 << 6 |
| | |
| | if 0: |
| | flags |= 1 << 16 |
| | |
| | if 0: |
| | flags |= 1 << 17 |
| | |
| | if 0: |
| | flags |= 1 << 18 |
| |
|
| | ft2font = get_font(fontfile) |
| |
|
| | descriptor = { |
| | 'Type': Name('FontDescriptor'), |
| | 'FontName': Name(t1font.prop['FontName']), |
| | 'Flags': flags, |
| | 'FontBBox': ft2font.bbox, |
| | 'ItalicAngle': italic_angle, |
| | 'Ascent': ft2font.ascender, |
| | 'Descent': ft2font.descender, |
| | 'CapHeight': 1000, |
| | 'XHeight': 500, |
| | 'FontFile': fontfileObject, |
| | 'FontFamily': t1font.prop['FamilyName'], |
| | 'StemV': 50, |
| | |
| | |
| | } |
| |
|
| | self.writeObject(fontdescObject, descriptor) |
| |
|
| | self.outputStream(fontfileObject, b"".join(t1font.parts[:2]), |
| | extra={'Length1': len(t1font.parts[0]), |
| | 'Length2': len(t1font.parts[1]), |
| | 'Length3': 0}) |
| |
|
| | return fontdescObject |
| |
|
| | def _get_xobject_glyph_name(self, filename, glyph_name): |
| | Fx = self.fontName(filename) |
| | return "-".join([ |
| | Fx.name.decode(), |
| | os.path.splitext(os.path.basename(filename))[0], |
| | glyph_name]) |
| |
|
| | _identityToUnicodeCMap = b"""/CIDInit /ProcSet findresource begin |
| | 12 dict begin |
| | begincmap |
| | /CIDSystemInfo |
| | << /Registry (Adobe) |
| | /Ordering (UCS) |
| | /Supplement 0 |
| | >> def |
| | /CMapName /Adobe-Identity-UCS def |
| | /CMapType 2 def |
| | 1 begincodespacerange |
| | <0000> <ffff> |
| | endcodespacerange |
| | %d beginbfrange |
| | %s |
| | endbfrange |
| | endcmap |
| | CMapName currentdict /CMap defineresource pop |
| | end |
| | end""" |
| |
|
| | def embedTTF(self, filename, characters): |
| | """Embed the TTF font from the named file into the document.""" |
| |
|
| | font = get_font(filename) |
| | fonttype = mpl.rcParams['pdf.fonttype'] |
| |
|
| | def cvt(length, upe=font.units_per_EM, nearest=True): |
| | """Convert font coordinates to PDF glyph coordinates.""" |
| | value = length / upe * 1000 |
| | if nearest: |
| | return round(value) |
| | |
| | if value < 0: |
| | return math.floor(value) |
| | else: |
| | return math.ceil(value) |
| |
|
| | def embedTTFType3(font, characters, descriptor): |
| | """The Type 3-specific part of embedding a Truetype font""" |
| | widthsObject = self.reserveObject('font widths') |
| | fontdescObject = self.reserveObject('font descriptor') |
| | fontdictObject = self.reserveObject('font dictionary') |
| | charprocsObject = self.reserveObject('character procs') |
| | differencesArray = [] |
| | firstchar, lastchar = 0, 255 |
| | bbox = [cvt(x, nearest=False) for x in font.bbox] |
| |
|
| | fontdict = { |
| | 'Type': Name('Font'), |
| | 'BaseFont': ps_name, |
| | 'FirstChar': firstchar, |
| | 'LastChar': lastchar, |
| | 'FontDescriptor': fontdescObject, |
| | 'Subtype': Name('Type3'), |
| | 'Name': descriptor['FontName'], |
| | 'FontBBox': bbox, |
| | 'FontMatrix': [.001, 0, 0, .001, 0, 0], |
| | 'CharProcs': charprocsObject, |
| | 'Encoding': { |
| | 'Type': Name('Encoding'), |
| | 'Differences': differencesArray}, |
| | 'Widths': widthsObject |
| | } |
| |
|
| | from encodings import cp1252 |
| |
|
| | |
| | def get_char_width(charcode): |
| | s = ord(cp1252.decoding_table[charcode]) |
| | width = font.load_char( |
| | s, flags=LOAD_NO_SCALE | LOAD_NO_HINTING).horiAdvance |
| | return cvt(width) |
| | with warnings.catch_warnings(): |
| | |
| | |
| | |
| | warnings.filterwarnings("ignore") |
| | widths = [get_char_width(charcode) |
| | for charcode in range(firstchar, lastchar+1)] |
| | descriptor['MaxWidth'] = max(widths) |
| |
|
| | |
| | |
| | |
| | glyph_ids = [] |
| | differences = [] |
| | multi_byte_chars = set() |
| | for c in characters: |
| | ccode = c |
| | gind = font.get_char_index(ccode) |
| | glyph_ids.append(gind) |
| | glyph_name = font.get_glyph_name(gind) |
| | if ccode <= 255: |
| | differences.append((ccode, glyph_name)) |
| | else: |
| | multi_byte_chars.add(glyph_name) |
| | differences.sort() |
| |
|
| | last_c = -2 |
| | for c, name in differences: |
| | if c != last_c + 1: |
| | differencesArray.append(c) |
| | differencesArray.append(Name(name)) |
| | last_c = c |
| |
|
| | |
| | rawcharprocs = _get_pdf_charprocs(filename, glyph_ids) |
| | charprocs = {} |
| | for charname in sorted(rawcharprocs): |
| | stream = rawcharprocs[charname] |
| | charprocDict = {} |
| | |
| | |
| | if charname in multi_byte_chars: |
| | charprocDict = {'Type': Name('XObject'), |
| | 'Subtype': Name('Form'), |
| | 'BBox': bbox} |
| | |
| | |
| | |
| | |
| | |
| | |
| | stream = stream[stream.find(b"d1") + 2:] |
| | charprocObject = self.reserveObject('charProc') |
| | self.outputStream(charprocObject, stream, extra=charprocDict) |
| |
|
| | |
| | |
| | if charname in multi_byte_chars: |
| | name = self._get_xobject_glyph_name(filename, charname) |
| | self.multi_byte_charprocs[name] = charprocObject |
| | else: |
| | charprocs[charname] = charprocObject |
| |
|
| | |
| | self.writeObject(fontdictObject, fontdict) |
| | self.writeObject(fontdescObject, descriptor) |
| | self.writeObject(widthsObject, widths) |
| | self.writeObject(charprocsObject, charprocs) |
| |
|
| | return fontdictObject |
| |
|
| | def embedTTFType42(font, characters, descriptor): |
| | """The Type 42-specific part of embedding a Truetype font""" |
| | fontdescObject = self.reserveObject('font descriptor') |
| | cidFontDictObject = self.reserveObject('CID font dictionary') |
| | type0FontDictObject = self.reserveObject('Type 0 font dictionary') |
| | cidToGidMapObject = self.reserveObject('CIDToGIDMap stream') |
| | fontfileObject = self.reserveObject('font file stream') |
| | wObject = self.reserveObject('Type 0 widths') |
| | toUnicodeMapObject = self.reserveObject('ToUnicode map') |
| |
|
| | subset_str = "".join(chr(c) for c in characters) |
| | _log.debug("SUBSET %s characters: %s", filename, subset_str) |
| | fontdata = _backend_pdf_ps.get_glyphs_subset(filename, subset_str) |
| | _log.debug( |
| | "SUBSET %s %d -> %d", filename, |
| | os.stat(filename).st_size, fontdata.getbuffer().nbytes |
| | ) |
| |
|
| | |
| | full_font = font |
| |
|
| | |
| | |
| | |
| | font = FT2Font(fontdata) |
| |
|
| | cidFontDict = { |
| | 'Type': Name('Font'), |
| | 'Subtype': Name('CIDFontType2'), |
| | 'BaseFont': ps_name, |
| | 'CIDSystemInfo': { |
| | 'Registry': 'Adobe', |
| | 'Ordering': 'Identity', |
| | 'Supplement': 0}, |
| | 'FontDescriptor': fontdescObject, |
| | 'W': wObject, |
| | 'CIDToGIDMap': cidToGidMapObject |
| | } |
| |
|
| | type0FontDict = { |
| | 'Type': Name('Font'), |
| | 'Subtype': Name('Type0'), |
| | 'BaseFont': ps_name, |
| | 'Encoding': Name('Identity-H'), |
| | 'DescendantFonts': [cidFontDictObject], |
| | 'ToUnicode': toUnicodeMapObject |
| | } |
| |
|
| | |
| | descriptor['FontFile2'] = fontfileObject |
| | self.outputStream( |
| | fontfileObject, fontdata.getvalue(), |
| | extra={'Length1': fontdata.getbuffer().nbytes}) |
| |
|
| | |
| | |
| | cid_to_gid_map = ['\0'] * 65536 |
| | widths = [] |
| | max_ccode = 0 |
| | for c in characters: |
| | ccode = c |
| | gind = font.get_char_index(ccode) |
| | glyph = font.load_char(ccode, |
| | flags=LOAD_NO_SCALE | LOAD_NO_HINTING) |
| | widths.append((ccode, cvt(glyph.horiAdvance))) |
| | if ccode < 65536: |
| | cid_to_gid_map[ccode] = chr(gind) |
| | max_ccode = max(ccode, max_ccode) |
| | widths.sort() |
| | cid_to_gid_map = cid_to_gid_map[:max_ccode + 1] |
| |
|
| | last_ccode = -2 |
| | w = [] |
| | max_width = 0 |
| | unicode_groups = [] |
| | for ccode, width in widths: |
| | if ccode != last_ccode + 1: |
| | w.append(ccode) |
| | w.append([width]) |
| | unicode_groups.append([ccode, ccode]) |
| | else: |
| | w[-1].append(width) |
| | unicode_groups[-1][1] = ccode |
| | max_width = max(max_width, width) |
| | last_ccode = ccode |
| |
|
| | unicode_bfrange = [] |
| | for start, end in unicode_groups: |
| | |
| | if start > 65535: |
| | continue |
| | end = min(65535, end) |
| |
|
| | unicode_bfrange.append( |
| | b"<%04x> <%04x> [%s]" % |
| | (start, end, |
| | b" ".join(b"<%04x>" % x for x in range(start, end+1)))) |
| | unicode_cmap = (self._identityToUnicodeCMap % |
| | (len(unicode_groups), b"\n".join(unicode_bfrange))) |
| |
|
| | |
| | glyph_ids = [] |
| | for ccode in characters: |
| | if not _font_supports_glyph(fonttype, ccode): |
| | gind = full_font.get_char_index(ccode) |
| | glyph_ids.append(gind) |
| |
|
| | bbox = [cvt(x, nearest=False) for x in full_font.bbox] |
| | rawcharprocs = _get_pdf_charprocs(filename, glyph_ids) |
| | for charname in sorted(rawcharprocs): |
| | stream = rawcharprocs[charname] |
| | charprocDict = {'Type': Name('XObject'), |
| | 'Subtype': Name('Form'), |
| | 'BBox': bbox} |
| | |
| | |
| | |
| | |
| | |
| | |
| | stream = stream[stream.find(b"d1") + 2:] |
| | charprocObject = self.reserveObject('charProc') |
| | self.outputStream(charprocObject, stream, extra=charprocDict) |
| |
|
| | name = self._get_xobject_glyph_name(filename, charname) |
| | self.multi_byte_charprocs[name] = charprocObject |
| |
|
| | |
| | cid_to_gid_map = "".join(cid_to_gid_map).encode("utf-16be") |
| | self.outputStream(cidToGidMapObject, cid_to_gid_map) |
| |
|
| | |
| | self.outputStream(toUnicodeMapObject, unicode_cmap) |
| |
|
| | descriptor['MaxWidth'] = max_width |
| |
|
| | |
| | self.writeObject(cidFontDictObject, cidFontDict) |
| | self.writeObject(type0FontDictObject, type0FontDict) |
| | self.writeObject(fontdescObject, descriptor) |
| | self.writeObject(wObject, w) |
| |
|
| | return type0FontDictObject |
| |
|
| | |
| |
|
| | ps_name = self._get_subsetted_psname( |
| | font.postscript_name, |
| | font.get_charmap() |
| | ) |
| | ps_name = ps_name.encode('ascii', 'replace') |
| | ps_name = Name(ps_name) |
| | pclt = font.get_sfnt_table('pclt') or {'capHeight': 0, 'xHeight': 0} |
| | post = font.get_sfnt_table('post') or {'italicAngle': (0, 0)} |
| | ff = font.face_flags |
| | sf = font.style_flags |
| |
|
| | flags = 0 |
| | symbolic = False |
| | if ff & FIXED_WIDTH: |
| | flags |= 1 << 0 |
| | if 0: |
| | flags |= 1 << 1 |
| | if symbolic: |
| | flags |= 1 << 2 |
| | else: |
| | flags |= 1 << 5 |
| | if sf & ITALIC: |
| | flags |= 1 << 6 |
| | if 0: |
| | flags |= 1 << 16 |
| | if 0: |
| | flags |= 1 << 17 |
| | if 0: |
| | flags |= 1 << 18 |
| |
|
| | descriptor = { |
| | 'Type': Name('FontDescriptor'), |
| | 'FontName': ps_name, |
| | 'Flags': flags, |
| | 'FontBBox': [cvt(x, nearest=False) for x in font.bbox], |
| | 'Ascent': cvt(font.ascender, nearest=False), |
| | 'Descent': cvt(font.descender, nearest=False), |
| | 'CapHeight': cvt(pclt['capHeight'], nearest=False), |
| | 'XHeight': cvt(pclt['xHeight']), |
| | 'ItalicAngle': post['italicAngle'][1], |
| | 'StemV': 0 |
| | } |
| |
|
| | if fonttype == 3: |
| | return embedTTFType3(font, characters, descriptor) |
| | elif fonttype == 42: |
| | return embedTTFType42(font, characters, descriptor) |
| |
|
| | def alphaState(self, alpha): |
| | """Return name of an ExtGState that sets alpha to the given value.""" |
| |
|
| | state = self.alphaStates.get(alpha, None) |
| | if state is not None: |
| | return state[0] |
| |
|
| | name = next(self._alpha_state_seq) |
| | self.alphaStates[alpha] = \ |
| | (name, {'Type': Name('ExtGState'), |
| | 'CA': alpha[0], 'ca': alpha[1]}) |
| | return name |
| |
|
| | def _soft_mask_state(self, smask): |
| | """ |
| | Return an ExtGState that sets the soft mask to the given shading. |
| | |
| | Parameters |
| | ---------- |
| | smask : Reference |
| | Reference to a shading in DeviceGray color space, whose luminosity |
| | is to be used as the alpha channel. |
| | |
| | Returns |
| | ------- |
| | Name |
| | """ |
| |
|
| | state = self._soft_mask_states.get(smask, None) |
| | if state is not None: |
| | return state[0] |
| |
|
| | name = next(self._soft_mask_seq) |
| | groupOb = self.reserveObject('transparency group for soft mask') |
| | self._soft_mask_states[smask] = ( |
| | name, |
| | { |
| | 'Type': Name('ExtGState'), |
| | 'AIS': False, |
| | 'SMask': { |
| | 'Type': Name('Mask'), |
| | 'S': Name('Luminosity'), |
| | 'BC': [1], |
| | 'G': groupOb |
| | } |
| | } |
| | ) |
| | self._soft_mask_groups.append(( |
| | groupOb, |
| | { |
| | 'Type': Name('XObject'), |
| | 'Subtype': Name('Form'), |
| | 'FormType': 1, |
| | 'Group': { |
| | 'S': Name('Transparency'), |
| | 'CS': Name('DeviceGray') |
| | }, |
| | 'Matrix': [1, 0, 0, 1, 0, 0], |
| | 'Resources': {'Shading': {'S': smask}}, |
| | 'BBox': [0, 0, 1, 1] |
| | }, |
| | [Name('S'), Op.shading] |
| | )) |
| | return name |
| |
|
| | def writeExtGSTates(self): |
| | self.writeObject( |
| | self._extGStateObject, |
| | dict([ |
| | *self.alphaStates.values(), |
| | *self._soft_mask_states.values() |
| | ]) |
| | ) |
| |
|
| | def _write_soft_mask_groups(self): |
| | for ob, attributes, content in self._soft_mask_groups: |
| | self.beginStream(ob.id, None, attributes) |
| | self.output(*content) |
| | self.endStream() |
| |
|
| | def hatchPattern(self, hatch_style): |
| | |
| | if hatch_style is not None: |
| | edge, face, hatch = hatch_style |
| | if edge is not None: |
| | edge = tuple(edge) |
| | if face is not None: |
| | face = tuple(face) |
| | hatch_style = (edge, face, hatch) |
| |
|
| | pattern = self.hatchPatterns.get(hatch_style, None) |
| | if pattern is not None: |
| | return pattern |
| |
|
| | name = next(self._hatch_pattern_seq) |
| | self.hatchPatterns[hatch_style] = name |
| | return name |
| |
|
| | def writeHatches(self): |
| | hatchDict = dict() |
| | sidelen = 72.0 |
| | for hatch_style, name in self.hatchPatterns.items(): |
| | ob = self.reserveObject('hatch pattern') |
| | hatchDict[name] = ob |
| | res = {'Procsets': |
| | [Name(x) for x in "PDF Text ImageB ImageC ImageI".split()]} |
| | self.beginStream( |
| | ob.id, None, |
| | {'Type': Name('Pattern'), |
| | 'PatternType': 1, 'PaintType': 1, 'TilingType': 1, |
| | 'BBox': [0, 0, sidelen, sidelen], |
| | 'XStep': sidelen, 'YStep': sidelen, |
| | 'Resources': res, |
| | |
| | 'Matrix': [1, 0, 0, 1, 0, self.height * 72]}) |
| |
|
| | stroke_rgb, fill_rgb, hatch = hatch_style |
| | self.output(stroke_rgb[0], stroke_rgb[1], stroke_rgb[2], |
| | Op.setrgb_stroke) |
| | if fill_rgb is not None: |
| | self.output(fill_rgb[0], fill_rgb[1], fill_rgb[2], |
| | Op.setrgb_nonstroke, |
| | 0, 0, sidelen, sidelen, Op.rectangle, |
| | Op.fill) |
| |
|
| | self.output(mpl.rcParams['hatch.linewidth'], Op.setlinewidth) |
| |
|
| | self.output(*self.pathOperations( |
| | Path.hatch(hatch), |
| | Affine2D().scale(sidelen), |
| | simplify=False)) |
| | self.output(Op.fill_stroke) |
| |
|
| | self.endStream() |
| | self.writeObject(self.hatchObject, hatchDict) |
| |
|
| | def addGouraudTriangles(self, points, colors): |
| | """ |
| | Add a Gouraud triangle shading. |
| | |
| | Parameters |
| | ---------- |
| | points : np.ndarray |
| | Triangle vertices, shape (n, 3, 2) |
| | where n = number of triangles, 3 = vertices, 2 = x, y. |
| | colors : np.ndarray |
| | Vertex colors, shape (n, 3, 1) or (n, 3, 4) |
| | as with points, but last dimension is either (gray,) |
| | or (r, g, b, alpha). |
| | |
| | Returns |
| | ------- |
| | Name, Reference |
| | """ |
| | name = Name('GT%d' % len(self.gouraudTriangles)) |
| | ob = self.reserveObject(f'Gouraud triangle {name}') |
| | self.gouraudTriangles.append((name, ob, points, colors)) |
| | return name, ob |
| |
|
| | def writeGouraudTriangles(self): |
| | gouraudDict = dict() |
| | for name, ob, points, colors in self.gouraudTriangles: |
| | gouraudDict[name] = ob |
| | shape = points.shape |
| | flat_points = points.reshape((shape[0] * shape[1], 2)) |
| | colordim = colors.shape[2] |
| | assert colordim in (1, 4) |
| | flat_colors = colors.reshape((shape[0] * shape[1], colordim)) |
| | if colordim == 4: |
| | |
| | colordim = 3 |
| | points_min = np.min(flat_points, axis=0) - (1 << 8) |
| | points_max = np.max(flat_points, axis=0) + (1 << 8) |
| | factor = 0xffffffff / (points_max - points_min) |
| |
|
| | self.beginStream( |
| | ob.id, None, |
| | {'ShadingType': 4, |
| | 'BitsPerCoordinate': 32, |
| | 'BitsPerComponent': 8, |
| | 'BitsPerFlag': 8, |
| | 'ColorSpace': Name( |
| | 'DeviceRGB' if colordim == 3 else 'DeviceGray' |
| | ), |
| | 'AntiAlias': False, |
| | 'Decode': ([points_min[0], points_max[0], |
| | points_min[1], points_max[1]] |
| | + [0, 1] * colordim), |
| | }) |
| |
|
| | streamarr = np.empty( |
| | (shape[0] * shape[1],), |
| | dtype=[('flags', 'u1'), |
| | ('points', '>u4', (2,)), |
| | ('colors', 'u1', (colordim,))]) |
| | streamarr['flags'] = 0 |
| | streamarr['points'] = (flat_points - points_min) * factor |
| | streamarr['colors'] = flat_colors[:, :colordim] * 255.0 |
| |
|
| | self.write(streamarr.tobytes()) |
| | self.endStream() |
| | self.writeObject(self.gouraudObject, gouraudDict) |
| |
|
| | def imageObject(self, image): |
| | """Return name of an image XObject representing the given image.""" |
| |
|
| | entry = self._images.get(id(image), None) |
| | if entry is not None: |
| | return entry[1] |
| |
|
| | name = next(self._image_seq) |
| | ob = self.reserveObject(f'image {name}') |
| | self._images[id(image)] = (image, name, ob) |
| | return name |
| |
|
| | def _unpack(self, im): |
| | """ |
| | Unpack image array *im* into ``(data, alpha)``, which have shape |
| | ``(height, width, 3)`` (RGB) or ``(height, width, 1)`` (grayscale or |
| | alpha), except that alpha is None if the image is fully opaque. |
| | """ |
| | im = im[::-1] |
| | if im.ndim == 2: |
| | return im, None |
| | else: |
| | rgb = im[:, :, :3] |
| | rgb = np.array(rgb, order='C') |
| | |
| | if im.shape[2] == 4: |
| | alpha = im[:, :, 3][..., None] |
| | if np.all(alpha == 255): |
| | alpha = None |
| | else: |
| | alpha = np.array(alpha, order='C') |
| | else: |
| | alpha = None |
| | return rgb, alpha |
| |
|
| | def _writePng(self, img): |
| | """ |
| | Write the image *img* into the pdf file using png |
| | predictors with Flate compression. |
| | """ |
| | buffer = BytesIO() |
| | img.save(buffer, format="png") |
| | buffer.seek(8) |
| | png_data = b'' |
| | bit_depth = palette = None |
| | while True: |
| | length, type = struct.unpack(b'!L4s', buffer.read(8)) |
| | if type in [b'IHDR', b'PLTE', b'IDAT']: |
| | data = buffer.read(length) |
| | if len(data) != length: |
| | raise RuntimeError("truncated data") |
| | if type == b'IHDR': |
| | bit_depth = int(data[8]) |
| | elif type == b'PLTE': |
| | palette = data |
| | elif type == b'IDAT': |
| | png_data += data |
| | elif type == b'IEND': |
| | break |
| | else: |
| | buffer.seek(length, 1) |
| | buffer.seek(4, 1) |
| | return png_data, bit_depth, palette |
| |
|
| | def _writeImg(self, data, id, smask=None): |
| | """ |
| | Write the image *data*, of shape ``(height, width, 1)`` (grayscale) or |
| | ``(height, width, 3)`` (RGB), as pdf object *id* and with the soft mask |
| | (alpha channel) *smask*, which should be either None or a ``(height, |
| | width, 1)`` array. |
| | """ |
| | height, width, color_channels = data.shape |
| | obj = {'Type': Name('XObject'), |
| | 'Subtype': Name('Image'), |
| | 'Width': width, |
| | 'Height': height, |
| | 'ColorSpace': Name({1: 'DeviceGray', 3: 'DeviceRGB'}[color_channels]), |
| | 'BitsPerComponent': 8} |
| | if smask: |
| | obj['SMask'] = smask |
| | if mpl.rcParams['pdf.compression']: |
| | if data.shape[-1] == 1: |
| | data = data.squeeze(axis=-1) |
| | png = {'Predictor': 10, 'Colors': color_channels, 'Columns': width} |
| | img = Image.fromarray(data) |
| | img_colors = img.getcolors(maxcolors=256) |
| | if color_channels == 3 and img_colors is not None: |
| | |
| | |
| | num_colors = len(img_colors) |
| | palette = np.array([comp for _, color in img_colors for comp in color], |
| | dtype=np.uint8) |
| | palette24 = ((palette[0::3].astype(np.uint32) << 16) | |
| | (palette[1::3].astype(np.uint32) << 8) | |
| | palette[2::3]) |
| | rgb24 = ((data[:, :, 0].astype(np.uint32) << 16) | |
| | (data[:, :, 1].astype(np.uint32) << 8) | |
| | data[:, :, 2]) |
| | indices = np.argsort(palette24).astype(np.uint8) |
| | rgb8 = indices[np.searchsorted(palette24, rgb24, sorter=indices)] |
| | img = Image.fromarray(rgb8, mode='P') |
| | img.putpalette(palette) |
| | png_data, bit_depth, palette = self._writePng(img) |
| | if bit_depth is None or palette is None: |
| | raise RuntimeError("invalid PNG header") |
| | palette = palette[:num_colors * 3] |
| | obj['ColorSpace'] = [Name('Indexed'), Name('DeviceRGB'), |
| | num_colors - 1, palette] |
| | obj['BitsPerComponent'] = bit_depth |
| | png['Colors'] = 1 |
| | png['BitsPerComponent'] = bit_depth |
| | else: |
| | png_data, _, _ = self._writePng(img) |
| | else: |
| | png = None |
| | self.beginStream( |
| | id, |
| | self.reserveObject('length of image stream'), |
| | obj, |
| | png=png |
| | ) |
| | if png: |
| | self.currentstream.write(png_data) |
| | else: |
| | self.currentstream.write(data.tobytes()) |
| | self.endStream() |
| |
|
| | def writeImages(self): |
| | for img, name, ob in self._images.values(): |
| | data, adata = self._unpack(img) |
| | if adata is not None: |
| | smaskObject = self.reserveObject("smask") |
| | self._writeImg(adata, smaskObject.id) |
| | else: |
| | smaskObject = None |
| | self._writeImg(data, ob.id, smaskObject) |
| |
|
| | def markerObject(self, path, trans, fill, stroke, lw, joinstyle, |
| | capstyle): |
| | """Return name of a marker XObject representing the given path.""" |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | pathops = self.pathOperations(path, trans, simplify=False) |
| | key = (tuple(pathops), bool(fill), bool(stroke), joinstyle, capstyle) |
| | result = self.markers.get(key) |
| | if result is None: |
| | name = Name('M%d' % len(self.markers)) |
| | ob = self.reserveObject('marker %d' % len(self.markers)) |
| | bbox = path.get_extents(trans) |
| | self.markers[key] = [name, ob, bbox, lw] |
| | else: |
| | if result[-1] < lw: |
| | result[-1] = lw |
| | name = result[0] |
| | return name |
| |
|
| | def writeMarkers(self): |
| | for ((pathops, fill, stroke, joinstyle, capstyle), |
| | (name, ob, bbox, lw)) in self.markers.items(): |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | bbox = bbox.padded(lw * 5) |
| | self.beginStream( |
| | ob.id, None, |
| | {'Type': Name('XObject'), 'Subtype': Name('Form'), |
| | 'BBox': list(bbox.extents)}) |
| | self.output(GraphicsContextPdf.joinstyles[joinstyle], |
| | Op.setlinejoin) |
| | self.output(GraphicsContextPdf.capstyles[capstyle], Op.setlinecap) |
| | self.output(*pathops) |
| | self.output(Op.paint_path(fill, stroke)) |
| | self.endStream() |
| |
|
| | def pathCollectionObject(self, gc, path, trans, padding, filled, stroked): |
| | name = Name('P%d' % len(self.paths)) |
| | ob = self.reserveObject('path %d' % len(self.paths)) |
| | self.paths.append( |
| | (name, path, trans, ob, gc.get_joinstyle(), gc.get_capstyle(), |
| | padding, filled, stroked)) |
| | return name |
| |
|
| | def writePathCollectionTemplates(self): |
| | for (name, path, trans, ob, joinstyle, capstyle, padding, filled, |
| | stroked) in self.paths: |
| | pathops = self.pathOperations(path, trans, simplify=False) |
| | bbox = path.get_extents(trans) |
| | if not np.all(np.isfinite(bbox.extents)): |
| | extents = [0, 0, 0, 0] |
| | else: |
| | bbox = bbox.padded(padding) |
| | extents = list(bbox.extents) |
| | self.beginStream( |
| | ob.id, None, |
| | {'Type': Name('XObject'), 'Subtype': Name('Form'), |
| | 'BBox': extents}) |
| | self.output(GraphicsContextPdf.joinstyles[joinstyle], |
| | Op.setlinejoin) |
| | self.output(GraphicsContextPdf.capstyles[capstyle], Op.setlinecap) |
| | self.output(*pathops) |
| | self.output(Op.paint_path(filled, stroked)) |
| | self.endStream() |
| |
|
| | @staticmethod |
| | def pathOperations(path, transform, clip=None, simplify=None, sketch=None): |
| | return [Verbatim(_path.convert_to_string( |
| | path, transform, clip, simplify, sketch, |
| | 6, |
| | [Op.moveto.value, Op.lineto.value, b'', Op.curveto.value, |
| | Op.closepath.value], |
| | True))] |
| |
|
| | def writePath(self, path, transform, clip=False, sketch=None): |
| | if clip: |
| | clip = (0.0, 0.0, self.width * 72, self.height * 72) |
| | simplify = path.should_simplify |
| | else: |
| | clip = None |
| | simplify = False |
| | cmds = self.pathOperations(path, transform, clip, simplify=simplify, |
| | sketch=sketch) |
| | self.output(*cmds) |
| |
|
| | def reserveObject(self, name=''): |
| | """ |
| | Reserve an ID for an indirect object. |
| | |
| | The name is used for debugging in case we forget to print out |
| | the object with writeObject. |
| | """ |
| | id = next(self._object_seq) |
| | self.xrefTable.append([None, 0, name]) |
| | return Reference(id) |
| |
|
| | def recordXref(self, id): |
| | self.xrefTable[id][0] = self.fh.tell() - self.tell_base |
| |
|
| | def writeObject(self, object, contents): |
| | self.recordXref(object.id) |
| | object.write(contents, self) |
| |
|
| | def writeXref(self): |
| | """Write out the xref table.""" |
| | self.startxref = self.fh.tell() - self.tell_base |
| | self.write(b"xref\n0 %d\n" % len(self.xrefTable)) |
| | for i, (offset, generation, name) in enumerate(self.xrefTable): |
| | if offset is None: |
| | raise AssertionError( |
| | 'No offset for object %d (%s)' % (i, name)) |
| | else: |
| | key = b"f" if name == 'the zero object' else b"n" |
| | text = b"%010d %05d %b \n" % (offset, generation, key) |
| | self.write(text) |
| |
|
| | def writeInfoDict(self): |
| | """Write out the info dictionary, checking it for good form""" |
| |
|
| | self.infoObject = self.reserveObject('info') |
| | self.writeObject(self.infoObject, self.infoDict) |
| |
|
| | def writeTrailer(self): |
| | """Write out the PDF trailer.""" |
| |
|
| | self.write(b"trailer\n") |
| | self.write(pdfRepr( |
| | {'Size': len(self.xrefTable), |
| | 'Root': self.rootObject, |
| | 'Info': self.infoObject})) |
| | |
| | self.write(b"\nstartxref\n%d\n%%%%EOF\n" % self.startxref) |
| |
|
| |
|
| | class RendererPdf(_backend_pdf_ps.RendererPDFPSBase): |
| |
|
| | _afm_font_dir = cbook._get_data_path("fonts/pdfcorefonts") |
| | _use_afm_rc_name = "pdf.use14corefonts" |
| |
|
| | def __init__(self, file, image_dpi, height, width): |
| | super().__init__(width, height) |
| | self.file = file |
| | self.gc = self.new_gc() |
| | self.image_dpi = image_dpi |
| |
|
| | def finalize(self): |
| | self.file.output(*self.gc.finalize()) |
| |
|
| | def check_gc(self, gc, fillcolor=None): |
| | orig_fill = getattr(gc, '_fillcolor', (0., 0., 0.)) |
| | gc._fillcolor = fillcolor |
| |
|
| | orig_alphas = getattr(gc, '_effective_alphas', (1.0, 1.0)) |
| |
|
| | if gc.get_rgb() is None: |
| | |
| | |
| | |
| | gc.set_foreground((0, 0, 0, 0), isRGBA=True) |
| |
|
| | if gc._forced_alpha: |
| | gc._effective_alphas = (gc._alpha, gc._alpha) |
| | elif fillcolor is None or len(fillcolor) < 4: |
| | gc._effective_alphas = (gc._rgb[3], 1.0) |
| | else: |
| | gc._effective_alphas = (gc._rgb[3], fillcolor[3]) |
| |
|
| | delta = self.gc.delta(gc) |
| | if delta: |
| | self.file.output(*delta) |
| |
|
| | |
| | gc._fillcolor = orig_fill |
| | gc._effective_alphas = orig_alphas |
| |
|
| | def get_image_magnification(self): |
| | return self.image_dpi/72.0 |
| |
|
| | def draw_image(self, gc, x, y, im, transform=None): |
| | |
| |
|
| | h, w = im.shape[:2] |
| | if w == 0 or h == 0: |
| | return |
| |
|
| | if transform is None: |
| | |
| | gc.set_alpha(1.0) |
| |
|
| | self.check_gc(gc) |
| |
|
| | w = 72.0 * w / self.image_dpi |
| | h = 72.0 * h / self.image_dpi |
| |
|
| | imob = self.file.imageObject(im) |
| |
|
| | if transform is None: |
| | self.file.output(Op.gsave, |
| | w, 0, 0, h, x, y, Op.concat_matrix, |
| | imob, Op.use_xobject, Op.grestore) |
| | else: |
| | tr1, tr2, tr3, tr4, tr5, tr6 = transform.frozen().to_values() |
| |
|
| | self.file.output(Op.gsave, |
| | 1, 0, 0, 1, x, y, Op.concat_matrix, |
| | tr1, tr2, tr3, tr4, tr5, tr6, Op.concat_matrix, |
| | imob, Op.use_xobject, Op.grestore) |
| |
|
| | def draw_path(self, gc, path, transform, rgbFace=None): |
| | |
| | self.check_gc(gc, rgbFace) |
| | self.file.writePath( |
| | path, transform, |
| | rgbFace is None and gc.get_hatch_path() is None, |
| | gc.get_sketch_params()) |
| | self.file.output(self.gc.paint()) |
| |
|
| | def draw_path_collection(self, gc, master_transform, paths, all_transforms, |
| | offsets, offset_trans, facecolors, edgecolors, |
| | linewidths, linestyles, antialiaseds, urls, |
| | offset_position): |
| | |
| | |
| | |
| | can_do_optimization = True |
| | facecolors = np.asarray(facecolors) |
| | edgecolors = np.asarray(edgecolors) |
| |
|
| | if not len(facecolors): |
| | filled = False |
| | can_do_optimization = not gc.get_hatch() |
| | else: |
| | if np.all(facecolors[:, 3] == facecolors[0, 3]): |
| | filled = facecolors[0, 3] != 0.0 |
| | else: |
| | can_do_optimization = False |
| |
|
| | if not len(edgecolors): |
| | stroked = False |
| | else: |
| | if np.all(np.asarray(linewidths) == 0.0): |
| | stroked = False |
| | elif np.all(edgecolors[:, 3] == edgecolors[0, 3]): |
| | stroked = edgecolors[0, 3] != 0.0 |
| | else: |
| | can_do_optimization = False |
| |
|
| | |
| | |
| | |
| | |
| | len_path = len(paths[0].vertices) if len(paths) > 0 else 0 |
| | uses_per_path = self._iter_collection_uses_per_path( |
| | paths, all_transforms, offsets, facecolors, edgecolors) |
| | should_do_optimization = \ |
| | len_path + uses_per_path + 5 < len_path * uses_per_path |
| |
|
| | if (not can_do_optimization) or (not should_do_optimization): |
| | return RendererBase.draw_path_collection( |
| | self, gc, master_transform, paths, all_transforms, |
| | offsets, offset_trans, facecolors, edgecolors, |
| | linewidths, linestyles, antialiaseds, urls, |
| | offset_position) |
| |
|
| | padding = np.max(linewidths) |
| | path_codes = [] |
| | for i, (path, transform) in enumerate(self._iter_collection_raw_paths( |
| | master_transform, paths, all_transforms)): |
| | name = self.file.pathCollectionObject( |
| | gc, path, transform, padding, filled, stroked) |
| | path_codes.append(name) |
| |
|
| | output = self.file.output |
| | output(*self.gc.push()) |
| | lastx, lasty = 0, 0 |
| | for xo, yo, path_id, gc0, rgbFace in self._iter_collection( |
| | gc, path_codes, offsets, offset_trans, |
| | facecolors, edgecolors, linewidths, linestyles, |
| | antialiaseds, urls, offset_position): |
| |
|
| | self.check_gc(gc0, rgbFace) |
| | dx, dy = xo - lastx, yo - lasty |
| | output(1, 0, 0, 1, dx, dy, Op.concat_matrix, path_id, |
| | Op.use_xobject) |
| | lastx, lasty = xo, yo |
| | output(*self.gc.pop()) |
| |
|
| | def draw_markers(self, gc, marker_path, marker_trans, path, trans, |
| | rgbFace=None): |
| | |
| |
|
| | |
| | len_marker_path = len(marker_path) |
| | uses = len(path) |
| | if len_marker_path * uses < len_marker_path + uses + 5: |
| | RendererBase.draw_markers(self, gc, marker_path, marker_trans, |
| | path, trans, rgbFace) |
| | return |
| |
|
| | self.check_gc(gc, rgbFace) |
| | fill = gc.fill(rgbFace) |
| | stroke = gc.stroke() |
| |
|
| | output = self.file.output |
| | marker = self.file.markerObject( |
| | marker_path, marker_trans, fill, stroke, self.gc._linewidth, |
| | gc.get_joinstyle(), gc.get_capstyle()) |
| |
|
| | output(Op.gsave) |
| | lastx, lasty = 0, 0 |
| | for vertices, code in path.iter_segments( |
| | trans, |
| | clip=(0, 0, self.file.width*72, self.file.height*72), |
| | simplify=False): |
| | if len(vertices): |
| | x, y = vertices[-2:] |
| | if not (0 <= x <= self.file.width * 72 |
| | and 0 <= y <= self.file.height * 72): |
| | continue |
| | dx, dy = x - lastx, y - lasty |
| | output(1, 0, 0, 1, dx, dy, Op.concat_matrix, |
| | marker, Op.use_xobject) |
| | lastx, lasty = x, y |
| | output(Op.grestore) |
| |
|
| | def draw_gouraud_triangle(self, gc, points, colors, trans): |
| | self.draw_gouraud_triangles(gc, points.reshape((1, 3, 2)), |
| | colors.reshape((1, 3, 4)), trans) |
| |
|
| | def draw_gouraud_triangles(self, gc, points, colors, trans): |
| | assert len(points) == len(colors) |
| | if len(points) == 0: |
| | return |
| | assert points.ndim == 3 |
| | assert points.shape[1] == 3 |
| | assert points.shape[2] == 2 |
| | assert colors.ndim == 3 |
| | assert colors.shape[1] == 3 |
| | assert colors.shape[2] in (1, 4) |
| |
|
| | shape = points.shape |
| | points = points.reshape((shape[0] * shape[1], 2)) |
| | tpoints = trans.transform(points) |
| | tpoints = tpoints.reshape(shape) |
| | name, _ = self.file.addGouraudTriangles(tpoints, colors) |
| | output = self.file.output |
| |
|
| | if colors.shape[2] == 1: |
| | |
| | gc.set_alpha(1.0) |
| | self.check_gc(gc) |
| | output(name, Op.shading) |
| | return |
| |
|
| | alpha = colors[0, 0, 3] |
| | if np.allclose(alpha, colors[:, :, 3]): |
| | |
| | gc.set_alpha(alpha) |
| | self.check_gc(gc) |
| | output(name, Op.shading) |
| | else: |
| | |
| | alpha = colors[:, :, 3][:, :, None] |
| | _, smask_ob = self.file.addGouraudTriangles(tpoints, alpha) |
| | gstate = self.file._soft_mask_state(smask_ob) |
| | output(Op.gsave, gstate, Op.setgstate, |
| | name, Op.shading, |
| | Op.grestore) |
| |
|
| | def _setup_textpos(self, x, y, angle, oldx=0, oldy=0, oldangle=0): |
| | if angle == oldangle == 0: |
| | self.file.output(x - oldx, y - oldy, Op.textpos) |
| | else: |
| | angle = math.radians(angle) |
| | self.file.output(math.cos(angle), math.sin(angle), |
| | -math.sin(angle), math.cos(angle), |
| | x, y, Op.textmatrix) |
| | self.file.output(0, 0, Op.textpos) |
| |
|
| | def draw_mathtext(self, gc, x, y, s, prop, angle): |
| | |
| | width, height, descent, glyphs, rects = \ |
| | self._text2path.mathtext_parser.parse(s, 72, prop) |
| |
|
| | if gc.get_url() is not None: |
| | self.file._annotations[-1][1].append(_get_link_annotation( |
| | gc, x, y, width, height, angle)) |
| |
|
| | fonttype = mpl.rcParams['pdf.fonttype'] |
| |
|
| | |
| | a = math.radians(angle) |
| | self.file.output(Op.gsave) |
| | self.file.output(math.cos(a), math.sin(a), |
| | -math.sin(a), math.cos(a), |
| | x, y, Op.concat_matrix) |
| |
|
| | self.check_gc(gc, gc._rgb) |
| | prev_font = None, None |
| | oldx, oldy = 0, 0 |
| | unsupported_chars = [] |
| |
|
| | self.file.output(Op.begin_text) |
| | for font, fontsize, num, ox, oy in glyphs: |
| | self.file._character_tracker.track_glyph(font, num) |
| | fontname = font.fname |
| | if not _font_supports_glyph(fonttype, num): |
| | |
| | |
| | unsupported_chars.append((font, fontsize, ox, oy, num)) |
| | else: |
| | self._setup_textpos(ox, oy, 0, oldx, oldy) |
| | oldx, oldy = ox, oy |
| | if (fontname, fontsize) != prev_font: |
| | self.file.output(self.file.fontName(fontname), fontsize, |
| | Op.selectfont) |
| | prev_font = fontname, fontsize |
| | self.file.output(self.encode_string(chr(num), fonttype), |
| | Op.show) |
| | self.file.output(Op.end_text) |
| |
|
| | for font, fontsize, ox, oy, num in unsupported_chars: |
| | self._draw_xobject_glyph( |
| | font, fontsize, font.get_char_index(num), ox, oy) |
| |
|
| | |
| | for ox, oy, width, height in rects: |
| | self.file.output(Op.gsave, ox, oy, width, height, |
| | Op.rectangle, Op.fill, Op.grestore) |
| |
|
| | |
| | self.file.output(Op.grestore) |
| |
|
| | def draw_tex(self, gc, x, y, s, prop, angle, *, mtext=None): |
| | |
| | texmanager = self.get_texmanager() |
| | fontsize = prop.get_size_in_points() |
| | dvifile = texmanager.make_dvi(s, fontsize) |
| | with dviread.Dvi(dvifile, 72) as dvi: |
| | page, = dvi |
| |
|
| | if gc.get_url() is not None: |
| | self.file._annotations[-1][1].append(_get_link_annotation( |
| | gc, x, y, page.width, page.height, angle)) |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | oldfont, seq = None, [] |
| | for x1, y1, dvifont, glyph, width in page.text: |
| | if dvifont != oldfont: |
| | pdfname = self.file.dviFontName(dvifont) |
| | seq += [['font', pdfname, dvifont.size]] |
| | oldfont = dvifont |
| | seq += [['text', x1, y1, [bytes([glyph])], x1+width]] |
| |
|
| | |
| | |
| | |
| | i, curx, fontsize = 0, 0, None |
| | while i < len(seq)-1: |
| | elt, nxt = seq[i:i+2] |
| | if elt[0] == 'font': |
| | fontsize = elt[2] |
| | elif elt[0] == nxt[0] == 'text' and elt[2] == nxt[2]: |
| | offset = elt[4] - nxt[1] |
| | if abs(offset) < 0.1: |
| | elt[3][-1] += nxt[3][0] |
| | elt[4] += nxt[4]-nxt[1] |
| | else: |
| | elt[3] += [offset*1000.0/fontsize, nxt[3][0]] |
| | elt[4] = nxt[4] |
| | del seq[i+1] |
| | continue |
| | i += 1 |
| |
|
| | |
| | mytrans = Affine2D().rotate_deg(angle).translate(x, y) |
| |
|
| | |
| | self.check_gc(gc, gc._rgb) |
| | self.file.output(Op.begin_text) |
| | curx, cury, oldx, oldy = 0, 0, 0, 0 |
| | for elt in seq: |
| | if elt[0] == 'font': |
| | self.file.output(elt[1], elt[2], Op.selectfont) |
| | elif elt[0] == 'text': |
| | curx, cury = mytrans.transform((elt[1], elt[2])) |
| | self._setup_textpos(curx, cury, angle, oldx, oldy) |
| | oldx, oldy = curx, cury |
| | if len(elt[3]) == 1: |
| | self.file.output(elt[3][0], Op.show) |
| | else: |
| | self.file.output(elt[3], Op.showkern) |
| | else: |
| | assert False |
| | self.file.output(Op.end_text) |
| |
|
| | |
| | |
| | boxgc = self.new_gc() |
| | boxgc.copy_properties(gc) |
| | boxgc.set_linewidth(0) |
| | pathops = [Path.MOVETO, Path.LINETO, Path.LINETO, Path.LINETO, |
| | Path.CLOSEPOLY] |
| | for x1, y1, h, w in page.boxes: |
| | path = Path([[x1, y1], [x1+w, y1], [x1+w, y1+h], [x1, y1+h], |
| | [0, 0]], pathops) |
| | self.draw_path(boxgc, path, mytrans, gc._rgb) |
| |
|
| | def encode_string(self, s, fonttype): |
| | if fonttype in (1, 3): |
| | return s.encode('cp1252', 'replace') |
| | return s.encode('utf-16be', 'replace') |
| |
|
| | def draw_text(self, gc, x, y, s, prop, angle, ismath=False, mtext=None): |
| | |
| |
|
| | |
| |
|
| | self.check_gc(gc, gc._rgb) |
| | if ismath: |
| | return self.draw_mathtext(gc, x, y, s, prop, angle) |
| |
|
| | fontsize = prop.get_size_in_points() |
| |
|
| | if mpl.rcParams['pdf.use14corefonts']: |
| | font = self._get_font_afm(prop) |
| | fonttype = 1 |
| | else: |
| | font = self._get_font_ttf(prop) |
| | self.file._character_tracker.track(font, s) |
| | fonttype = mpl.rcParams['pdf.fonttype'] |
| |
|
| | if gc.get_url() is not None: |
| | font.set_text(s) |
| | width, height = font.get_width_height() |
| | self.file._annotations[-1][1].append(_get_link_annotation( |
| | gc, x, y, width / 64, height / 64, angle)) |
| |
|
| | |
| | |
| | if fonttype not in [3, 42]: |
| | self.file.output(Op.begin_text, |
| | self.file.fontName(prop), fontsize, Op.selectfont) |
| | self._setup_textpos(x, y, angle) |
| | self.file.output(self.encode_string(s, fonttype), |
| | Op.show, Op.end_text) |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | else: |
| | |
| | |
| | singlebyte_chunks = [] |
| | |
| | multibyte_glyphs = [] |
| | prev_was_multibyte = True |
| | prev_font = font |
| | for item in _text_helpers.layout( |
| | s, font, kern_mode=KERNING_UNFITTED): |
| | if _font_supports_glyph(fonttype, ord(item.char)): |
| | if prev_was_multibyte or item.ft_object != prev_font: |
| | singlebyte_chunks.append((item.ft_object, item.x, [])) |
| | prev_font = item.ft_object |
| | if item.prev_kern: |
| | singlebyte_chunks[-1][2].append(item.prev_kern) |
| | singlebyte_chunks[-1][2].append(item.char) |
| | prev_was_multibyte = False |
| | else: |
| | multibyte_glyphs.append( |
| | (item.ft_object, item.x, item.glyph_idx) |
| | ) |
| | prev_was_multibyte = True |
| | |
| | |
| | self.file.output(Op.gsave) |
| | a = math.radians(angle) |
| | self.file.output(math.cos(a), math.sin(a), |
| | -math.sin(a), math.cos(a), |
| | x, y, Op.concat_matrix) |
| | |
| |
|
| | self.file.output(Op.begin_text) |
| | prev_start_x = 0 |
| | for ft_object, start_x, kerns_or_chars in singlebyte_chunks: |
| | ft_name = self.file.fontName(ft_object.fname) |
| | self.file.output(ft_name, fontsize, Op.selectfont) |
| | self._setup_textpos(start_x, 0, 0, prev_start_x, 0, 0) |
| | self.file.output( |
| | |
| | |
| | [-1000 * next(group) / fontsize if tp == float |
| | else self.encode_string("".join(group), fonttype) |
| | for tp, group in itertools.groupby(kerns_or_chars, type)], |
| | Op.showkern) |
| | prev_start_x = start_x |
| | self.file.output(Op.end_text) |
| | |
| | for ft_object, start_x, glyph_idx in multibyte_glyphs: |
| | self._draw_xobject_glyph( |
| | ft_object, fontsize, glyph_idx, start_x, 0 |
| | ) |
| | self.file.output(Op.grestore) |
| |
|
| | def _draw_xobject_glyph(self, font, fontsize, glyph_idx, x, y): |
| | """Draw a multibyte character from a Type 3 font as an XObject.""" |
| | glyph_name = font.get_glyph_name(glyph_idx) |
| | name = self.file._get_xobject_glyph_name(font.fname, glyph_name) |
| | self.file.output( |
| | Op.gsave, |
| | 0.001 * fontsize, 0, 0, 0.001 * fontsize, x, y, Op.concat_matrix, |
| | Name(name), Op.use_xobject, |
| | Op.grestore, |
| | ) |
| |
|
| | def new_gc(self): |
| | |
| | return GraphicsContextPdf(self.file) |
| |
|
| |
|
| | class GraphicsContextPdf(GraphicsContextBase): |
| |
|
| | def __init__(self, file): |
| | super().__init__() |
| | self._fillcolor = (0.0, 0.0, 0.0) |
| | self._effective_alphas = (1.0, 1.0) |
| | self.file = file |
| | self.parent = None |
| |
|
| | def __repr__(self): |
| | d = dict(self.__dict__) |
| | del d['file'] |
| | del d['parent'] |
| | return repr(d) |
| |
|
| | def stroke(self): |
| | """ |
| | Predicate: does the path need to be stroked (its outline drawn)? |
| | This tests for the various conditions that disable stroking |
| | the path, in which case it would presumably be filled. |
| | """ |
| | |
| | |
| | return (self._linewidth > 0 and self._alpha > 0 and |
| | (len(self._rgb) <= 3 or self._rgb[3] != 0.0)) |
| |
|
| | def fill(self, *args): |
| | """ |
| | Predicate: does the path need to be filled? |
| | |
| | An optional argument can be used to specify an alternative |
| | _fillcolor, as needed by RendererPdf.draw_markers. |
| | """ |
| | if len(args): |
| | _fillcolor = args[0] |
| | else: |
| | _fillcolor = self._fillcolor |
| | return (self._hatch or |
| | (_fillcolor is not None and |
| | (len(_fillcolor) <= 3 or _fillcolor[3] != 0.0))) |
| |
|
| | def paint(self): |
| | """ |
| | Return the appropriate pdf operator to cause the path to be |
| | stroked, filled, or both. |
| | """ |
| | return Op.paint_path(self.fill(), self.stroke()) |
| |
|
| | capstyles = {'butt': 0, 'round': 1, 'projecting': 2} |
| | joinstyles = {'miter': 0, 'round': 1, 'bevel': 2} |
| |
|
| | def capstyle_cmd(self, style): |
| | return [self.capstyles[style], Op.setlinecap] |
| |
|
| | def joinstyle_cmd(self, style): |
| | return [self.joinstyles[style], Op.setlinejoin] |
| |
|
| | def linewidth_cmd(self, width): |
| | return [width, Op.setlinewidth] |
| |
|
| | def dash_cmd(self, dashes): |
| | offset, dash = dashes |
| | if dash is None: |
| | dash = [] |
| | offset = 0 |
| | return [list(dash), offset, Op.setdash] |
| |
|
| | def alpha_cmd(self, alpha, forced, effective_alphas): |
| | name = self.file.alphaState(effective_alphas) |
| | return [name, Op.setgstate] |
| |
|
| | def hatch_cmd(self, hatch, hatch_color): |
| | if not hatch: |
| | if self._fillcolor is not None: |
| | return self.fillcolor_cmd(self._fillcolor) |
| | else: |
| | return [Name('DeviceRGB'), Op.setcolorspace_nonstroke] |
| | else: |
| | hatch_style = (hatch_color, self._fillcolor, hatch) |
| | name = self.file.hatchPattern(hatch_style) |
| | return [Name('Pattern'), Op.setcolorspace_nonstroke, |
| | name, Op.setcolor_nonstroke] |
| |
|
| | def rgb_cmd(self, rgb): |
| | if mpl.rcParams['pdf.inheritcolor']: |
| | return [] |
| | if rgb[0] == rgb[1] == rgb[2]: |
| | return [rgb[0], Op.setgray_stroke] |
| | else: |
| | return [*rgb[:3], Op.setrgb_stroke] |
| |
|
| | def fillcolor_cmd(self, rgb): |
| | if rgb is None or mpl.rcParams['pdf.inheritcolor']: |
| | return [] |
| | elif rgb[0] == rgb[1] == rgb[2]: |
| | return [rgb[0], Op.setgray_nonstroke] |
| | else: |
| | return [*rgb[:3], Op.setrgb_nonstroke] |
| |
|
| | def push(self): |
| | parent = GraphicsContextPdf(self.file) |
| | parent.copy_properties(self) |
| | parent.parent = self.parent |
| | self.parent = parent |
| | return [Op.gsave] |
| |
|
| | def pop(self): |
| | assert self.parent is not None |
| | self.copy_properties(self.parent) |
| | self.parent = self.parent.parent |
| | return [Op.grestore] |
| |
|
| | def clip_cmd(self, cliprect, clippath): |
| | """Set clip rectangle. Calls `.pop()` and `.push()`.""" |
| | cmds = [] |
| | |
| | while ((self._cliprect, self._clippath) != (cliprect, clippath) |
| | and self.parent is not None): |
| | cmds.extend(self.pop()) |
| | |
| | if ((self._cliprect, self._clippath) != (cliprect, clippath) or |
| | self.parent is None): |
| | cmds.extend(self.push()) |
| | if self._cliprect != cliprect: |
| | cmds.extend([cliprect, Op.rectangle, Op.clip, Op.endpath]) |
| | if self._clippath != clippath: |
| | path, affine = clippath.get_transformed_path_and_affine() |
| | cmds.extend( |
| | PdfFile.pathOperations(path, affine, simplify=False) + |
| | [Op.clip, Op.endpath]) |
| | return cmds |
| |
|
| | commands = ( |
| | |
| | (('_cliprect', '_clippath'), clip_cmd), |
| | (('_alpha', '_forced_alpha', '_effective_alphas'), alpha_cmd), |
| | (('_capstyle',), capstyle_cmd), |
| | (('_fillcolor',), fillcolor_cmd), |
| | (('_joinstyle',), joinstyle_cmd), |
| | (('_linewidth',), linewidth_cmd), |
| | (('_dashes',), dash_cmd), |
| | (('_rgb',), rgb_cmd), |
| | |
| | (('_hatch', '_hatch_color'), hatch_cmd), |
| | ) |
| |
|
| | def delta(self, other): |
| | """ |
| | Copy properties of other into self and return PDF commands |
| | needed to transform *self* into *other*. |
| | """ |
| | cmds = [] |
| | fill_performed = False |
| | for params, cmd in self.commands: |
| | different = False |
| | for p in params: |
| | ours = getattr(self, p) |
| | theirs = getattr(other, p) |
| | try: |
| | if ours is None or theirs is None: |
| | different = ours is not theirs |
| | else: |
| | different = bool(ours != theirs) |
| | except ValueError: |
| | ours = np.asarray(ours) |
| | theirs = np.asarray(theirs) |
| | different = (ours.shape != theirs.shape or |
| | np.any(ours != theirs)) |
| | if different: |
| | break |
| |
|
| | |
| | if params == ('_hatch', '_hatch_color') and fill_performed: |
| | different = True |
| |
|
| | if different: |
| | if params == ('_fillcolor',): |
| | fill_performed = True |
| | theirs = [getattr(other, p) for p in params] |
| | cmds.extend(cmd(self, *theirs)) |
| | for p in params: |
| | setattr(self, p, getattr(other, p)) |
| | return cmds |
| |
|
| | def copy_properties(self, other): |
| | """ |
| | Copy properties of other into self. |
| | """ |
| | super().copy_properties(other) |
| | fillcolor = getattr(other, '_fillcolor', self._fillcolor) |
| | effective_alphas = getattr(other, '_effective_alphas', |
| | self._effective_alphas) |
| | self._fillcolor = fillcolor |
| | self._effective_alphas = effective_alphas |
| |
|
| | def finalize(self): |
| | """ |
| | Make sure every pushed graphics state is popped. |
| | """ |
| | cmds = [] |
| | while self.parent is not None: |
| | cmds.extend(self.pop()) |
| | return cmds |
| |
|
| |
|
| | class PdfPages: |
| | """ |
| | A multi-page PDF file. |
| | |
| | Examples |
| | -------- |
| | >>> import matplotlib.pyplot as plt |
| | >>> # Initialize: |
| | >>> with PdfPages('foo.pdf') as pdf: |
| | ... # As many times as you like, create a figure fig and save it: |
| | ... fig = plt.figure() |
| | ... pdf.savefig(fig) |
| | ... # When no figure is specified the current figure is saved |
| | ... pdf.savefig() |
| | |
| | Notes |
| | ----- |
| | In reality `PdfPages` is a thin wrapper around `PdfFile`, in order to avoid |
| | confusion when using `~.pyplot.savefig` and forgetting the format argument. |
| | """ |
| | __slots__ = ('_file', 'keep_empty') |
| |
|
| | def __init__(self, filename, keep_empty=True, metadata=None): |
| | """ |
| | Create a new PdfPages object. |
| | |
| | Parameters |
| | ---------- |
| | filename : str or path-like or file-like |
| | Plots using `PdfPages.savefig` will be written to a file at this |
| | location. The file is opened at once and any older file with the |
| | same name is overwritten. |
| | |
| | keep_empty : bool, optional |
| | If set to False, then empty pdf files will be deleted automatically |
| | when closed. |
| | |
| | metadata : dict, optional |
| | Information dictionary object (see PDF reference section 10.2.1 |
| | 'Document Information Dictionary'), e.g.: |
| | ``{'Creator': 'My software', 'Author': 'Me', 'Title': 'Awesome'}``. |
| | |
| | The standard keys are 'Title', 'Author', 'Subject', 'Keywords', |
| | 'Creator', 'Producer', 'CreationDate', 'ModDate', and |
| | 'Trapped'. Values have been predefined for 'Creator', 'Producer' |
| | and 'CreationDate'. They can be removed by setting them to `None`. |
| | """ |
| | self._file = PdfFile(filename, metadata=metadata) |
| | self.keep_empty = keep_empty |
| |
|
| | def __enter__(self): |
| | return self |
| |
|
| | def __exit__(self, exc_type, exc_val, exc_tb): |
| | self.close() |
| |
|
| | def close(self): |
| | """ |
| | Finalize this object, making the underlying file a complete |
| | PDF file. |
| | """ |
| | self._file.finalize() |
| | self._file.close() |
| | if (self.get_pagecount() == 0 and not self.keep_empty and |
| | not self._file.passed_in_file_object): |
| | os.remove(self._file.fh.name) |
| | self._file = None |
| |
|
| | def infodict(self): |
| | """ |
| | Return a modifiable information dictionary object |
| | (see PDF reference section 10.2.1 'Document Information |
| | Dictionary'). |
| | """ |
| | return self._file.infoDict |
| |
|
| | def savefig(self, figure=None, **kwargs): |
| | """ |
| | Save a `.Figure` to this file as a new page. |
| | |
| | Any other keyword arguments are passed to `~.Figure.savefig`. |
| | |
| | Parameters |
| | ---------- |
| | figure : `.Figure` or int, default: the active figure |
| | The figure, or index of the figure, that is saved to the file. |
| | """ |
| | if not isinstance(figure, Figure): |
| | if figure is None: |
| | manager = Gcf.get_active() |
| | else: |
| | manager = Gcf.get_fig_manager(figure) |
| | if manager is None: |
| | raise ValueError("No figure {}".format(figure)) |
| | figure = manager.canvas.figure |
| | |
| | try: |
| | orig_canvas = figure.canvas |
| | figure.canvas = FigureCanvasPdf(figure) |
| | figure.savefig(self, format="pdf", **kwargs) |
| | finally: |
| | figure.canvas = orig_canvas |
| |
|
| | def get_pagecount(self): |
| | """Return the current number of pages in the multipage pdf file.""" |
| | return len(self._file.pageList) |
| |
|
| | def attach_note(self, text, positionRect=[-100, -100, 0, 0]): |
| | """ |
| | Add a new text note to the page to be saved next. The optional |
| | positionRect specifies the position of the new note on the |
| | page. It is outside the page per default to make sure it is |
| | invisible on printouts. |
| | """ |
| | self._file.newTextnote(text, positionRect) |
| |
|
| |
|
| | class FigureCanvasPdf(FigureCanvasBase): |
| | |
| |
|
| | fixed_dpi = 72 |
| | filetypes = {'pdf': 'Portable Document Format'} |
| |
|
| | def get_default_filetype(self): |
| | return 'pdf' |
| |
|
| | def print_pdf(self, filename, *, |
| | bbox_inches_restore=None, metadata=None): |
| |
|
| | dpi = self.figure.dpi |
| | self.figure.dpi = 72 |
| | width, height = self.figure.get_size_inches() |
| | if isinstance(filename, PdfPages): |
| | file = filename._file |
| | else: |
| | file = PdfFile(filename, metadata=metadata) |
| | try: |
| | file.newPage(width, height) |
| | renderer = MixedModeRenderer( |
| | self.figure, width, height, dpi, |
| | RendererPdf(file, dpi, height, width), |
| | bbox_inches_restore=bbox_inches_restore) |
| | self.figure.draw(renderer) |
| | renderer.finalize() |
| | if not isinstance(filename, PdfPages): |
| | file.finalize() |
| | finally: |
| | if isinstance(filename, PdfPages): |
| | file.endStream() |
| | else: |
| | file.close() |
| |
|
| | def draw(self): |
| | self.figure.draw_without_rendering() |
| | return super().draw() |
| |
|
| |
|
| | FigureManagerPdf = FigureManagerBase |
| |
|
| |
|
| | @_Backend.export |
| | class _BackendPdf(_Backend): |
| | FigureCanvas = FigureCanvasPdf |
| |
|