relative_path
stringclasses
812 values
section
stringclasses
339 values
filename
stringlengths
2
61
text
stringlengths
6
1.76M
PyTorch/SpeechSynthesis/Tacotron2/trtis_cpp/src/trt/plugins/taco2ProjectionPlugin
taco2ProjectionPlugin
taco2ProjectionLayerPlugin
/* * Copyright (c) 2019-2020, NVIDIA CORPORATION. All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * Neither the name of the NVIDIA CORPORATION nor the * names of its contributors may be used to endorse or promote products * derived from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL NVIDIA CORPORATION BE LIABLE FOR ANY * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #ifndef TT2I_PROJECTIONLAYERPLUGIN_H #define TT2I_PROJECTIONLAYERPLUGIN_H #include "NvInfer.h" #include <memory> #include <string> #include <vector> namespace nvinfer1 { namespace plugin { class Taco2ProjectionKernel; class Taco2ProjectionLayerPlugin : public nvinfer1::IPluginV2DynamicExt { public: using value_type = float; /** * @brief Get the name of this plugin. * * @return The name. */ static const char* getName(); /** * @brief Get the version of this plugin. * * @return The version. */ static const char* getVersion(); /** * @brief Create a new Taco2ProjectionLayerPlugin from serialized data. * * @param data The data. * @param length The length of the data in bytes. * * @return The instantiated plugin. */ static Taco2ProjectionLayerPlugin deserialize(const void* data, size_t length); /** * @brief Create a new Taco2ProjectionLayerPlugin. * * @param weightsChannel The weights for channel projection. * @param weightsGate The weights for gate projection. * @param biasChannel The bias for channels. * @param biasGate The bias for the gate. * @param hiddenIntputLength The hidden input lentgh. * @param contextIntputLength The context input length. * @param channelDimension The number of channels. * @param gateDimension The number of gates. */ Taco2ProjectionLayerPlugin(const nvinfer1::Weights& weightsChannel, const nvinfer1::Weights& weightsGate, const nvinfer1::Weights& biasChannel, const nvinfer1::Weights& biasGate, int hiddenIntputLength, int contextIntputLength, int channelDimension, int gateDimension); /** * @brief The move constructor. * * @param other The Taco2ProjectionLayerPlugin to move. */ Taco2ProjectionLayerPlugin(Taco2ProjectionLayerPlugin&& other); /** * @brief The move assignment operator. * * @param other The Taco2ProjectionLayerPlugin to move. * * @return This object. */ Taco2ProjectionLayerPlugin& operator=(Taco2ProjectionLayerPlugin&& other); /** * @brief Destructor. */ ~Taco2ProjectionLayerPlugin(); // disable copying Taco2ProjectionLayerPlugin(const Taco2ProjectionLayerPlugin& other) = delete; Taco2ProjectionLayerPlugin& operator=(const Taco2ProjectionLayerPlugin& other) = delete; /** * @brief Return the data type of the plugin output at the requested index. * * @param index The output index. * @param inputTypes The input data types. * @param nbInputs The number of inputs. * * @return The type of output. */ nvinfer1::DataType getOutputDataType(int index, const nvinfer1::DataType* inputTypes, int nbInputs) const override; /** * @brief Get the plugin type. * * @return The plugin type. */ const char* getPluginType() const override; /** * @brief Get the plugin version. * * @return The plugin version. */ const char* getPluginVersion() const override; /** * @brief Get the number of outputs. * * @return The number of outputs. */ int getNbOutputs() const override; /** * @brief Get the dimensions of an output tensor. * * @param outputIndex The index of the output tensor. * @param inputs Expressions for dimensions of the input tensors. * @param nbInputs The number of input tensors. * @param expBuilder Object for generating new expressions. * * @return The resulting dimensions. */ DimsExprs getOutputDimensions( int outputIndex, const DimsExprs* inputs, int nbInputs, IExprBuilder& expBuilder) override; /** * @brief Check if the given plugin format is supported. * * @param pos The format position/index in inOut.format[]. * @param inOut The input and output formats. * @param nbInputs The number of inputs. * @param nbOutputs The number of outputs. * * @return True if it is supported. */ bool supportsFormatCombination(int pos, const PluginTensorDesc* inOut, int nbInputs, int nbOutputs) override; /** * @brief Configure this plugin with the given inputs, outputs, and datat * types. * * @param in The input tensor attributes that used for configuration. * @param nbInputs The number of inputs. * @param out The output tensor attributes that are used for configuration. * @param nbOutputs The number of outputs. */ void configurePlugin( const DynamicPluginTensorDesc* in, int nbInputs, const DynamicPluginTensorDesc* out, int nbOutputs) override; /** * @brief Initialize the plugin. * * @return 0 if initialization was successful. Non-zero otherwise. */ int initialize() override; /** * @brief Terminate the plugin (deinitialize). */ void terminate() override; /** * @brief Get workspace size required by this plugin for up to the given * batch size. * * @param in The input tensor descriptors. * @param nbInputs The number of inputs. * @param out The output tensor descriptors. * @param nbOutputs The number of outputs. * * @return The workspace size in bytes. */ size_t getWorkspaceSize( const PluginTensorDesc* in, int nbInputs, const PluginTensorDesc* out, int nbOutputs) const override; /** * @brief Set this plugin for execution on the stream. * * @param inputDesc The input tensor descriptors. * @param outputDesc The output tensor descriptors. * @param inputs The input tensors. * @param outputs The output tensors. * @param workspace The workspace. * @param stream The stream to operate on. * * @return 0 if successfully queued, non-zero otherwise. */ int enqueue(const PluginTensorDesc* inputDesc, const PluginTensorDesc* outputDesc, const void* const* inputs, void* const* outputs, void* workspace, cudaStream_t stream); /** * @brief Get the number of bytes occupied by this plugin if serialized. * * @return The size in bytes. */ size_t getSerializationSize() const override; /** * @brief Serialize this plugin. * * @param buffer The buffer to write to. */ void serialize(void* buffer) const override; /** * @brief Destroy this plugin instance. */ void destroy() override; /** * @brief Clone this pulgin instance. * * @return The cloned plugin. */ IPluginV2DynamicExt* clone() const override; /** * @brief Set the namespace of this plugin. * * @param pluginNamespace The namespace. */ void setPluginNamespace(const char* pluginNamespace) override; /** * @brief Get the namespace of this plugin. * * @return The namespace. */ const char* getPluginNamespace() const override; private: static const float ONE; static const float ZERO; int mHiddenInputLength; int mContextInputLength; int mNumChannelDimension; int mNumGateDimension; std::vector<value_type> mWeightsChannel; std::vector<value_type> mWeightsGate; std::vector<value_type> mBiasChannel; std::vector<value_type> mBiasGate; std::unique_ptr<Taco2ProjectionKernel> mKernel; std::string mNamespace; int getTotalInputLength() const; int getTotalDimensions() const; }; } // namespace plugin } // namespace nvinfer1 #endif
TensorFlow/Classification/ConvNets/dataprep
dataprep
build_imagenet_data
#!/usr/bin/python # Copyright 2016 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Converts ImageNet data to TFRecords file format with Example protos. The raw ImageNet data set is expected to reside in JPEG files located in the following directory structure. data_dir/n01440764/ILSVRC2012_val_00000293.JPEG data_dir/n01440764/ILSVRC2012_val_00000543.JPEG ... where 'n01440764' is the unique synset label associated with these images. The training data set consists of 1000 sub-directories (i.e. labels) each containing 1200 JPEG images for a total of 1.2M JPEG images. The evaluation data set consists of 1000 sub-directories (i.e. labels) each containing 50 JPEG images for a total of 50K JPEG images. This TensorFlow script converts the training and evaluation data into a sharded data set consisting of 1024 and 128 TFRecord files, respectively. train_directory/train-00000-of-01024 train_directory/train-00001-of-01024 ... train_directory/train-01023-of-01024 and validation_directory/validation-00000-of-00128 validation_directory/validation-00001-of-00128 ... validation_directory/validation-00127-of-00128 Each validation TFRecord file contains ~390 records. Each training TFREcord file contains ~1250 records. Each record within the TFRecord file is a serialized Example proto. The Example proto contains the following fields: image/encoded: string containing JPEG encoded image in RGB colorspace image/height: integer, image height in pixels image/width: integer, image width in pixels image/colorspace: string, specifying the colorspace, always 'RGB' image/channels: integer, specifying the number of channels, always 3 image/format: string, specifying the format, always 'JPEG' image/filename: string containing the basename of the image file e.g. 'n01440764_10026.JPEG' or 'ILSVRC2012_val_00000293.JPEG' image/class/label: integer specifying the index in a classification layer. The label ranges from [1, 1000] where 0 is not used. image/class/synset: string specifying the unique ID of the label, e.g. 'n01440764' image/class/text: string specifying the human-readable version of the label e.g. 'red fox, Vulpes vulpes' image/object/bbox/xmin: list of integers specifying the 0+ human annotated bounding boxes image/object/bbox/xmax: list of integers specifying the 0+ human annotated bounding boxes image/object/bbox/ymin: list of integers specifying the 0+ human annotated bounding boxes image/object/bbox/ymax: list of integers specifying the 0+ human annotated bounding boxes image/object/bbox/label: integer specifying the index in a classification layer. The label ranges from [1, 1000] where 0 is not used. Note this is always identical to the image label. Note that the length of xmin is identical to the length of xmax, ymin and ymax for each example. Running this script using 16 threads may take around ~2.5 hours on an HP Z420. """ from __future__ import absolute_import from __future__ import division from __future__ import print_function from datetime import datetime import os import random import sys import threading import numpy as np import six import tensorflow as tf tf.app.flags.DEFINE_string('train_directory', '/tmp/', 'Training data directory') tf.app.flags.DEFINE_string('validation_directory', '/tmp/', 'Validation data directory') tf.app.flags.DEFINE_string('output_directory', '/tmp/', 'Output data directory') tf.app.flags.DEFINE_integer('train_shards', 1024, 'Number of shards in training TFRecord files.') tf.app.flags.DEFINE_integer('validation_shards', 128, 'Number of shards in validation TFRecord files.') tf.app.flags.DEFINE_integer('num_threads', 8, 'Number of threads to preprocess the images.') # The labels file contains a list of valid labels are held in this file. # Assumes that the file contains entries as such: # n01440764 # n01443537 # n01484850 # where each line corresponds to a label expressed as a synset. We map # each synset contained in the file to an integer (based on the alphabetical # ordering). See below for details. tf.app.flags.DEFINE_string('labels_file', 'imagenet_lsvrc_2015_synsets.txt', 'Labels file') # This file containing mapping from synset to human-readable label. # Assumes each line of the file looks like: # # n02119247 black fox # n02119359 silver fox # n02119477 red fox, Vulpes fulva # # where each line corresponds to a unique mapping. Note that each line is # formatted as <synset>\t<human readable label>. tf.app.flags.DEFINE_string('imagenet_metadata_file', 'imagenet_metadata.txt', 'ImageNet metadata file') # This file is the output of process_bounding_box.py # Assumes each line of the file looks like: # # n00007846_64193.JPEG,0.0060,0.2620,0.7545,0.9940 # # where each line corresponds to one bounding box annotation associated # with an image. Each line can be parsed as: # # <JPEG file name>, <xmin>, <ymin>, <xmax>, <ymax> # # Note that there might exist mulitple bounding box annotations associated # with an image file. tf.app.flags.DEFINE_string('bounding_box_file', './imagenet_2012_bounding_boxes.csv', 'Bounding box file') FLAGS = tf.app.flags.FLAGS def _int64_feature(value): """Wrapper for inserting int64 features into Example proto.""" if not isinstance(value, list): value = [value] return tf.train.Feature(int64_list=tf.train.Int64List(value=value)) def _float_feature(value): """Wrapper for inserting float features into Example proto.""" if not isinstance(value, list): value = [value] return tf.train.Feature(float_list=tf.train.FloatList(value=value)) def _bytes_feature(value): """Wrapper for inserting bytes features into Example proto.""" if six.PY3 and isinstance(value, six.text_type): value = six.binary_type(value, encoding='utf-8') return tf.train.Feature(bytes_list=tf.train.BytesList(value=[value])) def _convert_to_example(filename, image_buffer, label, synset, human, bbox, height, width): """Build an Example proto for an example. Args: filename: string, path to an image file, e.g., '/path/to/example.JPG' image_buffer: string, JPEG encoding of RGB image label: integer, identifier for the ground truth for the network synset: string, unique WordNet ID specifying the label, e.g., 'n02323233' human: string, human-readable label, e.g., 'red fox, Vulpes vulpes' bbox: list of bounding boxes; each box is a list of integers specifying [xmin, ymin, xmax, ymax]. All boxes are assumed to belong to the same label as the image label. height: integer, image height in pixels width: integer, image width in pixels Returns: Example proto """ xmin = [] ymin = [] xmax = [] ymax = [] for b in bbox: assert len(b) == 4 # pylint: disable=expression-not-assigned [l.append(point) for l, point in zip([xmin, ymin, xmax, ymax], b)] # pylint: enable=expression-not-assigned colorspace = 'RGB' channels = 3 image_format = 'JPEG' example = tf.train.Example(features=tf.train.Features(feature={ 'image/height': _int64_feature(height), 'image/width': _int64_feature(width), 'image/colorspace': _bytes_feature(colorspace), 'image/channels': _int64_feature(channels), 'image/class/label': _int64_feature(label), 'image/class/synset': _bytes_feature(synset), 'image/class/text': _bytes_feature(human), 'image/object/bbox/xmin': _float_feature(xmin), 'image/object/bbox/xmax': _float_feature(xmax), 'image/object/bbox/ymin': _float_feature(ymin), 'image/object/bbox/ymax': _float_feature(ymax), 'image/object/bbox/label': _int64_feature([label] * len(xmin)), 'image/format': _bytes_feature(image_format), 'image/filename': _bytes_feature(os.path.basename(filename)), 'image/encoded': _bytes_feature(image_buffer)})) return example class ImageCoder(object): """Helper class that provides TensorFlow image coding utilities.""" def __init__(self): # Create a single Session to run all image coding calls. self._sess = tf.Session() # Initializes function that converts PNG to JPEG data. self._png_data = tf.placeholder(dtype=tf.string) image = tf.image.decode_png(self._png_data, channels=3) self._png_to_jpeg = tf.image.encode_jpeg(image, format='rgb', quality=100) # Initializes function that converts CMYK JPEG data to RGB JPEG data. self._cmyk_data = tf.placeholder(dtype=tf.string) image = tf.image.decode_jpeg(self._cmyk_data, channels=0) self._cmyk_to_rgb = tf.image.encode_jpeg(image, format='rgb', quality=100) # Initializes function that decodes RGB JPEG data. self._decode_jpeg_data = tf.placeholder(dtype=tf.string) self._decode_jpeg = tf.image.decode_jpeg(self._decode_jpeg_data, channels=3) def png_to_jpeg(self, image_data): return self._sess.run(self._png_to_jpeg, feed_dict={self._png_data: image_data}) def cmyk_to_rgb(self, image_data): return self._sess.run(self._cmyk_to_rgb, feed_dict={self._cmyk_data: image_data}) def decode_jpeg(self, image_data): image = self._sess.run(self._decode_jpeg, feed_dict={self._decode_jpeg_data: image_data}) assert len(image.shape) == 3 assert image.shape[2] == 3 return image def _is_png(filename): """Determine if a file contains a PNG format image. Args: filename: string, path of the image file. Returns: boolean indicating if the image is a PNG. """ # File list from: # https://groups.google.com/forum/embed/?place=forum/torch7#!topic/torch7/fOSTXHIESSU return 'n02105855_2933.JPEG' in filename def _is_cmyk(filename): """Determine if file contains a CMYK JPEG format image. Args: filename: string, path of the image file. Returns: boolean indicating if the image is a JPEG encoded with CMYK color space. """ # File list from: # https://github.com/cytsai/ilsvrc-cmyk-image-list blacklist = ['n01739381_1309.JPEG', 'n02077923_14822.JPEG', 'n02447366_23489.JPEG', 'n02492035_15739.JPEG', 'n02747177_10752.JPEG', 'n03018349_4028.JPEG', 'n03062245_4620.JPEG', 'n03347037_9675.JPEG', 'n03467068_12171.JPEG', 'n03529860_11437.JPEG', 'n03544143_17228.JPEG', 'n03633091_5218.JPEG', 'n03710637_5125.JPEG', 'n03961711_5286.JPEG', 'n04033995_2932.JPEG', 'n04258138_17003.JPEG', 'n04264628_27969.JPEG', 'n04336792_7448.JPEG', 'n04371774_5854.JPEG', 'n04596742_4225.JPEG', 'n07583066_647.JPEG', 'n13037406_4650.JPEG'] return filename.split('/')[-1] in blacklist def _process_image(filename, coder): """Process a single image file. Args: filename: string, path to an image file e.g., '/path/to/example.JPG'. coder: instance of ImageCoder to provide TensorFlow image coding utils. Returns: image_buffer: string, JPEG encoding of RGB image. height: integer, image height in pixels. width: integer, image width in pixels. """ # Read the image file. with tf.gfile.FastGFile(filename, 'rb') as f: image_data = f.read() # Clean the dirty data. if _is_png(filename): # 1 image is a PNG. print('Converting PNG to JPEG for %s' % filename) image_data = coder.png_to_jpeg(image_data) elif _is_cmyk(filename): # 22 JPEG images are in CMYK colorspace. print('Converting CMYK to RGB for %s' % filename) image_data = coder.cmyk_to_rgb(image_data) # Decode the RGB JPEG. image = coder.decode_jpeg(image_data) # Check that image converted to RGB assert len(image.shape) == 3 height = image.shape[0] width = image.shape[1] assert image.shape[2] == 3 return image_data, height, width def _process_image_files_batch(coder, thread_index, ranges, name, filenames, synsets, labels, humans, bboxes, num_shards): """Processes and saves list of images as TFRecord in 1 thread. Args: coder: instance of ImageCoder to provide TensorFlow image coding utils. thread_index: integer, unique batch to run index is within [0, len(ranges)). ranges: list of pairs of integers specifying ranges of each batches to analyze in parallel. name: string, unique identifier specifying the data set filenames: list of strings; each string is a path to an image file synsets: list of strings; each string is a unique WordNet ID labels: list of integer; each integer identifies the ground truth humans: list of strings; each string is a human-readable label bboxes: list of bounding boxes for each image. Note that each entry in this list might contain from 0+ entries corresponding to the number of bounding box annotations for the image. num_shards: integer number of shards for this data set. """ # Each thread produces N shards where N = int(num_shards / num_threads). # For instance, if num_shards = 128, and the num_threads = 2, then the first # thread would produce shards [0, 64). num_threads = len(ranges) assert not num_shards % num_threads num_shards_per_batch = int(num_shards / num_threads) shard_ranges = np.linspace(ranges[thread_index][0], ranges[thread_index][1], num_shards_per_batch + 1).astype(int) num_files_in_thread = ranges[thread_index][1] - ranges[thread_index][0] counter = 0 for s in range(num_shards_per_batch): # Generate a sharded version of the file name, e.g. 'train-00002-of-00010' shard = thread_index * num_shards_per_batch + s output_filename = '%s-%.5d-of-%.5d' % (name, shard, num_shards) output_file = os.path.join(FLAGS.output_directory, output_filename) writer = tf.python_io.TFRecordWriter(output_file) shard_counter = 0 files_in_shard = np.arange(shard_ranges[s], shard_ranges[s + 1], dtype=int) for i in files_in_shard: filename = filenames[i] label = labels[i] synset = synsets[i] human = humans[i] bbox = bboxes[i] image_buffer, height, width = _process_image(filename, coder) example = _convert_to_example(filename, image_buffer, label, synset, human, bbox, height, width) writer.write(example.SerializeToString()) shard_counter += 1 counter += 1 if not counter % 1000: print('%s [thread %d]: Processed %d of %d images in thread batch.' % (datetime.now(), thread_index, counter, num_files_in_thread)) sys.stdout.flush() writer.close() print('%s [thread %d]: Wrote %d images to %s' % (datetime.now(), thread_index, shard_counter, output_file)) sys.stdout.flush() shard_counter = 0 print('%s [thread %d]: Wrote %d images to %d shards.' % (datetime.now(), thread_index, counter, num_files_in_thread)) sys.stdout.flush() def _process_image_files(name, filenames, synsets, labels, humans, bboxes, num_shards): """Process and save list of images as TFRecord of Example protos. Args: name: string, unique identifier specifying the data set filenames: list of strings; each string is a path to an image file synsets: list of strings; each string is a unique WordNet ID labels: list of integer; each integer identifies the ground truth humans: list of strings; each string is a human-readable label bboxes: list of bounding boxes for each image. Note that each entry in this list might contain from 0+ entries corresponding to the number of bounding box annotations for the image. num_shards: integer number of shards for this data set. """ assert len(filenames) == len(synsets) assert len(filenames) == len(labels) assert len(filenames) == len(humans) assert len(filenames) == len(bboxes) # Break all images into batches with a [ranges[i][0], ranges[i][1]]. spacing = np.linspace(0, len(filenames), FLAGS.num_threads + 1).astype(np.int) ranges = [] threads = [] for i in range(len(spacing) - 1): ranges.append([spacing[i], spacing[i + 1]]) # Launch a thread for each batch. print('Launching %d threads for spacings: %s' % (FLAGS.num_threads, ranges)) sys.stdout.flush() # Create a mechanism for monitoring when all threads are finished. coord = tf.train.Coordinator() # Create a generic TensorFlow-based utility for converting all image codings. coder = ImageCoder() threads = [] for thread_index in range(len(ranges)): args = (coder, thread_index, ranges, name, filenames, synsets, labels, humans, bboxes, num_shards) t = threading.Thread(target=_process_image_files_batch, args=args) t.start() threads.append(t) # Wait for all the threads to terminate. coord.join(threads) print('%s: Finished writing all %d images in data set.' % (datetime.now(), len(filenames))) sys.stdout.flush() def _find_image_files(data_dir, labels_file): """Build a list of all images files and labels in the data set. Args: data_dir: string, path to the root directory of images. Assumes that the ImageNet data set resides in JPEG files located in the following directory structure. data_dir/n01440764/ILSVRC2012_val_00000293.JPEG data_dir/n01440764/ILSVRC2012_val_00000543.JPEG where 'n01440764' is the unique synset label associated with these images. labels_file: string, path to the labels file. The list of valid labels are held in this file. Assumes that the file contains entries as such: n01440764 n01443537 n01484850 where each line corresponds to a label expressed as a synset. We map each synset contained in the file to an integer (based on the alphabetical ordering) starting with the integer 1 corresponding to the synset contained in the first line. The reason we start the integer labels at 1 is to reserve label 0 as an unused background class. Returns: filenames: list of strings; each string is a path to an image file. synsets: list of strings; each string is a unique WordNet ID. labels: list of integer; each integer identifies the ground truth. """ print('Determining list of input files and labels from %s.' % data_dir) challenge_synsets = [l.strip() for l in tf.gfile.FastGFile(labels_file, 'r').readlines()] labels = [] filenames = [] synsets = [] # Leave label index 0 empty as a background class. label_index = 1 # Construct the list of JPEG files and labels. for synset in challenge_synsets: jpeg_file_path = '%s/%s/*.JPEG' % (data_dir, synset) matching_files = tf.gfile.Glob(jpeg_file_path) labels.extend([label_index] * len(matching_files)) synsets.extend([synset] * len(matching_files)) filenames.extend(matching_files) if not label_index % 100: print('Finished finding files in %d of %d classes.' % ( label_index, len(challenge_synsets))) label_index += 1 # Shuffle the ordering of all image files in order to guarantee # random ordering of the images with respect to label in the # saved TFRecord files. Make the randomization repeatable. shuffled_index = list(range(len(filenames))) random.seed(12345) random.shuffle(shuffled_index) filenames = [filenames[i] for i in shuffled_index] synsets = [synsets[i] for i in shuffled_index] labels = [labels[i] for i in shuffled_index] print('Found %d JPEG files across %d labels inside %s.' % (len(filenames), len(challenge_synsets), data_dir)) return filenames, synsets, labels def _find_human_readable_labels(synsets, synset_to_human): """Build a list of human-readable labels. Args: synsets: list of strings; each string is a unique WordNet ID. synset_to_human: dict of synset to human labels, e.g., 'n02119022' --> 'red fox, Vulpes vulpes' Returns: List of human-readable strings corresponding to each synset. """ humans = [] for s in synsets: assert s in synset_to_human, ('Failed to find: %s' % s) humans.append(synset_to_human[s]) return humans def _find_image_bounding_boxes(filenames, image_to_bboxes): """Find the bounding boxes for a given image file. Args: filenames: list of strings; each string is a path to an image file. image_to_bboxes: dictionary mapping image file names to a list of bounding boxes. This list contains 0+ bounding boxes. Returns: List of bounding boxes for each image. Note that each entry in this list might contain from 0+ entries corresponding to the number of bounding box annotations for the image. """ num_image_bbox = 0 bboxes = [] for f in filenames: basename = os.path.basename(f) if basename in image_to_bboxes: bboxes.append(image_to_bboxes[basename]) num_image_bbox += 1 else: bboxes.append([]) print('Found %d images with bboxes out of %d images' % ( num_image_bbox, len(filenames))) return bboxes def _process_dataset(name, directory, num_shards, synset_to_human, image_to_bboxes): """Process a complete data set and save it as a TFRecord. Args: name: string, unique identifier specifying the data set. directory: string, root path to the data set. num_shards: integer number of shards for this data set. synset_to_human: dict of synset to human labels, e.g., 'n02119022' --> 'red fox, Vulpes vulpes' image_to_bboxes: dictionary mapping image file names to a list of bounding boxes. This list contains 0+ bounding boxes. """ filenames, synsets, labels = _find_image_files(directory, FLAGS.labels_file) humans = _find_human_readable_labels(synsets, synset_to_human) bboxes = _find_image_bounding_boxes(filenames, image_to_bboxes) _process_image_files(name, filenames, synsets, labels, humans, bboxes, num_shards) def _build_synset_lookup(imagenet_metadata_file): """Build lookup for synset to human-readable label. Args: imagenet_metadata_file: string, path to file containing mapping from synset to human-readable label. Assumes each line of the file looks like: n02119247 black fox n02119359 silver fox n02119477 red fox, Vulpes fulva where each line corresponds to a unique mapping. Note that each line is formatted as <synset>\t<human readable label>. Returns: Dictionary of synset to human labels, such as: 'n02119022' --> 'red fox, Vulpes vulpes' """ lines = tf.gfile.FastGFile(imagenet_metadata_file, 'r').readlines() synset_to_human = {} for l in lines: if l: parts = l.strip().split('\t') assert len(parts) == 2 synset = parts[0] human = parts[1] synset_to_human[synset] = human return synset_to_human def _build_bounding_box_lookup(bounding_box_file): """Build a lookup from image file to bounding boxes. Args: bounding_box_file: string, path to file with bounding boxes annotations. Assumes each line of the file looks like: n00007846_64193.JPEG,0.0060,0.2620,0.7545,0.9940 where each line corresponds to one bounding box annotation associated with an image. Each line can be parsed as: <JPEG file name>, <xmin>, <ymin>, <xmax>, <ymax> Note that there might exist mulitple bounding box annotations associated with an image file. This file is the output of process_bounding_boxes.py. Returns: Dictionary mapping image file names to a list of bounding boxes. This list contains 0+ bounding boxes. """ lines = tf.gfile.FastGFile(bounding_box_file, 'r').readlines() images_to_bboxes = {} num_bbox = 0 num_image = 0 for l in lines: if l: parts = l.split(',') assert len(parts) == 5, ('Failed to parse: %s' % l) filename = parts[0] xmin = float(parts[1]) ymin = float(parts[2]) xmax = float(parts[3]) ymax = float(parts[4]) box = [xmin, ymin, xmax, ymax] if filename not in images_to_bboxes: images_to_bboxes[filename] = [] num_image += 1 images_to_bboxes[filename].append(box) num_bbox += 1 print('Successfully read %d bounding boxes ' 'across %d images.' % (num_bbox, num_image)) return images_to_bboxes def main(unused_argv): assert not FLAGS.train_shards % FLAGS.num_threads, ( 'Please make the FLAGS.num_threads commensurate with FLAGS.train_shards') assert not FLAGS.validation_shards % FLAGS.num_threads, ( 'Please make the FLAGS.num_threads commensurate with ' 'FLAGS.validation_shards') print('Saving results to %s' % FLAGS.output_directory) # Build a map from synset to human-readable label. synset_to_human = _build_synset_lookup(FLAGS.imagenet_metadata_file) image_to_bboxes = _build_bounding_box_lookup(FLAGS.bounding_box_file) # Run it! _process_dataset('validation', FLAGS.validation_directory, FLAGS.validation_shards, synset_to_human, image_to_bboxes) _process_dataset('train', FLAGS.train_directory, FLAGS.train_shards, synset_to_human, image_to_bboxes) if __name__ == '__main__': tf.app.run()
TensorFlow/Translation/GNMT
GNMT
gnmt_model
# Copyright 2017 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== # # Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """GNMT attention sequence-to-sequence model with dynamic RNN support.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import numpy as np import tensorflow as tf from tensorflow.contrib.cudnn_rnn.python.layers import cudnn_rnn import attention_wrapper import block_lstm import model import model_helper from utils import misc_utils as utils class GNMTModel(model.BaseModel): """Sequence-to-sequence dynamic model with GNMT attention architecture. """ def __init__(self, hparams, mode, features, scope=None, extra_args=None): self.is_gnmt_attention = ( hparams.attention_architecture in ["gnmt", "gnmt_v2"]) super(GNMTModel, self).__init__( hparams=hparams, mode=mode, features=features, scope=scope, extra_args=extra_args) def _prepare_beam_search_decoder_inputs( self, beam_width, memory, source_sequence_length, encoder_state): memory = tf.contrib.seq2seq.tile_batch( memory, multiplier=beam_width) source_sequence_length = tf.contrib.seq2seq.tile_batch( source_sequence_length, multiplier=beam_width) encoder_state = tf.contrib.seq2seq.tile_batch( encoder_state, multiplier=beam_width) batch_size = self.batch_size * beam_width return memory, source_sequence_length, encoder_state, batch_size def _build_encoder(self, hparams): """Build a GNMT encoder.""" assert hparams.encoder_type == "gnmt" # Build GNMT encoder. num_bi_layers = 1 num_uni_layers = self.num_encoder_layers - num_bi_layers utils.print_out("# Build a GNMT encoder") utils.print_out(" num_bi_layers = %d" % num_bi_layers) utils.print_out(" num_uni_layers = %d" % num_uni_layers) # source is batch-majored source = self.features["source"] import sys print('source.shape: %s' % source.shape, file=sys.stderr) if self.time_major: # Later rnn would use time-majored inputs source = tf.transpose(source) with tf.variable_scope("encoder"): dtype = self.dtype encoder_emb_inp = tf.cast( self.encoder_emb_lookup_fn(self.embedding_encoder, source), dtype) # Build 1st bidi layer. bi_encoder_outputs, bi_encoder_state = self._build_encoder_layers_bidi( encoder_emb_inp, self.features["source_sequence_length"], hparams, dtype) # Build all the rest unidi layers encoder_state, encoder_outputs = self._build_encoder_layers_unidi( bi_encoder_outputs, self.features["source_sequence_length"], num_uni_layers, hparams, dtype) # Pass all encoder states to the decoder # except the first bi-directional layer encoder_state = (bi_encoder_state[1],) + ( (encoder_state,) if num_uni_layers == 1 else encoder_state) return encoder_outputs, encoder_state def _build_encoder_layers_bidi(self, inputs, sequence_length, hparams, dtype): """docstring.""" if hparams.use_fused_lstm: fn = self._build_bidi_rnn_fused elif hparams.use_cudnn_lstm: fn = self._build_bidi_rnn_cudnn else: fn = self._build_bidi_rnn_base return fn(inputs, sequence_length, hparams, dtype) def _build_bidi_rnn_fused(self, inputs, sequence_length, hparams, dtype): if (not np.isclose(hparams.dropout, 0.) and self.mode == tf.contrib.learn.ModeKeys.TRAIN): inputs = tf.nn.dropout(inputs, keep_prob=1-hparams.dropout) fwd_cell = block_lstm.LSTMBlockFusedCell( hparams.num_units, hparams.forget_bias, dtype=dtype) fwd_encoder_outputs, (fwd_final_c, fwd_final_h) = fwd_cell( inputs, dtype=dtype, sequence_length=sequence_length) inputs_r = tf.reverse_sequence( inputs, sequence_length, batch_axis=1, seq_axis=0) bak_cell = block_lstm.LSTMBlockFusedCell( hparams.num_units, hparams.forget_bias, dtype=dtype) bak_encoder_outputs, (bak_final_c, bak_final_h) = bak_cell( inputs_r, dtype=dtype, sequence_length=sequence_length) bak_encoder_outputs = tf.reverse_sequence( bak_encoder_outputs, sequence_length, batch_axis=1, seq_axis=0) bi_encoder_outputs = tf.concat( [fwd_encoder_outputs, bak_encoder_outputs], axis=-1) fwd_state = tf.nn.rnn_cell.LSTMStateTuple(fwd_final_c, fwd_final_h) bak_state = tf.nn.rnn_cell.LSTMStateTuple(bak_final_c, bak_final_h) bi_encoder_state = (fwd_state, bak_state) # mask aren't applied on outputs, but final states are post-masking. return bi_encoder_outputs, bi_encoder_state def _build_unidi_rnn_fused(self, inputs, state, sequence_length, hparams, dtype): if (not np.isclose(hparams.dropout, 0.) and self.mode == tf.contrib.learn.ModeKeys.TRAIN): inputs = tf.nn.dropout(inputs, keep_prob=1-hparams.dropout) cell = block_lstm.LSTMBlockFusedCell( hparams.num_units, hparams.forget_bias, dtype=dtype) outputs, (final_c, final_h) = cell( inputs, state, dtype=dtype, sequence_length=sequence_length) # mask aren't applied on outputs, but final states are post-masking. return outputs, tf.nn.rnn_cell.LSTMStateTuple(final_c, final_h) def _build_unidi_rnn_cudnn(self, inputs, state, sequence_length, dtype, hparams, num_layers, is_fwd): # cudnn inputs only support time-major if not self.time_major: inputs = tf.transpose(inputs, axis=[1, 0, 2]) if num_layers == 1 and not np.isclose(hparams.dropout, 0.): # Special case when drop is used and only one layer dropout = 0. inputs = tf.nn.dropout(inputs, keep_prob=1-dropout) else: dropout = hparams.dropout # the outputs would be in time-majored sequence_length = tf.transpose(sequence_length) if not is_fwd: inputs = tf.reverse_sequence( inputs, sequence_length, batch_axis=1, seq_axis=0) cell = tf.contrib.cudnn_rnn.CudnnLSTM( num_layers=num_layers, num_units=hparams.num_units, direction=cudnn_rnn.CUDNN_RNN_UNIDIRECTION, dtype=self.dtype, dropout=dropout) outputs, (h, c) = cell(inputs, initial_state=state) """ # Mask outputs # [batch, time] mask = tf.sequence_mask(sequence_length, dtype=self.dtype) # [time, batch] mask = tf.transpose(mask) outputs *= mask """ if not is_fwd: outputs = tf.reverse_sequence( inputs, sequence_length, batch_axis=1, seq_axis=0) # NOTICE! There's no way to get the "correct" masked cell state in cudnn # rnn. if num_layers == 1: h = tf.squeeze(h, axis=0) c = tf.squeeze(c, axis=0) return outputs, tf.nn.rnn_cell.LSTMStateTuple(c=c, h=h) # Split h and c to form a h.set_shape((num_layers, None, hparams.num_units)) c.set_shape((num_layers, None, hparams.num_units)) hs = tf.unstack(h) cs = tf.unstack(c) # The cell passed to bidi-dyanmic-rnn is a MultiRNNCell consisting 2 regular # LSTM, the state of each is a simple LSTMStateTuple. Thus the state of the # MultiRNNCell is a tuple of LSTMStateTuple. states = tuple( tf.nn.rnn_cell.LSTMStateTuple(c=c, h=h) for h, c in zip(hs, cs)) # No need to transpose back return outputs, states def _build_encoder_cell(self, hparams, num_layers, num_residual_layers, dtype=None): """Build a multi-layer RNN cell that can be used by encoder.""" return model_helper.create_rnn_cell( unit_type=hparams.unit_type, num_units=self.num_units, num_layers=num_layers, num_residual_layers=num_residual_layers, forget_bias=hparams.forget_bias, dropout=hparams.dropout, mode=self.mode, dtype=dtype, single_cell_fn=self.single_cell_fn, use_block_lstm=hparams.use_block_lstm) def _build_bidi_rnn_base(self, inputs, sequence_length, hparams, dtype): """Create and call biddirectional RNN cells.""" # num_residual_layers: Number of residual layers from top to bottom. For # example, if `num_bi_layers=4` and `num_residual_layers=2`, the last 2 # RNN layers in each RNN cell will be wrapped with `ResidualWrapper`. # Construct forward and backward cells fw_cell = self._build_encoder_cell(hparams, 1, # num_bi_layers, 0, # num_bi_residual_layers, dtype) bw_cell = self._build_encoder_cell(hparams, 1, # num_bi_layers, 0, # num_bi_residual_layers, dtype) if hparams.use_dynamic_rnn: bi_outputs, bi_state = tf.nn.bidirectional_dynamic_rnn( fw_cell, bw_cell, inputs, dtype=dtype, sequence_length=sequence_length, time_major=self.time_major, swap_memory=True) else: bi_outputs, bi_state = tf.contrib.recurrent.bidirectional_functional_rnn( fw_cell, bw_cell, inputs, dtype=dtype, sequence_length=sequence_length, time_major=self.time_major, use_tpu=False) return tf.concat(bi_outputs, -1), bi_state def _build_bidi_rnn_cudnn(self, inputs, sequence_length, hparams, dtype): # Notice cudnn rnn dropout is applied between layers. (if 1 layer only then # no dropout). if not np.isclose(hparams.dropout, 0.): inputs = tf.nn.dropout(inputs, keep_prob=1-hparams.dropout) if not hparams.use_loose_bidi_cudnn_lstm: fwd_outputs, fwd_states = self._build_unidi_rnn_cudnn( inputs, None, # initial_state sequence_length, dtype, hparams, 1, # num_layer is_fwd=True) bak_outputs, bak_states = self._build_unidi_rnn_cudnn( inputs, None, # initial_state sequence_length, dtype, hparams, 1, # num_layer is_fwd=False) bi_outputs = tf.concat([fwd_outputs, bak_outputs], axis=-1) return bi_outputs, (fwd_states, bak_states) else: # Cudnn only accept time-majored inputs if not self.time_major: inputs = tf.transpose(inputs, axis=[1, 0, 2]) bi_outputs, (bi_h, bi_c) = tf.contrib.cudnn_rnn.CudnnLSTM( num_layers=1, # num_bi_layers, num_units=hparams.num_units, direction=cudnn_rnn.CUDNN_RNN_BIDIRECTION, dropout=0., # one layer, dropout isn't applied anyway, seed=hparams.random_seed, dtype=self.dtype, kernel_initializer=tf.get_variable_scope().initializer, bias_initializer=tf.zeros_initializer())(inputs) # state shape is [num_layers * num_dir, batch, dim] bi_h.set_shape((2, None, hparams.num_units)) bi_c.set_shape((2, None, hparams.num_units)) fwd_h, bak_h = tf.unstack(bi_h) fwd_c, bak_c = tf.unstack(bi_c) # No need to transpose back return bi_outputs, (tf.nn.rnn_cell.LSTMStateTuple(c=fwd_c, h=fwd_h), tf.nn.rnn_cell.LSTMStateTuple(c=bak_c, h=bak_h)) def _build_encoder_layers_unidi(self, inputs, sequence_length, num_uni_layers, hparams, dtype): """Build encoder layers all at once.""" encoder_outputs = None encoder_state = tuple() if hparams.use_fused_lstm: for i in range(num_uni_layers): if (not np.isclose(hparams.dropout, 0.) and self.mode == tf.contrib.learn.ModeKeys.TRAIN): cell_inputs = tf.nn.dropout(inputs, keep_prob=1-hparams.dropout) else: cell_inputs = inputs cell = block_lstm.LSTMBlockFusedCell( hparams.num_units, hparams.forget_bias, dtype=dtype) encoder_outputs, (final_c, final_h) = cell( cell_inputs, dtype=dtype, sequence_length=sequence_length) encoder_state += (tf.nn.rnn_cell.LSTMStateTuple(final_c, final_h),) if i >= num_uni_layers - self.num_encoder_residual_layers: # Add the pre-dropout inputs. Residual wrapper is applied after # dropout wrapper. encoder_outputs += inputs inputs = encoder_outputs elif hparams.use_cudnn_lstm: # Single layer cudnn rnn, dropout isnt applied in the kernel for i in range(num_uni_layers): if (not np.isclose(hparams.dropout, 0.) and self.mode == tf.contrib.learn.ModeKeys.TRAIN): inputs = tf.nn.dropout(inputs, keep_prob=1-hparams.dropout) encoder_outputs, encoder_states = self._build_unidi_rnn_cudnn( inputs, None, # initial_state sequence_length, dtype, hparams, 1, # num_layer is_fwd=True) encoder_state += (tf.nn.rnn_cell.LSTMStateTuple(encoder_states.c, encoder_states.h),) if i >= num_uni_layers - self.num_encoder_residual_layers: encoder_outputs += inputs inputs = encoder_outputs else: uni_cell = model_helper.create_rnn_cell( unit_type=hparams.unit_type, num_units=hparams.num_units, num_layers=num_uni_layers, num_residual_layers=self.num_encoder_residual_layers, forget_bias=hparams.forget_bias, dropout=hparams.dropout, dtype=dtype, mode=self.mode, single_cell_fn=self.single_cell_fn, use_block_lstm=hparams.use_block_lstm) if hparams.use_dynamic_rnn: encoder_outputs, encoder_state = tf.nn.dynamic_rnn( uni_cell, inputs, dtype=dtype, sequence_length=sequence_length, time_major=self.time_major) else: encoder_outputs, encoder_state = tf.contrib.recurrent.functional_rnn( uni_cell, inputs, dtype=dtype, sequence_length=sequence_length, time_major=self.time_major, use_tpu=False) return encoder_state, encoder_outputs def _build_decoder_cell(self, hparams, encoder_outputs, encoder_state, source_sequence_length): """Build a RNN cell with GNMT attention architecture.""" # GNMT attention assert self.is_gnmt_attention attention_option = hparams.attention attention_architecture = hparams.attention_architecture assert attention_option == "normed_bahdanau" assert attention_architecture == "gnmt_v2" num_units = hparams.num_units infer_mode = hparams.infer_mode dtype = tf.float16 if hparams.use_fp16 else tf.float32 if self.time_major: memory = tf.transpose(encoder_outputs, [1, 0, 2]) else: memory = encoder_outputs if (self.mode == tf.contrib.learn.ModeKeys.INFER and infer_mode == "beam_search"): memory, source_sequence_length, encoder_state, batch_size = ( self._prepare_beam_search_decoder_inputs( hparams.beam_width, memory, source_sequence_length, encoder_state)) else: batch_size = self.batch_size attention_mechanism = model.create_attention_mechanism( num_units, memory, source_sequence_length, dtype=dtype) cell_list = model_helper._cell_list( # pylint: disable=protected-access unit_type=hparams.unit_type, num_units=num_units, num_layers=self.num_decoder_layers, num_residual_layers=self.num_decoder_residual_layers, forget_bias=hparams.forget_bias, dropout=hparams.dropout, mode=self.mode, dtype=dtype, single_cell_fn=self.single_cell_fn, residual_fn=gnmt_residual_fn, use_block_lstm=hparams.use_block_lstm) # Only wrap the bottom layer with the attention mechanism. attention_cell = cell_list.pop(0) # Only generate alignment in greedy INFER mode. alignment_history = (self.mode == tf.contrib.learn.ModeKeys.INFER and infer_mode != "beam_search") attention_cell = attention_wrapper.AttentionWrapper( attention_cell, attention_mechanism, attention_layer_size=None, # don't use attention layer. output_attention=False, alignment_history=alignment_history, name="attention") cell = GNMTAttentionMultiCell(attention_cell, cell_list) if hparams.pass_hidden_state: decoder_initial_state = tuple( zs.clone(cell_state=es) if isinstance(zs, attention_wrapper.AttentionWrapperState) else es for zs, es in zip( cell.zero_state(batch_size, dtype), encoder_state)) else: decoder_initial_state = cell.zero_state(batch_size, dtype) return cell, decoder_initial_state def _build_decoder_cudnn(self, encoder_outputs, encoder_state, hparams): pass """ # Training # Use dynamic_rnn to compute the 1st layer outputs and attention # GNMT attention with tf.variable_scope("decoder") as decoder_scope: assert self.is_gnmt_attention attention_option = hparams.attention attention_architecture = hparams.attention_architecture assert attention_option == "normed_bahdanau" assert attention_architecture == "gnmt_v2" num_units = hparams.num_units infer_mode = hparams.infer_mode dtype = tf.float16 if hparams.use_fp16 else tf.float32 if self.time_major: memory = tf.transpose(encoder_outputs, [1, 0, 2]) else: memory = encoder_outputs source_sequence_length = self.features["source_sequence_length"] if (self.mode == tf.contrib.learn.ModeKeys.INFER and infer_mode == "beam_search"): memory, source_sequence_length, encoder_state, batch_size = ( self._prepare_beam_search_decoder_inputs( hparams.beam_width, memory, source_sequence_length, encoder_state)) else: batch_size = self.batch_size attention_mechanism = model.create_attention_mechanism( num_units, memory, source_sequence_length, dtype=dtype) attention_cell = model_helper._cell_list( # pylint: disable=protected-access unit_type=hparams.unit_type, num_units=num_units, num_layers=1, # just one layer num_residual_layers=0, # 1st layer has no residual connection. forget_bias=hparams.forget_bias, dropout=hparams.dropout, mode=self.mode, dtype=dtype, single_cell_fn=self.single_cell_fn, residual_fn=gnmt_residual_fn, use_block_lstm=False)[0] # Only generate alignment in greedy INFER mode. alignment_history = (self.mode == tf.contrib.learn.ModeKeys.INFER and infer_mode != "beam_search") attention_cell = attention_wrapper.AttentionWrapper( attention_cell, attention_mechanism, attention_layer_size=None, # don't use attention layer. output_attention=False, alignment_history=alignment_history, name="attention") decoder_attention_cell_initial_state = attention_cell.zero_state( batch_size, dtype).clone(cell_state=encoder_state[0]) # TODO(jamesqin): support frnn # [batch, time] target_input = self.features["target_input"] if self.time_major: # If using time_major mode, then target_input should be [time, batch] # then the decoder_emb_inp would be [time, batch, dim] target_input = tf.transpose(target_input) decoder_emb_inp = tf.cast( tf.nn.embedding_lookup(self.embedding_decoder, target_input), self.dtype) attention_cell_outputs, attention_cell_state = tf.nn.dynamic_rnn( attention_cell, decoder_emb_inp, sequence_length=self.features["target_sequence_length"], initial_state=decoder_attention_cell_initial_state, dtype=self.dtype, scope=decoder_scope, parallel_iterations=hparams.parallel_iterations, time_major=self.time_major) attention = None inputs = tf.concat([target_input, attention_cell_outputs], axis=-1) initial_state = encoder_state[1:] num_bi_layers = 1 num_unidi_decoder_layers = self.num_decoder_layers = num_bi_layers # 3 layers of uni cudnn for i in range(num_unidi_decoder_layers): # Concat input with attention if (not np.isclose(hparams.dropout, 0.) and self.mode == tf.contrib.learn.ModeKeys.TRAIN): inputs = tf.nn.dropout(inputs, keep_prob=1 - hparams.dropout) outputs, states = self._build_unidi_rnn_cudnn( inputs, initial_state[i], self.features["target_sequence_length"], self.dtype, hparams, 1, # num_layer is_fwd=True) if i >= num_unidi_decoder_layers - self.num_decoder_residual_layers: outputs += inputs inputs = outputs pass """ def _build_decoder_fused_for_training(self, encoder_outputs, initial_state, decoder_emb_inp, hparams): assert self.mode == tf.contrib.learn.ModeKeys.TRAIN num_bi_layers = 1 num_unidi_decoder_layers = self.num_decoder_layers - num_bi_layers assert num_unidi_decoder_layers == 3 # The 1st LSTM layer if self.time_major: batch = tf.shape(encoder_outputs)[1] tgt_max_len = tf.shape(decoder_emb_inp)[0] # [batch_size] -> scalar initial_attention = tf.zeros( shape=[tgt_max_len, batch, hparams.num_units], dtype=self.dtype) else: batch = tf.shape(encoder_outputs)[0] tgt_max_len = tf.shape(decoder_emb_inp)[1] initial_attention = tf.zeros( shape=[batch, tgt_max_len, hparams.num_units], dtype=self.dtype) # Concat with initial attention dec_inp = tf.concat([decoder_emb_inp, initial_attention], axis=-1) # [tgt_time, batch, units] # var_scope naming chosen to agree with inference graph. with tf.variable_scope("multi_rnn_cell/cell_0_attention/attention"): outputs, _ = self._build_unidi_rnn_fused( dec_inp, initial_state[0], self.features["target_sequence_length"], hparams, self.dtype) # Get attention # Fused attention layer has memory of shape [batch, src_time, ...] if self.time_major: memory = tf.transpose(encoder_outputs, [1, 0, 2]) else: memory = encoder_outputs fused_attention_layer = attention_wrapper.BahdanauAttentionFusedLayer( hparams.num_units, memory, memory_sequence_length=self.features["source_sequence_length"], dtype=self.dtype) # [batch, tgt_time, units] if self.time_major: queries = tf.transpose(outputs, [1, 0, 2]) else: queries = outputs fused_attention = fused_attention_layer(queries) if self.time_major: # [tgt_time, batch, units] fused_attention = tf.transpose(fused_attention, [1, 0, 2]) # 2-4th layer inputs = outputs for i in range(num_unidi_decoder_layers): # [tgt_time, batch, 2 * units] concat_inputs = tf.concat([inputs, fused_attention], axis=-1) # var_scope naming chosen to agree with inference graph. with tf.variable_scope("multi_rnn_cell/cell_%d" % (i+1)): outputs, _ = self._build_unidi_rnn_fused( concat_inputs, initial_state[i + 1], self.features["target_sequence_length"], hparams, self.dtype) if i >= num_unidi_decoder_layers - self.num_decoder_residual_layers: # gnmt_v2 attention adds the original inputs. outputs += inputs inputs = outputs return outputs class GNMTAttentionMultiCell(tf.nn.rnn_cell.MultiRNNCell): """A MultiCell with GNMT attention style.""" def __init__(self, attention_cell, cells): """Creates a GNMTAttentionMultiCell. Args: attention_cell: An instance of AttentionWrapper. cells: A list of RNNCell wrapped with AttentionInputWrapper. """ cells = [attention_cell] + cells super(GNMTAttentionMultiCell, self).__init__(cells, state_is_tuple=True) def __call__(self, inputs, state, scope=None): """Run the cell with bottom layer's attention copied to all upper layers.""" if not tf.contrib.framework.nest.is_sequence(state): raise ValueError( "Expected state to be a tuple of length %d, but received: %s" % (len(self.state_size), state)) with tf.variable_scope(scope or "multi_rnn_cell"): new_states = [] with tf.variable_scope("cell_0_attention"): attention_cell = self._cells[0] attention_state = state[0] cur_inp, new_attention_state = attention_cell(inputs, attention_state) new_states.append(new_attention_state) for i in range(1, len(self._cells)): with tf.variable_scope("cell_%d" % i): cell = self._cells[i] cur_state = state[i] cur_inp = tf.concat([cur_inp, new_attention_state.attention], -1) cur_inp, new_state = cell(cur_inp, cur_state) new_states.append(new_state) return cur_inp, tuple(new_states) def gnmt_residual_fn(inputs, outputs): """Residual function that handles different inputs and outputs inner dims. Args: inputs: cell inputs, this is actual inputs concatenated with the attention vector. outputs: cell outputs Returns: outputs + actual inputs """ def split_input(inp, out): inp_dim = inp.get_shape().as_list()[-1] out_dim = out.get_shape().as_list()[-1] return tf.split(inp, [out_dim, inp_dim - out_dim], axis=-1) actual_inputs, _ = tf.contrib.framework.nest.map_structure( split_input, inputs, outputs) def assert_shape_match(inp, out): inp.get_shape().assert_is_compatible_with(out.get_shape()) tf.contrib.framework.nest.assert_same_structure(actual_inputs, outputs) tf.contrib.framework.nest.map_structure( assert_shape_match, actual_inputs, outputs) return tf.contrib.framework.nest.map_structure( lambda inp, out: inp + out, actual_inputs, outputs)
TensorFlow/Segmentation/UNet_Medical/utils/hooks
hooks
training_hook
# Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import tensorflow as tf import horovod.tensorflow as hvd class TrainingHook(tf.estimator.SessionRunHook): def __init__(self, logger, max_steps, log_every=1): self._log_every = log_every self._iter_idx = 0 self.logger = logger self.max_steps = max_steps def before_run(self, run_context): run_args = tf.estimator.SessionRunArgs( fetches=[ 'cross_loss_ref:0', 'dice_loss_ref:0', 'total_loss_ref:0', ] ) return run_args def after_run(self, run_context, run_values): cross_loss, dice_loss, total_loss = run_values.results if (self._iter_idx % self._log_every == 0) and (hvd.rank() == 0): self.logger.log(step=(self._iter_idx, self.max_steps), data={'train_ce_loss': float(cross_loss), 'train_dice_loss': float(dice_loss), 'train_total_loss': float(total_loss)}) self._iter_idx += 1
TensorFlow2/LanguageModeling/BERT/scripts
scripts
run_inference_benchmark
#!/usr/bin/env bash # Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== echo "Container nvidia build = " $NVIDIA_BUILD_ID bert_model=${1:-"large"} batch_size=${2:-"8"} precision=${3:-"fp16"} use_xla=${4:-"true"} squad_version="1.1" if [ "$bert_model" = "large" ] ; then export BERT_DIR=data/download/google_pretrained_weights/uncased_L-24_H-1024_A-16 else export BERT_DIR=data/download/google_pretrained_weights/uncased_L-12_H-768_A-12 fi init_checkpoint=$BERT_DIR/bert_model.ckpt export SQUAD_DIR=data/download/squad/v${squad_version} export SQUAD_VERSION=v$squad_version if [ "$squad_version" = "1.1" ] ; then version_2_with_negative="False" else version_2_with_negative="True" fi echo "Squad directory set as " $SQUAD_DIR " BERT directory set as " $BERT_DIR echo "Results directory set as " $RESULTS_DIR use_fp16="" if [ "$precision" = "fp16" ] ; then echo "fp16 activated!" use_fp16="--use_fp16" fi if [ "$use_xla" = "true" ] ; then use_xla_tag="--enable_xla" echo "XLA activated" else use_xla_tag="" fi ckpt_str=${init_checkpoint//\//-} printf -v TAG "squad_inference_benchmark_%s_%s_bs%d" "$bert_model" "$precision" $batch_size DATESTAMP=`date +'%y%m%d%H%M%S'` #Edit to save logs & checkpoints in a different directory RESULTS_DIR=/tmp/bert_inference_benchmark_${DATESTAMP} LOGFILE=/results/$TAG.log printf "Logs written to %s\n" "$LOGFILE" mkdir -p $RESULTS_DIR mkdir -p /results #Check if all necessary files are available before training for DIR_or_file in $SQUAD_DIR $RESULTS_DIR $BERT_DIR/vocab.txt $BERT_DIR/bert_config.json; do if [ ! -d "$DIR_or_file" ] && [ ! -f "$DIR_or_file" ]; then echo "Error! $DIR_or_file directory missing. Please mount correctly" exit -1 fi done python run_squad.py \ --mode=predict \ --input_meta_data_path=${SQUAD_DIR}/squad_${SQUAD_VERSION}_meta_data \ --vocab_file=$BERT_DIR/vocab.txt \ --bert_config_file=$BERT_DIR/bert_config.json \ --init_checkpoint=$init_checkpoint \ --predict_file=$SQUAD_DIR/dev-v${squad_version}.json \ --predict_batch_size=$batch_size \ --model_dir=$RESULTS_DIR \ --benchmark \ $use_fp16 $use_xla_tag |& tee $LOGFILE rm $RESULTS_DIR -r
PyTorch/DrugDiscovery/SE3Transformer
SE3Transformer
README
# [SE(3)-Transformers For PyTorch [migrated, click here]](https://github.com/NVIDIA/DeepLearningExamples/blob/master/DGLPyTorch/DrugDiscovery/SE3Transformer) ![Model high-level architecture](../../../DGLPyTorch/DrugDiscovery/SE3Transformer/images/se3-transformer.png)
TensorFlow/Detection/SSD/models/research/object_detection/utils
utils
per_image_evaluation_test
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Tests for object_detection.utils.per_image_evaluation.""" import numpy as np import tensorflow as tf from object_detection.utils import per_image_evaluation class SingleClassTpFpWithDifficultBoxesTest(tf.test.TestCase): def setUp(self): num_groundtruth_classes = 1 matching_iou_threshold = 0.5 nms_iou_threshold = 1.0 nms_max_output_boxes = 10000 self.eval = per_image_evaluation.PerImageEvaluation( num_groundtruth_classes, matching_iou_threshold, nms_iou_threshold, nms_max_output_boxes) self.detected_boxes = np.array([[0, 0, 1, 1], [0, 0, 2, 2], [0, 0, 3, 3]], dtype=float) self.detected_scores = np.array([0.6, 0.8, 0.5], dtype=float) detected_masks_0 = np.array([[0, 1, 1, 0], [0, 0, 1, 0], [0, 0, 0, 0]], dtype=np.uint8) detected_masks_1 = np.array([[1, 0, 0, 0], [1, 1, 0, 0], [0, 0, 0, 0]], dtype=np.uint8) detected_masks_2 = np.array([[0, 0, 0, 0], [0, 1, 1, 0], [0, 1, 0, 0]], dtype=np.uint8) self.detected_masks = np.stack( [detected_masks_0, detected_masks_1, detected_masks_2], axis=0) self.groundtruth_boxes = np.array([[0, 0, 1, 1], [0, 0, 10, 10]], dtype=float) groundtruth_masks_0 = np.array([[1, 1, 0, 0], [1, 1, 0, 0], [0, 0, 0, 0]], dtype=np.uint8) groundtruth_masks_1 = np.array([[0, 0, 0, 1], [0, 0, 0, 1], [0, 0, 0, 1]], dtype=np.uint8) self.groundtruth_masks = np.stack( [groundtruth_masks_0, groundtruth_masks_1], axis=0) def test_match_to_gt_box_0(self): groundtruth_groundtruth_is_difficult_list = np.array([False, True], dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array( [False, False], dtype=bool) scores, tp_fp_labels = self.eval._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, self.groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list) expected_scores = np.array([0.8, 0.6, 0.5], dtype=float) expected_tp_fp_labels = np.array([False, True, False], dtype=bool) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_mask_match_to_gt_mask_0(self): groundtruth_groundtruth_is_difficult_list = np.array([False, True], dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array( [False, False], dtype=bool) scores, tp_fp_labels = self.eval._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, self.groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list, detected_masks=self.detected_masks, groundtruth_masks=self.groundtruth_masks) expected_scores = np.array([0.8, 0.6, 0.5], dtype=float) expected_tp_fp_labels = np.array([True, False, False], dtype=bool) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_match_to_gt_box_1(self): groundtruth_groundtruth_is_difficult_list = np.array([True, False], dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array( [False, False], dtype=bool) scores, tp_fp_labels = self.eval._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, self.groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list) expected_scores = np.array([0.8, 0.5], dtype=float) expected_tp_fp_labels = np.array([False, False], dtype=bool) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_mask_match_to_gt_mask_1(self): groundtruth_groundtruth_is_difficult_list = np.array([True, False], dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array( [False, False], dtype=bool) scores, tp_fp_labels = self.eval._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, self.groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list, detected_masks=self.detected_masks, groundtruth_masks=self.groundtruth_masks) expected_scores = np.array([0.6, 0.5], dtype=float) expected_tp_fp_labels = np.array([False, False], dtype=bool) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) class SingleClassTpFpWithGroupOfBoxesTest(tf.test.TestCase): def setUp(self): num_groundtruth_classes = 1 matching_iou_threshold = 0.5 nms_iou_threshold = 1.0 nms_max_output_boxes = 10000 self.eval = per_image_evaluation.PerImageEvaluation( num_groundtruth_classes, matching_iou_threshold, nms_iou_threshold, nms_max_output_boxes) self.detected_boxes = np.array( [[0, 0, 1, 1], [0, 0, 2, 1], [0, 0, 3, 1]], dtype=float) self.detected_scores = np.array([0.8, 0.6, 0.5], dtype=float) detected_masks_0 = np.array([[0, 1, 1, 0], [0, 0, 1, 0], [0, 0, 0, 0]], dtype=np.uint8) detected_masks_1 = np.array([[1, 0, 0, 0], [1, 1, 0, 0], [0, 0, 0, 0]], dtype=np.uint8) detected_masks_2 = np.array([[0, 0, 0, 0], [0, 1, 1, 0], [0, 1, 0, 0]], dtype=np.uint8) self.detected_masks = np.stack( [detected_masks_0, detected_masks_1, detected_masks_2], axis=0) self.groundtruth_boxes = np.array( [[0, 0, 1, 1], [0, 0, 5, 5], [10, 10, 20, 20]], dtype=float) groundtruth_masks_0 = np.array([[1, 0, 0, 0], [1, 0, 0, 0], [1, 0, 0, 0]], dtype=np.uint8) groundtruth_masks_1 = np.array([[0, 0, 1, 0], [0, 0, 1, 0], [0, 0, 1, 0]], dtype=np.uint8) groundtruth_masks_2 = np.array([[0, 1, 0, 0], [0, 1, 0, 0], [0, 1, 0, 0]], dtype=np.uint8) self.groundtruth_masks = np.stack( [groundtruth_masks_0, groundtruth_masks_1, groundtruth_masks_2], axis=0) def test_match_to_non_group_of_and_group_of_box(self): groundtruth_groundtruth_is_difficult_list = np.array( [False, False, False], dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array( [False, True, True], dtype=bool) expected_scores = np.array([0.8], dtype=float) expected_tp_fp_labels = np.array([True], dtype=bool) scores, tp_fp_labels = self.eval._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, self.groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_mask_match_to_non_group_of_and_group_of_box(self): groundtruth_groundtruth_is_difficult_list = np.array( [False, False, False], dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array( [False, True, True], dtype=bool) expected_scores = np.array([0.6], dtype=float) expected_tp_fp_labels = np.array([True], dtype=bool) scores, tp_fp_labels = self.eval._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, self.groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list, detected_masks=self.detected_masks, groundtruth_masks=self.groundtruth_masks) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_match_two_to_group_of_box(self): groundtruth_groundtruth_is_difficult_list = np.array( [False, False, False], dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array( [True, False, True], dtype=bool) expected_scores = np.array([0.5], dtype=float) expected_tp_fp_labels = np.array([False], dtype=bool) scores, tp_fp_labels = self.eval._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, self.groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_mask_match_two_to_group_of_box(self): groundtruth_groundtruth_is_difficult_list = np.array( [False, False, False], dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array( [True, False, True], dtype=bool) expected_scores = np.array([0.8], dtype=float) expected_tp_fp_labels = np.array([True], dtype=bool) scores, tp_fp_labels = self.eval._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, self.groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list, detected_masks=self.detected_masks, groundtruth_masks=self.groundtruth_masks) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) class SingleClassTpFpWithGroupOfBoxesTestWeighted(tf.test.TestCase): def setUp(self): num_groundtruth_classes = 1 matching_iou_threshold = 0.5 nms_iou_threshold = 1.0 nms_max_output_boxes = 10000 self.group_of_weight = 0.5 self.eval = per_image_evaluation.PerImageEvaluation( num_groundtruth_classes, matching_iou_threshold, nms_iou_threshold, nms_max_output_boxes, self.group_of_weight) self.detected_boxes = np.array( [[0, 0, 1, 1], [0, 0, 2, 1], [0, 0, 3, 1]], dtype=float) self.detected_scores = np.array([0.8, 0.6, 0.5], dtype=float) detected_masks_0 = np.array( [[0, 1, 1, 0], [0, 0, 1, 0], [0, 0, 0, 0]], dtype=np.uint8) detected_masks_1 = np.array( [[1, 0, 0, 0], [1, 1, 0, 0], [0, 0, 0, 0]], dtype=np.uint8) detected_masks_2 = np.array( [[0, 0, 0, 0], [0, 1, 1, 0], [0, 1, 0, 0]], dtype=np.uint8) self.detected_masks = np.stack( [detected_masks_0, detected_masks_1, detected_masks_2], axis=0) self.groundtruth_boxes = np.array( [[0, 0, 1, 1], [0, 0, 5, 5], [10, 10, 20, 20]], dtype=float) groundtruth_masks_0 = np.array( [[1, 0, 0, 0], [1, 0, 0, 0], [1, 0, 0, 0]], dtype=np.uint8) groundtruth_masks_1 = np.array( [[0, 0, 1, 0], [0, 0, 1, 0], [0, 0, 1, 0]], dtype=np.uint8) groundtruth_masks_2 = np.array( [[0, 1, 0, 0], [0, 1, 0, 0], [0, 1, 0, 0]], dtype=np.uint8) self.groundtruth_masks = np.stack( [groundtruth_masks_0, groundtruth_masks_1, groundtruth_masks_2], axis=0) def test_match_to_non_group_of_and_group_of_box(self): groundtruth_groundtruth_is_difficult_list = np.array( [False, False, False], dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array( [False, True, True], dtype=bool) expected_scores = np.array([0.8, 0.6], dtype=float) expected_tp_fp_labels = np.array([1.0, self.group_of_weight], dtype=float) scores, tp_fp_labels = self.eval._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, self.groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_mask_match_to_non_group_of_and_group_of_box(self): groundtruth_groundtruth_is_difficult_list = np.array( [False, False, False], dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array( [False, True, True], dtype=bool) expected_scores = np.array([0.6, 0.8, 0.5], dtype=float) expected_tp_fp_labels = np.array( [1.0, self.group_of_weight, self.group_of_weight], dtype=float) scores, tp_fp_labels = self.eval._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, self.groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list, detected_masks=self.detected_masks, groundtruth_masks=self.groundtruth_masks) tf.logging.info( "test_mask_match_to_non_group_of_and_group_of_box {} {}".format( tp_fp_labels, expected_tp_fp_labels)) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_match_two_to_group_of_box(self): groundtruth_groundtruth_is_difficult_list = np.array( [False, False, False], dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array( [True, False, True], dtype=bool) expected_scores = np.array([0.5, 0.8], dtype=float) expected_tp_fp_labels = np.array([0.0, self.group_of_weight], dtype=float) scores, tp_fp_labels = self.eval._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, self.groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list) tf.logging.info("test_match_two_to_group_of_box {} {}".format( tp_fp_labels, expected_tp_fp_labels)) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_mask_match_two_to_group_of_box(self): groundtruth_groundtruth_is_difficult_list = np.array( [False, False, False], dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array( [True, False, True], dtype=bool) expected_scores = np.array([0.8, 0.6, 0.5], dtype=float) expected_tp_fp_labels = np.array( [1.0, self.group_of_weight, self.group_of_weight], dtype=float) scores, tp_fp_labels = self.eval._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, self.groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list, detected_masks=self.detected_masks, groundtruth_masks=self.groundtruth_masks) tf.logging.info("test_mask_match_two_to_group_of_box {} {}".format( tp_fp_labels, expected_tp_fp_labels)) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) class SingleClassTpFpNoDifficultBoxesTest(tf.test.TestCase): def setUp(self): num_groundtruth_classes = 1 matching_iou_threshold_high_iou = 0.5 matching_iou_threshold_low_iou = 0.1 nms_iou_threshold = 1.0 nms_max_output_boxes = 10000 self.eval_high_iou = per_image_evaluation.PerImageEvaluation( num_groundtruth_classes, matching_iou_threshold_high_iou, nms_iou_threshold, nms_max_output_boxes) self.eval_low_iou = per_image_evaluation.PerImageEvaluation( num_groundtruth_classes, matching_iou_threshold_low_iou, nms_iou_threshold, nms_max_output_boxes) self.detected_boxes = np.array([[0, 0, 1, 1], [0, 0, 2, 2], [0, 0, 3, 3]], dtype=float) self.detected_scores = np.array([0.6, 0.8, 0.5], dtype=float) detected_masks_0 = np.array([[0, 1, 1, 0], [0, 0, 1, 0], [0, 0, 0, 0]], dtype=np.uint8) detected_masks_1 = np.array([[1, 0, 0, 0], [1, 1, 0, 0], [0, 0, 0, 0]], dtype=np.uint8) detected_masks_2 = np.array([[0, 0, 0, 0], [0, 1, 1, 0], [0, 1, 0, 0]], dtype=np.uint8) self.detected_masks = np.stack( [detected_masks_0, detected_masks_1, detected_masks_2], axis=0) def test_no_true_positives(self): groundtruth_boxes = np.array([[100, 100, 105, 105]], dtype=float) groundtruth_groundtruth_is_difficult_list = np.zeros(1, dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array([False], dtype=bool) scores, tp_fp_labels = self.eval_high_iou._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list) expected_scores = np.array([0.8, 0.6, 0.5], dtype=float) expected_tp_fp_labels = np.array([False, False, False], dtype=bool) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_mask_no_true_positives(self): groundtruth_boxes = np.array([[100, 100, 105, 105]], dtype=float) groundtruth_masks_0 = np.array([[1, 1, 1, 1], [1, 1, 1, 1], [1, 1, 1, 1]], dtype=np.uint8) groundtruth_masks = np.stack([groundtruth_masks_0], axis=0) groundtruth_groundtruth_is_difficult_list = np.zeros(1, dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array([False], dtype=bool) scores, tp_fp_labels = self.eval_high_iou._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list, detected_masks=self.detected_masks, groundtruth_masks=groundtruth_masks) expected_scores = np.array([0.8, 0.6, 0.5], dtype=float) expected_tp_fp_labels = np.array([False, False, False], dtype=bool) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_one_true_positives_with_large_iou_threshold(self): groundtruth_boxes = np.array([[0, 0, 1, 1]], dtype=float) groundtruth_groundtruth_is_difficult_list = np.zeros(1, dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array([False], dtype=bool) scores, tp_fp_labels = self.eval_high_iou._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list) expected_scores = np.array([0.8, 0.6, 0.5], dtype=float) expected_tp_fp_labels = np.array([False, True, False], dtype=bool) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_mask_one_true_positives_with_large_iou_threshold(self): groundtruth_boxes = np.array([[0, 0, 1, 1]], dtype=float) groundtruth_masks_0 = np.array([[1, 0, 0, 0], [1, 1, 0, 0], [0, 0, 0, 0]], dtype=np.uint8) groundtruth_masks = np.stack([groundtruth_masks_0], axis=0) groundtruth_groundtruth_is_difficult_list = np.zeros(1, dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array([False], dtype=bool) scores, tp_fp_labels = self.eval_high_iou._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list, detected_masks=self.detected_masks, groundtruth_masks=groundtruth_masks) expected_scores = np.array([0.8, 0.6, 0.5], dtype=float) expected_tp_fp_labels = np.array([True, False, False], dtype=bool) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_one_true_positives_with_very_small_iou_threshold(self): groundtruth_boxes = np.array([[0, 0, 1, 1]], dtype=float) groundtruth_groundtruth_is_difficult_list = np.zeros(1, dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array([False], dtype=bool) scores, tp_fp_labels = self.eval_low_iou._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list) expected_scores = np.array([0.8, 0.6, 0.5], dtype=float) expected_tp_fp_labels = np.array([True, False, False], dtype=bool) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) def test_two_true_positives_with_large_iou_threshold(self): groundtruth_boxes = np.array([[0, 0, 1, 1], [0, 0, 3.5, 3.5]], dtype=float) groundtruth_groundtruth_is_difficult_list = np.zeros(2, dtype=bool) groundtruth_groundtruth_is_group_of_list = np.array( [False, False], dtype=bool) scores, tp_fp_labels = self.eval_high_iou._compute_tp_fp_for_single_class( self.detected_boxes, self.detected_scores, groundtruth_boxes, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list) expected_scores = np.array([0.8, 0.6, 0.5], dtype=float) expected_tp_fp_labels = np.array([False, True, True], dtype=bool) self.assertTrue(np.allclose(expected_scores, scores)) self.assertTrue(np.allclose(expected_tp_fp_labels, tp_fp_labels)) class MultiClassesTpFpTest(tf.test.TestCase): def test_tp_fp(self): num_groundtruth_classes = 3 matching_iou_threshold = 0.5 nms_iou_threshold = 1.0 nms_max_output_boxes = 10000 eval1 = per_image_evaluation.PerImageEvaluation(num_groundtruth_classes, matching_iou_threshold, nms_iou_threshold, nms_max_output_boxes) detected_boxes = np.array([[0, 0, 1, 1], [10, 10, 5, 5], [0, 0, 2, 2], [5, 10, 10, 5], [10, 5, 5, 10], [0, 0, 3, 3]], dtype=float) detected_scores = np.array([0.8, 0.1, 0.8, 0.9, 0.7, 0.8], dtype=float) detected_class_labels = np.array([0, 1, 1, 2, 0, 2], dtype=int) groundtruth_boxes = np.array([[0, 0, 1, 1], [0, 0, 3.5, 3.5]], dtype=float) groundtruth_class_labels = np.array([0, 2], dtype=int) groundtruth_groundtruth_is_difficult_list = np.zeros(2, dtype=float) groundtruth_groundtruth_is_group_of_list = np.array( [False, False], dtype=bool) scores, tp_fp_labels, _ = eval1.compute_object_detection_metrics( detected_boxes, detected_scores, detected_class_labels, groundtruth_boxes, groundtruth_class_labels, groundtruth_groundtruth_is_difficult_list, groundtruth_groundtruth_is_group_of_list) expected_scores = [np.array([0.8], dtype=float)] * 3 expected_tp_fp_labels = [np.array([True]), np.array([False]), np.array([True ])] for i in range(len(expected_scores)): self.assertTrue(np.allclose(expected_scores[i], scores[i])) self.assertTrue(np.array_equal(expected_tp_fp_labels[i], tp_fp_labels[i])) class CorLocTest(tf.test.TestCase): def test_compute_corloc_with_normal_iou_threshold(self): num_groundtruth_classes = 3 matching_iou_threshold = 0.5 nms_iou_threshold = 1.0 nms_max_output_boxes = 10000 eval1 = per_image_evaluation.PerImageEvaluation(num_groundtruth_classes, matching_iou_threshold, nms_iou_threshold, nms_max_output_boxes) detected_boxes = np.array([[0, 0, 1, 1], [0, 0, 2, 2], [0, 0, 3, 3], [0, 0, 5, 5]], dtype=float) detected_scores = np.array([0.9, 0.9, 0.1, 0.9], dtype=float) detected_class_labels = np.array([0, 1, 0, 2], dtype=int) groundtruth_boxes = np.array([[0, 0, 1, 1], [0, 0, 3, 3], [0, 0, 6, 6]], dtype=float) groundtruth_class_labels = np.array([0, 0, 2], dtype=int) is_class_correctly_detected_in_image = eval1._compute_cor_loc( detected_boxes, detected_scores, detected_class_labels, groundtruth_boxes, groundtruth_class_labels) expected_result = np.array([1, 0, 1], dtype=int) self.assertTrue(np.array_equal(expected_result, is_class_correctly_detected_in_image)) def test_compute_corloc_with_very_large_iou_threshold(self): num_groundtruth_classes = 3 matching_iou_threshold = 0.9 nms_iou_threshold = 1.0 nms_max_output_boxes = 10000 eval1 = per_image_evaluation.PerImageEvaluation(num_groundtruth_classes, matching_iou_threshold, nms_iou_threshold, nms_max_output_boxes) detected_boxes = np.array([[0, 0, 1, 1], [0, 0, 2, 2], [0, 0, 3, 3], [0, 0, 5, 5]], dtype=float) detected_scores = np.array([0.9, 0.9, 0.1, 0.9], dtype=float) detected_class_labels = np.array([0, 1, 0, 2], dtype=int) groundtruth_boxes = np.array([[0, 0, 1, 1], [0, 0, 3, 3], [0, 0, 6, 6]], dtype=float) groundtruth_class_labels = np.array([0, 0, 2], dtype=int) is_class_correctly_detected_in_image = eval1._compute_cor_loc( detected_boxes, detected_scores, detected_class_labels, groundtruth_boxes, groundtruth_class_labels) expected_result = np.array([1, 0, 0], dtype=int) self.assertTrue(np.array_equal(expected_result, is_class_correctly_detected_in_image)) if __name__ == "__main__": tf.test.main()
TensorFlow/Detection/SSD/models/research/object_detection/utils
utils
metrics
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Functions for computing metrics like precision, recall, CorLoc and etc.""" from __future__ import division import numpy as np def compute_precision_recall(scores, labels, num_gt): """Compute precision and recall. Args: scores: A float numpy array representing detection score labels: A float numpy array representing weighted true/false positive labels num_gt: Number of ground truth instances Raises: ValueError: if the input is not of the correct format Returns: precision: Fraction of positive instances over detected ones. This value is None if no ground truth labels are present. recall: Fraction of detected positive instance over all positive instances. This value is None if no ground truth labels are present. """ if not isinstance(labels, np.ndarray) or len(labels.shape) != 1: raise ValueError("labels must be single dimension numpy array") if labels.dtype != np.float and labels.dtype != np.bool: raise ValueError("labels type must be either bool or float") if not isinstance(scores, np.ndarray) or len(scores.shape) != 1: raise ValueError("scores must be single dimension numpy array") if num_gt < np.sum(labels): raise ValueError("Number of true positives must be smaller than num_gt.") if len(scores) != len(labels): raise ValueError("scores and labels must be of the same size.") if num_gt == 0: return None, None sorted_indices = np.argsort(scores) sorted_indices = sorted_indices[::-1] true_positive_labels = labels[sorted_indices] false_positive_labels = (true_positive_labels <= 0).astype(float) cum_true_positives = np.cumsum(true_positive_labels) cum_false_positives = np.cumsum(false_positive_labels) precision = cum_true_positives.astype(float) / ( cum_true_positives + cum_false_positives) recall = cum_true_positives.astype(float) / num_gt return precision, recall def compute_average_precision(precision, recall): """Compute Average Precision according to the definition in VOCdevkit. Precision is modified to ensure that it does not decrease as recall decrease. Args: precision: A float [N, 1] numpy array of precisions recall: A float [N, 1] numpy array of recalls Raises: ValueError: if the input is not of the correct format Returns: average_precison: The area under the precision recall curve. NaN if precision and recall are None. """ if precision is None: if recall is not None: raise ValueError("If precision is None, recall must also be None") return np.NAN if not isinstance(precision, np.ndarray) or not isinstance( recall, np.ndarray): raise ValueError("precision and recall must be numpy array") if precision.dtype != np.float or recall.dtype != np.float: raise ValueError("input must be float numpy array.") if len(precision) != len(recall): raise ValueError("precision and recall must be of the same size.") if not precision.size: return 0.0 if np.amin(precision) < 0 or np.amax(precision) > 1: raise ValueError("Precision must be in the range of [0, 1].") if np.amin(recall) < 0 or np.amax(recall) > 1: raise ValueError("recall must be in the range of [0, 1].") if not all(recall[i] <= recall[i + 1] for i in range(len(recall) - 1)): raise ValueError("recall must be a non-decreasing array") recall = np.concatenate([[0], recall, [1]]) precision = np.concatenate([[0], precision, [0]]) # Preprocess precision to be a non-decreasing array for i in range(len(precision) - 2, -1, -1): precision[i] = np.maximum(precision[i], precision[i + 1]) indices = np.where(recall[1:] != recall[:-1])[0] + 1 average_precision = np.sum( (recall[indices] - recall[indices - 1]) * precision[indices]) return average_precision def compute_cor_loc(num_gt_imgs_per_class, num_images_correctly_detected_per_class): """Compute CorLoc according to the definition in the following paper. https://www.robots.ox.ac.uk/~vgg/rg/papers/deselaers-eccv10.pdf Returns nans if there are no ground truth images for a class. Args: num_gt_imgs_per_class: 1D array, representing number of images containing at least one object instance of a particular class num_images_correctly_detected_per_class: 1D array, representing number of images that are correctly detected at least one object instance of a particular class Returns: corloc_per_class: A float numpy array represents the corloc score of each class """ return np.where( num_gt_imgs_per_class == 0, np.nan, num_images_correctly_detected_per_class / num_gt_imgs_per_class) def compute_median_rank_at_k(tp_fp_list, k): """Computes MedianRank@k, where k is the top-scoring labels. Args: tp_fp_list: a list of numpy arrays; each numpy array corresponds to the all detection on a single image, where the detections are sorted by score in descending order. Further, each numpy array element can have boolean or float values. True positive elements have either value >0.0 or True; any other value is considered false positive. k: number of top-scoring proposals to take. Returns: median_rank: median rank of all true positive proposals among top k by score. """ ranks = [] for i in range(len(tp_fp_list)): ranks.append( np.where(tp_fp_list[i][0:min(k, tp_fp_list[i].shape[0])] > 0)[0]) concatenated_ranks = np.concatenate(ranks) return np.median(concatenated_ranks) def compute_recall_at_k(tp_fp_list, num_gt, k): """Computes Recall@k, MedianRank@k, where k is the top-scoring labels. Args: tp_fp_list: a list of numpy arrays; each numpy array corresponds to the all detection on a single image, where the detections are sorted by score in descending order. Further, each numpy array element can have boolean or float values. True positive elements have either value >0.0 or True; any other value is considered false positive. num_gt: number of groundtruth anotations. k: number of top-scoring proposals to take. Returns: recall: recall evaluated on the top k by score detections. """ tp_fp_eval = [] for i in range(len(tp_fp_list)): tp_fp_eval.append(tp_fp_list[i][0:min(k, tp_fp_list[i].shape[0])]) tp_fp_eval = np.concatenate(tp_fp_eval) return np.sum(tp_fp_eval) / num_gt
TensorFlow2/Recommendation/WideAndDeep/scripts
scripts
preproc
#!/bin/bash # Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. set -e function usage() { echo "Usage: bash scripts/preproc.sh" } if [ ! -d "scripts" ] || [ ! "$(ls -A 'scripts')" ]; then echo "You are probably calling this script from wrong directory" usage exit 1 fi time python -m data.outbrain.nvtabular.preproc "$@"
TensorFlow/Detection/SSD/models/research/object_detection/utils
utils
dataset_util_test
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Tests for object_detection.utils.dataset_util.""" import os import tensorflow as tf from object_detection.utils import dataset_util class DatasetUtilTest(tf.test.TestCase): def test_read_examples_list(self): example_list_data = """example1 1\nexample2 2""" example_list_path = os.path.join(self.get_temp_dir(), 'examples.txt') with tf.gfile.Open(example_list_path, 'wb') as f: f.write(example_list_data) examples = dataset_util.read_examples_list(example_list_path) self.assertListEqual(['example1', 'example2'], examples) if __name__ == '__main__': tf.test.main()
PyTorch/LanguageModeling/Transformer-XL/pytorch/scripts/docker
docker
build
#!/bin/bash # Copyright (c) 2019-2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. docker build . --network=host --rm -t transformer-xl:latest
TensorFlow/Segmentation/UNet_3D_Medical/scripts
scripts
unet3d_train_full
# Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # This script launches 3D-UNet run 5-fold cross-validation FP32 training for 16000 iterations each. # Usage: # bash examples/unet3d_train_full.sh <number/of/gpus> <path/to/dataset> <path/to/results/directory> <batch/size> horovodrun -np $1 python main.py --data_dir $2 --model_dir $3 --log_dir $3/log.json --exec_mode train_and_evaluate --max_steps 16000 --augment --batch_size $4 --fold 0 --use_xla > $3/log_FP32_$1GPU_fold0.txt horovodrun -np $1 python main.py --data_dir $2 --model_dir $3 --log_dir $3/log.json --exec_mode train_and_evaluate --max_steps 16000 --augment --batch_size $4 --fold 1 --use_xla > $3/log_FP32_$1GPU_fold1.txt horovodrun -np $1 python main.py --data_dir $2 --model_dir $3 --log_dir $3/log.json --exec_mode train_and_evaluate --max_steps 16000 --augment --batch_size $4 --fold 2 --use_xla > $3/log_FP32_$1GPU_fold2.txt horovodrun -np $1 python main.py --data_dir $2 --model_dir $3 --log_dir $3/log.json --exec_mode train_and_evaluate --max_steps 16000 --augment --batch_size $4 --fold 3 --use_xla > $3/log_FP32_$1GPU_fold3.txt horovodrun -np $1 python main.py --data_dir $2 --model_dir $3 --log_dir $3/log.json --exec_mode train_and_evaluate --max_steps 16000 --augment --batch_size $4 --fold 4 --use_xla > $3/log_FP32_$1GPU_fold4.txt python runtime/parse_results.py --model_dir $3 --env FP32_$1GPU
PyTorch/SpeechSynthesis/HiFiGAN
HiFiGAN
inference
# Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import itertools import sys import time import warnings from pathlib import Path from tqdm import tqdm import torch import numpy as np from scipy.stats import norm from scipy.io.wavfile import write from torch.nn.functional import l1_loss from torch.nn.utils.rnn import pad_sequence import dllogger as DLLogger from dllogger import StdOutBackend, JSONStreamBackend, Verbosity import models from common import gpu_affinity from common.tb_dllogger import stdout_metric_format, unique_log_fpath from common.text import cmudict from common.text.text_processing import TextProcessing from common.utils import l2_promote from fastpitch.pitch_transform import pitch_transform_custom from hifigan.data_function import MAX_WAV_VALUE, mel_spectrogram from hifigan.logging import init_inference_metadata from hifigan.models import Denoiser CHECKPOINT_SPECIFIC_ARGS = [ 'sampling_rate', 'hop_length', 'win_length', 'p_arpabet', 'text_cleaners', 'symbol_set', 'max_wav_value', 'prepend_space_to_text', 'append_space_to_text'] def parse_args(parser): """ Parse commandline arguments. """ parser.add_argument('-i', '--input', type=str, required=True, help='Full path to the input text ' '(phareses separated by newlines)') parser.add_argument('-o', '--output', default=None, help='Output folder to save audio (file per phrase)') parser.add_argument('--log-file', type=str, default=None, help='Path to a DLLogger log file') parser.add_argument('--save-mels', action='store_true', help='Save generator outputs to disk') parser.add_argument('--cuda', action='store_true', help='Run inference on a GPU using CUDA') parser.add_argument('--cudnn-benchmark', action='store_true', help='Enable cudnn benchmark mode') parser.add_argument('--l2-promote', action='store_true', help='Increase max fetch granularity of GPU L2 cache') parser.add_argument('--fastpitch', type=str, default=None, required=False, help='Full path to the spectrogram generator .pt file ' '(skip to synthesize from ground truth mels)') parser.add_argument('--waveglow', type=str, default=None, required=False, help='Full path to a WaveGlow model .pt file') parser.add_argument('-s', '--waveglow-sigma-infer', default=0.9, type=float, help='WaveGlow sigma') parser.add_argument('--hifigan', type=str, default=None, required=False, help='Full path to a HiFi-GAN model .pt file') parser.add_argument('-d', '--denoising-strength', default=0.0, type=float, help='Capture and subtract model bias to enhance audio') parser.add_argument('--hop-length', type=int, default=256, help='STFT hop length for estimating audio length from mel size') parser.add_argument('--win-length', type=int, default=1024, help='STFT win length for denoiser and mel loss') parser.add_argument('-sr', '--sampling-rate', default=22050, type=int, choices=[22050, 44100], help='Sampling rate') parser.add_argument('--max_wav_value', default=32768.0, type=float, help='Maximum audiowave value') parser.add_argument('--amp', action='store_true', help='Inference with AMP') parser.add_argument('-bs', '--batch-size', type=int, default=64) parser.add_argument('--warmup-steps', type=int, default=0, help='Warmup iterations before measuring performance') parser.add_argument('--repeats', type=int, default=1, help='Repeat inference for benchmarking') parser.add_argument('--torchscript', action='store_true', help='Run inference with TorchScript model (convert to TS if needed)') parser.add_argument('--checkpoint-format', type=str, choices=['pyt', 'ts'], default='pyt', help='Input checkpoint format (PyT or TorchScript)') parser.add_argument('--torch-tensorrt', action='store_true', help='Run inference with Torch-TensorRT model (compile beforehand)') parser.add_argument('--report-mel-loss', action='store_true', help='Report mel loss in metrics') parser.add_argument('--ema', action='store_true', help='Use EMA averaged model (if saved in checkpoints)') parser.add_argument('--dataset-path', type=str, help='Path to dataset (for loading extra data fields)') parser.add_argument('--speaker', type=int, default=0, help='Speaker ID for a multi-speaker model') parser.add_argument('--affinity', type=str, default='single', choices=['socket', 'single', 'single_unique', 'socket_unique_interleaved', 'socket_unique_continuous', 'disabled'], help='type of CPU affinity') transf = parser.add_argument_group('transform') transf.add_argument('--fade-out', type=int, default=6, help='Number of fadeout frames at the end') transf.add_argument('--pace', type=float, default=1.0, help='Adjust the pace of speech') transf.add_argument('--pitch-transform-flatten', action='store_true', help='Flatten the pitch') transf.add_argument('--pitch-transform-invert', action='store_true', help='Invert the pitch wrt mean value') transf.add_argument('--pitch-transform-amplify', type=float, default=1.0, help='Multiplicative amplification of pitch variability. ' 'Typical values are in the range (1.0, 3.0).') transf.add_argument('--pitch-transform-shift', type=float, default=0.0, help='Raise/lower the pitch by <hz>') transf.add_argument('--pitch-transform-custom', action='store_true', help='Apply the transform from pitch_transform.py') txt = parser.add_argument_group('Text processing parameters') txt.add_argument('--text-cleaners', type=str, nargs='*', default=['english_cleaners_v2'], help='Type of text cleaners for input text') txt.add_argument('--symbol-set', type=str, default='english_basic', help='Define symbol set for input text') txt.add_argument('--p-arpabet', type=float, default=0.0, help='') txt.add_argument('--heteronyms-path', type=str, default='data/cmudict/heteronyms', help='') txt.add_argument('--cmudict-path', type=str, default='data/cmudict/cmudict-0.7b', help='') return parser def load_fields(fpath): lines = [l.strip() for l in open(fpath, encoding='utf-8')] if fpath.endswith('.tsv'): columns = lines[0].split('\t') fields = list(zip(*[t.split('\t') for t in lines[1:]])) else: columns = ['text'] fields = [lines] return {c: f for c, f in zip(columns, fields)} def prepare_input_sequence(fields, device, symbol_set, text_cleaners, batch_size=128, dataset=None, load_mels=False, load_pitch=False, p_arpabet=0.0): tp = TextProcessing(symbol_set, text_cleaners, p_arpabet=p_arpabet) fields['text'] = [torch.LongTensor(tp.encode_text(text)) for text in fields['text']] order = np.argsort([-t.size(0) for t in fields['text']]) fields['text'] = [fields['text'][i] for i in order] fields['text_lens'] = torch.LongTensor([t.size(0) for t in fields['text']]) for t in fields['text']: print(tp.sequence_to_text(t.numpy())) if load_mels: assert 'mel' in fields assert dataset is not None fields['mel'] = [ torch.load(Path(dataset, fields['mel'][i])).t() for i in order] fields['mel_lens'] = torch.LongTensor([t.size(0) for t in fields['mel']]) if load_pitch: assert 'pitch' in fields fields['pitch'] = [ torch.load(Path(dataset, fields['pitch'][i])) for i in order] fields['pitch_lens'] = torch.LongTensor([t.size(0) for t in fields['pitch']]) if 'output' in fields: fields['output'] = [fields['output'][i] for i in order] # cut into batches & pad batches = [] for b in range(0, len(order), batch_size): batch = {f: values[b:b+batch_size] for f, values in fields.items()} for f in batch: if f == 'text': batch[f] = pad_sequence(batch[f], batch_first=True) elif f == 'mel' and load_mels: batch[f] = pad_sequence(batch[f], batch_first=True).permute(0, 2, 1) elif f == 'pitch' and load_pitch: batch[f] = pad_sequence(batch[f], batch_first=True) if type(batch[f]) is torch.Tensor: batch[f] = batch[f].to(device) batches.append(batch) return batches def build_pitch_transformation(args): if args.pitch_transform_custom: def custom_(pitch, pitch_lens, mean, std): return (pitch_transform_custom(pitch * std + mean, pitch_lens) - mean) / std return custom_ fun = 'pitch' if args.pitch_transform_flatten: fun = f'({fun}) * 0.0' if args.pitch_transform_invert: fun = f'({fun}) * -1.0' if args.pitch_transform_amplify != 1.0: ampl = args.pitch_transform_amplify fun = f'({fun}) * {ampl}' if args.pitch_transform_shift != 0.0: hz = args.pitch_transform_shift fun = f'({fun}) + {hz} / std' if fun == 'pitch': return None return eval(f'lambda pitch, pitch_lens, mean, std: {fun}') def setup_mel_loss_reporting(args, voc_train_setup): if args.denoising_strength > 0.0: print('WARNING: denoising will be included in vocoder mel loss') num_mels = voc_train_setup.get('num_mels', 80) fmin = voc_train_setup.get('mel_fmin', 0) fmax = voc_train_setup.get('mel_fmax', 8000) # not mel_fmax_loss def compute_audio_mel_loss(gen_audios, gt_mels, mel_lens): gen_audios /= MAX_WAV_VALUE total_loss = 0 for gen_audio, gt_mel, mel_len in zip(gen_audios, gt_mels, mel_lens): mel_len = mel_len.item() gen_audio = gen_audio[None, :mel_len * args.hop_length] gen_mel = mel_spectrogram(gen_audio, args.win_length, num_mels, args.sampling_rate, args.hop_length, args.win_length, fmin, fmax)[0] total_loss += l1_loss(gen_mel, gt_mel[:, :mel_len]) return total_loss.item() return compute_audio_mel_loss def compute_mel_loss(mels, lens, gt_mels, gt_lens): total_loss = 0 for mel, len_, gt_mel, gt_len in zip(mels, lens, gt_mels, gt_lens): min_len = min(len_, gt_len) total_loss += l1_loss(gt_mel[:, :min_len], mel[:, :min_len]) return total_loss.item() class MeasureTime(list): def __init__(self, *args, cuda=True, **kwargs): super(MeasureTime, self).__init__(*args, **kwargs) self.cuda = cuda def __enter__(self): if self.cuda: torch.cuda.synchronize() self.t0 = time.time() def __exit__(self, exc_type, exc_value, exc_traceback): if self.cuda: torch.cuda.synchronize() self.append(time.time() - self.t0) def __add__(self, other): assert len(self) == len(other) return MeasureTime((sum(ab) for ab in zip(self, other)), cuda=self.cuda) def main(): """ Launches text-to-speech inference on a single GPU. """ parser = argparse.ArgumentParser(description='PyTorch FastPitch Inference', allow_abbrev=False) parser = parse_args(parser) args, unk_args = parser.parse_known_args() if args.affinity != 'disabled': nproc_per_node = torch.cuda.device_count() # print(nproc_per_node) affinity = gpu_affinity.set_affinity( 0, nproc_per_node, args.affinity ) print(f'Thread affinity: {affinity}') if args.l2_promote: l2_promote() torch.backends.cudnn.benchmark = args.cudnn_benchmark if args.output is not None: Path(args.output).mkdir(parents=False, exist_ok=True) log_fpath = args.log_file or str(Path(args.output, 'nvlog_infer.json')) DLLogger.init(backends=[ JSONStreamBackend(Verbosity.DEFAULT, log_fpath, append=True), JSONStreamBackend(Verbosity.DEFAULT, unique_log_fpath(log_fpath)), StdOutBackend(Verbosity.VERBOSE, metric_format=stdout_metric_format) ]) init_inference_metadata(args.batch_size) [DLLogger.log("PARAMETER", {k: v}) for k, v in vars(args).items()] device = torch.device('cuda' if args.cuda else 'cpu') gen_train_setup = {} voc_train_setup = {} generator = None vocoder = None denoiser = None is_ts_based_infer = args.torch_tensorrt or args.torchscript assert args.checkpoint_format == 'pyt' or is_ts_based_infer, \ 'TorchScript checkpoint can be used only for TS or Torch-TRT' \ ' inference. Please set --torchscript or --torch-tensorrt flag.' assert args.waveglow is None or args.hifigan is None, \ "Specify a single vocoder model" def _load_pyt_or_ts_model(model_name, ckpt_path): if args.checkpoint_format == 'ts': model = models.load_and_setup_ts_model(model_name, ckpt_path, args.amp, device) model_train_setup = {} return model, model_train_setup model, _, model_train_setup = models.load_and_setup_model( model_name, parser, ckpt_path, args.amp, device, unk_args=unk_args, forward_is_infer=True, jitable=is_ts_based_infer) if is_ts_based_infer: model = torch.jit.script(model) return model, model_train_setup if args.fastpitch is not None: gen_name = 'fastpitch' generator, gen_train_setup = _load_pyt_or_ts_model('FastPitch', args.fastpitch) if args.waveglow is not None: voc_name = 'waveglow' with warnings.catch_warnings(): warnings.simplefilter("ignore") vocoder, _, voc_train_setup = models.load_and_setup_model( 'WaveGlow', parser, args.waveglow, args.amp, device, unk_args=unk_args, forward_is_infer=True, jitable=False) if args.denoising_strength > 0.0: denoiser = Denoiser(vocoder, sigma=0.0, win_length=args.win_length).to(device) # if args.torchscript: # vocoder = torch.jit.script(vocoder) def generate_audio(mel): audios = vocoder(mel, sigma=args.waveglow_sigma_infer) if denoiser is not None: audios = denoiser(audios.float(), args.denoising_strength).squeeze(1) return audios elif args.hifigan is not None: voc_name = 'hifigan' vocoder, voc_train_setup = _load_pyt_or_ts_model('HiFi-GAN', args.hifigan) if args.denoising_strength > 0.0: denoiser = Denoiser(vocoder, win_length=args.win_length).to(device) if args.torch_tensorrt: vocoder = models.convert_ts_to_trt('HiFi-GAN', vocoder, parser, args.amp, unk_args) def generate_audio(mel): audios = vocoder(mel).float() if denoiser is not None: audios = denoiser(audios.squeeze(1), args.denoising_strength) return audios.squeeze(1) * args.max_wav_value if len(unk_args) > 0: raise ValueError(f'Invalid options {unk_args}') for k in CHECKPOINT_SPECIFIC_ARGS: v1 = gen_train_setup.get(k, None) v2 = voc_train_setup.get(k, None) assert v1 is None or v2 is None or v1 == v2, \ f'{k} mismatch in spectrogram generator and vocoder' val = v1 or v2 if val and getattr(args, k) != val: src = 'generator' if v2 is None else 'vocoder' print(f'Overwriting args.{k}={getattr(args, k)} with {val} ' f'from {src} checkpoint.') setattr(args, k, val) gen_kw = {'pace': args.pace, 'speaker': args.speaker, 'pitch_tgt': None, 'pitch_transform': build_pitch_transformation(args)} if is_ts_based_infer and generator is not None: gen_kw.pop('pitch_transform') print('Note: --pitch-transform-* args are disabled with TorchScript. ' 'To condition on pitch, pass pitch_tgt as input.') if args.p_arpabet > 0.0: cmudict.initialize(args.cmudict_path, args.heteronyms_path) if args.report_mel_loss: mel_loss_fn = setup_mel_loss_reporting(args, voc_train_setup) fields = load_fields(args.input) batches = prepare_input_sequence( fields, device, args.symbol_set, args.text_cleaners, args.batch_size, args.dataset_path, load_mels=(generator is None or args.report_mel_loss), p_arpabet=args.p_arpabet) cycle = itertools.cycle(batches) # Use real data rather than synthetic - FastPitch predicts len for _ in tqdm(range(args.warmup_steps), 'Warmup'): with torch.no_grad(): b = next(cycle) if generator is not None: mel, *_ = generator(b['text']) else: mel, mel_lens = b['mel'], b['mel_lens'] if args.amp: mel = mel.half() if vocoder is not None: audios = generate_audio(mel) gen_measures = MeasureTime(cuda=args.cuda) vocoder_measures = MeasureTime(cuda=args.cuda) all_utterances = 0 all_samples = 0 all_batches = 0 all_letters = 0 all_frames = 0 gen_mel_loss_sum = 0 voc_mel_loss_sum = 0 reps = args.repeats log_enabled = reps == 1 log = lambda s, d: DLLogger.log(step=s, data=d) if log_enabled else None for rep in (tqdm(range(reps), 'Inference') if reps > 1 else range(reps)): for b in batches: if generator is None: mel, mel_lens = b['mel'], b['mel_lens'] if args.amp: mel = mel.half() else: with torch.no_grad(), gen_measures: mel, mel_lens, *_ = generator(b['text'], **gen_kw) if args.report_mel_loss: gen_mel_loss_sum += compute_mel_loss( mel, mel_lens, b['mel'], b['mel_lens']) gen_infer_perf = mel.size(0) * mel.size(2) / gen_measures[-1] all_letters += b['text_lens'].sum().item() all_frames += mel.size(0) * mel.size(2) log(rep, {f"{gen_name}_frames/s": gen_infer_perf}) log(rep, {f"{gen_name}_latency": gen_measures[-1]}) if args.save_mels: for i, mel_ in enumerate(mel): m = mel_[:, :mel_lens[i].item()].permute(1, 0) fname = b['output'][i] if 'output' in b else f'mel_{i}.npy' mel_path = Path(args.output, Path(fname).stem + '.npy') np.save(mel_path, m.cpu().numpy()) if vocoder is not None: with torch.no_grad(), vocoder_measures: audios = generate_audio(mel) vocoder_infer_perf = ( audios.size(0) * audios.size(1) / vocoder_measures[-1]) log(rep, {f"{voc_name}_samples/s": vocoder_infer_perf}) log(rep, {f"{voc_name}_latency": vocoder_measures[-1]}) if args.report_mel_loss: voc_mel_loss_sum += mel_loss_fn(audios, mel, mel_lens) if args.output is not None and reps == 1: for i, audio in enumerate(audios): audio = audio[:mel_lens[i].item() * args.hop_length] if args.fade_out: fade_len = args.fade_out * args.hop_length fade_w = torch.linspace(1.0, 0.0, fade_len) audio[-fade_len:] *= fade_w.to(audio.device) audio = audio / torch.max(torch.abs(audio)) fname = b['output'][i] if 'output' in b else f'audio_{i}.wav' audio_path = Path(args.output, fname) write(audio_path, args.sampling_rate, audio.cpu().numpy()) if generator is not None: log(rep, {"latency": (gen_measures[-1] + vocoder_measures[-1])}) all_utterances += mel.size(0) all_samples += mel_lens.sum().item() * args.hop_length all_batches += 1 log_enabled = True if generator is not None: gm = np.sort(np.asarray(gen_measures)) rtf = all_samples / (all_utterances * gm.mean() * args.sampling_rate) rtf_at = all_samples / (all_batches * gm.mean() * args.sampling_rate) log((), {f"avg_{gen_name}_tokens/s": all_letters / gm.sum()}) log((), {f"avg_{gen_name}_frames/s": all_frames / gm.sum()}) log((), {f"avg_{gen_name}_latency": gm.mean()}) log((), {f"avg_{gen_name}_RTF": rtf}) log((), {f"avg_{gen_name}_RTF@{args.batch_size}": rtf_at}) log((), {f"90%_{gen_name}_latency": gm.mean() + norm.ppf((1.0 + 0.90) / 2) * gm.std()}) log((), {f"95%_{gen_name}_latency": gm.mean() + norm.ppf((1.0 + 0.95) / 2) * gm.std()}) log((), {f"99%_{gen_name}_latency": gm.mean() + norm.ppf((1.0 + 0.99) / 2) * gm.std()}) if args.report_mel_loss: log((), {f"avg_{gen_name}_mel-loss": gen_mel_loss_sum / all_utterances}) if vocoder is not None: vm = np.sort(np.asarray(vocoder_measures)) rtf = all_samples / (all_utterances * vm.mean() * args.sampling_rate) rtf_at = all_samples / (all_batches * vm.mean() * args.sampling_rate) log((), {f"avg_{voc_name}_samples/s": all_samples / vm.sum()}) log((), {f"avg_{voc_name}_latency": vm.mean()}) log((), {f"avg_{voc_name}_RTF": rtf}) log((), {f"avg_{voc_name}_RTF@{args.batch_size}": rtf_at}) log((), {f"90%_{voc_name}_latency": vm.mean() + norm.ppf((1.0 + 0.90) / 2) * vm.std()}) log((), {f"95%_{voc_name}_latency": vm.mean() + norm.ppf((1.0 + 0.95) / 2) * vm.std()}) log((), {f"99%_{voc_name}_latency": vm.mean() + norm.ppf((1.0 + 0.99) / 2) * vm.std()}) if args.report_mel_loss: log((), {f"avg_{voc_name}_mel-loss": voc_mel_loss_sum / all_utterances}) if generator is not None and vocoder is not None: m = gm + vm rtf = all_samples / (all_utterances * m.mean() * args.sampling_rate) rtf_at = all_samples / (all_batches * m.mean() * args.sampling_rate) log((), {"avg_samples/s": all_samples / m.sum()}) log((), {"avg_letters/s": all_letters / m.sum()}) log((), {"avg_latency": m.mean()}) log((), {"avg_RTF": rtf}) log((), {f"avg_RTF@{args.batch_size}": rtf_at}) log((), {"90%_latency": m.mean() + norm.ppf((1.0 + 0.90) / 2) * m.std()}) log((), {"95%_latency": m.mean() + norm.ppf((1.0 + 0.95) / 2) * m.std()}) log((), {"99%_latency": m.mean() + norm.ppf((1.0 + 0.99) / 2) * m.std()}) DLLogger.flush() if __name__ == '__main__': main()
PaddlePaddle/LanguageModeling/BERT/scripts/docker
docker
launch
#!/bin/bash # Copyright (c) 2022 NVIDIA Corporation. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. CMD=${1:-/bin/bash} NV_VISIBLE_DEVICES=${2:-"all"} DOCKER_BRIDGE=${3:-"host"} docker run -it --rm \ --gpus device=$NV_VISIBLE_DEVICES \ --net=$DOCKER_BRIDGE \ --shm-size=1g \ --ulimit memlock=-1 \ --ulimit stack=67108864 \ -v $PWD:/workspace/bert \ -v $PWD/results:/results \ bert $CMD
PyTorch/SpeechSynthesis/FastPitch/waveglow
waveglow
model
# ***************************************************************************** # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # * Neither the name of the NVIDIA CORPORATION nor the # names of its contributors may be used to endorse or promote products # derived from this software without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL NVIDIA CORPORATION BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. # # ***************************************************************************** import torch from torch.autograd import Variable import torch.nn.functional as F @torch.jit.script def fused_add_tanh_sigmoid_multiply(input_a, input_b, n_channels): n_channels_int = n_channels[0] in_act = input_a + input_b t_act = torch.tanh(in_act[:, :n_channels_int, :]) s_act = torch.sigmoid(in_act[:, n_channels_int:, :]) acts = t_act * s_act return acts class Invertible1x1Conv(torch.nn.Module): """ The layer outputs both the convolution, and the log determinant of its weight matrix. If reverse=True it does convolution with inverse """ def __init__(self, c): super(Invertible1x1Conv, self).__init__() self.conv = torch.nn.Conv1d(c, c, kernel_size=1, stride=1, padding=0, bias=False) # Sample a random orthonormal matrix to initialize weights W = torch.qr(torch.FloatTensor(c, c).normal_())[0] # Ensure determinant is 1.0 not -1.0 if torch.det(W) < 0: W[:, 0] = -1 * W[:, 0] W = W.view(c, c, 1) W = W.contiguous() self.conv.weight.data = W def forward(self, z): # shape batch_size, group_size, n_of_groups = z.size() W = self.conv.weight.squeeze() # Forward computation log_det_W = batch_size * n_of_groups * torch.logdet(W.unsqueeze(0).float()).squeeze() z = self.conv(z) return z, log_det_W def infer(self, z): # shape batch_size, group_size, n_of_groups = z.size() W = self.conv.weight.squeeze() if not hasattr(self, 'W_inverse'): # Reverse computation W_inverse = W.float().inverse() W_inverse = Variable(W_inverse[..., None]) if z.type() == 'torch.cuda.HalfTensor' or z.type() == 'torch.HalfTensor': W_inverse = W_inverse.half() self.W_inverse = W_inverse z = F.conv1d(z, self.W_inverse, bias=None, stride=1, padding=0) return z class WN(torch.nn.Module): """ This is the WaveNet like layer for the affine coupling. The primary difference from WaveNet is the convolutions need not be causal. There is also no dilation size reset. The dilation only doubles on each layer """ def __init__(self, n_in_channels, n_mel_channels, n_layers, n_channels, kernel_size): super(WN, self).__init__() assert(kernel_size % 2 == 1) assert(n_channels % 2 == 0) self.n_layers = n_layers self.n_channels = n_channels self.in_layers = torch.nn.ModuleList() self.res_skip_layers = torch.nn.ModuleList() self.cond_layers = torch.nn.ModuleList() start = torch.nn.Conv1d(n_in_channels, n_channels, 1) start = torch.nn.utils.weight_norm(start, name='weight') self.start = start # Initializing last layer to 0 makes the affine coupling layers # do nothing at first. This helps with training stability end = torch.nn.Conv1d(n_channels, 2 * n_in_channels, 1) end.weight.data.zero_() end.bias.data.zero_() self.end = end for i in range(n_layers): dilation = 2 ** i padding = int((kernel_size * dilation - dilation) / 2) in_layer = torch.nn.Conv1d(n_channels, 2 * n_channels, kernel_size, dilation=dilation, padding=padding) in_layer = torch.nn.utils.weight_norm(in_layer, name='weight') self.in_layers.append(in_layer) cond_layer = torch.nn.Conv1d(n_mel_channels, 2 * n_channels, 1) cond_layer = torch.nn.utils.weight_norm(cond_layer, name='weight') self.cond_layers.append(cond_layer) # last one is not necessary if i < n_layers - 1: res_skip_channels = 2 * n_channels else: res_skip_channels = n_channels res_skip_layer = torch.nn.Conv1d(n_channels, res_skip_channels, 1) res_skip_layer = torch.nn.utils.weight_norm( res_skip_layer, name='weight') self.res_skip_layers.append(res_skip_layer) def forward(self, forward_input): audio, spect = forward_input audio = self.start(audio) for i in range(self.n_layers): acts = fused_add_tanh_sigmoid_multiply( self.in_layers[i](audio), self.cond_layers[i](spect), torch.IntTensor([self.n_channels])) res_skip_acts = self.res_skip_layers[i](acts) if i < self.n_layers - 1: audio = res_skip_acts[:, :self.n_channels, :] + audio skip_acts = res_skip_acts[:, self.n_channels:, :] else: skip_acts = res_skip_acts if i == 0: output = skip_acts else: output = skip_acts + output return self.end(output) class WaveGlow(torch.nn.Module): def __init__(self, n_mel_channels, n_flows, n_group, n_early_every, n_early_size, WN_config): super(WaveGlow, self).__init__() self.upsample = torch.nn.ConvTranspose1d(n_mel_channels, n_mel_channels, 1024, stride=256) assert(n_group % 2 == 0) self.n_flows = n_flows self.n_group = n_group self.n_early_every = n_early_every self.n_early_size = n_early_size self.WN = torch.nn.ModuleList() self.convinv = torch.nn.ModuleList() n_half = int(n_group / 2) # Set up layers with the right sizes based on how many dimensions # have been output already n_remaining_channels = n_group for k in range(n_flows): if k % self.n_early_every == 0 and k > 0: n_half = n_half - int(self.n_early_size / 2) n_remaining_channels = n_remaining_channels - self.n_early_size self.convinv.append(Invertible1x1Conv(n_remaining_channels)) self.WN.append(WN(n_half, n_mel_channels * n_group, **WN_config)) self.n_remaining_channels = n_remaining_channels def forward(self, forward_input): """ forward_input[0] = mel_spectrogram: batch x n_mel_channels x frames forward_input[1] = audio: batch x time """ spect, audio = forward_input # Upsample spectrogram to size of audio spect = self.upsample(spect) assert(spect.size(2) >= audio.size(1)) if spect.size(2) > audio.size(1): spect = spect[:, :, :audio.size(1)] spect = spect.unfold(2, self.n_group, self.n_group).permute(0, 2, 1, 3) spect = spect.contiguous().view(spect.size(0), spect.size(1), -1) spect = spect.permute(0, 2, 1) audio = audio.unfold(1, self.n_group, self.n_group).permute(0, 2, 1) output_audio = [] log_s_list = [] log_det_W_list = [] for k in range(self.n_flows): if k % self.n_early_every == 0 and k > 0: output_audio.append(audio[:, :self.n_early_size, :]) audio = audio[:, self.n_early_size:, :] audio, log_det_W = self.convinv[k](audio) log_det_W_list.append(log_det_W) n_half = int(audio.size(1) / 2) audio_0 = audio[:, :n_half, :] audio_1 = audio[:, n_half:, :] output = self.WN[k]((audio_0, spect)) log_s = output[:, n_half:, :] b = output[:, :n_half, :] audio_1 = torch.exp(log_s) * audio_1 + b log_s_list.append(log_s) audio = torch.cat([audio_0, audio_1], 1) output_audio.append(audio) return torch.cat(output_audio, 1), log_s_list, log_det_W_list def infer(self, spect, sigma=1.0): spect = self.upsample(spect) # trim conv artifacts. maybe pad spec to kernel multiple time_cutoff = self.upsample.kernel_size[0] - self.upsample.stride[0] spect = spect[:, :, :-time_cutoff] spect = spect.unfold(2, self.n_group, self.n_group).permute(0, 2, 1, 3) spect = spect.contiguous().view(spect.size(0), spect.size(1), -1) spect = spect.permute(0, 2, 1) audio = torch.randn(spect.size(0), self.n_remaining_channels, spect.size(2), device=spect.device).to(spect.dtype) audio = torch.autograd.Variable(sigma * audio) for k in reversed(range(self.n_flows)): n_half = int(audio.size(1) / 2) audio_0 = audio[:, :n_half, :] audio_1 = audio[:, n_half:, :] output = self.WN[k]((audio_0, spect)) s = output[:, n_half:, :] b = output[:, :n_half, :] audio_1 = (audio_1 - b) / torch.exp(s) audio = torch.cat([audio_0, audio_1], 1) audio = self.convinv[k].infer(audio) if k % self.n_early_every == 0 and k > 0: z = torch.randn(spect.size(0), self.n_early_size, spect.size( 2), device=spect.device).to(spect.dtype) audio = torch.cat((sigma * z, audio), 1) audio = audio.permute( 0, 2, 1).contiguous().view( audio.size(0), -1).data return audio def infer_onnx(self, spect, z, sigma=0.9): spect = self.upsample(spect) # trim conv artifacts. maybe pad spec to kernel multiple time_cutoff = self.upsample.kernel_size[0] - self.upsample.stride[0] spect = spect[:, :, :-time_cutoff] length_spect_group = spect.size(2)//8 mel_dim = 80 batch_size = spect.size(0) spect = spect.view((batch_size, mel_dim, length_spect_group, self.n_group)) spect = spect.permute(0, 2, 1, 3) spect = spect.contiguous() spect = spect.view((batch_size, length_spect_group, self.n_group*mel_dim)) spect = spect.permute(0, 2, 1) spect = spect.contiguous() audio = z[:, :self.n_remaining_channels, :] z = z[:, self.n_remaining_channels:self.n_group, :] audio = sigma*audio for k in reversed(range(self.n_flows)): n_half = int(audio.size(1) // 2) audio_0 = audio[:, :n_half, :] audio_1 = audio[:, n_half:(n_half+n_half), :] output = self.WN[k]((audio_0, spect)) s = output[:, n_half:(n_half+n_half), :] b = output[:, :n_half, :] audio_1 = (audio_1 - b) / torch.exp(s) audio = torch.cat([audio_0, audio_1], 1) audio = self.convinv[k].infer(audio) if k % self.n_early_every == 0 and k > 0: audio = torch.cat((z[:, :self.n_early_size, :], audio), 1) z = z[:, self.n_early_size:self.n_group, :] audio = audio.permute(0,2,1).contiguous().view(batch_size, (length_spect_group * self.n_group)) return audio @staticmethod def remove_weightnorm(model): waveglow = model for WN in waveglow.WN: WN.start = torch.nn.utils.remove_weight_norm(WN.start) WN.in_layers = remove(WN.in_layers) WN.cond_layers = remove(WN.cond_layers) WN.res_skip_layers = remove(WN.res_skip_layers) return waveglow def remove(conv_list): new_conv_list = torch.nn.ModuleList() for old_conv in conv_list: old_conv = torch.nn.utils.remove_weight_norm(old_conv) new_conv_list.append(old_conv) return new_conv_list
PyTorch/SpeechSynthesis/HiFiGAN/common
common
stft
""" BSD 3-Clause License Copyright (c) 2017, Prem Seetharaman All rights reserved. * Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. * Neither the name of the copyright holder nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. """ import numpy as np import torch import torch.nn.functional as F from librosa.util import pad_center, tiny from scipy.signal import get_window from torch.autograd import Variable from common.audio_processing import window_sumsquare class STFT(torch.nn.Module): """adapted from Prem Seetharaman's https://github.com/pseeth/pytorch-stft""" def __init__(self, filter_length=800, hop_length=200, win_length=800, window='hann'): super(STFT, self).__init__() self.filter_length = filter_length self.hop_length = hop_length self.win_length = win_length self.window = window self.forward_transform = None scale = self.filter_length / self.hop_length fourier_basis = np.fft.fft(np.eye(self.filter_length)) cutoff = int((self.filter_length / 2 + 1)) fourier_basis = np.vstack([np.real(fourier_basis[:cutoff, :]), np.imag(fourier_basis[:cutoff, :])]) forward_basis = torch.FloatTensor(fourier_basis[:, None, :]) inverse_basis = torch.FloatTensor( np.linalg.pinv(scale * fourier_basis).T[:, None, :].copy()) if window is not None: assert(filter_length >= win_length) # get window and zero center pad it to filter_length fft_window = get_window(window, win_length, fftbins=True) fft_window = pad_center(fft_window, size=filter_length) fft_window = torch.from_numpy(fft_window).float() # window the bases forward_basis *= fft_window inverse_basis *= fft_window self.register_buffer('forward_basis', forward_basis.float()) self.register_buffer('inverse_basis', inverse_basis.float()) def transform(self, input_data): num_batches = input_data.size(0) num_samples = input_data.size(1) self.num_samples = num_samples # similar to librosa, reflect-pad the input input_data = input_data.view(num_batches, 1, num_samples) input_data = F.pad( input_data.unsqueeze(1), (int(self.filter_length / 2), int(self.filter_length / 2), 0, 0), mode='reflect') input_data = input_data.squeeze(1) forward_transform = F.conv1d( input_data, Variable(self.forward_basis, requires_grad=False), stride=self.hop_length, padding=0) cutoff = int((self.filter_length / 2) + 1) real_part = forward_transform[:, :cutoff, :] imag_part = forward_transform[:, cutoff:, :] magnitude = torch.sqrt(real_part**2 + imag_part**2) phase = torch.autograd.Variable( torch.atan2(imag_part.data, real_part.data)) return magnitude, phase def inverse(self, magnitude, phase): recombine_magnitude_phase = torch.cat( [magnitude*torch.cos(phase), magnitude*torch.sin(phase)], dim=1) with torch.no_grad(): inverse_transform = F.conv_transpose1d( recombine_magnitude_phase, self.inverse_basis, stride=self.hop_length, padding=0) if self.window is not None: window_sum = window_sumsquare( self.window, magnitude.size(-1), hop_length=self.hop_length, win_length=self.win_length, n_fft=self.filter_length, dtype=np.float32) # remove modulation effects approx_nonzero_indices = torch.from_numpy( np.where(window_sum > tiny(window_sum))[0]) window_sum = torch.autograd.Variable( torch.from_numpy(window_sum), requires_grad=False) window_sum = window_sum.cuda() if magnitude.is_cuda else window_sum inverse_transform[:, :, approx_nonzero_indices] /= window_sum[approx_nonzero_indices] # scale by hop ratio inverse_transform *= float(self.filter_length) / self.hop_length inverse_transform = inverse_transform[:, :, int(self.filter_length/2):] inverse_transform = inverse_transform[:, :, :-int(self.filter_length/2):] return inverse_transform def forward(self, input_data): self.magnitude, self.phase = self.transform(input_data) reconstruction = self.inverse(self.magnitude, self.phase) return reconstruction
Tools/DGLPyTorch/SyntheticGraphGeneration/syngen/preprocessing/datasets
datasets
ogbn_mag
# Copyright (c) 2023, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os import shutil from typing import Optional import cudf import cupy as cp import numpy as np import pandas as pd from ogb.nodeproppred import NodePropPredDataset from syngen.configuration import SynGenDatasetFeatureSpec from syngen.preprocessing.base_preprocessing import BasePreprocessing from syngen.utils.io_utils import dump_dataframe from syngen.utils.types import MetaData class OGBN_MAG_Preprocessing(BasePreprocessing): """ The OGBN_MAG_Preprocessing class includes the transformation operation for a subset of the Microsoft Academic Graph (MAG). It's a heterogeneous network that contains four types of entities—papers (736,389 nodes), authors (1,134,649 nodes), institutions (8,740 nodes), and fields of study (59,965 nodes)—as well as four types of directed relations connecting two types of entities—an author is “affiliated with” an institution, an author “writes” a paper, a paper “cites” a paper, and a paper “has a topic of” a field of study. For more information, please check https://ogb.stanford.edu/docs/nodeprop/ """ def __init__( self, source_path: str, destination_path: Optional[str] = None, download: bool = False, **kwargs, ): super().__init__(source_path, destination_path, download, **kwargs) def download(self): NodePropPredDataset(name="ogbn-mag", root=self.source_path) def _check_files(self) -> bool: return True def transform(self, gpu=False, use_cache=False): tabular_operator = cudf if gpu else pd operator = cp if gpu else np if use_cache and os.path.exists(self.destination_path): return SynGenDatasetFeatureSpec.instantiate_from_preprocessed(self.destination_path) shutil.rmtree(self.destination_path, ignore_errors=True) os.makedirs(self.destination_path) dataset = NodePropPredDataset(name="ogbn-mag", root=self.source_path)[0] data = dataset[0] labels = dataset[1]["paper"] graph_metadata = { MetaData.NODES: [], MetaData.EDGES: [], } connections = {} for e, edges in data["edge_index_dict"].items(): structural_data = pd.DataFrame(edges.T, columns=[MetaData.SRC, MetaData.DST]) connections[e[1]] = tabular_operator.DataFrame({ "src_id": edges[0, :], "dst_id": edges[1, :], }) edata = data["edge_reltype"][e] edge_type = { MetaData.NAME: e[1], MetaData.COUNT: len(structural_data), MetaData.SRC_NODE_TYPE: e[0], MetaData.DST_NODE_TYPE: e[2], MetaData.DIRECTED: False, MetaData.FEATURES: [{ MetaData.NAME: 'feat', MetaData.DTYPE: str(edata.dtype), MetaData.FEATURE_TYPE: MetaData.CATEGORICAL, }], MetaData.FEATURES_PATH: f"{e[1]}_features.parquet", MetaData.STRUCTURE_PATH: f"{e[1]}_list.parquet", } dump_dataframe(tabular_operator.DataFrame(edata, columns=['feat']), os.path.join(self.destination_path, edge_type[MetaData.FEATURES_PATH])) dump_dataframe(structural_data, os.path.join(self.destination_path, edge_type[MetaData.STRUCTURE_PATH])) graph_metadata[MetaData.EDGES].append(edge_type) # paper node type continuous_column_names = ["feat_" + str(i) for i in range(data["node_feat_dict"]["paper"].shape[1])] paper_features_dataframe = tabular_operator.DataFrame( data["node_feat_dict"]["paper"], columns=continuous_column_names, ).astype("float32") paper_features_dataframe["year"] = tabular_operator.DataFrame(data["node_year"]["paper"]).astype("int32") paper_features_dataframe["venue"] = tabular_operator.DataFrame(labels).astype("int32") paper_node_type = { MetaData.NAME: "paper", MetaData.COUNT: data["num_nodes_dict"]['paper'], MetaData.FEATURES: [ { MetaData.NAME: name, MetaData.DTYPE: str(dtype), MetaData.FEATURE_TYPE: MetaData.CATEGORICAL if str(dtype).startswith('int') else MetaData.CONTINUOUS, } for name, dtype in paper_features_dataframe.dtypes.items() ], MetaData.FEATURES_PATH: "paper.parquet", } dump_dataframe(paper_features_dataframe, os.path.join(self.destination_path, paper_node_type[MetaData.FEATURES_PATH])) graph_metadata[MetaData.NODES].append(paper_node_type) # author node type paper_features_dataframe["paper_id"] = operator.arange(paper_features_dataframe.shape[0]) author_feat = connections["writes"].merge( paper_features_dataframe, left_on="dst_id", right_on="paper_id", how="left" ).groupby("src_id", sort=True).mean() author_features_dataframe = author_feat[continuous_column_names] author_node_type = { MetaData.NAME: "author", MetaData.COUNT: data["num_nodes_dict"]['author'], MetaData.FEATURES: [ { MetaData.NAME: name, MetaData.DTYPE: str(dtype), MetaData.FEATURE_TYPE: MetaData.CONTINUOUS, } for name, dtype in author_features_dataframe.dtypes.items() ], MetaData.FEATURES_PATH: "author.parquet", } dump_dataframe(author_features_dataframe, os.path.join(self.destination_path, author_node_type[MetaData.FEATURES_PATH])) graph_metadata[MetaData.NODES].append(author_node_type) # institution node type author_features_dataframe["author_id"] = operator.arange(author_features_dataframe.shape[0]) institution_feat = connections["affiliated_with"].merge( author_features_dataframe, left_on="src_id", right_on="author_id" ).groupby("dst_id", sort=True).mean() institution_dataframe = institution_feat[continuous_column_names] institution_node_type = { MetaData.NAME: "institution", MetaData.COUNT: data["num_nodes_dict"]['institution'], MetaData.FEATURES: [ { MetaData.NAME: name, MetaData.DTYPE: str(dtype), MetaData.FEATURE_TYPE: MetaData.CONTINUOUS, } for name, dtype in institution_dataframe.dtypes.items() ], MetaData.FEATURES_PATH: "institution.parquet", } dump_dataframe(institution_dataframe, os.path.join(self.destination_path, institution_node_type[MetaData.FEATURES_PATH])) graph_metadata[MetaData.NODES].append(institution_node_type) # field_of_study node type field_of_study_feat = connections["has_topic"].merge( paper_features_dataframe, left_on="src_id", right_on="paper_id" ).groupby("dst_id", sort=True).mean() field_of_study_dataframe = field_of_study_feat[continuous_column_names] field_of_study_node_type = { MetaData.NAME: "field_of_study", MetaData.COUNT: data["num_nodes_dict"]['field_of_study'], MetaData.FEATURES: [ { MetaData.NAME: name, MetaData.DTYPE: str(dtype), MetaData.FEATURE_TYPE: MetaData.CONTINUOUS, } for name, dtype in field_of_study_dataframe.dtypes.items() ], MetaData.FEATURES_PATH: "field_of_study.parquet", } dump_dataframe(field_of_study_dataframe, os.path.join(self.destination_path, field_of_study_node_type[MetaData.FEATURES_PATH])) graph_metadata[MetaData.NODES].append(field_of_study_node_type) with open(os.path.join(self.destination_path, 'graph_metadata.json'), 'w') as f: json.dump(graph_metadata, f, indent=4) graph_metadata[MetaData.PATH] = self.destination_path return SynGenDatasetFeatureSpec(graph_metadata)
CUDA-Optimized/FastSpeech/tacotron2
tacotron2
layers
# BSD 3-Clause License # Copyright (c) 2018-2020, NVIDIA Corporation # All rights reserved. # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # * Redistributions of source code must retain the above copyright notice, this # list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright notice, # this list of conditions and the following disclaimer in the documentation # and/or other materials provided with the distribution. # * Neither the name of the copyright holder nor the names of its # contributors may be used to endorse or promote products derived from # this software without specific prior written permission. # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE # FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL # DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR # SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER # CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, # OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. """https://github.com/NVIDIA/tacotron2""" import torch from librosa.filters import mel as librosa_mel_fn class LinearNorm(torch.nn.Module): def __init__(self, in_dim, out_dim, bias=True, w_init_gain='linear'): super(LinearNorm, self).__init__() self.linear_layer = torch.nn.Linear(in_dim, out_dim, bias=bias) torch.nn.init.xavier_uniform_( self.linear_layer.weight, gain=torch.nn.init.calculate_gain(w_init_gain)) def forward(self, x): return self.linear_layer(x) class ConvNorm(torch.nn.Module): def __init__(self, in_channels, out_channels, kernel_size=1, stride=1, padding=None, dilation=1, bias=True, w_init_gain='linear'): super(ConvNorm, self).__init__() if padding is None: assert(kernel_size % 2 == 1) padding = int(dilation * (kernel_size - 1) / 2) self.conv = torch.nn.Conv1d(in_channels, out_channels, kernel_size=kernel_size, stride=stride, padding=padding, dilation=dilation, bias=bias) torch.nn.init.xavier_uniform_( self.conv.weight, gain=torch.nn.init.calculate_gain(w_init_gain)) def forward(self, signal): conv_signal = self.conv(signal) return conv_signal
TensorFlow/Recommendation/WideAndDeep/utils
utils
schedulers
# Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import tensorflow as tf __all__ = ['learning_rate_scheduler'] def learning_rate_scheduler(lr_init, warmup_steps, global_step): warmup_lr = (lr_init * tf.cast(global_step, tf.float32) / tf.cast(warmup_steps, tf.float32)) return tf.cond(global_step < warmup_steps, lambda: warmup_lr, lambda: lr_init)
PyTorch/LanguageModeling/BERT/triton/deployment_toolkit/bermuda
bermuda
pyt
# Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import logging import typing from collections import Counter from pathlib import Path from typing import Dict, Optional, Union import numpy as np import torch # pytype: disable=import-error import yaml from model_navigator.model import ModelSignatureConfig from model_navigator.tensor import TensorSpec from model_navigator.utils.config import YamlConfigFile from ..core import ( GET_MODEL_FN_NAME, BaseLoader, BaseRunner, BaseRunnerSession, BaseSaver, Format, Model, Precision, load_from_file, ) from ..extensions import loaders, runners, savers from .utils import get_dynamic_axes, get_shapes_with_dynamic_axes LOGGER = logging.getLogger(__name__) def get_sample_input(dataloader, device): for batch in dataloader: _, x, _ = batch break if isinstance(x, dict): sample_input = list(x.values()) elif isinstance(x, list): sample_input = x else: raise TypeError("The first element (x) of batch returned by dataloader must be a list or a dict") for idx, s in enumerate(sample_input): sample_input[idx] = torch.from_numpy(s).to(device) return tuple(sample_input) def get_model_device(torch_model): if next(torch_model.parameters()).is_cuda: return "cuda" else: return "cpu" def infer_model_precision(model): counter = Counter() for param in model.parameters(): counter[param.dtype] += 1 if counter[torch.float16] > 0: return Precision.FP16 else: return Precision.FP32 def _get_tensor_dtypes(dataloader, precision): def _get_dtypes(t): def _get_dtype(v): dtype = str(v.dtype) if dtype == "float64": dtype = "float32" if precision == Precision.FP16 and dtype == "float32": dtype = "float16" return np.dtype(dtype) return {k: _get_dtype(v) for k, v in t.items()} batch = next(dataloader) _, x, y = batch input_dtypes = _get_dtypes(x) output_dtypes = _get_dtypes(y) return input_dtypes, output_dtypes ### TODO assumption: floating point input ### type has same precision as the model def _get_model_signature( inputs_names: typing.List[str], outputs_names: typing.List[str], precision, dataloader_fn, batch_size_dim: typing.Optional[int] = None, ): dataloader = dataloader_fn() input_dtypes, output_dtypes = _get_tensor_dtypes(dataloader, precision) input_shapes, output_shapes = get_shapes_with_dynamic_axes(dataloader, batch_size_dim=batch_size_dim) inputs = { name: TensorSpec(name=name, dtype=input_dtypes[name], shape=tuple(input_shapes[name])) for name in inputs_names } outputs = { name: TensorSpec(name=name, dtype=output_dtypes[name], shape=tuple(output_shapes[name])) for name in outputs_names } return ModelSignatureConfig(inputs, outputs) class PyTorchModelLoader(BaseLoader): required_fn_name_for_signature_parsing: Optional[str] = GET_MODEL_FN_NAME def __init__(self, **kwargs): self._model_args = kwargs def load(self, model_path: Union[str, Path], **kwargs) -> Model: if isinstance(model_path, Path): model_path = model_path.as_posix() get_model = load_from_file(model_path, "model", GET_MODEL_FN_NAME) model, io_names_dict = get_model(**self._model_args) dataloader_fn = kwargs.get("dataloader_fn", None) output_type = kwargs.get("output_type", None) precision = infer_model_precision(model) batch_axis = getattr(model, "bermuda_batch_axis", 0) # by default models supports batching; batch_axis=0 model_signature = _get_model_signature( inputs_names=io_names_dict["inputs"], outputs_names=io_names_dict["outputs"], precision=precision, dataloader_fn=dataloader_fn, batch_size_dim=batch_axis, ) model = Model(handle=model, precision=precision, inputs=model_signature.inputs, outputs=model_signature.outputs) if output_type == Format.TS_TRACE.value: return self._trace(model, dataloader_fn) elif output_type == Format.TS_SCRIPT.value: return self._script(model) elif output_type == Format.ONNX.value: return model else: raise ValueError(f"Not supported PyTorch format: {output_type}") def _trace(self, model: Model, dataloader_fn) -> Model: device = get_model_device(model.handle) dummy_input = get_sample_input(dataloader_fn(), device) traced_model = torch.jit.trace_module(model.handle, {"forward": dummy_input}) return Model(traced_model, precision=model.precision, inputs=model.inputs, outputs=model.outputs) def _script(self, model: Model) -> Model: scripted_model = torch.jit.script(model.handle) return Model(scripted_model, precision=model.precision, inputs=model.inputs, outputs=model.outputs) class TorchScriptLoader(BaseLoader): def __init__(self, tensor_names_path: str = None, **kwargs): self._model_args = kwargs self._io_spec = None if tensor_names_path is not None: with Path(tensor_names_path).open("r") as fh: tensor_infos = yaml.load(fh, Loader=yaml.SafeLoader) self._io_spec = ModelSignatureConfig(tensor_infos["inputs"], tensor_infos["outputs"]) def load(self, model_path: Union[str, Path], **_) -> Model: if not isinstance(model_path, Path): model_path = Path(model_path) model = torch.jit.load(model_path.as_posix()) precision = infer_model_precision(model) io_spec = self._io_spec if not io_spec: yaml_path = model_path.parent / f"{model_path.name}.yaml" if not yaml_path.is_file(): raise ValueError( f"If `--tensor-names-path is not provided, " f"TorchScript model loader expects file {yaml_path} with tensor information." ) with yaml_path.open("r") as fh: tensor_info = yaml.load(fh, Loader=yaml.SafeLoader) io_spec = ModelSignatureConfig(tensor_info["inputs"], tensor_info["outputs"]) return Model(handle=model, precision=precision, inputs=io_spec.inputs, outputs=io_spec.outputs) class PYT2ONNXSaver(BaseSaver): def __init__(self, onnx_opset: int = None): self._onnx_opset = onnx_opset def save(self, model: Model, model_path: Union[str, Path], dataloader_fn) -> Model: if isinstance(model_path, Path): model_path = model_path.as_posix() assert isinstance(model.handle, torch.jit.ScriptModule) or isinstance( model.handle, torch.nn.Module ), "The model must be of type 'torch.jit.ScriptModule' or 'torch.nn.Module'. Converter aborted." batch_axis = getattr(model.handle, "bermuda_batch_axis", 0) # by default models supports batching; batch_axis=0 dynamic_axes = get_dynamic_axes(dataloader_fn(), batch_size_dim=batch_axis) device = get_model_device(model.handle) dummy_input = get_sample_input(dataloader_fn(), device) with torch.no_grad(): torch.onnx.export( model.handle, dummy_input, model_path, do_constant_folding=True, input_names=list(model.inputs), output_names=list(model.outputs), dynamic_axes=dynamic_axes, opset_version=self._onnx_opset, enable_onnx_checker=True, ) class TorchScriptSaver(BaseSaver): def save(self, model: Model, model_path: Union[str, Path], dataloader_fn) -> None: if not isinstance(model_path, Path): model_path = Path(model_path) if isinstance(model.handle, torch.jit.ScriptModule): torch.jit.save(model.handle, model_path.as_posix()) else: raise RuntimeError("The model must be of type 'torch.jit.ScriptModule'. Saving aborted.") signature_config = ModelSignatureConfig(inputs=model.inputs, outputs=model.outputs) annotation_path = model_path.parent / f"{model_path.name}.yaml" with YamlConfigFile(annotation_path) as config_file: config_file.save_config(signature_config) class PyTorchRunner(BaseRunner): def __init__(self): pass def init_inference(self, model: Model): return PyTorchRunnerSession(model=model) class PyTorchRunnerSession(BaseRunnerSession): def __init__(self, model: Model): super().__init__(model) assert isinstance(model.handle, torch.jit.ScriptModule) or isinstance( model.handle, torch.nn.Module ), "The model must be of type 'torch.jit.ScriptModule' or 'torch.nn.Module'. Runner aborted." self._model = model self._output_names = None def __enter__(self): self._output_names = list(self._model.outputs) return self def __exit__(self, exc_type, exc_value, traceback): self._output_names = None self._model = None def __call__(self, x: Dict[str, object]): with torch.no_grad(): feed_list = [torch.from_numpy(v).cuda() for k, v in x.items()] y_pred = self._model.handle(*feed_list) if isinstance(y_pred, torch.Tensor): y_pred = (y_pred,) y_pred = [t.cpu().numpy() for t in y_pred] y_pred = dict(zip(self._output_names, y_pred)) return y_pred loaders.register_extension(Format.PYT.value, PyTorchModelLoader) loaders.register_extension(Format.TS_TRACE.value, TorchScriptLoader) loaders.register_extension(Format.TS_SCRIPT.value, TorchScriptLoader) savers.register_extension(Format.TS_SCRIPT.value, TorchScriptSaver) savers.register_extension(Format.TS_TRACE.value, TorchScriptSaver) savers.register_extension(f"{Format.PYT.value}--{Format.ONNX.value}", PYT2ONNXSaver) runners.register_extension(Format.PYT.value, PyTorchRunner) runners.register_extension(Format.TS_SCRIPT.value, PyTorchRunner) runners.register_extension(Format.TS_TRACE.value, PyTorchRunner)
PyTorch/SpeechSynthesis/FastPitch/triton/scripts/docker
docker
interactive
#!/usr/bin/env bash # Copyright (c) 2021 NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. docker run -it --rm \ --gpus "device=all" \ --net=host \ --shm-size=1g \ --ulimit memlock=-1 \ --ulimit stack=67108864 \ --privileged \ -e WORKDIR=$(pwd) \ -v $(pwd):$(pwd) \ -v /var/run/docker.sock:/var/run/docker.sock \ fastpitch:latest bash
TensorFlow/Detection/SSD/models/research/object_detection/meta_architectures
meta_architectures
faster_rcnn_meta_arch
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Faster R-CNN meta-architecture definition. General tensorflow implementation of Faster R-CNN detection models. See Faster R-CNN: Ren, Shaoqing, et al. "Faster R-CNN: Towards real-time object detection with region proposal networks." Advances in neural information processing systems. 2015. We allow for three modes: number_of_stages={1, 2, 3}. In case of 1 stage, all of the user facing methods (e.g., predict, postprocess, loss) can be used as if the model consisted only of the RPN, returning class agnostic proposals (these can be thought of as approximate detections with no associated class information). In case of 2 stages, proposals are computed, then passed through a second stage "box classifier" to yield (multi-class) detections. Finally, in case of 3 stages which is only used during eval, proposals are computed, then passed through a second stage "box classifier" that will compute refined boxes and classes, and then features are pooled from the refined and non-maximum suppressed boxes and are passed through the box classifier again. If number of stages is 3 during training it will be reduced to two automatically. Implementations of Faster R-CNN models must define a new FasterRCNNFeatureExtractor and override three methods: `preprocess`, `_extract_proposal_features` (the first stage of the model), and `_extract_box_classifier_features` (the second stage of the model). Optionally, the `restore_fn` method can be overridden. See tests for an example. A few important notes: + Batching conventions: We support batched inference and training where all images within a batch have the same resolution. Batch sizes are determined dynamically via the shape of the input tensors (rather than being specified directly as, e.g., a model constructor). A complication is that due to non-max suppression, we are not guaranteed to get the same number of proposals from the first stage RPN (region proposal network) for each image (though in practice, we should often get the same number of proposals). For this reason we pad to a max number of proposals per image within a batch. This `self.max_num_proposals` property is set to the `first_stage_max_proposals` parameter at inference time and the `second_stage_batch_size` at training time since we subsample the batch to be sent through the box classifier during training. For the second stage of the pipeline, we arrange the proposals for all images within the batch along a single batch dimension. For example, the input to _extract_box_classifier_features is a tensor of shape `[total_num_proposals, crop_height, crop_width, depth]` where total_num_proposals is batch_size * self.max_num_proposals. (And note that per the above comment, a subset of these entries correspond to zero paddings.) + Coordinate representations: Following the API (see model.DetectionModel definition), our outputs after postprocessing operations are always normalized boxes however, internally, we sometimes convert to absolute --- e.g. for loss computation. In particular, anchors and proposal_boxes are both represented as absolute coordinates. Images are resized in the `preprocess` method. The Faster R-CNN meta architecture has two post-processing methods `_postprocess_rpn` which is applied after first stage and `_postprocess_box_classifier` which is applied after second stage. There are three different ways post-processing can happen depending on number_of_stages configured in the meta architecture: 1. When number_of_stages is 1: `_postprocess_rpn` is run as part of the `postprocess` method where true_image_shapes is used to clip proposals, perform non-max suppression and normalize them. 2. When number of stages is 2: `_postprocess_rpn` is run as part of the `_predict_second_stage` method where `resized_image_shapes` is used to clip proposals, perform non-max suppression and normalize them. In this case `postprocess` method skips `_postprocess_rpn` and only runs `_postprocess_box_classifier` using `true_image_shapes` to clip detections, perform non-max suppression and normalize them. 3. When number of stages is 3: `_postprocess_rpn` is run as part of the `_predict_second_stage` using `resized_image_shapes` to clip proposals, perform non-max suppression and normalize them. Subsequently, `_postprocess_box_classifier` is run as part of `_predict_third_stage` using `true_image_shapes` to clip detections, peform non-max suppression and normalize them. In this case, the `postprocess` method skips both `_postprocess_rpn` and `_postprocess_box_classifier`. """ from abc import abstractmethod from functools import partial import tensorflow as tf from object_detection.anchor_generators import grid_anchor_generator from object_detection.builders import box_predictor_builder from object_detection.core import box_list from object_detection.core import box_list_ops from object_detection.core import box_predictor from object_detection.core import losses from object_detection.core import model from object_detection.core import standard_fields as fields from object_detection.core import target_assigner from object_detection.utils import ops from object_detection.utils import shape_utils slim = tf.contrib.slim class FasterRCNNFeatureExtractor(object): """Faster R-CNN Feature Extractor definition.""" def __init__(self, is_training, first_stage_features_stride, batch_norm_trainable=False, reuse_weights=None, weight_decay=0.0): """Constructor. Args: is_training: A boolean indicating whether the training version of the computation graph should be constructed. first_stage_features_stride: Output stride of extracted RPN feature map. batch_norm_trainable: Whether to update batch norm parameters during training or not. When training with a relative large batch size (e.g. 8), it could be desirable to enable batch norm update. reuse_weights: Whether to reuse variables. Default is None. weight_decay: float weight decay for feature extractor (default: 0.0). """ self._is_training = is_training self._first_stage_features_stride = first_stage_features_stride self._train_batch_norm = (batch_norm_trainable and is_training) self._reuse_weights = reuse_weights self._weight_decay = weight_decay @abstractmethod def preprocess(self, resized_inputs): """Feature-extractor specific preprocessing (minus image resizing).""" pass def extract_proposal_features(self, preprocessed_inputs, scope): """Extracts first stage RPN features. This function is responsible for extracting feature maps from preprocessed images. These features are used by the region proposal network (RPN) to predict proposals. Args: preprocessed_inputs: A [batch, height, width, channels] float tensor representing a batch of images. scope: A scope name. Returns: rpn_feature_map: A tensor with shape [batch, height, width, depth] activations: A dictionary mapping activation tensor names to tensors. """ with tf.variable_scope(scope, values=[preprocessed_inputs]): return self._extract_proposal_features(preprocessed_inputs, scope) @abstractmethod def _extract_proposal_features(self, preprocessed_inputs, scope): """Extracts first stage RPN features, to be overridden.""" pass def extract_box_classifier_features(self, proposal_feature_maps, scope): """Extracts second stage box classifier features. Args: proposal_feature_maps: A 4-D float tensor with shape [batch_size * self.max_num_proposals, crop_height, crop_width, depth] representing the feature map cropped to each proposal. scope: A scope name. Returns: proposal_classifier_features: A 4-D float tensor with shape [batch_size * self.max_num_proposals, height, width, depth] representing box classifier features for each proposal. """ with tf.variable_scope( scope, values=[proposal_feature_maps], reuse=tf.AUTO_REUSE): return self._extract_box_classifier_features(proposal_feature_maps, scope) @abstractmethod def _extract_box_classifier_features(self, proposal_feature_maps, scope): """Extracts second stage box classifier features, to be overridden.""" pass def restore_from_classification_checkpoint_fn( self, first_stage_feature_extractor_scope, second_stage_feature_extractor_scope): """Returns a map of variables to load from a foreign checkpoint. Args: first_stage_feature_extractor_scope: A scope name for the first stage feature extractor. second_stage_feature_extractor_scope: A scope name for the second stage feature extractor. Returns: A dict mapping variable names (to load from a checkpoint) to variables in the model graph. """ variables_to_restore = {} for variable in tf.global_variables(): for scope_name in [first_stage_feature_extractor_scope, second_stage_feature_extractor_scope]: if variable.op.name.startswith(scope_name): var_name = variable.op.name.replace(scope_name + '/', '') variables_to_restore[var_name] = variable return variables_to_restore class FasterRCNNMetaArch(model.DetectionModel): """Faster R-CNN Meta-architecture definition.""" def __init__(self, is_training, num_classes, image_resizer_fn, feature_extractor, number_of_stages, first_stage_anchor_generator, first_stage_target_assigner, first_stage_atrous_rate, first_stage_box_predictor_arg_scope_fn, first_stage_box_predictor_kernel_size, first_stage_box_predictor_depth, first_stage_minibatch_size, first_stage_sampler, first_stage_non_max_suppression_fn, first_stage_max_proposals, first_stage_localization_loss_weight, first_stage_objectness_loss_weight, crop_and_resize_fn, initial_crop_size, maxpool_kernel_size, maxpool_stride, second_stage_target_assigner, second_stage_mask_rcnn_box_predictor, second_stage_batch_size, second_stage_sampler, second_stage_non_max_suppression_fn, second_stage_score_conversion_fn, second_stage_localization_loss_weight, second_stage_classification_loss_weight, second_stage_classification_loss, second_stage_mask_prediction_loss_weight=1.0, hard_example_miner=None, parallel_iterations=16, add_summaries=True, clip_anchors_to_image=False, use_static_shapes=False, resize_masks=True): """FasterRCNNMetaArch Constructor. Args: is_training: A boolean indicating whether the training version of the computation graph should be constructed. num_classes: Number of classes. Note that num_classes *does not* include the background category, so if groundtruth labels take values in {0, 1, .., K-1}, num_classes=K (and not K+1, even though the assigned classification targets can range from {0,... K}). image_resizer_fn: A callable for image resizing. This callable takes a rank-3 image tensor of shape [height, width, channels] (corresponding to a single image), an optional rank-3 instance mask tensor of shape [num_masks, height, width] and returns a resized rank-3 image tensor, a resized mask tensor if one was provided in the input. In addition this callable must also return a 1-D tensor of the form [height, width, channels] containing the size of the true image, as the image resizer can perform zero padding. See protos/image_resizer.proto. feature_extractor: A FasterRCNNFeatureExtractor object. number_of_stages: An integer values taking values in {1, 2, 3}. If 1, the function will construct only the Region Proposal Network (RPN) part of the model. If 2, the function will perform box refinement and other auxiliary predictions all in the second stage. If 3, it will extract features from refined boxes and perform the auxiliary predictions on the non-maximum suppressed refined boxes. If is_training is true and the value of number_of_stages is 3, it is reduced to 2 since all the model heads are trained in parallel in second stage during training. first_stage_anchor_generator: An anchor_generator.AnchorGenerator object (note that currently we only support grid_anchor_generator.GridAnchorGenerator objects) first_stage_target_assigner: Target assigner to use for first stage of Faster R-CNN (RPN). first_stage_atrous_rate: A single integer indicating the atrous rate for the single convolution op which is applied to the `rpn_features_to_crop` tensor to obtain a tensor to be used for box prediction. Some feature extractors optionally allow for producing feature maps computed at denser resolutions. The atrous rate is used to compensate for the denser feature maps by using an effectively larger receptive field. (This should typically be set to 1). first_stage_box_predictor_arg_scope_fn: A function to construct tf-slim arg_scope for conv2d, separable_conv2d and fully_connected ops for the RPN box predictor. first_stage_box_predictor_kernel_size: Kernel size to use for the convolution op just prior to RPN box predictions. first_stage_box_predictor_depth: Output depth for the convolution op just prior to RPN box predictions. first_stage_minibatch_size: The "batch size" to use for computing the objectness and location loss of the region proposal network. This "batch size" refers to the number of anchors selected as contributing to the loss function for any given image within the image batch and is only called "batch_size" due to terminology from the Faster R-CNN paper. first_stage_sampler: Sampler to use for first stage loss (RPN loss). first_stage_non_max_suppression_fn: batch_multiclass_non_max_suppression callable that takes `boxes`, `scores` and optional `clip_window`(with all other inputs already set) and returns a dictionary containing tensors with keys: `detection_boxes`, `detection_scores`, `detection_classes`, `num_detections`. This is used to perform non max suppression on the boxes predicted by the Region Proposal Network (RPN). See `post_processing.batch_multiclass_non_max_suppression` for the type and shape of these tensors. first_stage_max_proposals: Maximum number of boxes to retain after performing Non-Max Suppression (NMS) on the boxes predicted by the Region Proposal Network (RPN). first_stage_localization_loss_weight: A float first_stage_objectness_loss_weight: A float crop_and_resize_fn: A differentiable resampler to use for cropping RPN proposal features. initial_crop_size: A single integer indicating the output size (width and height are set to be the same) of the initial bilinear interpolation based cropping during ROI pooling. maxpool_kernel_size: A single integer indicating the kernel size of the max pool op on the cropped feature map during ROI pooling. maxpool_stride: A single integer indicating the stride of the max pool op on the cropped feature map during ROI pooling. second_stage_target_assigner: Target assigner to use for second stage of Faster R-CNN. If the model is configured with multiple prediction heads, this target assigner is used to generate targets for all heads (with the correct `unmatched_class_label`). second_stage_mask_rcnn_box_predictor: Mask R-CNN box predictor to use for the second stage. second_stage_batch_size: The batch size used for computing the classification and refined location loss of the box classifier. This "batch size" refers to the number of proposals selected as contributing to the loss function for any given image within the image batch and is only called "batch_size" due to terminology from the Faster R-CNN paper. second_stage_sampler: Sampler to use for second stage loss (box classifier loss). second_stage_non_max_suppression_fn: batch_multiclass_non_max_suppression callable that takes `boxes`, `scores`, optional `clip_window` and optional (kwarg) `mask` inputs (with all other inputs already set) and returns a dictionary containing tensors with keys: `detection_boxes`, `detection_scores`, `detection_classes`, `num_detections`, and (optionally) `detection_masks`. See `post_processing.batch_multiclass_non_max_suppression` for the type and shape of these tensors. second_stage_score_conversion_fn: Callable elementwise nonlinearity (that takes tensors as inputs and returns tensors). This is usually used to convert logits to probabilities. second_stage_localization_loss_weight: A float indicating the scale factor for second stage localization loss. second_stage_classification_loss_weight: A float indicating the scale factor for second stage classification loss. second_stage_classification_loss: Classification loss used by the second stage classifier. Either losses.WeightedSigmoidClassificationLoss or losses.WeightedSoftmaxClassificationLoss. second_stage_mask_prediction_loss_weight: A float indicating the scale factor for second stage mask prediction loss. This is applicable only if second stage box predictor is configured to predict masks. hard_example_miner: A losses.HardExampleMiner object (can be None). parallel_iterations: (Optional) The number of iterations allowed to run in parallel for calls to tf.map_fn. add_summaries: boolean (default: True) controlling whether summary ops should be added to tensorflow graph. clip_anchors_to_image: Normally, anchors generated for a given image size are pruned during training if they lie outside the image window. This option clips the anchors to be within the image instead of pruning. use_static_shapes: If True, uses implementation of ops with static shape guarantees. resize_masks: Indicates whether the masks presend in the groundtruth should be resized in the model with `image_resizer_fn` Raises: ValueError: If `second_stage_batch_size` > `first_stage_max_proposals` at training time. ValueError: If first_stage_anchor_generator is not of type grid_anchor_generator.GridAnchorGenerator. """ # TODO(rathodv): add_summaries is currently unused. Respect that directive # in the future. super(FasterRCNNMetaArch, self).__init__(num_classes=num_classes) if not isinstance(first_stage_anchor_generator, grid_anchor_generator.GridAnchorGenerator): raise ValueError('first_stage_anchor_generator must be of type ' 'grid_anchor_generator.GridAnchorGenerator.') self._is_training = is_training self._image_resizer_fn = image_resizer_fn self._resize_masks = resize_masks self._feature_extractor = feature_extractor self._number_of_stages = number_of_stages self._proposal_target_assigner = first_stage_target_assigner self._detector_target_assigner = second_stage_target_assigner # Both proposal and detector target assigners use the same box coder self._box_coder = self._proposal_target_assigner.box_coder # (First stage) Region proposal network parameters self._first_stage_anchor_generator = first_stage_anchor_generator self._first_stage_atrous_rate = first_stage_atrous_rate self._first_stage_box_predictor_arg_scope_fn = ( first_stage_box_predictor_arg_scope_fn) self._first_stage_box_predictor_kernel_size = ( first_stage_box_predictor_kernel_size) self._first_stage_box_predictor_depth = first_stage_box_predictor_depth self._first_stage_minibatch_size = first_stage_minibatch_size self._first_stage_sampler = first_stage_sampler self._first_stage_box_predictor = ( box_predictor_builder.build_convolutional_box_predictor( is_training=self._is_training, num_classes=1, conv_hyperparams_fn=self._first_stage_box_predictor_arg_scope_fn, use_dropout=False, dropout_keep_prob=1.0, box_code_size=self._box_coder.code_size, kernel_size=1, num_layers_before_predictor=0, min_depth=0, max_depth=0)) self._first_stage_nms_fn = first_stage_non_max_suppression_fn self._first_stage_max_proposals = first_stage_max_proposals self._use_static_shapes = use_static_shapes self._first_stage_localization_loss = ( losses.WeightedSmoothL1LocalizationLoss()) self._first_stage_objectness_loss = ( losses.WeightedSoftmaxClassificationLoss()) self._first_stage_loc_loss_weight = first_stage_localization_loss_weight self._first_stage_obj_loss_weight = first_stage_objectness_loss_weight # Per-region cropping parameters self._crop_and_resize_fn = crop_and_resize_fn self._initial_crop_size = initial_crop_size self._maxpool_kernel_size = maxpool_kernel_size self._maxpool_stride = maxpool_stride self._mask_rcnn_box_predictor = second_stage_mask_rcnn_box_predictor self._second_stage_batch_size = second_stage_batch_size self._second_stage_sampler = second_stage_sampler self._second_stage_nms_fn = second_stage_non_max_suppression_fn self._second_stage_score_conversion_fn = second_stage_score_conversion_fn self._second_stage_localization_loss = ( losses.WeightedSmoothL1LocalizationLoss()) self._second_stage_classification_loss = second_stage_classification_loss self._second_stage_mask_loss = ( losses.WeightedSigmoidClassificationLoss()) self._second_stage_loc_loss_weight = second_stage_localization_loss_weight self._second_stage_cls_loss_weight = second_stage_classification_loss_weight self._second_stage_mask_loss_weight = ( second_stage_mask_prediction_loss_weight) self._hard_example_miner = hard_example_miner self._parallel_iterations = parallel_iterations self.clip_anchors_to_image = clip_anchors_to_image if self._number_of_stages <= 0 or self._number_of_stages > 3: raise ValueError('Number of stages should be a value in {1, 2, 3}.') @property def first_stage_feature_extractor_scope(self): return 'FirstStageFeatureExtractor' @property def second_stage_feature_extractor_scope(self): return 'SecondStageFeatureExtractor' @property def first_stage_box_predictor_scope(self): return 'FirstStageBoxPredictor' @property def second_stage_box_predictor_scope(self): return 'SecondStageBoxPredictor' @property def max_num_proposals(self): """Max number of proposals (to pad to) for each image in the input batch. At training time, this is set to be the `second_stage_batch_size` if hard example miner is not configured, else it is set to `first_stage_max_proposals`. At inference time, this is always set to `first_stage_max_proposals`. Returns: A positive integer. """ if self._is_training and not self._hard_example_miner: return self._second_stage_batch_size return self._first_stage_max_proposals @property def anchors(self): if not self._anchors: raise RuntimeError('anchors have not been constructed yet!') if not isinstance(self._anchors, box_list.BoxList): raise RuntimeError('anchors should be a BoxList object, but is not.') return self._anchors def preprocess(self, inputs): """Feature-extractor specific preprocessing. See base class. For Faster R-CNN, we perform image resizing in the base class --- each class subclassing FasterRCNNMetaArch is responsible for any additional preprocessing (e.g., scaling pixel values to be in [-1, 1]). Args: inputs: a [batch, height_in, width_in, channels] float tensor representing a batch of images with values between 0 and 255.0. Returns: preprocessed_inputs: a [batch, height_out, width_out, channels] float tensor representing a batch of images. true_image_shapes: int32 tensor of shape [batch, 3] where each row is of the form [height, width, channels] indicating the shapes of true images in the resized images, as resized images can be padded with zeros. Raises: ValueError: if inputs tensor does not have type tf.float32 """ if inputs.dtype is not tf.float32: raise ValueError('`preprocess` expects a tf.float32 tensor') with tf.name_scope('Preprocessor'): outputs = shape_utils.static_or_dynamic_map_fn( self._image_resizer_fn, elems=inputs, dtype=[tf.float32, tf.int32], parallel_iterations=self._parallel_iterations) resized_inputs = outputs[0] true_image_shapes = outputs[1] return (self._feature_extractor.preprocess(resized_inputs), true_image_shapes) def _compute_clip_window(self, image_shapes): """Computes clip window for non max suppression based on image shapes. This function assumes that the clip window's left top corner is at (0, 0). Args: image_shapes: A 2-D int32 tensor of shape [batch_size, 3] containing shapes of images in the batch. Each row represents [height, width, channels] of an image. Returns: A 2-D float32 tensor of shape [batch_size, 4] containing the clip window for each image in the form [ymin, xmin, ymax, xmax]. """ clip_heights = image_shapes[:, 0] clip_widths = image_shapes[:, 1] clip_window = tf.to_float(tf.stack([tf.zeros_like(clip_heights), tf.zeros_like(clip_heights), clip_heights, clip_widths], axis=1)) return clip_window def predict(self, preprocessed_inputs, true_image_shapes): """Predicts unpostprocessed tensors from input tensor. This function takes an input batch of images and runs it through the forward pass of the network to yield "raw" un-postprocessed predictions. If `number_of_stages` is 1, this function only returns first stage RPN predictions (un-postprocessed). Otherwise it returns both first stage RPN predictions as well as second stage box classifier predictions. Other remarks: + Anchor pruning vs. clipping: following the recommendation of the Faster R-CNN paper, we prune anchors that venture outside the image window at training time and clip anchors to the image window at inference time. + Proposal padding: as described at the top of the file, proposals are padded to self._max_num_proposals and flattened so that proposals from all images within the input batch are arranged along the same batch dimension. Args: preprocessed_inputs: a [batch, height, width, channels] float tensor representing a batch of images. true_image_shapes: int32 tensor of shape [batch, 3] where each row is of the form [height, width, channels] indicating the shapes of true images in the resized images, as resized images can be padded with zeros. Returns: prediction_dict: a dictionary holding "raw" prediction tensors: 1) rpn_box_predictor_features: A 4-D float32 tensor with shape [batch_size, height, width, depth] to be used for predicting proposal boxes and corresponding objectness scores. 2) rpn_features_to_crop: A 4-D float32 tensor with shape [batch_size, height, width, depth] representing image features to crop using the proposal boxes predicted by the RPN. 3) image_shape: a 1-D tensor of shape [4] representing the input image shape. 4) rpn_box_encodings: 3-D float tensor of shape [batch_size, num_anchors, self._box_coder.code_size] containing predicted boxes. 5) rpn_objectness_predictions_with_background: 3-D float tensor of shape [batch_size, num_anchors, 2] containing class predictions (logits) for each of the anchors. Note that this tensor *includes* background class predictions (at class index 0). 6) anchors: A 2-D tensor of shape [num_anchors, 4] representing anchors for the first stage RPN (in absolute coordinates). Note that `num_anchors` can differ depending on whether the model is created in training or inference mode. (and if number_of_stages > 1): 7) refined_box_encodings: a 3-D tensor with shape [total_num_proposals, num_classes, self._box_coder.code_size] representing predicted (final) refined box encodings, where total_num_proposals=batch_size*self._max_num_proposals. If using a shared box across classes the shape will instead be [total_num_proposals, 1, self._box_coder.code_size]. 8) class_predictions_with_background: a 3-D tensor with shape [total_num_proposals, num_classes + 1] containing class predictions (logits) for each of the anchors, where total_num_proposals=batch_size*self._max_num_proposals. Note that this tensor *includes* background class predictions (at class index 0). 9) num_proposals: An int32 tensor of shape [batch_size] representing the number of proposals generated by the RPN. `num_proposals` allows us to keep track of which entries are to be treated as zero paddings and which are not since we always pad the number of proposals to be `self.max_num_proposals` for each image. 10) proposal_boxes: A float32 tensor of shape [batch_size, self.max_num_proposals, 4] representing decoded proposal bounding boxes in absolute coordinates. 11) mask_predictions: (optional) a 4-D tensor with shape [total_num_padded_proposals, num_classes, mask_height, mask_width] containing instance mask predictions. Raises: ValueError: If `predict` is called before `preprocess`. """ (rpn_box_predictor_features, rpn_features_to_crop, anchors_boxlist, image_shape) = self._extract_rpn_feature_maps(preprocessed_inputs) (rpn_box_encodings, rpn_objectness_predictions_with_background ) = self._predict_rpn_proposals(rpn_box_predictor_features) # The Faster R-CNN paper recommends pruning anchors that venture outside # the image window at training time and clipping at inference time. clip_window = tf.to_float(tf.stack([0, 0, image_shape[1], image_shape[2]])) if self._is_training: if self.clip_anchors_to_image: anchors_boxlist = box_list_ops.clip_to_window( anchors_boxlist, clip_window, filter_nonoverlapping=False) else: (rpn_box_encodings, rpn_objectness_predictions_with_background, anchors_boxlist) = self._remove_invalid_anchors_and_predictions( rpn_box_encodings, rpn_objectness_predictions_with_background, anchors_boxlist, clip_window) else: anchors_boxlist = box_list_ops.clip_to_window( anchors_boxlist, clip_window, filter_nonoverlapping=not self._use_static_shapes) self._anchors = anchors_boxlist prediction_dict = { 'rpn_box_predictor_features': rpn_box_predictor_features, 'rpn_features_to_crop': rpn_features_to_crop, 'image_shape': image_shape, 'rpn_box_encodings': rpn_box_encodings, 'rpn_objectness_predictions_with_background': rpn_objectness_predictions_with_background, 'anchors': self._anchors.get() } if self._number_of_stages >= 2: # If mixed-precision training on TPU is enabled, rpn_box_encodings and # rpn_objectness_predictions_with_background are bfloat16 tensors. # Considered prediction results, they need to be casted to float32 # tensors for correct postprocess_rpn computation in predict_second_stage. prediction_dict.update(self._predict_second_stage( tf.to_float(rpn_box_encodings), tf.to_float(rpn_objectness_predictions_with_background), rpn_features_to_crop, self._anchors.get(), image_shape, true_image_shapes)) if self._number_of_stages == 3: prediction_dict = self._predict_third_stage( prediction_dict, true_image_shapes) return prediction_dict def _image_batch_shape_2d(self, image_batch_shape_1d): """Takes a 1-D image batch shape tensor and converts it to a 2-D tensor. Example: If 1-D image batch shape tensor is [2, 300, 300, 3]. The corresponding 2-D image batch tensor would be [[300, 300, 3], [300, 300, 3]] Args: image_batch_shape_1d: 1-D tensor of the form [batch_size, height, width, channels]. Returns: image_batch_shape_2d: 2-D tensor of shape [batch_size, 3] were each row is of the form [height, width, channels]. """ return tf.tile(tf.expand_dims(image_batch_shape_1d[1:], 0), [image_batch_shape_1d[0], 1]) def _predict_second_stage(self, rpn_box_encodings, rpn_objectness_predictions_with_background, rpn_features_to_crop, anchors, image_shape, true_image_shapes): """Predicts the output tensors from second stage of Faster R-CNN. Args: rpn_box_encodings: 4-D float tensor of shape [batch_size, num_valid_anchors, self._box_coder.code_size] containing predicted boxes. rpn_objectness_predictions_with_background: 2-D float tensor of shape [batch_size, num_valid_anchors, 2] containing class predictions (logits) for each of the anchors. Note that this tensor *includes* background class predictions (at class index 0). rpn_features_to_crop: A 4-D float32 or bfloat16 tensor with shape [batch_size, height, width, depth] representing image features to crop using the proposal boxes predicted by the RPN. anchors: 2-D float tensor of shape [num_anchors, self._box_coder.code_size]. image_shape: A 1D int32 tensors of size [4] containing the image shape. true_image_shapes: int32 tensor of shape [batch, 3] where each row is of the form [height, width, channels] indicating the shapes of true images in the resized images, as resized images can be padded with zeros. Returns: prediction_dict: a dictionary holding "raw" prediction tensors: 1) refined_box_encodings: a 3-D tensor with shape [total_num_proposals, num_classes, self._box_coder.code_size] representing predicted (final) refined box encodings, where total_num_proposals=batch_size*self._max_num_proposals. If using a shared box across classes the shape will instead be [total_num_proposals, 1, self._box_coder.code_size]. 2) class_predictions_with_background: a 3-D tensor with shape [total_num_proposals, num_classes + 1] containing class predictions (logits) for each of the anchors, where total_num_proposals=batch_size*self._max_num_proposals. Note that this tensor *includes* background class predictions (at class index 0). 3) num_proposals: An int32 tensor of shape [batch_size] representing the number of proposals generated by the RPN. `num_proposals` allows us to keep track of which entries are to be treated as zero paddings and which are not since we always pad the number of proposals to be `self.max_num_proposals` for each image. 4) proposal_boxes: A float32 tensor of shape [batch_size, self.max_num_proposals, 4] representing decoded proposal bounding boxes in absolute coordinates. 5) proposal_boxes_normalized: A float32 tensor of shape [batch_size, self.max_num_proposals, 4] representing decoded proposal bounding boxes in normalized coordinates. Can be used to override the boxes proposed by the RPN, thus enabling one to extract features and get box classification and prediction for externally selected areas of the image. 6) box_classifier_features: a 4-D float32 or bfloat16 tensor representing the features for each proposal. """ image_shape_2d = self._image_batch_shape_2d(image_shape) proposal_boxes_normalized, _, num_proposals = self._postprocess_rpn( rpn_box_encodings, rpn_objectness_predictions_with_background, anchors, image_shape_2d, true_image_shapes) # If mixed-precision training on TPU is enabled, the dtype of # rpn_features_to_crop is bfloat16, otherwise it is float32. tf.cast is # used to match the dtype of proposal_boxes_normalized to that of # rpn_features_to_crop for further computation. flattened_proposal_feature_maps = ( self._compute_second_stage_input_feature_maps( rpn_features_to_crop, tf.cast(proposal_boxes_normalized, rpn_features_to_crop.dtype))) box_classifier_features = ( self._feature_extractor.extract_box_classifier_features( flattened_proposal_feature_maps, scope=self.second_stage_feature_extractor_scope)) if self._mask_rcnn_box_predictor.is_keras_model: box_predictions = self._mask_rcnn_box_predictor( [box_classifier_features], prediction_stage=2) else: box_predictions = self._mask_rcnn_box_predictor.predict( [box_classifier_features], num_predictions_per_location=[1], scope=self.second_stage_box_predictor_scope, prediction_stage=2) refined_box_encodings = tf.squeeze( box_predictions[box_predictor.BOX_ENCODINGS], axis=1, name='all_refined_box_encodings') class_predictions_with_background = tf.squeeze( box_predictions[box_predictor.CLASS_PREDICTIONS_WITH_BACKGROUND], axis=1, name='all_class_predictions_with_background') absolute_proposal_boxes = ops.normalized_to_image_coordinates( proposal_boxes_normalized, image_shape, self._parallel_iterations) prediction_dict = { 'refined_box_encodings': refined_box_encodings, 'class_predictions_with_background': class_predictions_with_background, 'num_proposals': num_proposals, 'proposal_boxes': absolute_proposal_boxes, 'box_classifier_features': box_classifier_features, 'proposal_boxes_normalized': proposal_boxes_normalized, } return prediction_dict def _predict_third_stage(self, prediction_dict, image_shapes): """Predicts non-box, non-class outputs using refined detections. For training, masks as predicted directly on the box_classifier_features, which are region-features from the initial anchor boxes. For inference, this happens after calling the post-processing stage, such that masks are only calculated for the top scored boxes. Args: prediction_dict: a dictionary holding "raw" prediction tensors: 1) refined_box_encodings: a 3-D tensor with shape [total_num_proposals, num_classes, self._box_coder.code_size] representing predicted (final) refined box encodings, where total_num_proposals=batch_size*self._max_num_proposals. If using a shared box across classes the shape will instead be [total_num_proposals, 1, self._box_coder.code_size]. 2) class_predictions_with_background: a 3-D tensor with shape [total_num_proposals, num_classes + 1] containing class predictions (logits) for each of the anchors, where total_num_proposals=batch_size*self._max_num_proposals. Note that this tensor *includes* background class predictions (at class index 0). 3) num_proposals: An int32 tensor of shape [batch_size] representing the number of proposals generated by the RPN. `num_proposals` allows us to keep track of which entries are to be treated as zero paddings and which are not since we always pad the number of proposals to be `self.max_num_proposals` for each image. 4) proposal_boxes: A float32 tensor of shape [batch_size, self.max_num_proposals, 4] representing decoded proposal bounding boxes in absolute coordinates. 5) box_classifier_features: a 4-D float32 tensor representing the features for each proposal. image_shapes: A 2-D int32 tensors of shape [batch_size, 3] containing shapes of images in the batch. Returns: prediction_dict: a dictionary that in addition to the input predictions does hold the following predictions as well: 1) mask_predictions: a 4-D tensor with shape [batch_size, max_detection, mask_height, mask_width] containing instance mask predictions. """ if self._is_training: curr_box_classifier_features = prediction_dict['box_classifier_features'] detection_classes = prediction_dict['class_predictions_with_background'] if self._mask_rcnn_box_predictor.is_keras_model: mask_predictions = self._mask_rcnn_box_predictor( [curr_box_classifier_features], prediction_stage=3) else: mask_predictions = self._mask_rcnn_box_predictor.predict( [curr_box_classifier_features], num_predictions_per_location=[1], scope=self.second_stage_box_predictor_scope, prediction_stage=3) prediction_dict['mask_predictions'] = tf.squeeze(mask_predictions[ box_predictor.MASK_PREDICTIONS], axis=1) else: detections_dict = self._postprocess_box_classifier( prediction_dict['refined_box_encodings'], prediction_dict['class_predictions_with_background'], prediction_dict['proposal_boxes'], prediction_dict['num_proposals'], image_shapes) prediction_dict.update(detections_dict) detection_boxes = detections_dict[ fields.DetectionResultFields.detection_boxes] detection_classes = detections_dict[ fields.DetectionResultFields.detection_classes] rpn_features_to_crop = prediction_dict['rpn_features_to_crop'] batch_size = tf.shape(detection_boxes)[0] max_detection = tf.shape(detection_boxes)[1] flattened_detected_feature_maps = ( self._compute_second_stage_input_feature_maps( rpn_features_to_crop, detection_boxes)) curr_box_classifier_features = ( self._feature_extractor.extract_box_classifier_features( flattened_detected_feature_maps, scope=self.second_stage_feature_extractor_scope)) if self._mask_rcnn_box_predictor.is_keras_model: mask_predictions = self._mask_rcnn_box_predictor( [curr_box_classifier_features], prediction_stage=3) else: mask_predictions = self._mask_rcnn_box_predictor.predict( [curr_box_classifier_features], num_predictions_per_location=[1], scope=self.second_stage_box_predictor_scope, prediction_stage=3) detection_masks = tf.squeeze(mask_predictions[ box_predictor.MASK_PREDICTIONS], axis=1) _, num_classes, mask_height, mask_width = ( detection_masks.get_shape().as_list()) _, max_detection = detection_classes.get_shape().as_list() prediction_dict['mask_predictions'] = tf.reshape( detection_masks, [-1, num_classes, mask_height, mask_width]) if num_classes > 1: detection_masks = self._gather_instance_masks( detection_masks, detection_classes) prediction_dict[fields.DetectionResultFields.detection_masks] = ( tf.reshape(tf.sigmoid(detection_masks), [batch_size, max_detection, mask_height, mask_width])) return prediction_dict def _gather_instance_masks(self, instance_masks, classes): """Gathers the masks that correspond to classes. Args: instance_masks: A 4-D float32 tensor with shape [K, num_classes, mask_height, mask_width]. classes: A 2-D int32 tensor with shape [batch_size, max_detection]. Returns: masks: a 3-D float32 tensor with shape [K, mask_height, mask_width]. """ _, num_classes, height, width = instance_masks.get_shape().as_list() k = tf.shape(instance_masks)[0] instance_masks = tf.reshape(instance_masks, [-1, height, width]) classes = tf.to_int32(tf.reshape(classes, [-1])) gather_idx = tf.range(k) * num_classes + classes return tf.gather(instance_masks, gather_idx) def _extract_rpn_feature_maps(self, preprocessed_inputs): """Extracts RPN features. This function extracts two feature maps: a feature map to be directly fed to a box predictor (to predict location and objectness scores for proposals) and a feature map from which to crop regions which will then be sent to the second stage box classifier. Args: preprocessed_inputs: a [batch, height, width, channels] image tensor. Returns: rpn_box_predictor_features: A 4-D float32 tensor with shape [batch, height, width, depth] to be used for predicting proposal boxes and corresponding objectness scores. rpn_features_to_crop: A 4-D float32 tensor with shape [batch, height, width, depth] representing image features to crop using the proposals boxes. anchors: A BoxList representing anchors (for the RPN) in absolute coordinates. image_shape: A 1-D tensor representing the input image shape. """ image_shape = tf.shape(preprocessed_inputs) rpn_features_to_crop, self.endpoints = ( self._feature_extractor.extract_proposal_features( preprocessed_inputs, scope=self.first_stage_feature_extractor_scope)) feature_map_shape = tf.shape(rpn_features_to_crop) anchors = box_list_ops.concatenate( self._first_stage_anchor_generator.generate([(feature_map_shape[1], feature_map_shape[2])])) with slim.arg_scope(self._first_stage_box_predictor_arg_scope_fn()): kernel_size = self._first_stage_box_predictor_kernel_size reuse = tf.get_variable_scope().reuse rpn_box_predictor_features = slim.conv2d( rpn_features_to_crop, self._first_stage_box_predictor_depth, kernel_size=[kernel_size, kernel_size], rate=self._first_stage_atrous_rate, activation_fn=tf.nn.relu6, scope='Conv', reuse=reuse) return (rpn_box_predictor_features, rpn_features_to_crop, anchors, image_shape) def _predict_rpn_proposals(self, rpn_box_predictor_features): """Adds box predictors to RPN feature map to predict proposals. Note resulting tensors will not have been postprocessed. Args: rpn_box_predictor_features: A 4-D float32 tensor with shape [batch, height, width, depth] to be used for predicting proposal boxes and corresponding objectness scores. Returns: box_encodings: 3-D float tensor of shape [batch_size, num_anchors, self._box_coder.code_size] containing predicted boxes. objectness_predictions_with_background: 3-D float tensor of shape [batch_size, num_anchors, 2] containing class predictions (logits) for each of the anchors. Note that this tensor *includes* background class predictions (at class index 0). Raises: RuntimeError: if the anchor generator generates anchors corresponding to multiple feature maps. We currently assume that a single feature map is generated for the RPN. """ num_anchors_per_location = ( self._first_stage_anchor_generator.num_anchors_per_location()) if len(num_anchors_per_location) != 1: raise RuntimeError('anchor_generator is expected to generate anchors ' 'corresponding to a single feature map.') if self._first_stage_box_predictor.is_keras_model: box_predictions = self._first_stage_box_predictor( [rpn_box_predictor_features]) else: box_predictions = self._first_stage_box_predictor.predict( [rpn_box_predictor_features], num_anchors_per_location, scope=self.first_stage_box_predictor_scope) box_encodings = tf.concat( box_predictions[box_predictor.BOX_ENCODINGS], axis=1) objectness_predictions_with_background = tf.concat( box_predictions[box_predictor.CLASS_PREDICTIONS_WITH_BACKGROUND], axis=1) return (tf.squeeze(box_encodings, axis=2), objectness_predictions_with_background) def _remove_invalid_anchors_and_predictions( self, box_encodings, objectness_predictions_with_background, anchors_boxlist, clip_window): """Removes anchors that (partially) fall outside an image. Also removes associated box encodings and objectness predictions. Args: box_encodings: 3-D float tensor of shape [batch_size, num_anchors, self._box_coder.code_size] containing predicted boxes. objectness_predictions_with_background: 3-D float tensor of shape [batch_size, num_anchors, 2] containing class predictions (logits) for each of the anchors. Note that this tensor *includes* background class predictions (at class index 0). anchors_boxlist: A BoxList representing num_anchors anchors (for the RPN) in absolute coordinates. clip_window: a 1-D tensor representing the [ymin, xmin, ymax, xmax] extent of the window to clip/prune to. Returns: box_encodings: 4-D float tensor of shape [batch_size, num_valid_anchors, self._box_coder.code_size] containing predicted boxes, where num_valid_anchors <= num_anchors objectness_predictions_with_background: 2-D float tensor of shape [batch_size, num_valid_anchors, 2] containing class predictions (logits) for each of the anchors, where num_valid_anchors <= num_anchors. Note that this tensor *includes* background class predictions (at class index 0). anchors: A BoxList representing num_valid_anchors anchors (for the RPN) in absolute coordinates. """ pruned_anchors_boxlist, keep_indices = box_list_ops.prune_outside_window( anchors_boxlist, clip_window) def _batch_gather_kept_indices(predictions_tensor): return shape_utils.static_or_dynamic_map_fn( partial(tf.gather, indices=keep_indices), elems=predictions_tensor, dtype=tf.float32, parallel_iterations=self._parallel_iterations, back_prop=True) return (_batch_gather_kept_indices(box_encodings), _batch_gather_kept_indices(objectness_predictions_with_background), pruned_anchors_boxlist) def _flatten_first_two_dimensions(self, inputs): """Flattens `K-d` tensor along batch dimension to be a `(K-1)-d` tensor. Converts `inputs` with shape [A, B, ..., depth] into a tensor of shape [A * B, ..., depth]. Args: inputs: A float tensor with shape [A, B, ..., depth]. Note that the first two and last dimensions must be statically defined. Returns: A float tensor with shape [A * B, ..., depth] (where the first and last dimension are statically defined. """ combined_shape = shape_utils.combined_static_and_dynamic_shape(inputs) flattened_shape = tf.stack([combined_shape[0] * combined_shape[1]] + combined_shape[2:]) return tf.reshape(inputs, flattened_shape) def postprocess(self, prediction_dict, true_image_shapes): """Convert prediction tensors to final detections. This function converts raw predictions tensors to final detection results. See base class for output format conventions. Note also that by default, scores are to be interpreted as logits, but if a score_converter is used, then scores are remapped (and may thus have a different interpretation). If number_of_stages=1, the returned results represent proposals from the first stage RPN and are padded to have self.max_num_proposals for each image; otherwise, the results can be interpreted as multiclass detections from the full two-stage model and are padded to self._max_detections. Args: prediction_dict: a dictionary holding prediction tensors (see the documentation for the predict method. If number_of_stages=1, we expect prediction_dict to contain `rpn_box_encodings`, `rpn_objectness_predictions_with_background`, `rpn_features_to_crop`, and `anchors` fields. Otherwise we expect prediction_dict to additionally contain `refined_box_encodings`, `class_predictions_with_background`, `num_proposals`, `proposal_boxes` and, optionally, `mask_predictions` fields. true_image_shapes: int32 tensor of shape [batch, 3] where each row is of the form [height, width, channels] indicating the shapes of true images in the resized images, as resized images can be padded with zeros. Returns: detections: a dictionary containing the following fields detection_boxes: [batch, max_detection, 4] detection_scores: [batch, max_detections] detection_classes: [batch, max_detections] (this entry is only created if rpn_mode=False) num_detections: [batch] Raises: ValueError: If `predict` is called before `preprocess`. """ with tf.name_scope('FirstStagePostprocessor'): if self._number_of_stages == 1: proposal_boxes, proposal_scores, num_proposals = self._postprocess_rpn( prediction_dict['rpn_box_encodings'], prediction_dict['rpn_objectness_predictions_with_background'], prediction_dict['anchors'], true_image_shapes, true_image_shapes) return { fields.DetectionResultFields.detection_boxes: proposal_boxes, fields.DetectionResultFields.detection_scores: proposal_scores, fields.DetectionResultFields.num_detections: tf.to_float(num_proposals), } # TODO(jrru): Remove mask_predictions from _post_process_box_classifier. if (self._number_of_stages == 2 or (self._number_of_stages == 3 and self._is_training)): with tf.name_scope('SecondStagePostprocessor'): mask_predictions = prediction_dict.get(box_predictor.MASK_PREDICTIONS) detections_dict = self._postprocess_box_classifier( prediction_dict['refined_box_encodings'], prediction_dict['class_predictions_with_background'], prediction_dict['proposal_boxes'], prediction_dict['num_proposals'], true_image_shapes, mask_predictions=mask_predictions) if 'rpn_features_to_crop' in prediction_dict and self._initial_crop_size: self._add_detection_features_output_node( detections_dict[fields.DetectionResultFields.detection_boxes], prediction_dict['rpn_features_to_crop']) return detections_dict if self._number_of_stages == 3: # Post processing is already performed in 3rd stage. We need to transfer # postprocessed tensors from `prediction_dict` to `detections_dict`. return prediction_dict def _add_detection_features_output_node(self, detection_boxes, rpn_features_to_crop): """Add the detection features to the output node. The detection features are from cropping rpn_features with boxes. Each bounding box has one feature vector of length depth, which comes from mean_pooling of the cropped rpn_features. Args: detection_boxes: a 3-D float32 tensor of shape [batch_size, max_detection, 4] which represents the bounding boxes. rpn_features_to_crop: A 4-D float32 tensor with shape [batch, height, width, depth] representing image features to crop using the proposals boxes. """ with tf.name_scope('SecondStageDetectionFeaturesExtract'): flattened_detected_feature_maps = ( self._compute_second_stage_input_feature_maps( rpn_features_to_crop, detection_boxes)) detection_features_unpooled = ( self._feature_extractor.extract_box_classifier_features( flattened_detected_feature_maps, scope=self.second_stage_feature_extractor_scope)) batch_size = tf.shape(detection_boxes)[0] max_detection = tf.shape(detection_boxes)[1] detection_features_pool = tf.reduce_mean( detection_features_unpooled, axis=[1, 2]) detection_features = tf.reshape( detection_features_pool, [batch_size, max_detection, tf.shape(detection_features_pool)[-1]]) detection_features = tf.identity( detection_features, 'detection_features') def _postprocess_rpn(self, rpn_box_encodings_batch, rpn_objectness_predictions_with_background_batch, anchors, image_shapes, true_image_shapes): """Converts first stage prediction tensors from the RPN to proposals. This function decodes the raw RPN predictions, runs non-max suppression on the result. Note that the behavior of this function is slightly modified during training --- specifically, we stop the gradient from passing through the proposal boxes and we only return a balanced sampled subset of proposals with size `second_stage_batch_size`. Args: rpn_box_encodings_batch: A 3-D float32 tensor of shape [batch_size, num_anchors, self._box_coder.code_size] containing predicted proposal box encodings. rpn_objectness_predictions_with_background_batch: A 3-D float tensor of shape [batch_size, num_anchors, 2] containing objectness predictions (logits) for each of the anchors with 0 corresponding to background and 1 corresponding to object. anchors: A 2-D tensor of shape [num_anchors, 4] representing anchors for the first stage RPN. Note that `num_anchors` can differ depending on whether the model is created in training or inference mode. image_shapes: A 2-D tensor of shape [batch, 3] containing the shapes of images in the batch. true_image_shapes: int32 tensor of shape [batch, 3] where each row is of the form [height, width, channels] indicating the shapes of true images in the resized images, as resized images can be padded with zeros. Returns: proposal_boxes: A float tensor with shape [batch_size, max_num_proposals, 4] representing the (potentially zero padded) proposal boxes for all images in the batch. These boxes are represented as normalized coordinates. proposal_scores: A float tensor with shape [batch_size, max_num_proposals] representing the (potentially zero padded) proposal objectness scores for all images in the batch. num_proposals: A Tensor of type `int32`. A 1-D tensor of shape [batch] representing the number of proposals predicted for each image in the batch. """ rpn_box_encodings_batch = tf.expand_dims(rpn_box_encodings_batch, axis=2) rpn_encodings_shape = shape_utils.combined_static_and_dynamic_shape( rpn_box_encodings_batch) tiled_anchor_boxes = tf.tile( tf.expand_dims(anchors, 0), [rpn_encodings_shape[0], 1, 1]) proposal_boxes = self._batch_decode_boxes(rpn_box_encodings_batch, tiled_anchor_boxes) proposal_boxes = tf.squeeze(proposal_boxes, axis=2) rpn_objectness_softmax_without_background = tf.nn.softmax( rpn_objectness_predictions_with_background_batch)[:, :, 1] clip_window = self._compute_clip_window(image_shapes) (proposal_boxes, proposal_scores, _, _, _, num_proposals) = self._first_stage_nms_fn( tf.expand_dims(proposal_boxes, axis=2), tf.expand_dims(rpn_objectness_softmax_without_background, axis=2), clip_window=clip_window) if self._is_training: proposal_boxes = tf.stop_gradient(proposal_boxes) if not self._hard_example_miner: (groundtruth_boxlists, groundtruth_classes_with_background_list, _, groundtruth_weights_list ) = self._format_groundtruth_data(true_image_shapes) (proposal_boxes, proposal_scores, num_proposals) = self._sample_box_classifier_batch( proposal_boxes, proposal_scores, num_proposals, groundtruth_boxlists, groundtruth_classes_with_background_list, groundtruth_weights_list) # normalize proposal boxes def normalize_boxes(args): proposal_boxes_per_image = args[0] image_shape = args[1] normalized_boxes_per_image = box_list_ops.to_normalized_coordinates( box_list.BoxList(proposal_boxes_per_image), image_shape[0], image_shape[1], check_range=False).get() return normalized_boxes_per_image normalized_proposal_boxes = shape_utils.static_or_dynamic_map_fn( normalize_boxes, elems=[proposal_boxes, image_shapes], dtype=tf.float32) return normalized_proposal_boxes, proposal_scores, num_proposals def _sample_box_classifier_batch( self, proposal_boxes, proposal_scores, num_proposals, groundtruth_boxlists, groundtruth_classes_with_background_list, groundtruth_weights_list): """Samples a minibatch for second stage. Args: proposal_boxes: A float tensor with shape [batch_size, num_proposals, 4] representing the (potentially zero padded) proposal boxes for all images in the batch. These boxes are represented in absolute coordinates. proposal_scores: A float tensor with shape [batch_size, num_proposals] representing the (potentially zero padded) proposal objectness scores for all images in the batch. num_proposals: A Tensor of type `int32`. A 1-D tensor of shape [batch] representing the number of proposals predicted for each image in the batch. groundtruth_boxlists: A list of BoxLists containing (absolute) coordinates of the groundtruth boxes. groundtruth_classes_with_background_list: A list of 2-D one-hot (or k-hot) tensors of shape [num_boxes, num_classes+1] containing the class targets with the 0th index assumed to map to the background class. groundtruth_weights_list: A list of 1-D tensors of shape [num_boxes] indicating the weight associated with the groundtruth boxes. Returns: proposal_boxes: A float tensor with shape [batch_size, second_stage_batch_size, 4] representing the (potentially zero padded) proposal boxes for all images in the batch. These boxes are represented in absolute coordinates. proposal_scores: A float tensor with shape [batch_size, second_stage_batch_size] representing the (potentially zero padded) proposal objectness scores for all images in the batch. num_proposals: A Tensor of type `int32`. A 1-D tensor of shape [batch] representing the number of proposals predicted for each image in the batch. """ single_image_proposal_box_sample = [] single_image_proposal_score_sample = [] single_image_num_proposals_sample = [] for (single_image_proposal_boxes, single_image_proposal_scores, single_image_num_proposals, single_image_groundtruth_boxlist, single_image_groundtruth_classes_with_background, single_image_groundtruth_weights) in zip( tf.unstack(proposal_boxes), tf.unstack(proposal_scores), tf.unstack(num_proposals), groundtruth_boxlists, groundtruth_classes_with_background_list, groundtruth_weights_list): single_image_boxlist = box_list.BoxList(single_image_proposal_boxes) single_image_boxlist.add_field(fields.BoxListFields.scores, single_image_proposal_scores) sampled_boxlist = self._sample_box_classifier_minibatch_single_image( single_image_boxlist, single_image_num_proposals, single_image_groundtruth_boxlist, single_image_groundtruth_classes_with_background, single_image_groundtruth_weights) sampled_padded_boxlist = box_list_ops.pad_or_clip_box_list( sampled_boxlist, num_boxes=self._second_stage_batch_size) single_image_num_proposals_sample.append(tf.minimum( sampled_boxlist.num_boxes(), self._second_stage_batch_size)) bb = sampled_padded_boxlist.get() single_image_proposal_box_sample.append(bb) single_image_proposal_score_sample.append( sampled_padded_boxlist.get_field(fields.BoxListFields.scores)) return (tf.stack(single_image_proposal_box_sample), tf.stack(single_image_proposal_score_sample), tf.stack(single_image_num_proposals_sample)) def _format_groundtruth_data(self, true_image_shapes): """Helper function for preparing groundtruth data for target assignment. In order to be consistent with the model.DetectionModel interface, groundtruth boxes are specified in normalized coordinates and classes are specified as label indices with no assumed background category. To prepare for target assignment, we: 1) convert boxes to absolute coordinates, 2) add a background class at class index 0 3) groundtruth instance masks, if available, are resized to match image_shape. Args: true_image_shapes: int32 tensor of shape [batch, 3] where each row is of the form [height, width, channels] indicating the shapes of true images in the resized images, as resized images can be padded with zeros. Returns: groundtruth_boxlists: A list of BoxLists containing (absolute) coordinates of the groundtruth boxes. groundtruth_classes_with_background_list: A list of 2-D one-hot (or k-hot) tensors of shape [num_boxes, num_classes+1] containing the class targets with the 0th index assumed to map to the background class. groundtruth_masks_list: If present, a list of 3-D tf.float32 tensors of shape [num_boxes, image_height, image_width] containing instance masks. This is set to None if no masks exist in the provided groundtruth. """ groundtruth_boxlists = [ box_list_ops.to_absolute_coordinates( box_list.BoxList(boxes), true_image_shapes[i, 0], true_image_shapes[i, 1]) for i, boxes in enumerate( self.groundtruth_lists(fields.BoxListFields.boxes)) ] groundtruth_classes_with_background_list = [ tf.to_float( tf.pad(one_hot_encoding, [[0, 0], [1, 0]], mode='CONSTANT')) for one_hot_encoding in self.groundtruth_lists( fields.BoxListFields.classes)] groundtruth_masks_list = self._groundtruth_lists.get( fields.BoxListFields.masks) # TODO(rathodv): Remove mask resizing once the legacy pipeline is deleted. if groundtruth_masks_list is not None and self._resize_masks: resized_masks_list = [] for mask in groundtruth_masks_list: _, resized_mask, _ = self._image_resizer_fn( # Reuse the given `image_resizer_fn` to resize groundtruth masks. # `mask` tensor for an image is of the shape [num_masks, # image_height, image_width]. Below we create a dummy image of the # the shape [image_height, image_width, 1] to use with # `image_resizer_fn`. image=tf.zeros(tf.stack([tf.shape(mask)[1], tf.shape(mask)[2], 1])), masks=mask) resized_masks_list.append(resized_mask) groundtruth_masks_list = resized_masks_list if self.groundtruth_has_field(fields.BoxListFields.weights): groundtruth_weights_list = self.groundtruth_lists( fields.BoxListFields.weights) else: # Set weights for all batch elements equally to 1.0 groundtruth_weights_list = [] for groundtruth_classes in groundtruth_classes_with_background_list: num_gt = tf.shape(groundtruth_classes)[0] groundtruth_weights = tf.ones(num_gt) groundtruth_weights_list.append(groundtruth_weights) return (groundtruth_boxlists, groundtruth_classes_with_background_list, groundtruth_masks_list, groundtruth_weights_list) def _sample_box_classifier_minibatch_single_image( self, proposal_boxlist, num_valid_proposals, groundtruth_boxlist, groundtruth_classes_with_background, groundtruth_weights): """Samples a mini-batch of proposals to be sent to the box classifier. Helper function for self._postprocess_rpn. Args: proposal_boxlist: A BoxList containing K proposal boxes in absolute coordinates. num_valid_proposals: Number of valid proposals in the proposal boxlist. groundtruth_boxlist: A Boxlist containing N groundtruth object boxes in absolute coordinates. groundtruth_classes_with_background: A tensor with shape `[N, self.num_classes + 1]` representing groundtruth classes. The classes are assumed to be k-hot encoded, and include background as the zero-th class. groundtruth_weights: Weights attached to the groundtruth_boxes. Returns: a BoxList contained sampled proposals. """ (cls_targets, cls_weights, _, _, _) = self._detector_target_assigner.assign( proposal_boxlist, groundtruth_boxlist, groundtruth_classes_with_background, unmatched_class_label=tf.constant( [1] + self._num_classes * [0], dtype=tf.float32), groundtruth_weights=groundtruth_weights) # Selects all boxes as candidates if none of them is selected according # to cls_weights. This could happen as boxes within certain IOU ranges # are ignored. If triggered, the selected boxes will still be ignored # during loss computation. cls_weights = tf.reduce_mean(cls_weights, axis=-1) positive_indicator = tf.greater(tf.argmax(cls_targets, axis=1), 0) valid_indicator = tf.logical_and( tf.range(proposal_boxlist.num_boxes()) < num_valid_proposals, cls_weights > 0 ) selected_positions = self._second_stage_sampler.subsample( valid_indicator, self._second_stage_batch_size, positive_indicator) return box_list_ops.boolean_mask( proposal_boxlist, selected_positions, use_static_shapes=self._use_static_shapes, indicator_sum=(self._second_stage_batch_size if self._use_static_shapes else None)) def _compute_second_stage_input_feature_maps(self, features_to_crop, proposal_boxes_normalized): """Crops to a set of proposals from the feature map for a batch of images. Helper function for self._postprocess_rpn. This function calls `tf.image.crop_and_resize` to create the feature map to be passed to the second stage box classifier for each proposal. Args: features_to_crop: A float32 tensor with shape [batch_size, height, width, depth] proposal_boxes_normalized: A float32 tensor with shape [batch_size, num_proposals, box_code_size] containing proposal boxes in normalized coordinates. Returns: A float32 tensor with shape [K, new_height, new_width, depth]. """ cropped_regions = self._flatten_first_two_dimensions( self._crop_and_resize_fn( features_to_crop, proposal_boxes_normalized, [self._initial_crop_size, self._initial_crop_size])) return slim.max_pool2d( cropped_regions, [self._maxpool_kernel_size, self._maxpool_kernel_size], stride=self._maxpool_stride) def _postprocess_box_classifier(self, refined_box_encodings, class_predictions_with_background, proposal_boxes, num_proposals, image_shapes, mask_predictions=None): """Converts predictions from the second stage box classifier to detections. Args: refined_box_encodings: a 3-D float tensor with shape [total_num_padded_proposals, num_classes, self._box_coder.code_size] representing predicted (final) refined box encodings. If using a shared box across classes the shape will instead be [total_num_padded_proposals, 1, 4] class_predictions_with_background: a 3-D tensor float with shape [total_num_padded_proposals, num_classes + 1] containing class predictions (logits) for each of the proposals. Note that this tensor *includes* background class predictions (at class index 0). proposal_boxes: a 3-D float tensor with shape [batch_size, self.max_num_proposals, 4] representing decoded proposal bounding boxes in absolute coordinates. num_proposals: a 1-D int32 tensor of shape [batch] representing the number of proposals predicted for each image in the batch. image_shapes: a 2-D int32 tensor containing shapes of input image in the batch. mask_predictions: (optional) a 4-D float tensor with shape [total_num_padded_proposals, num_classes, mask_height, mask_width] containing instance mask prediction logits. Returns: A dictionary containing: `detection_boxes`: [batch, max_detection, 4] in normalized co-ordinates. `detection_scores`: [batch, max_detections] `detection_classes`: [batch, max_detections] `num_detections`: [batch] `detection_masks`: (optional) [batch, max_detections, mask_height, mask_width]. Note that a pixel-wise sigmoid score converter is applied to the detection masks. """ refined_box_encodings_batch = tf.reshape( refined_box_encodings, [-1, self.max_num_proposals, refined_box_encodings.shape[1], self._box_coder.code_size]) class_predictions_with_background_batch = tf.reshape( class_predictions_with_background, [-1, self.max_num_proposals, self.num_classes + 1] ) refined_decoded_boxes_batch = self._batch_decode_boxes( refined_box_encodings_batch, proposal_boxes) class_predictions_with_background_batch = ( self._second_stage_score_conversion_fn( class_predictions_with_background_batch)) class_predictions_batch = tf.reshape( tf.slice(class_predictions_with_background_batch, [0, 0, 1], [-1, -1, -1]), [-1, self.max_num_proposals, self.num_classes]) clip_window = self._compute_clip_window(image_shapes) mask_predictions_batch = None if mask_predictions is not None: mask_height = mask_predictions.shape[2].value mask_width = mask_predictions.shape[3].value mask_predictions = tf.sigmoid(mask_predictions) mask_predictions_batch = tf.reshape( mask_predictions, [-1, self.max_num_proposals, self.num_classes, mask_height, mask_width]) (nmsed_boxes, nmsed_scores, nmsed_classes, nmsed_masks, _, num_detections) = self._second_stage_nms_fn( refined_decoded_boxes_batch, class_predictions_batch, clip_window=clip_window, change_coordinate_frame=True, num_valid_boxes=num_proposals, masks=mask_predictions_batch) detections = { fields.DetectionResultFields.detection_boxes: nmsed_boxes, fields.DetectionResultFields.detection_scores: nmsed_scores, fields.DetectionResultFields.detection_classes: nmsed_classes, fields.DetectionResultFields.num_detections: tf.to_float(num_detections) } if nmsed_masks is not None: detections[fields.DetectionResultFields.detection_masks] = nmsed_masks return detections def _batch_decode_boxes(self, box_encodings, anchor_boxes): """Decodes box encodings with respect to the anchor boxes. Args: box_encodings: a 4-D tensor with shape [batch_size, num_anchors, num_classes, self._box_coder.code_size] representing box encodings. anchor_boxes: [batch_size, num_anchors, self._box_coder.code_size] representing decoded bounding boxes. If using a shared box across classes the shape will instead be [total_num_proposals, 1, self._box_coder.code_size]. Returns: decoded_boxes: a [batch_size, num_anchors, num_classes, self._box_coder.code_size] float tensor representing bounding box predictions (for each image in batch, proposal and class). If using a shared box across classes the shape will instead be [batch_size, num_anchors, 1, self._box_coder.code_size]. """ combined_shape = shape_utils.combined_static_and_dynamic_shape( box_encodings) num_classes = combined_shape[2] tiled_anchor_boxes = tf.tile( tf.expand_dims(anchor_boxes, 2), [1, 1, num_classes, 1]) tiled_anchors_boxlist = box_list.BoxList( tf.reshape(tiled_anchor_boxes, [-1, 4])) decoded_boxes = self._box_coder.decode( tf.reshape(box_encodings, [-1, self._box_coder.code_size]), tiled_anchors_boxlist) return tf.reshape(decoded_boxes.get(), tf.stack([combined_shape[0], combined_shape[1], num_classes, 4])) def loss(self, prediction_dict, true_image_shapes, scope=None): """Compute scalar loss tensors given prediction tensors. If number_of_stages=1, only RPN related losses are computed (i.e., `rpn_localization_loss` and `rpn_objectness_loss`). Otherwise all losses are computed. Args: prediction_dict: a dictionary holding prediction tensors (see the documentation for the predict method. If number_of_stages=1, we expect prediction_dict to contain `rpn_box_encodings`, `rpn_objectness_predictions_with_background`, `rpn_features_to_crop`, `image_shape`, and `anchors` fields. Otherwise we expect prediction_dict to additionally contain `refined_box_encodings`, `class_predictions_with_background`, `num_proposals`, and `proposal_boxes` fields. true_image_shapes: int32 tensor of shape [batch, 3] where each row is of the form [height, width, channels] indicating the shapes of true images in the resized images, as resized images can be padded with zeros. scope: Optional scope name. Returns: a dictionary mapping loss keys (`first_stage_localization_loss`, `first_stage_objectness_loss`, 'second_stage_localization_loss', 'second_stage_classification_loss') to scalar tensors representing corresponding loss values. """ with tf.name_scope(scope, 'Loss', prediction_dict.values()): (groundtruth_boxlists, groundtruth_classes_with_background_list, groundtruth_masks_list, groundtruth_weights_list ) = self._format_groundtruth_data(true_image_shapes) loss_dict = self._loss_rpn( prediction_dict['rpn_box_encodings'], prediction_dict['rpn_objectness_predictions_with_background'], prediction_dict['anchors'], groundtruth_boxlists, groundtruth_classes_with_background_list, groundtruth_weights_list) if self._number_of_stages > 1: loss_dict.update( self._loss_box_classifier( prediction_dict['refined_box_encodings'], prediction_dict['class_predictions_with_background'], prediction_dict['proposal_boxes'], prediction_dict['num_proposals'], groundtruth_boxlists, groundtruth_classes_with_background_list, groundtruth_weights_list, prediction_dict['image_shape'], prediction_dict.get('mask_predictions'), groundtruth_masks_list, prediction_dict.get( fields.DetectionResultFields.detection_boxes), prediction_dict.get( fields.DetectionResultFields.num_detections))) return loss_dict def _loss_rpn(self, rpn_box_encodings, rpn_objectness_predictions_with_background, anchors, groundtruth_boxlists, groundtruth_classes_with_background_list, groundtruth_weights_list): """Computes scalar RPN loss tensors. Uses self._proposal_target_assigner to obtain regression and classification targets for the first stage RPN, samples a "minibatch" of anchors to participate in the loss computation, and returns the RPN losses. Args: rpn_box_encodings: A 4-D float tensor of shape [batch_size, num_anchors, self._box_coder.code_size] containing predicted proposal box encodings. rpn_objectness_predictions_with_background: A 2-D float tensor of shape [batch_size, num_anchors, 2] containing objectness predictions (logits) for each of the anchors with 0 corresponding to background and 1 corresponding to object. anchors: A 2-D tensor of shape [num_anchors, 4] representing anchors for the first stage RPN. Note that `num_anchors` can differ depending on whether the model is created in training or inference mode. groundtruth_boxlists: A list of BoxLists containing coordinates of the groundtruth boxes. groundtruth_classes_with_background_list: A list of 2-D one-hot (or k-hot) tensors of shape [num_boxes, num_classes+1] containing the class targets with the 0th index assumed to map to the background class. groundtruth_weights_list: A list of 1-D tf.float32 tensors of shape [num_boxes] containing weights for groundtruth boxes. Returns: a dictionary mapping loss keys (`first_stage_localization_loss`, `first_stage_objectness_loss`) to scalar tensors representing corresponding loss values. """ with tf.name_scope('RPNLoss'): (batch_cls_targets, batch_cls_weights, batch_reg_targets, batch_reg_weights, _) = target_assigner.batch_assign_targets( target_assigner=self._proposal_target_assigner, anchors_batch=box_list.BoxList(anchors), gt_box_batch=groundtruth_boxlists, gt_class_targets_batch=(len(groundtruth_boxlists) * [None]), gt_weights_batch=groundtruth_weights_list) batch_cls_weights = tf.reduce_mean(batch_cls_weights, axis=2) batch_cls_targets = tf.squeeze(batch_cls_targets, axis=2) def _minibatch_subsample_fn(inputs): cls_targets, cls_weights = inputs return self._first_stage_sampler.subsample( tf.cast(cls_weights, tf.bool), self._first_stage_minibatch_size, tf.cast(cls_targets, tf.bool)) batch_sampled_indices = tf.to_float(shape_utils.static_or_dynamic_map_fn( _minibatch_subsample_fn, [batch_cls_targets, batch_cls_weights], dtype=tf.bool, parallel_iterations=self._parallel_iterations, back_prop=True)) # Normalize by number of examples in sampled minibatch normalizer = tf.reduce_sum(batch_sampled_indices, axis=1) batch_one_hot_targets = tf.one_hot( tf.to_int32(batch_cls_targets), depth=2) sampled_reg_indices = tf.multiply(batch_sampled_indices, batch_reg_weights) losses_mask = None if self.groundtruth_has_field(fields.InputDataFields.is_annotated): losses_mask = tf.stack(self.groundtruth_lists( fields.InputDataFields.is_annotated)) localization_losses = self._first_stage_localization_loss( rpn_box_encodings, batch_reg_targets, weights=sampled_reg_indices, losses_mask=losses_mask) objectness_losses = self._first_stage_objectness_loss( rpn_objectness_predictions_with_background, batch_one_hot_targets, weights=tf.expand_dims(batch_sampled_indices, axis=-1), losses_mask=losses_mask) localization_loss = tf.reduce_mean( tf.reduce_sum(localization_losses, axis=1) / normalizer) objectness_loss = tf.reduce_mean( tf.reduce_sum(objectness_losses, axis=1) / normalizer) localization_loss = tf.multiply(self._first_stage_loc_loss_weight, localization_loss, name='localization_loss') objectness_loss = tf.multiply(self._first_stage_obj_loss_weight, objectness_loss, name='objectness_loss') loss_dict = {localization_loss.op.name: localization_loss, objectness_loss.op.name: objectness_loss} return loss_dict def _loss_box_classifier(self, refined_box_encodings, class_predictions_with_background, proposal_boxes, num_proposals, groundtruth_boxlists, groundtruth_classes_with_background_list, groundtruth_weights_list, image_shape, prediction_masks=None, groundtruth_masks_list=None, detection_boxes=None, num_detections=None): """Computes scalar box classifier loss tensors. Uses self._detector_target_assigner to obtain regression and classification targets for the second stage box classifier, optionally performs hard mining, and returns losses. All losses are computed independently for each image and then averaged across the batch. Please note that for boxes and masks with multiple labels, the box regression and mask prediction losses are only computed for one label. This function assumes that the proposal boxes in the "padded" regions are actually zero (and thus should not be matched to). Args: refined_box_encodings: a 3-D tensor with shape [total_num_proposals, num_classes, box_coder.code_size] representing predicted (final) refined box encodings. If using a shared box across classes this will instead have shape [total_num_proposals, 1, box_coder.code_size]. class_predictions_with_background: a 2-D tensor with shape [total_num_proposals, num_classes + 1] containing class predictions (logits) for each of the anchors. Note that this tensor *includes* background class predictions (at class index 0). proposal_boxes: [batch_size, self.max_num_proposals, 4] representing decoded proposal bounding boxes. num_proposals: A Tensor of type `int32`. A 1-D tensor of shape [batch] representing the number of proposals predicted for each image in the batch. groundtruth_boxlists: a list of BoxLists containing coordinates of the groundtruth boxes. groundtruth_classes_with_background_list: a list of 2-D one-hot (or k-hot) tensors of shape [num_boxes, num_classes + 1] containing the class targets with the 0th index assumed to map to the background class. groundtruth_weights_list: A list of 1-D tf.float32 tensors of shape [num_boxes] containing weights for groundtruth boxes. image_shape: a 1-D tensor of shape [4] representing the image shape. prediction_masks: an optional 4-D tensor with shape [total_num_proposals, num_classes, mask_height, mask_width] containing the instance masks for each box. groundtruth_masks_list: an optional list of 3-D tensors of shape [num_boxes, image_height, image_width] containing the instance masks for each of the boxes. detection_boxes: 3-D float tensor of shape [batch, max_total_detections, 4] containing post-processed detection boxes in normalized co-ordinates. num_detections: 1-D int32 tensor of shape [batch] containing number of valid detections in `detection_boxes`. Returns: a dictionary mapping loss keys ('second_stage_localization_loss', 'second_stage_classification_loss') to scalar tensors representing corresponding loss values. Raises: ValueError: if `predict_instance_masks` in second_stage_mask_rcnn_box_predictor is True and `groundtruth_masks_list` is not provided. """ with tf.name_scope('BoxClassifierLoss'): paddings_indicator = self._padded_batched_proposals_indicator( num_proposals, proposal_boxes.shape[1]) proposal_boxlists = [ box_list.BoxList(proposal_boxes_single_image) for proposal_boxes_single_image in tf.unstack(proposal_boxes)] batch_size = len(proposal_boxlists) num_proposals_or_one = tf.to_float(tf.expand_dims( tf.maximum(num_proposals, tf.ones_like(num_proposals)), 1)) normalizer = tf.tile(num_proposals_or_one, [1, self.max_num_proposals]) * batch_size (batch_cls_targets_with_background, batch_cls_weights, batch_reg_targets, batch_reg_weights, _) = target_assigner.batch_assign_targets( target_assigner=self._detector_target_assigner, anchors_batch=proposal_boxlists, gt_box_batch=groundtruth_boxlists, gt_class_targets_batch=groundtruth_classes_with_background_list, unmatched_class_label=tf.constant( [1] + self._num_classes * [0], dtype=tf.float32), gt_weights_batch=groundtruth_weights_list) class_predictions_with_background = tf.reshape( class_predictions_with_background, [batch_size, self.max_num_proposals, -1]) flat_cls_targets_with_background = tf.reshape( batch_cls_targets_with_background, [batch_size * self.max_num_proposals, -1]) one_hot_flat_cls_targets_with_background = tf.argmax( flat_cls_targets_with_background, axis=1) one_hot_flat_cls_targets_with_background = tf.one_hot( one_hot_flat_cls_targets_with_background, flat_cls_targets_with_background.get_shape()[1]) # If using a shared box across classes use directly if refined_box_encodings.shape[1] == 1: reshaped_refined_box_encodings = tf.reshape( refined_box_encodings, [batch_size, self.max_num_proposals, self._box_coder.code_size]) # For anchors with multiple labels, picks refined_location_encodings # for just one class to avoid over-counting for regression loss and # (optionally) mask loss. else: reshaped_refined_box_encodings = ( self._get_refined_encodings_for_postitive_class( refined_box_encodings, one_hot_flat_cls_targets_with_background, batch_size)) losses_mask = None if self.groundtruth_has_field(fields.InputDataFields.is_annotated): losses_mask = tf.stack(self.groundtruth_lists( fields.InputDataFields.is_annotated)) second_stage_loc_losses = self._second_stage_localization_loss( reshaped_refined_box_encodings, batch_reg_targets, weights=batch_reg_weights, losses_mask=losses_mask) / normalizer second_stage_cls_losses = ops.reduce_sum_trailing_dimensions( self._second_stage_classification_loss( class_predictions_with_background, batch_cls_targets_with_background, weights=batch_cls_weights, losses_mask=losses_mask), ndims=2) / normalizer second_stage_loc_loss = tf.reduce_sum( second_stage_loc_losses * tf.to_float(paddings_indicator)) second_stage_cls_loss = tf.reduce_sum( second_stage_cls_losses * tf.to_float(paddings_indicator)) if self._hard_example_miner: (second_stage_loc_loss, second_stage_cls_loss ) = self._unpad_proposals_and_apply_hard_mining( proposal_boxlists, second_stage_loc_losses, second_stage_cls_losses, num_proposals) localization_loss = tf.multiply(self._second_stage_loc_loss_weight, second_stage_loc_loss, name='localization_loss') classification_loss = tf.multiply(self._second_stage_cls_loss_weight, second_stage_cls_loss, name='classification_loss') loss_dict = {localization_loss.op.name: localization_loss, classification_loss.op.name: classification_loss} second_stage_mask_loss = None if prediction_masks is not None: if groundtruth_masks_list is None: raise ValueError('Groundtruth instance masks not provided. ' 'Please configure input reader.') if not self._is_training: (proposal_boxes, proposal_boxlists, paddings_indicator, one_hot_flat_cls_targets_with_background ) = self._get_mask_proposal_boxes_and_classes( detection_boxes, num_detections, image_shape, groundtruth_boxlists, groundtruth_classes_with_background_list, groundtruth_weights_list) unmatched_mask_label = tf.zeros(image_shape[1:3], dtype=tf.float32) (batch_mask_targets, _, _, batch_mask_target_weights, _) = target_assigner.batch_assign_targets( target_assigner=self._detector_target_assigner, anchors_batch=proposal_boxlists, gt_box_batch=groundtruth_boxlists, gt_class_targets_batch=groundtruth_masks_list, unmatched_class_label=unmatched_mask_label, gt_weights_batch=groundtruth_weights_list) # Pad the prediction_masks with to add zeros for background class to be # consistent with class predictions. if prediction_masks.get_shape().as_list()[1] == 1: # Class agnostic masks or masks for one-class prediction. Logic for # both cases is the same since background predictions are ignored # through the batch_mask_target_weights. prediction_masks_masked_by_class_targets = prediction_masks else: prediction_masks_with_background = tf.pad( prediction_masks, [[0, 0], [1, 0], [0, 0], [0, 0]]) prediction_masks_masked_by_class_targets = tf.boolean_mask( prediction_masks_with_background, tf.greater(one_hot_flat_cls_targets_with_background, 0)) mask_height = prediction_masks.shape[2].value mask_width = prediction_masks.shape[3].value reshaped_prediction_masks = tf.reshape( prediction_masks_masked_by_class_targets, [batch_size, -1, mask_height * mask_width]) batch_mask_targets_shape = tf.shape(batch_mask_targets) flat_gt_masks = tf.reshape(batch_mask_targets, [-1, batch_mask_targets_shape[2], batch_mask_targets_shape[3]]) # Use normalized proposals to crop mask targets from image masks. flat_normalized_proposals = box_list_ops.to_normalized_coordinates( box_list.BoxList(tf.reshape(proposal_boxes, [-1, 4])), image_shape[1], image_shape[2]).get() flat_cropped_gt_mask = self._crop_and_resize_fn( tf.expand_dims(flat_gt_masks, -1), tf.expand_dims(flat_normalized_proposals, axis=1), [mask_height, mask_width]) # Without stopping gradients into cropped groundtruth masks the # performance with 100-padded groundtruth masks when batch size > 1 is # about 4% worse. # TODO(rathodv): Investigate this since we don't expect any variables # upstream of flat_cropped_gt_mask. flat_cropped_gt_mask = tf.stop_gradient(flat_cropped_gt_mask) batch_cropped_gt_mask = tf.reshape( flat_cropped_gt_mask, [batch_size, -1, mask_height * mask_width]) mask_losses_weights = ( batch_mask_target_weights * tf.to_float(paddings_indicator)) mask_losses = self._second_stage_mask_loss( reshaped_prediction_masks, batch_cropped_gt_mask, weights=tf.expand_dims(mask_losses_weights, axis=-1), losses_mask=losses_mask) total_mask_loss = tf.reduce_sum(mask_losses) normalizer = tf.maximum( tf.reduce_sum(mask_losses_weights * mask_height * mask_width), 1.0) second_stage_mask_loss = total_mask_loss / normalizer if second_stage_mask_loss is not None: mask_loss = tf.multiply(self._second_stage_mask_loss_weight, second_stage_mask_loss, name='mask_loss') loss_dict[mask_loss.op.name] = mask_loss return loss_dict def _get_mask_proposal_boxes_and_classes( self, detection_boxes, num_detections, image_shape, groundtruth_boxlists, groundtruth_classes_with_background_list, groundtruth_weights_list): """Returns proposal boxes and class targets to compute evaluation mask loss. During evaluation, detection boxes are used to extract features for mask prediction. Therefore, to compute mask loss during evaluation detection boxes must be used to compute correct class and mask targets. This function returns boxes and classes in the correct format for computing mask targets during evaluation. Args: detection_boxes: A 3-D float tensor of shape [batch, max_detection_boxes, 4] containing detection boxes in normalized co-ordinates. num_detections: A 1-D float tensor of shape [batch] containing number of valid boxes in `detection_boxes`. image_shape: A 1-D tensor of shape [4] containing image tensor shape. groundtruth_boxlists: A list of groundtruth boxlists. groundtruth_classes_with_background_list: A list of groundtruth classes. groundtruth_weights_list: A list of groundtruth weights. Return: mask_proposal_boxes: detection boxes to use for mask proposals in absolute co-ordinates. mask_proposal_boxlists: `mask_proposal_boxes` in a list of BoxLists in absolute co-ordinates. mask_proposal_paddings_indicator: a tensor indicating valid boxes. mask_proposal_one_hot_flat_cls_targets_with_background: Class targets computed using detection boxes. """ batch, max_num_detections, _ = detection_boxes.shape.as_list() proposal_boxes = tf.reshape(box_list_ops.to_absolute_coordinates( box_list.BoxList(tf.reshape(detection_boxes, [-1, 4])), image_shape[1], image_shape[2]).get(), [batch, max_num_detections, 4]) proposal_boxlists = [ box_list.BoxList(detection_boxes_single_image) for detection_boxes_single_image in tf.unstack(proposal_boxes) ] paddings_indicator = self._padded_batched_proposals_indicator( tf.to_int32(num_detections), detection_boxes.shape[1]) (batch_cls_targets_with_background, _, _, _, _) = target_assigner.batch_assign_targets( target_assigner=self._detector_target_assigner, anchors_batch=proposal_boxlists, gt_box_batch=groundtruth_boxlists, gt_class_targets_batch=groundtruth_classes_with_background_list, unmatched_class_label=tf.constant( [1] + self._num_classes * [0], dtype=tf.float32), gt_weights_batch=groundtruth_weights_list) flat_cls_targets_with_background = tf.reshape( batch_cls_targets_with_background, [-1, self._num_classes + 1]) one_hot_flat_cls_targets_with_background = tf.argmax( flat_cls_targets_with_background, axis=1) one_hot_flat_cls_targets_with_background = tf.one_hot( one_hot_flat_cls_targets_with_background, flat_cls_targets_with_background.get_shape()[1]) return (proposal_boxes, proposal_boxlists, paddings_indicator, one_hot_flat_cls_targets_with_background) def _get_refined_encodings_for_postitive_class( self, refined_box_encodings, flat_cls_targets_with_background, batch_size): # We only predict refined location encodings for the non background # classes, but we now pad it to make it compatible with the class # predictions refined_box_encodings_with_background = tf.pad(refined_box_encodings, [[0, 0], [1, 0], [0, 0]]) refined_box_encodings_masked_by_class_targets = ( box_list_ops.boolean_mask( box_list.BoxList( tf.reshape(refined_box_encodings_with_background, [-1, self._box_coder.code_size])), tf.reshape(tf.greater(flat_cls_targets_with_background, 0), [-1]), use_static_shapes=self._use_static_shapes, indicator_sum=batch_size * self.max_num_proposals if self._use_static_shapes else None).get()) return tf.reshape( refined_box_encodings_masked_by_class_targets, [ batch_size, self.max_num_proposals, self._box_coder.code_size ]) def _padded_batched_proposals_indicator(self, num_proposals, max_num_proposals): """Creates indicator matrix of non-pad elements of padded batch proposals. Args: num_proposals: Tensor of type tf.int32 with shape [batch_size]. max_num_proposals: Maximum number of proposals per image (integer). Returns: A Tensor of type tf.bool with shape [batch_size, max_num_proposals]. """ batch_size = tf.size(num_proposals) tiled_num_proposals = tf.tile( tf.expand_dims(num_proposals, 1), [1, max_num_proposals]) tiled_proposal_index = tf.tile( tf.expand_dims(tf.range(max_num_proposals), 0), [batch_size, 1]) return tf.greater(tiled_num_proposals, tiled_proposal_index) def _unpad_proposals_and_apply_hard_mining(self, proposal_boxlists, second_stage_loc_losses, second_stage_cls_losses, num_proposals): """Unpads proposals and applies hard mining. Args: proposal_boxlists: A list of `batch_size` BoxLists each representing `self.max_num_proposals` representing decoded proposal bounding boxes for each image. second_stage_loc_losses: A Tensor of type `float32`. A tensor of shape `[batch_size, self.max_num_proposals]` representing per-anchor second stage localization loss values. second_stage_cls_losses: A Tensor of type `float32`. A tensor of shape `[batch_size, self.max_num_proposals]` representing per-anchor second stage classification loss values. num_proposals: A Tensor of type `int32`. A 1-D tensor of shape [batch] representing the number of proposals predicted for each image in the batch. Returns: second_stage_loc_loss: A scalar float32 tensor representing the second stage localization loss. second_stage_cls_loss: A scalar float32 tensor representing the second stage classification loss. """ for (proposal_boxlist, single_image_loc_loss, single_image_cls_loss, single_image_num_proposals) in zip( proposal_boxlists, tf.unstack(second_stage_loc_losses), tf.unstack(second_stage_cls_losses), tf.unstack(num_proposals)): proposal_boxlist = box_list.BoxList( tf.slice(proposal_boxlist.get(), [0, 0], [single_image_num_proposals, -1])) single_image_loc_loss = tf.slice(single_image_loc_loss, [0], [single_image_num_proposals]) single_image_cls_loss = tf.slice(single_image_cls_loss, [0], [single_image_num_proposals]) return self._hard_example_miner( location_losses=tf.expand_dims(single_image_loc_loss, 0), cls_losses=tf.expand_dims(single_image_cls_loss, 0), decoded_boxlist_list=[proposal_boxlist]) def regularization_losses(self): """Returns a list of regularization losses for this model. Returns a list of regularization losses for this model that the estimator needs to use during training/optimization. Returns: A list of regularization loss tensors. """ return tf.get_collection(tf.GraphKeys.REGULARIZATION_LOSSES) def restore_map(self, fine_tune_checkpoint_type='detection', load_all_detection_checkpoint_vars=False): """Returns a map of variables to load from a foreign checkpoint. See parent class for details. Args: fine_tune_checkpoint_type: whether to restore from a full detection checkpoint (with compatible variable names) or to restore from a classification checkpoint for initialization prior to training. Valid values: `detection`, `classification`. Default 'detection'. load_all_detection_checkpoint_vars: whether to load all variables (when `fine_tune_checkpoint_type` is `detection`). If False, only variables within the feature extractor scopes are included. Default False. Returns: A dict mapping variable names (to load from a checkpoint) to variables in the model graph. Raises: ValueError: if fine_tune_checkpoint_type is neither `classification` nor `detection`. """ if fine_tune_checkpoint_type not in ['detection', 'classification']: raise ValueError('Not supported fine_tune_checkpoint_type: {}'.format( fine_tune_checkpoint_type)) if fine_tune_checkpoint_type == 'classification': return self._feature_extractor.restore_from_classification_checkpoint_fn( self.first_stage_feature_extractor_scope, self.second_stage_feature_extractor_scope) variables_to_restore = tf.global_variables() variables_to_restore.append(slim.get_or_create_global_step()) # Only load feature extractor variables to be consistent with loading from # a classification checkpoint. include_patterns = None if not load_all_detection_checkpoint_vars: include_patterns = [ self.first_stage_feature_extractor_scope, self.second_stage_feature_extractor_scope ] feature_extractor_variables = tf.contrib.framework.filter_variables( variables_to_restore, include_patterns=include_patterns) return {var.op.name: var for var in feature_extractor_variables} def updates(self): """Returns a list of update operators for this model. Returns a list of update operators for this model that must be executed at each training step. The estimator's train op needs to have a control dependency on these updates. Returns: A list of update operators. """ return tf.get_collection(tf.GraphKeys.UPDATE_OPS)
TensorFlow/Classification/ConvNets/model/layers
layers
squeeze_excitation_layer
#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright 2018 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== import tensorflow as tf from model import layers from model import blocks __all__ = ['squeeze_excitation_layer'] def squeeze_excitation_layer( inputs, ratio, training=True, data_format='NCHW', kernel_initializer=tf.compat.v1.variance_scaling_initializer(), bias_initializer=tf.zeros_initializer(), name="squeeze_excitation_layer" ): if data_format not in ['NHWC', 'NCHW']: raise ValueError("Unknown data format: `%s` (accepted: ['NHWC', 'NCHW'])" % data_format) in_shape = inputs.get_shape() num_channels = in_shape[1] if data_format == "NCHW" else in_shape[-1] with tf.variable_scope(name): net = inputs # squeeze squeeze = layers.reduce_mean( net, keepdims=False, data_format=data_format, name='squeeze_spatial_mean' ) # fc + relu excitation = layers.dense( inputs=squeeze, units=num_channels // ratio, use_bias=True, trainable=training, kernel_initializer=kernel_initializer, bias_initializer=bias_initializer ) excitation = layers.relu(excitation) # fc + sigmoid excitation = layers.dense( inputs=excitation, units=num_channels, use_bias=True, trainable=training, kernel_initializer=kernel_initializer, bias_initializer=bias_initializer ) excitation = layers.sigmoid(excitation) out_shape = [-1, num_channels, 1, 1] if data_format == "NCHW" else [-1, 1, 1, num_channels] excitation = tf.reshape(excitation, out_shape) net = net * excitation return net
PyTorch/Classification/ConvNets/efficientnet/training/FP32
FP32
DGX1V-16G_efficientnet-b0_FP32
python ./multiproc.py --nproc_per_node 8 ./launch.py --model efficientnet-b0 --precision FP32 --mode convergence --platform DGX1V-16G /imagenet --workspace ${1:-./} --raport-file raport.json
PyTorch/SpeechSynthesis/FastPitch/hifigan
hifigan
arg_parser
# Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from ast import literal_eval def parse_hifigan_args(parent, add_help=False): """ Parse model specific commandline arguments. """ parser = argparse.ArgumentParser(parents=[parent], add_help=add_help, allow_abbrev=False) hfg = parser.add_argument_group('HiFi-GAN generator parameters') hfg.add_argument('--upsample_rates', default=[8, 8, 2, 2], type=literal_eval_arg, help='Upsample rates') hfg.add_argument('--upsample_kernel_sizes', default=[16, 16, 4, 4], type=literal_eval_arg, help='Upsample kernel sizes') hfg.add_argument('--upsample_initial_channel', default=512, type=int, help='Upsample initial channel') hfg.add_argument('--resblock', default='1', type=str, help='Resblock module version') hfg.add_argument('--resblock_kernel_sizes', default=[3, 7, 11], type=literal_eval_arg, help='Resblock kernel sizes') hfg.add_argument('--resblock_dilation_sizes', type=literal_eval_arg, default=[[1, 3, 5], [1, 3, 5], [1, 3, 5]], help='Resblock dilation sizes'), hfg = parser.add_argument_group('HiFi-GAN discriminator parameters') hfg.add_argument('--mpd_periods', default=[2, 3, 5, 7, 11], type=literal_eval_arg, help='Periods of MultiPeriodDiscriminator') hfg.add_argument('--concat_fwd', action='store_true', help='Faster Discriminators (requires more GPU memory)') hfg.add_argument('--hifigan-config', type=str, default=None, required=False, help='Path to a HiFi-GAN config .json' ' (if provided, overrides model architecture flags)') return parser def literal_eval_arg(val): try: return literal_eval(val) except SyntaxError as e: # Argparse does not handle SyntaxError raise ValueError(str(e)) from e
TensorFlow2/Recommendation/DLRM_and_DCNv2/tensorflow-dot-based-interact/tensorflow_dot_based_interact/cc/kernels/launchers
launchers
dot_based_interact_fp16_launcher
// Copyright (c) 2023, NVIDIA CORPORATION. All rights reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. // #ifndef FP16_LAUNCHER_CU #define FP16_LAUNCHER_CU #include "../cuda_kernels/dot_based_interact_fp16.cu" inline void dotBasedInteractFP16Fwd(const void *input, const void *bottom_mlp_output, void *output, uint batch_size, uint num_rows, uint num_cols, cudaStream_t stream) { // num_rows is num cat variables + 1 // num_cols is embedding dim const uint kWarpSize = 32; const uint kWarpSizeLog2 = Log2<kWarpSize>::value; const uint kTileDim = 16; const uint kTileDimLog2 = Log2<kTileDim>::value; const uint warps_per_threadblock = 4; const uint threadblock_size = warps_per_threadblock * 32; const uint kRowTilesPerStep = 2; const uint kColTilesPerStep = 1; // num tiles - divide rounding up uint num_row_tiles = (num_rows + kTileDim - 1) >> kTileDimLog2; uint num_col_tiles = (num_cols + kTileDim - 1) >> kTileDimLog2; // number of rows and columns after padding uint num_rows_after_padding = kTileDim << 1; //TODO here we assume that kTileDim*2 >= num_rows. uint num_cols_after_padding = num_col_tiles << kTileDimLog2; // uint num_row_steps = num_row_tiles / kRowTilesPerStep; uint num_col_steps = num_col_tiles / kColTilesPerStep; const uint K_BLOCKS = 8; //todo what is this? is this 128//16, ie num_col_tiles? const uint M_BLOCKS = 2; //todo what is this? is this 32//16, ie num_row_tiles? const uint SKEW_HALF = ((K_BLOCKS % 2) == 0) ? 8 : 0; const uint SMEM_STRIDE = (K_BLOCKS * 16 + SKEW_HALF); //num_cols padded for tensor cores, then padded again for alignment // multiple of 2 to guarantee 256-bit alignment for start of the row, at least 16 to safeload a tile const uint smem_rows_per_warp = M_BLOCKS << 4; //is this m_blocks * tile dim? how many rows we multiply in a single pass const uint smem_elems_per_warp_mat = smem_rows_per_warp * SMEM_STRIDE; const uint SKEW_HALF_ACC = ((M_BLOCKS % 2) == 0) ? 8 : 0; const uint SMEM_STRIDE_ACC = (M_BLOCKS * 16 + SKEW_HALF_ACC); const uint smem_elems_per_warp_acc = M_BLOCKS * 16 * SMEM_STRIDE_ACC * 2; const uint smem_elems_per_warp = (smem_elems_per_warp_mat > smem_elems_per_warp_acc) ? smem_elems_per_warp_mat : smem_elems_per_warp_acc; uint raw_output_size = ((num_rows * (num_rows - 1)) >> 1) + num_cols; uint output_size = ((raw_output_size-1)/8 + 1)*8; //round up to multiple of 8 uint padding_size = output_size-raw_output_size; bool float4_predicate = !((num_cols & 7) || (output_size & 7)); if (float4_predicate) { //aligned dotBasedInteractFwdKernelFP16<warps_per_threadblock, threadblock_size, M_BLOCKS, K_BLOCKS, SMEM_STRIDE, SMEM_STRIDE_ACC, kWarpSize, kWarpSizeLog2, kTileDim, kTileDimLog2, true> <<<(batch_size + warps_per_threadblock - 1) / warps_per_threadblock, threadblock_size, warps_per_threadblock * smem_elems_per_warp * sizeof(__half), stream>>>((const __half *)input, (half *)output, batch_size, num_rows, num_cols, num_rows_after_padding, num_cols_after_padding, smem_elems_per_warp, smem_rows_per_warp, output_size, num_row_steps, num_col_steps, padding_size); } else { //not aligned dotBasedInteractFwdKernelFP16<warps_per_threadblock, threadblock_size, M_BLOCKS, K_BLOCKS, SMEM_STRIDE, SMEM_STRIDE_ACC, kWarpSize, kWarpSizeLog2, kTileDim, kTileDimLog2, false> <<<(batch_size + warps_per_threadblock - 1) / warps_per_threadblock, threadblock_size, warps_per_threadblock * smem_elems_per_warp * sizeof(__half), stream>>>((const __half *)input, (half *)output, batch_size, num_rows, num_cols, num_rows_after_padding, num_cols_after_padding, smem_elems_per_warp, smem_rows_per_warp, output_size, num_row_steps, num_col_steps, padding_size); } } inline void dotBasedInteractFP16Bwd(const void *input, const void *upstream_grad, void *grad, void *bottom_mlp_grad, uint batch_size, uint num_rows, uint num_cols, cudaStream_t stream) { const uint kWarpSize = 32; const uint kWarpSizeLog2 = Log2<kWarpSize>::value; const uint kTileDim = 16; const uint kTileDimLog2 = Log2<kTileDim>::value; const uint mem_skew_size = 8; const uint kWarpsPerBlock = 4; const uint kWarpsPerBlockLog2 = Log2<kWarpsPerBlock>::value; const uint kNumThreads = kWarpsPerBlock * kWarpSize; const uint kRowTilesPerStep = 2; const uint kColTilesPerStep = 1; uint row_tiles_per_step = num_rows > kTileDim ? kRowTilesPerStep : 1; // num tiles uint num_row_tiles = (num_rows + kTileDim - 1) >> kTileDimLog2; uint num_col_tiles = (num_cols + kTileDim - 1) >> kTileDimLog2; // number of rows and columns after padding uint num_rows_after_padding = kTileDim << 1; uint num_cols_after_padding = num_col_tiles << kTileDimLog2; // 2D ugrad size and stride uint interaction_ugrad_2D_stride = num_rows_after_padding + mem_skew_size; uint interaction_ugrad_2D_size_elems = num_rows_after_padding * interaction_ugrad_2D_stride; uint interaction_ugrad_2D_size_bytes = interaction_ugrad_2D_size_elems * sizeof(half); // 1D ugrad size uint interaction_ugrad_size = num_rows * (num_rows - 1) >> 1; uint interaction_ugrad_size_with_padding = ((interaction_ugrad_size-1)/8 + 1)*8; //round up to multiple of 8 // in_out place size and stride uint input_stride = num_cols_after_padding + mem_skew_size; uint input_size_elems = num_rows_after_padding * input_stride; uint input_size_bytes = input_size_elems * sizeof(half); // sample size uint sample_size = num_rows * num_cols; // output size uint output_size_elems = kTileDim * kTileDim * kRowTilesPerStep * kColTilesPerStep; uint output_size_bytes = output_size_elems * sizeof(float); // staging area size uint staging_area_size_bytes = output_size_bytes > interaction_ugrad_2D_size_bytes ? output_size_bytes : interaction_ugrad_2D_size_bytes; // Shared memory size uint shared_mem_per_warp_size_byte = input_size_bytes + staging_area_size_bytes; uint shared_mem_size_bytes = kWarpsPerBlock * shared_mem_per_warp_size_byte; uint num_blocks = (batch_size + kWarpsPerBlock - 1) >> kWarpsPerBlockLog2; uint num_row_steps = num_row_tiles / row_tiles_per_step; uint num_col_steps = num_col_tiles / kColTilesPerStep; bool float4_predicate = !((interaction_ugrad_size_with_padding & 7) || (num_cols & 7)); if (float4_predicate) { //aligned dotBasedInteractBwdKernelFP16<kWarpsPerBlock, kNumThreads, kRowTilesPerStep, kColTilesPerStep, kWarpSize, kWarpSizeLog2, kTileDim, kTileDimLog2, true> <<<num_blocks, kNumThreads, shared_mem_size_bytes, stream>>>((const half *)input, (const half *)upstream_grad, (half *)grad, (half *)bottom_mlp_grad, batch_size, num_rows, num_cols, num_rows_after_padding, num_cols_after_padding, sample_size, interaction_ugrad_size, interaction_ugrad_size_with_padding, interaction_ugrad_2D_size_elems, interaction_ugrad_2D_stride, input_size_elems, input_stride, num_row_steps, num_col_steps, row_tiles_per_step, shared_mem_per_warp_size_byte); } else { //unaligned dotBasedInteractBwdKernelFP16<kWarpsPerBlock, kNumThreads, kRowTilesPerStep, kColTilesPerStep, kWarpSize, kWarpSizeLog2, kTileDim, kTileDimLog2, false> <<<num_blocks, kNumThreads, shared_mem_size_bytes, stream>>>((const half *)input, (const half *)upstream_grad, (half *)grad, (half *)bottom_mlp_grad, batch_size, num_rows, num_cols, num_rows_after_padding, num_cols_after_padding, sample_size, interaction_ugrad_size, interaction_ugrad_size_with_padding, interaction_ugrad_2D_size_elems, interaction_ugrad_2D_stride, input_size_elems, input_stride, num_row_steps, num_col_steps, row_tiles_per_step, shared_mem_per_warp_size_byte); } } #endif /* FP16_LAUNCHER_CU */
PaddlePaddle/Classification/RN50v1.5
RN50v1.5
program
# Copyright (c) 2022 NVIDIA Corporation. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import logging import time from profile import Profiler import dllogger import models import numpy as np from lr_scheduler import build_lr_scheduler from optimizer import build_optimizer from utils.misc import AverageMeter from utils.mode import Mode, RunScope from utils.utility import get_num_trainers import paddle import paddle.nn.functional as F from paddle.distributed import fleet from paddle.distributed.fleet import DistributedStrategy from paddle.distributed.fleet.meta_optimizers.common import CollectiveHelper from paddle.incubate import asp as sparsity def create_feeds(image_shape): """ Create feeds mapping for the inputs of Pragrm execution. Args: image_shape(list[int]): Model input shape, such as [4, 224, 224]. Returns: feeds(dict): A dict to map variables'name to their values. key (string): Name of variable to feed. Value (tuple): paddle.static.data. """ feeds = {} feeds['data'] = paddle.static.data( name="data", shape=[None] + image_shape, dtype="float32" ) feeds['label'] = paddle.static.data( name="label", shape=[None, 1], dtype="int64" ) return feeds def create_fetchs(out, feeds, class_num, label_smoothing=0, mode=Mode.TRAIN): """ Create fetchs to obtain specific outputs from Pragrm execution (included loss and measures). Args: out(variable): The model output variable. feeds(dict): A dict of mapping variables'name to their values (The input of Program execution). class_num(int): The number of classes. label_smoothing(float, optional): Epsilon of label smoothing. Default: 0. mode(utils.Mode, optional): Train or eval mode. Default: Mode.TRAIN Returns: fetchs(dict): A dict of outputs from Program execution (included loss and measures). key (string): Name of variable to fetch. Value (tuple): (variable, AverageMeter). """ fetchs = {} target = paddle.reshape(feeds['label'], [-1, 1]) if mode == Mode.TRAIN: if label_smoothing == 0: loss = F.cross_entropy(out, target) else: label_one_hot = F.one_hot(target, class_num) soft_target = F.label_smooth(label_one_hot, epsilon=label_smoothing) soft_target = paddle.reshape(soft_target, shape=[-1, class_num]) log_softmax = -F.log_softmax(out, axis=-1) loss = paddle.sum(log_softmax * soft_target, axis=-1) else: loss = F.cross_entropy(out, target) label = paddle.argmax(out, axis=-1, dtype='int32') fetchs['label'] = (label, None) loss = loss.mean() fetchs['loss'] = (loss, AverageMeter('loss', '7.4f', need_avg=True)) acc_top1 = paddle.metric.accuracy(input=out, label=target, k=1) acc_top5 = paddle.metric.accuracy(input=out, label=target, k=5) metric_dict = {} metric_dict["top1"] = acc_top1 metric_dict["top5"] = acc_top5 for key in metric_dict: if mode != Mode.TRAIN and paddle.distributed.get_world_size() > 1: paddle.distributed.all_reduce( metric_dict[key], op=paddle.distributed.ReduceOp.SUM ) metric_dict[key] = ( metric_dict[key] / paddle.distributed.get_world_size() ) fetchs[key] = ( metric_dict[key], AverageMeter(key, '7.4f', need_avg=True), ) return fetchs def create_strategy(args, is_train=True): """ Create paddle.static.BuildStrategy and paddle.static.ExecutionStrategy with arguments. Args: args(Namespace): Arguments obtained from ArgumentParser. is_train(bool, optional): Indicate the prupose of strategy is for training of not. Default is True. Returns: build_strategy(paddle.static.BuildStrategy): A instance of BuildStrategy. exec_strategy(paddle.static.ExecutionStrategy): A instance of ExecutionStrategy. """ build_strategy = paddle.static.BuildStrategy() exec_strategy = paddle.static.ExecutionStrategy() exec_strategy.num_threads = 1 exec_strategy.num_iteration_per_drop_scope = ( 10000 if args.amp and args.use_pure_fp16 else 10 ) paddle.set_flags( { 'FLAGS_cudnn_exhaustive_search': True, 'FLAGS_conv_workspace_size_limit': 4096, } ) if not is_train: build_strategy.fix_op_run_order = True if args.amp: build_strategy.fuse_bn_act_ops = True build_strategy.fuse_elewise_add_act_ops = True build_strategy.fuse_bn_add_act_ops = True build_strategy.enable_addto = True if args.fuse_resunit and is_train: build_strategy.fuse_resunit = True return build_strategy, exec_strategy def dist_optimizer(args, optimizer): """ Create a distributed optimizer based on a given optimizer. Args: args(Namespace): Arguments obtained from ArgumentParser. optimizer(paddle.optimizer): A normal optimizer. Returns: optimizer(fleet.distributed_optimizer): A distributed optimizer. """ build_strategy, exec_strategy = create_strategy(args) dist_strategy = DistributedStrategy() dist_strategy.execution_strategy = exec_strategy dist_strategy.build_strategy = build_strategy dist_strategy.fuse_all_reduce_ops = True all_reduce_size = 16 dist_strategy.fuse_grad_size_in_MB = all_reduce_size dist_strategy.nccl_comm_num = 1 dist_strategy.sync_nccl_allreduce = True if args.amp: dist_strategy.cudnn_batchnorm_spatial_persistent = True dist_strategy.amp = True dist_strategy.amp_configs = { "init_loss_scaling": args.scale_loss, "use_dynamic_loss_scaling": args.use_dynamic_loss_scaling, "use_pure_fp16": args.use_pure_fp16, } dist_strategy.asp = args.asp optimizer = fleet.distributed_optimizer(optimizer, strategy=dist_strategy) return optimizer def build(args, main_prog, startup_prog, step_each_epoch, is_train=True): """ Build a executable paddle.static.Program via following four steps: 1. Create feeds. 2. Create a model. 3. Create fetchs. 4. Create an optimizer if is_train==True. Args: args(Namespace): Arguments obtained from ArgumentParser. main_prog(paddle.static.Program):The main program. startup_prog(paddle.static.Program):The startup program. step_each_epoch(int): The number of steps in each epoch. is_train(bool, optional): Whether the main programe created is for training. Default: True. Returns: fetchs(dict): A dict of outputs from Program execution (included loss and measures). lr_scheduler(paddle.optimizer.lr.LRScheduler): A learning rate scheduler. feeds(dict): A dict to map variables'name to their values. optimizer(Optimizer): An optimizer with distributed/AMP/ASP strategy. """ with paddle.static.program_guard(main_prog, startup_prog): with paddle.utils.unique_name.guard(): mode = Mode.TRAIN if is_train else Mode.EVAL feeds = create_feeds(args.image_shape) model_name = args.model_arch_name class_num = args.num_of_class input_image_channel = args.image_channel data_format = args.data_layout use_pure_fp16 = args.use_pure_fp16 bn_weight_decay = args.bn_weight_decay model = models.__dict__[model_name]( class_num=class_num, input_image_channel=input_image_channel, data_format=data_format, use_pure_fp16=use_pure_fp16, bn_weight_decay=bn_weight_decay, ) out = model(feeds["data"]) fetchs = create_fetchs( out, feeds, class_num, args.label_smoothing, mode=mode ) if args.asp: sparsity.set_excluded_layers(main_program=main_prog, param_names=[model.fc.weight.name]) lr_scheduler = None optimizer = None if is_train: lr_scheduler = build_lr_scheduler(args, step_each_epoch) optimizer = build_optimizer(args, lr_scheduler) optimizer = dist_optimizer(args, optimizer) optimizer.minimize(fetchs['loss'][0], startup_prog) # This is a workaround to "Communicator of ring id 0 has not been initialized.". # Since Paddle's design, the initialization would be done inside train program, # eval_only need to manually call initialization. if ( args.run_scope == RunScope.EVAL_ONLY and paddle.distributed.get_world_size() > 1 ): collective_helper = CollectiveHelper( role_maker=fleet.PaddleCloudRoleMaker(is_collective=True) ) collective_helper.update_startup_program(startup_prog) return fetchs, lr_scheduler, feeds, optimizer def compile_prog(args, program, loss_name=None, is_train=True): """ Compile the given program, which would fuse computing ops or optimize memory footprint based building strategy in config. Args: args(Namespace): Arguments obtained from ArgumentParser. program(paddle.static.Program): The main program to be compiled. loss_name(str, optional): The name of loss variable. Default: None. is_train(bool, optional): Indicate the prupose of strategy is for training of not. Default is True. Returns: compiled_program(paddle.static.CompiledProgram): A compiled program. """ build_strategy, exec_strategy = create_strategy(args, is_train) compiled_program = paddle.static.CompiledProgram( program, build_strategy=build_strategy ) return compiled_program def run( args, dataloader, exe, program, fetchs, epoch, mode=Mode.TRAIN, lr_scheduler=None, ): """ Execute program. Args: args(Namespace): Arguments obtained from ArgumentParser. dataloader(nvidia.dali.plugin.paddle.DALIGenericIterator): Iteratable output of NVIDIA DALI pipeline, please refer to dali_dataloader in dali.py for details. exe(paddle.static.Executor): A executor to run program. program(paddle.static.Program): The program to be executed. fetchs(dict): A dict of outputs from Program execution (included loss and measures). epoch(int): Current epoch id to run. mode(utils.Mode, optional): Train or eval mode. Default: Mode.TRAIN. lr_scheduler(paddle.optimizer.lr.LRScheduler, optional): A learning rate scheduler. Default: None. Returns: metrics (dict): A dictionary to collect values of metrics. """ num_trainers = get_num_trainers() fetch_list = [f[0] for f in fetchs.values()] metric_dict = {"lr": AverageMeter('lr', 'f', postfix=",", need_avg=False)} for k in fetchs: if fetchs[k][1] is not None: metric_dict[k] = fetchs[k][1] metric_dict["batch_time"] = AverageMeter('batch_time', '.5f', postfix=" s,") metric_dict["data_time"] = AverageMeter('data_time', '.5f', postfix=" s,") metric_dict["compute_time"] = AverageMeter( 'compute_time', '.5f', postfix=" s," ) for m in metric_dict.values(): m.reset() profiler = Profiler() tic = time.perf_counter() idx = 0 batch_size = None latency = [] total_benchmark_steps = args.benchmark_steps + args.benchmark_warmup_steps dataloader.reset() while True: # profiler.profile_setup return True only when # profile is enable and idx == stop steps if profiler.profile_setup(idx): break idx += 1 try: batch = next(dataloader) except StopIteration: # Reset dataloader when run benchmark to fill required steps. if args.benchmark and (idx < total_benchmark_steps): dataloader.reset() # Reset tic timestamp to ignore exception handling time. tic = time.perf_counter() continue break except RuntimeError: logging.warning( "Except RuntimeError when reading data from dataloader, try to read once again..." ) continue reader_toc = time.perf_counter() metric_dict['data_time'].update(reader_toc - tic) batch_size = batch[0]["data"].shape()[0] feed_dict = batch[0] with profiler.profile_tag( idx, "Training" if mode == Mode.TRAIN else "Evaluation" ): results = exe.run( program=program, feed=feed_dict, fetch_list=fetch_list ) for name, m in zip(fetchs.keys(), results): if name in metric_dict: metric_dict[name].update(np.mean(m), batch_size) metric_dict["compute_time"].update(time.perf_counter() - reader_toc) metric_dict["batch_time"].update(time.perf_counter() - tic) if mode == Mode.TRAIN: metric_dict['lr'].update(lr_scheduler.get_lr()) if lr_scheduler is not None: with profiler.profile_tag(idx, "LR Step"): lr_scheduler.step() tic = time.perf_counter() if idx % args.print_interval == 0: log_msg = {} log_msg['loss'] = metric_dict['loss'].val.item() log_msg['top1'] = metric_dict['top1'].val.item() log_msg['top5'] = metric_dict['top5'].val.item() log_msg['data_time'] = metric_dict['data_time'].val log_msg['compute_time'] = metric_dict['compute_time'].val log_msg['batch_time'] = metric_dict['batch_time'].val log_msg['ips'] = ( batch_size * num_trainers / metric_dict['batch_time'].val ) if mode == Mode.TRAIN: log_msg['lr'] = metric_dict['lr'].val log_info((epoch, idx), log_msg, mode) if args.benchmark: latency.append(metric_dict['batch_time'].val) # Ignore the warmup iters if idx == args.benchmark_warmup_steps: metric_dict["compute_time"].reset() metric_dict["data_time"].reset() metric_dict["batch_time"].reset() latency.clear() logging.info("Begin benchmark at step %d", idx + 1) if idx == total_benchmark_steps: benchmark_data = {} benchmark_data['ips'] = ( batch_size * num_trainers / metric_dict['batch_time'].avg ) if mode == mode.EVAL: latency = np.array(latency) * 1000 quantile = np.quantile(latency, [0.9, 0.95, 0.99]) benchmark_data['latency_avg'] = np.mean(latency) benchmark_data['latency_p90'] = quantile[0] benchmark_data['latency_p95'] = quantile[1] benchmark_data['latency_p99'] = quantile[2] logging.info("End benchmark at epoch step %d", idx) return benchmark_data epoch_data = {} epoch_data['loss'] = metric_dict['loss'].avg.item() epoch_data['epoch_time'] = metric_dict['batch_time'].total epoch_data['ips'] = ( batch_size * num_trainers * metric_dict["batch_time"].count / metric_dict["batch_time"].sum ) if mode == Mode.EVAL: epoch_data['top1'] = metric_dict['top1'].avg.item() epoch_data['top5'] = metric_dict['top5'].avg.item() log_info((epoch,), epoch_data, mode) return epoch_data def log_info(step, metrics, mode): """ Log metrics with step and mode information. Args: step(tuple): Step, coulbe (epoch-id, iter-id). Use tuple() for summary. metrics(dict): A dictionary collected values of metrics. mode(utils.Mode): Train or eval mode. """ prefix = 'train' if mode == Mode.TRAIN else 'val' dllogger_iter_data = {} for key in metrics: dllogger_iter_data[f"{prefix}.{key}"] = metrics[key] dllogger.log(step=step, data=dllogger_iter_data)
PyTorch/Segmentation/MaskRCNN/pytorch/maskrcnn_benchmark/utils
utils
README
# Utility functions This folder contain utility functions that are not used in the core library, but are useful for building models or training code using the config system.
PyTorch/Segmentation/MaskRCNN/pytorch/configs/gn_baselines
gn_baselines
scratch_e2e_mask_rcnn_R_50_FPN_3x_gn
INPUT: MIN_SIZE_TRAIN: 800 MAX_SIZE_TRAIN: 1333 MIN_SIZE_TEST: 800 MAX_SIZE_TEST: 1333 MODEL: META_ARCHITECTURE: "GeneralizedRCNN" WEIGHT: "" # no pretrained model BACKBONE: CONV_BODY: "R-50-FPN" OUT_CHANNELS: 256 FREEZE_CONV_BODY_AT: 0 # finetune all layers RESNETS: # use GN for backbone TRANS_FUNC: "BottleneckWithGN" STEM_FUNC: "StemWithGN" FPN: USE_GN: True # use GN for FPN RPN: USE_FPN: True ANCHOR_STRIDE: (4, 8, 16, 32, 64) PRE_NMS_TOP_N_TRAIN: 2000 PRE_NMS_TOP_N_TEST: 1000 POST_NMS_TOP_N_TEST: 1000 FPN_POST_NMS_TOP_N_TEST: 1000 ROI_HEADS: USE_FPN: True BATCH_SIZE_PER_IMAGE: 512 POSITIVE_FRACTION: 0.25 ROI_BOX_HEAD: USE_GN: True # use GN for bbox head POOLER_RESOLUTION: 7 POOLER_SCALES: (0.25, 0.125, 0.0625, 0.03125) POOLER_SAMPLING_RATIO: 2 FEATURE_EXTRACTOR: "FPN2MLPFeatureExtractor" PREDICTOR: "FPNPredictor" ROI_MASK_HEAD: USE_GN: True # use GN for mask head POOLER_SCALES: (0.25, 0.125, 0.0625, 0.03125) CONV_LAYERS: (256, 256, 256, 256) FEATURE_EXTRACTOR: "MaskRCNNFPNFeatureExtractor" PREDICTOR: "MaskRCNNC4Predictor" POOLER_RESOLUTION: 14 POOLER_SAMPLING_RATIO: 2 RESOLUTION: 28 SHARE_BOX_FEATURE_EXTRACTOR: False MASK_ON: True DATASETS: TRAIN: ("coco_2014_train", "coco_2014_valminusminival") TEST: ("coco_2014_minival",) DATALOADER: SIZE_DIVISIBILITY: 32 SOLVER: # Assume 8 gpus BASE_LR: 0.02 WEIGHT_DECAY: 0.0001 STEPS: (210000, 250000) MAX_ITER: 270000 IMS_PER_BATCH: 16 TEST: IMS_PER_BATCH: 8
TensorFlow/Detection/SSD/models/research/object_detection/configs
configs
mask_rcnn_resnet50_atrous_coco_8GPU
# Mask R-CNN with Resnet-50 (v1), Atrous version # Configured for MSCOCO Dataset. # Users should configure the fine_tune_checkpoint field in the train config as # well as the label_map_path and input_path fields in the train_input_reader and # eval_input_reader. Search for "PATH_TO_BE_CONFIGURED" to find the fields that # should be configured. # # Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. model { faster_rcnn { num_classes: 90 image_resizer { keep_aspect_ratio_resizer { min_dimension: 800 max_dimension: 1365 } } number_of_stages: 3 feature_extractor { type: 'faster_rcnn_resnet50' first_stage_features_stride: 8 } first_stage_anchor_generator { grid_anchor_generator { scales: [0.25, 0.5, 1.0, 2.0] aspect_ratios: [0.5, 1.0, 2.0] height_stride: 8 width_stride: 8 } } first_stage_atrous_rate: 2 first_stage_box_predictor_conv_hyperparams { op: CONV regularizer { l2_regularizer { weight: 0.0 } } initializer { truncated_normal_initializer { stddev: 0.01 } } } first_stage_nms_score_threshold: 0.0 first_stage_nms_iou_threshold: 0.7 first_stage_max_proposals: 300 first_stage_localization_loss_weight: 2.0 first_stage_objectness_loss_weight: 1.0 initial_crop_size: 14 maxpool_kernel_size: 2 maxpool_stride: 2 second_stage_box_predictor { mask_rcnn_box_predictor { use_dropout: false dropout_keep_probability: 1.0 predict_instance_masks: true mask_height: 33 mask_width: 33 mask_prediction_conv_depth: 0 mask_prediction_num_conv_layers: 4 fc_hyperparams { op: FC regularizer { l2_regularizer { weight: 0.0 } } initializer { variance_scaling_initializer { factor: 1.0 uniform: true mode: FAN_AVG } } } conv_hyperparams { op: CONV regularizer { l2_regularizer { weight: 0.0 } } initializer { truncated_normal_initializer { stddev: 0.01 } } } } } second_stage_post_processing { batch_non_max_suppression { score_threshold: 0.0 iou_threshold: 0.6 max_detections_per_class: 100 max_total_detections: 300 } score_converter: SOFTMAX } second_stage_localization_loss_weight: 2.0 second_stage_classification_loss_weight: 1.0 second_stage_mask_prediction_loss_weight: 4.0 } } train_config: { batch_size: 1 optimizer { momentum_optimizer: { learning_rate: { manual_step_learning_rate { initial_learning_rate: 0.0024 schedule { step: 900000 learning_rate: .00024 } schedule { step: 1200000 learning_rate: .000024 } } } momentum_optimizer_value: 0.9 } use_moving_average: false } gradient_clipping_by_norm: 10.0 fine_tune_checkpoint: "/checkpoints/resnet_v1_50/model.ckpt" fine_tune_checkpoint_type: "classification" # Note: The below line limits the training process to 200K steps, which we # empirically found to be sufficient enough to train the pets dataset. This # effectively bypasses the learning rate schedule (the learning rate will # never decay). Remove the below line to train indefinitely. num_steps: 200000 data_augmentation_options { random_horizontal_flip { } } } train_input_reader: { tf_record_input_reader { input_path: "/data/coco_train.record*" } label_map_path: "object_detection/data/mscoco_label_map.pbtxt" } eval_config: { metrics_set: "coco_detection_metrics" use_moving_averages: false num_examples: 8000 } eval_input_reader: { tf_record_input_reader { input_path: "/data/coco_val.record*" } label_map_path: "object_detection/data/mscoco_label_map.pbtxt" shuffle: false num_readers: 1 }
PyTorch/Translation/GNMT/seq2seq/train
train
table
# Copyright (c) 2018-2020, NVIDIA CORPORATION. All rights reserved. # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all # copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. from pytablewriter import MarkdownTableWriter class TrainingTable: def __init__(self, acc_unit='BLEU', time_unit='min', perf_unit='tok/s'): self.data = [] self.acc_unit = acc_unit self.time_unit = time_unit self.perf_unit = perf_unit self.time_unit_convert = {'s': 1, 'min': 1/60, 'h': 1/3600} def add(self, gpus, batch_size, accuracy, perf, time_to_train): time_to_train *= self.time_unit_convert[self.time_unit] if not accuracy: accuracy = 0.0 accuracy = round(accuracy, 2) self.data.append([gpus, batch_size, accuracy, perf, time_to_train]) def write(self, title, math): writer = MarkdownTableWriter() writer.table_name = f'{title}' header = [f'**GPUs**', f'**Batch Size / GPU**', f'**Accuracy - {math.upper()} ({self.acc_unit})**', f'**Throughput - {math.upper()} ({self.perf_unit})**', f'**Time to Train - {math.upper()} ({self.time_unit})**', ] writer.headers = header writer.value_matrix = self.data writer.write_table()
PyTorch/SpeechSynthesis/Tacotron2/platform
platform
DGX1_tacotron2_AMP_8NGPU_train
mkdir -p output python -m multiproc train.py -m Tacotron2 -o output/ --amp -lr 1e-3 --epochs 1501 -bs 104 --weight-decay 1e-6 --grad-clip-thresh 1.0 --cudnn-enabled --load-mel-from-disk --training-files=filelists/ljs_mel_text_train_filelist.txt --validation-files=filelists/ljs_mel_text_val_filelist.txt --log-file nvlog.json --anneal-steps 500 1000 1500 --anneal-factor 0.3
TensorFlow2/LanguageModeling/BERT/official/utils/misc
misc
callstack_sampler
"""A simple Python callstack sampler.""" import contextlib import datetime import signal import traceback class CallstackSampler(object): """A simple signal-based Python callstack sampler. """ def __init__(self, interval=None): self.stacks = [] self.interval = 0.001 if interval is None else interval def _sample(self, signum, frame): """Samples the current stack.""" del signum stack = traceback.extract_stack(frame) formatted_stack = [] formatted_stack.append(datetime.datetime.utcnow()) for filename, lineno, function_name, text in stack: formatted_frame = '{}:{}({})({})'.format(filename, lineno, function_name, text) formatted_stack.append(formatted_frame) self.stacks.append(formatted_stack) signal.setitimer(signal.ITIMER_VIRTUAL, self.interval, 0) @contextlib.contextmanager def profile(self): signal.signal(signal.SIGVTALRM, self._sample) signal.setitimer(signal.ITIMER_VIRTUAL, self.interval, 0) try: yield finally: signal.setitimer(signal.ITIMER_VIRTUAL, 0) def save(self, fname): with open(fname, 'w') as f: for s in self.stacks: for l in s: f.write('%s\n' % l) f.write('\n') @contextlib.contextmanager def callstack_sampling(filename, interval=None): """Periodically samples the Python callstack. Args: filename: the filename interval: the sampling interval, in seconds. Defaults to 0.001. Yields: nothing """ sampler = CallstackSampler(interval=interval) with sampler.profile(): yield sampler.save(filename)
TensorFlow/Detection/SSD/models/research/object_detection/utils
utils
ops
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """A module for helper tensorflow ops.""" import collections import math import numpy as np import six import tensorflow as tf from object_detection.core import standard_fields as fields from object_detection.utils import shape_utils from object_detection.utils import static_shape def expanded_shape(orig_shape, start_dim, num_dims): """Inserts multiple ones into a shape vector. Inserts an all-1 vector of length num_dims at position start_dim into a shape. Can be combined with tf.reshape to generalize tf.expand_dims. Args: orig_shape: the shape into which the all-1 vector is added (int32 vector) start_dim: insertion position (int scalar) num_dims: length of the inserted all-1 vector (int scalar) Returns: An int32 vector of length tf.size(orig_shape) + num_dims. """ with tf.name_scope('ExpandedShape'): start_dim = tf.expand_dims(start_dim, 0) # scalar to rank-1 before = tf.slice(orig_shape, [0], start_dim) add_shape = tf.ones(tf.reshape(num_dims, [1]), dtype=tf.int32) after = tf.slice(orig_shape, start_dim, [-1]) new_shape = tf.concat([before, add_shape, after], 0) return new_shape def normalized_to_image_coordinates(normalized_boxes, image_shape, parallel_iterations=32): """Converts a batch of boxes from normal to image coordinates. Args: normalized_boxes: a float32 tensor of shape [None, num_boxes, 4] in normalized coordinates. image_shape: a float32 tensor of shape [4] containing the image shape. parallel_iterations: parallelism for the map_fn op. Returns: absolute_boxes: a float32 tensor of shape [None, num_boxes, 4] containing the boxes in image coordinates. """ x_scale = tf.cast(image_shape[2], tf.float32) y_scale = tf.cast(image_shape[1], tf.float32) def _to_absolute_coordinates(normalized_boxes): y_min, x_min, y_max, x_max = tf.split( value=normalized_boxes, num_or_size_splits=4, axis=1) y_min = y_scale * y_min y_max = y_scale * y_max x_min = x_scale * x_min x_max = x_scale * x_max scaled_boxes = tf.concat([y_min, x_min, y_max, x_max], 1) return scaled_boxes absolute_boxes = shape_utils.static_or_dynamic_map_fn( _to_absolute_coordinates, elems=(normalized_boxes), dtype=tf.float32, parallel_iterations=parallel_iterations, back_prop=True) return absolute_boxes def meshgrid(x, y): """Tiles the contents of x and y into a pair of grids. Multidimensional analog of numpy.meshgrid, giving the same behavior if x and y are vectors. Generally, this will give: xgrid(i1, ..., i_m, j_1, ..., j_n) = x(j_1, ..., j_n) ygrid(i1, ..., i_m, j_1, ..., j_n) = y(i_1, ..., i_m) Keep in mind that the order of the arguments and outputs is reverse relative to the order of the indices they go into, done for compatibility with numpy. The output tensors have the same shapes. Specifically: xgrid.get_shape() = y.get_shape().concatenate(x.get_shape()) ygrid.get_shape() = y.get_shape().concatenate(x.get_shape()) Args: x: A tensor of arbitrary shape and rank. xgrid will contain these values varying in its last dimensions. y: A tensor of arbitrary shape and rank. ygrid will contain these values varying in its first dimensions. Returns: A tuple of tensors (xgrid, ygrid). """ with tf.name_scope('Meshgrid'): x = tf.convert_to_tensor(x) y = tf.convert_to_tensor(y) x_exp_shape = expanded_shape(tf.shape(x), 0, tf.rank(y)) y_exp_shape = expanded_shape(tf.shape(y), tf.rank(y), tf.rank(x)) xgrid = tf.tile(tf.reshape(x, x_exp_shape), y_exp_shape) ygrid = tf.tile(tf.reshape(y, y_exp_shape), x_exp_shape) new_shape = y.get_shape().concatenate(x.get_shape()) xgrid.set_shape(new_shape) ygrid.set_shape(new_shape) return xgrid, ygrid def fixed_padding(inputs, kernel_size, rate=1): """Pads the input along the spatial dimensions independently of input size. Args: inputs: A tensor of size [batch, height_in, width_in, channels]. kernel_size: The kernel to be used in the conv2d or max_pool2d operation. Should be a positive integer. rate: An integer, rate for atrous convolution. Returns: output: A tensor of size [batch, height_out, width_out, channels] with the input, either intact (if kernel_size == 1) or padded (if kernel_size > 1). """ kernel_size_effective = kernel_size + (kernel_size - 1) * (rate - 1) pad_total = kernel_size_effective - 1 pad_beg = pad_total // 2 pad_end = pad_total - pad_beg padded_inputs = tf.pad(inputs, [[0, 0], [pad_beg, pad_end], [pad_beg, pad_end], [0, 0]]) return padded_inputs def pad_to_multiple(tensor, multiple): """Returns the tensor zero padded to the specified multiple. Appends 0s to the end of the first and second dimension (height and width) of the tensor until both dimensions are a multiple of the input argument 'multiple'. E.g. given an input tensor of shape [1, 3, 5, 1] and an input multiple of 4, PadToMultiple will append 0s so that the resulting tensor will be of shape [1, 4, 8, 1]. Args: tensor: rank 4 float32 tensor, where tensor -> [batch_size, height, width, channels]. multiple: the multiple to pad to. Returns: padded_tensor: the tensor zero padded to the specified multiple. """ if multiple == 1: return tensor tensor_shape = tensor.get_shape() batch_size = static_shape.get_batch_size(tensor_shape) tensor_height = static_shape.get_height(tensor_shape) tensor_width = static_shape.get_width(tensor_shape) tensor_depth = static_shape.get_depth(tensor_shape) if batch_size is None: batch_size = tf.shape(tensor)[0] if tensor_height is None: tensor_height = tf.shape(tensor)[1] padded_tensor_height = tf.to_int32( tf.ceil(tf.to_float(tensor_height) / tf.to_float(multiple))) * multiple else: padded_tensor_height = int( math.ceil(float(tensor_height) / multiple) * multiple) if tensor_width is None: tensor_width = tf.shape(tensor)[2] padded_tensor_width = tf.to_int32( tf.ceil(tf.to_float(tensor_width) / tf.to_float(multiple))) * multiple else: padded_tensor_width = int( math.ceil(float(tensor_width) / multiple) * multiple) if tensor_depth is None: tensor_depth = tf.shape(tensor)[3] # Use tf.concat instead of tf.pad to preserve static shape if padded_tensor_height != tensor_height: height_pad = tf.zeros([ batch_size, padded_tensor_height - tensor_height, tensor_width, tensor_depth ]) tensor = tf.concat([tensor, height_pad], 1) if padded_tensor_width != tensor_width: width_pad = tf.zeros([ batch_size, padded_tensor_height, padded_tensor_width - tensor_width, tensor_depth ]) tensor = tf.concat([tensor, width_pad], 2) return tensor def padded_one_hot_encoding(indices, depth, left_pad): """Returns a zero padded one-hot tensor. This function converts a sparse representation of indices (e.g., [4]) to a zero padded one-hot representation (e.g., [0, 0, 0, 0, 1] with depth = 4 and left_pad = 1). If `indices` is empty, the result will simply be a tensor of shape (0, depth + left_pad). If depth = 0, then this function just returns `None`. Args: indices: an integer tensor of shape [num_indices]. depth: depth for the one-hot tensor (integer). left_pad: number of zeros to left pad the one-hot tensor with (integer). Returns: padded_onehot: a tensor with shape (num_indices, depth + left_pad). Returns `None` if the depth is zero. Raises: ValueError: if `indices` does not have rank 1 or if `left_pad` or `depth are either negative or non-integers. TODO(rathodv): add runtime checks for depth and indices. """ if depth < 0 or not isinstance(depth, six.integer_types): raise ValueError('`depth` must be a non-negative integer.') if left_pad < 0 or not isinstance(left_pad, six.integer_types): raise ValueError('`left_pad` must be a non-negative integer.') if depth == 0: return None rank = len(indices.get_shape().as_list()) if rank != 1: raise ValueError('`indices` must have rank 1, but has rank=%s' % rank) def one_hot_and_pad(): one_hot = tf.cast(tf.one_hot(tf.cast(indices, tf.int64), depth, on_value=1, off_value=0), tf.float32) return tf.pad(one_hot, [[0, 0], [left_pad, 0]], mode='CONSTANT') result = tf.cond(tf.greater(tf.size(indices), 0), one_hot_and_pad, lambda: tf.zeros((depth + left_pad, 0))) return tf.reshape(result, [-1, depth + left_pad]) def dense_to_sparse_boxes(dense_locations, dense_num_boxes, num_classes): """Converts bounding boxes from dense to sparse form. Args: dense_locations: a [max_num_boxes, 4] tensor in which only the first k rows are valid bounding box location coordinates, where k is the sum of elements in dense_num_boxes. dense_num_boxes: a [max_num_classes] tensor indicating the counts of various bounding box classes e.g. [1, 0, 0, 2] means that the first bounding box is of class 0 and the second and third bounding boxes are of class 3. The sum of elements in this tensor is the number of valid bounding boxes. num_classes: number of classes Returns: box_locations: a [num_boxes, 4] tensor containing only valid bounding boxes (i.e. the first num_boxes rows of dense_locations) box_classes: a [num_boxes] tensor containing the classes of each bounding box (e.g. dense_num_boxes = [1, 0, 0, 2] => box_classes = [0, 3, 3] """ num_valid_boxes = tf.reduce_sum(dense_num_boxes) box_locations = tf.slice(dense_locations, tf.constant([0, 0]), tf.stack([num_valid_boxes, 4])) tiled_classes = [tf.tile([i], tf.expand_dims(dense_num_boxes[i], 0)) for i in range(num_classes)] box_classes = tf.concat(tiled_classes, 0) box_locations.set_shape([None, 4]) return box_locations, box_classes def indices_to_dense_vector(indices, size, indices_value=1., default_value=0, dtype=tf.float32): """Creates dense vector with indices set to specific value and rest to zeros. This function exists because it is unclear if it is safe to use tf.sparse_to_dense(indices, [size], 1, validate_indices=False) with indices which are not ordered. This function accepts a dynamic size (e.g. tf.shape(tensor)[0]) Args: indices: 1d Tensor with integer indices which are to be set to indices_values. size: scalar with size (integer) of output Tensor. indices_value: values of elements specified by indices in the output vector default_value: values of other elements in the output vector. dtype: data type. Returns: dense 1D Tensor of shape [size] with indices set to indices_values and the rest set to default_value. """ size = tf.to_int32(size) zeros = tf.ones([size], dtype=dtype) * default_value values = tf.ones_like(indices, dtype=dtype) * indices_value return tf.dynamic_stitch([tf.range(size), tf.to_int32(indices)], [zeros, values]) def reduce_sum_trailing_dimensions(tensor, ndims): """Computes sum across all dimensions following first `ndims` dimensions.""" return tf.reduce_sum(tensor, axis=tuple(range(ndims, tensor.shape.ndims))) def retain_groundtruth(tensor_dict, valid_indices): """Retains groundtruth by valid indices. Args: tensor_dict: a dictionary of following groundtruth tensors - fields.InputDataFields.groundtruth_boxes fields.InputDataFields.groundtruth_classes fields.InputDataFields.groundtruth_keypoints fields.InputDataFields.groundtruth_instance_masks fields.InputDataFields.groundtruth_is_crowd fields.InputDataFields.groundtruth_area fields.InputDataFields.groundtruth_label_types fields.InputDataFields.groundtruth_difficult valid_indices: a tensor with valid indices for the box-level groundtruth. Returns: a dictionary of tensors containing only the groundtruth for valid_indices. Raises: ValueError: If the shape of valid_indices is invalid. ValueError: field fields.InputDataFields.groundtruth_boxes is not present in tensor_dict. """ input_shape = valid_indices.get_shape().as_list() if not (len(input_shape) == 1 or (len(input_shape) == 2 and input_shape[1] == 1)): raise ValueError('The shape of valid_indices is invalid.') valid_indices = tf.reshape(valid_indices, [-1]) valid_dict = {} if fields.InputDataFields.groundtruth_boxes in tensor_dict: # Prevents reshape failure when num_boxes is 0. num_boxes = tf.maximum(tf.shape( tensor_dict[fields.InputDataFields.groundtruth_boxes])[0], 1) for key in tensor_dict: if key in [fields.InputDataFields.groundtruth_boxes, fields.InputDataFields.groundtruth_classes, fields.InputDataFields.groundtruth_keypoints, fields.InputDataFields.groundtruth_instance_masks]: valid_dict[key] = tf.gather(tensor_dict[key], valid_indices) # Input decoder returns empty tensor when these fields are not provided. # Needs to reshape into [num_boxes, -1] for tf.gather() to work. elif key in [fields.InputDataFields.groundtruth_is_crowd, fields.InputDataFields.groundtruth_area, fields.InputDataFields.groundtruth_difficult, fields.InputDataFields.groundtruth_label_types]: valid_dict[key] = tf.reshape( tf.gather(tf.reshape(tensor_dict[key], [num_boxes, -1]), valid_indices), [-1]) # Fields that are not associated with boxes. else: valid_dict[key] = tensor_dict[key] else: raise ValueError('%s not present in input tensor dict.' % ( fields.InputDataFields.groundtruth_boxes)) return valid_dict def retain_groundtruth_with_positive_classes(tensor_dict): """Retains only groundtruth with positive class ids. Args: tensor_dict: a dictionary of following groundtruth tensors - fields.InputDataFields.groundtruth_boxes fields.InputDataFields.groundtruth_classes fields.InputDataFields.groundtruth_keypoints fields.InputDataFields.groundtruth_instance_masks fields.InputDataFields.groundtruth_is_crowd fields.InputDataFields.groundtruth_area fields.InputDataFields.groundtruth_label_types fields.InputDataFields.groundtruth_difficult Returns: a dictionary of tensors containing only the groundtruth with positive classes. Raises: ValueError: If groundtruth_classes tensor is not in tensor_dict. """ if fields.InputDataFields.groundtruth_classes not in tensor_dict: raise ValueError('`groundtruth classes` not in tensor_dict.') keep_indices = tf.where(tf.greater( tensor_dict[fields.InputDataFields.groundtruth_classes], 0)) return retain_groundtruth(tensor_dict, keep_indices) def replace_nan_groundtruth_label_scores_with_ones(label_scores): """Replaces nan label scores with 1.0. Args: label_scores: a tensor containing object annoation label scores. Returns: a tensor where NaN label scores have been replaced by ones. """ return tf.where( tf.is_nan(label_scores), tf.ones(tf.shape(label_scores)), label_scores) def filter_groundtruth_with_crowd_boxes(tensor_dict): """Filters out groundtruth with boxes corresponding to crowd. Args: tensor_dict: a dictionary of following groundtruth tensors - fields.InputDataFields.groundtruth_boxes fields.InputDataFields.groundtruth_classes fields.InputDataFields.groundtruth_keypoints fields.InputDataFields.groundtruth_instance_masks fields.InputDataFields.groundtruth_is_crowd fields.InputDataFields.groundtruth_area fields.InputDataFields.groundtruth_label_types Returns: a dictionary of tensors containing only the groundtruth that have bounding boxes. """ if fields.InputDataFields.groundtruth_is_crowd in tensor_dict: is_crowd = tensor_dict[fields.InputDataFields.groundtruth_is_crowd] is_not_crowd = tf.logical_not(is_crowd) is_not_crowd_indices = tf.where(is_not_crowd) tensor_dict = retain_groundtruth(tensor_dict, is_not_crowd_indices) return tensor_dict def filter_groundtruth_with_nan_box_coordinates(tensor_dict): """Filters out groundtruth with no bounding boxes. Args: tensor_dict: a dictionary of following groundtruth tensors - fields.InputDataFields.groundtruth_boxes fields.InputDataFields.groundtruth_classes fields.InputDataFields.groundtruth_keypoints fields.InputDataFields.groundtruth_instance_masks fields.InputDataFields.groundtruth_is_crowd fields.InputDataFields.groundtruth_area fields.InputDataFields.groundtruth_label_types Returns: a dictionary of tensors containing only the groundtruth that have bounding boxes. """ groundtruth_boxes = tensor_dict[fields.InputDataFields.groundtruth_boxes] nan_indicator_vector = tf.greater(tf.reduce_sum(tf.to_int32( tf.is_nan(groundtruth_boxes)), reduction_indices=[1]), 0) valid_indicator_vector = tf.logical_not(nan_indicator_vector) valid_indices = tf.where(valid_indicator_vector) return retain_groundtruth(tensor_dict, valid_indices) def normalize_to_target(inputs, target_norm_value, dim, epsilon=1e-7, trainable=True, scope='NormalizeToTarget', summarize=True): """L2 normalizes the inputs across the specified dimension to a target norm. This op implements the L2 Normalization layer introduced in Liu, Wei, et al. "SSD: Single Shot MultiBox Detector." and Liu, Wei, Andrew Rabinovich, and Alexander C. Berg. "Parsenet: Looking wider to see better." and is useful for bringing activations from multiple layers in a convnet to a standard scale. Note that the rank of `inputs` must be known and the dimension to which normalization is to be applied should be statically defined. TODO(jonathanhuang): Add option to scale by L2 norm of the entire input. Args: inputs: A `Tensor` of arbitrary size. target_norm_value: A float value that specifies an initial target norm or a list of floats (whose length must be equal to the depth along the dimension to be normalized) specifying a per-dimension multiplier after normalization. dim: The dimension along which the input is normalized. epsilon: A small value to add to the inputs to avoid dividing by zero. trainable: Whether the norm is trainable or not scope: Optional scope for variable_scope. summarize: Whether or not to add a tensorflow summary for the op. Returns: The input tensor normalized to the specified target norm. Raises: ValueError: If dim is smaller than the number of dimensions in 'inputs'. ValueError: If target_norm_value is not a float or a list of floats with length equal to the depth along the dimension to be normalized. """ with tf.variable_scope(scope, 'NormalizeToTarget', [inputs]): if not inputs.get_shape(): raise ValueError('The input rank must be known.') input_shape = inputs.get_shape().as_list() input_rank = len(input_shape) if dim < 0 or dim >= input_rank: raise ValueError( 'dim must be non-negative but smaller than the input rank.') if not input_shape[dim]: raise ValueError('input shape should be statically defined along ' 'the specified dimension.') depth = input_shape[dim] if not (isinstance(target_norm_value, float) or (isinstance(target_norm_value, list) and len(target_norm_value) == depth) and all([isinstance(val, float) for val in target_norm_value])): raise ValueError('target_norm_value must be a float or a list of floats ' 'with length equal to the depth along the dimension to ' 'be normalized.') if isinstance(target_norm_value, float): initial_norm = depth * [target_norm_value] else: initial_norm = target_norm_value target_norm = tf.contrib.framework.model_variable( name='weights', dtype=tf.float32, initializer=tf.constant(initial_norm, dtype=tf.float32), trainable=trainable) if summarize: mean = tf.reduce_mean(target_norm) mean = tf.Print(mean, ['NormalizeToTarget:', mean]) tf.summary.scalar(tf.get_variable_scope().name, mean) lengths = epsilon + tf.sqrt(tf.reduce_sum(tf.square(inputs), dim, True)) mult_shape = input_rank*[1] mult_shape[dim] = depth return tf.reshape(target_norm, mult_shape) * tf.truediv(inputs, lengths) def batch_position_sensitive_crop_regions(images, boxes, crop_size, num_spatial_bins, global_pool, parallel_iterations=64): """Position sensitive crop with batches of images and boxes. This op is exactly like `position_sensitive_crop_regions` below but operates on batches of images and boxes. See `position_sensitive_crop_regions` function below for the operation applied per batch element. Args: images: A `Tensor`. Must be one of the following types: `uint8`, `int8`, `int16`, `int32`, `int64`, `half`, `float32`, `float64`. A 4-D tensor of shape `[batch, image_height, image_width, depth]`. Both `image_height` and `image_width` need to be positive. boxes: A `Tensor` of type `float32`. A 3-D tensor of shape `[batch, num_boxes, 4]`. Each box is specified in normalized coordinates `[y1, x1, y2, x2]`. A normalized coordinate value of `y` is mapped to the image coordinate at `y * (image_height - 1)`, so as the `[0, 1]` interval of normalized image height is mapped to `[0, image_height - 1] in image height coordinates. We do allow y1 > y2, in which case the sampled crop is an up-down flipped version of the original image. The width dimension is treated similarly. crop_size: See `position_sensitive_crop_regions` below. num_spatial_bins: See `position_sensitive_crop_regions` below. global_pool: See `position_sensitive_crop_regions` below. parallel_iterations: Number of batch items to process in parallel. Returns: """ def _position_sensitive_crop_fn(inputs): images, boxes = inputs return position_sensitive_crop_regions( images, boxes, crop_size=crop_size, num_spatial_bins=num_spatial_bins, global_pool=global_pool) return shape_utils.static_or_dynamic_map_fn( _position_sensitive_crop_fn, elems=[images, boxes], dtype=tf.float32, parallel_iterations=parallel_iterations) def position_sensitive_crop_regions(image, boxes, crop_size, num_spatial_bins, global_pool): """Position-sensitive crop and pool rectangular regions from a feature grid. The output crops are split into `spatial_bins_y` vertical bins and `spatial_bins_x` horizontal bins. For each intersection of a vertical and a horizontal bin the output values are gathered by performing `tf.image.crop_and_resize` (bilinear resampling) on a a separate subset of channels of the image. This reduces `depth` by a factor of `(spatial_bins_y * spatial_bins_x)`. When global_pool is True, this function implements a differentiable version of position-sensitive RoI pooling used in [R-FCN detection system](https://arxiv.org/abs/1605.06409). When global_pool is False, this function implements a differentiable version of position-sensitive assembling operation used in [instance FCN](https://arxiv.org/abs/1603.08678). Args: image: A `Tensor`. Must be one of the following types: `uint8`, `int8`, `int16`, `int32`, `int64`, `half`, `float32`, `float64`. A 3-D tensor of shape `[image_height, image_width, depth]`. Both `image_height` and `image_width` need to be positive. boxes: A `Tensor` of type `float32`. A 2-D tensor of shape `[num_boxes, 4]`. Each box is specified in normalized coordinates `[y1, x1, y2, x2]`. A normalized coordinate value of `y` is mapped to the image coordinate at `y * (image_height - 1)`, so as the `[0, 1]` interval of normalized image height is mapped to `[0, image_height - 1] in image height coordinates. We do allow y1 > y2, in which case the sampled crop is an up-down flipped version of the original image. The width dimension is treated similarly. crop_size: A list of two integers `[crop_height, crop_width]`. All cropped image patches are resized to this size. The aspect ratio of the image content is not preserved. Both `crop_height` and `crop_width` need to be positive. num_spatial_bins: A list of two integers `[spatial_bins_y, spatial_bins_x]`. Represents the number of position-sensitive bins in y and x directions. Both values should be >= 1. `crop_height` should be divisible by `spatial_bins_y`, and similarly for width. The number of image channels should be divisible by (spatial_bins_y * spatial_bins_x). Suggested value from R-FCN paper: [3, 3]. global_pool: A boolean variable. If True, we perform average global pooling on the features assembled from the position-sensitive score maps. If False, we keep the position-pooled features without global pooling over the spatial coordinates. Note that using global_pool=True is equivalent to but more efficient than running the function with global_pool=False and then performing global average pooling. Returns: position_sensitive_features: A 4-D tensor of shape `[num_boxes, K, K, crop_channels]`, where `crop_channels = depth / (spatial_bins_y * spatial_bins_x)`, where K = 1 when global_pool is True (Average-pooled cropped regions), and K = crop_size when global_pool is False. Raises: ValueError: Raised in four situations: `num_spatial_bins` is not >= 1; `num_spatial_bins` does not divide `crop_size`; `(spatial_bins_y*spatial_bins_x)` does not divide `depth`; `bin_crop_size` is not square when global_pool=False due to the constraint in function space_to_depth. """ total_bins = 1 bin_crop_size = [] for (num_bins, crop_dim) in zip(num_spatial_bins, crop_size): if num_bins < 1: raise ValueError('num_spatial_bins should be >= 1') if crop_dim % num_bins != 0: raise ValueError('crop_size should be divisible by num_spatial_bins') total_bins *= num_bins bin_crop_size.append(crop_dim // num_bins) if not global_pool and bin_crop_size[0] != bin_crop_size[1]: raise ValueError('Only support square bin crop size for now.') ymin, xmin, ymax, xmax = tf.unstack(boxes, axis=1) spatial_bins_y, spatial_bins_x = num_spatial_bins # Split each box into spatial_bins_y * spatial_bins_x bins. position_sensitive_boxes = [] for bin_y in range(spatial_bins_y): step_y = (ymax - ymin) / spatial_bins_y for bin_x in range(spatial_bins_x): step_x = (xmax - xmin) / spatial_bins_x box_coordinates = [ymin + bin_y * step_y, xmin + bin_x * step_x, ymin + (bin_y + 1) * step_y, xmin + (bin_x + 1) * step_x, ] position_sensitive_boxes.append(tf.stack(box_coordinates, axis=1)) image_splits = tf.split(value=image, num_or_size_splits=total_bins, axis=2) image_crops = [] for (split, box) in zip(image_splits, position_sensitive_boxes): if split.shape.is_fully_defined() and box.shape.is_fully_defined(): crop = tf.squeeze( matmul_crop_and_resize( tf.expand_dims(split, axis=0), tf.expand_dims(box, axis=0), bin_crop_size), axis=0) else: crop = tf.image.crop_and_resize( tf.expand_dims(split, 0), box, tf.zeros(tf.shape(boxes)[0], dtype=tf.int32), bin_crop_size) image_crops.append(crop) if global_pool: # Average over all bins. position_sensitive_features = tf.add_n(image_crops) / len(image_crops) # Then average over spatial positions within the bins. position_sensitive_features = tf.reduce_mean( position_sensitive_features, [1, 2], keep_dims=True) else: # Reorder height/width to depth channel. block_size = bin_crop_size[0] if block_size >= 2: image_crops = [tf.space_to_depth( crop, block_size=block_size) for crop in image_crops] # Pack image_crops so that first dimension is for position-senstive boxes. position_sensitive_features = tf.stack(image_crops, axis=0) # Unroll the position-sensitive boxes to spatial positions. position_sensitive_features = tf.squeeze( tf.batch_to_space_nd(position_sensitive_features, block_shape=[1] + num_spatial_bins, crops=tf.zeros((3, 2), dtype=tf.int32)), squeeze_dims=[0]) # Reorder back the depth channel. if block_size >= 2: position_sensitive_features = tf.depth_to_space( position_sensitive_features, block_size=block_size) return position_sensitive_features def reframe_box_masks_to_image_masks(box_masks, boxes, image_height, image_width): """Transforms the box masks back to full image masks. Embeds masks in bounding boxes of larger masks whose shapes correspond to image shape. Args: box_masks: A tf.float32 tensor of size [num_masks, mask_height, mask_width]. boxes: A tf.float32 tensor of size [num_masks, 4] containing the box corners. Row i contains [ymin, xmin, ymax, xmax] of the box corresponding to mask i. Note that the box corners are in normalized coordinates. image_height: Image height. The output mask will have the same height as the image height. image_width: Image width. The output mask will have the same width as the image width. Returns: A tf.float32 tensor of size [num_masks, image_height, image_width]. """ # TODO(rathodv): Make this a public function. def reframe_box_masks_to_image_masks_default(): """The default function when there are more than 0 box masks.""" def transform_boxes_relative_to_boxes(boxes, reference_boxes): boxes = tf.reshape(boxes, [-1, 2, 2]) min_corner = tf.expand_dims(reference_boxes[:, 0:2], 1) max_corner = tf.expand_dims(reference_boxes[:, 2:4], 1) transformed_boxes = (boxes - min_corner) / (max_corner - min_corner) return tf.reshape(transformed_boxes, [-1, 4]) box_masks_expanded = tf.expand_dims(box_masks, axis=3) num_boxes = tf.shape(box_masks_expanded)[0] unit_boxes = tf.concat( [tf.zeros([num_boxes, 2]), tf.ones([num_boxes, 2])], axis=1) reverse_boxes = transform_boxes_relative_to_boxes(unit_boxes, boxes) return tf.image.crop_and_resize( image=box_masks_expanded, boxes=reverse_boxes, box_ind=tf.range(num_boxes), crop_size=[image_height, image_width], extrapolation_value=0.0) image_masks = tf.cond( tf.shape(box_masks)[0] > 0, reframe_box_masks_to_image_masks_default, lambda: tf.zeros([0, image_height, image_width, 1], dtype=tf.float32)) return tf.squeeze(image_masks, axis=3) def merge_boxes_with_multiple_labels(boxes, classes, confidences, num_classes, quantization_bins=10000): """Merges boxes with same coordinates and returns K-hot encoded classes. Args: boxes: A tf.float32 tensor with shape [N, 4] holding N boxes. Only normalized coordinates are allowed. classes: A tf.int32 tensor with shape [N] holding class indices. The class index starts at 0. confidences: A tf.float32 tensor with shape [N] holding class confidences. num_classes: total number of classes to use for K-hot encoding. quantization_bins: the number of bins used to quantize the box coordinate. Returns: merged_boxes: A tf.float32 tensor with shape [N', 4] holding boxes, where N' <= N. class_encodings: A tf.int32 tensor with shape [N', num_classes] holding K-hot encodings for the merged boxes. confidence_encodings: A tf.float32 tensor with shape [N', num_classes] holding encodings of confidences for the merged boxes. merged_box_indices: A tf.int32 tensor with shape [N'] holding original indices of the boxes. """ boxes_shape = tf.shape(boxes) classes_shape = tf.shape(classes) confidences_shape = tf.shape(confidences) box_class_shape_assert = shape_utils.assert_shape_equal_along_first_dimension( boxes_shape, classes_shape) box_confidence_shape_assert = ( shape_utils.assert_shape_equal_along_first_dimension( boxes_shape, confidences_shape)) box_dimension_assert = tf.assert_equal(boxes_shape[1], 4) box_normalized_assert = shape_utils.assert_box_normalized(boxes) with tf.control_dependencies( [box_class_shape_assert, box_confidence_shape_assert, box_dimension_assert, box_normalized_assert]): quantized_boxes = tf.to_int64(boxes * (quantization_bins - 1)) ymin, xmin, ymax, xmax = tf.unstack(quantized_boxes, axis=1) hashcodes = ( ymin + xmin * quantization_bins + ymax * quantization_bins * quantization_bins + xmax * quantization_bins * quantization_bins * quantization_bins) unique_hashcodes, unique_indices = tf.unique(hashcodes) num_boxes = tf.shape(boxes)[0] num_unique_boxes = tf.shape(unique_hashcodes)[0] merged_box_indices = tf.unsorted_segment_min( tf.range(num_boxes), unique_indices, num_unique_boxes) merged_boxes = tf.gather(boxes, merged_box_indices) def map_box_encodings(i): """Produces box K-hot and score encodings for each class index.""" box_mask = tf.equal( unique_indices, i * tf.ones(num_boxes, dtype=tf.int32)) box_mask = tf.reshape(box_mask, [-1]) box_indices = tf.boolean_mask(classes, box_mask) box_confidences = tf.boolean_mask(confidences, box_mask) box_class_encodings = tf.sparse_to_dense( box_indices, [num_classes], 1, validate_indices=False) box_confidence_encodings = tf.sparse_to_dense( box_indices, [num_classes], box_confidences, validate_indices=False) return box_class_encodings, box_confidence_encodings class_encodings, confidence_encodings = tf.map_fn( map_box_encodings, tf.range(num_unique_boxes), back_prop=False, dtype=(tf.int32, tf.float32)) merged_boxes = tf.reshape(merged_boxes, [-1, 4]) class_encodings = tf.reshape(class_encodings, [-1, num_classes]) confidence_encodings = tf.reshape(confidence_encodings, [-1, num_classes]) merged_box_indices = tf.reshape(merged_box_indices, [-1]) return (merged_boxes, class_encodings, confidence_encodings, merged_box_indices) def nearest_neighbor_upsampling(input_tensor, scale=None, height_scale=None, width_scale=None): """Nearest neighbor upsampling implementation. Nearest neighbor upsampling function that maps input tensor with shape [batch_size, height, width, channels] to [batch_size, height * scale , width * scale, channels]. This implementation only uses reshape and broadcasting to make it TPU compatible. Args: input_tensor: A float32 tensor of size [batch, height_in, width_in, channels]. scale: An integer multiple to scale resolution of input data in both height and width dimensions. height_scale: An integer multiple to scale the height of input image. This option when provided overrides `scale` option. width_scale: An integer multiple to scale the width of input image. This option when provided overrides `scale` option. Returns: data_up: A float32 tensor of size [batch, height_in*scale, width_in*scale, channels]. Raises: ValueError: If both scale and height_scale or if both scale and width_scale are None. """ if not scale and (height_scale is None or width_scale is None): raise ValueError('Provide either `scale` or `height_scale` and' ' `width_scale`.') with tf.name_scope('nearest_neighbor_upsampling'): h_scale = scale if height_scale is None else height_scale w_scale = scale if width_scale is None else width_scale (batch_size, height, width, channels) = shape_utils.combined_static_and_dynamic_shape(input_tensor) output_tensor = tf.reshape( input_tensor, [batch_size, height, 1, width, 1, channels]) * tf.ones( [1, 1, h_scale, 1, w_scale, 1], dtype=input_tensor.dtype) return tf.reshape(output_tensor, [batch_size, height * h_scale, width * w_scale, channels]) def matmul_gather_on_zeroth_axis(params, indices, scope=None): """Matrix multiplication based implementation of tf.gather on zeroth axis. TODO(rathodv, jonathanhuang): enable sparse matmul option. Args: params: A float32 Tensor. The tensor from which to gather values. Must be at least rank 1. indices: A Tensor. Must be one of the following types: int32, int64. Must be in range [0, params.shape[0]) scope: A name for the operation (optional). Returns: A Tensor. Has the same type as params. Values from params gathered from indices given by indices, with shape indices.shape + params.shape[1:]. """ with tf.name_scope(scope, 'MatMulGather'): params_shape = shape_utils.combined_static_and_dynamic_shape(params) indices_shape = shape_utils.combined_static_and_dynamic_shape(indices) params2d = tf.reshape(params, [params_shape[0], -1]) indicator_matrix = tf.one_hot(indices, params_shape[0]) gathered_result_flattened = tf.matmul(indicator_matrix, params2d) return tf.reshape(gathered_result_flattened, tf.stack(indices_shape + params_shape[1:])) def matmul_crop_and_resize(image, boxes, crop_size, scope=None): """Matrix multiplication based implementation of the crop and resize op. Extracts crops from the input image tensor and bilinearly resizes them (possibly with aspect ratio change) to a common output size specified by crop_size. This is more general than the crop_to_bounding_box op which extracts a fixed size slice from the input image and does not allow resizing or aspect ratio change. Returns a tensor with crops from the input image at positions defined at the bounding box locations in boxes. The cropped boxes are all resized (with bilinear interpolation) to a fixed size = `[crop_height, crop_width]`. The result is a 5-D tensor `[batch, num_boxes, crop_height, crop_width, depth]`. Running time complexity: O((# channels) * (# boxes) * (crop_size)^2 * M), where M is the number of pixels of the longer edge of the image. Note that this operation is meant to replicate the behavior of the standard tf.image.crop_and_resize operation but there are a few differences. Specifically: 1) The extrapolation value (the values that are interpolated from outside the bounds of the image window) is always zero 2) Only XLA supported operations are used (e.g., matrix multiplication). 3) There is no `box_indices` argument --- to run this op on multiple images, one must currently call this op independently on each image. 4) All shapes and the `crop_size` parameter are assumed to be statically defined. Moreover, the number of boxes must be strictly nonzero. Args: image: A `Tensor`. Must be one of the following types: `uint8`, `int8`, `int16`, `int32`, `int64`, `half`, 'bfloat16', `float32`, `float64`. A 4-D tensor of shape `[batch, image_height, image_width, depth]`. Both `image_height` and `image_width` need to be positive. boxes: A `Tensor` of type `float32` or 'bfloat16'. A 3-D tensor of shape `[batch, num_boxes, 4]`. The boxes are specified in normalized coordinates and are of the form `[y1, x1, y2, x2]`. A normalized coordinate value of `y` is mapped to the image coordinate at `y * (image_height - 1)`, so as the `[0, 1]` interval of normalized image height is mapped to `[0, image_height - 1] in image height coordinates. We do allow y1 > y2, in which case the sampled crop is an up-down flipped version of the original image. The width dimension is treated similarly. Normalized coordinates outside the `[0, 1]` range are allowed, in which case we use `extrapolation_value` to extrapolate the input image values. crop_size: A list of two integers `[crop_height, crop_width]`. All cropped image patches are resized to this size. The aspect ratio of the image content is not preserved. Both `crop_height` and `crop_width` need to be positive. scope: A name for the operation (optional). Returns: A 5-D tensor of shape `[batch, num_boxes, crop_height, crop_width, depth]` Raises: ValueError: if image tensor does not have shape `[batch, image_height, image_width, depth]` and all dimensions statically defined. ValueError: if boxes tensor does not have shape `[batch, num_boxes, 4]` where num_boxes > 0. ValueError: if crop_size is not a list of two positive integers """ img_shape = image.shape.as_list() boxes_shape = boxes.shape.as_list() _, img_height, img_width, _ = img_shape if not isinstance(crop_size, list) or len(crop_size) != 2: raise ValueError('`crop_size` must be a list of length 2') dimensions = img_shape + crop_size + boxes_shape if not all([isinstance(dim, int) for dim in dimensions]): raise ValueError('all input shapes must be statically defined') if len(boxes_shape) != 3 or boxes_shape[2] != 4: raise ValueError('`boxes` should have shape `[batch, num_boxes, 4]`') if len(img_shape) != 4: raise ValueError('image should have shape ' '`[batch, image_height, image_width, depth]`') num_crops = boxes_shape[0] if not num_crops > 0: raise ValueError('number of boxes must be > 0') if not (crop_size[0] > 0 and crop_size[1] > 0): raise ValueError('`crop_size` must be a list of two positive integers.') def _lin_space_weights(num, img_size): if num > 1: start_weights = tf.linspace(img_size - 1.0, 0.0, num) stop_weights = img_size - 1 - start_weights else: start_weights = tf.constant(num * [.5 * (img_size - 1)], dtype=tf.float32) stop_weights = tf.constant(num * [.5 * (img_size - 1)], dtype=tf.float32) return (start_weights, stop_weights) with tf.name_scope(scope, 'MatMulCropAndResize'): y1_weights, y2_weights = _lin_space_weights(crop_size[0], img_height) x1_weights, x2_weights = _lin_space_weights(crop_size[1], img_width) y1_weights = tf.cast(y1_weights, boxes.dtype) y2_weights = tf.cast(y2_weights, boxes.dtype) x1_weights = tf.cast(x1_weights, boxes.dtype) x2_weights = tf.cast(x2_weights, boxes.dtype) [y1, x1, y2, x2] = tf.unstack(boxes, axis=2) # Pixel centers of input image and grid points along height and width image_idx_h = tf.constant( np.reshape(np.arange(img_height), (1, 1, 1, img_height)), dtype=boxes.dtype) image_idx_w = tf.constant( np.reshape(np.arange(img_width), (1, 1, 1, img_width)), dtype=boxes.dtype) grid_pos_h = tf.expand_dims( tf.einsum('ab,c->abc', y1, y1_weights) + tf.einsum( 'ab,c->abc', y2, y2_weights), axis=3) grid_pos_w = tf.expand_dims( tf.einsum('ab,c->abc', x1, x1_weights) + tf.einsum( 'ab,c->abc', x2, x2_weights), axis=3) # Create kernel matrices of pairwise kernel evaluations between pixel # centers of image and grid points. kernel_h = tf.nn.relu(1 - tf.abs(image_idx_h - grid_pos_h)) kernel_w = tf.nn.relu(1 - tf.abs(image_idx_w - grid_pos_w)) # Compute matrix multiplication between the spatial dimensions of the image # and height-wise kernel using einsum. intermediate_image = tf.einsum('abci,aiop->abcop', kernel_h, image) # Compute matrix multiplication between the spatial dimensions of the # intermediate_image and width-wise kernel using einsum. return tf.einsum('abno,abcop->abcnp', kernel_w, intermediate_image) def native_crop_and_resize(image, boxes, crop_size, scope=None): """Same as `matmul_crop_and_resize` but uses tf.image.crop_and_resize.""" def get_box_inds(proposals): proposals_shape = proposals.get_shape().as_list() if any(dim is None for dim in proposals_shape): proposals_shape = tf.shape(proposals) ones_mat = tf.ones(proposals_shape[:2], dtype=tf.int32) multiplier = tf.expand_dims( tf.range(start=0, limit=proposals_shape[0]), 1) return tf.reshape(ones_mat * multiplier, [-1]) with tf.name_scope(scope, 'CropAndResize'): cropped_regions = tf.image.crop_and_resize( image, tf.reshape(boxes, [-1] + boxes.shape.as_list()[2:]), get_box_inds(boxes), crop_size) final_shape = tf.concat([tf.shape(boxes)[:2], tf.shape(cropped_regions)[1:]], axis=0) return tf.reshape(cropped_regions, final_shape) EqualizationLossConfig = collections.namedtuple('EqualizationLossConfig', ['weight', 'exclude_prefixes'])
Tools/DGLPyTorch/SyntheticGraphGeneration/syngen/configuration
configuration
configuration
# Copyright (c) 2023, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import copy import json import os import warnings from typing import Dict, Optional, Union from syngen.configuration.utils import optional_comparison, one_field_from_list_of_dicts from syngen.utils.io_utils import load_dataframe, load_graph from syngen.utils.types import MetaData, DataSourceInputType class SynGenDatasetFeatureSpec(dict): """ SynGenDatasetFeatureSpec is an util class to simply the work with SynGen Dataset Format Args: graph_metadata (Dict): dict in SynGen Format """ def __init__(self, graph_metadata: Dict): super().__init__(graph_metadata) @staticmethod def instantiate_from_preprocessed(path: str): """ Creates a SynGenDatasetFeatureSpec and checks all specified files Args: path: path to the directory with a dataset in SynGen Format """ if os.path.isfile(path): file_path = path dir_path = os.path.dirname(file_path) elif os.path.isdir(path): file_path = os.path.join(path, 'graph_metadata.json') dir_path = path else: raise ValueError(f"expected path to existing file or directory. got {path}") with open(file_path, 'r') as f: graph_metadata = json.load(f) graph_metadata[MetaData.PATH] = dir_path config = SynGenDatasetFeatureSpec(graph_metadata) config.validate() return config def get_tabular_data(self, part, name, cache=False, absolute_path=None, return_cat_feats=False): part_info = self.get_info(part, name) if MetaData.FEATURES_DATA in part_info: return part_info[MetaData.FEATURES_DATA] part_features_info = part_info[MetaData.FEATURES] part_features_path = part_info[MetaData.FEATURES_PATH] if part_features_path is None: raise ValueError() if MetaData.PATH not in self: if absolute_path is None: raise ValueError("Please specify the absolute path for the feature spec: " "by passing absolute_path argument or specifying MetaData.PATH in the Feature Spec") else: self[MetaData.PATH] = absolute_path features_df = load_dataframe(os.path.join(self[MetaData.PATH], part_features_path), feature_info=part_features_info) if cache: part_info[MetaData.FEATURES_DATA] = features_df if return_cat_feats: cat_features = [ feature_info[MetaData.NAME] for feature_info in part_info[MetaData.FEATURES] if feature_info[MetaData.FEATURE_TYPE] == MetaData.CATEGORICAL ] return features_df, cat_features return features_df def get_structural_data(self, edge_name, cache=False, absolute_path=None, ): edge_info = self.get_edge_info(edge_name) if MetaData.STRUCTURE_DATA in edge_info: return edge_info[MetaData.STRUCTURE_DATA] structure_path = edge_info[MetaData.STRUCTURE_PATH] if structure_path is None: raise ValueError() if MetaData.PATH not in self: if absolute_path is None: raise ValueError("Please specify the absolute path for the feature spec: " "by passing absolute_path argument or specifying MetaData.PATH in the Feature Spec") else: self[MetaData.PATH] = absolute_path graph = load_graph(os.path.join(self[MetaData.PATH], structure_path)) if cache: edge_info[MetaData.STRUCTURE_DATA] = graph return graph def get_edge_info(self, name: Union[str, list], src_node_type: Optional[str] = None, dst_node_type: Optional[str] = None): if isinstance(name, list): src_node_type, name, dst_node_type = name for edge_type in self[MetaData.EDGES]: if edge_type[MetaData.NAME] == name \ and optional_comparison(src_node_type, edge_type[MetaData.SRC_NODE_TYPE]) \ and optional_comparison(dst_node_type, edge_type[MetaData.DST_NODE_TYPE]): return edge_type def get_node_info(self, name: str): for node_type in self[MetaData.NODES]: if node_type[MetaData.NAME] == name: return node_type def get_info(self, part, name): if part == MetaData.NODES: return self.get_node_info(name) elif part == MetaData.EDGES: return self.get_edge_info(name) else: raise ValueError(f"unsupported FeatureSpec part expected [{MetaData.NODES}, {MetaData.EDGES}], got {part}") def validate(self): for part in [MetaData.NODES, MetaData.EDGES]: for part_info in self[part]: if part_info[MetaData.FEATURES_PATH]: tab_path = os.path.join(self[MetaData.PATH], part_info[MetaData.FEATURES_PATH]) assert os.path.exists(tab_path), f"{part}-{part_info[MetaData.NAME]}: {tab_path} does not exist" assert len(part_info[MetaData.FEATURES]) > 0, \ f"{part}-{part_info[MetaData.NAME]}: tabular features are not specified" feature_files = one_field_from_list_of_dicts( part_info[MetaData.FEATURES], MetaData.FEATURE_FILE, res_aggregator=set) if len(feature_files) > 1: assert os.path.isdir(tab_path), \ "different feature files are specified MetaData. FEATURES_PATH should be a directory" for ff in feature_files: ff_path = os.path.join(tab_path, ff) assert os.path.exists(ff_path), \ f"{part}-{part_info[MetaData.NAME]}: {ff_path} does not exist" if part == MetaData.EDGES: struct_path = os.path.join(self[MetaData.PATH], part_info[MetaData.STRUCTURE_PATH]) assert os.path.exists(struct_path), \ f"{part}-{part_info[MetaData.NAME]}: {struct_path} does not exist" def copy(self): res = {} keys_to_ignore = {MetaData.STRUCTURE_DATA, MetaData.FEATURES_DATA} for part in (MetaData.EDGES, MetaData.NODES): res[part] = [ { k: copy.deepcopy(v) for k, v in part_info.items() if k not in keys_to_ignore } for part_info in self[part] ] return SynGenDatasetFeatureSpec(res) class SynGenConfiguration(SynGenDatasetFeatureSpec): """ SynGen Configuration """ def __init__(self, configuration: Dict): super().__init__(configuration) self._fill_missing_values() self.validate() def validate(self): if MetaData.ALIGNERS in self: for aligner_info in self[MetaData.ALIGNERS]: for edge_name in aligner_info[MetaData.EDGES]: if not self.get_edge_info(edge_name)[MetaData.FEATURES_PATH].endswith(".parquet"): raise ValueError("Alignment supports only .parquet files right now") for node_name in aligner_info[MetaData.NODES]: if not self.get_node_info(node_name)[MetaData.FEATURES_PATH].endswith(".parquet"): raise ValueError("Alignment supports only .parquet files right now") def _process_tabular_generators(self, graph_part_info, part): if MetaData.TABULAR_GENERATORS not in graph_part_info: return if graph_part_info[MetaData.FEATURES] == -1: assert len(graph_part_info[MetaData.TABULAR_GENERATORS]) == 1 tab_gen_cfg = graph_part_info[MetaData.TABULAR_GENERATORS][0] assert tab_gen_cfg[MetaData.DATA_SOURCE][MetaData.TYPE] == DataSourceInputType.CONFIGURATION cfg = SynGenConfiguration.instantiate_from_preprocessed(tab_gen_cfg[MetaData.DATA_SOURCE][MetaData.PATH]) data_source_part_info = cfg.get_info(part, tab_gen_cfg[MetaData.DATA_SOURCE][MetaData.NAME]) graph_part_info[MetaData.FEATURES] = data_source_part_info[MetaData.FEATURES] for tab_gen_cfg in graph_part_info[MetaData.TABULAR_GENERATORS]: if tab_gen_cfg[MetaData.FEATURES_LIST] == -1: assert len(graph_part_info[MetaData.TABULAR_GENERATORS]) == 1, \ "you may use mimic value (-1) only if you specify a single tabular generator" tab_gen_cfg[MetaData.FEATURES_LIST] = [f[MetaData.NAME] for f in graph_part_info[MetaData.FEATURES]] if tab_gen_cfg[MetaData.DATA_SOURCE][MetaData.TYPE] == DataSourceInputType.RANDOM: edge_features = [f[MetaData.NAME] for f in graph_part_info[MetaData.FEATURES]] for feature_name in tab_gen_cfg[MetaData.FEATURES_LIST]: if feature_name not in edge_features: graph_part_info[MetaData.FEATURES].append( { MetaData.NAME: feature_name, MetaData.DTYPE: 'float32', MetaData.FEATURE_TYPE: MetaData.CONTINUOUS, # Now random generator supports only continuous features } ) def _fill_missing_values(self): for part in [MetaData.NODES, MetaData.EDGES]: for part_info in self[part]: if MetaData.FEATURES not in part_info: part_info[MetaData.FEATURES] = [] warnings.warn( f"{part}-{part_info[MetaData.NAME]}: no {MetaData.FEATURES} specified, default is []") if MetaData.FEATURES_PATH not in part_info: part_info[MetaData.FEATURES_PATH] = None warnings.warn( f"{part}-{part_info[MetaData.NAME]}: no {MetaData.FEATURES_PATH} specified, default is None") if MetaData.COUNT not in part_info: part_info[MetaData.COUNT] = -1 warnings.warn( f"{part}-{part_info[MetaData.NAME]}: no {MetaData.COUNT} specified, " f"try to mimic based on generators data") self._process_tabular_generators(part_info, part) if part == MetaData.EDGES: if MetaData.DIRECTED not in part_info: part_info[MetaData.DIRECTED] = False if part_info[MetaData.COUNT] == -1: data_source_info = part_info[MetaData.STRUCTURE_GENERATOR][MetaData.DATA_SOURCE] if data_source_info[MetaData.TYPE] == DataSourceInputType.CONFIGURATION: cfg = SynGenConfiguration.instantiate_from_preprocessed(data_source_info[MetaData.PATH]) data_source_part_info = cfg.get_info(part, data_source_info[MetaData.NAME]) elif data_source_info[MetaData.TYPE] == DataSourceInputType.RANDOM: raise ValueError('Can\'t fill the ') else: raise ValueError("unsupported structure generator datasource type") if part_info[MetaData.COUNT] == -1: part_info[MetaData.COUNT] = data_source_part_info[MetaData.COUNT] def copy(self): res = super().copy() if MetaData.ALIGNERS in self: res[MetaData.ALIGNERS] = copy.deepcopy(self[MetaData.ALIGNERS]) return SynGenConfiguration(res)
PaddlePaddle/LanguageModeling/BERT/vocab
vocab
bert-large-cased-vocab
[PAD] [unused1] [unused2] [unused3] [unused4] [unused5] [unused6] [unused7] [unused8] [unused9] [unused10] [unused11] [unused12] [unused13] [unused14] [unused15] [unused16] [unused17] [unused18] [unused19] [unused20] [unused21] [unused22] [unused23] [unused24] [unused25] [unused26] [unused27] [unused28] [unused29] [unused30] [unused31] [unused32] [unused33] [unused34] [unused35] [unused36] [unused37] [unused38] [unused39] [unused40] [unused41] [unused42] [unused43] [unused44] [unused45] [unused46] [unused47] [unused48] [unused49] [unused50] [unused51] [unused52] [unused53] [unused54] [unused55] [unused56] [unused57] [unused58] [unused59] [unused60] [unused61] [unused62] [unused63] [unused64] [unused65] [unused66] [unused67] [unused68] [unused69] [unused70] [unused71] [unused72] [unused73] [unused74] [unused75] [unused76] [unused77] [unused78] [unused79] [unused80] [unused81] [unused82] [unused83] [unused84] [unused85] [unused86] [unused87] [unused88] [unused89] [unused90] [unused91] [unused92] [unused93] [unused94] [unused95] [unused96] [unused97] [unused98] [unused99] [UNK] [CLS] [SEP] [MASK] [unused100] [unused101] ! " # $ % & ' ( ) * + , - . / 0 1 2 3 4 5 6 7 8 9 : ; < = > ? @ A B C D E F G H I J K L M N O P Q R S T U V W X Y Z [ \ ] ^ _ ` a b c d e f g h i j k l m n o p q r s t u v w x y z { | } ~ ¡ ¢ £ ¥ § ¨ © ª « ¬ ® ° ± ² ³ ´ µ ¶ · ¹ º » ¼ ½ ¾ ¿ À Á Â Ä Å Æ Ç È É Í Î Ñ Ó Ö × Ø Ú Ü Þ ß à á â ã ä å æ ç è é ê ë ì í î ï ð ñ ò ó ô õ ö ÷ ø ù ú û ü ý þ ÿ Ā ā ă ą Ć ć Č č ď Đ đ ē ė ę ě ğ ġ Ħ ħ ĩ Ī ī İ ı ļ Ľ ľ Ł ł ń ņ ň ŋ Ō ō ŏ ő Œ œ ř Ś ś Ş ş Š š Ţ ţ ť ũ ū ŭ ů ű ų ŵ ŷ ź Ż ż Ž ž Ə ƒ ơ ư ǎ ǐ ǒ ǔ ǫ Ș ș Ț ț ɐ ɑ ɔ ɕ ə ɛ ɡ ɣ ɨ ɪ ɲ ɾ ʀ ʁ ʂ ʃ ʊ ʋ ʌ ʐ ʑ ʒ ʔ ʰ ʲ ʳ ʷ ʻ ʼ ʾ ʿ ˈ ː ˡ ˢ ˣ ́ ̃ ̍ ̯ ͡ Α Β Γ Δ Ε Η Θ Ι Κ Λ Μ Ν Ο Π Σ Τ Φ Χ Ψ Ω ά έ ή ί α β γ δ ε ζ η θ ι κ λ μ ν ξ ο π ρ ς σ τ υ φ χ ψ ω ό ύ ώ І Ј А Б В Г Д Е Ж З И К Л М Н О П Р С Т У Ф Х Ц Ч Ш Э Ю Я а б в г д е ж з и й к л м н о п р с т у ф х ц ч ш щ ъ ы ь э ю я ё і ї ј њ ћ Ա Հ ա ե ի կ մ յ ն ո ս տ ր ւ ְ ִ ֵ ֶ ַ ָ ֹ ּ א ב ג ד ה ו ז ח ט י כ ל ם מ ן נ ס ע פ צ ק ר ש ת ، ء آ أ إ ئ ا ب ة ت ث ج ح خ د ذ ر ز س ش ص ض ط ظ ع غ ف ق ك ل م ن ه و ى ي َ ِ ٹ پ چ ک گ ہ ی ے ं आ क ग च ज ण त द ध न प ब भ म य र ल व श ष स ह ा ि ी ु े ो ् । ॥ আ ই এ ও ক খ গ চ ছ জ ট ত থ দ ধ ন প ব ম য র ল শ স হ ় া ি ী ু ে ো ্ য় க த ப ம ய ர ல வ ா ி ு ் ร ་ ག ང ད ན བ མ ར ལ ས ི ུ ེ ོ ა ე ი ლ ნ ო რ ს ᴬ ᴵ ᵀ ᵃ ᵇ ᵈ ᵉ ᵍ ᵏ ᵐ ᵒ ᵖ ᵗ ᵘ ᵢ ᵣ ᵤ ᵥ ᶜ ᶠ ḍ Ḥ ḥ Ḩ ḩ ḳ ṃ ṅ ṇ ṛ ṣ ṭ ạ ả ấ ầ ẩ ậ ắ ế ề ể ễ ệ ị ọ ố ồ ổ ộ ớ ờ ợ ụ ủ ứ ừ ử ữ ự ỳ ỹ ἀ ἐ ὁ ὐ ὰ ὶ ὸ ῆ ῖ ῦ ῶ ‐ ‑ ‒ – — ― ‖ ‘ ’ ‚ “ ” „ † ‡ • … ‰ ′ ″ ⁄ ⁰ ⁱ ⁴ ⁵ ⁶ ⁷ ⁸ ⁹ ⁺ ⁻ ⁿ ₀ ₁ ₂ ₃ ₄ ₅ ₆ ₇ ₈ ₉ ₊ ₍ ₎ ₐ ₑ ₒ ₓ ₕ ₖ ₘ ₙ ₚ ₛ ₜ ₤ € ₱ ₹ ℓ № ℝ ⅓ ← ↑ → ↔ ⇌ ⇒ ∂ ∈ − ∗ ∘ √ ∞ ∧ ∨ ∩ ∪ ≈ ≠ ≡ ≤ ≥ ⊂ ⊆ ⊕ ⋅ ─ │ ■ ● ★ ☆ ☉ ♠ ♣ ♥ ♦ ♭ ♯ ⟨ ⟩ ⱼ 、 。 《 》 「 」 『 』 〜 い う え お か き く け こ さ し す せ そ た ち つ て と な に の は ひ ま み む め も や ゆ よ ら り る れ ん ア ィ イ ウ エ オ カ ガ キ ク グ コ サ シ ジ ス ズ タ ダ ッ テ デ ト ド ナ ニ ハ バ パ フ ブ プ マ ミ ム ャ ュ ラ リ ル レ ロ ン ・ ー 一 三 上 下 中 事 二 井 京 人 亻 仁 佐 侍 光 公 力 北 十 南 原 口 史 司 吉 同 和 囗 国 國 土 城 士 大 天 太 夫 女 子 宀 安 宮 宿 小 尚 山 島 川 州 平 年 心 愛 戸 文 新 方 日 明 星 書 月 木 本 李 村 東 松 林 正 武 氏 水 氵 江 河 海 版 犬 王 生 田 白 皇 省 真 石 社 神 竹 美 義 花 藤 西 谷 車 辶 道 郎 郡 部 野 金 長 門 陽 青 食 馬 高 龍 龸 사 씨 의 이 한 fi fl ! ( ) , - / : the of and to in was The is for as on with that ##s his by he at from it her He had an were you be In she are but which It not or have my him one this me has also up their first out who been they She into all would its ##ing time two ##a ##e said about when over more other can after back them then ##ed there like so only ##n could ##d ##i ##y what no ##o where This made than if You ##ly through we before ##r just some ##er years do New ##t down between new now will three most On around year used such being well during They know against under later did part known off while His re ... ##l people until way American didn University your both many get United became head There second As work any But still again born even eyes After including de took And long team season family see right same called name because film don 10 found much school ##es going won place away We day left John 000 hand since World these how make number each life area man four go No here very National ##m played released never began States album home last too held several May own ##on take end School ##h ll series What want use another city When 2010 side At may That came face June think game those high March early September ##al 2011 looked July state small thought went January October ##u based August ##us world good April York us 12 2012 2008 For 2009 group along few South little ##k following November something 2013 December set 2007 old 2006 2014 located ##an music County City former ##in room ve next All ##man got father house ##g body 15 20 18 started If 2015 town our line War large population named British company member five My single ##en age State moved February 11 Her should century government built come best show However within look men door without need wasn 2016 water One system knew every died League turned asked North St wanted building received song served though felt ##ia station band ##ers local public himself different death say ##1 30 ##2 2005 16 night behind children English members near saw together son 14 voice village 13 hands help ##3 due French London top told open published third 2017 play across During put final often include 25 ##le main having 2004 once ever let book led gave late front find club ##4 German included species College form opened mother women enough West must 2000 power really 17 making half ##6 order might ##is given million times days point full service With km major ##7 original become seen II north six ##te love ##0 national International ##5 24 So District lost run couldn career always ##9 2003 ##th country ##z House air tell south worked woman player ##A almost war River ##ic married continued Then James close black short ##8 ##na using history returned light car ##ra sure William things General ##ry 2002 better support 100 among From feet King anything 21 19 established district 2001 feel great ##ton level Cup These written games others already title story ##p law thing US record role however By students England white control least inside land ##C 22 give community hard ##ie non ##c produced George round period Park business various ##ne does present wife far taken per reached David able version working young live created joined East living appeared case High done 23 important President Award France position office looking total general class To production ##S football party brother keep mind free Street hair announced development either nothing moment Church followed wrote why India San election 1999 lead How ##ch ##rs words European course considered America arms Army political ##la 28 26 west east ground further church less site First Not Australia toward California ##ness described works An Council heart past military 27 ##or heard field human soon founded 1998 playing trying ##x ##ist ##ta television mouth although taking win fire Division ##ity Party Royal program Some Don Association According tried TV Paul outside daughter Best While someone match recorded Canada closed region Air above months elected ##da ##ian road ##ar brought move 1997 leave ##um Thomas 1996 am low Robert formed person services points Mr miles ##b stop rest doing needed international release floor start sound call killed real dark research finished language Michael professional change sent 50 upon 29 track hit event 2018 term example Germany similar return ##ism fact pulled stood says ran information yet result developed girl ##re God 1995 areas signed decided ##ment Company seemed ##el co turn race common video Charles Indian ##ation blood art red ##able added rather 1994 met director addition design average minutes ##ies ##ted available bed coming friend idea kind Union Road remained ##ting everything ##ma running care finally Chinese appointed 1992 Australian ##ley popular mean teams probably ##land usually project social Championship possible word Russian instead mi herself ##T Peter Hall Center seat style money 1993 else Department table Music current 31 features special events character Two square sold debut ##v process Although Since ##ka 40 Central currently education placed lot China quickly forward seven ##ling Europe arm performed Japanese 1991 Henry Now Dr ##ion week Group myself big UK Washington ten deep 1990 Club Japan space La directed smile episode hours whole ##de ##less Why wouldn designed strong training changed Society stage involved hadn towards leading police eight kept Institute study largest child eventually private modern Court throughout getting originally attack ##E talk Great longer songs alone ##ine wide dead walked shot ##ri Oh force ##st Art today friends Island Richard 1989 center construction believe size White ship completed ##B gone Just rock sat ##R radio below entire families league includes type lived official range hold featured Most ##ter president passed means ##f forces lips Mary Do guitar ##ce food wall Of spent Its performance hear ##P Western reported sister ##et morning ##M especially ##ive Minister itself post bit groups 1988 ##tion Black ##ng Well raised sometimes Canadian Paris Spanish replaced schools Academy leaving central female Christian Jack whose college onto provided ##D ##ville players actually stopped ##son Museum doesn ##ts books fight allowed ##ur beginning Records awarded parents coach ##os Red saying ##ck Smith Yes Lake ##L aircraft 1987 ##ble previous ft action Italian African happened vocals Act future court ##ge 1986 degree phone ##ro Is countries winning breath Love river matter Lord Other list self parts ##ate provide cut shows plan 1st interest ##ized Africa stated Sir fell owned earlier ended competition attention 1985 lower nearly bad older stay Saint ##se certain 1984 fingers blue try fourth Grand ##as king ##nt makes chest movement states moving data introduced model date section Los deal ##I skin entered middle success Texas ##w summer island ##N Republic length husband 1980 ##ey reason anyone forced via base 500 job covered Festival Roman successful rights cover Man writing Ireland ##F related goal takes buildings true weeks 1983 Because opening novel ISBN meet gold ##ous mid km² standing Football Chicago shook whom ##ki 1982 Day feeling scored boy higher Force leader heavy fall question sense army Second energy meeting themselves kill ##am board census ##ya ##ns mine meant market required battle campaign attended approximately Kingdom runs active ##ha contract clear previously health 1979 Arts complete Catholic couple units ##ll ##ty Committee shoulder sea systems listed ##O caught tournament ##G northern author Film Your ##men holding offered personal 1981 southern artist traditional studio 200 capital ##ful regular ask giving organization month news Are read managed helped studied student defeated natural industry Year noted decision Government quite ##id smiled 1972 Maybe tracks ##ke Mark al media engine hour Their relationship plays property structure 1976 ago Hill Martin 1978 ready Many Like Bay immediately generally Italy Greek practice caused division significant Joseph speed Let thinking completely 1974 primary mostly ##field ##K 1975 ##to Even writer ##led dropped magazine collection understand route highest particular films lines network Science loss carried direction green 1977 location producer according Women Queen neck thus independent view 1970 Angeles Soviet distance problem Board tour western income appearance access Mexico nodded street surface arrived believed Old 1968 1973 becoming whether 1945 figure singer stand Following issue window wrong pain everyone lives issues park slowly la act ##va bring Lee operations key comes fine cold famous Navy 1971 Me additional individual ##ner Zealand goals county contains Service minute 2nd reach talking particularly ##ham movie Director glass paper studies ##co railway standard Education 45 represented Chief Louis launched Star terms 60 1969 experience watched Another Press Tom staff starting subject break Virginia nine eye ##age evidence foot ##est companies Prince ##V gun create Big People guy Green simply numerous ##line increased twenty ##ga ##do 1967 award officer stone Before material Northern grew male plant Life legs step Al unit 35 except answer ##U report response Edward commercial edition trade science ##ca Irish Law shown rate failed ##ni remains changes mm limited larger Later cause waiting Time ##wood cost Bill manager activities likely allow operated retired ##ping 65 directly Who associated effect hell Florida straight hot Valley management girls expected eastern Mike chance cast centre chair hurt problems ##li walk programs Team characters Battle edge pay maybe corner majority medical Joe Summer ##io attempt Pacific command Radio ##by names municipality 1964 train economic Brown feature sex source agreed remember Three 1966 1965 Pennsylvania victory senior annual III Southern results Sam serving religious Jones appears ##der despite claimed Both musical matches fast security selected Young double complex hospital chief Times ##ve Championships filled Public Despite beautiful Research plans Province ##ally Wales ##ko artists metal nearby Spain ##il 32 houses supported piece ##no stared recording nature legal Russia ##ization remaining looks ##sh bridge closer cases scene marriage Little ##é uses Earth specific Frank theory Good discovered referred bass culture university presented Congress ##go metres continue 1960 isn Awards meaning cell composed separate Series forms Blue cross ##tor increase test computer slightly Where Jewish Town tree status 1944 variety responsible pretty initially ##way realized pass provides Captain Alexander recent score broke Scott drive financial showed Line stories ordered soldiers genus operation gaze sitting society Only hope actor follow Empire Yeah technology happy focus policy spread situation ##ford ##ba Mrs watch Can 1963 Commission touch earned troops Under 1962 individuals cannot 19th ##lin mile expression exactly suddenly weight dance stepped places appear difficult Railway anti numbers kilometres star ##ier department ice Britain removed Once ##lo Boston value ##ant mission trees Order sports join serve Major poor Poland mainly Theatre pushed Station ##it Lady federal silver ##ler foreign ##ard Eastern ##den box hall subsequently lies acquired 1942 ancient CD History Jean beyond ##ger El ##les growing championship native Parliament Williams watching direct overall offer Also 80 Secretary spoke Latin ability ##ated safe presence ##ial headed regional planned 1961 Johnson throat consists ##W extended Or bar walls Chris stations politician Olympics influence share fighting speak hundred Carolina die stars ##tic color Chapter ##ish fear sleep goes Francisco oil Bank sign physical ##berg Dutch seasons ##rd Games Governor sorry lack Centre memory baby smaller charge Did multiple ships shirt Assembly amount leaves 3rd Foundation conditions 1943 Rock Democratic Daniel ##at winner products ##ina store latter Professor civil prior host 1956 soft vote needs Each rules 1958 pressure letter normal proposed levels records 1959 paid intended Victoria purpose okay historical issued 1980s broadcast rule simple picked firm Sea 1941 Elizabeth 1940 serious featuring highly graduated mentioned choice 1948 replied percent Scotland ##hi females constructed 1957 settled Steve recognized cities crew glanced kiss competed flight knowledge editor More Conference ##H fifth elements ##ee ##tes function newspaper recently Miss cultural brown twice Office 1939 truth Creek 1946 households USA 1950 quality ##tt border seconds destroyed pre wait ahead build image 90 cars ##mi 33 promoted professor et bank medal text broken Middle revealed sides wing seems channel 1970s Ben loved effort officers Will ##ff 70 Israel Jim upper fully label Jr assistant powerful pair positive ##ary gives 1955 20th races remain kitchen primarily ##ti Sydney easy Tour whispered buried 300 News Polish 1952 Duke Columbia produce accepted 00 approach minor 1947 Special 44 Asian basis visit Fort Civil finish formerly beside leaned ##ite median rose coast effects supposed Cross ##hip Corps residents Jackson ##ir Bob basketball 36 Asia seem Bishop Book ##ber ring ##ze owner BBC ##ja transferred acting De appearances walking Le press grabbed 1954 officially 1953 ##pe risk taught review ##X lay ##well council Avenue seeing losing Ohio Super province ones travel ##sa projects equipment spot Berlin administrative heat potential shut capacity elections growth fought Republican mixed Andrew teacher turning strength shoulders beat wind 1949 Health follows camp suggested perhaps Alex mountain contact divided candidate fellow 34 Show necessary workers ball horse ways questions protect gas activity younger bottom founder Scottish screen treatment easily com ##house dedicated Master warm Night Georgia Long von ##me perfect website 1960s piano efforts ##ide Tony sort offers Development Simon executive ##nd save Over Senate 1951 1990s draw master Police ##ius renamed boys initial prominent damage Co ##ov ##za online begin occurred captured youth Top account tells Justice conducted forest ##town bought teeth Jersey ##di purchased agreement Michigan ##ure campus prison becomes product secret guess Route huge types drums 64 split defeat estate housing ##ot brothers Coast declared happen titled therefore sun commonly alongside Stadium library Home article steps telling slow assigned refused laughed wants Nick wearing Rome Open ##ah Hospital pointed Taylor lifted escape participated ##j drama parish Santa ##per organized mass pick Airport gets Library unable pull Live ##ging surrounding ##ries focused Adam facilities ##ning ##ny 38 ##ring notable era connected gained operating laid Regiment branch defined Christmas machine Four academic Iran adopted concept Men compared search traffic Max Maria greater ##ding widely ##burg serves 1938 37 Go hotel shared typically scale 1936 leg suffered yards pieces Ministry Wilson episodes empty 1918 safety continues yellow historic settlement 400 Come Corporation enemy content picture evening territory method trial solo driver Here ##ls entrance Prize spring whatever ##ent 75 ##ji reading Arthur ##cy Our clothes Prime Illinois Kong code ##ria sit Harry Federal chosen administration bodies begins stomach Though seats Hong density Sun leaders Field museum chart platform languages ##ron birth holds Gold ##un fish combined ##ps 4th 1937 largely captain trust Game van boat Oxford basic beneath Islands painting nice Toronto path males sources block conference parties murder clubs crowd calling About Business peace knows lake speaking stayed Brazil allowing Born unique thick Technology ##que receive des semi alive noticed format ##ped coffee digital ##ned handed guard tall faced setting plants partner claim reduced temple animals determined classes ##out estimated ##ad Olympic providing Massachusetts learned Inc Philadelphia Social carry 42 possibly hosted tonight respectively Today shape Mount roles designated brain etc Korea thoughts Brian Highway doors background drew models footballer tone turns 1935 quiet tower wood bus write software weapons flat marked 1920 newly tight Eric finger Journal FC Van rise critical Atlantic granted returning communities humans quick 39 48 ranked sight pop Swedish Stephen card analysis attacked ##wa Sunday identified Jason champion situated 1930 expanded tears ##nce reaching Davis protection Emperor positions nominated Bridge tax dress allows avoid leadership killing actress guest steel knowing electric cells disease grade unknown ##ium resulted Pakistan confirmed ##ged tongue covers ##Y roof entirely applied votes drink interview exchange Township reasons ##ised page calls dog agent nose teaching ##ds ##ists advanced wish Golden existing vehicle del 1919 develop attacks pressed Sports planning resulting facility Sarah notes 1933 Class Historic winter ##mo audience Community household Netherlands creation ##ize keeping 1914 claims dry guys opposite ##ak explained Ontario secondary difference Francis actions organizations yard animal Up Lewis titles Several 1934 Ryan 55 Supreme rolled 1917 distribution figures afraid rural yourself ##rt sets barely Instead passing awards 41 silence authority occupied environment windows engineering surprised flying crime reports Mountain powers driving succeeded reviews 1929 Head missing Song Jesus opportunity inspired ends albums conversation impact injury surprise billion learning heavily oldest union creating ##ky festival literature letters sexual ##tte apartment Final comedy nation orders ##sen contemporary Power drawn existence connection ##ating Post Junior remembered message Medal castle note engineer sounds Beach crossed ##dy ear scientific sales ##ai theme starts clearly ##ut trouble ##gan bag ##han BC sons 1928 silent versions daily Studies ending Rose guns 1932 headquarters reference obtained Squadron concert none du Among ##don prevent Member answered staring Between ##lla portion drug liked association performances Nations formation Castle lose learn scoring relatively quarter 47 Premier ##ors Sweden baseball attempted trip worth perform airport fields enter honor Medical rear commander officials condition supply materials 52 Anna volume threw Persian 43 interested Gallery achieved visited laws relief Area Matt singles Lieutenant Country fans Cambridge sky Miller effective tradition Port ##ana minister extra entitled System sites authorities acres committee racing 1931 desk trains ass weren Family farm ##ance industrial ##head iron 49 abandoned Out Holy chairman waited frequently display Light transport starring Patrick Engineering eat FM judge reaction centuries price ##tive Korean defense Get arrested 1927 send urban ##ss pilot Okay Media reality arts soul thirty ##be catch generation ##nes apart Anne drop See ##ving sixth trained Management magic cm height Fox Ian resources vampire principal Was haven ##au Walter Albert rich 1922 causing entry ##ell shortly 46 worry doctor composer rank Network bright showing regions 1924 wave carrying kissed finding missed Earl lying target vehicles Military controlled dinner ##board briefly lyrics motion duty strange attempts invited kg villages 5th Land ##mer Christ prepared twelve check thousand earth copies en transfer citizens Americans politics nor theatre Project ##bo clean rooms laugh ##ran application contained anyway containing Sciences 1925 rare speech exist 1950s falling passenger ##im stands 51 ##ol ##ow phase governor kids details methods Vice employed performing counter Jane heads Channel wine opposition aged 1912 Every 1926 highway ##ura 1921 aired 978 permanent Forest finds joint approved ##pur brief doubt acts brand wild closely Ford Kevin chose shall port sweet fun asking Be ##bury sought Dave Mexican mom Right Howard Moscow Charlie Stone ##mann admitted ##ver wooden 1923 Officer relations Hot combat publication chain shop inhabitants proved ideas address 1915 Memorial explain increasing conflict Anthony Melbourne narrow temperature slid 1916 worse selling documentary Ali Ray opposed vision dad extensive Infantry commissioned Doctor offices programming core respect storm ##pa ##ay ##om promotion der struck anymore shit Region receiving DVD alternative ##ue ride maximum 1910 ##ious Third Affairs cancer Executive ##op dream 18th Due ##ker ##worth economy IV Billboard identity subsequent statement skills ##back funding ##ons Round Foreign truck Please lights wondered ##ms frame yes Still districts fiction Colonel converted 150 grown accident critics fit Information architecture Point Five armed Billy poet functions consisted suit Turkish Band object desire ##ities sounded flow Norwegian articles Marie pulling thin singing Hunter Human Battalion Federation Kim origin represent dangerous weather fuel ex ##sing Last bedroom aid knees Alan angry assumed plane Something founding concerned global Fire di please Portuguese touched Roger nuclear Register Jeff fixed royal lie finals NFL Manchester towns handle shaped Chairman Dean launch understanding Children violence failure sector Brigade wrapped fired sharp tiny developing expansion Free institutions technical Nothing otherwise Main inch Saturday wore Senior attached cheek representing Kansas ##chi ##kin actual advantage Dan Austria ##dale hoped multi squad Norway streets 1913 Services hired grow pp wear painted Minnesota stuff Building 54 Philippines 1900 ##ties educational Khan Magazine ##port Cape signal Gordon sword Anderson cool engaged Commander images Upon tied Security cup rail Vietnam successfully ##red Muslim gain bringing Native hers occurs negative Philip Kelly Colorado category ##lan 600 Have supporting wet 56 stairs Grace observed ##ung funds restaurant 1911 Jews ##ments ##che Jake Back 53 asks journalist accept bands bronze helping ##ice decades mayor survived usual influenced Douglas Hey ##izing surrounded retirement Temple derived Pope registered producing ##ral structures Johnny contributed finishing buy specifically ##king patients Jordan internal regarding Samuel Clark ##q afternoon Finally scenes notice refers quietly threat Water Those Hamilton promise freedom Turkey breaking maintained device lap ultimately Champion Tim Bureau expressed investigation extremely capable qualified recognition items ##up Indiana adult rain greatest architect Morgan dressed equal Antonio collected drove occur Grant graduate anger Sri worried standards ##ore injured somewhere damn Singapore Jimmy pocket homes stock religion aware regarded Wisconsin ##tra passes fresh ##ea argued Ltd EP Diego importance Census incident Egypt Missouri domestic leads ceremony Early camera Father challenge Switzerland lands familiar hearing spend educated Tennessee Thank ##ram Thus concern putting inches map classical Allen crazy valley Space softly ##my pool worldwide climate experienced neighborhood scheduled neither fleet 1908 Girl ##J Part engines locations darkness Revolution establishment lawyer objects apparently Queensland Entertainment bill mark Television ##ong pale demand Hotel selection ##rn ##ino Labour Liberal burned Mom merged Arizona request ##lia ##light hole employees ##ical incorporated 95 independence Walker covering joining ##ica task papers backing sell biggest 6th strike establish ##ō gently 59 Orchestra Winter protein Juan locked dates Boy aren shooting Luke solid charged Prior resigned interior garden spoken improve wonder promote hidden ##med combination Hollywood Swiss consider ##ks Lincoln literary drawing Marine weapon Victor Trust Maryland properties ##ara exhibition understood hung Tell installed loud fashion affected junior landing flowers ##he Internet beach Heart tries Mayor programme 800 wins noise ##ster ##ory 58 contain fair delivered ##ul wedding Square advance behavior Program Oregon ##rk residence realize certainly hill Houston 57 indicated ##water wounded Village massive Moore thousands personnel dating opera poetry ##her causes feelings Frederick applications push approached foundation pleasure sale fly gotten northeast costs raise paintings ##ney views horses formal Arab hockey typical representative rising ##des clock stadium shifted Dad peak Fame vice disappeared users Way Naval prize hoping values evil Bell consisting ##ón Regional ##ics improved circle carefully broad ##ini Fine maintain operate offering mention Death stupid Through Princess attend interests ruled somewhat wings roads grounds ##ual Greece Champions facing hide voted require Dark Matthew credit sighed separated manner ##ile Boys 1905 committed impossible lip candidates 7th Bruce arranged Islamic courses criminal ##ened smell ##bed 08 consecutive ##ening proper purchase weak Prix 1906 aside introduction Look ##ku changing budget resistance factory Forces agency ##tone northwest user 1907 stating ##one sport Design environmental cards concluded Carl 250 accused ##ology Girls sick intelligence Margaret responsibility Guard ##tus 17th sq goods 1909 hate ##ek capture stores Gray comic Modern Silver Andy electronic wheel ##ied Deputy ##bs Czech zone choose constant reserve ##lle Tokyo spirit sub degrees flew pattern compete Dance ##ik secretary Imperial 99 reduce Hungarian confused ##rin Pierre describes regularly Rachel 85 landed passengers ##ise ##sis historian meters Youth ##ud participate ##cing arrival tired Mother ##gy jumped Kentucky faces feed Israeli Ocean ##Q ##án plus snow techniques plate sections falls jazz ##ris tank loan repeated opinion ##res unless rugby journal Lawrence moments shock distributed ##ded adjacent Argentina crossing uncle ##ric Detroit communication mental tomorrow session Emma Without ##gen Miami charges Administration hits coat protected Cole invasion priest 09 Gary enjoyed plot measure bound friendly throw musician ##lon ##ins Age knife damaged birds driven lit ears breathing Arabic Jan faster Jonathan ##gate Independent starred Harris teachers Alice sequence mph file translated decide determine Review documents sudden threatened ##ft bear distinct decade burning ##sky 1930s replace begun extension ##time 1904 equivalent accompanied Christopher Danish ##ye Besides ##more persons fallen Rural roughly saved willing ensure Belgium 05 musicians ##ang giant Six Retrieved worst purposes ##bly mountains seventh slipped brick 07 ##py somehow Carter Iraq cousin favor islands journey FIFA contrast planet vs calm ##ings concrete branches gray profit Russell ##ae ##ux ##ens philosophy businesses talked parking ##ming owners Place ##tle agricultural Kate 06 southeast draft Eddie earliest forget Dallas Commonwealth edited 66 inner ed operates 16th Harvard assistance ##si designs Take bathroom indicate CEO Command Louisiana 1902 Dublin Books 1901 tropical 1903 ##tors Places tie progress forming solution 62 letting ##ery studying ##jo duties Baseball taste Reserve ##ru Ann ##gh visible ##vi notably link NCAA southwest Never storage mobile writers favorite Pro pages truly count ##tta string kid 98 Ross row ##idae Kennedy ##tan Hockey hip waist grandfather listen ##ho feels busy 72 stream obvious cycle shaking Knight ##ren Carlos painter trail web linked 04 Palace existed ##ira responded closing End examples Marshall weekend jaw Denmark lady township medium chin Story option fifteen Moon represents makeup investment jump childhood Oklahoma roll normally Ten Operation Graham Seattle Atlanta paused promised rejected treated returns flag ##ita Hungary danger glad movements visual subjects credited soldier Norman ill translation José Quebec medicine warning theater praised municipal 01 commune churches acid folk 8th testing add survive Sound devices residential severe presidential Mississippi Austin Perhaps Charlotte hanging Montreal grin ##ten racial partnership shoot shift ##nie Les downtown Brothers Garden matters restored mirror forever winners rapidly poverty ##ible Until DC faith hundreds Real Ukraine Nelson balance Adams contest relative ethnic Edinburgh composition ##nts emergency ##van marine reputation Down pack 12th Communist Mountains pro stages measures ##ld ABC Li victims benefit Iowa Broadway gathered rating Defense classic ##ily ceiling ##ions snapped Everything constituency Franklin Thompson Stewart entering Judge forth ##sk wanting smiling moves tunnel premiered grass unusual Ukrainian bird Friday tail Portugal coal element Fred guards Senator collaboration beauty Wood chemical beer justice signs ##Z sees ##zi Puerto ##zed 96 smooth Bowl gift limit 97 heading Source wake requires Ed Constitution factor Lane factors adding Note cleared pictures pink ##ola Kent Local Singh moth Ty ##ture courts Seven temporary involving Vienna emerged fishing agree defensive stuck secure Tamil ##ick bottle 03 Player instruments Spring patient flesh contributions cry Malaysia 120 Global da Alabama Within ##work debuted expect Cleveland concerns retained horror 10th spending Peace Transport grand Crown instance institution acted Hills mounted Campbell shouldn 1898 ##ably chamber soil 88 Ethan sand cheeks ##gi marry 61 weekly classification DNA Elementary Roy definitely Soon Rights gate suggests aspects imagine golden beating Studios Warren differences significantly glance occasionally ##od clothing Assistant depth sending possibility mode prisoners requirements daughters dated Representatives prove guilty interesting smoke cricket 93 ##ates rescue Connecticut underground Opera 13th reign ##ski thanks leather equipped routes fan ##ans script Wright bishop Welsh jobs faculty eleven Railroad appearing anniversary Upper ##down anywhere Rugby Metropolitan Meanwhile Nicholas champions forehead mining drinking 76 Jerry membership Brazilian Wild Rio scheme Unlike strongly ##bility fill ##rian easier MP Hell ##sha Stanley banks Baron ##ique Robinson 67 Gabriel Austrian Wayne exposed ##wan Alfred 1899 manage mix visitors eating ##rate Sean commission Cemetery policies Camp parallel traveled guitarist 02 supplies couples poem blocks Rick Training Energy achieve appointment Wing Jamie 63 novels ##em 1890 songwriter Base Jay ##gar naval scared miss labor technique crisis Additionally backed destroy seriously tools tennis 91 god ##ington continuing steam obviously Bobby adapted fifty enjoy Jacob publishing column ##ular Baltimore Donald Liverpool 92 drugs movies ##ock Heritage ##je ##istic vocal strategy gene advice ##bi Ottoman riding ##side Agency Indonesia 11th laughing sleeping und muttered listening deck tip 77 ownership grey Claire deeply provincial popularity Cooper ##á Emily ##sed designer Murray describe Danny Around Parker ##dae 68 rates suffering considerable 78 nervous powered tons circumstances wished belonged Pittsburgh flows 9th ##use belt 81 useful 15th context List Dead Iron seek Season worn frequency legislation replacement memories Tournament Again Barry organisation copy Gulf waters meets struggle Oliver 1895 Susan protest kick Alliance components 1896 Tower Windows demanded regiment sentence Woman Logan Referee hosts debate knee Blood ##oo universities practices Ward ranking correct happening Vincent attracted classified ##stic processes immediate waste increasingly Helen ##po Lucas Phil organ 1897 tea suicide actors lb crash approval waves ##ered hated grip 700 amongst 69 74 hunting dying lasted illegal ##rum stare defeating ##gs shrugged °C Jon Count Orleans 94 affairs formally ##and ##ves criticized Disney Vol successor tests scholars palace Would celebrated rounds grant Schools Such commanded demon Romania ##all Karl 71 ##yn 84 Daily totally Medicine fruit Die upset Lower Conservative 14th Mitchell escaped shoes Morris ##tz queen harder prime Thanks indeed Sky authors rocks definition Nazi accounts printed experiences ##ters divisions Cathedral denied depending Express ##let 73 appeal loose colors filed ##isation gender ##ew throne forests Finland domain boats Baker squadron shore remove ##ification careful wound railroad 82 seeking agents ##ved Blues ##off customers ignored net ##ction hiding Originally declined ##ess franchise eliminated NBA merely pure appropriate visiting forty markets offensive coverage cave ##nia spell ##lar Benjamin ##ire Convention filmed Trade ##sy ##ct Having palm 1889 Evans intense plastic Julia document jeans vessel SR ##fully proposal Birmingham le ##ative assembly 89 fund lock 1893 AD meetings occupation modified Years odd aimed reform Mission Works shake cat exception convinced executed pushing dollars replacing soccer manufacturing ##ros expensive kicked minimum Josh coastal Chase ha Thailand publications deputy Sometimes Angel effectively ##illa criticism conduct Serbian landscape NY absence passage ##ula Blake Indians 1892 admit Trophy ##ball Next ##rated ##ians charts kW orchestra 79 heritage 1894 rough exists boundary Bible Legislative moon medieval ##over cutting print ##ett birthday ##hood destruction Julian injuries influential sisters raising statue colour dancing characteristics orange ##ok ##aries Ken colonial twin Larry surviving ##shi Barbara personality entertainment assault ##ering talent happens license 86 couch Century soundtrack shower swimming cash Staff bent 1885 bay lunch ##lus dozen vessels CBS greatly critic Test symbol panel shell output reaches 87 Front motor ocean ##era ##ala maintenance violent scent Limited Las Hope Theater Which survey Robin recordings compilation ##ward bomb insurance Authority sponsored satellite Jazz refer stronger blow whilst Wrestling suggest ##rie climbed ##els voices shopping 1891 Neil discovery ##vo ##ations burst Baby peaked Brooklyn knocked lift ##try false nations Hugh Catherine preserved distinguished terminal resolution ratio pants cited competitions completion DJ bone uniform schedule shouted 83 1920s rarely Basketball Taiwan artistic bare vampires arrest Utah Marcus assist gradually qualifying Victorian vast rival Warner Terry Economic ##cia losses boss versus audio runner apply surgery Play twisted comfortable ##cs Everyone guests ##lt Harrison UEFA lowered occasions ##lly ##cher chapter youngest eighth Culture ##room ##stone 1888 Songs Seth Digital involvement expedition relationships signing 1000 fault annually circuit afterwards meat creature ##ou cable Bush ##net Hispanic rapid gonna figured extent considering cried ##tin sigh dynasty ##ration cabinet Richmond stable ##zo 1864 Admiral Unit occasion shares badly longest ##ify Connor extreme wondering girlfriend Studio ##tions 1865 tribe exact muscles hat Luis Orthodox decisions amateur description ##lis hips kingdom ##ute Portland whereas Bachelor outer discussion partly Arkansas 1880 dreams perfectly Lloyd ##bridge asleep ##tti Greg permission trading pitch mill Stage liquid Keith ##tal wolf processing stick Jerusalem profile rushed spiritual argument Ice Guy till Delhi roots Section missions Glasgow penalty NBC encouraged identify keyboards ##zing ##ston disc plain informed Bernard thinks fled Justin ##day newspapers ##wick Ralph ##zer unlike Stars artillery ##ified recovered arrangement searching ##pers ##tory ##rus deaths Egyptian diameter ##í marketing corporate teach marks Turner staying hallway Sebastian chapel naked mistake possession 1887 dominated jacket creative Fellow Falls Defence suspended employment ##rry Hebrew Hudson Week Wars recognize Natural controversial Tommy thank Athletic benefits decline intention ##ets Lost Wall participation elevation supports parliament 1861 concentration Movement ##IS competing stops behalf ##mm limits funded discuss Collins departure obtain woods latest universe alcohol Laura rush blade funny Dennis forgotten Amy Symphony apparent graduating 1862 Rob Grey collections Mason emotions ##ugh literally Any counties 1863 nomination fighter habitat respond external Capital exit Video carbon sharing Bad opportunities Perry photo ##mus Orange posted remainder transportation portrayed Labor recommended percussion rated Grade rivers partially suspected strip adults button struggled intersection Canal ##ability poems claiming Madrid 1886 Together ##our Much Vancouver instrument instrumental 1870 mad angle Control Phoenix Leo Communications mail ##ette ##ev preferred adaptation alleged discussed deeper ##ane Yet Monday volumes thrown Zane ##logy displayed rolling dogs Along Todd ##ivity withdrew representation belief ##sia crown Late Short hardly grinned romantic Pete ##ken networks enemies Colin Eventually Side donated ##su steady grab guide Finnish Milan pregnant controversy reminded 1884 Stuart ##bach ##ade Race Belgian LP Production Zone lieutenant infantry Child confusion sang resident ##ez victim 1881 channels Ron businessman ##gle Dick colony pace producers ##ese agencies Craig Lucy Very centers Yorkshire photography ##ched Album championships Metro substantial Standard terrible directors contribution advertising emotional ##its layer segment sir folded Roberts ceased Hampshire ##ray detailed partners m² ##pt Beth genre commented generated remote aim Hans credits concerts periods breakfast gay shadow defence Too Had transition Afghanistan ##book eggs defend ##lli writes Systems bones mess seed scientists Shortly Romanian ##zy Freedom muscle hero parent agriculture checked Islam Bristol Freyja Arena cabin Germans electricity ranks viewed medals Wolf associate Madison Sorry fort Chile detail widespread attorney boyfriend ##nan Students Spencer ##ig bite Maine demolished Lisa erected Someone operational Commissioner NHL Coach Bar forcing Dream Rico cargo Murphy ##fish ##ase distant ##master ##ora Organization doorway Steven traded electrical frequent ##wn Branch Sure 1882 placing Manhattan attending attributed excellent pounds ruling principles component Mediterranean Vegas machines percentage infrastructure throwing affiliated Kings secured Caribbean Track Ted honour opponent Virgin Construction grave produces Challenge stretched paying murmured ##ata integrated waved Nathan ##ator transmission videos ##yan ##hu Nova descent AM Harold conservative Therefore venue competitive ##ui conclusion funeral confidence releases scholar ##sson Treaty stress mood ##sm Mac residing Action Fund ##ship animated fitted ##kar defending voting tend ##berry answers believes ##ci helps Aaron ##tis themes ##lay populations Players stroke Trinity electoral paint abroad charity keys Fair ##pes interrupted participants murdered Days supporters ##ab expert borders mate ##llo solar architectural tension ##bling Parish tape operator Cultural Clinton indicates publisher ordinary sugar arrive rifle acoustic ##uring assets ##shire SS sufficient options HMS Classic bars rebuilt governments Beijing reporter screamed Abbey crying mechanical instantly communications Political cemetery Cameron Stop representatives USS texts mathematics innings civilian Serbia ##hill practical patterns dust Faculty debt ##end ##cus junction suppose experimental Computer Food wrist abuse dealing bigger cap principle ##pin Muhammad Fleet Collection attempting dismissed ##burn regime Herbert ##ua shadows 1883 Eve Lanka 1878 Performance fictional ##lock Noah Run Voivodeship exercise broadcasting ##fer RAF Magic Bangladesh suitable ##low ##del styles toured Code identical links insisted 110 flash Model slave Derek Rev fairly Greater sole ##lands connecting zero bench ##ome switched Fall Owen yours Electric shocked convention ##bra climb memorial swept Racing decides belong ##nk parliamentary ##und ages proof ##dan delivery 1860 ##ów sad publicly leaning Archbishop dirt ##ose categories 1876 burn ##bing requested Guinea Historical rhythm relation ##heim ye pursue merchant ##mes lists continuous frowned colored tool gods involves Duncan photographs Cricket slight Gregory atmosphere wider Cook ##tar essential Being FA emperor wealthy nights ##bar licensed Hawaii viewers Language load nearest milk kilometers platforms ##ys territories Rogers sheet Rangers contested ##lation isolated assisted swallowed Small Contemporary Technical Edwards express Volume endemic ##ei tightly Whatever indigenous Colombia ##ulation hp characterized ##ida Nigeria Professional duo Soccer slaves Farm smart Attorney Attendance Common salt ##vin tribes nod sentenced bid sample Drive switch instant 21st Cuba drunk Alaska proud awareness hitting sessions Thai locally elsewhere Dragon gentle touching ##lee Springs Universal Latino spin 1871 Chart recalled Type pointing ##ii lowest ##ser grandmother Adelaide Jacques spotted Buffalo restoration Son Joan farmers Lily 1879 lucky ##dal luck eldest ##rant Market drummer deployed warned prince sing amazing sailed ##oon 1875 Primary traveling Masters Sara cattle Trail gang Further desert relocated ##tch ##ord Flight illness Munich ninth repair Singles ##lated Tyler tossed boots Work sized earning shoved magazines housed dam researchers Former spun premiere spaces organised wealth crimes devoted stones Urban automatic hop affect outstanding tanks mechanism Muslims Ms shots argue Jeremy connections Armenian increases rubbed 1867 retail gear Pan bonus jurisdiction weird concerning whisper ##gal Microsoft tenure hills www Gmina porch files reportedly venture Storm ##ence Nature killer panic fate Secret Wang scream drivers belongs Chamber clan monument mixing Peru bet Riley Friends Isaac submarine 1877 130 judges harm ranging affair prepare pupils householder Policy decorated Nation slammed activist implemented Room qualify Publishing establishing Baptist touring subsidiary ##nal legend 1872 laughter PC Athens settlers ties dual dear Draft strategic Ivan reveal closest dominant Ah ##ult Denver bond boundaries drafted tables ##TV eyed Edition ##ena 1868 belonging 1874 Industrial cream Ridge Hindu scholarship Ma opens initiated ##ith yelled compound random Throughout grades physics sank grows exclusively settle Saints brings Amsterdam Make Hart walks battery violin ##born explanation ##ware 1873 ##har provinces thrust exclusive sculpture shops ##fire VI constitution Barcelona monster Devon Jefferson Sullivan bow ##din desperate ##ć Julie ##mon ##ising terminus Jesse abilities golf ##ple ##via ##away Raymond measured jury firing revenue suburb Bulgarian 1866 ##cha timber Things ##weight Morning spots Alberta Data explains Kyle friendship raw tube demonstrated aboard immigrants reply breathe Manager ease ##ban ##dia Diocese ##vy ##ía pit ongoing ##lie Gilbert Costa 1940s Report voters cloud traditions ##MS gallery Jennifer swung Broadcasting Does diverse reveals arriving initiative ##ani Give Allied Pat Outstanding monastery blind Currently ##war bloody stopping focuses managing Florence Harvey creatures 900 breast internet Artillery purple ##mate alliance excited fee Brisbane lifetime Private ##aw ##nis ##gue ##ika phrase regulations reflected manufactured conventional pleased client ##ix ##ncy Pedro reduction ##con welcome jail comfort Iranian Norfolk Dakota ##tein evolution everywhere Initially sensitive Olivia Oscar implementation sits stolen demands slide grandson ##ich merger ##mic Spirit ##° ticket root difficulty Nevada ##als lined Dylan Original Call biological EU dramatic ##hn Operations treaty gap ##list Am Romanized moral Butler perspective Furthermore Manuel absolutely unsuccessful disaster dispute preparation tested discover ##ach shield squeezed brushed battalion Arnold ##ras superior treat clinical ##so Apple Syria Cincinnati package flights editions Leader minority wonderful hang Pop Philippine telephone bell honorary ##mar balls Democrat dirty thereafter collapsed Inside slip wrestling ##ín listened regard bowl None Sport completing trapped ##view copper Wallace Honor blame Peninsula ##ert ##oy Anglo bearing simultaneously honest ##ias Mix Got speaker voiced impressed prices error 1869 ##feld trials Nine Industry substitute Municipal departed slept ##ama Junction Socialist flower dropping comment fantasy ##ress arrangements travelled furniture fist relieved ##tics Leonard linear earn expand Soul Plan Leeds Sierra accessible innocent Winner Fighter Range winds vertical Pictures 101 charter cooperation prisoner interviews recognised sung manufacturer exposure submitted Mars leaf gauge screaming likes eligible ##ac gathering columns ##dra belly UN maps messages speakers ##ants garage unincorporated Number Watson sixteen lots beaten Could Municipality ##ano Horse talks Drake scores Venice genetic ##mal ##ère Cold Jose nurse traditionally ##bus Territory Key Nancy ##win thumb São index dependent carries controls Comics coalition physician referring Ruth Based restricted inherited internationally stretch THE plates margin Holland knock significance valuable Kenya carved emotion conservation municipalities overseas resumed Finance graduation blinked temperatures constantly productions scientist ghost cuts permitted ##ches firmly ##bert patrol ##yo Croatian attacking 1850 portrait promoting sink conversion ##kov locomotives Guide ##val nephew relevant Marc drum originated Chair visits dragged Price favour corridor properly respective Caroline reporting inaugural 1848 industries ##ching edges Christianity Maurice Trent Economics carrier Reed ##gon tribute Pradesh ##ale extend attitude Yale ##lu settlements glasses taxes targets ##ids quarters ##ological connect hence metre collapse underneath banned Future clients alternate explosion kinds Commons hungry dragon Chapel Buddhist lover depression pulls ##ges ##uk origins computers crosses kissing assume emphasis lighting ##ites personally crashed beam touchdown lane comparison ##mont Hitler ##las execution ##ene acre sum Pearl ray ##point essentially worker convicted tear Clay recovery Literature Unfortunately ##row partial Petersburg Bulgaria coaching evolved reception enters narrowed elevator therapy defended pairs ##lam breaks Bennett Uncle cylinder ##ison passion bases Actor cancelled battles extensively oxygen Ancient specialized negotiations ##rat acquisition convince interpretation ##00 photos aspect colleges Artist keeps ##wing Croatia ##ona Hughes Otto comments ##du Ph Sweet adventure describing Student Shakespeare scattered objective Aviation Phillips Fourth athletes ##hal ##tered Guitar intensity née dining curve Obama topics legislative Mill Cruz ##ars Members recipient Derby inspiration corresponding fed YouTube coins pressing intent Karen cinema Delta destination shorter Christians imagined canal Newcastle Shah Adrian super Males 160 liberal lord bat supplied Claude meal worship ##atic Han wire °F ##tha punishment thirteen fighters ##ibility 1859 Ball gardens ##ari Ottawa pole indicating Twenty Higher Bass Ivy farming ##urs certified Saudi plenty ##ces restaurants Representative Miles payment ##inger ##rit Confederate festivals references ##ić Mario PhD playoffs witness rice mask saving opponents enforcement automatically relegated ##oe radar whenever Financial imperial uncredited influences Abraham skull Guardian Haven Bengal impressive input mixture Warsaw altitude distinction 1857 collective Annie ##ean ##bal directions Flying ##nic faded ##ella contributing ##ó employee ##lum ##yl ruler oriented conductor focusing ##die Giants Mills mines Deep curled Jessica guitars Louise procedure Machine failing attendance Nepal Brad Liam tourist exhibited Sophie depicted Shaw Chuck ##can expecting challenges ##nda equally resignation ##logical Tigers loop pitched outdoor reviewed hopes True temporarily Borough torn jerked collect Berkeley Independence cotton retreat campaigns participating Intelligence Heaven ##ked situations borough Democrats Harbor ##len Liga serial circles fourteen ##lot seized filling departments finance absolute Roland Nate floors raced struggling deliver protests ##tel Exchange efficient experiments ##dar faint 3D binding Lions lightly skill proteins difficulties ##cal monthly camps flood loves Amanda Commerce ##oid ##lies elementary ##tre organic ##stein ##ph receives Tech enormous distinctive Joint experiment Circuit citizen ##hy shelter ideal practically formula addressed Foster Productions ##ax variable punk Voice fastest concentrated ##oma ##yer stored surrender vary Sergeant Wells ward Wait ##ven playoff reducing cavalry ##dle Venezuela tissue amounts sweat ##we Non ##nik beetle ##bu ##tu Jared Hunt ##₂ fat Sultan Living Circle Secondary Suddenly reverse ##min Travel ##bin Lebanon ##mas virus Wind dissolved enrolled holiday Keep helicopter Clarke constitutional technologies doubles instructions ##ace Azerbaijan ##ill occasional frozen trick wiped writings Shanghai preparing challenged mainstream summit 180 ##arian ##rating designation ##ada revenge filming tightened Miguel Montana reflect celebration bitch flashed signals rounded peoples ##tation renowned Google characteristic Campaign sliding ##rman usage Record Using woke solutions holes theories logo Protestant relaxed brow nickname Reading marble ##tro symptoms Overall capita ##ila outbreak revolution deemed Principal Hannah approaches inducted Wellington vulnerable Environmental Drama incumbent Dame 1854 travels samples accurate physically Sony Nashville ##sville ##lic ##og Producer Lucky tough Stanford resort repeatedly eyebrows Far choir commenced ##ep ##ridge rage swing sequel heir buses ad Grove ##late ##rick updated ##SA Delaware ##fa Athletics warmth Off excitement verse Protection Villa corruption intellectual Jenny ##lyn mystery prayer healthy ##ologist Bear lab Ernest Remix register basement Montgomery consistent tier 1855 Preston Brooks ##maker vocalist laboratory delayed wheels rope bachelor pitcher Block Nevertheless suspect efficiency Nebraska siege FBI planted ##AC Newton breeding ##ain eighteen Argentine encounter servant 1858 elder Shadow Episode fabric doctors survival removal chemistry volunteers Kane variant arrives Eagle Left ##fe Jo divorce ##ret yesterday Bryan handling diseases customer Sheriff Tiger Harper ##oi resting Linda Sheffield gasped sexy economics alien tale footage Liberty yeah fundamental Ground flames Actress photographer Maggie Additional joke custom Survey Abu silk consumption Ellis bread ##uous engagement puts Dog ##hr poured guilt CDP boxes hardware clenched ##cio stem arena extending ##com examination Steel encountered revised 140 picking Car hasn Minor pride Roosevelt boards ##mia blocked curious drag narrative brigade Prefecture mysterious namely connects Devil historians CHAPTER quit installation Golf empire elevated ##eo releasing Bond ##uri harsh ban ##BA contracts cloth presents stake chorus ##eau swear ##mp allies generations Motor meter pen warrior veteran ##EC comprehensive missile interaction instruction Renaissance rested Dale fix fluid les investigate loaded widow exhibit artificial select rushing tasks signature nowhere Engineer feared Prague bother extinct gates Bird climbing heels striking artwork hunt awake ##hin Formula thereby commitment imprisoned Beyond ##MA transformed Agriculture Low Movie radical complicated Yellow Auckland mansion tenth Trevor predecessor ##eer disbanded sucked circular witch gaining lean Behind illustrated rang celebrate bike consist framework ##cent Shane owns 350 comprises collaborated colleagues ##cast engage fewer ##ave 1856 observation diplomatic legislature improvements Interstate craft MTV martial administered jet approaching permanently attraction manuscript numbered Happy Andrea shallow Gothic Anti ##bad improvement trace preserve regardless rode dies achievement maintaining Hamburg spine ##air flowing encourage widened posts ##bound 125 Southeast Santiago ##bles impression receiver Single closure ##unt communist honors Northwest 105 ##ulated cared un hug magnetic seeds topic perceived prey prevented Marvel Eight Michel Transportation rings Gate ##gne Byzantine accommodate floating ##dor equation ministry ##ito ##gled Rules earthquake revealing Brother Celtic blew chairs Panama Leon attractive descendants Care Ambassador tours breathed threatening ##cho smiles Lt Beginning ##iness fake assists fame strings Mobile Liu parks http 1852 brush Aunt bullet consciousness ##sta ##ther consequences gather dug 1851 bridges Doug ##sion Artists ignore Carol brilliant radiation temples basin clouds ##cted Stevens spite soap consumer Damn Snow recruited ##craft Advanced tournaments Quinn undergraduate questioned Palmer Annual Others feeding Spider printing ##orn cameras functional Chester readers Alpha universal Faith Brandon François authored Ring el aims athletic possessed Vermont programmes ##uck bore Fisher statements shed saxophone neighboring pronounced barrel bags ##dge organisations pilots casualties Kenneth ##brook silently Malcolm span Essex anchor ##hl virtual lessons Henri Trump Page pile locomotive wounds uncomfortable sustained Diana Eagles ##pi 2000s documented ##bel Cassie delay kisses ##ines variation ##ag growled ##mark ##ways Leslie studios Friedrich aunt actively armor eaten historically Better purse honey ratings ##ée naturally 1840 peer Kenny Cardinal database Looking runners handsome Double PA ##boat ##sted protecting ##jan Diamond concepts interface ##aki Watch Article Columbus dialogue pause ##rio extends blanket pulse 1853 affiliate ladies Ronald counted kills demons ##zation Airlines Marco Cat companion mere Yugoslavia Forum Allan pioneer Competition Methodist patent nobody Stockholm ##ien regulation ##ois accomplished ##itive washed sake Vladimir crops prestigious humor Sally labour tributary trap altered examined Mumbai bombing Ash noble suspension ruins ##bank spare displays guided dimensional Iraqi ##hon sciences Franz relating fence followers Palestine invented proceeded Batman Bradley ##yard ##ova crystal Kerala ##ima shipping handled Want abolished Drew ##tter Powell Half ##table ##cker exhibitions Were assignment assured ##rine Indonesian Grammy acknowledged Kylie coaches structural clearing stationed Say Total Rail besides glow threats afford Tree Musical ##pp elite centered explore Engineers Stakes Hello tourism severely assessment ##tly crack politicians ##rrow sheets volunteer ##borough ##hold announcement recover contribute lungs ##ille mainland presentation Johann Writing 1849 ##bird Study Boulevard coached fail airline Congo Plus Syrian introduce ridge Casey manages ##fi searched Support succession progressive coup cultures ##lessly sensation Cork Elena Sofia Philosophy mini trunk academy Mass Liz practiced Reid ##ule satisfied experts Wilhelm Woods invitation Angels calendar joy Sr Dam packed ##uan bastard Workers broadcasts logic cooking backward ##ack Chen creates enzyme ##xi Davies aviation VII Conservation fucking Knights ##kan requiring hectares wars ate ##box Mind desired oak absorbed Really Vietnamese Paulo athlete ##car ##eth Talk Wu ##cks survivors Yang Joel Almost Holmes Armed Joshua priests discontinued ##sey blond Rolling suggesting CA clay exterior Scientific ##sive Giovanni Hi farther contents Winners animation neutral mall Notes layers professionals Armstrong Against Piano involve monitor angel parked bears seated feat beliefs ##kers Version suffer ##ceae guidance ##eur honored raid alarm Glen Ellen Jamaica trio enabled ##ils procedures ##hus moderate upstairs ##ses torture Georgian rebellion Fernando Nice ##are Aires Campus beast ##hing 1847 ##FA Isle ##logist Princeton cathedral Oakland Solomon ##tto Milwaukee upcoming midfielder Neither sacred Eyes appreciate Brunswick secrets Rice Somerset Chancellor Curtis ##gel Rich separation grid ##los ##bon urge ##ees ##ree freight towers psychology requirement dollar ##fall ##sman exile tomb Salt Stefan Buenos Revival Porter tender diesel chocolate Eugene Legion Laboratory sheep arched hospitals orbit Full ##hall drinks ripped ##RS tense Hank leagues ##nberg PlayStation fool Punjab relatives Comedy sur 1846 Tonight Sox ##if Rabbi org speaks institute defender painful wishes Weekly literacy portions snake item deals ##tum autumn sharply reforms thighs prototype ##ition argues disorder Physics terror provisions refugees predominantly independently march ##graphy Arabia Andrews Bus Money drops ##zar pistol matrix revolutionary ##ust Starting ##ptic Oak Monica ##ides servants ##hed archaeological divorced rocket enjoying fires ##nel assembled qualification retiring ##fied Distinguished handful infection Durham ##itz fortune renewed Chelsea ##sley curved gesture retain exhausted ##ifying Perth jumping Palestinian Simpson colonies steal ##chy corners Finn arguing Martha ##var Betty emerging Heights Hindi Manila pianist founders regret Napoleon elbow overhead bold praise humanity ##ori Revolutionary ##ere fur ##ole Ashley Official ##rm lovely Architecture ##sch Baronet virtually ##OS descended immigration ##das ##kes Holly Wednesday maintains theatrical Evan Gardens citing ##gia segments Bailey Ghost ##city governing graphics ##ined privately potentially transformation Crystal Cabinet sacrifice hesitated mud Apollo Desert bin victories Editor Railways Web Case tourists Brussels Franco compiled topped Gene engineers commentary egg escort nerve arch necessarily frustration Michelle democracy genes Facebook halfway ##ient 102 flipped Won ##mit NASA Lynn Provincial ambassador Inspector glared Change McDonald developments tucked noting Gibson circulation dubbed armies resource Headquarters ##iest Mia Albanian Oil Albums excuse intervention Grande Hugo integration civilians depends reserves Dee compositions identification restrictions quarterback Miranda Universe favourite ranges hint loyal Op entity Manual quoted dealt specialist Zhang download Westminster Rebecca streams Anglican variations Mine detective Films reserved ##oke ##key sailing ##gger expanding recall discovers particles behaviour Gavin blank permit Java Fraser Pass ##non ##TA panels statistics notion courage dare venues ##roy Box Newport travelling Thursday warriors Glenn criteria 360 mutual restore varied bitter Katherine ##lant ritual bits ##à Henderson trips Richardson Detective curse psychological Il midnight streak facts Dawn Indies Edmund roster Gen ##nation 1830 congregation shaft ##ically ##mination Indianapolis Sussex loving ##bit sounding horrible Continental Griffin advised magical millions ##date 1845 Safety lifting determination valid dialect Penn Know triple avoided dancer judgment sixty farmer lakes blast aggressive Abby tag chains inscription ##nn conducting Scout buying ##wich spreading ##OC array hurried Environment improving prompted fierce Taking Away tune pissed Bull catching ##ying eyebrow metropolitan terrain ##rel Lodge manufacturers creator ##etic happiness ports ##ners Relations fortress targeted ##ST allegedly blues ##osa Bosnia ##dom burial similarly stranger pursued symbols rebels reflection routine traced indoor eventual ##ska ##ão ##una MD ##phone oh grants Reynolds rid operators ##nus Joey vital siblings keyboard br removing societies drives solely princess lighter Various Cavalry believing SC underwent relay smelled syndrome welfare authorized seemingly Hard chicken ##rina Ages Bo democratic barn Eye shorts ##coming ##hand disappointed unexpected centres Exhibition Stories Site banking accidentally Agent conjunction André Chloe resist width Queens provision ##art Melissa Honorary Del prefer abruptly duration ##vis Glass enlisted ##ado discipline Sisters carriage ##ctor ##sburg Lancashire log fuck ##iz closet collecting holy rape trusted cleaning inhabited Rocky 104 editorial ##yu ##ju succeed strict Cuban ##iya Bronze outcome ##ifies ##set corps Hero barrier Kumar groaned Nina Burton enable stability Milton knots ##ination slavery ##borg curriculum trailer warfare Dante Edgar revival Copenhagen define advocate Garrett Luther overcome pipe 750 construct Scotia kings flooding ##hard Ferdinand Felix forgot Fish Kurt elaborate ##BC graphic gripped colonel Sophia Advisory Self ##uff ##lio monitoring seal senses rises peaceful journals 1837 checking legendary Ghana ##power ammunition Rosa Richards nineteenth ferry aggregate Troy inter ##wall Triple steep tent Cyprus 1844 ##woman commanding farms doi navy specified na cricketer transported Think comprising grateful solve ##core beings clerk grain vector discrimination ##TC Katie reasonable drawings veins consideration Monroe repeat breed dried witnessed ordained Current spirits remarkable consultant urged Remember anime singers phenomenon Rhode Carlo demanding findings manual varying Fellowship generate safely heated withdrawn ##ao headquartered ##zon ##lav ##ency Col Memphis imposed rivals Planet healing ##hs ensemble Warriors ##bone cult Frankfurt ##HL diversity Gerald intermediate ##izes reactions Sister ##ously ##lica quantum awkward mentions pursuit ##ography varies profession molecular consequence lectures cracked 103 slowed ##tsu cheese upgraded suite substance Kingston 1800 Idaho Theory ##een ain Carson Molly ##OR configuration Whitney reads audiences ##tie Geneva Outside ##nen ##had transit volleyball Randy Chad rubber motorcycle respected eager Level coin ##lets neighbouring ##wski confident ##cious poll uncertain punch thesis Tucker IATA Alec ##ographic ##law 1841 desperately 1812 Lithuania accent Cox lightning skirt ##load Burns Dynasty ##ug chapters Working dense Morocco ##kins casting Set activated oral Brien horn HIV dawn stumbled altar tore considerably Nicole interchange registration biography Hull Stan bulk consent Pierce ##ER Fifth marched terrorist ##piece ##itt Presidential Heather staged Plant relegation sporting joins ##ced Pakistani dynamic Heat ##lf ourselves Except Elliott nationally goddess investors Burke Jackie ##ā ##RA Tristan Associate Tuesday scope Near bunch ##abad ##ben sunlight ##aire manga Willie trucks boarding Lion lawsuit Learning Der pounding awful ##mine IT Legend romance Serie AC gut precious Robertson hometown realm Guards Tag batting ##vre halt conscious 1838 acquire collar ##gg ##ops Herald nationwide citizenship Aircraft decrease em Fiction Female corporation Located ##ip fights unconscious Tampa Poetry lobby Malta ##sar ##bie layout Tate reader stained ##bre ##rst ##ulate loudly Eva Cohen exploded Merit Maya ##rable Rovers ##IC Morrison Should vinyl ##mie onwards ##gie vicinity Wildlife probability Mar Barnes ##ook spinning Moses ##vie Surrey Planning conferences protective Plaza deny Canterbury manor Estate tilted comics IBM destroying server Dorothy ##horn Oslo lesser heaven Marshal scales strikes ##ath firms attract ##BS controlling Bradford southeastern Amazon Travis Janet governed 1842 Train Holden bleeding gifts rent 1839 palms ##ū judicial Ho Finals conflicts unlikely draws ##cies compensation adds elderly Anton lasting Nintendo codes ministers pot associations capabilities ##cht libraries ##sie chances performers runway ##af ##nder Mid Vocals ##uch ##eon interpreted priority Uganda ruined Mathematics cook AFL Lutheran AIDS Capitol chase axis Moreover María Saxon storyline ##ffed Tears Kid cent colours Sex ##long pm blonde Edwin CE diocese ##ents ##boy Inn ##ller Saskatchewan ##kh stepping Windsor ##oka ##eri Xavier Resources 1843 ##top ##rad ##lls Testament poorly 1836 drifted slope CIA remix Lords mature hosting diamond beds ##ncies luxury trigger ##lier preliminary hybrid journalists Enterprise proven expelled insects Beautiful lifestyle vanished ##ake ##ander matching surfaces Dominican Kids referendum Orlando Truth Sandy privacy Calgary Speaker sts Nobody shifting ##gers Roll Armenia Hand ##ES 106 ##ont Guild larvae Stock flame gravity enhanced Marion surely ##tering Tales algorithm Emmy darker VIII ##lash hamlet deliberately occurring choices Gage fees settling ridiculous ##ela Sons cop custody ##ID proclaimed Cardinals ##pm Metal Ana 1835 clue Cardiff riders observations MA sometime ##och performer intact Points allegations rotation Tennis tenor Directors ##ats Transit thigh Complex ##works twentieth Factory doctrine Daddy ##ished pretend Winston cigarette ##IA specimens hydrogen smoking mathematical arguments openly developer ##iro fists somebody ##san Standing Caleb intelligent Stay Interior echoed Valentine varieties Brady cluster Ever voyage ##of deposits ultimate Hayes horizontal proximity ##ás estates exploration NATO Classical ##most bills condemned 1832 hunger ##ato planes deserve offense sequences rendered acceptance ##ony manufacture Plymouth innovative predicted ##RC Fantasy ##une supporter absent Picture bassist rescued ##MC Ahmed Monte ##sts ##rius insane novelist ##és agrees Antarctic Lancaster Hopkins calculated startled ##star tribal Amendment ##hoe invisible patron deer Walk tracking Lyon tickets ##ED philosopher compounds chuckled ##wi pound loyalty Academic petition refuses marking Mercury northeastern dimensions scandal Canyon patch publish ##oning Peak minds ##boro Presbyterian Hardy theoretical magnitude bombs cage ##ders ##kai measuring explaining avoiding touchdowns Card theology ##ured Popular export suspicious Probably photograph Lou Parks Arms compact Apparently excess Banks lied stunned territorial Filipino spectrum learns wash imprisonment ugly ##rose Albany Erik sends ##hara ##rid consumed ##gling Belgrade Da opposing Magnus footsteps glowing delicate Alexandria Ludwig gorgeous Bros Index ##PA customs preservation bonds ##mond environments ##nto instructed parted adoption locality workshops goalkeeper ##rik ##uma Brighton Slovenia ##ulating ##tical towel hugged stripped Bears upright Wagner ##aux secretly Adventures nest Course Lauren Boeing Abdul Lakes 450 ##cu USSR caps Chan ##nna conceived Actually Belfast Lithuanian concentrate possess militia pine protagonist Helena ##PS ##band Belle Clara Reform currency pregnancy 1500 ##rim Isabella hull Name trend journalism diet ##mel Recording acclaimed Tang Jace steering vacant suggestion costume laser ##š ##ink ##pan ##vić integral achievements wise classroom unions southwestern ##uer Garcia toss Tara Large ##tate evident responsibilities populated satisfaction ##bia casual Ecuador ##ght arose ##ović Cornwall embrace refuse Heavyweight XI Eden activists ##uation biology ##shan fraud Fuck matched legacy Rivers missionary extraordinary Didn holder wickets crucial Writers Hurricane Iceland gross trumpet accordance hurry flooded doctorate Albania ##yi united deceased jealous grief flute portraits ##а pleasant Founded Face crowned Raja advisor Salem ##ec Achievement admission freely minimal Sudan developers estimate disabled ##lane downstairs Bruno ##pus pinyin ##ude lecture deadly underlying optical witnesses Combat Julius tapped variants ##like Colonial Critics Similarly mouse voltage sculptor Concert salary Frances ##ground hook premises Software instructor nominee ##ited fog slopes ##zu vegetation sail ##rch Body Apart atop View utility ribs cab migration ##wyn bounded 2019 pillow trails ##ub Halifax shade Rush ##lah ##dian Notre interviewed Alexandra Springfield Indeed rubbing dozens amusement legally ##lers Jill Cinema ignoring Choice ##ures pockets ##nell laying Blair tackles separately ##teen Criminal performs theorem Communication suburbs ##iel competitors rows ##hai Manitoba Eleanor interactions nominations assassination ##dis Edmonton diving ##dine essay ##tas AFC Edge directing imagination sunk implement Theodore trembling sealed ##rock Nobel ##ancy ##dorf ##chen genuine apartments Nicolas AA Bach Globe Store 220 ##10 Rochester ##ño alert 107 Beck ##nin Naples Basin Crawford fears Tracy ##hen disk ##pped seventeen Lead backup reconstruction ##lines terrified sleeve nicknamed popped ##making ##ern Holiday Gospel ibn ##ime convert divine resolved ##quet ski realizing ##RT Legislature reservoir Rain sinking rainfall elimination challenging tobacco ##outs Given smallest Commercial pin rebel comedian exchanged airing dish Salvador promising ##wl relax presenter toll aerial ##eh Fletcher brass disappear zones adjusted contacts ##lk sensed Walt mild toes flies shame considers wildlife Hanna Arsenal Ladies naming ##ishing anxiety discussions cute undertaken Cash strain Wyoming dishes precise Angela ##ided hostile twins 115 Built ##pel Online tactics Newman ##bourne unclear repairs embarrassed listing tugged Vale ##gin Meredith bout ##cle velocity tips froze evaluation demonstrate ##card criticised Nash lineup Rao monks bacteria lease ##lish frightened den revived finale ##rance flee Letters decreased ##oh Sounds wrap Sharon incidents renovated everybody stole Bath boxing 1815 withdraw backs interim react murders Rhodes Copa framed flown Estonia Heavy explored ##rra ##GA ##ali Istanbul 1834 ##rite ##aging ##ues Episcopal arc orientation Maxwell infected ##rot BCE Brook grasp Roberto Excellence 108 withdrawal Marines rider Lo ##sin ##run Subsequently garrison hurricane facade Prussia crushed enterprise ##mber Twitter Generation Physical Sugar editing communicate Ellie ##hurst Ernst wagon promotional conquest Parliamentary courtyard lawyers Superman email Prussian lately lecturer Singer Majesty Paradise sooner Heath slot curves convoy ##vian induced synonym breeze ##plane ##ox peered Coalition ##hia odds ##esh ##lina Tomorrow Nadu ##ico ##rah damp autonomous console Victory counts Luxembourg intimate Archived Carroll spy Zero habit Always faction teenager Johnston chaos ruin commerce blog ##shed ##the reliable Word Yu Norton parade Catholics damned ##iling surgeon ##tia Allison Jonas remarked ##ès idiot Making proposals Industries strategies artifacts batteries reward ##vers Agricultural distinguish lengths Jeffrey Progressive kicking Patricia ##gio ballot ##ios skilled ##gation Colt limestone ##AS peninsula ##itis LA hotels shapes Crime depicting northwestern HD silly Das ##² ##ws ##ash ##matic thermal Has forgive surrendered Palm Nacional drank haired Mercedes ##foot loading Timothy ##roll mechanisms traces digging discussing Natalie ##zhou Forbes landmark Anyway Manor conspiracy gym knocking viewing Formation Pink Beauty limbs Phillip sponsor Joy granite Harbour ##ero payments Ballet conviction ##dam Hood estimates lacked Mad Jorge ##wen refuge ##LA invaded Kat suburban ##fold investigated Ari complained creek Georges ##uts powder accepting deserved carpet Thunder molecules Legal cliff strictly enrollment ranch ##rg ##mba proportion renovation crop grabbing ##liga finest entries receptor helmet blown Listen flagship workshop resolve nails Shannon portal jointly shining Violet overwhelming upward Mick proceedings ##dies ##aring Laurence Churchill ##rice commit 170 inclusion Examples ##verse ##rma fury paths ##SC ankle nerves Chemistry rectangular sworn screenplay cake Mann Seoul Animal sizes Speed vol Population Southwest Hold continuously Qualified wishing Fighting Made disappointment Portsmouth Thirty ##beck Ahmad teammate MLB graph Charleston realizes ##dium exhibits preventing ##int fever rivalry Male mentally dull ##lor ##rich consistently ##igan Madame certificate suited Krishna accuracy Webb Budapest Rex 1831 Cornell OK surveillance ##gated habitats Adventure Conrad Superior Gay sofa aka boot Statistics Jessie Liberation ##lip ##rier brands saint Heinrich Christine bath Rhine ballet Jin consensus chess Arctic stack furious cheap toy ##yre ##face ##gging gastropod ##nne Romans membrane answering 25th architects sustainable ##yne Hon 1814 Baldwin dome ##awa ##zen celebrity enclosed ##uit ##mmer Electronic locals ##CE supervision mineral Chemical Slovakia alley hub ##az heroes Creative ##AM incredible politically ESPN yanked halls Aboriginal Greatest yield ##20 congressional robot Kiss welcomed MS speeds proceed Sherman eased Greene Walsh Geoffrey variables rocky ##print acclaim Reverend Wonder tonnes recurring Dawson continent finite AP continental ID facilitate essays Rafael Neal 1833 ancestors ##met ##gic Especially teenage frustrated Jules cock expense ##oli ##old blocking Notable prohibited ca dock organize ##wald Burma Gloria dimension aftermath choosing Mickey torpedo pub ##used manuscripts laps Ulster staircase sphere Insurance Contest lens risks investigations ERA glare ##play Graduate auction Chronicle ##tric ##50 Coming seating Wade seeks inland Thames Rather butterfly contracted positioned consumers contestants fragments Yankees Santos administrator hypothesis retire Denis agreements Winnipeg ##rill 1820 trophy crap shakes Jenkins ##rium ya twist labels Maritime ##lings ##iv 111 ##ensis Cairo Anything ##fort opinions crowded ##nian abandon ##iff drained imported ##rr tended ##rain Going introducing sculptures bankruptcy danced demonstration stance settings gazed abstract pet Calvin stiff strongest wrestler ##dre Republicans grace allocated cursed snail advancing Return errors Mall presenting eliminate Amateur Institution counting ##wind warehouse ##nde Ethiopia trailed hollow ##press Literary capability nursing preceding lamp Thomson Morton ##ctic Crew Close composers boom Clare missiles 112 hunter snap ##oni ##tail Us declaration ##cock rally huh lion straightened Philippe Sutton alpha valued maker navigation detected favorable perception Charter ##ña Ricky rebounds tunnels slapped Emergency supposedly ##act deployment socialist tubes anybody corn ##NA Seminary heating pump ##AA achieving souls ##ass Link ##ele ##smith greeted Bates Americas Elder cure contestant 240 fold Runner Uh licked Politics committees neighbors fairy Silva Leipzig tipped correctly exciting electronics foundations cottage governmental ##hat allied claws presidency cruel Agreement slender accompanying precisely ##pass driveway swim Stand crews ##mission rely everyday Wings demo ##hic recreational min nationality ##duction Easter ##hole canvas Kay Leicester talented Discovery shells ##ech Kerry Ferguson Leave ##place altogether adopt butt wolves ##nsis ##ania modest soprano Boris ##ught electron depicts hid cruise differ treasure ##nch Gun Mama Bengali trainer merchants innovation presumably Shirley bottles proceeds Fear invested Pirates particle Dominic blamed Fight Daisy ##pper ##graphic nods knight Doyle tales Carnegie Evil Inter Shore Nixon transform Savannah ##gas Baltic stretching worlds protocol Percy Toby Heroes brave dancers ##aria backwards responses Chi Gaelic Berry crush embarked promises Madonna researcher realised inaugurated Cherry Mikhail Nottingham reinforced subspecies rapper ##kie Dreams Re Damon Minneapolis monsters suspicion Tel surroundings afterward complaints OF sectors Algeria lanes Sabha objectives Donna bothered distracted deciding ##ives ##CA ##onia bishops Strange machinery Voiced synthesis reflects interference ##TS ##ury keen ##ign frown freestyle ton Dixon Sacred Ruby Prison ##ión 1825 outfit ##tain curiosity ##ight frames steadily emigrated horizon ##erly Doc philosophical Table UTC Marina ##DA secular ##eed Zimbabwe cops Mack sheriff Sanskrit Francesco catches questioning streaming Kill testimony hissed tackle countryside copyright ##IP Buddhism ##rator ladder ##ON Past rookie depths ##yama ##ister ##HS Samantha Dana Educational brows Hammond raids envelope ##sco ##hart ##ulus epic detection Streets Potter statistical für ni accounting ##pot employer Sidney Depression commands Tracks averaged lets Ram longtime suits branded chip Shield loans ought Said sip ##rome requests Vernon bordered veterans ##ament Marsh Herzegovina Pine ##igo mills anticipation reconnaissance ##ef expectations protested arrow guessed depot maternal weakness ##ap projected pour Carmen provider newer remind freed ##rily ##wal ##tones intentions Fiji timing Match managers Kosovo Herman Wesley Chang 135 semifinals shouting Indo Janeiro Chess Macedonia Buck ##onies rulers Mail ##vas ##sel MHz Programme Task commercially subtle propaganda spelled bowling basically Raven 1828 Colony 109 ##ingham ##wara anticipated 1829 ##iers graduates ##rton ##fication endangered ISO diagnosed ##tage exercises Battery bolt poison cartoon ##ción hood bowed heal Meyer Reagan ##wed subfamily ##gent momentum infant detect ##sse Chapman Darwin mechanics NSW Cancer Brooke Nuclear comprised hire sanctuary wingspan contrary remembering surprising Basic stealing OS hatred ##lled masters violation Rule ##nger assuming conquered louder robe Beatles legitimate ##vation massacre Rica unsuccessfully poets ##enberg careers doubled premier battalions Dubai Paper Louisville gestured dressing successive mumbled Vic referee pupil ##cated ##rre ceremonies picks ##IN diplomat alike geographical rays ##HA ##read harbour factories pastor playwright Ultimate nationalist uniforms obtaining kit Amber ##pling screenwriter ancestry ##cott Fields PR Coleman rat Bavaria squeeze highlighted Adult reflecting Mel 1824 bicycle organizing sided Previously Underground Prof athletics coupled mortal Hampton worthy immune Ava ##gun encouraging simplified ##ssa ##nte ##ann Providence entities Pablo Strong Housing ##ista ##ators kidnapped mosque Kirk whispers fruits shattered fossil Empress Johns Webster Thing refusing differently specimen Ha ##EN ##tina ##elle ##night Horn neighbourhood Bolivia ##rth genres Pre ##vich Amelia swallow Tribune Forever Psychology Use ##bers Gazette ash ##usa Monster ##cular delegation blowing Oblast retreated automobile ##ex profits shirts devil Treasury ##backs Drums Ronnie gameplay expertise Evening resides Caesar unity Crazy linking Vision donations Isabel valve Sue WWE logical availability fitting revolt ##mill Linux taxi Access pollution statues Augustus ##pen cello ##some lacking ##ati Gwen ##aka ##ovich 1821 Wow initiatives Uruguay Cain stroked examine ##ī mentor moist disorders buttons ##tica ##anna Species Lynch museums scorer Poor eligibility op unveiled cats Title wheat critically Syracuse ##osis marketed enhance Ryder ##NG ##ull ##rna embedded throws foods happily ##ami lesson formats punched ##rno expressions qualities ##sal Gods ##lity elect wives ##lling jungle Toyota reversed Grammar Cloud Agnes ##ules disputed verses Lucien threshold ##rea scanned ##bled ##dley ##lice Kazakhstan Gardner Freeman ##rz inspection Rita accommodation advances chill Elliot thriller Constantinople ##mos debris whoever 1810 Santo Carey remnants Guatemala ##irs carriers equations mandatory ##WA anxious measurement Summit Terminal Erin ##zes LLC ##uo glancing sin ##₃ Downtown flowering Euro Leigh Lance warn decent recommendations ##ote Quartet ##rrell Clarence colleague guarantee 230 Clayton Beast addresses prospect destroyer vegetables Leadership fatal prints 190 ##makers Hyde persuaded illustrations Southampton Joyce beats editors mount ##grave Malaysian Bombay endorsed ##sian ##bee applying Religion nautical bomber Na airfield gravel ##rew Cave bye dig decree burden Election Hawk Fe ##iled reunited ##tland liver Teams Put delegates Ella ##fect Cal invention Castro bored ##kawa ##ail Trinidad NASCAR pond develops ##pton expenses Zoe Released ##rf organs beta parameters Neill ##lene lateral Beat blades Either ##hale Mitch ##ET ##vous Rod burnt phones Rising ##front investigating ##dent Stephanie ##keeper screening ##uro Swan Sinclair modes bullets Nigerian melody ##ques Rifle ##12 128 ##jin charm Venus ##tian fusion advocated visitor pinned genera 3000 Ferry Solo quantity regained platinum shoots narrowly preceded update ##ichi equality unaware regiments ally ##tos transmitter locks Seeing outlets feast reopened ##ows struggles Buddy 1826 bark elegant amused Pretty themed schemes Lisbon Te patted terrorism Mystery ##croft ##imo Madagascar Journey dealer contacted ##quez ITV vacation Wong Sacramento organisms ##pts balcony coloured sheer defines MC abortion forbidden accredited Newfoundland tendency entrepreneur Benny Tanzania needing finalist mythology weakened gown sentences Guest websites Tibetan UFC voluntary annoyed Welcome honestly correspondence geometry Deutsche Biology Help ##aya Lines Hector ##ael reluctant ##ages wears inquiry ##dell Holocaust Tourism Wei volcanic ##mates Visual sorts neighborhoods Running apple shy Laws bend Northeast feminist Speedway Murder visa stuffed fangs transmitted fiscal Ain enlarged ##ndi Cecil Peterson Benson Bedford acceptable ##CC ##wer purely triangle foster Alberto educator Highland acute LGBT Tina Mi adventures Davidson Honda translator monk enacted summoned ##ional collector Genesis Un liner Di Statistical ##CS filter Knox Religious Stella Estonian Turn ##ots primitive parishes ##lles complexity autobiography rigid cannon pursuing exploring ##gram ##mme freshman caves Expedition Traditional iTunes certification cooling ##ort ##gna ##IT ##lman ##VA Motion explosive licence boxer shrine loosely Brigadier Savage Brett MVP heavier ##elli ##gged Buddha Easy spells fails incredibly Georg stern compatible Perfect applies cognitive excessive nightmare neighbor Sicily appealed static ##₁ Aberdeen ##leigh slipping bride ##guard Um Clyde 1818 ##gible Hal Frost Sanders interactive Hour ##vor hurting bull termed shelf capturing ##pace rolls 113 ##bor Chilean teaches ##rey exam shipped Twin borrowed ##lift Shit ##hot Lindsay Below Kiev Lin leased ##sto Eli Diane Val subtropical shoe Bolton Dragons ##rification Vatican ##pathy Crisis dramatically talents babies ##ores surname ##AP ##cology cubic opted Archer sweep tends Karnataka Judy stint Similar ##nut explicitly ##nga interact Mae portfolio clinic abbreviated Counties ##iko hearts ##ı providers screams Individual ##etti Monument ##iana accessed encounters gasp ##rge defunct Avery ##rne nobility useless Phase Vince senator ##FL 1813 surprisingly ##illo ##chin Boyd rumors equity Gone Hearts chassis overnight Trek wrists submit civic designers ##rity prominence decorative derives starter ##AF wisdom Powers reluctantly measurements doctoral Noel Gideon Baden Cologne lawn Hawaiian anthology ##rov Raiders embassy Sterling ##pal Telugu troubled ##FC ##bian fountain observe ore ##uru ##gence spelling Border grinning sketch Benedict Xbox dialects readily immigrant Constitutional aided nevertheless SE tragedy ##ager ##rden Flash ##MP Europa emissions ##ield panties Beverly Homer curtain ##oto toilet Isn Jerome Chiefs Hermann supernatural juice integrity Scots auto Patriots Strategic engaging prosecution cleaned Byron investments adequate vacuum laughs ##inus ##nge Usually Roth Cities Brand corpse ##ffy Gas rifles Plains sponsorship Levi tray owed della commanders ##ead tactical ##rion García harbor discharge ##hausen gentleman endless highways ##itarian pleaded ##eta archive Midnight exceptions instances Gibraltar cart ##NS Darren Bonnie ##yle ##iva OCLC bra Jess ##EA consulting Archives Chance distances commissioner ##AR LL sailors ##sters enthusiasm Lang ##zia Yugoslav confirm possibilities Suffolk ##eman banner 1822 Supporting fingertips civilization ##gos technically 1827 Hastings sidewalk strained monuments Floyd Chennai Elvis villagers Cumberland strode albeit Believe planets combining Mohammad container ##mouth ##tures verb BA Tank Midland screened Gang Democracy Helsinki screens thread charitable ##version swiftly ma rational combine ##SS ##antly dragging Cliff Tasmania quest professionally ##aj rap ##lion livestock ##hua informal specially lonely Matthews Dictionary 1816 Observatory correspondent constitute homeless waving appreciated Analysis Meeting dagger ##AL Gandhi flank Giant Choir ##not glimpse toe Writer teasing springs ##dt Glory healthcare regulated complaint math Publications makers ##hips cement Need apologize disputes finishes Partners boring ups gains 1793 Congressional clergy Folk ##made ##nza Waters stays encoded spider betrayed Applied inception ##urt ##zzo wards bells UCLA Worth bombers Mo trademark Piper ##vel incorporates 1801 ##cial dim Twelve ##word Appeals tighter spacecraft ##tine coordinates ##iac mistakes Zach laptop Teresa ##llar ##yr favored Nora sophisticated Irving hammer División corporations niece ##rley Patterson UNESCO trafficking Ming balanced plaque Latvia broader ##owed Save confined ##vable Dalton tide ##right ##ural ##num swords caring ##eg IX Acting paved ##moto launching Antoine substantially Pride Philharmonic grammar Indoor Ensemble enabling 114 resided Angelo publicity chaired crawled Maharashtra Telegraph lengthy preference differential anonymous Honey ##itation wage ##iki consecrated Bryant regulatory Carr ##én functioning watches ##ú shifts diagnosis Search app Peters ##SE ##cat Andreas honours temper counsel Urdu Anniversary maritime ##uka harmony ##unk essence Lorenzo choked Quarter indie ##oll loses ##prints amendment Adolf scenario similarities ##rade ##LC technological metric Russians thoroughly ##tead cruiser 1806 ##nier 1823 Teddy ##psy au progressed exceptional broadcaster partnered fitness irregular placement mothers unofficial Garion Johannes 1817 regain Solar publishes Gates Broken thirds conversations dive Raj contributor quantities Worcester governance ##flow generating pretending Belarus ##voy radius skating Marathon 1819 affection undertook ##wright los ##bro locate PS excluded recreation tortured jewelry moaned ##logue ##cut Complete ##rop 117 ##II plantation whipped slower crater ##drome Volunteer attributes celebrations regards Publishers oath utilized Robbie Giuseppe fiber indication melted archives Damien storey affecting identifying dances alumni comparable upgrade rented sprint ##kle Marty ##lous treating railways Lebanese erupted occupy sympathy Jude Darling Qatar drainage McCarthy heel Klein computing wireless flip Du Bella ##ast ##ssen narrator mist sings alignment 121 2020 securing ##rail Progress missionaries brutal mercy ##shing Hip ##ache ##olo switching ##here Malay ##ob constituted Mohammed Often standings surge teachings ink detached systematic Trial Myanmar ##wo offs Reyes decoration translations wherever reviewer speculation Bangkok terminated ##ester beard RCA Aidan Associated Emerson Charity 1803 generous Dudley ATP ##haven prizes toxic gloves ##iles ##dos Turning myth Parade ##building Hits ##eva teamed Above Duchess Holt ##oth Sub Ace atomic inform Ship depend Jun ##bes Norwich globe Baroque Christina Cotton Tunnel kidding Concerto Brittany tasted phases stems angles ##TE ##nam ##40 charted Alison intensive Willis glory ##lit Bergen est taller ##dicate labeled ##ido commentator Warrior Viscount shortened aisle Aria Spike spectators goodbye overlooking mammals ##lude wholly Barrett ##gus accompany seventy employ ##mb ambitious beloved basket ##mma ##lding halted descendant pad exclaimed cloak ##pet Strait Bang Aviv sadness ##ffer Donovan 1880s agenda swinging ##quin jerk Boat ##rist nervously Silence Echo shout implies ##iser ##cking Shiva Weston damages ##tist effectiveness Horace cycling Rey ache Photography PDF Dear leans Lea ##vision booth attained disbelief ##eus ##ution Hop pension toys Eurovision faithful ##heads Andre owe default Atlas Megan highlights lovers Constantine Sixth masses ##garh emerge Auto Slovak ##oa ##vert Superintendent flicked inventor Chambers Frankie Romeo pottery companions Rudolf ##liers diary Unless tap alter Randall ##ddle ##eal limitations ##boards utterly knelt guaranteed Cowboys Islander horns ##ike Wendy sexually Smart breasts ##cian compromise Duchy AT Galaxy analog Style ##aking weighed Nigel optional Czechoslovakia practicing Ham ##0s feedback batted uprising operative applicable criminals classrooms Somehow ##ode ##OM Naomi Winchester ##pping Bart Regina competitor Recorded Yuan Vera lust Confederation ##test suck 1809 Lambert 175 Friend ##ppa Slowly ##⁺ Wake Dec ##aneous chambers Color Gus ##site Alternative ##world Exeter Omaha celebrities striker 210 dwarf meals Oriental Pearson financing revenues underwater Steele screw Feeling Mt acids badge swore theaters Moving admired lung knot penalties 116 fork ##cribed Afghan outskirts Cambodia oval wool fossils Ned Countess Darkness delicious ##nica Evelyn Recordings guidelines ##CP Sandra meantime Antarctica modeling granddaughter ##rial Roma Seventh Sunshine Gabe ##nton Shop Turks prolific soup parody ##nta Judith disciplines resign Companies Libya Jets inserted Mile retrieve filmmaker ##rand realistic unhappy ##30 sandstone ##nas ##lent ##ush ##rous Brent trash Rescue ##unted Autumn disgust flexible infinite sideways ##oss ##vik trailing disturbed 50th Newark posthumously ##rol Schmidt Josef ##eous determining menu Pole Anita Luc peaks 118 Yard warrant generic deserted Walking stamp tracked ##berger paired surveyed sued Rainbow ##isk Carpenter submarines realization touches sweeping Fritz module Whether resembles ##form ##lop unsure hunters Zagreb unemployment Senators Georgetown ##onic Barker foul commercials Dresden Words collision Carlton Fashion doubted ##ril precision MIT Jacobs mob Monk retaining gotta ##rod remake Fast chips ##pled sufficiently ##lights delivering ##enburg Dancing Barton Officers metals ##lake religions ##ré motivated differs dorsal ##birds ##rts Priest polished ##aling Saxony Wyatt knockout ##hor Lopez RNA ##link metallic ##kas daylight Montenegro ##lining wrapping resemble Jam Viking uncertainty angels enables ##fy Stuttgart tricks tattoo 127 wicked asset breach ##yman MW breaths Jung im 1798 noon vowel ##qua calmly seasonal chat ingredients cooled Randolph ensuring ##ib ##idal flashing 1808 Macedonian Cool councils ##lick advantages Immediately Madras ##cked Pain fancy chronic Malayalam begged ##nese Inner feathers ##vey Names dedication Sing pan Fischer nurses Sharp inning stamps Meg ##ello edged motioned Jacksonville ##ffle ##dic ##US divide garnered Ranking chasing modifications ##oc clever midst flushed ##DP void ##sby ambulance beaches groan isolation strengthen prevention ##ffs Scouts reformed geographic squadrons Fiona Kai Consequently ##uss overtime ##yas Fr ##BL Papua Mixed glances Haiti Sporting sandy confronted René Tanner 1811 ##IM advisory trim ##ibe González gambling Jupiter ##ility ##owski ##nar 122 apology teased Pool feminine wicket eagle shiny ##lator blend peaking nasty nodding fraction tech Noble Kuwait brushing Italia Canberra duet Johan 1805 Written cameo Stalin pig cord ##zio Surely SA owing holidays 123 Ranger lighthouse ##ige miners 1804 ##ë ##gren ##ried crashing ##atory wartime highlight inclined Torres Tax ##zel ##oud Own ##corn Divine EMI Relief Northwestern ethics BMW click plasma Christie coordinator Shepherd washing cooked ##dio ##eat Cerambycidae algebra Engine costumes Vampire vault submission virtue assumption ##rell Toledo ##oting ##rva crept emphasized ##lton ##ood Greeks surgical crest Patrol Beta Tessa ##GS pizza traits rats Iris spray ##GC Lightning binary escapes ##take Clary crowds ##zong hauled maid ##fen Manning ##yang Nielsen aesthetic sympathetic affiliation soaked Mozart personalities begging ##iga clip Raphael yearly Lima abundant ##lm 1794 strips Initiative reporters ##vsky consolidated ##itated Civic rankings mandate symbolic ##ively 1807 rental duck nave complications ##nor Irene Nazis haunted scholarly Pratt Gran Embassy Wave pity genius bats canton Tropical marker ##cos escorted Climate ##posed appreciation freezing puzzle Internal pools Shawn pathway Daniels Fitzgerald extant olive Vanessa marriages cocked ##dging prone chemicals doll drawer ##HF Stark Property ##tai flowed Sheridan ##uated Less Omar remarks catalogue Seymour wreck Carrie ##bby Mercer displaced sovereignty rip Flynn Archie Quarterfinals Hassan ##ards vein Osaka pouring wages Romance ##cript ##phere 550 ##eil ##stown Documentary ancestor CNN Panthers publishers Rise ##mu biting Bright String succeeding 119 loaned Warwick Sheikh Von Afterwards Jax Camden helicopters Hence Laurel ##ddy transaction Corp clause ##owing ##kel Investment cups Lucia Moss Giles chef López decisive 30th distress linguistic surveys Ready maiden Touch frontier incorporate exotic mollusk Leopold Ride ##wain ##ndo teammates tones drift ordering Feb Penny Normandy Present Flag pipes ##rro delight motto Tibet leap Eliza Produced teenagers sitcom Try Hansen Cody wandered terrestrial frog scare resisted employers coined ##DS resistant Fly captive dissolution judged associates defining ##court Hale ##mbo raises clusters twelfth ##metric Roads ##itude satisfy Android Reds Gloucester Category Valencia Daemon stabbed Luna Churches Canton ##eller Attack Kashmir annexed grabs asteroid Hartford recommendation Rodriguez handing stressed frequencies delegate Bones Erie Weber Hands Acts millimetres 24th Fat Howe casually ##SL convent 1790 IF ##sity 1795 yelling ##ises drain addressing amino Marcel Sylvia Paramount Gerard Volleyball butter 124 Albion ##GB triggered 1792 folding accepts ##ße preparations Wimbledon dose ##grass escaping ##tling import charging ##dation 280 Nolan ##fried Calcutta ##pool Cove examining minded heartbeat twisting domains bush Tunisia Purple Leone ##code evacuated battlefield tiger Electrical ##ared chased ##cre cultivated Jet solved shrug ringing Impact ##iant kilometre ##log commemorate migrated singular designing promptly Higgins ##own ##aves freshwater Marketing Payne beg locker pray implied AAA corrected Trans Europeans Ashe acknowledge Introduction ##writer ##llen Munster auxiliary growl Hours Poems ##AT reduces Plain plague canceled detention polite necklace Gustav ##gu ##lance En Angola ##bb dwelling ##hea 5000 Qing Dodgers rim ##ored ##haus spilled Elisabeth Viktor backpack 1802 amended ##worthy Phantom ##ctive keeper ##loom Vikings ##gua employs Tehran specialty ##bate Marx Mirror Jenna rides needle prayers clarinet forewings ##walk Midlands convincing advocacy Cao Birds cycles Clement Gil bubble Maximum humanitarian Tan cries ##SI Parsons Trio offshore Innovation clutched 260 ##mund ##duct Prairie relied Falcon ##ste Kolkata Gill Swift Negro Zoo valleys ##OL Opening beams MPs outline Bermuda Personal exceed productive ##MT republic forum ##sty tornado Known dipped Edith folks mathematician watershed Ricardo synthetic ##dication deity ##₄ gaming subjected suspects Foot swollen Motors ##tty ##ý aloud ceremonial es nuts intend Carlisle tasked hesitation sponsors unified inmates ##ctions ##stan tiles jokes whereby outcomes Lights scary Stoke Portrait Blind sergeant violations cultivation fuselage Mister Alfonso candy sticks teen agony Enough invite Perkins Appeal mapping undergo Glacier Melanie affects incomplete ##dd Colombian ##nate CBC purchasing bypass Drug Electronics Frontier Coventry ##aan autonomy scrambled Recent bounced cow experiencing Rouge cuisine Elite disability Ji inheritance wildly Into ##wig confrontation Wheeler shiver Performing aligned consequently Alexis Sin woodland executives Stevenson Ferrari inevitable ##cist ##dha ##base Corner comeback León ##eck ##urus MacDonald pioneering breakdown landscapes Veterans Rican Theological stirred participant Credit Hyderabad snails Claudia ##ocene compliance ##MI Flags Middlesex storms winding asserted er ##ault ##kal waking ##rates abbey Augusta tooth trustees Commodore ##uded Cunningham NC Witch marching Sword Same spiral Harley ##ahan Zack Audio 1890s ##fit Simmons Kara Veronica negotiated Speaking FIBA Conservatory formations constituencies explicit facial eleventh ##ilt villain ##dog ##case ##hol armored tin hairs ##umi ##rai mattress Angus cease verbal Recreation savings Aurora peers Monastery Airways drowned additions downstream sticking Shi mice skiing ##CD Raw Riverside warming hooked boost memorable posed treatments 320 ##dai celebrating blink helpless circa Flowers PM uncommon Oct Hawks overwhelmed Sparhawk repaired Mercy pose counterpart compare survives ##½ ##eum coordinate Lil grandchildren notorious Yi Judaism Juliet accusations 1789 floated marathon roar fortified reunion 145 Nov Paula ##fare ##toria tearing Cedar disappearance Si gifted scar 270 PBS Technologies Marvin 650 roller cupped negotiate ##erman passport tram miracle styled ##tier necessity Des rehabilitation Lara USD psychic wipe ##lem mistaken ##lov charming Rider pageant dynamics Cassidy ##icus defenses ##tadt ##vant aging ##inal declare mistress supervised ##alis ##rest Ashton submerged sack Dodge grocery ramp Teacher lineage imagery arrange inscriptions Organisation Siege combines pounded Fleming legends columnist Apostolic prose insight Arabian expired ##uses ##nos Alone elbows ##asis ##adi ##combe Step Waterloo Alternate interval Sonny plains Goals incorporating recruit adjoining Cheshire excluding marrying ducked Cherokee par ##inate hiking Coal ##bow natives ribbon Allies con descriptions positively ##lal defendant 22nd Vivian ##beat Weather possessions Date sweetheart inability Salisbury adviser ideology Nordic ##eu Cubs IP Administrative ##nick facto liberation Burnett Javier fashioned Electoral Turin theft unanimous Per 1799 Clan Hawkins Teachers ##wes Cameroon Parkway ##gment demolition atoms nucleus ##thi recovering ##yte ##vice lifts Must deposit Hancock Semi darkened Declaration moan muscular Myers attractions sauce simulation ##weed Alps barriers ##baum Barack galleries Min holders Greenwich donation Everybody Wolfgang sandwich Kendra Collegiate casino Slavic ensuing Porto ##grapher Jesuit suppressed tires Ibrahim protesters Ibn Amos 1796 phenomena Hayden Paraguay Squad Reilly complement aluminum ##eers doubts decay demise Practice patience fireplace transparent monarchy ##person Rodney mattered rotating Clifford disposal Standards paced ##llie arise tallest tug documentation node freeway Nikolai ##cite clicked imaging Lorraine Tactical Different Regular Holding 165 Pilot guarded ##polis Classics Mongolia Brock monarch cellular receptors Mini Chandler financed financially Lives erection Fuller unnamed Kannada cc passive plateau ##arity freak ##rde retrieved transactions ##sus 23rd swimmer beef fulfill Arlington offspring reasoning Rhys saves pseudonym centimetres shivered shuddered ##ME Feel ##otic professors Blackburn ##eng ##life ##haw interred lodge fragile Della guardian ##bbled catalog clad observer tract declaring ##headed Lok dean Isabelle 1776 irrigation spectacular shuttle mastering ##aro Nathaniel Retired ##lves Brennan ##kha dick ##dated ##hler Rookie leapt televised weekends Baghdad Yemen ##fo factions ion Lab mortality passionate Hammer encompasses confluence demonstrations Ki derivative soils ##unch Ranch Universities conventions outright aiming hierarchy reside illusion graves rituals 126 Antwerp Dover ##ema campuses Hobart lifelong aliens ##vity Memory coordination alphabet ##mina Titans pushes Flanders ##holder Normal excellence capped profound Taipei portrayal sparked scratch se ##eas ##hir Mackenzie ##cation Neo Shin ##lined magnificent poster batsman ##rgent persuade ##ement Icelandic miserable collegiate Feature geography ##mura Comic Circus processor barracks Tale ##11 Bulls ##rap strengthened ##bell injection miniature broadly Letter fare hostage traders ##nium ##mere Fortune Rivera Lu triumph Browns Bangalore cooperative Basel announcing Sawyer ##him ##cco ##kara darted ##AD ##nova sucking ##position perimeter flung Holdings ##NP Basque sketches Augustine Silk Elijah analyst armour riots acquiring ghosts ##ems 132 Pioneer Colleges Simone Economy Author semester Soldier il ##unting ##bid freaking Vista tumor ##bat murderer ##eda unreleased ##grove ##sser ##té edit statute sovereign ##gawa Killer stares Fury comply ##lord ##nant barrels Andhra Maple generator mascot unusually eds ##ante ##runner rod ##tles Historically Jennings dumped Established resemblance ##lium ##cise ##body ##voke Lydia ##hou ##iring nonetheless 1797 corrupt patrons physicist sneak Livingston Citizens Architects Werner trends Melody eighty markings brakes ##titled oversaw processed mock Midwest intervals ##EF stretches werewolf ##MG Pack controller ##dition Honours cane Griffith vague repertoire Courtney orgasm Abdullah dominance occupies Ya introduces Lester instinct collaborative Indigenous refusal ##rank outlet debts spear 155 ##keeping ##ulu Catalan ##osh tensions ##OT bred crude Dunn abdomen accurately ##fu ##lough accidents Row Audrey rude Getting promotes replies Paolo merge ##nock trans Evangelical automated Canon ##wear ##ggy ##gma Broncos foolish icy Voices knives Aside dreamed generals molecule AG rejection insufficient ##nagar deposited sacked Landing arches helpful devotion intake Flower PGA dragons evolutionary ##mail 330 GM tissues ##tree arcade composite lid Across implications lacks theological assessed concentrations Den ##mans ##ulous Fu homeland ##stream Harriet ecclesiastical troop ecological winked ##xed eighteenth Casino specializing ##sworth unlocked supreme devastated snatched trauma GDP Nord saddle Wes convenient competes ##nu ##iss Marian subway ##rri successes umbrella ##far ##ually Dundee ##cence spark ##rix ##я Quality Geological cockpit rpm Cam Bucharest riot ##PM Leah ##dad ##pose Ka m³ Bundesliga Wolfe grim textile quartet expressing fantastic destroyers eternal picnic ##oro contractor 1775 spanning declining ##cating Lowe Sutherland Emirates downward nineteen violently scout viral melting enterprises ##cer Crosby Jubilee antenna urgent Rory ##uin ##sure wandering ##gler ##vent Suzuki Lifetime Dirty occupying ##quent Disc Guru mound Lennon Humanities listeners Walton uh Braves Bologna ##bis ##gra Dwight crawl flags memoir Thorne Archdiocese dairy ##uz ##tery roared adjust patches inn Knowing ##bbed ##zan scan Papa precipitation angrily passages postal Phi embraced blacks economist triangular Sen shooter punished Millennium Swimming confessed Aston defeats Era cousins Williamson ##rer daytime dumb ##rek underway specification Buchanan prayed concealed activation ##issa canon awesome Starr plural summers ##fields Slam unnecessary 1791 resume trilogy compression ##rough selective dignity Yan ##xton immense ##yun lone seeded hiatus lightweight summary Yo approve Galway rejoined Elise garbage burns speeches 129 Honduras ##liness inventory jersey FK assure slumped Lionel Suite ##sbury Lena continuation ##AN brightly ##nti GT Knowledge ##park ##lius lethal ##tribution ##sions Certificate Mara ##lby algorithms Jade blows pirates fleeing wheelchair Stein sophomore Alt Territorial diploma snakes ##olic ##tham Tiffany Pius flush urging Hanover Reich ##olate Unity Pike collectively Theme ballad kindergarten rocked zoo ##page whip Rodríguez strokes checks Becky Stern upstream ##uta Silent volunteered Sigma ##ingen ##tract ##ede Gujarat screwed entertaining ##action ##ryn defenders innocence lesbian que Richie nodes Lie juvenile Jakarta safer confront Bert breakthrough gospel Cable ##zie institutional Archive brake liquor feeds ##iate chancellor Encyclopedia Animation scanning teens ##mother Core Rear Wine ##flower reactor Ave cardinal sodium strands Olivier crouched Vaughan Sammy Image scars Emmanuel flour bias nipple revelation ##ucci Denny ##ssy Form Runners admits Rama violated Burmese feud underwear Mohamed Named swift statewide Door Recently comparing Hundred ##idge ##nity ##rds Rally Reginald Auburn solving waitress Treasurer ##ilization Halloween Ministers Boss Shut ##listic Rahman demonstrating ##pies Gaza Yuri installations Math schooling ##bble Bronx exiled gasoline 133 bundle humid FCC proportional relate VFL ##dez continuity ##cene syndicated atmospheric arrows Wanderers reinforcements Willow Lexington Rotten ##yon discovering Serena portable ##lysis targeting £1 Goodman Steam sensors detachment Malik ##erie attitudes Goes Kendall Read Sleep beans Nikki modification Jeanne knuckles Eleven ##iously Gross Jaime dioxide moisture Stones UCI displacement Metacritic Jury lace rendering elephant Sergei ##quire GP Abbott ##type projection Mouse Bishops whispering Kathleen Rams ##jar whites ##oran assess dispatched ##hire kin ##mir Nursing advocates tremendous sweater assisting ##bil Farmer prominently reddish Hague cyclone ##SD Sage Lawson Sanctuary discharged retains ##ube shotgun wilderness Reformed similarity Entry Watts Bahá Quest Looks visions Reservoir Arabs curls Blu dripping accomplish Verlag drill sensor Dillon physicians smashed ##dir painters Renault straw fading Directorate lounge commissions Brain ##graph neo ##urg plug coordinated ##houses Critical lamps illustrator Returning erosion Crow ##ciation blessing Thought Wife medalist synthesizer Pam Thornton Esther HBO fond Associates ##raz pirate permits Wide tire ##PC Ernie Nassau transferring RFC ##ntly um spit AS ##mps Mining polar villa anchored ##zzi embarrassment relates ##ă Rupert counterparts 131 Baxter ##18 Igor recognizes Clive ##hane ##eries ##ibly occurrence ##scope fin colorful Rapids banker tile ##rative ##dus delays destinations ##llis Pond Dane grandparents rewarded socially motorway ##hof ##lying ##human modeled Dayton Forward conscience Sharma whistle Mayer Sasha ##pical circuits Zhou ##ça Latvian finalists predators Lafayette closes obligations Resolution ##vier Trustees reminiscent ##hos Highlands Protected asylum evacuation ##acy Chevrolet confession Somalia emergence separating ##rica alright calcium Laurent Welfare Leonardo ashes dental Deal minerals ##lump ##mount accounted staggered slogan photographic builder ##imes ##raft tragic 144 SEC Hit tailed ##ples ##rring ##rson ethical wrestlers concludes lunar ##ept nitrogen Aid cyclist quarterfinals ##ه harvest ##hem Pasha IL ##mis continually ##forth Intel bucket ##ended witches pretended dresses viewer peculiar lowering volcano Marilyn Qualifier clung ##sher Cut modules Bowie ##lded onset transcription residences ##pie ##itor scrapped ##bic Monaco Mayo eternity Strike uncovered skeleton ##wicz Isles bug Promoted ##rush Mechanical XII ##ivo gripping stubborn velvet TD decommissioned operas spatial unstable Congressman wasted ##aga ##ume advertisements ##nya obliged Cannes Conway bricks ##gnant ##mity ##uise jumps Clear ##cine ##sche chord utter Su podium spokesman Royce assassin confirmation licensing liberty ##rata Geographic individually detained ##ffe Saturn crushing airplane bushes knights ##PD Lilly hurts unexpectedly Conservatives pumping Forty candle Pérez peasants supplement Sundays ##ggs ##rries risen enthusiastic corresponds pending ##IF Owens floods Painter inflation presumed inscribed Chamberlain bizarre 1200 liability reacted tub Legacy ##eds ##pted shone ##litz ##NC Tiny genome bays Eduardo robbery stall hatch Depot Variety Flora reprinted trembled outlined CR Theresa spans ##plication Jensen ##eering posting ##rky pays ##ost Marcos fortifications inferior ##ential Devi despair Talbot ##chus updates ego Booth Darius tops ##lau Scene ##DC Harlem Trey Generally candles ##α Neville Admiralty ##hong iconic victorious 1600 Rowan abundance miniseries clutching sanctioned ##words obscure ##ision ##rle ##EM disappearing Resort Obviously ##eb exceeded 1870s Adults ##cts Cry Kerr ragged selfish ##lson circled pillars galaxy ##asco ##mental rebuild caution Resistance Start bind splitting Baba Hogan ps partnerships slam Peggy courthouse ##OD organizational packages Angie ##nds possesses ##rp Expressway Gould Terror Him Geoff nobles ##ope shark ##nh identifies ##oor testified Playing ##ump ##isa stool Idol ##pice ##tana Byrne Gerry grunted 26th observing habits privilege immortal wagons ##thy dot Bring ##lian ##witz newest ##uga constraints Screen Issue ##RNA ##vil reminder ##gles addiction piercing stunning var ##rita Signal accumulated ##wide float devastating viable cartoons Uttar flared ##encies Theology patents ##bahn privileges ##ava ##CO 137 ##oped ##NT orchestral medication 225 erect Nadia École fried Sales scripts ##rease airs Cage inadequate structured countless Avengers Kathy disguise mirrors Investigation reservation ##nson Legends humorous Mona decorations attachment Via motivation Browne strangers ##ński Shadows Twins ##pressed Alma Nominated ##ott Sergio canopy 152 Semifinals devised ##irk upwards Traffic Goddess Move beetles 138 spat ##anne holdings ##SP tangled Whilst Fowler anthem ##ING ##ogy snarled moonlight songwriting tolerance Worlds exams ##pia notices sensitivity poetic Stephens Boone insect reconstructed Fresh 27th balloon ##ables Brendan mug ##gee 1780 apex exports slides Lahore hiring Shell electorate sexuality poker nonprofit ##imate cone ##uce Okinawa superintendent ##HC referenced turret Sprint Citizen equilibrium Stafford curb Driver Valerie ##rona aching impacts ##bol observers Downs Shri ##uth airports ##uda assignments curtains solitary icon patrols substances Jasper mountainous Published ached ##ingly announce dove damaging ##tism Primera Dexter limiting batch ##uli undergoing refugee Ye admiral pavement ##WR ##reed pipeline desires Ramsey Sheila thickness Brotherhood Tea instituted Belt Break plots ##ais masculine ##where Theo ##aged ##mined Experience scratched Ethiopian Teaching ##nov Aiden Abe Samoa conditioning ##mous Otherwise fade Jenks ##encing Nat ##lain Anyone ##kis smirk Riding ##nny Bavarian blessed potatoes Hook ##wise likewise hardened Merry amid persecution ##sten Elections Hoffman Pitt ##vering distraction exploitation infamous quote averaging healed Rhythm Germanic Mormon illuminated guides ##ische interfere ##ilized rector perennial ##ival Everett courtesy ##nham Kirby Mk ##vic Medieval ##tale Luigi limp ##diction Alive greeting shove ##force ##fly Jasmine Bend Capt Suzanne ditch 134 ##nning Host fathers rebuilding Vocal wires ##manship tan Factor fixture ##LS Māori Plate pyramid ##umble slap Schneider yell ##ulture ##tional Goodbye sore ##pher depressed ##dox pitching Find Lotus ##wang strand Teen debates prevalent ##bilities exposing hears billed ##rse reorganized compelled disturbing displaying ##tock Clinical emotionally ##iah Derbyshire grouped ##quel Bahrain Journalism IN persistent blankets Crane camping Direct proving Lola ##dding Corporate birthplace ##boats ##ender Figure dared Assam precursor ##nched Tribe Restoration slate Meyrick hunted stroking Earlier Kind polls appeals monetary ##reate Kira Langdon explores GPS extensions squares Results draped announcer merit ##ennial ##tral ##roved ##cion robots supervisor snorted ##group Cannon procession monkey freeze sleeves Nile verdict ropes firearms extraction tensed EC Saunders ##tches diamonds Marriage ##amble curling Amazing ##haling unrelated ##roads Daughter cum discarded kidney cliffs forested Candy ##lap authentic tablet notation ##nburg Bulldogs Callum Meet mouths coated ##xe Truman combinations ##mation Steelers Fan Than paternal ##father ##uti Rebellion inviting Fun theatres ##ي ##rom curator ##cision networking Oz drought ##ssel granting MBA Shelby Elaine jealousy Kyoto shores signaling tenants debated Intermediate Wise ##hes ##pu Havana duke vicious exited servers Nonetheless Reports explode ##beth Nationals offerings Oval conferred eponymous folklore ##NR Shire planting 1783 Zeus accelerated Constable consuming troubles McCartney texture bust Immigration excavated hopefully ##cession ##coe ##name ##ully lining Einstein Venezuelan reissued minorities Beatrice crystals ##nies circus lava Beirut extinction ##shu Becker ##uke issuing Zurich extract ##esta ##rred regulate progression hut alcoholic plea AB Norse Hubert Mansfield ashamed ##put Bombardment stripes electrons Denise horrified Nor arranger Hay Koch ##ddling ##iner Birthday Josie deliberate explorer ##jiang ##signed Arrow wiping satellites baritone mobility ##rals Dorset turbine Coffee 185 ##lder Cara Colts pits Crossing coral ##birth Tai zombie smoothly ##hp mates ##ady Marguerite ##tary puzzled tapes overly Sonic Prayer Thinking ##uf IEEE obligation ##cliffe Basil redesignated ##mmy nostrils Barney XIII ##phones vacated unused Berg ##roid Towards viola 136 Event subdivided rabbit recruiting ##nery Namibia ##16 ##ilation recruits Famous Francesca ##hari Goa ##lat Karachi haul biblical ##cible MGM ##rta horsepower profitable Grandma importantly Martinez incoming ##kill beneficial nominal praying ##isch gable nail noises ##ttle Polytechnic rub ##cope Thor audition erotic ##ending ##iano Ultimately armoured ##mum presently pedestrian ##tled Ipswich offence ##ffin ##borne Flemish ##hman echo ##cting auditorium gentlemen winged ##tched Nicaragua Unknown prosperity exhaust pie Peruvian compartment heights disabilities ##pole Harding Humphrey postponed moths Mathematical Mets posters axe ##nett Nights Typically chuckle councillors alternating 141 Norris ##ately ##etus deficit dreaming cooler oppose Beethoven ##esis Marquis flashlight headache investor responding appointments ##shore Elias ideals shades torch lingering ##real pier fertile Diploma currents Snake ##horse ##15 Briggs ##ota ##hima ##romatic Coastal Kuala ankles Rae slice Hilton locking Approximately Workshop Niagara strangely ##scence functionality advertisement Rapid Anders ho Soviets packing basal Sunderland Permanent ##fting rack tying Lowell ##ncing Wizard mighty tertiary pencil dismissal torso grasped ##yev Sand gossip ##nae Beer implementing ##19 ##riya Fork Bee ##eria Win ##cid sailor pressures ##oping speculated Freddie originating ##DF ##SR ##outh 28th melt Brenda lump Burlington USC marginal ##bine Dogs swamp cu Ex uranium metro spill Pietro seize Chorus partition ##dock ##media engineered ##oria conclusions subdivision ##uid Illustrated Leading ##hora Berkshire definite ##books ##cin ##suke noun winced Doris dissertation Wilderness ##quest braced arbitrary kidnapping Kurdish ##but clearance excavations wanna Allmusic insult presided yacht ##SM Honour Tin attracting explosives Gore Bride ##ience Packers Devils Observer ##course Loser ##erry ##hardt ##mble Cyrillic undefeated ##stra subordinate ##ame Wigan compulsory Pauline Cruise Opposition ##ods Period dispersed expose ##60 ##has Certain Clerk Wolves ##hibition apparatus allegiance orbital justified thanked ##ević Biblical Carolyn Graves ##tton Hercules backgrounds replica 1788 aquatic Mega Stirling obstacles filing Founder vowels Deborah Rotterdam surpassed Belarusian ##ologists Zambia Ren Olga Alpine bi councillor Oaks Animals eliminating digit Managing ##GE laundry ##rdo presses slamming Tudor thief posterior ##bas Rodgers smells ##ining Hole SUV trombone numbering representations Domingo Paralympics cartridge ##rash Combined shelves Kraków revision ##frame Sánchez ##tracted ##bler Alain townships sic trousers Gibbs anterior symmetry vaguely Castile IRA resembling Penguin ##ulent infections ##stant raped ##pressive worrying brains bending JR Evidence Venetian complexes Jonah 850 exported Ambrose Gap philanthropist ##atus Marxist weighing ##KO ##nath Soldiers chiefs reject repeating shaky Zürich preserving ##xin cigarettes ##break mortar ##fin Already reproduction socks Waiting amazed ##aca dash ##path Airborne ##harf ##get descending OBE Sant Tess Lucius enjoys ##ttered ##ivation ##ete Leinster Phillies execute geological unfinished Courts SP Beaver Duck motions Platinum friction ##aud ##bet Parts Stade entirety sprang Smithsonian coffin prolonged Borneo ##vise unanimously ##uchi Cars Cassandra Australians ##CT ##rgen Louisa spur Constance ##lities Patent racism tempo ##ssion ##chard ##nology ##claim Million Nichols ##dah Numerous ing Pure plantations donor ##EP ##rip convenience ##plate dots indirect ##written Dong failures adapt wizard unfortunately ##gion practitioners economically Enrique unchanged kingdoms refined definitions lazy worries railing ##nay Kaiser ##lug cracks sells ninety ##WC Directed denotes developmental papal unfortunate disappointing sixteenth Jen ##urier NWA drifting Horror ##chemical behaviors bury surfaced foreigners slick AND ##rene ##ditions ##teral scrap kicks comprise buddy ##anda Mental ##ype Dom wines Limerick Luca Rand ##won Tomatoes homage geometric ##nted telescope Shelley poles ##fan shareholders Autonomous cope intensified Genoa Reformation grazing ##tern Zhao provisional ##bies Con ##riel Cynthia Raleigh vivid threaten Length subscription roses Müller ##isms robin ##tial Laos Stanton nationalism ##clave ##ND ##17 ##zz staging Busch Cindy relieve ##spective packs neglected CBE alpine Evolution uneasy coastline Destiny Barber Julio ##tted informs unprecedented Pavilion ##bei ##ference betrayal awaiting leaked V8 puppet adverse Bourne Sunset collectors ##glass ##sque copied Demon conceded resembled Rafe Levy prosecutor ##ject flora manned deaf Mosque reminds Lizzie Products Funny cassette congress ##rong Rover tossing prompting chooses Satellite cautiously Reese ##UT Huang Gloucestershire giggled Kitty ##å Pleasant Aye ##ond judging 1860s intentionally Hurling aggression ##xy transfers employing ##fies ##oda Archibald Blessed Ski flavor Rosie ##burgh sunset Scholarship WC surround ranged ##jay Degree Houses squeezing limb premium Leningrad steals ##inated ##ssie madness vacancy hydraulic Northampton ##prise Marks Boxing ##fying academics ##lich ##TY CDs ##lma hardcore monitors paperback cables Dimitri upside advent Ra ##clusive Aug Christchurch objected stalked Simple colonists ##laid CT discusses fellowship Carnival cares Miracle pastoral rooted shortage borne Quentin meditation tapping Novel ##ades Alicia Burn famed residency Fernández Johannesburg Zhu offended Mao outward ##inas XV denial noticing ##ís quarry ##hound ##amo Bernie Bentley Joanna mortgage ##rdi ##sumption lenses extracted depiction ##RE Networks Broad Revenue flickered virgin flanked ##о Enterprises probable Liberals Falcons drowning phrases loads assumes inhaled awe logs slightest spiders waterfall ##pate rocking shrub ##uil roofs ##gard prehistoric wary ##rak TO clips sustain treason microphone voter Lamb psychologist wrinkled ##ères mating Carrier 340 ##lbert sensing ##rino destiny distract weaker UC Nearly neurons spends Apache ##rem genuinely wells ##lanted stereo ##girl Lois Leaving consul fungi Pier Cyril 80s Jungle ##tani illustration Split ##hana Abigail ##patrick 1787 diminished Selected packaging ##EG Martínez communal Manufacturing sentiment 143 unwilling praising Citation pills ##iti ##rax muffled neatly workforce Yep leisure Tu ##nding Wakefield ancestral ##uki destructive seas Passion showcase ##ceptive heroic 142 exhaustion Customs ##aker Scholar sliced ##inian Direction ##OW Swansea aluminium ##eep ceramic McCoy Career Sector chartered Damascus pictured Interest stiffened Plateau obsolete ##tant irritated inappropriate overs ##nko bail Talent Sur ours ##nah barred legged sociology Bud dictionary ##luk Cover obey ##oring annoying ##dong apprentice Cyrus Role ##GP ##uns ##bag Greenland Porsche Rocket ##32 organism ##ntary reliability ##vocation ##й Found ##hine motors promoter unfair ##oms ##note distribute eminent rails appealing chiefly meaningful Stephan ##rehension Consumer psychiatric bowler saints ##iful ##н 1777 Pol Dorian Townsend hastily ##jima Quincy Sol fascinated Scarlet alto Avon certainty ##eding Keys ##chu Chu ##VE ions tributaries Thanksgiving ##fusion astronomer oxide pavilion Supply Casa Bollywood sadly mutations Keller ##wave nationals ##rgo ##ym predict Catholicism Vega ##eration ##ums Mali tuned Lankan Plans radial Bosnian Lexi ##14 ##ü sacks unpleasant Empty handles ##taking Bon switches intently tuition antique ##jk fraternity notebook Desmond ##sei prostitution ##how deed ##OP 501 Somewhere Rocks ##mons campaigned frigate gases suppress ##hang Merlin Northumberland dominate expeditions thunder ##ups ##rical Cap thorough Ariel ##kind renewable constructing pacing terrorists Bowen documentaries westward ##lass ##nage Merchant ##ued Beaumont Din ##hian Danube peasant Garrison encourages gratitude reminding stormed ##ouse pronunciation ##ailed Weekend suggestions ##ffing ##DI Active Colombo ##logists Merrill ##cens Archaeological Medina captained ##yk duel cracking Wilkinson Guam pickup renovations ##ël ##izer delighted ##iri Weaver ##ctional tens ##hab Clint ##usion ##each petals Farrell ##sable caste ##will Ezra ##qi ##standing thrilled ambush exhaled ##SU Resource blur forearm specifications contingent cafe ##iology Antony fundraising grape ##rgy turnout ##udi Clifton laboratories Irvine ##opus ##lid Monthly Bihar statutory Roses Emil ##rig lumber optimal ##DR pumps plaster Mozambique ##aco nightclub propelled ##hun ked surplus wax ##urai pioneered Sunny imprint Forget Eliot approximate patronage ##bek ##ely ##mbe Partnership curl snapping 29th Patriarch ##jord seldom ##ature astronomy Bremen XIV airborne 205 1778 recognizing stranded arrogant bombardment destined ensured 146 robust Davenport Interactive Offensive Fi prevents probe propeller sorrow Blade mounting automotive ##dged wallet 201 lashes Forrest ##ift Cell Younger shouts ##cki folds ##chet Epic yields homosexual tunes ##minate ##text Manny chemist hindwings ##urn pilgrimage ##sfield ##riff MLS ##rive Huntington translates Path slim ##ndra ##oz climax commuter desperation ##reet denying ##rious daring seminary polo ##clamation Teatro Torah Cats identities Poles photographed fiery popularly ##cross winters Hesse ##vio Nurse Senegal Salon prescribed justify ##gues ##и ##orted HQ ##hiro evaluated momentarily ##unts Debbie ##licity ##TP Mighty Rabbit ##chal Events Savoy ##ht Brandenburg Bordeaux ##laus Release ##IE ##kowski 1900s SK Strauss ##aly Sonia Updated synagogue McKay flattened 370 clutch contests toast evaluate pope heirs jam tutor reverted ##ading nonsense hesitate Lars Ceylon Laurie ##guchi accordingly customary 148 Ethics Multiple instincts IGN ##ä bullshit ##hit ##par desirable ##ducing ##yam alias ashore licenses ##lification misery 147 Cola assassinated fiercely ##aft las goat substrate lords Cass Bridges ICC lasts sights reproductive ##asi Ivory Clean fixing ##lace seeming aide 1850s harassment ##FF ##LE reasonably ##coat ##cano NYC 1784 Fifty immunity Canadians Cheng comforting meanwhile ##tera ##blin breeds glowed ##vour Aden ##verted ##aded ##oral neat enforced poisoning ##ews ##hone enforce predecessors survivor Month unfamiliar pierced waived dump responds Mai Declan angular Doesn interpretations ##yar invest Dhaka policeman Congregation Eighth painfully ##este ##vior Württemberg ##cles blockade encouragement ##fie Caucasus Malone Universidad utilize Nissan inherent 151 agreeing syllable determines Protocol conclude ##gara 40th Xu Taiwanese ##ather boiler printer Lacey titular Klaus Fallon Wembley fox Chandra Governorate obsessed ##Ps micro ##25 Cooke gymnasium weaving Shall Hussein glaring softball Reader Dominion Trouble varsity Cooperation Chaos Kang Kramer Eisenhower proves Connie consortium governors Bethany opener Normally Willy linebacker Regent Used AllMusic Twilight ##shaw Companion Tribunal simpler ##gam Experimental Slovenian cellar deadline trout Hubbard ads idol ##hetto Granada clues salmon 1700 Omega Caldwell softened Bills Honolulu ##gn Terrace suitcase ##IL frantic ##oons Abbot Sitting Fortress Riders sickness enzymes trustee Bern forged ##13 ##ruff ##rl ##versity inspector champagne ##held ##FI hereditary Taliban handball ##wine Sioux ##dicated honoured 139 ##tude Skye meanings ##rkin cardiac analyzed vegetable ##FS Royals dial freelance ##fest partisan petroleum ridden Lincolnshire panting ##comb presidents Haley ##chs contributes Jew discoveries panicked Woody eyelids Fate Tulsa mg whiskey zombies Wii ##udge investigators ##bull centred ##screen Bone Lana ##oise forts ##ske Conan Lyons ##writing SH ##ride rhythmic 154 ##llah pioneers ##bright captivity Sanchez Oman ##mith Flint Platform ##ioned emission packet Persia ##formed takeover tempted Vance Few Toni receptions ##ن exchanges Camille whale Chronicles ##rent ##ushing ##rift Alto Genus ##asing onward foremost longing Rockefeller containers ##cribe intercepted ##olt pleading Bye bee ##umbling 153 undertake Izzy cheaper Ultra validity ##pse Sa hovering ##pert vintage engraved ##rise farmland ##ever ##ifier Atlantis propose Catalonia plunged ##edly demonstrates gig ##cover 156 Osborne cowboy herd investigator loops Burning rests Instrumental embarrassing focal install readings swirling Chatham parameter ##zin ##holders Mandarin Moody converting Escape warnings ##chester incarnation ##ophone adopting ##lins Cromwell ##laws Axis Verde Kappa Schwartz Serbs caliber Wanna Chung ##ality nursery principally Bulletin likelihood logging ##erty Boyle supportive twitched ##usive builds Marseille omitted motif Lands ##lusion ##ssed Barrow Airfield Harmony WWF endured merging convey branding examinations 167 Italians ##dh dude 1781 ##teau crawling thoughtful clasped concluding brewery Moldova Wan Towers Heidelberg 202 ##ict Lagos imposing ##eval ##serve Bacon frowning thirteenth conception calculations ##ович ##mile ##ivated mutation strap ##lund demographic nude perfection stocks ##renched ##dit Alejandro bites fragment ##hack ##rchy GB Surgery Berger punish boiling consume Elle Sid Dome relies Crescent treasurer Bloody 1758 upheld Guess Restaurant signatures font millennium mural stakes Abel hailed insists Alumni Breton ##jun digits ##FM ##thal Talking motive reigning babe masks ##ø Shaun potato sour whitish Somali ##derman ##rab ##wy chancel telecommunications Noise messenger tidal grinding ##ogenic Rebel constituent peripheral recruitment ##ograph ##tler pumped Ravi poked ##gley Olive diabetes discs liking sting fits stir Mari Sega creativity weights Macau mandated Bohemia disastrous Katrina Baku Rajasthan waiter ##psis Siberia verbs ##truction patented 1782 ##ndon Relegated Hunters Greenwood Shock accusing skipped Sessions markers subset monumental Viola comparative Alright Barbados setup Session standardized ##ík ##sket appoint AFB Nationalist ##WS Troop leaped Treasure goodness weary originates 100th compassion expresses recommend 168 composing seventeenth Tex Atlético bald Finding Presidency Sharks favoured inactive ##lter suffix princes brighter ##ctus classics defendants culminated terribly Strategy evenings ##ção ##iver ##urance absorb ##rner Territories RBI soothing Martín concurrently ##tr Nicholson fibers swam ##oney Allie Algerian Dartmouth Mafia ##bos ##tts Councillor vocabulary ##bla ##lé intending ##dler Guerrero sunshine pedal ##TO administrators periodic scholarships Loop Madeline exaggerated ##ressed Regan ##cellular Explorer ##oids Alexandre vows Reporter Unable Average absorption ##bedience Fortunately Auxiliary Grandpa ##HP ##ovo potent temporal adrenaline ##udo confusing guiding Dry qualifications joking wherein heavyweight ##ices nightmares pharmaceutical Commanding ##aled ##ove Gregor ##UP censorship degradation glorious Austro ##rench 380 Miriam sped ##orous offset ##KA fined specialists Pune João ##dina propped fungus ##ς frantically Gabrielle Hare committing ##plied Ask Wilmington stunt numb warmer preacher earnings ##lating integer ##ija federation homosexuality ##cademia epidemic grumbled shoving Milk Satan Tobias innovations ##dington geology memoirs ##IR spared culminating Daphne Focus severed stricken Paige Mans flats Russo communes litigation strengthening ##powered Staffordshire Wiltshire Painting Watkins ##د specializes Select ##rane ##aver Fulton playable ##VN openings sampling ##coon ##21 Allah travelers allocation ##arily Loch ##hm commentators fulfilled ##troke Emeritus Vanderbilt Vijay pledged ##tative diagram drilling ##MD ##plain Edison productivity 31st ##rying ##ption ##gano ##oration ##bara posture bothering platoon politely ##inating redevelopment Job ##vale stark incorrect Mansion renewal threatens Bahamas fridge ##tata Uzbekistan ##edia Sainte ##mio gaps neural ##storm overturned Preservation shields ##ngo ##physics ah gradual killings ##anza consultation premiership Felipe coincidence ##ène ##any Handbook ##loaded Edit Guns arguably ##ş compressed depict seller ##qui Kilkenny ##kling Olympia librarian ##acles dramas JP Kit Maj ##lists proprietary ##nged ##ettes ##tok exceeding Lock induction numerical ##vist Straight foyer imaginary ##pop violinist Carla bouncing ##ashi abolition ##uction restoring scenic ##č Doom overthrow para ##vid ##ughty Concord HC cocaine deputies ##aul visibility ##wart Kapoor Hutchinson ##agan flashes kn decreasing ##ronology quotes vain satisfying ##iam ##linger 310 Hanson fauna ##zawa ##rrel Trenton ##VB Employment vocational Exactly bartender butterflies tow ##chers ##ocks pigs merchandise ##game ##pine Shea ##gration Connell Josephine monopoly ##dled Cobb warships cancellation someday stove ##Cs candidacy superhero unrest Toulouse admiration undergone whirled Reconnaissance costly ##ships 290 Cafe amber Tory ##mpt definitive ##dress proposes redesigned acceleration ##asa ##raphy Presley exits Languages ##cel Mode spokesperson ##tius Ban forthcoming grounded ACC compelling logistics retailers abused ##gating soda ##yland ##lution Landmark XVI blush ##tem hurling dread Tobago Foley ##uad scenarios ##mentation ##rks Score fatigue hairy correspond ##iard defences confiscated ##rudence 1785 Formerly Shot advertised 460 Text ridges Promise Dev exclusion NHS tuberculosis rockets ##offs sparkling 256 disappears mankind ##hore HP ##omo taxation Multi DS Virgil ##ams Dell stacked guessing Jump Nope cheer hates ballots overlooked analyses Prevention maturity dos ##cards ##lect Mare ##yssa Petty ##wning differing iOS ##ior Joachim Sentinel ##nstein 90s Pamela 480 Asher ##lary Vicente landings portray ##rda ##xley Virtual ##uary finances Jain Somebody Tri behave Michele ##ider dwellings FAA Gallagher ##lide Monkey 195 aforementioned ##rism ##bey ##kim ##puted Mesa hopped unopposed recipients Reality Been gritted 149 playground pillar ##rone Guinness ##tad Théâtre depended Tipperary Reuben frightening wooded Target globally ##uted Morales Baptiste drunken Institut characterised ##chemistry Strip discrete Premiership ##zzling gazing Outer ##quisition Sikh Booker ##yal contemporaries Jericho ##chan ##physical ##witch Militia ##rez ##zard dangers ##utter ##₀ Programs darling participates railroads ##ienne behavioral bureau ##rook 161 Hicks ##rises Comes inflicted bees kindness norm ##ković generators ##pard ##omy ##ili methodology Alvin façade latitude ##plified DE Morse ##mered educate intersects ##MF ##cz ##vated AL ##graded ##fill constitutes artery feudal avant cautious ##ogue immigrated ##chenko Saul Clinic Fang choke Cornelius flexibility temperate pins ##erson oddly inequality 157 Natasha Sal ##uter 215 aft blinking ##ntino northward Exposition cookies Wedding impulse Overseas terrifying ##ough Mortimer ##see 440 https og imagining ##cars Nicola exceptionally threads ##cup Oswald Provisional dismantled deserves 1786 Fairy discourse Counsel departing Arc guarding ##orse 420 alterations vibrant Em squinted terrace rowing Led accessories SF Sgt cheating Atomic ##raj Blackpool ##iary boarded substituted bestowed lime kernel ##jah Belmont shaken sticky retrospective Louie migrants weigh sunglasses thumbs ##hoff excavation ##nks Extra Polo motives Drum infrared tastes berth verge ##stand programmed warmed Shankar Titan chromosome cafeteria dividing pepper CPU Stevie satirical Nagar scowled Died backyard ##gata ##reath ##bir Governors portraying ##yah Revenge ##acing 1772 margins Bahn OH lowland ##razed catcher replay ##yoshi Seriously ##licit Aristotle ##ald Habsburg weekday Secretariat CO ##dly ##joy ##stad litre ultra ##cke Mongol Tucson correlation compose traps Groups Hai Salvatore ##dea cents ##eese concession clash Trip Panzer Moroccan cruisers torque Ba grossed ##arate restriction concentrating FDA ##Leod ##ones Scholars ##esi throbbing specialised ##heses Chicken ##fia ##ificant Erich Residence ##trate manipulation namesake ##tom Hoover cue Lindsey Lonely 275 ##HT combustion subscribers Punjabi respects Jeremiah penned ##gor ##rilla suppression ##tration Crimson piston Derry crimson lyrical oversee portrays CF Districts Lenin Cora searches clans VHS ##hel Jacqueline Redskins Clubs desktop indirectly alternatives marijuana suffrage ##smos Irwin ##liff Process ##hawks Sloane ##bson Sonata yielded Flores ##ares armament adaptations integrate neighbours shelters ##tour Skinner ##jet ##tations 1774 Peterborough ##elles ripping Liang Dickinson charities Rwanda monasteries crossover racist barked guerrilla ##ivate Grayson ##iques ##vious ##got Rolls denominations atom affinity ##delity Wish ##inted ##inae interrogation ##cey ##erina ##lifting 192 Sands 1779 mast Likewise ##hyl ##oft contempt ##por assaulted fills establishments Mal consulted ##omi ##sight greet ##roma ##egan Pulitzer ##rried ##dius ##ractical ##voked Hasan CB ##zzy Romanesque Panic wheeled recorder ##tters ##warm ##gly botanist Balkan Lockheed Polly farewell suffers purchases Eaton ##80 Quick commenting Saga beasts hides motifs ##icks Alonso Springer Wikipedia circulated encoding jurisdictions snout UAE Integrated unmarried Heinz ##lein ##figured deleted ##tley Zen Cycling Fuel Scandinavian ##rants Conner reef Marino curiously lingered Gina manners activism Mines Expo Micah promotions Server booked derivatives eastward detailing reelection ##chase 182 Campeonato Po 158 Peel winger ##itch canyon ##pit LDS A1 ##shin Giorgio pathetic ##rga ##mist Aren ##lag confronts motel textbook shine turbines 1770 Darcy ##cot Southeastern ##lessness Banner recognise stray Kitchen paperwork realism Chrysler filmmakers fishermen ##hetic variously Vishnu fiddle Eddy Origin ##tec ##ulin Flames Rs bankrupt Extreme Pomeranian ##emption ratified ##iu jockey Stratford ##ivating ##oire Babylon pardon AI affordable deities disturbance Trying ##sai Ida Papers advancement 70s archbishop Luftwaffe announces tugging ##lphin ##sistence ##eel ##ishes ambition aura ##fled ##lected ##vue Prasad boiled clarity Violin investigative routing Yankee ##uckle McMahon bugs eruption ##rooms Minutes relics ##ckle ##nse sipped valves weakly ##ital Middleton collided ##quer bamboo insignia Tyne exercised Ninth echoing polynomial considerations lunged ##bius objections complain disguised plaza ##VC institutes Judicial ascent imminent Waterford hello Lumpur Niger Goldman vendors Kensington Wren browser ##bner ##tri ##mize ##pis ##lea Cheyenne Bold Settlement Hollow Paralympic axle ##toire ##actic impose perched utilizing slips Benz Michaels manipulate Chiang ##mian Dolphins prohibition attacker ecology Estadio ##SB ##uild attracts recalls glacier lad ##rima Barlow kHz melodic ##aby ##iracy assumptions Cornish ##aru DOS Maddie ##mers lyric Luton nm ##tron Reno Fin YOU Broadcast Finch sensory ##bent Jeep ##uman additionally Buildings businessmen treaties 235 Stranger gateway Charlton accomplishments Diary apologized zinc histories supplier ##tting 162 asphalt Treatment Abbas ##pating ##yres Bloom sedan soloist ##cum antagonist denounced Fairfax ##aving ##enko noticeable Budget Buckingham Snyder retreating Jai spoon invading giggle woven gunfire arrests ##vered ##come respiratory violet ##aws Byrd shocking tenant Jamaican Ottomans Seal theirs ##isse ##48 cooperate peering ##nius 163 Composer organist Mongolian Bauer Spy collects prophecy congregations ##moor Brick calculation fixtures exempt ##dden Ada Thousand ##lue tracing ##achi bodyguard vicar supplying Łódź interception monitored ##heart Paso overlap annoyance ##dice yellowish stables elders illegally honesty ##oar skinny spinal ##puram Bourbon ##cor flourished Medium ##stics ##aba Follow ##ckey stationary ##scription dresser scrutiny Buckley Clearly ##SF Lyrics ##heimer drying Oracle internally rains ##last Enemy ##oes McLean Ole phosphate Rosario Rifles ##mium battered Pepper Presidents conquer Château castles ##aldo ##ulf Depending Lesser Boom trades Peyton 164 emphasize accustomed SM Ai Classification ##mins ##35 ##rons leak piled deeds lush ##self beginnings breathless 1660 McGill ##ago ##chaft ##gies humour Bomb securities Might ##zone ##eves Matthias Movies Levine vengeance ##ads Challenger Misty Traditionally constellation ##rass deepest workplace ##oof ##vina impatient ##ML Mughal Alessandro scenery Slater postseason troupe ##ń Volunteers Facility militants Reggie sanctions Expeditionary Nam countered interpret Basilica coding expectation Duffy def Tong wakes Bowling Vehicle Adler salad intricate stronghold medley ##uries ##bur joints ##rac ##yx ##IO Ordnance Welch distributor Ark cavern trench Weiss Mauritius decreases docks eagerly irritation Matilda biographer Visiting ##marked ##iter ##ear ##gong Moreno attendant Bury instrumentation theologian clit nuns symphony translate 375 loser ##user ##VR ##meter ##orious harmful ##yuki Commissioners Mendoza sniffed Hulk ##dded ##ulator ##nz Donnell ##eka deported Met SD Aerospace ##cultural ##odes Fantastic cavity remark emblem fearing ##iance ICAO Liberia stab ##yd Pac Gymnasium IS Everton ##vanna mantle ##ief Ramon ##genic Shooting Smoke Random Africans MB tavern bargain voluntarily Ion Peoples Rusty attackers Patton sins ##cake Hat moderately ##hala ##alia requesting mechanic ##eae Seine Robbins ##ulum susceptible Bravo Slade Strasbourg rubble entrusted Creation ##amp smoothed ##uintet evenly reviewers skip Sculpture 177 Rough ##rrie Reeves ##cede Administrator garde minus carriages grenade Ninja fuscous ##kley Punk contributors Aragon Tottenham ##cca ##sir VA laced dealers ##sonic crisp harmonica Artistic Butch Andes Farmers corridors unseen ##tium Countries Lone envisioned Katy ##lang ##cc Quarterly ##neck consort ##aceae bidding Corey concurrent ##acts ##gum Highness ##lient ##rators arising ##unta pathways 49ers bolted complaining ecosystem libretto Ser narrated 212 Soft influx ##dder incorporation plagued tents ##ddled 1750 Risk citation Tomas hostilities seals Bruins Dominique attic competent ##UR ##cci hugging Breuning bacterial Shrewsbury vowed eh elongated hangs render centimeters ##ficient Mu turtle besieged ##gaard grapes bravery collaborations deprived ##amine ##using ##gins arid ##uve coats hanged ##sting Pa prefix ##ranged Exit Chain Flood Materials suspicions ##ö hovered Hidden ##state Malawi ##24 Mandy norms fascinating airlines delivers ##rust Cretaceous spanned pillows ##onomy jar ##kka regent fireworks morality discomfort lure uneven ##jack Lucian 171 archaeology ##til mornings Billie Marquess impending spilling tombs ##volved Celia Coke underside ##bation Vaughn Daytona Godfrey Pascal Alien ##sign 172 ##lage iPhone Gonna genocide ##rber oven endure dashed simultaneous ##phism Wally ##rō ants predator reissue ##aper Speech funk Rudy claw Hindus Numbers Bing lantern ##aurus scattering poisoned ##active Andrei algebraic baseman ##ritz Gregg ##cola selections ##putation lick Laguna ##IX Sumatra Warning turf buyers Burgess Oldham exploit worm initiate strapped tuning filters haze ##е ##ledge ##ydro ##culture amendments Promotion ##union Clair ##uria petty shutting ##eveloped Phoebe Zeke conducts grains clashes ##latter illegitimate willingly Deer Lakers Reference chaplain commitments interrupt salvation Panther Qualifying Assessment cancel efficiently attorneys Dynamo impress accession clinging randomly reviewing Romero Cathy charting clapped rebranded Azerbaijani coma indicator punches ##tons Sami monastic prospects Pastor ##rville electrified ##CI ##utical tumbled Chef muzzle selecting UP Wheel protocols ##tat Extended beautifully nests ##stal Andersen ##anu ##³ ##rini kneeling ##reis ##xia anatomy dusty Safe turmoil Bianca ##elo analyze ##ر ##eran podcast Slovene Locke Rue ##retta ##uni Person Prophet crooked disagreed Versailles Sarajevo Utrecht ##ogen chewing ##ception ##iidae Missile attribute majors Arch intellectuals ##andra ideological Cory Salzburg ##fair Lot electromagnetic Distribution ##oper ##pered Russ Terra repeats fluttered Riga ##ific ##gt cows Hair labelled protects Gale Personnel Düsseldorf Moran rematch ##OE Slow forgiveness ##ssi proudly Macmillan insist undoubtedly Québec Violence ##yuan ##aine mourning linen accidental ##iol ##arium grossing lattice maneuver ##marine prestige petrol gradient invasive militant Galerie widening ##aman ##quist disagreement ##ales creepy remembers buzz ##erial Exempt Dirk mon Addison ##inen deposed ##agon fifteenth Hang ornate slab ##lades Fountain contractors das Warwickshire 1763 ##rc Carly Essays Indy Ligue greenhouse slit ##sea chewed wink ##azi Playhouse ##kon Gram Ko Samson creators revive ##rians spawned seminars Craft Tall diverted assistants computational enclosure ##acity Coca ##eve databases Drop ##loading ##hage Greco Privy entrances pork prospective Memories robes ##market transporting ##lik Rudolph Horton visually ##uay ##nja Centro Tor Howell ##rsey admitting postgraduate herbs ##att Chin Rutherford ##bot ##etta Seasons explanations ##bery Friedman heap ##ryl ##sberg jaws ##agh Choi Killing Fanny ##suming ##hawk hopeful ##aid Monty gum remarkably Secrets disco harp advise ##avia Marathi ##cycle Truck abbot sincere urine ##mology masked bathing ##tun Fellows ##TM ##gnetic owl ##jon hymn ##leton 208 hostility ##cée baked Bottom ##AB shudder ##ater ##von ##hee reorganization Cycle ##phs Lex ##style ##rms Translation ##erick ##imeter ##ière attested Hillary ##DM gal wander Salle ##laming Perez Pit ##LP USAF contexts Disease blazing aroused razor walled Danielle Mont Funk royalty thee 203 donors ##erton famously processors reassigned welcoming Goldberg ##quities undisclosed Orient Patty vaccine refrigerator Cypriot consonant ##waters 176 sober ##lement Racecourse ##uate Luckily Selection conceptual vines Breaking wa lions oversight sheltered Dancer ponds borrow ##BB ##pulsion Daly ##eek fertility spontaneous Worldwide gasping ##tino 169 ABS Vickers ambient energetic prisons ##eson Stacy ##roach GmbH Afro Marin farmhouse pinched ##cursion ##sp Sabine ##pire 181 nak swelling humble perfume ##balls Rai cannons ##taker Married Maltese canals interceptions hats lever slowing ##ppy Nike Silas Scarborough skirts 166 inauguration Shuttle alloy beads belts Compton Cause battling critique surf Dock roommate ##ulet invade Garland ##slow nutrition persona ##zam Wichita acquaintance coincided ##cate Dracula clamped ##gau overhaul ##broken ##rrier melodies ventures Paz convex Roots ##holding Tribute transgender ##ò chimney ##riad Ajax Thereafter messed nowadays pH ##100 ##alog Pomerania ##yra Rossi glove ##TL Races ##asily tablets Jase ##ttes diner ##rns Hu Mohan anytime weighted remixes Dove cherry imports ##urity GA ##TT ##iated ##sford Clarkson evidently rugged Dust siding ##ometer acquitted choral ##mite infants Domenico gallons Atkinson gestures slated ##xa Archaeology unwanted ##ibes ##duced premise Colby Geelong disqualified ##pf ##voking simplicity Walkover Qaeda Warden ##bourg ##ān Invasion Babe harness 183 ##tated maze Burt bedrooms ##nsley Horizon ##oast minimize peeked MLA Trains tractor nudged ##iform Growth Benton separates ##about ##kari buffer anthropology brigades foil ##wu Domain licking whore ##rage ##sham Initial Courthouse Rutgers dams villains supermarket ##brush Brunei Palermo arises Passenger outreach ##gill Labrador McLaren ##uy Lori ##fires Heads magistrate ¹⁄₂ Weapons ##wai ##roke projecting ##ulates bordering McKenzie Pavel midway Guangzhou streamed racer ##lished eccentric spectral 206 ##mism Wilde Grange preparatory lent ##tam starving Gertrude ##cea ##ricted Breakfast Mira blurted derive ##lair blunt sob Cheltenham Henrik reinstated intends ##istan unite ##ector playful sparks mapped Cadet luggage prosperous ##ein salon ##utes Biological ##rland Tyrone buyer ##lose amounted Saw smirked Ronan Reviews Adele trait ##proof Bhutan Ginger ##junct digitally stirring ##isted coconut Hamlet Dinner Scale pledge ##RP Wrong Goal Panel therapeutic elevations infectious priesthood ##inda Guyana diagnostic ##mbre Blackwell sails ##arm literal periodically gleaming Robot Rector ##abulous ##tres Reaching Romantic CP Wonderful ##tur ornamental ##nges traitor ##zilla genetics mentioning ##eim resonance Areas Shopping ##nard Gail Solid ##rito ##mara Willem Chip Matches Volkswagen obstacle Organ invites Coral attain ##anus ##dates Midway shuffled Cecilia dessert Gateway Ch Napoleonic Petroleum jets goose striped bowls vibration Sims nickel Thirteen problematic intervene ##grading ##unds Mum semifinal Radical ##izations refurbished ##sation ##harine Maximilian cites Advocate Potomac surged preserves Curry angled ordination ##pad Cade ##DE ##sko researched torpedoes Resident wetlands hay applicants depart Bernstein ##pic ##ario ##rae favourable ##wari ##р metabolism nobleman Defaulted calculate ignition Celebrity Belize sulfur Flat Sc USB flicker Hertfordshire Sept CFL Pasadena Saturdays Titus ##nir Canary Computing Isaiah ##mler formidable pulp orchid Called Solutions kilograms steamer ##hil Doncaster successors Stokes Holstein ##sius sperm API Rogue instability Acoustic ##rag 159 undercover Wouldn ##pra ##medical Eliminated honorable ##chel denomination abrupt Buffy blouse fi Regardless Subsequent ##rdes Lover ##tford bacon ##emia carving ##cripts Massacre Ramos Latter ##ulp ballroom ##gement richest bruises Rest Wiley ##aster explosions ##lastic Edo ##LD Mir choking disgusted faintly Barracks blasted headlights Tours ensued presentations ##cale wrought ##oat ##coa Quaker ##sdale recipe ##gny corpses ##liance comfortably ##wat Landscape niche catalyst ##leader Securities messy ##RL Rodrigo backdrop ##opping treats Emilio Anand bilateral meadow VC socialism ##grad clinics ##itating ##ppe ##ymphonic seniors Advisor Armoured Method Alley ##orio Sad fueled raided Axel NH rushes Dixie Otis wrecked ##22 capitalism café ##bbe ##pion ##forcing Aubrey Lublin Whenever Sears Scheme ##lana Meadows treatise ##RI ##ustic sacrifices sustainability Biography mystical Wanted multiplayer Applications disliked ##tisfied impaired empirical forgetting Fairfield Sunni blurred Growing Avalon coil Camera Skin bruised terminals ##fted ##roving Commando ##hya ##sper reservations needles dangling ##rsch ##rsten ##spect ##mbs yoga regretted Bliss Orion Rufus glucose Olsen autobiographical ##dened 222 humidity Shan ##ifiable supper ##rou flare ##MO campaigning descend socio declares Mounted Gracie Arte endurance ##ety Copper costa airplay ##MB Proceedings dislike grimaced occupants births glacial oblivious cans installment muddy ##ł captains pneumonia Quiet Sloan Excuse ##nine Geography gymnastics multimedia drains Anthology Gear cylindrical Fry undertaking ##pler ##tility Nan ##recht Dub philosophers piss Atari ##pha Galicia México ##nking Continuing bump graveyard persisted Shrine ##erapy defects Advance Bomber ##oil ##ffling cheerful ##lix scrub ##eto awkwardly collaborator fencing ##alo prophet Croix coughed ##lication roadway slaughter elephants ##erated Simpsons vulnerability ivory Birth lizard scarce cylinders fortunes ##NL Hate Priory ##lai McBride ##copy Lenny liaison Triangle coronation sampled savage amidst Grady whatsoever instinctively Reconstruction insides seizure Drawing ##rlin Antioch Gao Díaz 1760 Sparks ##tien ##bidae rehearsal ##bbs botanical ##hers compensate wholesale Seville shareholder prediction astronomical Reddy hardest circling whereabouts termination Rep Assistance Dramatic Herb ##ghter climbs 188 Poole 301 ##pable wit ##istice Walters relying Jakob ##redo proceeding Langley affiliates ou ##allo ##holm Samsung ##ishi Missing Xi vertices Claus foam restless ##uating ##sso ##ttering Philips delta bombed Catalogue coaster Ling Willard satire 410 Composition Net Orioles ##ldon fins Palatinate Woodward tease tilt brightness ##70 ##bbling ##loss ##dhi ##uilt Whoever ##yers hitter Elton Extension ace Affair restructuring ##loping Paterson hi ##rya spouse Shay Himself piles preaching ##gical bikes Brave expulsion Mirza stride Trees commemorated famine masonry Selena Watt Banking Rancho Stockton dip tattoos Vlad acquainted Flyers ruthless fourteenth illustrate ##akes EPA ##rows ##uiz bumped Designed Leaders mastered Manfred swirled McCain ##rout Artemis rabbi flinched upgrades penetrate shipyard transforming caretaker ##eiro Maureen tightening ##founded RAM ##icular ##mper ##rung Fifteen exploited consistency interstate ##ynn Bridget contamination Mistress ##rup coating ##FP ##jective Libyan 211 Gemma dependence shrubs ##ggled Germain retaliation traction ##PP Dangerous terminology psychiatrist ##garten hurdles Natal wasting Weir revolves stripe ##reased preferences ##entation ##lde ##áil ##otherapy Flame ##ologies viruses Label Pandora veil ##ogical Coliseum Cottage creeping Jong lectured ##çaise shoreline ##fference ##hra Shade Clock Faye bilingual Humboldt Operating ##fter ##was algae towed amphibious Parma impacted smacked Piedmont Monsters ##omb Moor ##lberg sinister Postal 178 Drummond Sign textbooks hazardous Brass Rosemary Pick Sit Architect transverse Centennial confess polling ##aia Julien ##mand consolidation Ethel ##ulse severity Yorker choreographer 1840s ##ltry softer versa ##geny ##quila ##jō Caledonia Friendship Visa rogue ##zzle bait feather incidence Foods Ships ##uto ##stead arousal ##rote Hazel ##bolic Swing ##ej ##cule ##jana ##metry ##uity Valuable ##ₙ Shropshire ##nect 365 Ones realise Café Albuquerque ##grown ##stadt 209 ##ᵢ prefers withstand Lillian MacArthur Hara ##fulness domination ##VO ##school Freddy ethnicity ##while adorned hormone Calder Domestic Freud Shields ##phus ##rgan BP Segunda Mustang ##GI Bonn patiently remarried ##umbria Crete Elephant Nuremberg tolerate Tyson ##evich Programming ##lander Bethlehem segregation Constituency quarterly blushed photographers Sheldon porcelain Blanche goddamn lively ##fused bumps ##eli curated coherent provoked ##vet Madeleine ##isco rainy Bethel accusation ponytail gag ##lington quicker scroll ##vate Bow Gender Ira crashes ACT Maintenance ##aton ##ieu bitterly strains rattled vectors ##arina ##ishly 173 parole ##nx amusing Gonzalez ##erative Caucus sensual Penelope coefficient Mateo ##mani proposition Duty lacrosse proportions Plato profiles Botswana Brandt reins mandolin encompassing ##gens Kahn prop summon ##MR ##yrian ##zaki Falling conditional thy ##bao ##ych radioactive ##nics Newspaper ##people ##nded Gaming sunny ##look Sherwood crafted NJ awoke 187 timeline giants possessing ##ycle Cheryl ng Ruiz polymer potassium Ramsay relocation ##leen Sociology ##bana Franciscan propulsion denote ##erjee registers headline Tests emerges Articles Mint livery breakup kits Rap Browning Bunny ##mington ##watch Anastasia Zachary arranging biographical Erica Nippon ##membrance Carmel ##sport ##xes Paddy ##holes Issues Spears compliment ##stro ##graphs Castillo ##MU ##space Corporal ##nent 174 Gentlemen ##ilize ##vage convinces Carmine Crash ##hashi Files Doctors brownish sweating goats ##conductor rendition ##bt NL ##spiration generates ##cans obsession ##noy Danger Diaz heats Realm priorities ##phon 1300 initiation pagan bursts archipelago chloride Screenplay Hewitt Khmer bang judgement negotiating ##ait Mabel densely Boulder knob 430 Alfredo ##kt pitches ##ées ##ان Macdonald ##llum imply ##mot Smile spherical ##tura Derrick Kelley Nico cortex launches differed parallels Navigation ##child ##rming canoe forestry reinforce ##mote confirming tasting scaled ##resh ##eting Understanding prevailing Pearce CW earnest Gaius asserts denoted landmarks Chargers warns ##flies Judges jagged ##dain tails Historian Millie ##sler 221 ##uard absurd Dion ##ially makeshift Specifically ignorance Eat ##ieri comparisons forensic 186 Giro skeptical disciplinary battleship ##45 Libby 520 Odyssey ledge ##post Eternal Missionary deficiency settler wonders ##gai raging ##cis Romney Ulrich annexation boxers sect 204 ARIA dei Hitchcock te Varsity ##fic CC lending ##nial ##tag ##rdy ##obe Defensive ##dson ##pore stellar Lam Trials contention Sung ##uminous Poe superiority ##plicate 325 bitten conspicuous ##olly Lila Pub Petit distorted ISIL distinctly ##family Cowboy mutant ##cats ##week Changes Sinatra epithet neglect Innocent gamma thrill reggae ##adia ##ational ##due landlord ##leaf visibly ##ì Darlington Gomez ##iting scarf ##lade Hinduism Fever scouts ##roi convened ##oki 184 Lao boycott unemployed ##lore ##ß ##hammer Curran disciples odor ##ygiene Lighthouse Played whales discretion Yves ##ceived pauses coincide ##nji dizzy ##scopic routed Guardians Kellan carnival nasal 224 ##awed Mitsubishi 640 Cast silky Projects joked Huddersfield Rothschild zu ##olar Divisions mildly ##eni ##lge Appalachian Sahara pinch ##roon wardrobe ##dham ##etal Bubba ##lini ##rumbling Communities Poznań unification Beau Kris SV Rowing Minh reconciliation ##saki ##sor taped ##reck certificates gubernatorial rainbow ##uing litter ##lique ##oted Butterfly benefited Images induce Balkans Velvet ##90 ##xon Bowman ##breaker penis ##nitz ##oint ##otive crust ##pps organizers Outdoor nominees ##rika TX ##ucks Protestants ##imation appetite Baja awaited ##points windshield ##igh ##zled Brody Buster stylized Bryce ##sz Dollar vest mold ounce ok receivers ##uza Purdue Harrington Hodges captures ##ggio Reservation ##ssin ##tman cosmic straightforward flipping remixed ##athed Gómez Lim motorcycles economies owning Dani ##rosis myths sire kindly 1768 Bean graphs ##mee ##RO ##geon puppy Stephenson notified ##jer Watching ##rama Sino urgency Islanders ##mash Plata fumble ##chev ##stance ##rack ##she facilitated swings akin enduring payload ##phine Deputies murals ##tooth 610 Jays eyeing ##quito transparency ##cote Timor negatively ##isan battled ##fected thankful Rage hospitality incorrectly 207 entrepreneurs ##cula ##wley hedge ##cratic Corpus Odessa Whereas ##ln fetch happier Amherst bullying graceful Height Bartholomew willingness qualifier 191 Syed Wesleyan Layla ##rrence Webber ##hum Rat ##cket ##herence Monterey contaminated Beside Mustafa Nana 213 ##pruce Reason ##spense spike ##gé AU disciple charcoal ##lean formulated Diesel Mariners accreditation glossy 1800s ##ih Mainz unison Marianne shear overseeing vernacular bowled ##lett unpopular ##ckoned ##monia Gaston ##TI ##oters Cups ##bones ##ports Museo minors 1773 Dickens ##EL ##NBC Presents ambitions axes Río Yukon bedside Ribbon Units faults conceal ##lani prevailed 214 Goodwin Jaguar crumpled Cullen Wireless ceded remotely Bin mocking straps ceramics ##avi ##uding ##ader Taft twenties ##aked Problem quasi Lamar ##ntes ##avan Barr ##eral hooks sa ##ône 194 ##ross Nero Caine trance Homeland benches Guthrie dismiss ##lex César foliage ##oot ##alty Assyrian Ahead Murdoch dictatorship wraps ##ntal Corridor Mackay respectable jewels understands ##pathic Bryn ##tep ON capsule intrigued Sleeping communists ##chayat ##current ##vez doubling booklet ##uche Creed ##NU spies ##sef adjusting 197 Imam heaved Tanya canonical restraint senators stainless ##gnate Matter cache restrained conflicting stung ##ool Sustainable antiquity 193 heavens inclusive ##ador fluent 303 911 archaeologist superseded ##plex Tammy inspire ##passing ##lub Lama Mixing ##activated ##yote parlor tactic 198 Stefano prostitute recycling sorted banana Stacey Musée aristocratic cough ##rting authorised gangs runoff thoughtfully ##nish Fisheries Provence detector hum ##zhen pill ##árez Map Leaves Peabody skater vent ##color 390 cerebral hostages mare Jurassic swell ##isans Knoxville Naked Malaya scowl Cobra ##anga Sexual ##dron ##iae 196 ##drick Ravens Blaine ##throp Ismail symmetric ##lossom Leicestershire Sylvester glazed ##tended Radar fused Families Blacks Sale Zion foothills microwave slain Collingwood ##pants ##dling killers routinely Janice hearings ##chanted ##ltration continents ##iving ##yster ##shot ##yna injected Guillaume ##ibi kinda Confederacy Barnett disasters incapable ##grating rhythms betting draining ##hak Callie Glover ##iliated Sherlock hearted punching Wolverhampton Leaf Pi builders furnished knighted Photo ##zle Touring fumbled pads ##ий Bartlett Gunner eerie Marius Bonus pots ##hino ##pta Bray Frey Ortiz stalls belongings Subway fascination metaphor Bat Boer Colchester sway ##gro rhetoric ##dheim Fool PMID admire ##hsil Strand TNA ##roth Nottinghamshire ##mat ##yler Oxfordshire ##nacle ##roner BS ##nces stimulus transports Sabbath ##postle Richter 4000 ##grim ##shima ##lette deteriorated analogous ##ratic UHF energies inspiring Yiddish Activities ##quential ##boe Melville ##ilton Judd consonants labs smuggling ##fari avid ##uc truce undead ##raith Mostly bracelet Connection Hussain awhile ##UC ##vention liable genetically ##phic Important Wildcats daddy transmit ##cas conserved Yesterday ##lite Nicky Guys Wilder Lay skinned Communists Garfield Nearby organizer Loss crafts walkway Chocolate Sundance Synod ##enham modify swayed Surface analysts brackets drone parachute smelling Andrés filthy frogs vertically ##OK localities marries AHL 35th ##pian Palazzo cube dismay relocate ##на Hear ##digo ##oxide prefecture converts hangar ##oya ##ucking Spectrum deepened spoiled Keeping ##phobic Verona outrage Improvement ##UI masterpiece slung Calling chant Haute mediated manipulated affirmed ##hesis Hangul skies ##llan Worcestershire ##kos mosaic ##bage ##wned Putnam folder ##LM guts noteworthy ##rada AJ sculpted ##iselle ##rang recognizable ##pent dolls lobbying impatiently Se staple Serb tandem Hiroshima thieves ##ynx faculties Norte ##alle ##trusion chords ##ylon Gareth ##lops ##escu FIA Levin auspices groin Hui nun Listed Honourable Larsen rigorous ##erer Tonga ##pment ##rave ##track ##aa ##enary 540 clone sediment esteem sighted cruelty ##boa inverse violating Amtrak Status amalgamated vertex AR harmless Amir mounts Coronation counseling Audi CO₂ splits ##eyer Humans Salmon ##have ##rado ##čić 216 takeoff classmates psychedelic ##gni Gypsy 231 Anger GAA ME ##nist ##tals Lissa Odd baptized Fiat fringe ##hren 179 elevators perspectives ##TF ##ngle Question frontal 950 thicker Molecular ##nological Sixteen Baton Hearing commemorative dorm Architectural purity ##erse risky Georgie relaxing ##ugs downed ##rar Slim ##phy IUCN ##thorpe Parkinson 217 Marley Shipping sweaty Jesuits Sindh Janata implying Armenians intercept Ankara commissioners ascended sniper Grass Walls salvage Dewey generalized learnt PT ##fighter ##tech DR ##itrus ##zza mercenaries slots ##burst ##finger ##nsky Princes Rhodesia ##munication ##strom Fremantle homework ins ##Os ##hao ##uffed Thorpe Xiao exquisite firstly liberated technician Oilers Phyllis herb sharks MBE ##stock Product banjo ##morandum ##than Visitors unavailable unpublished oxidation Vogue ##copic ##etics Yates ##ppard Leiden Trading cottages Principles ##Millan ##wife ##hiva Vicar nouns strolled ##eorological ##eton ##science precedent Armand Guido rewards ##ilis ##tise clipped chick ##endra averages tentatively 1830s ##vos Certainly 305 Société Commandant ##crats ##dified ##nka marsh angered ventilation Hutton Ritchie ##having Eclipse flick motionless Amor Fest Loire lays ##icit ##sband Guggenheim Luck disrupted ##ncia Disco ##vigator criticisms grins ##lons ##vial ##ody salute Coaches junk saxophonist ##eology Uprising Diet ##marks chronicles robbed ##iet ##ahi Bohemian magician wavelength Kenyan augmented fashionable ##ogies Luce F1 Monmouth ##jos ##loop enjoyment exemption Centers ##visor Soundtrack blinding practitioner solidarity sacrificed ##oso ##cture ##riated blended Abd Copyright ##nob 34th ##reak Claudio hectare rotor testify ##ends ##iably ##sume landowner ##cess ##ckman Eduard Silesian backseat mutually ##abe Mallory bounds Collective Poet Winkler pertaining scraped Phelps crane flickering Proto bubbles popularized removes ##86 Cadillac Warfare audible rites shivering ##sist ##nst ##biotic Mon fascist Bali Kathryn ambiguous furiously morale patio Sang inconsistent topology Greens monkeys Köppen 189 Toy vow ##ías bombings ##culus improvised lodged subsidiaries garment startling practised Hume Thorn categorized Till Eileen wedge ##64 Federico patriotic unlock ##oshi badminton Compared Vilnius ##KE Crimean Kemp decks spaced resolutions sighs ##mind Imagine Cartoon huddled policemen forwards ##rouch equals ##nter inspected Charley MG ##rte pamphlet Arturo dans scarcely ##ulton ##rvin parental unconstitutional watts Susannah Dare ##sitive Rowland Valle invalid ##ué Detachment acronym Yokohama verified ##lsson groove Liza clarified compromised 265 ##rgon ##orf hesitant Fruit Application Mathias icons ##cell Qin interventions ##uron punt remnant ##rien Ames manifold spines floral ##zable comrades Fallen orbits Annals hobby Auditorium implicated researching Pueblo Ta terminate ##pella Rings approximation fuzzy ##ús thriving ##ket Conor alarmed etched Cary ##rdon Ally ##rington Pay mint ##hasa ##unity ##dman ##itate Oceania furrowed trams ##aq Wentworth ventured choreography prototypes Patel mouthed trenches ##licing ##yya Lies deception ##erve ##vations Bertrand earthquakes ##tography Southwestern ##aja token Gupta ##yō Beckett initials ironic Tsar subdued shootout sobbing liar Scandinavia Souls ch therapist trader Regulation Kali busiest ##pation 32nd Telephone Vargas ##moky ##nose ##uge Favorite abducted bonding 219 255 correction mat drown fl unbeaten Pocket Summers Quite rods Percussion ##ndy buzzing cadet Wilkes attire directory utilities naive populous Hendrix ##actor disadvantage 1400 Landon Underworld ##ense Occasionally mercury Davey Morley spa wrestled ##vender eclipse Sienna supplemented thou Stream liturgical ##gall ##berries ##piration 1769 Bucks abandoning ##jutant ##nac 232 venom ##31 Roche dotted Currie Córdoba Milo Sharif divides justification prejudice fortunate ##vide ##ābād Rowe inflammatory ##eld avenue Sources ##rimal Messenger Blanco advocating formulation ##pute emphasizes nut Armored ##ented nutrients ##tment insistence Martins landowners ##RB comparatively headlines snaps ##qing Celebration ##mad republican ##NE Trace ##500 1771 proclamation NRL Rubin Buzz Weimar ##AG 199 posthumous ##ental ##deacon Distance intensely overheard Arcade diagonal hazard Giving weekdays ##ù Verdi actresses ##hare Pulling ##erries ##pores catering shortest ##ctors ##cure ##restle ##reta ##runch ##brecht ##uddin Moments senate Feng Prescott ##thest 218 divisional Bertie sparse surrounds coupling gravitational werewolves ##lax Rankings ##mated ##tries Shia ##mart ##23 ##vocative interfaces morphology newscast ##bide inputs solicitor Olaf cabinets puzzles ##tains Unified ##firmed WA solemn ##opy Tito Jaenelle Neolithic horseback ##ires pharmacy prevalence ##lint Swami ##bush ##tudes Philipp mythical divers Scouting aperture progressively ##bay ##nio bounce Floor ##elf Lucan adulthood helm Bluff Passage Salvation lemon napkin scheduling ##gets Elements Mina Novak stalled ##llister Infrastructure ##nky ##tania ##uished Katz Norma sucks trusting 1765 boilers Accordingly ##hered 223 Crowley ##fight ##ulo Henrietta ##hani pounder surprises ##chor ##glia Dukes ##cracy ##zier ##fs Patriot silicon ##VP simulcast telegraph Mysore cardboard Len ##QL Auguste accordion analytical specify ineffective hunched abnormal Transylvania ##dn ##tending Emilia glittering Maddy ##wana 1762 External Lecture endorsement Hernández Anaheim Ware offences ##phorus Plantation popping Bonaparte disgusting neared ##notes Identity heroin nicely ##raverse apron congestion ##PR padded ##fts invaders ##came freshly Halle endowed fracture ROM ##max sediments diffusion dryly ##tara Tam Draw Spin Talon Anthropology ##lify nausea ##shirt insert Fresno capitalist indefinitely apples Gift scooped 60s Cooperative mistakenly ##lover murmur ##iger Equipment abusive orphanage ##9th ##lterweight ##unda Baird ant saloon 33rd Chesapeake ##chair ##sound ##tend chaotic pornography brace ##aret heiress SSR resentment Arbor headmaster ##uren unlimited ##with ##jn Bram Ely Pokémon pivotal ##guous Database Marta Shine stumbling ##ovsky ##skin Henley Polk functioned ##layer ##pas ##udd ##MX blackness cadets feral Damian ##actions 2D ##yla Apocalypse ##aic inactivated ##china ##kovic ##bres destroys nap Macy sums Madhya Wisdom rejects ##amel 60th Cho bandwidth ##sons ##obbing ##orama Mutual shafts ##estone ##rsen accord replaces waterfront ##gonal ##rida convictions ##ays calmed suppliers Cummings GMA fearful Scientist Sinai examines experimented Netflix Enforcement Scarlett ##lasia Healthcare ##onte Dude inverted ##36 ##regation ##lidae Munro ##angay Airbus overlapping Drivers lawsuits bodily ##udder Wanda Effects Fathers ##finery ##islav Ridley observatory pod ##utrition Electricity landslide ##mable ##zoic ##imator ##uration Estates sleepy Nickelodeon steaming irony schedules snack spikes Hmm ##nesia ##bella ##hibit Greenville plucked Harald ##ono Gamma infringement roaring deposition ##pol ##orum 660 seminal passports engagements Akbar rotated ##bina ##gart Hartley ##lown ##truct uttered traumatic Dex ##ôme Holloway MV apartheid ##nee Counter Colton OR 245 Spaniards Regency Schedule scratching squads verify ##alk keyboardist rotten Forestry aids commemorating ##yed ##érie Sting ##elly Dai ##fers ##berley ##ducted Melvin cannabis glider ##enbach ##rban Costello Skating cartoonist AN audit ##pectator distributing 226 312 interpreter header Alternatively ##ases smug ##kumar cabins remastered Connolly Kelsey LED tentative Check Sichuan shaved ##42 Gerhard Harvest inward ##rque Hopefully hem ##34 Typical binds wrath Woodstock forcibly Fergus ##charged ##tured prepares amenities penetration ##ghan coarse ##oned enthusiasts ##av ##twined fielded ##cky Kiel ##obia 470 beers tremble youths attendees ##cademies ##sex Macon communism dir ##abi Lennox Wen differentiate jewel ##SO activate assert laden unto Gillespie Guillermo accumulation ##GM NGO Rosenberg calculating drastically ##omorphic peeled Liège insurgents outdoors ##enia Aspen Sep awakened ##eye Consul Maiden insanity ##brian furnace Colours distributions longitudinal syllables ##scent Martian accountant Atkins husbands sewage zur collaborate highlighting ##rites ##PI colonization nearer ##XT dunes positioning Ku multitude luxurious Volvo linguistics plotting squared ##inder outstretched ##uds Fuji ji ##feit ##ahu ##loat ##gado ##luster ##oku América ##iza Residents vine Pieces DD Vampires ##ová smoked harshly spreads ##turn ##zhi betray electors ##settled Considering exploits stamped Dusty enraged Nairobi ##38 intervened ##luck orchestras ##lda Hereford Jarvis calf ##itzer ##CH salesman Lovers cigar Angelica doomed heroine ##tible Sanford offenders ##ulously articulated ##oam Emanuel Gardiner Edna Shu gigantic ##stable Tallinn coasts Maker ale stalking ##oga ##smus lucrative southbound ##changing Reg ##lants Schleswig discount grouping physiological ##OH ##sun Galen assurance reconcile rib scarlet Thatcher anarchist ##oom Turnpike ##ceding cocktail Sweeney Allegheny concessions oppression reassuring ##poli ##ticus ##TR ##VI ##uca ##zione directional strikeouts Beneath Couldn Kabul ##national hydroelectric ##jit Desire ##riot enhancing northbound ##PO Ok Routledge volatile Bernardo Python 333 ample chestnut automobiles ##innamon ##care ##hering BWF salaries Turbo acquisitions ##stituting strengths pilgrims Ponce Pig Actors Beard sanitation ##RD ##mett Telecommunications worms ##idas Juno Larson Ventura Northeastern weighs Houghton collaborating lottery ##rano Wonderland gigs ##lmer ##zano ##edd ##nife mixtape predominant tripped ##ruly Alexei investing Belgarath Brasil hiss ##crat ##xham Côte 560 kilometer ##cological analyzing ##As engined listener ##cakes negotiation ##hisky Santana ##lemma IAAF Seneca skeletal Covenant Steiner ##lev ##uen Neptune retention ##upon Closing Czechoslovak chalk Navarre NZ ##IG ##hop ##oly ##quatorial ##sad Brewery Conflict Them renew turrets disagree Petra Slave ##reole adjustment ##dela ##regard ##sner framing stature ##rca ##sies ##46 ##mata Logic inadvertently naturalist spheres towering heightened Dodd rink ##fle Keyboards bulb diver ul ##tsk Exodus Deacon España Canadiens oblique thud reigned rug Whitman Dash ##iens Haifa pets ##arland manually dart ##bial Sven textiles subgroup Napier graffiti revolver humming Babu protector typed Provinces Sparta Wills subjective ##rella temptation ##liest FL Sadie manifest Guangdong Transfer entertain eve recipes ##33 Benedictine retailer ##dence establishes ##cluded ##rked Ursula ##ltz ##lars ##rena qualifiers ##curement colt depictions ##oit Spiritual differentiation staffed transitional ##lew 1761 fatalities ##oan Bayern Northamptonshire Weeks ##CU Fife capacities hoarse ##latt ##ة evidenced ##HD ##ographer assessing evolve hints 42nd streaked ##lve Yahoo ##estive ##rned ##zas baggage Elected secrecy ##champ Character Pen Decca cape Bernardino vapor Dolly counselor ##isers Benin ##khar ##CR notch ##thus ##racy bounty lend grassland ##chtenstein ##dating pseudo golfer simplest ##ceive Lucivar Triumph dinosaur dinosaurs ##šić Seahawks ##nco resorts reelected 1766 reproduce universally ##OA ER tendencies Consolidated Massey Tasmanian reckless ##icz ##ricks 1755 questionable Audience ##lates preseason Quran trivial Haitian Freeway dialed Appointed Heard ecosystems ##bula hormones Carbon Rd ##arney ##working Christoph presiding pu ##athy Morrow Dar ensures posing remedy EA disclosed ##hui ##rten rumours surveying ##ficiency Aziz Jewel Plays ##smatic Bernhard Christi ##eanut ##friend jailed ##dr govern neighbour butler Acheron murdering oils mac Editorial detectives bolts ##ulon Guitars malaria 36th Pembroke Opened ##hium harmonic serum ##sio Franks fingernails ##gli culturally evolving scalp VP deploy uploaded mater ##evo Jammu Spa ##icker flirting ##cursions Heidi Majority sprawled ##alytic Zheng bunker ##lena ST ##tile Jiang ceilings ##ently ##ols Recovery dire ##good Manson Honestly Montréal 1764 227 quota Lakshmi incentive Accounting ##cilla Eureka Reaper buzzed ##uh courtroom dub ##mberg KC Gong Theodor Académie NPR criticizing protesting ##pired ##yric abuses fisheries ##minated 1767 yd Gemini Subcommittee ##fuse Duff Wasn Wight cleaner ##tite planetary Survivor Zionist mounds ##rary landfall disruption yielding ##yana bids unidentified Garry Ellison Elmer Fishing Hayward demos modelling ##anche ##stick caressed entertained ##hesion piers Crimea ##mass WHO boulder trunks 1640 Biennale Palestinians Pursuit ##udes Dora contender ##dridge Nanjing ##ezer ##former ##ibel Whole proliferation ##tide ##weiler fuels predictions ##ente ##onium Filming absorbing Ramón strangled conveyed inhabit prostitutes recession bonded clinched ##eak ##iji ##edar Pleasure Rite Christy Therapy sarcasm ##collegiate hilt probation Sarawak coefficients underworld biodiversity SBS groom brewing dungeon ##claiming Hari turnover ##ntina ##omer ##opped orthodox styling ##tars ##ulata priced Marjorie ##eley ##abar Yong ##tically Crambidae Hernandez ##ego ##rricular ##ark ##lamour ##llin ##augh ##tens Advancement Loyola ##4th ##hh goin marshes Sardinia ##ša Ljubljana Singing suspiciously ##hesive Félix Regarding flap stimulation ##raught Apr Yin gaping tighten skier ##itas ##lad ##rani 264 Ashes Olson Problems Tabitha ##rading balancing sunrise ##ease ##iture ##ritic Fringe ##iciency Inspired Linnaeus PBA disapproval ##kles ##rka ##tails ##urger Disaster Laboratories apps paradise Aero Came sneaking Gee Beacon ODI commodity Ellington graphical Gretchen spire ##skaya ##trine RTÉ efficacy plc tribunal ##ytic downhill flu medications ##kaya widen Sunrise ##nous distinguishing pawn ##BO ##irn ##ssing ##ν Easton ##vila Rhineland ##aque defect ##saurus Goose Ju ##classified Middlesbrough shaping preached 1759 ##erland Ein Hailey musicals ##altered Galileo Hilda Fighters Lac ##ometric 295 Leafs Milano ##lta ##VD ##ivist penetrated Mask Orchard plaintiff ##icorn Yvonne ##fred outfielder peek Collier Caracas repealed Bois dell restrict Dolores Hadley peacefully ##LL condom Granny Orders sabotage ##toon ##rings compass marshal gears brigadier dye Yunnan communicating donate emerald vitamin administer Fulham ##classical ##llas Buckinghamshire Held layered disclosure Akira programmer shrimp Crusade ##ximal Luzon bakery ##cute Garth Citadel uniquely Curling info mum Para ##ști sleek ##ione hey Lantern mesh ##lacing ##lizzard ##gade prosecuted Alba Gilles greedy twists ##ogged Viper ##kata Appearances Skyla hymns ##pelled curving predictable Grave Watford ##dford ##liptic ##vary Westwood fluids Models statutes ##ynamite 1740 ##culate Framework Johanna ##gression Vuelta imp ##otion ##raga ##thouse Ciudad festivities ##love Beyoncé italics ##vance DB ##haman outs Singers ##ueva ##urning ##51 ##ntiary ##mobile 285 Mimi emeritus nesting Keeper Ways ##onal ##oux Edmond MMA ##bark ##oop Hampson ##ñez ##rets Gladstone wreckage Pont Playboy reluctance ##ná apprenticeship preferring Value originate ##wei ##olio Alexia ##rog Parachute jammed stud Eton vols ##ganized 1745 straining creep indicators ##mán humiliation hinted alma tanker ##egation Haynes Penang amazement branched rumble ##ddington archaeologists paranoid expenditure Absolutely Musicians banished ##fining baptism Joker Persons hemisphere ##tieth ##ück flock ##xing lbs Kung crab ##dak ##tinent Regulations barrage parcel ##ós Tanaka ##rsa Natalia Voyage flaws stepfather ##aven ##eological Botanical Minsk ##ckers Cinderella Feast Loving Previous Shark ##took barrister collaborators ##nnes Croydon Graeme Juniors ##7th ##formation ##ulos ##ák £2 ##hwa ##rove ##ș Whig demeanor Otago ##TH ##ooster Faber instructors ##ahl ##bha emptied ##schen saga ##lora exploding ##rges Crusaders ##caster ##uations streaks CBN bows insights ka 1650 diversion LSU Wingspan ##liva Response sanity Producers imitation ##fine Lange Spokane splash weed Siberian magnet ##rocodile capitals ##rgus swelled Rani Bells Silesia arithmetic rumor ##hampton favors Weird marketplace ##orm tsunami unpredictable ##citation ##ferno Tradition postwar stench succeeds ##roup Anya Users oversized totaling pouch ##nat Tripoli leverage satin ##cline Bathurst Lund Niall thereof ##quid Bangor barge Animated ##53 ##alan Ballard utilizes Done ballistic NDP gatherings ##elin ##vening Rockets Sabrina Tamara Tribal WTA ##citing blinded flux Khalid Una prescription ##jee Parents ##otics ##food Silicon cured electro perpendicular intimacy ##rified Lots ##ceiving ##powder incentives McKenna ##arma ##ounced ##rinkled Alzheimer ##tarian 262 Seas ##cam Novi ##hout ##morphic ##hazar ##hul ##nington Huron Bahadur Pirate pursed Griffiths indicted swap refrain ##mulating Lal stomped ##Pad ##mamoto Reef disposed plastered weeping ##rato Minas hourly tumors ##ruising Lyle ##yper ##sol Odisha credibility ##Dowell Braun Graphic lurched muster ##nex ##ührer ##connected ##iek ##ruba Carthage Peck maple bursting ##lava Enrico rite ##jak Moment ##skar Styx poking Spartan ##urney Hepburn Mart Titanic newsletter waits Mecklenburg agitated eats ##dious Chow matrices Maud ##sexual sermon 234 ##sible ##lung Qi cemeteries mined sprinter ##ckett coward ##gable ##hell ##thin ##FB Contact ##hay rainforest 238 Hemisphere boasts ##nders ##verance ##kat Convent Dunedin Lecturer lyricist ##bject Iberian comune ##pphire chunk ##boo thrusting fore informing pistols echoes Tier battleships substitution ##belt moniker ##charya ##lland Thoroughbred 38th ##01 ##tah parting tongues Cale ##seau Unionist modular celebrates preview steamed Bismarck 302 737 vamp ##finity ##nbridge weaknesses husky ##berman absently ##icide Craven tailored Tokugawa VIP syntax Kazan captives doses filtered overview Cleopatra Conversely stallion Burger Suez Raoul th ##reaves Dickson Nell Rate anal colder ##sław Arm Semitic ##green reflective 1100 episcopal journeys ##ours ##pository ##dering residue Gunn ##27 ##ntial ##crates ##zig Astros Renee Emerald ##vili connectivity undrafted Sampson treasures ##kura ##theon ##vern Destroyer ##iable ##ener Frederic briefcase confinement Bree ##WD Athena 233 Padres Thom speeding ##hali Dental ducks Putin ##rcle ##lou Asylum ##usk dusk pasture Institutes ONE jack ##named diplomacy Intercontinental Leagues Towns comedic premature ##edic ##mona ##ories trimmed Charge Cream guarantees Dmitry splashed Philosophical tramway ##cape Maynard predatory redundant ##gratory ##wry sobs Burgundy edible outfits Handel dazed dangerously idle Operational organizes ##sional blackish broker weddings ##halt Becca McGee ##gman protagonists ##pelling Keynes aux stumble ##ordination Nokia reel sexes ##woods ##pheric ##quished ##voc ##oir ##pathian ##ptus ##sma ##tating ##ê fulfilling sheath ##ayne Mei Ordinary Collin Sharpe grasses interdisciplinary ##OX Background ##ignment Assault transforms Hamas Serge ratios ##sik swaying ##rcia Rosen ##gant ##versible cinematographer curly penny Kamal Mellon Sailor Spence phased Brewers amassed Societies ##ropriations ##buted mythological ##SN ##byss ##ired Sovereign preface Parry ##ife altitudes crossings ##28 Crewe southernmost taut McKinley ##owa ##tore 254 ##ckney compiling Shelton ##hiko 228 Poll Shepard Labs Pace Carlson grasping ##ов Delaney Winning robotic intentional shattering ##boarding ##git ##grade Editions Reserves ignorant proposing ##hanna cutter Mongols NW ##eux Codex Cristina Daughters Rees forecast ##hita NGOs Stations Beaux Erwin ##jected ##EX ##trom Schumacher ##hrill ##rophe Maharaja Oricon ##sul ##dynamic ##fighting Ce Ingrid rumbled Prospect stairwell Barnard applause complementary ##uba grunt ##mented Bloc Carleton loft noisy ##hey 490 contrasted ##inator ##rief ##centric ##fica Cantonese Blanc Lausanne License artifact ##ddin rot Amongst Prakash RF ##topia milestone ##vard Winters Mead churchyard Lulu estuary ##ind Cha Infinity Meadow subsidies ##valent CONCACAF Ching medicinal navigate Carver Twice abdominal regulating RB toilets Brewer weakening ambushed ##aut ##vignon Lansing unacceptable reliance stabbing ##mpo ##naire Interview ##ested ##imed bearings ##lts Rashid ##iation authenticity vigorous ##frey ##uel biologist NFC ##rmaid ##wash Makes ##aunt ##steries withdrawing ##qa Buccaneers bleed inclination stain ##ilo ##ppel Torre privileged cereal trailers alumnus neon Cochrane Mariana caress ##47 ##ients experimentation Window convict signaled ##YP rower Pharmacy interacting 241 Strings dominating kinase Dinamo Wire pains sensations ##suse Twenty20 ##39 spotlight ##hend elemental ##pura Jameson Swindon honoring pained ##ediatric ##lux Psychological assemblies ingredient Martial Penguins beverage Monitor mysteries ##ION emigration mused ##sique crore AMC Funding Chinatown Establishment Finalist enjoyable 1756 ##mada ##rams NO newborn CS comprehend Invisible Siemens ##acon 246 contraction ##volving ##moration ##rok montane ##ntation Galloway ##llow Verity directorial pearl Leaning ##rase Fernandez swallowing Automatic Madness haunting paddle ##UE ##rrows ##vies ##zuki ##bolt ##iber Fender emails paste ##lancing hind homestead hopeless ##dles Rockies garlic fatty shrieked ##ismic Gillian Inquiry Schultz XML ##cius ##uld Domesday grenades northernmost ##igi Tbilisi optimistic ##poon Refuge stacks Bose smash surreal Nah Straits Conquest ##roo ##weet ##kell Gladys CH ##lim ##vitation Doctorate NRHP knocks Bey Romano ##pile 242 Diamonds strides eclectic Betsy clade ##hady ##leashed dissolve moss Suburban silvery ##bria tally turtles ##uctive finely industrialist ##nary Ernesto oz pact loneliness ##hov Tomb multinational risked Layne USL ne ##quiries Ad Message Kamen Kristen reefs implements ##itative educators garments gunshot ##essed ##rve Montevideo vigorously Stamford assemble packaged ##same état Viva paragraph ##eter ##wire Stick Navajo MCA ##pressing ensembles ABA ##zor ##llus Partner raked ##BI Iona thump Celeste Kiran ##iscovered ##rith inflammation ##arel Features loosened ##yclic Deluxe Speak economical Frankenstein Picasso showcased ##zad ##eira ##planes ##linear ##overs monsoon prosecutors slack Horses ##urers Angry coughing ##truder Questions ##tō ##zak challenger clocks ##ieving Newmarket ##acle cursing stimuli ##mming ##qualified slapping ##vasive narration ##kini Advertising CSI alliances mixes ##yes covert amalgamation reproduced ##ardt ##gis 1648 id Annette Boots Champagne Brest Daryl ##emon ##jou ##llers Mean adaptive technicians ##pair ##usal Yoga fronts leaping Jul harvesting keel ##44 petitioned ##lved yells Endowment proponent ##spur ##tised ##zal Homes Includes ##ifer ##oodoo ##rvette awarding mirrored ransom Flute outlook ##ganj DVDs Sufi frontman Goddard barren ##astic Suicide hillside Harlow Lau notions Amnesty Homestead ##irt GE hooded umpire mustered Catch Masonic ##erd Dynamics Equity Oro Charts Mussolini populace muted accompaniment ##lour ##ndes ignited ##iferous ##laced ##atch anguish registry ##tub ##hards ##neer 251 Hooker uncomfortably ##6th ##ivers Catalina MiG giggling 1754 Dietrich Kaladin pricing ##quence Sabah ##lving ##nical Gettysburg Vita Telecom Worst Palais Pentagon ##brand ##chichte Graf unnatural 1715 bio ##26 Radcliffe ##utt chatting spices ##aus untouched ##eper Doll turkey Syndicate ##rlene ##JP ##roots Como clashed modernization 1757 fantasies ##iating dissipated Sicilian inspect sensible reputed ##final Milford poised RC metabolic Tobacco Mecca optimization ##heat lobe rabbits NAS geologist ##liner Kilda carpenter nationalists ##brae summarized ##venge Designer misleading beamed ##meyer Matrix excuses ##aines ##biology 401 Moose drafting Sai ##ggle Comprehensive dripped skate ##WI ##enan ##ruk narrower outgoing ##enter ##nounce overseen ##structure travellers banging scarred ##thing ##arra Ebert Sometime ##nated BAFTA Hurricanes configurations ##MLL immortality ##heus gothic ##mpest clergyman viewpoint Maxim Instituto emitted quantitative 1689 Consortium ##rsk Meat Tao swimmers Shaking Terence mainline ##linity Quantum ##rogate Nair banquet 39th reprised lagoon subdivisions synonymous incurred password sprung ##vere Credits Petersen Faces ##vu statesman Zombie gesturing ##going Sergey dormant possessive totals southward Ángel ##odies HM Mariano Ramirez Wicked impressions ##Net ##cap ##ème Transformers Poker RIAA Redesignated ##chuk Harcourt Peña spacious tinged alternatively narrowing Brigham authorization Membership Zeppelin ##amed Handball steer ##orium ##rnal ##rops Committees endings ##MM ##yung ejected grams ##relli Birch Hilary Stadion orphan clawed ##kner Motown Wilkins ballads outspoken ##ancipation ##bankment ##cheng Advances harvested novelty ineligible oversees ##´s obeyed inevitably Kingdoms burying Fabian relevance Tatiana ##MCA sarcastic ##onda Akron 229 sandwiches Adobe Maddox ##azar Hunting ##onized Smiling ##tology Juventus Leroy Poets attach lo ##rly ##film Structure ##igate olds projections SMS outnumbered ##tase judiciary paramilitary playfully ##rsing ##tras Chico Vin informally abandonment ##russ Baroness injuring octagonal deciduous ##nea ##olm Hz Norwood poses Marissa alerted willed ##KS Dino ##ddler ##vani Barbie Thankfully 625 bicycles shimmering ##tinuum ##wolf Chesterfield ##idy ##urgency Knowles sweetly Ventures ##ponents ##valence Darryl Powerplant RAAF ##pec Kingsley Parramatta penetrating spectacle ##inia Marlborough residual compatibility hike Underwood depleted ministries ##odus ##ropriation rotting Faso ##inn Happiness Lille Suns cookie rift warmly ##lvin Bugs Gotham Gothenburg Properties ##seller ##ubi Created MAC Noelle Requiem Ulysses ##ails franchises ##icious ##rwick celestial kinetic 720 STS transmissions amplitude forums freeing reptiles tumbling ##continent ##rising ##tropy physiology ##uster Loves bodied neutrality Neumann assessments Vicky ##hom hampered ##uku Custom timed ##eville ##xious elastic ##section rig stilled shipment 243 artworks boulders Bournemouth ##hly ##LF ##linary rumored ##bino ##drum Chun Freiburg ##dges Equality 252 Guadalajara ##sors ##taire Roach cramped ##ultural Logistics Punch fines Lai caravan ##55 lame Collector pausing 315 migrant hawk signalling ##erham ##oughs Demons surfing Rana insisting Wien adolescent ##jong ##rera ##umba Regis brushes ##iman residues storytelling Consider contrasting regeneration ##elling ##hlete afforded reactors costing ##biotics ##gat ##евич chanting secondly confesses ##ikos ##uang ##ronological ##− Giacomo ##eca vaudeville weeds rejecting revoked affluent fullback progresses geologic proprietor replication gliding recounted ##bah ##igma Flow ii newcomer ##lasp ##miya Candace fractured interiors confidential Inverness footing ##robe Coordinator Westphalia jumper ##chism dormitory ##gno 281 acknowledging leveled ##éra Algiers migrate Frog Rare ##iovascular ##urous DSO nomadic ##iera woken lifeless ##graphical ##ifications Dot Sachs crow nmi Tacoma Weight mushroom RS conditioned ##zine Tunisian altering ##mizing Handicap Patti Monsieur clicking gorge interrupting ##powerment drawers Serra ##icides Specialist ##itte connector worshipped ##ask consoles tags ##iler glued ##zac fences Bratislava honeymoon 313 A2 disposition Gentleman Gilmore glaciers ##scribed Calhoun convergence Aleppo shortages ##43 ##orax ##worm ##codes ##rmal neutron ##ossa Bloomberg Salford periodicals ##ryan Slayer ##ynasties credentials ##tista surveyor File stinging unnoticed Medici ecstasy espionage Jett Leary circulating bargaining concerto serviced 37th HK ##fueling Delilah Marcia graded ##join Kaplan feasible ##nale ##yt Burnley dreadful ministerial Brewster Judah ##ngled ##rrey recycled Iroquois backstage parchment ##numbered Kern Motorsports Organizations ##mini Seems Warrington Dunbar Ezio ##eor paralyzed Ara yeast ##olis cheated reappeared banged ##ymph ##dick Lyndon glide Mat ##natch Hotels Household parasite irrelevant youthful ##smic ##tero ##anti 2d Ignacio squash ##nets shale ##اد Abrams ##oese assaults ##dier ##otte Swamp 287 Spurs ##economic Fargo auditioned ##mé Haas une abbreviation Turkic ##tisfaction favorites specials ##lial Enlightenment Burkina ##vir Comparative Lacrosse elves ##lerical ##pear Borders controllers ##villa excelled ##acher ##varo camouflage perpetual ##ffles devoid schooner ##bered ##oris Gibbons Lia discouraged sue ##gnition Excellent Layton noir smack ##ivable ##evity ##lone Myra weaken weaponry ##azza Shake backbone Certified clown occupational caller enslaved soaking Wexford perceive shortlisted ##pid feminism Bari Indie ##avelin ##ldo Hellenic Hundreds Savings comedies Honors Mohawk Told coded Incorporated hideous trusts hose Calais Forster Gabon Internationale AK Colour ##UM ##heist McGregor localized ##tronomy Darrell ##iara squirrel freaked ##eking ##manned ##ungen radiated ##dua commence Donaldson ##iddle MR SAS Tavern Teenage admissions Instruments ##ilizer Konrad contemplated ##ductor Jing Reacher recalling Dhabi emphasizing illumination ##tony legitimacy Goethe Ritter McDonnell Polar Seconds aspiring derby tunic ##rmed outlines Changing distortion ##cter Mechanics ##urly ##vana Egg Wolverine Stupid centralized knit ##Ms Saratoga Ogden storylines ##vres lavish beverages ##grarian Kyrgyzstan forcefully superb Elm Thessaloniki follower Plants slang trajectory Nowadays Bengals Ingram perch coloring carvings doubtful ##aph ##gratulations ##41 Curse 253 nightstand Campo Meiji decomposition ##giri McCormick Yours ##amon ##bang Texans injunction organise periodical ##peculative oceans ##aley Success Lehigh ##guin 1730 Davy allowance obituary ##tov treasury ##wayne euros readiness systematically ##stered ##igor ##xen ##cliff ##lya Send ##umatic Celtics Judiciary 425 propagation rebellious ##ims ##lut Dal ##ayman ##cloth Boise pairing Waltz torment Hatch aspirations diaspora ##hame Rank 237 Including Muir chained toxicity Université ##aroo Mathews meadows ##bio Editing Khorasan ##them ##ahn ##bari ##umes evacuate ##sium gram kidnap pinning ##diation ##orms beacon organising McGrath ##ogist Qur Tango ##ceptor ##rud ##cend ##cie ##jas ##sided Tuscany Venture creations exhibiting ##rcerer ##tten Butcher Divinity Pet Whitehead falsely perished handy Moines cyclists synthesizers Mortal notoriety ##ronic Dialogue expressive uk Nightingale grimly vineyards Driving relentless compiler ##district ##tuated Hades medicines objection Answer Soap Chattanooga ##gogue Haryana Parties Turtle ##ferred explorers stakeholders ##aar ##rbonne tempered conjecture ##tee ##hur Reeve bumper stew ##church ##generate ##ilitating ##chanized ##elier ##enne translucent ##lows Publisher evangelical inherit ##rted 247 SmackDown bitterness lesions ##worked mosques wed ##lashes Ng Rebels booking ##nail Incident Sailing yo confirms Chaplin baths ##kled modernist pulsing Cicero slaughtered boasted ##losure zipper ##hales aristocracy halftime jolt unlawful Marching sustaining Yerevan bracket ram Markus ##zef butcher massage ##quisite Leisure Pizza collapsing ##lante commentaries scripted ##disciplinary ##sused eroded alleging vase Chichester Peacock commencement dice hotter poisonous executions ##occo frost fielding vendor Counts Troops maize Divisional analogue shadowy Nuevo Ville radiating worthless Adriatic Buy blaze brutally horizontally longed ##matical federally Rolf Root exclude rag agitation Lounge astonished ##wirl Impossible transformations ##IVE ##ceded ##slav downloaded fucked Egyptians Welles ##ffington U2 befriended radios ##jid archaic compares ##ccelerator ##imated ##tosis Hung Scientists Thousands geographically ##LR Macintosh fluorescent ##ipur Wehrmacht ##BR ##firmary Chao ##ague Boyer ##grounds ##hism ##mento ##taining infancy ##cton 510 Boca ##loy 1644 ben dong stresses Sweat expressway graders ochreous nets Lawn thirst Uruguayan satisfactory ##tracts baroque rusty ##ław Shen Gdańsk chickens ##graving Hodge Papal SAT bearer ##ogo ##rger merits Calendar Highest Skills ##ortex Roberta paradigm recounts frigates swamps unitary ##oker balloons Hawthorne Muse spurred advisors reclaimed stimulate fibre pat repeal ##dgson ##iar ##rana anthropologist descends flinch reared ##chang ##eric ##lithic commissioning ##cumenical ##lume ##rchen Wolff ##tsky Eurasian Nepali Nightmare ZIP playback ##latz ##vington Warm ##75 Martina Rollins Saetan Variations sorting ##م 530 Joaquin Ptolemy thinner ##iator ##pticism Cebu Highlanders Linden Vanguard ##SV ##mor ##ulge ISSN cartridges repression Étienne 311 Lauderdale commodities null ##rb 1720 gearbox ##reator Ang Forgotten dubious ##rls ##dicative ##phate Groove Herrera ##çais Collections Maximus ##published Fell Qualification filtering ##tized Roe hazards ##37 ##lative ##tröm Guadalupe Tajikistan Preliminary fronted glands ##paper ##iche ##iding Cairns rallies Location seduce ##mple BYU ##itic ##FT Carmichael Prentice songwriters forefront Physicians ##rille ##zee Preparatory ##cherous UV ##dized Navarro misses ##nney Inland resisting ##sect Hurt ##lino galaxies ##raze Institutions devote ##lamp ##ciating baron ##bracing Hess operatic ##CL ##ος Chevalier Guiana ##lattered Fed ##cuted ##smo Skull denies 236 Waller ##mah Sakura mole nominate sermons ##bering widowed ##röm Cavendish ##struction Nehru Revelation doom Gala baking Nr Yourself banning Individuals Sykes orchestrated 630 Phone steered 620 specialising starvation ##AV ##alet ##upation seductive ##jects ##zure Tolkien Benito Wizards Submarine dictator Duo Caden approx basins ##nc shrink ##icles ##sponsible 249 mit outpost ##bayashi ##rouse ##tl Jana Lombard RBIs finalized humanities ##function Honorable tomato ##iot Pie tee ##pect Beaufort Ferris bucks ##graduate ##ocytes Directory anxiously ##nating flanks ##Ds virtues ##believable Grades criterion manufactures sourced ##balt ##dance ##tano Ying ##BF ##sett adequately blacksmith totaled trapping expanse Historia Worker Sense ascending housekeeper ##oos Crafts Resurrection ##verty encryption ##aris ##vat ##pox ##runk ##iability gazes spying ##ths helmets wired ##zophrenia Cheung WR downloads stereotypes 239 Lucknow bleak Bragg hauling ##haft prohibit ##ermined ##castle barony ##hta Typhoon antibodies ##ascism Hawthorn Kurdistan Minority Gorge Herr appliances disrupt Drugs Lazarus ##ilia ##ryo ##tany Gotta Masovian Roxy choreographed ##rissa turbulent ##listed Anatomy exiting ##det ##isław 580 Kaufman sage ##apa Symposium ##rolls Kaye ##ptera ##rocław jerking ##menclature Guo M1 resurrected trophies ##lard Gathering nestled serpent Dow reservoirs Claremont arbitration chronicle eki ##arded ##zers ##mmoth Congregational Astronomical NE RA Robson Scotch modelled slashed ##imus exceeds ##roper ##utile Laughing vascular superficial ##arians Barclay Caucasian classmate sibling Kimberly Shreveport ##ilde ##liche Cheney Deportivo Veracruz berries ##lase Bed MI Anatolia Mindanao broadband ##olia ##arte ##wab darts ##immer ##uze believers ordinance violate ##wheel ##ynth Alongside Coupe Hobbs arrondissement earl townland ##dote ##lihood ##sla Ghosts midfield pulmonary ##eno cues ##gol ##zda 322 Siena Sultanate Bradshaw Pieter ##thical Raceway bared competence ##ssent Bet ##urer ##ła Alistair Göttingen appropriately forge ##osterone ##ugen DL 345 convoys inventions ##resses ##cturnal Fay Integration slash ##roats Widow barking ##fant 1A Hooper ##cona ##runched unreliable ##emont ##esign ##stabulary ##stop Journalists bony ##iba ##trata ##ège horrific ##bish Jocelyn ##rmon ##apon ##cier trainers ##ulatory 1753 BR corpus synthesized ##bidden ##rafford Elgin ##entry Doherty clockwise ##played spins ##ample ##bley Cope constructions seater warlord Voyager documenting fairies ##viator Lviv jewellery suites ##gold Maia NME ##eavor ##kus Eugène furnishings ##risto MCC Metropolis Older Telangana ##mpus amplifier supervising 1710 buffalo cushion terminating ##powering steak Quickly contracting dem sarcastically Elsa ##hein bastards narratives Takes 304 composure typing variance ##ifice Softball ##rations McLaughlin gaped shrines ##hogany Glamorgan ##icle ##nai ##ntin Fleetwood Woodland ##uxe fictitious shrugs ##iper BWV conform ##uckled Launch ##ductory ##mized Tad ##stituted ##free Bel Chávez messing quartz ##iculate ##folia ##lynn ushered ##29 ##ailing dictated Pony ##opsis precinct 802 Plastic ##ughter ##uno ##porated Denton Matters SPD hating ##rogen Essential Deck Dortmund obscured ##maging Earle ##bred ##ittle ##ropolis saturated ##fiction ##ression Pereira Vinci mute warehouses ##ún biographies ##icking sealing ##dered executing pendant ##wives murmurs ##oko substrates symmetrical Susie ##mare Yusuf analogy ##urage Lesley limitation ##rby ##ío disagreements ##mise embroidered nape unarmed Sumner Stores dwell Wilcox creditors ##rivatization ##shes ##amia directs recaptured scouting McGuire cradle ##onnell Sato insulin mercenary tolerant Macquarie transitions cradled ##berto ##ivism ##yotes FF Ke Reach ##dbury 680 ##bill ##oja ##sui prairie ##ogan reactive ##icient ##rits Cyclone Sirius Survival Pak ##coach ##trar halves Agatha Opus contrasts ##jection ominous ##iden Baylor Woodrow duct fortification intercourse ##rois Colbert envy ##isi Afterward geared ##flections accelerate ##lenching Witness ##rrer Angelina Material assertion misconduct Nix cringed tingling ##eti ##gned Everest disturb sturdy ##keepers ##vied Profile heavenly ##kova ##victed translating ##sses 316 Invitational Mention martyr ##uristic Barron hardness Nakamura 405 Genevieve reflections ##falls jurist ##LT Pyramid ##yme Shoot heck linguist ##tower Ives superiors ##leo Achilles ##phological Christophe Padma precedence grassy Oral resurrection ##itting clumsy ##lten ##rue huts ##stars Equal ##queduct Devin Gaga diocesan ##plating ##upe ##graphers Patch Scream hail moaning tracts ##hdi Examination outsider ##ergic ##oter Archipelago Havilland greenish tilting Aleksandr Konstantin warship ##emann ##gelist ##ought billionaire ##blivion 321 Hungarians transplant ##jured ##fters Corbin autism pitchers Garner thence Scientology transitioned integrating repetitive ##dant Rene vomit ##burne 1661 Researchers Wallis insulted wavy ##wati Ewing excitedly ##kor frescoes injustice ##achal ##lumber ##úl novella ##sca Liv ##enstein ##river monstrous topping downfall looming sinks trillion ##pont Effect ##phi ##urley Sites catchment ##H1 Hopper ##raiser 1642 Maccabi lance ##chia ##sboro NSA branching retorted tensor Immaculate drumming feeder ##mony Dyer homicide Temeraire fishes protruding skins orchards ##nso inlet ventral ##finder Asiatic Sul 1688 Melinda assigns paranormal gardening Tau calming ##inge ##crow regimental Nik fastened correlated ##gene ##rieve Sick ##minster ##politan hardwood hurled ##ssler Cinematography rhyme Montenegrin Packard debating ##itution Helens Trick Museums defiance encompassed ##EE ##TU ##nees ##uben ##ünster ##nosis 435 Hagen cinemas Corbett commended ##fines ##oman bosses ripe scraping ##loc filly Saddam pointless Faust Orléans Syriac ##♭ longitude ##ropic Alfa bliss gangster ##ckling SL blending ##eptide ##nner bends escorting ##bloid ##quis burials ##sle ##è Ambulance insults ##gth Antrim unfolded ##missible splendid Cure warily Saigon Waste astonishment boroughs ##VS ##dalgo ##reshing ##usage rue marital versatile unpaid allotted bacterium ##coil ##cue Dorothea IDF ##location ##yke RPG ##tropical devotees liter ##pree Johnstone astronaut attends pollen periphery doctrines meta showered ##tyn GO Huh laude 244 Amar Christensen Ping Pontifical Austen raiding realities ##dric urges ##dek Cambridgeshire ##otype Cascade Greenberg Pact ##cognition ##aran ##urion Riot mimic Eastwood ##imating reversal ##blast ##henian Pitchfork ##sunderstanding Staten WCW lieu ##bard ##sang experimenting Aquino ##lums TNT Hannibal catastrophic ##lsive 272 308 ##otypic 41st Highways aggregator ##fluenza Featured Reece dispatch simulated ##BE Communion Vinnie hardcover inexpensive til ##adores groundwater kicker blogs frenzy ##wala dealings erase Anglia ##umour Hapoel Marquette ##raphic ##tives consult atrocities concussion ##érard Decree ethanol ##aen Rooney ##chemist ##hoot 1620 menacing Schuster ##bearable laborers sultan Juliana erased onstage ##ync Eastman ##tick hushed ##yrinth Lexie Wharton Lev ##PL Testing Bangladeshi ##bba ##usions communicated integers internship societal ##odles Loki ET Ghent broadcasters Unix ##auer Kildare Yamaha ##quencing ##zman chilled ##rapped ##uant Duval sentiments Oliveira packets Horne ##rient Harlan Mirage invariant ##anger ##tensive flexed sweetness ##wson alleviate insulting limo Hahn ##llars ##hesia ##lapping buys ##oaming mocked pursuits scooted ##conscious ##ilian Ballad jackets ##kra hilly ##cane Scenic McGraw silhouette whipping ##roduced ##wark ##chess ##rump Lemon calculus demonic ##latine Bharatiya Govt Que Trilogy Ducks Suit stairway ##ceipt Isa regulator Automobile flatly ##buster ##lank Spartans topography Tavi usable Chartered Fairchild ##sance ##vyn Digest nuclei typhoon ##llon Alvarez DJs Grimm authoritative firearm ##chschule Origins lair unmistakable ##xial ##cribing Mouth ##genesis ##shū ##gaon ##ulter Jaya Neck ##UN ##oing ##static relativity ##mott ##utive ##esan ##uveau BT salts ##roa Dustin preoccupied Novgorod ##asus Magnum tempting ##histling ##ilated Musa ##ghty Ashland pubs routines ##etto Soto 257 Featuring Augsburg ##alaya Bit loomed expects ##abby ##ooby Auschwitz Pendleton vodka ##sent rescuing systemic ##inet ##leg Yun applicant revered ##nacht ##ndas Muller characterization ##patient ##roft Carole ##asperated Amiga disconnected gel ##cologist Patriotic rallied assign veterinary installing ##cedural 258 Jang Parisian incarcerated stalk ##iment Jamal McPherson Palma ##oken ##viation 512 Rourke irrational ##rippled Devlin erratic ##NI ##payers Ni engages Portal aesthetics ##rrogance Milne assassins ##rots 335 385 Cambodian Females fellows si ##block ##otes Jayne Toro flutter ##eera Burr ##lanche relaxation ##fra Fitzroy ##undy 1751 261 comb conglomerate ribbons veto ##Es casts ##ege 1748 Ares spears spirituality comet ##nado ##yeh Veterinary aquarium yer Councils ##oked ##ynamic Malmö remorse auditions drilled Hoffmann Moe Nagoya Yacht ##hakti ##race ##rrick Talmud coordinating ##EI ##bul ##his ##itors ##ligent ##uerra Narayan goaltender taxa ##asures Det ##mage Infinite Maid bean intriguing ##cription gasps socket ##mentary ##reus sewing transmitting ##different ##furbishment ##traction Grimsby sprawling Shipyard ##destine ##hropic ##icked trolley ##agi ##lesh Josiah invasions Content firefighters intro Lucifer subunit Sahib Myrtle inhibitor maneuvers ##teca Wrath slippery ##versing Shoes ##dial ##illiers ##luded ##mmal ##pack handkerchief ##edestal ##stones Fusion cumulative ##mell ##cacia ##rudge ##utz foe storing swiped ##meister ##orra batter strung ##venting ##kker Doo Taste immensely Fairbanks Jarrett Boogie 1746 mage Kick legislators medial ##ilon ##logies ##ranton Hybrid ##uters Tide deportation Metz ##secration ##virus UFO ##fell ##orage ##raction ##rrigan 1747 fabricated ##BM ##GR ##rter muttering theorist ##tamine BMG Kincaid solvent ##azed Thin adorable Wendell ta ##viour pulses ##pologies counters exposition sewer Luciano Clancy ##angelo ##riars Showtime observes frankly ##oppy Bergman lobes timetable ##bri ##uest FX ##dust ##genus Glad Helmut Meridian ##besity ##ontaine Revue miracles ##titis PP bluff syrup 307 Messiah ##erne interfering picturesque unconventional dipping hurriedly Kerman 248 Ethnic Toward acidic Harrisburg ##65 intimidating ##aal Jed Pontiac munitions ##nchen growling mausoleum ##ération ##wami Cy aerospace caucus Doing ##around ##miring Cuthbert ##poradic ##rovisation ##wth evaluating ##scraper Belinda owes ##sitic ##thermal ##fast economists ##lishing ##uerre ##ân credible ##koto Fourteen cones ##ebrates bookstore towels ##phony Appearance newscasts ##olin Karin Bingham ##elves 1680 306 disks ##lston ##secutor Levant ##vout Micro snuck ##ogel ##racker Exploration drastic ##kening Elsie endowment ##utnant Blaze ##rrosion leaking 45th ##rug ##uernsey 760 Shapiro cakes ##ehan ##mei ##ité ##kla repetition successively Friendly Île Koreans Au Tirana flourish Spirits Yao reasoned ##leam Consort cater marred ordeal supremacy ##ritable Paisley euro healer portico wetland ##kman restart ##habilitation ##zuka ##Script emptiness communion ##CF ##inhabited ##wamy Casablanca pulsed ##rrible ##safe 395 Dual Terrorism ##urge ##found ##gnolia Courage patriarch segregated intrinsic ##liography ##phe PD convection ##icidal Dharma Jimmie texted constituents twitch ##calated ##mitage ##ringing 415 milling ##geons Armagh Geometridae evergreen needy reflex template ##pina Schubert ##bruck ##icted ##scher ##wildered 1749 Joanne clearer ##narl 278 Print automation consciously flashback occupations ##ests Casimir differentiated policing repay ##aks ##gnesium Evaluation commotion ##CM ##smopolitan Clapton mitochondrial Kobe 1752 Ignoring Vincenzo Wet bandage ##rassed ##unate Maris ##eted ##hetical figuring ##eit ##nap leopard strategically ##reer Fen Iain ##ggins ##pipe Matteo McIntyre ##chord ##feng Romani asshole flopped reassure Founding Styles Torino patrolling ##erging ##ibrating ##ructural sincerity ##ät ##teacher Juliette ##cé ##hog ##idated ##span Winfield ##fender ##nast ##pliant 1690 Bai Je Saharan expands Bolshevik rotate ##root Britannia Severn ##cini ##gering ##say sly Steps insertion rooftop Piece cuffs plausible ##zai Provost semantic ##data ##vade ##cimal IPA indictment Libraries flaming highlands liberties ##pio Elders aggressively ##pecific Decision pigeon nominally descriptive adjustments equestrian heaving ##mour ##dives ##fty ##yton intermittent ##naming ##sets Calvert Casper Tarzan ##kot Ramírez ##IB ##erus Gustavo Roller vaulted ##solation ##formatics ##tip Hunger colloquially handwriting hearth launcher ##idian ##ilities ##lind ##locating Magdalena Soo clubhouse ##kushima ##ruit Bogotá Organic Worship ##Vs ##wold upbringing ##kick groundbreaking ##urable ##ván repulsed ##dira ##ditional ##ici melancholy ##bodied ##cchi 404 concurrency H₂O bouts ##gami 288 Leto troll ##lak advising bundled ##nden lipstick littered ##leading ##mogeneous Experiment Nikola grove ##ogram Mace ##jure cheat Annabelle Tori lurking Emery Walden ##riz paints Markets brutality overrun ##agu ##sat din ostensibly Fielding flees ##eron Pound ornaments tornadoes ##nikov ##organisation ##reen ##Works ##ldred ##olten ##stillery soluble Mata Grimes Léon ##NF coldly permitting ##inga ##reaked Agents hostess ##dl Dyke Kota avail orderly ##saur ##sities Arroyo ##ceps ##egro Hawke Noctuidae html seminar ##ggles ##wasaki Clube recited ##sace Ascension Fitness dough ##ixel Nationale ##solidate pulpit vassal 570 Annapolis bladder phylogenetic ##iname convertible ##ppan Comet paler ##definite Spot ##dices frequented Apostles slalom ##ivision ##mana ##runcated Trojan ##agger ##iq ##league Concept Controller ##barian ##curate ##spersed ##tring engulfed inquired ##hmann 286 ##dict ##osy ##raw MacKenzie su ##ienced ##iggs ##quitaine bisexual ##noon runways subsp ##! ##" ### ##$ ##% ##& ##' ##( ##) ##* ##+ ##, ##- ##. ##/ ##: ##; ##< ##= ##> ##? ##@ ##[ ##\ ##] ##^ ##_ ##` ##{ ##| ##} ##~ ##¡ ##¢ ##£ ##¥ ##§ ##¨ ##© ##ª ##« ##¬ ##® ##± ##´ ##µ ##¶ ##· ##¹ ##º ##» ##¼ ##¾ ##¿ ##À ##Á ## ##Ä ##Å ##Æ ##Ç ##È ##É ##Í ##Î ##Ñ ##Ó ##Ö ##× ##Ø ##Ú ##Ü ##Þ ##â ##ã ##æ ##ç ##î ##ï ##ð ##ñ ##ô ##õ ##÷ ##û ##þ ##ÿ ##Ā ##ą ##Ć ##Č ##ď ##Đ ##đ ##ē ##ė ##ę ##ě ##ğ ##ġ ##Ħ ##ħ ##ĩ ##Ī ##İ ##ļ ##Ľ ##ľ ##Ł ##ņ ##ň ##ŋ ##Ō ##ŏ ##ő ##Œ ##œ ##ř ##Ś ##ś ##Ş ##Š ##Ţ ##ţ ##ť ##ũ ##ŭ ##ů ##ű ##ų ##ŵ ##ŷ ##ź ##Ż ##ż ##Ž ##ž ##Ə ##ƒ ##ơ ##ư ##ǎ ##ǐ ##ǒ ##ǔ ##ǫ ##Ș ##Ț ##ț ##ɐ ##ɑ ##ɔ ##ɕ ##ə ##ɛ ##ɡ ##ɣ ##ɨ ##ɪ ##ɲ ##ɾ ##ʀ ##ʁ ##ʂ ##ʃ ##ʊ ##ʋ ##ʌ ##ʐ ##ʑ ##ʒ ##ʔ ##ʰ ##ʲ ##ʳ ##ʷ ##ʻ ##ʼ ##ʾ ##ʿ ##ˈ ##ː ##ˡ ##ˢ ##ˣ ##́ ##̃ ##̍ ##̯ ##͡ ##Α ##Β ##Γ ##Δ ##Ε ##Η ##Θ ##Ι ##Κ ##Λ ##Μ ##Ν ##Ο ##Π ##Σ ##Τ ##Φ ##Χ ##Ψ ##Ω ##ά ##έ ##ή ##ί ##β ##γ ##δ ##ε ##ζ ##η ##θ ##ι ##κ ##λ ##μ ##ξ ##ο ##π ##ρ ##σ ##τ ##υ ##φ ##χ ##ψ ##ω ##ό ##ύ ##ώ ##І ##Ј ##А ##Б ##В ##Г ##Д ##Е ##Ж ##З ##И ##К ##Л ##М ##Н ##О ##П ##Р ##С ##Т ##У ##Ф ##Х ##Ц ##Ч ##Ш ##Э ##Ю ##Я ##б ##в ##г ##д ##ж ##з ##к ##л ##м ##п ##с ##т ##у ##ф ##х ##ц ##ч ##ш ##щ ##ъ ##ы ##ь ##э ##ю ##ё ##і ##ї ##ј ##њ ##ћ ##Ա ##Հ ##ա ##ե ##ի ##կ ##մ ##յ ##ն ##ո ##ս ##տ ##ր ##ւ ##ְ ##ִ ##ֵ ##ֶ ##ַ ##ָ ##ֹ ##ּ ##א ##ב ##ג ##ד ##ה ##ו ##ז ##ח ##ט ##י ##כ ##ל ##ם ##מ ##ן ##נ ##ס ##ע ##פ ##צ ##ק ##ר ##ש ##ת ##، ##ء ##آ ##أ ##إ ##ئ ##ا ##ب ##ت ##ث ##ج ##ح ##خ ##ذ ##ز ##س ##ش ##ص ##ض ##ط ##ظ ##ع ##غ ##ف ##ق ##ك ##ل ##و ##ى ##َ ##ِ ##ٹ ##پ ##چ ##ک ##گ ##ہ ##ی ##ے ##ं ##आ ##क ##ग ##च ##ज ##ण ##त ##द ##ध ##न ##प ##ब ##भ ##म ##य ##र ##ल ##व ##श ##ष ##स ##ह ##ा ##ि ##ी ##ु ##े ##ो ##् ##। ##॥ ##আ ##ই ##এ ##ও ##ক ##খ ##গ ##চ ##ছ ##জ ##ট ##ত ##থ ##দ ##ধ ##ন ##প ##ব ##ম ##য ##র ##ল ##শ ##স ##হ ##় ##া ##ি ##ী ##ু ##ে ##ো ##্ ##য় ##க ##த ##ப ##ம ##ய ##ர ##ல ##வ ##ா ##ி ##ு ##் ##ร ##་ ##ག ##ང ##ད ##ན ##བ ##མ ##ར ##ལ ##ས ##ི ##ུ ##ེ ##ོ ##ა ##ე ##ი ##ლ ##ნ ##ო ##რ ##ს ##ᴬ ##ᴵ ##ᵀ ##ᵃ ##ᵇ ##ᵈ ##ᵉ ##ᵍ ##ᵏ ##ᵐ ##ᵒ ##ᵖ ##ᵗ ##ᵘ ##ᵣ ##ᵤ ##ᵥ ##ᶜ ##ᶠ ##ḍ ##Ḥ ##ḥ ##Ḩ ##ḩ ##ḳ ##ṃ ##ṅ ##ṇ ##ṛ ##ṣ ##ṭ ##ạ ##ả ##ấ ##ầ ##ẩ ##ậ ##ắ ##ế ##ề ##ể ##ễ ##ệ ##ị ##ọ ##ố ##ồ ##ổ ##ộ ##ớ ##ờ ##ợ ##ụ ##ủ ##ứ ##ừ ##ử ##ữ ##ự ##ỳ ##ỹ ##ἀ ##ἐ ##ὁ ##ὐ ##ὰ ##ὶ ##ὸ ##ῆ ##ῖ ##ῦ ##ῶ ##‐ ##‑ ##‒ ##– ##— ##― ##‖ ##‘ ##’ ##‚ ##“ ##” ##„ ##† ##‡ ##• ##… ##‰ ##′ ##″ ##⁄ ##⁰ ##ⁱ ##⁴ ##⁵ ##⁶ ##⁷ ##⁸ ##⁹ ##⁻ ##ⁿ ##₅ ##₆ ##₇ ##₈ ##₉ ##₊ ##₍ ##₎ ##ₐ ##ₑ ##ₒ ##ₓ ##ₕ ##ₖ ##ₘ ##ₚ ##ₛ ##ₜ ##₤ ##€ ##₱ ##₹ ##ℓ ##№ ##ℝ ##⅓ ##← ##↑ ##→ ##↔ ##⇌ ##⇒ ##∂ ##∈ ##∗ ##∘ ##√ ##∞ ##∧ ##∨ ##∩ ##∪ ##≈ ##≠ ##≡ ##≤ ##≥ ##⊂ ##⊆ ##⊕ ##⋅ ##─ ##│ ##■ ##● ##★ ##☆ ##☉ ##♠ ##♣ ##♥ ##♦ ##♯ ##⟨ ##⟩ ##ⱼ ##、 ##。 ##《 ##》 ##「 ##」 ##『 ##』 ##〜 ##い ##う ##え ##お ##か ##き ##く ##け ##こ ##さ ##し ##す ##せ ##そ ##た ##ち ##つ ##て ##と ##な ##に ##の ##は ##ひ ##ま ##み ##む ##め ##も ##や ##ゆ ##よ ##ら ##り ##る ##れ ##ん ##ア ##ィ ##イ ##ウ ##エ ##オ ##カ ##ガ ##キ ##ク ##グ ##コ ##サ ##シ ##ジ ##ス ##ズ ##タ ##ダ ##ッ ##テ ##デ ##ト ##ド ##ナ ##ニ ##ハ ##バ ##パ ##フ ##ブ ##プ ##マ ##ミ ##ム ##ャ ##ュ ##ラ ##リ ##ル ##レ ##ロ ##ン ##・ ##ー ##一 ##三 ##上 ##下 ##中 ##事 ##二 ##井 ##京 ##人 ##亻 ##仁 ##佐 ##侍 ##光 ##公 ##力 ##北 ##十 ##南 ##原 ##口 ##史 ##司 ##吉 ##同 ##和 ##囗 ##国 ##國 ##土 ##城 ##士 ##大 ##天 ##太 ##夫 ##女 ##子 ##宀 ##安 ##宮 ##宿 ##小 ##尚 ##山 ##島 ##川 ##州 ##平 ##年 ##心 ##愛 ##戸 ##文 ##新 ##方 ##日 ##明 ##星 ##書 ##月 ##木 ##本 ##李 ##村 ##東 ##松 ##林 ##正 ##武 ##氏 ##水 ##氵 ##江 ##河 ##海 ##版 ##犬 ##王 ##生 ##田 ##白 ##皇 ##省 ##真 ##石 ##社 ##神 ##竹 ##美 ##義 ##花 ##藤 ##西 ##谷 ##車 ##辶 ##道 ##郎 ##郡 ##部 ##野 ##金 ##長 ##門 ##陽 ##青 ##食 ##馬 ##高 ##龍 ##龸 ##사 ##씨 ##의 ##이 ##한 ##fi ##fl ##! ##( ##) ##, ##- ##/ ##:
TensorFlow/Detection/SSD/models/research/object_detection/samples/configs
configs
faster_rcnn_inception_v2_coco
# Faster R-CNN with Inception v2, configuration for MSCOCO Dataset. # Users should configure the fine_tune_checkpoint field in the train config as # well as the label_map_path and input_path fields in the train_input_reader and # eval_input_reader. Search for "PATH_TO_BE_CONFIGURED" to find the fields that # should be configured. model { faster_rcnn { num_classes: 90 image_resizer { keep_aspect_ratio_resizer { min_dimension: 600 max_dimension: 1024 } } feature_extractor { type: 'faster_rcnn_inception_v2' first_stage_features_stride: 16 } first_stage_anchor_generator { grid_anchor_generator { scales: [0.25, 0.5, 1.0, 2.0] aspect_ratios: [0.5, 1.0, 2.0] height_stride: 16 width_stride: 16 } } first_stage_box_predictor_conv_hyperparams { op: CONV regularizer { l2_regularizer { weight: 0.0 } } initializer { truncated_normal_initializer { stddev: 0.01 } } } first_stage_nms_score_threshold: 0.0 first_stage_nms_iou_threshold: 0.7 first_stage_max_proposals: 300 first_stage_localization_loss_weight: 2.0 first_stage_objectness_loss_weight: 1.0 initial_crop_size: 14 maxpool_kernel_size: 2 maxpool_stride: 2 second_stage_box_predictor { mask_rcnn_box_predictor { use_dropout: false dropout_keep_probability: 1.0 fc_hyperparams { op: FC regularizer { l2_regularizer { weight: 0.0 } } initializer { variance_scaling_initializer { factor: 1.0 uniform: true mode: FAN_AVG } } } } } second_stage_post_processing { batch_non_max_suppression { score_threshold: 0.0 iou_threshold: 0.6 max_detections_per_class: 100 max_total_detections: 300 } score_converter: SOFTMAX } second_stage_localization_loss_weight: 2.0 second_stage_classification_loss_weight: 1.0 } } train_config: { batch_size: 1 optimizer { momentum_optimizer: { learning_rate: { manual_step_learning_rate { initial_learning_rate: 0.0002 schedule { step: 900000 learning_rate: .00002 } schedule { step: 1200000 learning_rate: .000002 } } } momentum_optimizer_value: 0.9 } use_moving_average: false } gradient_clipping_by_norm: 10.0 fine_tune_checkpoint: "PATH_TO_BE_CONFIGURED/model.ckpt" from_detection_checkpoint: true # Note: The below line limits the training process to 200K steps, which we # empirically found to be sufficient enough to train the COCO dataset. This # effectively bypasses the learning rate schedule (the learning rate will # never decay). Remove the below line to train indefinitely. num_steps: 200000 data_augmentation_options { random_horizontal_flip { } } } train_input_reader: { tf_record_input_reader { input_path: "PATH_TO_BE_CONFIGURED/mscoco_train.record-?????-of-00100" } label_map_path: "PATH_TO_BE_CONFIGURED/mscoco_label_map.pbtxt" } eval_config: { num_examples: 8000 # Note: The below line limits the evaluation process to 10 evaluations. # Remove the below line to evaluate indefinitely. max_evals: 10 } eval_input_reader: { tf_record_input_reader { input_path: "PATH_TO_BE_CONFIGURED/mscoco_val.record-?????-of-00010" } label_map_path: "PATH_TO_BE_CONFIGURED/mscoco_label_map.pbtxt" shuffle: false num_readers: 1 }
TensorFlow/Detection/SSD/models/research/object_detection/dataset_tools
dataset_tools
oid_hierarchical_labels_expansion
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== r"""An executable to expand hierarchically image-level labels and boxes. Example usage: python models/research/object_detection/dataset_tools/\ oid_hierarchical_labels_expansion.py \ --json_hierarchy_file=<path to JSON hierarchy> \ --input_annotations=<input csv file> \ --output_annotations=<output csv file> \ --annotation_type=<1 (for boxes) or 2 (for image-level labels)> """ from __future__ import print_function import argparse import json def _update_dict(initial_dict, update): """Updates dictionary with update content. Args: initial_dict: initial dictionary. update: updated dictionary. """ for key, value_list in update.iteritems(): if key in initial_dict: initial_dict[key].extend(value_list) else: initial_dict[key] = value_list def _build_plain_hierarchy(hierarchy, skip_root=False): """Expands tree hierarchy representation to parent-child dictionary. Args: hierarchy: labels hierarchy as JSON file. skip_root: if true skips root from the processing (done for the case when all classes under hierarchy are collected under virtual node). Returns: keyed_parent - dictionary of parent - all its children nodes. keyed_child - dictionary of children - all its parent nodes children - all children of the current node. """ all_children = [] all_keyed_parent = {} all_keyed_child = {} if 'Subcategory' in hierarchy: for node in hierarchy['Subcategory']: keyed_parent, keyed_child, children = _build_plain_hierarchy(node) # Update is not done through dict.update() since some children have multi- # ple parents in the hiearchy. _update_dict(all_keyed_parent, keyed_parent) _update_dict(all_keyed_child, keyed_child) all_children.extend(children) if not skip_root: all_keyed_parent[hierarchy['LabelName']] = all_children all_children = [hierarchy['LabelName']] + all_children for child, _ in all_keyed_child.iteritems(): all_keyed_child[child].append(hierarchy['LabelName']) all_keyed_child[hierarchy['LabelName']] = [] return all_keyed_parent, all_keyed_child, all_children class OIDHierarchicalLabelsExpansion(object): """ Main class to perform labels hierachical expansion.""" def __init__(self, hierarchy): """Constructor. Args: hierarchy: labels hierarchy as JSON object. """ self._hierarchy_keyed_parent, self._hierarchy_keyed_child, _ = ( _build_plain_hierarchy(hierarchy, skip_root=True)) def expand_boxes_from_csv(self, csv_row): """Expands a row containing bounding boxes from CSV file. Args: csv_row: a single row of Open Images released groundtruth file. Returns: a list of strings (including the initial row) corresponding to the ground truth expanded to multiple annotation for evaluation with Open Images Challenge 2018 metric. """ # Row header is expected to be exactly: # ImageID,Source,LabelName,Confidence,XMin,XMax,YMin,YMax,IsOccluded, # IsTruncated,IsGroupOf,IsDepiction,IsInside cvs_row_splitted = csv_row.split(',') assert len(cvs_row_splitted) == 13 result = [csv_row] assert cvs_row_splitted[2] in self._hierarchy_keyed_child parent_nodes = self._hierarchy_keyed_child[cvs_row_splitted[2]] for parent_node in parent_nodes: cvs_row_splitted[2] = parent_node result.append(','.join(cvs_row_splitted)) return result def expand_labels_from_csv(self, csv_row): """Expands a row containing bounding boxes from CSV file. Args: csv_row: a single row of Open Images released groundtruth file. Returns: a list of strings (including the initial row) corresponding to the ground truth expanded to multiple annotation for evaluation with Open Images Challenge 2018 metric. """ # Row header is expected to be exactly: # ImageID,Source,LabelName,Confidence cvs_row_splited = csv_row.split(',') assert len(cvs_row_splited) == 4 result = [csv_row] if int(cvs_row_splited[3]) == 1: assert cvs_row_splited[2] in self._hierarchy_keyed_child parent_nodes = self._hierarchy_keyed_child[cvs_row_splited[2]] for parent_node in parent_nodes: cvs_row_splited[2] = parent_node result.append(','.join(cvs_row_splited)) else: assert cvs_row_splited[2] in self._hierarchy_keyed_parent child_nodes = self._hierarchy_keyed_parent[cvs_row_splited[2]] for child_node in child_nodes: cvs_row_splited[2] = child_node result.append(','.join(cvs_row_splited)) return result def main(parsed_args): with open(parsed_args.json_hierarchy_file) as f: hierarchy = json.load(f) expansion_generator = OIDHierarchicalLabelsExpansion(hierarchy) labels_file = False if parsed_args.annotation_type == 2: labels_file = True elif parsed_args.annotation_type != 1: print('--annotation_type expected value is 1 or 2.') return -1 with open(parsed_args.input_annotations, 'r') as source: with open(parsed_args.output_annotations, 'w') as target: header = None for line in source: if not header: header = line target.writelines(header) continue if labels_file: expanded_lines = expansion_generator.expand_labels_from_csv(line) else: expanded_lines = expansion_generator.expand_boxes_from_csv(line) target.writelines(expanded_lines) if __name__ == '__main__': parser = argparse.ArgumentParser( description='Hierarchically expand annotations (excluding root node).') parser.add_argument( '--json_hierarchy_file', required=True, help='Path to the file containing label hierarchy in JSON format.') parser.add_argument( '--input_annotations', required=True, help="""Path to Open Images annotations file (either bounding boxes or image-level labels).""") parser.add_argument( '--output_annotations', required=True, help="""Path to the output file.""") parser.add_argument( '--annotation_type', type=int, required=True, help="""Type of the input annotations: 1 - boxes, 2 - image-level labels""" ) args = parser.parse_args() main(args)
TensorFlow2/LanguageModeling/BERT/official/nlp/modeling/layers
layers
transformer_scaffold_test
# Copyright 2019 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Tests for Keras-based transformer block layer.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import json import numpy as np import tensorflow as tf from tensorflow.python.keras import keras_parameterized # pylint: disable=g-direct-tensorflow-import from official.nlp.modeling.layers import attention from official.nlp.modeling.layers import transformer_scaffold # Test class that wraps a standard attention layer. If this layer is called # at any point, the list passed to the config object will be filled with a # boolean 'True'. We register this class as a Keras serializable so we can # test serialization below. # @tf.keras.utils.register_keras_serializable(package='TestOnly') class ValidatedAttentionLayer(attention.Attention): def __init__(self, call_list, **kwargs): super(ValidatedAttentionLayer, self).__init__(**kwargs) self.list = call_list def call(self, inputs): self.list.append(True) return super(ValidatedAttentionLayer, self).call(inputs) def get_config(self): config = super(ValidatedAttentionLayer, self).get_config() config['call_list'] = [] return config # This decorator runs the test in V1, V2-Eager, and V2-Functional mode. It # guarantees forward compatibility of this code for the V2 switchover. @keras_parameterized.run_all_keras_modes class TransformerLayerTest(keras_parameterized.TestCase): def test_layer_creation(self): sequence_length = 21 width = 80 call_list = [] attention_layer_cfg = { 'num_heads': 10, 'head_size': 8, 'call_list': call_list, } test_layer = transformer_scaffold.TransformerScaffold( attention_cls=ValidatedAttentionLayer, attention_cfg=attention_layer_cfg, num_attention_heads=10, intermediate_size=2048, intermediate_activation='relu') # Create a 3-dimensional input (the first dimension is implicit). data_tensor = tf.keras.Input(shape=(sequence_length, width)) output_tensor = test_layer(data_tensor) # The default output of a transformer layer should be the same as the input. self.assertEqual(data_tensor.shape.as_list(), output_tensor.shape.as_list()) # If call_list[0] exists and is True, the passed layer class was # instantiated from the given config properly. self.assertNotEmpty(call_list) self.assertTrue(call_list[0], "The passed layer class wasn't instantiated.") def test_layer_creation_with_mask(self): sequence_length = 21 width = 80 call_list = [] attention_layer_cfg = { 'num_heads': 10, 'head_size': 8, 'call_list': call_list, } test_layer = transformer_scaffold.TransformerScaffold( attention_cls=ValidatedAttentionLayer, attention_cfg=attention_layer_cfg, num_attention_heads=10, intermediate_size=2048, intermediate_activation='relu') # Create a 3-dimensional input (the first dimension is implicit). data_tensor = tf.keras.Input(shape=(sequence_length, width)) # Create a 2-dimensional input (the first dimension is implicit). mask_tensor = tf.keras.Input(shape=(sequence_length, sequence_length)) output_tensor = test_layer([data_tensor, mask_tensor]) # The default output of a transformer layer should be the same as the input. self.assertEqual(data_tensor.shape.as_list(), output_tensor.shape.as_list()) # If call_list[0] exists and is True, the passed layer class was # instantiated from the given config properly. self.assertNotEmpty(call_list) self.assertTrue(call_list[0], "The passed layer class wasn't instantiated.") def test_layer_creation_with_incorrect_mask_fails(self): sequence_length = 21 width = 80 call_list = [] attention_layer_cfg = { 'num_heads': 10, 'head_size': 8, 'call_list': call_list, } test_layer = transformer_scaffold.TransformerScaffold( attention_cls=ValidatedAttentionLayer, attention_cfg=attention_layer_cfg, num_attention_heads=10, intermediate_size=2048, intermediate_activation='relu') # Create a 3-dimensional input (the first dimension is implicit). data_tensor = tf.keras.Input(shape=(sequence_length, width)) # Create a 2-dimensional input (the first dimension is implicit). mask_tensor = tf.keras.Input(shape=(sequence_length, sequence_length - 3)) with self.assertRaisesRegex(ValueError, 'When passing a mask tensor.*'): _ = test_layer([data_tensor, mask_tensor]) def test_layer_invocation(self): sequence_length = 21 width = 80 call_list = [] attention_layer_cfg = { 'num_heads': 10, 'head_size': 8, 'call_list': call_list, } test_layer = transformer_scaffold.TransformerScaffold( attention_cls=ValidatedAttentionLayer, attention_cfg=attention_layer_cfg, num_attention_heads=10, intermediate_size=2048, intermediate_activation='relu') # Create a 3-dimensional input (the first dimension is implicit). data_tensor = tf.keras.Input(shape=(sequence_length, width)) output_tensor = test_layer(data_tensor) # Create a model from the test layer. model = tf.keras.Model(data_tensor, output_tensor) # Invoke the model on test data. We can't validate the output data itself # (the NN is too complex) but this will rule out structural runtime errors. batch_size = 6 input_data = 10 * np.random.random_sample( (batch_size, sequence_length, width)) _ = model.predict(input_data) # If call_list[0] exists and is True, the passed layer class was # instantiated from the given config properly. self.assertNotEmpty(call_list) self.assertTrue(call_list[0], "The passed layer class wasn't instantiated.") def test_layer_invocation_with_mask(self): sequence_length = 21 width = 80 call_list = [] attention_layer_cfg = { 'num_heads': 10, 'head_size': 8, 'call_list': call_list, } test_layer = transformer_scaffold.TransformerScaffold( attention_cls=ValidatedAttentionLayer, attention_cfg=attention_layer_cfg, num_attention_heads=10, intermediate_size=2048, intermediate_activation='relu') # Create a 3-dimensional input (the first dimension is implicit). data_tensor = tf.keras.Input(shape=(sequence_length, width)) # Create a 2-dimensional input (the first dimension is implicit). mask_tensor = tf.keras.Input(shape=(sequence_length, sequence_length)) output_tensor = test_layer([data_tensor, mask_tensor]) # Create a model from the test layer. model = tf.keras.Model([data_tensor, mask_tensor], output_tensor) # Invoke the model on test data. We can't validate the output data itself # (the NN is too complex) but this will rule out structural runtime errors. batch_size = 6 input_data = 10 * np.random.random_sample( (batch_size, sequence_length, width)) # The attention mask should be of shape (batch, from_seq_len, to_seq_len), # which here is (batch, sequence_length, sequence_length) mask_data = np.random.randint( 2, size=(batch_size, sequence_length, sequence_length)) _ = model.predict([input_data, mask_data]) # If call_list[0] exists and is True, the passed layer class was # instantiated from the given config properly. self.assertNotEmpty(call_list) self.assertTrue(call_list[0], "The passed layer class wasn't instantiated.") def test_layer_invocation_with_float16_dtype(self): sequence_length = 21 width = 80 call_list = [] attention_layer_cfg = { 'num_heads': 10, 'head_size': 8, 'call_list': call_list, } test_layer = transformer_scaffold.TransformerScaffold( attention_cls=ValidatedAttentionLayer, attention_cfg=attention_layer_cfg, num_attention_heads=10, intermediate_size=2048, intermediate_activation='relu', dtype='float16') # Create a 3-dimensional input (the first dimension is implicit). data_tensor = tf.keras.Input( shape=(sequence_length, width), dtype=tf.float16) # Create a 2-dimensional input (the first dimension is implicit). mask_tensor = tf.keras.Input(shape=(sequence_length, sequence_length)) output_tensor = test_layer([data_tensor, mask_tensor]) # Create a model from the test layer. model = tf.keras.Model([data_tensor, mask_tensor], output_tensor) # Invoke the model on test data. We can't validate the output data itself # (the NN is too complex) but this will rule out structural runtime errors. batch_size = 6 input_data = (10 * np.random.random_sample( (batch_size, sequence_length, width))).astype(np.float16) # The attention mask should be of shape (batch, from_seq_len, to_seq_len), # which here is (batch, sequence_length, sequence_length) mask_data = np.random.randint( 2, size=(batch_size, sequence_length, sequence_length)) _ = model.predict([input_data, mask_data]) # If call_list[0] exists and is True, the passed layer class was # instantiated from the given config properly. self.assertNotEmpty(call_list) self.assertTrue(call_list[0], "The passed layer class wasn't instantiated.") def test_transform_with_initializer(self): sequence_length = 21 width = 80 call_list = [] attention_layer_cfg = { 'num_heads': 10, 'head_size': 8, 'call_list': call_list, } test_layer = transformer_scaffold.TransformerScaffold( attention_cls=ValidatedAttentionLayer, attention_cfg=attention_layer_cfg, num_attention_heads=10, intermediate_size=2048, intermediate_activation='relu', kernel_initializer=tf.keras.initializers.TruncatedNormal(stddev=0.02)) # Create a 3-dimensional input (the first dimension is implicit). data_tensor = tf.keras.Input(shape=(sequence_length, width)) output = test_layer(data_tensor) # The default output of a transformer layer should be the same as the input. self.assertEqual(data_tensor.shape.as_list(), output.shape.as_list()) # If call_list[0] exists and is True, the passed layer class was # instantiated from the given config properly. self.assertNotEmpty(call_list) self.assertTrue(call_list[0]) def test_layer_restoration_from_config(self): sequence_length = 21 width = 80 call_list = [] attention_layer_cfg = { 'num_heads': 10, 'head_size': 8, 'call_list': call_list, 'name': 'test_layer', } test_layer = transformer_scaffold.TransformerScaffold( attention_cls=ValidatedAttentionLayer, attention_cfg=attention_layer_cfg, num_attention_heads=10, intermediate_size=2048, intermediate_activation='relu') # Create a 3-dimensional input (the first dimension is implicit). data_tensor = tf.keras.Input(shape=(sequence_length, width)) # Create a 2-dimensional input (the first dimension is implicit). mask_tensor = tf.keras.Input(shape=(sequence_length, sequence_length)) output_tensor = test_layer([data_tensor, mask_tensor]) # Create a model from the test layer. model = tf.keras.Model([data_tensor, mask_tensor], output_tensor) # Invoke the model on test data. We can't validate the output data itself # (the NN is too complex) but this will rule out structural runtime errors. batch_size = 6 input_data = 10 * np.random.random_sample( (batch_size, sequence_length, width)) # The attention mask should be of shape (batch, from_seq_len, to_seq_len), # which here is (batch, sequence_length, sequence_length) mask_data = np.random.randint( 2, size=(batch_size, sequence_length, sequence_length)) pre_serialization_output = model.predict([input_data, mask_data]) # Serialize the model config. Pass the serialized data through json to # ensure that we can serialize this layer to disk. serialized_data = json.dumps(model.get_config()) post_string_serialized_data = json.loads(serialized_data) # Create a new model from the old config, and copy the weights. These models # should have identical outputs. new_model = tf.keras.Model.from_config(post_string_serialized_data) new_model.set_weights(model.get_weights()) output = new_model.predict([input_data, mask_data]) self.assertAllClose(pre_serialization_output, output) # If the layer was configured correctly, it should have a list attribute # (since it should have the custom class and config passed to it). new_model.summary() new_call_list = new_model.get_layer( name='transformer_scaffold')._attention_layer.list self.assertNotEmpty(new_call_list) self.assertTrue(new_call_list[0], "The passed layer class wasn't instantiated.") if __name__ == '__main__': tf.test.main()
TensorFlow/Detection/SSD/models/research/object_detection/dataset_tools
dataset_tools
create_pascal_tf_record
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== r"""Convert raw PASCAL dataset to TFRecord for object_detection. Example usage: python object_detection/dataset_tools/create_pascal_tf_record.py \ --data_dir=/home/user/VOCdevkit \ --year=VOC2012 \ --output_path=/home/user/pascal.record """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import hashlib import io import logging import os from lxml import etree import PIL.Image import tensorflow as tf from object_detection.utils import dataset_util from object_detection.utils import label_map_util flags = tf.app.flags flags.DEFINE_string('data_dir', '', 'Root directory to raw PASCAL VOC dataset.') flags.DEFINE_string('set', 'train', 'Convert training set, validation set or ' 'merged set.') flags.DEFINE_string('annotations_dir', 'Annotations', '(Relative) path to annotations directory.') flags.DEFINE_string('year', 'VOC2007', 'Desired challenge year.') flags.DEFINE_string('output_path', '', 'Path to output TFRecord') flags.DEFINE_string('label_map_path', 'data/pascal_label_map.pbtxt', 'Path to label map proto') flags.DEFINE_boolean('ignore_difficult_instances', False, 'Whether to ignore ' 'difficult instances') FLAGS = flags.FLAGS SETS = ['train', 'val', 'trainval', 'test'] YEARS = ['VOC2007', 'VOC2012', 'merged'] def dict_to_tf_example(data, dataset_directory, label_map_dict, ignore_difficult_instances=False, image_subdirectory='JPEGImages'): """Convert XML derived dict to tf.Example proto. Notice that this function normalizes the bounding box coordinates provided by the raw data. Args: data: dict holding PASCAL XML fields for a single image (obtained by running dataset_util.recursive_parse_xml_to_dict) dataset_directory: Path to root directory holding PASCAL dataset label_map_dict: A map from string label names to integers ids. ignore_difficult_instances: Whether to skip difficult instances in the dataset (default: False). image_subdirectory: String specifying subdirectory within the PASCAL dataset directory holding the actual image data. Returns: example: The converted tf.Example. Raises: ValueError: if the image pointed to by data['filename'] is not a valid JPEG """ img_path = os.path.join(data['folder'], image_subdirectory, data['filename']) full_path = os.path.join(dataset_directory, img_path) with tf.gfile.GFile(full_path, 'rb') as fid: encoded_jpg = fid.read() encoded_jpg_io = io.BytesIO(encoded_jpg) image = PIL.Image.open(encoded_jpg_io) if image.format != 'JPEG': raise ValueError('Image format not JPEG') key = hashlib.sha256(encoded_jpg).hexdigest() width = int(data['size']['width']) height = int(data['size']['height']) xmin = [] ymin = [] xmax = [] ymax = [] classes = [] classes_text = [] truncated = [] poses = [] difficult_obj = [] if 'object' in data: for obj in data['object']: difficult = bool(int(obj['difficult'])) if ignore_difficult_instances and difficult: continue difficult_obj.append(int(difficult)) xmin.append(float(obj['bndbox']['xmin']) / width) ymin.append(float(obj['bndbox']['ymin']) / height) xmax.append(float(obj['bndbox']['xmax']) / width) ymax.append(float(obj['bndbox']['ymax']) / height) classes_text.append(obj['name'].encode('utf8')) classes.append(label_map_dict[obj['name']]) truncated.append(int(obj['truncated'])) poses.append(obj['pose'].encode('utf8')) example = tf.train.Example(features=tf.train.Features(feature={ 'image/height': dataset_util.int64_feature(height), 'image/width': dataset_util.int64_feature(width), 'image/filename': dataset_util.bytes_feature( data['filename'].encode('utf8')), 'image/source_id': dataset_util.bytes_feature( data['filename'].encode('utf8')), 'image/key/sha256': dataset_util.bytes_feature(key.encode('utf8')), 'image/encoded': dataset_util.bytes_feature(encoded_jpg), 'image/format': dataset_util.bytes_feature('jpeg'.encode('utf8')), 'image/object/bbox/xmin': dataset_util.float_list_feature(xmin), 'image/object/bbox/xmax': dataset_util.float_list_feature(xmax), 'image/object/bbox/ymin': dataset_util.float_list_feature(ymin), 'image/object/bbox/ymax': dataset_util.float_list_feature(ymax), 'image/object/class/text': dataset_util.bytes_list_feature(classes_text), 'image/object/class/label': dataset_util.int64_list_feature(classes), 'image/object/difficult': dataset_util.int64_list_feature(difficult_obj), 'image/object/truncated': dataset_util.int64_list_feature(truncated), 'image/object/view': dataset_util.bytes_list_feature(poses), })) return example def main(_): if FLAGS.set not in SETS: raise ValueError('set must be in : {}'.format(SETS)) if FLAGS.year not in YEARS: raise ValueError('year must be in : {}'.format(YEARS)) data_dir = FLAGS.data_dir years = ['VOC2007', 'VOC2012'] if FLAGS.year != 'merged': years = [FLAGS.year] writer = tf.python_io.TFRecordWriter(FLAGS.output_path) label_map_dict = label_map_util.get_label_map_dict(FLAGS.label_map_path) for year in years: logging.info('Reading from PASCAL %s dataset.', year) examples_path = os.path.join(data_dir, year, 'ImageSets', 'Main', 'aeroplane_' + FLAGS.set + '.txt') annotations_dir = os.path.join(data_dir, year, FLAGS.annotations_dir) examples_list = dataset_util.read_examples_list(examples_path) for idx, example in enumerate(examples_list): if idx % 100 == 0: logging.info('On image %d of %d', idx, len(examples_list)) path = os.path.join(annotations_dir, example + '.xml') with tf.gfile.GFile(path, 'r') as fid: xml_str = fid.read() xml = etree.fromstring(xml_str) data = dataset_util.recursive_parse_xml_to_dict(xml)['annotation'] tf_example = dict_to_tf_example(data, FLAGS.data_dir, label_map_dict, FLAGS.ignore_difficult_instances) writer.write(tf_example.SerializeToString()) writer.close() if __name__ == '__main__': tf.app.run()
PyTorch/SpeechRecognition/QuartzNet/common
common
optimizers
# Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from torch.optim import Optimizer import math def lr_policy(step, epoch, initial_lr, optimizer, steps_per_epoch, warmup_epochs, hold_epochs, num_epochs=None, policy='linear', min_lr=1e-5, exp_gamma=None): """ learning rate decay Args: initial_lr: base learning rate step: current iteration number N: total number of iterations over which learning rate is decayed lr_steps: list of steps to apply exp_gamma """ warmup_steps = warmup_epochs * steps_per_epoch hold_steps = hold_epochs * steps_per_epoch if policy == 'legacy': assert num_epochs is not None tot_steps = num_epochs * steps_per_epoch if step < warmup_steps: a = (step + 1) / (warmup_steps + 1) elif step < warmup_steps + hold_steps: a = 1.0 else: a = (((tot_steps - step) / (tot_steps - warmup_steps - hold_steps)) ** 2) elif policy == 'exponential': assert exp_gamma is not None if step < warmup_steps: a = (step + 1) / (warmup_steps + 1) elif step < warmup_steps + hold_steps: a = 1.0 else: a = exp_gamma ** (epoch - warmup_epochs - hold_epochs) else: raise ValueError new_lr = max(a * initial_lr, min_lr) for param_group in optimizer.param_groups: param_group['lr'] = new_lr class AdamW(Optimizer): """Implements AdamW algorithm. It has been proposed in `Adam: A Method for Stochastic Optimization`_. Arguments: params (iterable): iterable of parameters to optimize or dicts defining parameter groups lr (float, optional): learning rate (default: 1e-3) betas (Tuple[float, float], optional): coefficients used for computing running averages of gradient and its square (default: (0.9, 0.999)) eps (float, optional): term added to the denominator to improve numerical stability (default: 1e-8) weight_decay (float, optional): weight decay (L2 penalty) (default: 0) amsgrad (boolean, optional): whether to use the AMSGrad variant of this algorithm from the paper `On the Convergence of Adam and Beyond`_ Adam: A Method for Stochastic Optimization: https://arxiv.org/abs/1412.6980 On the Convergence of Adam and Beyond: https://openreview.net/forum?id=ryQu7f-RZ """ def __init__(self, params, lr=1e-3, betas=(0.9, 0.999), eps=1e-8, weight_decay=0, amsgrad=False): if not 0.0 <= lr: raise ValueError("Invalid learning rate: {}".format(lr)) if not 0.0 <= eps: raise ValueError("Invalid epsilon value: {}".format(eps)) if not 0.0 <= betas[0] < 1.0: raise ValueError("Invalid beta parameter at index 0: {}".format(betas[0])) if not 0.0 <= betas[1] < 1.0: raise ValueError("Invalid beta parameter at index 1: {}".format(betas[1])) defaults = dict(lr=lr, betas=betas, eps=eps, weight_decay=weight_decay, amsgrad=amsgrad) super(AdamW, self).__init__(params, defaults) def __setstate__(self, state): super(AdamW, self).__setstate__(state) for group in self.param_groups: group.setdefault('amsgrad', False) def step(self, closure=None): """Performs a single optimization step. Arguments: closure (callable, optional): A closure that reevaluates the model and returns the loss. """ loss = None if closure is not None: loss = closure() for group in self.param_groups: for p in group['params']: if p.grad is None: continue grad = p.grad.data if grad.is_sparse: raise RuntimeError('Adam does not support sparse gradients, please consider SparseAdam instead') amsgrad = group['amsgrad'] state = self.state[p] # State initialization if len(state) == 0: state['step'] = 0 # Exponential moving average of gradient values state['exp_avg'] = torch.zeros_like(p.data) # Exponential moving average of squared gradient values state['exp_avg_sq'] = torch.zeros_like(p.data) if amsgrad: # Maintains max of all exp. moving avg. of sq. grad. values state['max_exp_avg_sq'] = torch.zeros_like(p.data) exp_avg, exp_avg_sq = state['exp_avg'], state['exp_avg_sq'] if amsgrad: max_exp_avg_sq = state['max_exp_avg_sq'] beta1, beta2 = group['betas'] state['step'] += 1 # Decay the first and second moment running average coefficient exp_avg.mul_(beta1).add_(grad, alpha=1 - beta1) exp_avg_sq.mul_(beta2).addcmul_(1 - beta2, grad, grad) if amsgrad: # Maintains the maximum of all 2nd moment running avg. till now torch.max(max_exp_avg_sq, exp_avg_sq, out=max_exp_avg_sq) # Use the max. for normalizing running avg. of gradient denom = max_exp_avg_sq.sqrt().add_(group['eps']) else: denom = exp_avg_sq.sqrt().add_(group['eps']) bias_correction1 = 1 - beta1 ** state['step'] bias_correction2 = 1 - beta2 ** state['step'] step_size = group['lr'] * math.sqrt(bias_correction2) / bias_correction1 p.data.add_(torch.mul(p.data, group['weight_decay']).addcdiv_(1, exp_avg, denom), alpha=-step_size) return loss class Novograd(Optimizer): """ Implements Novograd algorithm. Args: params (iterable): iterable of parameters to optimize or dicts defining parameter groups lr (float, optional): learning rate (default: 1e-3) betas (Tuple[float, float], optional): coefficients used for computing running averages of gradient and its square (default: (0.95, 0)) eps (float, optional): term added to the denominator to improve numerical stability (default: 1e-8) weight_decay (float, optional): weight decay (L2 penalty) (default: 0) grad_averaging: gradient averaging amsgrad (boolean, optional): whether to use the AMSGrad variant of this algorithm from the paper `On the Convergence of Adam and Beyond`_ (default: False) """ def __init__(self, params, lr=1e-3, betas=(0.95, 0), eps=1e-8, weight_decay=0, grad_averaging=False, amsgrad=False): if not 0.0 <= lr: raise ValueError("Invalid learning rate: {}".format(lr)) if not 0.0 <= eps: raise ValueError("Invalid epsilon value: {}".format(eps)) if not 0.0 <= betas[0] < 1.0: raise ValueError("Invalid beta parameter at index 0: {}".format(betas[0])) if not 0.0 <= betas[1] < 1.0: raise ValueError("Invalid beta parameter at index 1: {}".format(betas[1])) defaults = dict(lr=lr, betas=betas, eps=eps, weight_decay=weight_decay, grad_averaging=grad_averaging, amsgrad=amsgrad) super(Novograd, self).__init__(params, defaults) def __setstate__(self, state): super(Novograd, self).__setstate__(state) for group in self.param_groups: group.setdefault('amsgrad', False) def step(self, closure=None): """Performs a single optimization step. Arguments: closure (callable, optional): A closure that reevaluates the model and returns the loss. """ loss = None if closure is not None: loss = closure() for group in self.param_groups: for p in group['params']: if p.grad is None: continue grad = p.grad.data if grad.is_sparse: raise RuntimeError('Sparse gradients are not supported.') amsgrad = group['amsgrad'] state = self.state[p] # State initialization if len(state) == 0: state['step'] = 0 # Exponential moving average of gradient values state['exp_avg'] = torch.zeros_like(p.data) # Exponential moving average of squared gradient values state['exp_avg_sq'] = torch.zeros([]).to(state['exp_avg'].device) if amsgrad: # Maintains max of all exp. moving avg. of sq. grad. values state['max_exp_avg_sq'] = torch.zeros([]).to(state['exp_avg'].device) exp_avg, exp_avg_sq = state['exp_avg'], state['exp_avg_sq'] if amsgrad: max_exp_avg_sq = state['max_exp_avg_sq'] beta1, beta2 = group['betas'] state['step'] += 1 norm = torch.sum(torch.pow(grad, 2)) if exp_avg_sq == 0: exp_avg_sq.copy_(norm) else: exp_avg_sq.mul_(beta2).add_(norm, alpha=1 - beta2) if amsgrad: # Maintains the maximum of all 2nd moment running avg. till now torch.max(max_exp_avg_sq, exp_avg_sq, out=max_exp_avg_sq) # Use the max. for normalizing running avg. of gradient denom = max_exp_avg_sq.sqrt().add_(group['eps']) else: denom = exp_avg_sq.sqrt().add_(group['eps']) grad.div_(denom) if group['weight_decay'] != 0: grad.add_(p.data, alpha=group['weight_decay']) if group['grad_averaging']: grad.mul_(1 - beta1) exp_avg.mul_(beta1).add_(grad) p.data.add_(exp_avg, alpha=-group['lr']) return loss
PyTorch/LanguageModeling/BERT/data
data
WikiDownloader
# Copyright (c) 2019-2021 NVIDIA CORPORATION. All rights reserved. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import bz2 import os import urllib.request import subprocess import sys import subprocess class WikiDownloader: def __init__(self, language, save_path): self.save_path = save_path + '/wikicorpus_' + language if not os.path.exists(self.save_path): os.makedirs(self.save_path) self.language = language # Use a mirror from https://dumps.wikimedia.org/mirrors.html if the below links do not work self.download_urls = { 'en' : 'https://dumps.wikimedia.org/enwiki/latest/enwiki-latest-pages-articles.xml.bz2', 'zh' : 'https://dumps.wikimedia.org/zhwiki/latest/zhwiki-latest-pages-articles.xml.bz2' } self.output_files = { 'en' : 'wikicorpus_en.xml.bz2', 'zh' : 'wikicorpus_zh.xml.bz2' } def download(self): if self.language in self.download_urls: url = self.download_urls[self.language] filename = self.output_files[self.language] print('Downloading:', url) if os.path.isfile(self.save_path + '/' + filename): print('** Download file already exists, skipping download') else: cmd = ['wget', url, '--output-document={}'.format(self.save_path + '/' + filename)] print('Running:', cmd) status = subprocess.run(cmd) if status.returncode != 0: raise RuntimeError('Wiki download not successful') # Always unzipping since this is relatively fast and will overwrite print('Unzipping:', self.output_files[self.language]) subprocess.run('bzip2 -dk ' + self.save_path + '/' + filename, shell=True, check=True) else: assert False, 'WikiDownloader not implemented for this language yet.'
TensorFlow/Segmentation/UNet_Industrial/scripts/benchmarking
benchmarking
UNet_evalbench
#!/usr/bin/env bash # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # This script launches UNet evaluation benchmark in FP32/TF32 on 1 GPUs using 16 batch size # Usage ./UNet_evalbench.sh <path to dataset> <dagm classID (1-10)> BASEDIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" >/dev/null 2>&1 && pwd )" export TF_CPP_MIN_LOG_LEVEL=3 # Cleaning up for benchmark RESULT_DIR="/tmp" rm -rf "${RESULT_DIR}" python "${BASEDIR}/../../main.py" \ --unet_variant='tinyUNet' \ --activation_fn='relu' \ --exec_mode='inference_benchmark' \ --iter_unit='batch' \ --num_iter=1500 \ --batch_size=16 \ --warmup_step=500 \ --results_dir="${RESULT_DIR}" \ --data_dir="${1}" \ --dataset_name='DAGM2007' \ --dataset_classID="${2}" \ --data_format='NCHW' \ --use_auto_loss_scaling \ --noamp \ --xla \ --learning_rate=1e-4 \ --learning_rate_decay_factor=0.8 \ --learning_rate_decay_steps=500 \ --rmsprop_decay=0.9 \ --rmsprop_momentum=0.8 \ --loss_fn_name='adaptive_loss' \ --weight_decay=1e-5 \ --weight_init_method='he_uniform' \ --augment_data \ --display_every=250 \ --debug_verbosity=0
PyTorch/Detection/SSD/examples
examples
SSD300_A100_FP16_1GPU
# This script launches SSD300 training in FP16 on 1 GPUs using 256 batch size # Usage bash SSD300_FP16_1GPU.sh <path to this repository> <path to dataset> <additional flags> python $1/main.py --backbone resnet50 --warmup 300 --bs 256 --data $2 ${@:3}
TensorFlow/Detection/SSD/models/research/object_detection/builders
builders
box_coder_builder
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """A function to build an object detection box coder from configuration.""" from object_detection.box_coders import faster_rcnn_box_coder from object_detection.box_coders import keypoint_box_coder from object_detection.box_coders import mean_stddev_box_coder from object_detection.box_coders import square_box_coder from object_detection.protos import box_coder_pb2 def build(box_coder_config): """Builds a box coder object based on the box coder config. Args: box_coder_config: A box_coder.proto object containing the config for the desired box coder. Returns: BoxCoder based on the config. Raises: ValueError: On empty box coder proto. """ if not isinstance(box_coder_config, box_coder_pb2.BoxCoder): raise ValueError('box_coder_config not of type box_coder_pb2.BoxCoder.') if box_coder_config.WhichOneof('box_coder_oneof') == 'faster_rcnn_box_coder': return faster_rcnn_box_coder.FasterRcnnBoxCoder(scale_factors=[ box_coder_config.faster_rcnn_box_coder.y_scale, box_coder_config.faster_rcnn_box_coder.x_scale, box_coder_config.faster_rcnn_box_coder.height_scale, box_coder_config.faster_rcnn_box_coder.width_scale ]) if box_coder_config.WhichOneof('box_coder_oneof') == 'keypoint_box_coder': return keypoint_box_coder.KeypointBoxCoder( box_coder_config.keypoint_box_coder.num_keypoints, scale_factors=[ box_coder_config.keypoint_box_coder.y_scale, box_coder_config.keypoint_box_coder.x_scale, box_coder_config.keypoint_box_coder.height_scale, box_coder_config.keypoint_box_coder.width_scale ]) if (box_coder_config.WhichOneof('box_coder_oneof') == 'mean_stddev_box_coder'): return mean_stddev_box_coder.MeanStddevBoxCoder( stddev=box_coder_config.mean_stddev_box_coder.stddev) if box_coder_config.WhichOneof('box_coder_oneof') == 'square_box_coder': return square_box_coder.SquareBoxCoder(scale_factors=[ box_coder_config.square_box_coder.y_scale, box_coder_config.square_box_coder.x_scale, box_coder_config.square_box_coder.length_scale ]) raise ValueError('Empty box coder.')
PyTorch/SpeechSynthesis/Tacotron2/notebooks/conversationalai/client/speech_ai_demo/utils/tacotron2
tacotron2
cleaners
""" from https://github.com/keithito/tacotron """ ''' Cleaners are transformations that run over the input text at both training and eval time. Cleaners can be selected by passing a comma-delimited list of cleaner names as the "cleaners" hyperparameter. Some cleaners are English-specific. You'll typically want to use: 1. "english_cleaners" for English text 2. "transliteration_cleaners" for non-English text that can be transliterated to ASCII using the Unidecode library (https://pypi.python.org/pypi/Unidecode) 3. "basic_cleaners" if you do not want to transliterate (in this case, you should also update the symbols in symbols.py to match your data). ''' import re from .numbers import normalize_numbers from .unidecoder import unidecoder # Regular expression matching whitespace: _whitespace_re = re.compile(r'\s+') # List of (regular expression, replacement) pairs for abbreviations: _abbreviations = [(re.compile('\\b%s\\.' % x[0], re.IGNORECASE), x[1]) for x in [ ('mrs', 'misess'), ('mr', 'mister'), ('dr', 'doctor'), ('st', 'saint'), ('co', 'company'), ('jr', 'junior'), ('maj', 'major'), ('gen', 'general'), ('drs', 'doctors'), ('rev', 'reverend'), ('lt', 'lieutenant'), ('hon', 'honorable'), ('sgt', 'sergeant'), ('capt', 'captain'), ('esq', 'esquire'), ('ltd', 'limited'), ('col', 'colonel'), ('ft', 'fort'), ]] def expand_abbreviations(text): for regex, replacement in _abbreviations: text = re.sub(regex, replacement, text) return text def expand_numbers(text): return normalize_numbers(text) def lowercase(text): return text.lower() def collapse_whitespace(text): return re.sub(_whitespace_re, ' ', text) def convert_to_ascii(text): return unidecoder(text) def basic_cleaners(text): '''Basic pipeline that lowercases and collapses whitespace without transliteration.''' text = lowercase(text) text = collapse_whitespace(text) return text def transliteration_cleaners(text): '''Pipeline for non-English text that transliterates to ASCII.''' text = convert_to_ascii(text) text = lowercase(text) text = collapse_whitespace(text) return text def english_cleaners(text): '''Pipeline for English text, including number and abbreviation expansion.''' text = convert_to_ascii(text) text = lowercase(text) text = expand_numbers(text) text = expand_abbreviations(text) text = collapse_whitespace(text) return text
PyTorch/Segmentation/MaskRCNN/pytorch
pytorch
CODE_OF_CONDUCT
# Code of Conduct Facebook has adopted a Code of Conduct that we expect project participants to adhere to. Please read the [full text](https://code.fb.com/codeofconduct/) so that you can understand what actions will and will not be tolerated.
PyTorch/Detection/Efficientdet
Efficientdet
.gitignore
# Byte-compiled / optimized / DLL files __pycache__/ *.py[cod] *$py.class # C extensions *.so # Distribution / packaging .Python build/ develop-eggs/ dist/ downloads/ eggs/ .eggs/ lib/ lib64/ parts/ sdist/ var/ wheels/ pip-wheel-metadata/ share/python-wheels/ *.egg-info/ .installed.cfg *.egg MANIFEST # PyInstaller # Usually these files are written by a python script from a template # before PyInstaller builds the exe, so as to inject date/other infos into it. *.manifest *.spec # Installer logs pip-log.txt pip-delete-this-directory.txt # Unit test / coverage reports htmlcov/ .tox/ .nox/ .coverage .coverage.* .cache nosetests.xml coverage.xml *.cover *.py,cover .hypothesis/ .pytest_cache/ # Translations *.mo *.pot # Django stuff: *.log local_settings.py db.sqlite3 db.sqlite3-journal # Flask stuff: instance/ .webassets-cache # Scrapy stuff: .scrapy # Sphinx documentation docs/_build/ # PyBuilder target/ # Jupyter Notebook .ipynb_checkpoints # IPython profile_default/ ipython_config.py # pyenv .python-version # pipenv # According to pypa/pipenv#598, it is recommended to include Pipfile.lock in version control. # However, in case of collaboration, if having platform-specific dependencies or dependencies # having no cross-platform support, pipenv may install dependencies that don't work, or not # install all needed dependencies. #Pipfile.lock # PEP 582; used by e.g. github.com/David-OConnor/pyflow __pypackages__/ # Celery stuff celerybeat-schedule celerybeat.pid # SageMath parsed files *.sage.py # Environments .env .venv env/ venv/ ENV/ env.bak/ venv.bak/ # Spyder project settings .spyderproject .spyproject # Rope project settings .ropeproject # mkdocs documentation /site # mypy .mypy_cache/ .dmypy.json dmypy.json # Pyre type checker .pyre/ # PyCharm .idea # PyTorch weights *.tar *.pth *.gz
PyTorch/SpeechRecognition/wav2vec2
wav2vec2
requirements
editdistance==0.6.0 librosa==0.10.1 omegaconf==2.0.6 # optional for handling certain Fairseq ckpts pyarrow==6.0.1 soundfile==0.12.1 sox==1.4.1 tqdm==4.53.0 git+https://github.com/NVIDIA/dllogger@v1.0.0#egg=dllogger
TensorFlow/Detection/SSD/models/research/object_detection/anchor_generators
anchor_generators
multiscale_grid_anchor_generator
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Generates grid anchors on the fly corresponding to multiple CNN layers. Generates grid anchors on the fly corresponding to multiple CNN layers as described in: "Focal Loss for Dense Object Detection" (https://arxiv.org/abs/1708.02002) T.-Y. Lin, P. Goyal, R. Girshick, K. He, P. Dollar """ from object_detection.anchor_generators import grid_anchor_generator from object_detection.core import anchor_generator from object_detection.core import box_list_ops class MultiscaleGridAnchorGenerator(anchor_generator.AnchorGenerator): """Generate a grid of anchors for multiple CNN layers of different scale.""" def __init__(self, min_level, max_level, anchor_scale, aspect_ratios, scales_per_octave, normalize_coordinates=True): """Constructs a MultiscaleGridAnchorGenerator. To construct anchors, at multiple scale resolutions, one must provide a the minimum level and maximum levels on a scale pyramid. To define the size of anchor, the anchor scale is provided to decide the size relatively to the stride of the corresponding feature map. The generator allows one pixel location on feature map maps to multiple anchors, that have different aspect ratios and intermediate scales. Args: min_level: minimum level in feature pyramid. max_level: maximum level in feature pyramid. anchor_scale: anchor scale and feature stride define the size of the base anchor on an image. For example, given a feature pyramid with strides [2^3, ..., 2^7] and anchor scale 4. The base anchor size is 4 * [2^3, ..., 2^7]. aspect_ratios: list or tuple of (float) aspect ratios to place on each grid point. scales_per_octave: integer number of intermediate scales per scale octave. normalize_coordinates: whether to produce anchors in normalized coordinates. (defaults to True). """ self._anchor_grid_info = [] self._aspect_ratios = aspect_ratios self._scales_per_octave = scales_per_octave self._normalize_coordinates = normalize_coordinates scales = [2**(float(scale) / scales_per_octave) for scale in range(scales_per_octave)] aspects = list(aspect_ratios) for level in range(min_level, max_level + 1): anchor_stride = [2**level, 2**level] base_anchor_size = [2**level * anchor_scale, 2**level * anchor_scale] self._anchor_grid_info.append({ 'level': level, 'info': [scales, aspects, base_anchor_size, anchor_stride] }) def name_scope(self): return 'MultiscaleGridAnchorGenerator' def num_anchors_per_location(self): """Returns the number of anchors per spatial location. Returns: a list of integers, one for each expected feature map to be passed to the Generate function. """ return len(self._anchor_grid_info) * [ len(self._aspect_ratios) * self._scales_per_octave] def _generate(self, feature_map_shape_list, im_height=1, im_width=1): """Generates a collection of bounding boxes to be used as anchors. Currently we require the input image shape to be statically defined. That is, im_height and im_width should be integers rather than tensors. Args: feature_map_shape_list: list of pairs of convnet layer resolutions in the format [(height_0, width_0), (height_1, width_1), ...]. For example, setting feature_map_shape_list=[(8, 8), (7, 7)] asks for anchors that correspond to an 8x8 layer followed by a 7x7 layer. im_height: the height of the image to generate the grid for. If both im_height and im_width are 1, anchors can only be generated in absolute coordinates. im_width: the width of the image to generate the grid for. If both im_height and im_width are 1, anchors can only be generated in absolute coordinates. Returns: boxes_list: a list of BoxLists each holding anchor boxes corresponding to the input feature map shapes. Raises: ValueError: if im_height and im_width are not integers. ValueError: if im_height and im_width are 1, but normalized coordinates were requested. """ anchor_grid_list = [] for feat_shape, grid_info in zip(feature_map_shape_list, self._anchor_grid_info): # TODO(rathodv) check the feature_map_shape_list is consistent with # self._anchor_grid_info level = grid_info['level'] stride = 2**level scales, aspect_ratios, base_anchor_size, anchor_stride = grid_info['info'] feat_h = feat_shape[0] feat_w = feat_shape[1] anchor_offset = [0, 0] if isinstance(im_height, int) and isinstance(im_width, int): if im_height % 2.0**level == 0 or im_height == 1: anchor_offset[0] = stride / 2.0 if im_width % 2.0**level == 0 or im_width == 1: anchor_offset[1] = stride / 2.0 ag = grid_anchor_generator.GridAnchorGenerator( scales, aspect_ratios, base_anchor_size=base_anchor_size, anchor_stride=anchor_stride, anchor_offset=anchor_offset) (anchor_grid,) = ag.generate(feature_map_shape_list=[(feat_h, feat_w)]) if self._normalize_coordinates: if im_height == 1 or im_width == 1: raise ValueError( 'Normalized coordinates were requested upon construction of the ' 'MultiscaleGridAnchorGenerator, but a subsequent call to ' 'generate did not supply dimension information.') anchor_grid = box_list_ops.to_normalized_coordinates( anchor_grid, im_height, im_width, check_range=False) anchor_grid_list.append(anchor_grid) return anchor_grid_list
TensorFlow/Detection/SSD/models/research/object_detection/models
models
faster_rcnn_nas_feature_extractor
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """NASNet Faster R-CNN implementation. Learning Transferable Architectures for Scalable Image Recognition Barret Zoph, Vijay Vasudevan, Jonathon Shlens, Quoc V. Le https://arxiv.org/abs/1707.07012 """ import tensorflow as tf from object_detection.meta_architectures import faster_rcnn_meta_arch from nets.nasnet import nasnet from nets.nasnet import nasnet_utils arg_scope = tf.contrib.framework.arg_scope slim = tf.contrib.slim def nasnet_large_arg_scope_for_detection(is_batch_norm_training=False): """Defines the default arg scope for the NASNet-A Large for object detection. This provides a small edit to switch batch norm training on and off. Args: is_batch_norm_training: Boolean indicating whether to train with batch norm. Returns: An `arg_scope` to use for the NASNet Large Model. """ imagenet_scope = nasnet.nasnet_large_arg_scope() with arg_scope(imagenet_scope): with arg_scope([slim.batch_norm], is_training=is_batch_norm_training) as sc: return sc # Note: This is largely a copy of _build_nasnet_base inside nasnet.py but # with special edits to remove instantiation of the stem and the special # ability to receive as input a pair of hidden states. def _build_nasnet_base(hidden_previous, hidden, normal_cell, reduction_cell, hparams, true_cell_num, start_cell_num): """Constructs a NASNet image model.""" # Find where to place the reduction cells or stride normal cells reduction_indices = nasnet_utils.calc_reduction_layers( hparams.num_cells, hparams.num_reduction_layers) # Note: The None is prepended to match the behavior of _imagenet_stem() cell_outputs = [None, hidden_previous, hidden] net = hidden # NOTE: In the nasnet.py code, filter_scaling starts at 1.0. We instead # start at 2.0 because 1 reduction cell has been created which would # update the filter_scaling to 2.0. filter_scaling = 2.0 # Run the cells for cell_num in range(start_cell_num, hparams.num_cells): stride = 1 if hparams.skip_reduction_layer_input: prev_layer = cell_outputs[-2] if cell_num in reduction_indices: filter_scaling *= hparams.filter_scaling_rate net = reduction_cell( net, scope='reduction_cell_{}'.format(reduction_indices.index(cell_num)), filter_scaling=filter_scaling, stride=2, prev_layer=cell_outputs[-2], cell_num=true_cell_num) true_cell_num += 1 cell_outputs.append(net) if not hparams.skip_reduction_layer_input: prev_layer = cell_outputs[-2] net = normal_cell( net, scope='cell_{}'.format(cell_num), filter_scaling=filter_scaling, stride=stride, prev_layer=prev_layer, cell_num=true_cell_num) true_cell_num += 1 cell_outputs.append(net) # Final nonlinearity. # Note that we have dropped the final pooling, dropout and softmax layers # from the default nasnet version. with tf.variable_scope('final_layer'): net = tf.nn.relu(net) return net # TODO(shlens): Only fixed_shape_resizer is currently supported for NASNet # featurization. The reason for this is that nasnet.py only supports # inputs with fully known shapes. We need to update nasnet.py to handle # shapes not known at compile time. class FasterRCNNNASFeatureExtractor( faster_rcnn_meta_arch.FasterRCNNFeatureExtractor): """Faster R-CNN with NASNet-A feature extractor implementation.""" def __init__(self, is_training, first_stage_features_stride, batch_norm_trainable=False, reuse_weights=None, weight_decay=0.0): """Constructor. Args: is_training: See base class. first_stage_features_stride: See base class. batch_norm_trainable: See base class. reuse_weights: See base class. weight_decay: See base class. Raises: ValueError: If `first_stage_features_stride` is not 16. """ if first_stage_features_stride != 16: raise ValueError('`first_stage_features_stride` must be 16.') super(FasterRCNNNASFeatureExtractor, self).__init__( is_training, first_stage_features_stride, batch_norm_trainable, reuse_weights, weight_decay) def preprocess(self, resized_inputs): """Faster R-CNN with NAS preprocessing. Maps pixel values to the range [-1, 1]. Args: resized_inputs: A [batch, height_in, width_in, channels] float32 tensor representing a batch of images with values between 0 and 255.0. Returns: preprocessed_inputs: A [batch, height_out, width_out, channels] float32 tensor representing a batch of images. """ return (2.0 / 255.0) * resized_inputs - 1.0 def _extract_proposal_features(self, preprocessed_inputs, scope): """Extracts first stage RPN features. Extracts features using the first half of the NASNet network. We construct the network in `align_feature_maps=True` mode, which means that all VALID paddings in the network are changed to SAME padding so that the feature maps are aligned. Args: preprocessed_inputs: A [batch, height, width, channels] float32 tensor representing a batch of images. scope: A scope name. Returns: rpn_feature_map: A tensor with shape [batch, height, width, depth] end_points: A dictionary mapping feature extractor tensor names to tensors Raises: ValueError: If the created network is missing the required activation. """ del scope if len(preprocessed_inputs.get_shape().as_list()) != 4: raise ValueError('`preprocessed_inputs` must be 4 dimensional, got a ' 'tensor of shape %s' % preprocessed_inputs.get_shape()) with slim.arg_scope(nasnet_large_arg_scope_for_detection( is_batch_norm_training=self._train_batch_norm)): with arg_scope([slim.conv2d, slim.batch_norm, slim.separable_conv2d], reuse=self._reuse_weights): _, end_points = nasnet.build_nasnet_large( preprocessed_inputs, num_classes=None, is_training=self._is_training, final_endpoint='Cell_11') # Note that both 'Cell_10' and 'Cell_11' have equal depth = 2016. rpn_feature_map = tf.concat([end_points['Cell_10'], end_points['Cell_11']], 3) # nasnet.py does not maintain the batch size in the first dimension. # This work around permits us retaining the batch for below. batch = preprocessed_inputs.get_shape().as_list()[0] shape_without_batch = rpn_feature_map.get_shape().as_list()[1:] rpn_feature_map_shape = [batch] + shape_without_batch rpn_feature_map.set_shape(rpn_feature_map_shape) return rpn_feature_map, end_points def _extract_box_classifier_features(self, proposal_feature_maps, scope): """Extracts second stage box classifier features. This function reconstructs the "second half" of the NASNet-A network after the part defined in `_extract_proposal_features`. Args: proposal_feature_maps: A 4-D float tensor with shape [batch_size * self.max_num_proposals, crop_height, crop_width, depth] representing the feature map cropped to each proposal. scope: A scope name. Returns: proposal_classifier_features: A 4-D float tensor with shape [batch_size * self.max_num_proposals, height, width, depth] representing box classifier features for each proposal. """ del scope # Note that we always feed into 2 layers of equal depth # where the first N channels corresponds to previous hidden layer # and the second N channels correspond to the final hidden layer. hidden_previous, hidden = tf.split(proposal_feature_maps, 2, axis=3) # Note that what follows is largely a copy of build_nasnet_large() within # nasnet.py. We are copying to minimize code pollution in slim. # TODO(shlens,skornblith): Determine the appropriate drop path schedule. # For now the schedule is the default (1.0->0.7 over 250,000 train steps). hparams = nasnet.large_imagenet_config() if not self._is_training: hparams.set_hparam('drop_path_keep_prob', 1.0) # Calculate the total number of cells in the network # -- Add 2 for the reduction cells. total_num_cells = hparams.num_cells + 2 # -- And add 2 for the stem cells for ImageNet training. total_num_cells += 2 normal_cell = nasnet_utils.NasNetANormalCell( hparams.num_conv_filters, hparams.drop_path_keep_prob, total_num_cells, hparams.total_training_steps) reduction_cell = nasnet_utils.NasNetAReductionCell( hparams.num_conv_filters, hparams.drop_path_keep_prob, total_num_cells, hparams.total_training_steps) with arg_scope([slim.dropout, nasnet_utils.drop_path], is_training=self._is_training): with arg_scope([slim.batch_norm], is_training=self._train_batch_norm): with arg_scope([slim.avg_pool2d, slim.max_pool2d, slim.conv2d, slim.batch_norm, slim.separable_conv2d, nasnet_utils.factorized_reduction, nasnet_utils.global_avg_pool, nasnet_utils.get_channel_index, nasnet_utils.get_channel_dim], data_format=hparams.data_format): # This corresponds to the cell number just past 'Cell_11' used by # by _extract_proposal_features(). start_cell_num = 12 # Note that this number equals: # start_cell_num + 2 stem cells + 1 reduction cell true_cell_num = 15 with slim.arg_scope(nasnet.nasnet_large_arg_scope()): net = _build_nasnet_base(hidden_previous, hidden, normal_cell=normal_cell, reduction_cell=reduction_cell, hparams=hparams, true_cell_num=true_cell_num, start_cell_num=start_cell_num) proposal_classifier_features = net return proposal_classifier_features def restore_from_classification_checkpoint_fn( self, first_stage_feature_extractor_scope, second_stage_feature_extractor_scope): """Returns a map of variables to load from a foreign checkpoint. Note that this overrides the default implementation in faster_rcnn_meta_arch.FasterRCNNFeatureExtractor which does not work for NASNet-A checkpoints. Args: first_stage_feature_extractor_scope: A scope name for the first stage feature extractor. second_stage_feature_extractor_scope: A scope name for the second stage feature extractor. Returns: A dict mapping variable names (to load from a checkpoint) to variables in the model graph. """ # Note that the NAS checkpoint only contains the moving average version of # the Variables so we need to generate an appropriate dictionary mapping. variables_to_restore = {} for variable in tf.global_variables(): if variable.op.name.startswith( first_stage_feature_extractor_scope): var_name = variable.op.name.replace( first_stage_feature_extractor_scope + '/', '') var_name += '/ExponentialMovingAverage' variables_to_restore[var_name] = variable if variable.op.name.startswith( second_stage_feature_extractor_scope): var_name = variable.op.name.replace( second_stage_feature_extractor_scope + '/', '') var_name += '/ExponentialMovingAverage' variables_to_restore[var_name] = variable return variables_to_restore
DGLPyTorch/DrugDiscovery/SE3Transformer/se3_transformer/model
model
fiber
# Copyright (c) 2021-2022, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # Permission is hereby granted, free of charge, to any person obtaining a # copy of this software and associated documentation files (the "Software"), # to deal in the Software without restriction, including without limitation # the rights to use, copy, modify, merge, publish, distribute, sublicense, # and/or sell copies of the Software, and to permit persons to whom the # Software is furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL # THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER # DEALINGS IN THE SOFTWARE. # # SPDX-FileCopyrightText: Copyright (c) 2021-2022 NVIDIA CORPORATION & AFFILIATES # SPDX-License-Identifier: MIT from collections import namedtuple from itertools import product from typing import Dict import torch from torch import Tensor from se3_transformer.runtime.utils import degree_to_dim FiberEl = namedtuple('FiberEl', ['degree', 'channels']) class Fiber(dict): """ Describes the structure of some set of features. Features are split into types (0, 1, 2, 3, ...). A feature of type k has a dimension of 2k+1. Type-0 features: invariant scalars Type-1 features: equivariant 3D vectors Type-2 features: equivariant symmetric traceless matrices ... As inputs to a SE3 layer, there can be many features of the same types, and many features of different types. The 'multiplicity' or 'number of channels' is the number of features of a given type. This class puts together all the degrees and their multiplicities in order to describe the inputs, outputs or hidden features of SE3 layers. """ def __init__(self, structure): if isinstance(structure, dict): structure = [FiberEl(int(d), int(m)) for d, m in sorted(structure.items(), key=lambda x: x[1])] elif not isinstance(structure[0], FiberEl): structure = list(map(lambda t: FiberEl(*t), sorted(structure, key=lambda x: x[1]))) self.structure = structure super().__init__({d: m for d, m in self.structure}) @property def degrees(self): return sorted([t.degree for t in self.structure]) @property def channels(self): return [self[d] for d in self.degrees] @property def num_features(self): """ Size of the resulting tensor if all features were concatenated together """ return sum(t.channels * degree_to_dim(t.degree) for t in self.structure) @staticmethod def create(num_degrees: int, num_channels: int): """ Create a Fiber with degrees 0..num_degrees-1, all with the same multiplicity """ return Fiber([(degree, num_channels) for degree in range(num_degrees)]) @staticmethod def from_features(feats: Dict[str, Tensor]): """ Infer the Fiber structure from a feature dict """ structure = {} for k, v in feats.items(): degree = int(k) assert len(v.shape) == 3, 'Feature shape should be (N, C, 2D+1)' assert v.shape[-1] == degree_to_dim(degree) structure[degree] = v.shape[-2] return Fiber(structure) def __getitem__(self, degree: int): """ fiber[degree] returns the multiplicity for this degree """ return dict(self.structure).get(degree, 0) def __iter__(self): """ Iterate over namedtuples (degree, channels) """ return iter(self.structure) def __mul__(self, other): """ If other in an int, multiplies all the multiplicities by other. If other is a fiber, returns the cartesian product. """ if isinstance(other, Fiber): return product(self.structure, other.structure) elif isinstance(other, int): return Fiber({t.degree: t.channels * other for t in self.structure}) def __add__(self, other): """ If other in an int, add other to all the multiplicities. If other is a fiber, add the multiplicities of the fibers together. """ if isinstance(other, Fiber): return Fiber({t.degree: t.channels + other[t.degree] for t in self.structure}) elif isinstance(other, int): return Fiber({t.degree: t.channels + other for t in self.structure}) def __repr__(self): return str(self.structure) @staticmethod def combine_max(f1, f2): """ Combine two fiber by taking the maximum multiplicity for each degree in both fibers """ new_dict = dict(f1.structure) for k, m in f2.structure: new_dict[k] = max(new_dict.get(k, 0), m) return Fiber(list(new_dict.items())) @staticmethod def combine_selectively(f1, f2): """ Combine two fiber by taking the sum of multiplicities for each degree in the first fiber """ # only use orders which occur in fiber f1 new_dict = dict(f1.structure) for k in f1.degrees: if k in f2.degrees: new_dict[k] += f2[k] return Fiber(list(new_dict.items())) def to_attention_heads(self, tensors: Dict[str, Tensor], num_heads: int): # dict(N, num_channels, 2d+1) -> (N, num_heads, -1) fibers = [tensors[str(degree)].reshape(*tensors[str(degree)].shape[:-2], num_heads, -1) for degree in self.degrees] fibers = torch.cat(fibers, -1) return fibers
Tools/DGLPyTorch/SyntheticGraphGeneration/demos/performance
performance
struct_generator
#!/usr/bin/env python # coding: utf-8 # Copyright 2023 NVIDIA Corporation. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== # # Graph structure generation demo # ## Overview # # In this notebbok we compare the performance (throughput) of graph structure generators presented in the SynGen tool. # # Available generators: # # 1. [Exact RMAT generator (GPU)](#1) # 1. [Exact RMAT generator (CPU)](#2) # 1. [Approximate RMAT generator (GPU)](#3) # 1. [Approximate RMAT generator (CPU)](#4) # In[1]: # Generator from syngen.generator.graph import RMATGenerator # Others import numpy as np import pandas as pd from matplotlib import pyplot as plt from matplotlib.pyplot import set_loglevel set_loglevel('warning') import time from itertools import product # In[2]: def get_xy(data): x = [edges for edges, _, _ in data] y = [eps for _, eps, _ in data] return x, y # ## Exact generator # <a id="1"></a> # ### GPU # In[3]: static_graph_generator=RMATGenerator() static_graph_generator.gpu=True # In[4]: static_graph_generator._fit_results = 0.4, 0.25, 0.2, 0.15 # In[5]: get_ipython().run_cell_magic('time', '', '\nstart = time.perf_counter()\ndata_proper = static_graph_generator.generate(num_nodes=6_797_556, \n num_edges=168_114, \n is_directed=False, \n has_self_loop=False)\nelapsed = time.perf_counter() - start\n\nprint(elapsed)\n') # In[6]: n_range = [15, 20, 30, 40] n_range = [2 ** x for x in n_range] edges_range = [1e6, 1e7, 1e8] edges_range = [int(x // 2) for x in edges_range] gpu_res = { int(np.log2(n)): [] for n in n_range } # Random run data_proper = static_graph_generator.generate(num_nodes=168_114, num_edges=6_797_556, is_directed=False, has_self_loop=False) for n, edges in product(n_range, edges_range): max_edges = (n * (n - 1)) // 2 density = edges / max_edges if density > 0.75: res = "FAIL" else: res = "GOOD" f_string = f"{n:<13} | {edges:<13} | {density:>8.3f} | {res}" print(f_string) if res == "FAIL": continue start = time.perf_counter() data_proper = static_graph_generator.generate(num_nodes=n, num_edges=edges, is_directed=False, has_self_loop=False) elapsed = time.perf_counter() - start gen_edges = data_proper.shape[0] edges_per_second = data_proper.shape[0] / elapsed calculated = (gen_edges, edges_per_second, elapsed) f_string = f"{n:<13} | {edges:<13} | {edges_per_second}" print(f_string) l = gpu_res[int(np.log2(n))] l.append(calculated) # In[7]: plt.figure(figsize=(15, 6)) for n, data in gpu_res.items(): x, y = get_xy(data) plt.plot(x, y, '--x', label=f'{n}') plt.xlabel('Total edges to generate') plt.ylabel('Edges per seconds') plt.legend() plt.legend() plt.yscale('log') plt.xscale('log') # <a id="2"></a> # ### CPU # In[8]: static_graph_generator=RMATGenerator() static_graph_generator.gpu=False # In[9]: static_graph_generator._fit_results = 0.4, 0.25, 0.2, 0.15 # In[10]: n_range = [15, 20, 30, 40] n_range = [2 ** x for x in n_range] edges_range = [1e6, 1e7, 1e8] edges_range = [int(x // 2) for x in edges_range] cpu_res = { int(np.log2(n)): [] for n in n_range } # Random run data_proper = static_graph_generator.generate(num_nodes=168_114, num_edges=6_797_556, is_directed=False, has_self_loop=False) for n, edges in product(n_range, edges_range): max_edges = (n * (n - 1)) // 2 density = edges / max_edges if density > 0.75: res = "FAIL" else: res = "GOOD" f_string = f"{n:<13} | {edges:<13} | {density:>8.3f} | {res}" print(f_string) if res == "FAIL": continue start = time.perf_counter() data_proper = static_graph_generator.generate(num_nodes=n, num_edges=edges, is_directed=False, has_self_loop=False) elapsed = time.perf_counter() - start gen_edges = data_proper.shape[0] edges_per_second = data_proper.shape[0] / elapsed calculated = (gen_edges, edges_per_second, elapsed) f_string = f"{n:<13} | {edges:<13} | {edges_per_second}" print(f_string) l = cpu_res[int(np.log2(n))] l.append(calculated) # In[11]: plt.figure(figsize=(15, 6)) for n, data in cpu_res.items(): x, y = get_xy(data) plt.plot(x, y, '--x', label=f'{n}') plt.xlabel('Total edges to generate') plt.ylabel('Edges per seconds') plt.legend() plt.legend() plt.yscale('log') plt.xscale('log') # In[12]: plt.figure(figsize=(15, 6), dpi=800) for n, data in gpu_res.items(): x, y = get_xy(data) plt.plot(x, y, '--x', label=f'2^{n} nodes') for n, data in cpu_res.items(): x, y = get_xy(data) plt.plot(x, y, '--o', label=f'2^{n} nodes') plt.annotate('dense \n graph', (cpu_res[15][2][0], cpu_res[15][2][1])) plt.annotate('gpu generators', (gpu_res[15][1][0], gpu_res[15][1][1]+2e7)) plt.annotate('cpu generators', (cpu_res[15][1][0], cpu_res[15][1][1]+7e4)) plt.title('Exact generator performance') plt.xlabel('Total edges to generate') plt.ylabel('Edges per seconds') plt.legend() plt.yscale('log') plt.xscale('log') # plt.savefig("myImage.png", format="png", dpi=800) plt.show() # ## Approximate generator # <a id="3"></a> # ### GPU # In[13]: import cupy as cp import numpy as np from pylibraft.random import rmat def generate_gpu_rmat_approx(a, b, c, d, r_scale, c_scale, n_edges, noise=0.5): gen_graph = None theta_len = max(r_scale, c_scale) * 4 base_theta = [a, b, c, d] if noise > 0: full_theta = [] for i in range(theta_len): noise_uniform = noise * np.random.uniform(-1, 1, size=len(base_theta)) noise_to_add = np.multiply(base_theta, noise_uniform) theta_n = base_theta + noise_to_add theta_n = theta_n / np.sum(theta_n) full_theta.append(theta_n) else: full_theta = base_theta * theta_len theta_cpu = np.array(full_theta, dtype=np.float32) theta = cp.asarray(theta_cpu) tmp = cp.empty((n_edges, 2), dtype=cp.int32) rmat(tmp, theta, r_scale, c_scale) return tmp # In[14]: a, b, c, d = 0.4, 0.25, 0.2, 0.15 # In[15]: n_range = [15, 20, 30, 40] edges_range = [1e6, 1e7, 1e8] edges_range = [int(x) for x in edges_range] gpu_res_approx = { n: [] for n in n_range } # Random run data_proper = generate_gpu_rmat_approx(a, b, c, d, 18, 18, 6_797_556) for n, edges in product(n_range, edges_range): max_edges = (2**n * (2**n - 1)) // 2 density = edges / max_edges if density > 0.75: res = "FAIL" else: res = "GOOD" f_string = f"{n:<13} | {edges:<13} | {density:>8.3f} | {res}" print(f_string) if res == "FAIL": continue start = time.perf_counter() data_proper = generate_gpu_rmat_approx(a, b, c, d, n, n, edges) elapsed = time.perf_counter() - start gen_edges = data_proper.shape[0] edges_per_second = data_proper.shape[0] / elapsed calculated = (gen_edges, edges_per_second, elapsed) f_string = f"{n:<13} | {edges:<13} | {edges_per_second}" print(f_string) l = gpu_res_approx[n] l.append(calculated) # In[16]: plt.figure(figsize=(15, 6)) for n, data in gpu_res_approx.items(): x, y = get_xy(data) plt.plot(x, y, '--x', label=f'{n}') plt.xlabel('Total edges to generate') plt.ylabel('Edges per seconds') plt.legend() plt.legend() plt.yscale('log') plt.xscale('log') # <a id="4"></a> # ### CPU # In[17]: from syngen.generator.graph.utils import effective_nonsquare_rmat_approximate # In[18]: a, b, c, d = 0.4, 0.25, 0.2, 0.15 theta = np.array([[a, b], [c, d]]) theta /= a + b + c + d # In[19]: part, _, _ = effective_nonsquare_rmat_approximate( theta, 6_797_556, (18, 18), noise_scaling=0.5, batch_size=1_000_000, dtype=np.int64, custom_samplers=None, generate_back_edges=False ) # In[20]: part.shape # In[21]: n_range = [15, 20, 30, 40] #n_range = [2 ** x for x in n_range] edges_range = [1e6, 1e7, 1e8] edges_range = [int(x) for x in edges_range] cpu_res_approx = { n: [] for n in n_range } # Random run part, _, _ = effective_nonsquare_rmat_approximate( theta, 6_797_556, (18, 18), noise_scaling=0.5, batch_size=1_000_000, dtype=np.int64, custom_samplers=None, generate_back_edges=False ) for n, edges in product(n_range, edges_range): max_edges = (2**n * (2**n - 1)) // 2 density = edges / max_edges if density > 0.75: res = "FAIL" else: res = "GOOD" f_string = f"{n:<13} | {edges:<13} | {density:>8.3f} | {res}" print(f_string) if res == "FAIL": continue start = time.perf_counter() data_proper, _, _ = effective_nonsquare_rmat_approximate( theta, edges, (n, n), noise_scaling=0.5, batch_size=1_000_000, dtype=np.int64, custom_samplers=None, generate_back_edges=False ) elapsed = time.perf_counter() - start gen_edges = data_proper.shape[0] edges_per_second = data_proper.shape[0] / elapsed calculated = (gen_edges, edges_per_second, elapsed) f_string = f"{n:<13} | {edges:<13} | {edges_per_second}" print(f_string) l = cpu_res_approx[n] l.append(calculated) # In[22]: plt.figure(figsize=(15, 6)) for n, data in cpu_res_approx.items(): x, y = get_xy(data) plt.plot(x, y, '--x', label=f'{n}') plt.xlabel('Total edges to generate') plt.ylabel('Edges per seconds') plt.legend() plt.legend() plt.yscale('log') plt.xscale('log') # In[24]: plt.figure(figsize=(15, 6), dpi=800) for n, data in gpu_res_approx.items(): x, y = get_xy(data) plt.plot(x, y, '--x', label=f'2^{n} nodes') for n, data in cpu_res_approx.items(): x, y = get_xy(data) plt.plot(x, y, '--o', label=f'2^{n} nodes') plt.annotate('gpu generators', (gpu_res_approx[15][1][0], gpu_res_approx[15][1][1]+5e9)) plt.annotate('cpu generators', (cpu_res_approx[15][1][0], cpu_res_approx[15][1][1]+5e5)) plt.title('Approximate generator performance') plt.xlabel('Total edges to generate') plt.ylabel('Edges per seconds') plt.legend() plt.yscale('log') plt.xscale('log') plt.savefig("/workspace/img/edge_perf.png") plt.show() # In[ ]:
CUDA-Optimized/FastSpeech/fastspeech/trt/plugins/add_pos_enc
add_pos_enc
test_add_pos_enc_plugin
# Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # * Neither the name of the NVIDIA CORPORATION nor the # names of its contributors may be used to endorse or promote products # derived from this software without specific prior written permission. # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL NVIDIA CORPORATION BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. from __future__ import print_function import numpy as np import tensorrt as trt import pycuda.autoinit import pycuda.driver as cuda import ctypes import os import sys import time logger = trt.Logger(trt.Logger.INFO) PLUGIN_PATH = '/home/dahn/git/fastspeech/fastspeech/trt/plugins/add_pos_enc/AddPosEncPlugin.so' ctypes.cdll.LoadLibrary(PLUGIN_PATH) def get_plugin_creator(plugin_name): trt.init_libnvinfer_plugins(logger, '') plugin_creator_list = trt.get_plugin_registry().plugin_creator_list plugin_creator = None for c in plugin_creator_list: if c.name == plugin_name: plugin_creator = c return plugin_creator def build_engine(shape): plugin_creator = get_plugin_creator('AddPosEncPlugin') if plugin_creator == None: print('Plugin not found. Exiting') exit() builder = trt.Builder(logger) builder.max_batch_size = 1024 builder.max_workspace_size = 1 << 20 builder.fp16_mode = use_fp16 network = builder.create_network() tensor = network.add_input('data', trt.DataType.FLOAT, shape) tensor = network.add_plugin_v2( [tensor], plugin_creator.create_plugin('AddPosEncPlugin', trt.PluginFieldCollection()) ).get_output(0) network.mark_output(tensor) return builder.build_cuda_engine(network) def run_trt(data): engine = build_engine(data.shape[1:]) context = engine.create_execution_context() d_data = cuda.mem_alloc(data.nbytes) output = np.zeros_like(data, dtype=np.float32) d_output = cuda.mem_alloc(output.nbytes) cuda.memcpy_htod(d_data, data) bindings = [int(d_data), int(d_output)] start = time.time() context.execute(data.shape[0], bindings) end = time.time() time_elapsed = end - start print("time elapsed: {:06f}".format(time_elapsed)) cuda.memcpy_dtoh(output, d_output) return output use_fp16 = len(sys.argv) > 1 and sys.argv[1].isdigit() and int(sys.argv[1]) == 1 print('Use FP16:', use_fp16) output = run_trt(np.zeros((16, 128, 384), np.float32)) print(output) print(output.shape)
TensorFlow2/Recommendation/WideAndDeep/triton/runner
runner
stages
# Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import pathlib from typing import List, Optional, Tuple, Union # method from PEP-366 to support relative import in executed modules if __name__ == "__main__" and __package__ is None: __package__ = pathlib.Path(__file__).parent.name from .core import Command class ResultsType: """ Results types generated by runner """ TRITON_PERFORMANCE_OFFLINE = "triton_performance_offline" TRITON_PERFORMANCE_ONLINE = "triton_performance_online" class Stage: """ Stage definition """ label: str commands: List[Command] result_path: Optional[str] result_type: Optional[str] def __init__( self, commands: Union[Tuple[str, ...], List[str]], result_path: Optional[str] = None, result_type: Optional[str] = None, ): """ Args: commands: List or Tuple of commands provided as raw string result_path: Path to results file generated by stage result_type: Type of results generated by stage """ if type(commands) not in [tuple, list]: raise ValueError("""Incorrect type of commands list. Please, provide list of commands as tuple.""") self.commands = list(map(lambda command: Command(data=command), commands)) self.result_path = result_path self.result_type = result_type class ExportStage(Stage): label = "Export Model" class ConversionStage(Stage): label = "Convert Model" class DeployStage(Stage): label = "Deploy Model" class CorrectnessStage(Stage): label = "Model Correctness" class TritonPreparePerformanceProfilingDataStage(Stage): label = "Prepare Triton Profiling Data" class TritonPerformanceOfflineStage(Stage): label = "Triton Performance Offline" class TritonPerformanceOnlineStage(Stage): label = "Triton Performance Online"
PyTorch/SpeechSynthesis/FastPitch/common/text
text
numerical
""" adapted from https://github.com/keithito/tacotron """ import inflect import re _magnitudes = ['trillion', 'billion', 'million', 'thousand', 'hundred', 'm', 'b', 't'] _magnitudes_key = {'m': 'million', 'b': 'billion', 't': 'trillion'} _measurements = '(f|c|k|d|m)' _measurements_key = {'f': 'fahrenheit', 'c': 'celsius', 'k': 'thousand', 'm': 'meters'} _currency_key = {'$': 'dollar', '£': 'pound', '€': 'euro', '₩': 'won'} _inflect = inflect.engine() _comma_number_re = re.compile(r'([0-9][0-9\,]+[0-9])') _decimal_number_re = re.compile(r'([0-9]+\.[0-9]+)') _currency_re = re.compile(r'([\$€£₩])([0-9\.\,]*[0-9]+)(?:[ ]?({})(?=[^a-zA-Z]|$))?'.format("|".join(_magnitudes)), re.IGNORECASE) _measurement_re = re.compile(r'([0-9\.\,]*[0-9]+(\s)?{}\b)'.format(_measurements), re.IGNORECASE) _ordinal_re = re.compile(r'[0-9]+(st|nd|rd|th)') # _range_re = re.compile(r'(?<=[0-9])+(-)(?=[0-9])+.*?') _roman_re = re.compile(r'\b(?=[MDCLXVI]+\b)M{0,4}(CM|CD|D?C{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{2,3})\b') # avoid I _multiply_re = re.compile(r'(\b[0-9]+)(x)([0-9]+)') _number_re = re.compile(r"[0-9]+'s|[0-9]+s|[0-9]+") def _remove_commas(m): return m.group(1).replace(',', '') def _expand_decimal_point(m): return m.group(1).replace('.', ' point ') def _expand_currency(m): currency = _currency_key[m.group(1)] quantity = m.group(2) magnitude = m.group(3) # remove commas from quantity to be able to convert to numerical quantity = quantity.replace(',', '') # check for million, billion, etc... if magnitude is not None and magnitude.lower() in _magnitudes: if len(magnitude) == 1: magnitude = _magnitudes_key[magnitude.lower()] return "{} {} {}".format(_expand_hundreds(quantity), magnitude, currency+'s') parts = quantity.split('.') if len(parts) > 2: return quantity + " " + currency + "s" # Unexpected format dollars = int(parts[0]) if parts[0] else 0 cents = int(parts[1]) if len(parts) > 1 and parts[1] else 0 if dollars and cents: dollar_unit = currency if dollars == 1 else currency+'s' cent_unit = 'cent' if cents == 1 else 'cents' return "{} {}, {} {}".format( _expand_hundreds(dollars), dollar_unit, _inflect.number_to_words(cents), cent_unit) elif dollars: dollar_unit = currency if dollars == 1 else currency+'s' return "{} {}".format(_expand_hundreds(dollars), dollar_unit) elif cents: cent_unit = 'cent' if cents == 1 else 'cents' return "{} {}".format(_inflect.number_to_words(cents), cent_unit) else: return 'zero' + ' ' + currency + 's' def _expand_hundreds(text): number = float(text) if 1000 < number < 10000 and (number % 100 == 0) and (number % 1000 != 0): return _inflect.number_to_words(int(number / 100)) + " hundred" else: return _inflect.number_to_words(text) def _expand_ordinal(m): return _inflect.number_to_words(m.group(0)) def _expand_measurement(m): _, number, measurement = re.split('(\d+(?:\.\d+)?)', m.group(0)) number = _inflect.number_to_words(number) measurement = "".join(measurement.split()) measurement = _measurements_key[measurement.lower()] return "{} {}".format(number, measurement) def _expand_range(m): return ' to ' def _expand_multiply(m): left = m.group(1) right = m.group(3) return "{} by {}".format(left, right) def _expand_roman(m): # from https://stackoverflow.com/questions/19308177/converting-roman-numerals-to-integers-in-python roman_numerals = {'I':1, 'V':5, 'X':10, 'L':50, 'C':100, 'D':500, 'M':1000} result = 0 num = m.group(0) for i, c in enumerate(num): if (i+1) == len(num) or roman_numerals[c] >= roman_numerals[num[i+1]]: result += roman_numerals[c] else: result -= roman_numerals[c] return str(result) def _expand_number(m): _, number, suffix = re.split(r"(\d+(?:'?\d+)?)", m.group(0)) number = int(number) if number > 1000 < 10000 and (number % 100 == 0) and (number % 1000 != 0): text = _inflect.number_to_words(number // 100) + " hundred" elif number > 1000 and number < 3000: if number == 2000: text = 'two thousand' elif number > 2000 and number < 2010: text = 'two thousand ' + _inflect.number_to_words(number % 100) elif number % 100 == 0: text = _inflect.number_to_words(number // 100) + ' hundred' else: number = _inflect.number_to_words(number, andword='', zero='oh', group=2).replace(', ', ' ') number = re.sub(r'-', ' ', number) text = number else: number = _inflect.number_to_words(number, andword='and') number = re.sub(r'-', ' ', number) number = re.sub(r',', '', number) text = number if suffix in ("'s", "s"): if text[-1] == 'y': text = text[:-1] + 'ies' else: text = text + suffix return text def normalize_numbers(text): text = re.sub(_comma_number_re, _remove_commas, text) text = re.sub(_currency_re, _expand_currency, text) text = re.sub(_decimal_number_re, _expand_decimal_point, text) text = re.sub(_ordinal_re, _expand_ordinal, text) # text = re.sub(_range_re, _expand_range, text) # text = re.sub(_measurement_re, _expand_measurement, text) text = re.sub(_roman_re, _expand_roman, text) text = re.sub(_multiply_re, _expand_multiply, text) text = re.sub(_number_re, _expand_number, text) return text
PyTorch/Segmentation/nnUNet/triton
triton
preprocess
# Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os import time from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser from data_preprocessing.preprocessor import Preprocessor from utils.utils import get_task_code parser = ArgumentParser(formatter_class=ArgumentDefaultsHelpFormatter) parser.add_argument("--data", type=str, default="/data", help="Path to data directory") parser.add_argument("--results", type=str, default="/data", help="Path for saving results directory") parser.add_argument( "--exec_mode", type=str, default="training", choices=["training", "val", "test"], help="Mode for data preprocessing", ) parser.add_argument("--dilation", action="store_true", help="Perform morphological label dilation") parser.add_argument("--task", type=str, help="Number of task to be run. MSD uses numbers 01-10") parser.add_argument("--dim", type=int, default=3, choices=[2, 3], help="Data dimension to prepare") parser.add_argument("--n_jobs", type=int, default=-1, help="Number of parallel jobs for data preprocessing") if __name__ == "__main__": args = parser.parse_args() start = time.time() Preprocessor(args).run() task_code = get_task_code(args) path = os.path.join(args.data, task_code) if args.exec_mode == "test": path = os.path.join(path, "test") end = time.time() print(f"Preprocessing time: {(end - start):.2f}")
TensorFlow/Detection/SSD/models/research/object_detection/samples/configs
configs
ssd_mobilenet_v2_quantized_300x300_coco
# Quantized trained SSD with Mobilenet v2 on MSCOCO Dataset. # Users should configure the fine_tune_checkpoint field in the train config as # well as the label_map_path and input_path fields in the train_input_reader and # eval_input_reader. Search for "PATH_TO_BE_CONFIGURED" to find the fields that # should be configured. model { ssd { num_classes: 90 box_coder { faster_rcnn_box_coder { y_scale: 10.0 x_scale: 10.0 height_scale: 5.0 width_scale: 5.0 } } matcher { argmax_matcher { matched_threshold: 0.5 unmatched_threshold: 0.5 ignore_thresholds: false negatives_lower_than_unmatched: true force_match_for_each_row: true } } similarity_calculator { iou_similarity { } } anchor_generator { ssd_anchor_generator { num_layers: 6 min_scale: 0.2 max_scale: 0.95 aspect_ratios: 1.0 aspect_ratios: 2.0 aspect_ratios: 0.5 aspect_ratios: 3.0 aspect_ratios: 0.3333 } } image_resizer { fixed_shape_resizer { height: 300 width: 300 } } box_predictor { convolutional_box_predictor { min_depth: 0 max_depth: 0 num_layers_before_predictor: 0 use_dropout: false dropout_keep_probability: 0.8 kernel_size: 1 box_code_size: 4 apply_sigmoid_to_scores: false conv_hyperparams { activation: RELU_6, regularizer { l2_regularizer { weight: 0.00004 } } initializer { truncated_normal_initializer { stddev: 0.03 mean: 0.0 } } batch_norm { train: true, scale: true, center: true, decay: 0.9997, epsilon: 0.001, } } } } feature_extractor { type: 'ssd_mobilenet_v2' min_depth: 16 depth_multiplier: 1.0 conv_hyperparams { activation: RELU_6, regularizer { l2_regularizer { weight: 0.00004 } } initializer { truncated_normal_initializer { stddev: 0.03 mean: 0.0 } } batch_norm { train: true, scale: true, center: true, decay: 0.9997, epsilon: 0.001, } } } loss { classification_loss { weighted_sigmoid { } } localization_loss { weighted_smooth_l1 { } } hard_example_miner { num_hard_examples: 3000 iou_threshold: 0.99 loss_type: CLASSIFICATION max_negatives_per_positive: 3 min_negatives_per_image: 3 } classification_weight: 1.0 localization_weight: 1.0 } normalize_loss_by_num_matches: true post_processing { batch_non_max_suppression { score_threshold: 1e-8 iou_threshold: 0.6 max_detections_per_class: 100 max_total_detections: 100 } score_converter: SIGMOID } } } train_config: { batch_size: 24 optimizer { rms_prop_optimizer: { learning_rate: { exponential_decay_learning_rate { initial_learning_rate: 0.004 decay_steps: 800720 decay_factor: 0.95 } } momentum_optimizer_value: 0.9 decay: 0.9 epsilon: 1.0 } } fine_tune_checkpoint: "PATH_TO_BE_CONFIGURED/model.ckpt" fine_tune_checkpoint_type: "detection" # Note: The below line limits the training process to 200K steps, which we # empirically found to be sufficient enough to train the pets dataset. This # effectively bypasses the learning rate schedule (the learning rate will # never decay). Remove the below line to train indefinitely. num_steps: 200000 data_augmentation_options { random_horizontal_flip { } } data_augmentation_options { ssd_random_crop { } } } train_input_reader: { tf_record_input_reader { input_path: "PATH_TO_BE_CONFIGURED/mscoco_train.record-?????-of-00100" } label_map_path: "PATH_TO_BE_CONFIGURED/mscoco_label_map.pbtxt" } eval_config: { num_examples: 8000 # Note: The below line limits the evaluation process to 10 evaluations. # Remove the below line to evaluate indefinitely. max_evals: 10 } eval_input_reader: { tf_record_input_reader { input_path: "PATH_TO_BE_CONFIGURED/mscoco_val.record-?????-of-00010" } label_map_path: "PATH_TO_BE_CONFIGURED/mscoco_label_map.pbtxt" shuffle: false num_readers: 1 } graph_rewriter { quantization { delay: 48000 weight_bits: 8 activation_bits: 8 } }
TensorFlow2/Segmentation/Contrib/UNet3P/models
models
model
""" Returns Unet3+ model """ import tensorflow as tf from omegaconf import DictConfig from .backbones import vgg16_backbone, vgg19_backbone, unet3plus_backbone from .unet3plus import unet3plus from .unet3plus_deep_supervision import unet3plus_deepsup from .unet3plus_deep_supervision_cgm import unet3plus_deepsup_cgm def prepare_model(cfg: DictConfig, training=False): """ Creates and return model object based on given model type. """ input_shape = [cfg.INPUT.HEIGHT, cfg.INPUT.WIDTH, cfg.INPUT.CHANNELS] input_layer = tf.keras.layers.Input( shape=input_shape, name="input_layer" ) # 320*320*3 filters = [64, 128, 256, 512, 1024] # create backbone if cfg.MODEL.BACKBONE.TYPE == "unet3plus": backbone_layers = unet3plus_backbone( input_layer, filters ) elif cfg.MODEL.BACKBONE.TYPE == "vgg16": backbone_layers = vgg16_backbone(input_layer, ) elif cfg.MODEL.BACKBONE.TYPE == "vgg19": backbone_layers = vgg19_backbone(input_layer, ) else: raise ValueError( "Wrong backbone type passed." "\nPlease check config file for possible options." ) print(f"Using {cfg.MODEL.BACKBONE.TYPE} as a backbone.") if cfg.MODEL.TYPE == "unet3plus": # training parameter does not matter in this case outputs, model_name = unet3plus( backbone_layers, cfg.OUTPUT.CLASSES, filters ) elif cfg.MODEL.TYPE == "unet3plus_deepsup": outputs, model_name = unet3plus_deepsup( backbone_layers, cfg.OUTPUT.CLASSES, filters, training ) elif cfg.MODEL.TYPE == "unet3plus_deepsup_cgm": if cfg.OUTPUT.CLASSES != 1: raise ValueError( "UNet3+ with Deep Supervision and Classification Guided Module" "\nOnly works when model output classes are equal to 1" ) outputs, model_name = unet3plus_deepsup_cgm( backbone_layers, cfg.OUTPUT.CLASSES, filters, training ) else: raise ValueError( "Wrong model type passed." "\nPlease check config file for possible options." ) return tf.keras.Model( inputs=input_layer, outputs=outputs, name=model_name ) if __name__ == "__main__": """## Test model Compilation,""" from omegaconf import OmegaConf cfg = { "WORK_DIR": "H:\\Projects\\UNet3P", "INPUT": {"HEIGHT": 320, "WIDTH": 320, "CHANNELS": 3}, "OUTPUT": {"CLASSES": 1}, # available variants are unet3plus, unet3plus_deepsup, unet3plus_deepsup_cgm "MODEL": {"TYPE": "unet3plus", # available variants are unet3plus, vgg16, vgg19 "BACKBONE": {"TYPE": "vgg19", } } } unet_3P = prepare_model(OmegaConf.create(cfg), True) unet_3P.summary() # tf.keras.utils.plot_model(unet_3P, show_layer_names=True, show_shapes=True) # unet_3P.save("unet_3P.hdf5")
TensorFlow/Detection/SSD/models/research/object_detection/utils
utils
np_box_list_ops_test
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Tests for object_detection.utils.np_box_list_ops.""" import numpy as np import tensorflow as tf from object_detection.utils import np_box_list from object_detection.utils import np_box_list_ops class AreaRelatedTest(tf.test.TestCase): def setUp(self): boxes1 = np.array([[4.0, 3.0, 7.0, 5.0], [5.0, 6.0, 10.0, 7.0]], dtype=float) boxes2 = np.array([[3.0, 4.0, 6.0, 8.0], [14.0, 14.0, 15.0, 15.0], [0.0, 0.0, 20.0, 20.0]], dtype=float) self.boxlist1 = np_box_list.BoxList(boxes1) self.boxlist2 = np_box_list.BoxList(boxes2) def test_area(self): areas = np_box_list_ops.area(self.boxlist1) expected_areas = np.array([6.0, 5.0], dtype=float) self.assertAllClose(expected_areas, areas) def test_intersection(self): intersection = np_box_list_ops.intersection(self.boxlist1, self.boxlist2) expected_intersection = np.array([[2.0, 0.0, 6.0], [1.0, 0.0, 5.0]], dtype=float) self.assertAllClose(intersection, expected_intersection) def test_iou(self): iou = np_box_list_ops.iou(self.boxlist1, self.boxlist2) expected_iou = np.array([[2.0 / 16.0, 0.0, 6.0 / 400.0], [1.0 / 16.0, 0.0, 5.0 / 400.0]], dtype=float) self.assertAllClose(iou, expected_iou) def test_ioa(self): boxlist1 = np_box_list.BoxList( np.array( [[0.25, 0.25, 0.75, 0.75], [0.0, 0.0, 0.5, 0.75]], dtype= np.float32)) boxlist2 = np_box_list.BoxList( np.array( [[0.5, 0.25, 1.0, 1.0], [0.0, 0.0, 1.0, 1.0]], dtype=np.float32)) ioa21 = np_box_list_ops.ioa(boxlist2, boxlist1) expected_ioa21 = np.array([[0.5, 0.0], [1.0, 1.0]], dtype=np.float32) self.assertAllClose(ioa21, expected_ioa21) def test_scale(self): boxlist = np_box_list.BoxList( np.array( [[0.25, 0.25, 0.75, 0.75], [0.0, 0.0, 0.5, 0.75]], dtype= np.float32)) boxlist_scaled = np_box_list_ops.scale(boxlist, 2.0, 3.0) expected_boxlist_scaled = np_box_list.BoxList( np.array( [[0.5, 0.75, 1.5, 2.25], [0.0, 0.0, 1.0, 2.25]], dtype=np.float32)) self.assertAllClose(expected_boxlist_scaled.get(), boxlist_scaled.get()) def test_clip_to_window(self): boxlist = np_box_list.BoxList( np.array( [[0.25, 0.25, 0.75, 0.75], [0.0, 0.0, 0.5, 0.75], [-0.2, -0.3, 0.7, 1.5]], dtype=np.float32)) boxlist_clipped = np_box_list_ops.clip_to_window(boxlist, [0.0, 0.0, 1.0, 1.0]) expected_boxlist_clipped = np_box_list.BoxList( np.array( [[0.25, 0.25, 0.75, 0.75], [0.0, 0.0, 0.5, 0.75], [0.0, 0.0, 0.7, 1.0]], dtype=np.float32)) self.assertAllClose(expected_boxlist_clipped.get(), boxlist_clipped.get()) def test_prune_outside_window(self): boxlist = np_box_list.BoxList( np.array( [[0.25, 0.25, 0.75, 0.75], [0.0, 0.0, 0.5, 0.75], [-0.2, -0.3, 0.7, 1.5]], dtype=np.float32)) boxlist_pruned, _ = np_box_list_ops.prune_outside_window( boxlist, [0.0, 0.0, 1.0, 1.0]) expected_boxlist_pruned = np_box_list.BoxList( np.array( [[0.25, 0.25, 0.75, 0.75], [0.0, 0.0, 0.5, 0.75]], dtype= np.float32)) self.assertAllClose(expected_boxlist_pruned.get(), boxlist_pruned.get()) def test_concatenate(self): boxlist1 = np_box_list.BoxList( np.array( [[0.25, 0.25, 0.75, 0.75], [0.0, 0.0, 0.5, 0.75]], dtype= np.float32)) boxlist2 = np_box_list.BoxList( np.array( [[0.5, 0.25, 1.0, 1.0], [0.0, 0.0, 1.0, 1.0]], dtype=np.float32)) boxlists = [boxlist1, boxlist2] boxlist_concatenated = np_box_list_ops.concatenate(boxlists) boxlist_concatenated_expected = np_box_list.BoxList( np.array( [[0.25, 0.25, 0.75, 0.75], [0.0, 0.0, 0.5, 0.75], [0.5, 0.25, 1.0, 1.0], [0.0, 0.0, 1.0, 1.0]], dtype=np.float32)) self.assertAllClose(boxlist_concatenated_expected.get(), boxlist_concatenated.get()) def test_change_coordinate_frame(self): boxlist = np_box_list.BoxList( np.array( [[0.25, 0.25, 0.75, 0.75], [0.0, 0.0, 0.5, 0.75]], dtype= np.float32)) boxlist_coord = np_box_list_ops.change_coordinate_frame( boxlist, np.array([0, 0, 0.5, 0.5], dtype=np.float32)) expected_boxlist_coord = np_box_list.BoxList( np.array([[0.5, 0.5, 1.5, 1.5], [0, 0, 1.0, 1.5]], dtype=np.float32)) self.assertAllClose(boxlist_coord.get(), expected_boxlist_coord.get()) def test_filter_scores_greater_than(self): boxlist = np_box_list.BoxList( np.array( [[0.25, 0.25, 0.75, 0.75], [0.0, 0.0, 0.5, 0.75]], dtype= np.float32)) boxlist.add_field('scores', np.array([0.8, 0.2], np.float32)) boxlist_greater = np_box_list_ops.filter_scores_greater_than(boxlist, 0.5) expected_boxlist_greater = np_box_list.BoxList( np.array([[0.25, 0.25, 0.75, 0.75]], dtype=np.float32)) self.assertAllClose(boxlist_greater.get(), expected_boxlist_greater.get()) class GatherOpsTest(tf.test.TestCase): def setUp(self): boxes = np.array([[3.0, 4.0, 6.0, 8.0], [14.0, 14.0, 15.0, 15.0], [0.0, 0.0, 20.0, 20.0]], dtype=float) self.boxlist = np_box_list.BoxList(boxes) self.boxlist.add_field('scores', np.array([0.5, 0.7, 0.9], dtype=float)) self.boxlist.add_field('labels', np.array([[0, 0, 0, 1, 0], [0, 1, 0, 0, 0], [0, 0, 0, 0, 1]], dtype=int)) def test_gather_with_out_of_range_indices(self): indices = np.array([3, 1], dtype=int) boxlist = self.boxlist with self.assertRaises(ValueError): np_box_list_ops.gather(boxlist, indices) def test_gather_with_invalid_multidimensional_indices(self): indices = np.array([[0, 1], [1, 2]], dtype=int) boxlist = self.boxlist with self.assertRaises(ValueError): np_box_list_ops.gather(boxlist, indices) def test_gather_without_fields_specified(self): indices = np.array([2, 0, 1], dtype=int) boxlist = self.boxlist subboxlist = np_box_list_ops.gather(boxlist, indices) expected_scores = np.array([0.9, 0.5, 0.7], dtype=float) self.assertAllClose(expected_scores, subboxlist.get_field('scores')) expected_boxes = np.array([[0.0, 0.0, 20.0, 20.0], [3.0, 4.0, 6.0, 8.0], [14.0, 14.0, 15.0, 15.0]], dtype=float) self.assertAllClose(expected_boxes, subboxlist.get()) expected_labels = np.array([[0, 0, 0, 0, 1], [0, 0, 0, 1, 0], [0, 1, 0, 0, 0]], dtype=int) self.assertAllClose(expected_labels, subboxlist.get_field('labels')) def test_gather_with_invalid_field_specified(self): indices = np.array([2, 0, 1], dtype=int) boxlist = self.boxlist with self.assertRaises(ValueError): np_box_list_ops.gather(boxlist, indices, 'labels') with self.assertRaises(ValueError): np_box_list_ops.gather(boxlist, indices, ['objectness']) def test_gather_with_fields_specified(self): indices = np.array([2, 0, 1], dtype=int) boxlist = self.boxlist subboxlist = np_box_list_ops.gather(boxlist, indices, ['labels']) self.assertFalse(subboxlist.has_field('scores')) expected_boxes = np.array([[0.0, 0.0, 20.0, 20.0], [3.0, 4.0, 6.0, 8.0], [14.0, 14.0, 15.0, 15.0]], dtype=float) self.assertAllClose(expected_boxes, subboxlist.get()) expected_labels = np.array([[0, 0, 0, 0, 1], [0, 0, 0, 1, 0], [0, 1, 0, 0, 0]], dtype=int) self.assertAllClose(expected_labels, subboxlist.get_field('labels')) class SortByFieldTest(tf.test.TestCase): def setUp(self): boxes = np.array([[3.0, 4.0, 6.0, 8.0], [14.0, 14.0, 15.0, 15.0], [0.0, 0.0, 20.0, 20.0]], dtype=float) self.boxlist = np_box_list.BoxList(boxes) self.boxlist.add_field('scores', np.array([0.5, 0.9, 0.4], dtype=float)) self.boxlist.add_field('labels', np.array([[0, 0, 0, 1, 0], [0, 1, 0, 0, 0], [0, 0, 0, 0, 1]], dtype=int)) def test_with_invalid_field(self): with self.assertRaises(ValueError): np_box_list_ops.sort_by_field(self.boxlist, 'objectness') with self.assertRaises(ValueError): np_box_list_ops.sort_by_field(self.boxlist, 'labels') def test_with_invalid_sorting_order(self): with self.assertRaises(ValueError): np_box_list_ops.sort_by_field(self.boxlist, 'scores', 'Descending') def test_with_descending_sorting(self): sorted_boxlist = np_box_list_ops.sort_by_field(self.boxlist, 'scores') expected_boxes = np.array([[14.0, 14.0, 15.0, 15.0], [3.0, 4.0, 6.0, 8.0], [0.0, 0.0, 20.0, 20.0]], dtype=float) self.assertAllClose(expected_boxes, sorted_boxlist.get()) expected_scores = np.array([0.9, 0.5, 0.4], dtype=float) self.assertAllClose(expected_scores, sorted_boxlist.get_field('scores')) def test_with_ascending_sorting(self): sorted_boxlist = np_box_list_ops.sort_by_field( self.boxlist, 'scores', np_box_list_ops.SortOrder.ASCEND) expected_boxes = np.array([[0.0, 0.0, 20.0, 20.0], [3.0, 4.0, 6.0, 8.0], [14.0, 14.0, 15.0, 15.0],], dtype=float) self.assertAllClose(expected_boxes, sorted_boxlist.get()) expected_scores = np.array([0.4, 0.5, 0.9], dtype=float) self.assertAllClose(expected_scores, sorted_boxlist.get_field('scores')) class NonMaximumSuppressionTest(tf.test.TestCase): def setUp(self): self._boxes = np.array([[0, 0, 1, 1], [0, 0.1, 1, 1.1], [0, -0.1, 1, 0.9], [0, 10, 1, 11], [0, 10.1, 1, 11.1], [0, 100, 1, 101]], dtype=float) self._boxlist = np_box_list.BoxList(self._boxes) def test_with_no_scores_field(self): boxlist = np_box_list.BoxList(self._boxes) max_output_size = 3 iou_threshold = 0.5 with self.assertRaises(ValueError): np_box_list_ops.non_max_suppression( boxlist, max_output_size, iou_threshold) def test_nms_disabled_max_output_size_equals_three(self): boxlist = np_box_list.BoxList(self._boxes) boxlist.add_field('scores', np.array([.9, .75, .6, .95, .2, .3], dtype=float)) max_output_size = 3 iou_threshold = 1. # No NMS expected_boxes = np.array([[0, 10, 1, 11], [0, 0, 1, 1], [0, 0.1, 1, 1.1]], dtype=float) nms_boxlist = np_box_list_ops.non_max_suppression( boxlist, max_output_size, iou_threshold) self.assertAllClose(nms_boxlist.get(), expected_boxes) def test_select_from_three_clusters(self): boxlist = np_box_list.BoxList(self._boxes) boxlist.add_field('scores', np.array([.9, .75, .6, .95, .2, .3], dtype=float)) max_output_size = 3 iou_threshold = 0.5 expected_boxes = np.array([[0, 10, 1, 11], [0, 0, 1, 1], [0, 100, 1, 101]], dtype=float) nms_boxlist = np_box_list_ops.non_max_suppression( boxlist, max_output_size, iou_threshold) self.assertAllClose(nms_boxlist.get(), expected_boxes) def test_select_at_most_two_from_three_clusters(self): boxlist = np_box_list.BoxList(self._boxes) boxlist.add_field('scores', np.array([.9, .75, .6, .95, .5, .3], dtype=float)) max_output_size = 2 iou_threshold = 0.5 expected_boxes = np.array([[0, 10, 1, 11], [0, 0, 1, 1]], dtype=float) nms_boxlist = np_box_list_ops.non_max_suppression( boxlist, max_output_size, iou_threshold) self.assertAllClose(nms_boxlist.get(), expected_boxes) def test_select_at_most_thirty_from_three_clusters(self): boxlist = np_box_list.BoxList(self._boxes) boxlist.add_field('scores', np.array([.9, .75, .6, .95, .5, .3], dtype=float)) max_output_size = 30 iou_threshold = 0.5 expected_boxes = np.array([[0, 10, 1, 11], [0, 0, 1, 1], [0, 100, 1, 101]], dtype=float) nms_boxlist = np_box_list_ops.non_max_suppression( boxlist, max_output_size, iou_threshold) self.assertAllClose(nms_boxlist.get(), expected_boxes) def test_select_from_ten_indentical_boxes(self): boxes = np.array(10 * [[0, 0, 1, 1]], dtype=float) boxlist = np_box_list.BoxList(boxes) boxlist.add_field('scores', np.array(10 * [0.8])) iou_threshold = .5 max_output_size = 3 expected_boxes = np.array([[0, 0, 1, 1]], dtype=float) nms_boxlist = np_box_list_ops.non_max_suppression( boxlist, max_output_size, iou_threshold) self.assertAllClose(nms_boxlist.get(), expected_boxes) def test_different_iou_threshold(self): boxes = np.array([[0, 0, 20, 100], [0, 0, 20, 80], [200, 200, 210, 300], [200, 200, 210, 250]], dtype=float) boxlist = np_box_list.BoxList(boxes) boxlist.add_field('scores', np.array([0.9, 0.8, 0.7, 0.6])) max_output_size = 4 iou_threshold = .4 expected_boxes = np.array([[0, 0, 20, 100], [200, 200, 210, 300],], dtype=float) nms_boxlist = np_box_list_ops.non_max_suppression( boxlist, max_output_size, iou_threshold) self.assertAllClose(nms_boxlist.get(), expected_boxes) iou_threshold = .5 expected_boxes = np.array([[0, 0, 20, 100], [200, 200, 210, 300], [200, 200, 210, 250]], dtype=float) nms_boxlist = np_box_list_ops.non_max_suppression( boxlist, max_output_size, iou_threshold) self.assertAllClose(nms_boxlist.get(), expected_boxes) iou_threshold = .8 expected_boxes = np.array([[0, 0, 20, 100], [0, 0, 20, 80], [200, 200, 210, 300], [200, 200, 210, 250]], dtype=float) nms_boxlist = np_box_list_ops.non_max_suppression( boxlist, max_output_size, iou_threshold) self.assertAllClose(nms_boxlist.get(), expected_boxes) def test_multiclass_nms(self): boxlist = np_box_list.BoxList( np.array( [[0.2, 0.4, 0.8, 0.8], [0.4, 0.2, 0.8, 0.8], [0.6, 0.0, 1.0, 1.0]], dtype=np.float32)) scores = np.array([[-0.2, 0.1, 0.5, -0.4, 0.3], [0.7, -0.7, 0.6, 0.2, -0.9], [0.4, 0.34, -0.9, 0.2, 0.31]], dtype=np.float32) boxlist.add_field('scores', scores) boxlist_clean = np_box_list_ops.multi_class_non_max_suppression( boxlist, score_thresh=0.25, iou_thresh=0.1, max_output_size=3) scores_clean = boxlist_clean.get_field('scores') classes_clean = boxlist_clean.get_field('classes') boxes = boxlist_clean.get() expected_scores = np.array([0.7, 0.6, 0.34, 0.31]) expected_classes = np.array([0, 2, 1, 4]) expected_boxes = np.array([[0.4, 0.2, 0.8, 0.8], [0.4, 0.2, 0.8, 0.8], [0.6, 0.0, 1.0, 1.0], [0.6, 0.0, 1.0, 1.0]], dtype=np.float32) self.assertAllClose(scores_clean, expected_scores) self.assertAllClose(classes_clean, expected_classes) self.assertAllClose(boxes, expected_boxes) if __name__ == '__main__': tf.test.main()
TensorFlow2/Recommendation/DLRM_and_DCNv2/preproc
preproc
prepare_dataset
#! /bin/bash # Copyright (c) 2021 NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # Examples: # to run on a DGX2 with a frequency limit of 3 (will need 8xV100-32GB to fit the model in GPU memory) # ./prepare_dataset.sh DGX2 3 # # to run on a DGX2 with a frequency limit of 15 (should fit on a single V100-32GB): # ./prepare_dataset.sh DGX2 15 # # to run on CPU with a frequency limit of 15: # ./prepare_dataset.sh CPU 15 set -e set -x ls -ltrash download_dir=${download_dir:-'/data/criteo_orig'} ./verify_criteo_downloaded.sh ${download_dir} spark_output_path=${spark_output_path:-'/data/spark/output'} if [ -f ${spark_output_path}/train/_SUCCESS ] \ && [ -f ${spark_output_path}/validation/_SUCCESS ] \ && [ -f ${spark_output_path}/test/_SUCCESS ]; then echo "Spark preprocessing already carried out" else echo "Performing spark preprocessing" ./run_spark.sh $1 ${download_dir} ${spark_output_path} $2 fi conversion_intermediate_dir=${conversion_intermediate_dir:-'/data/intermediate_binary'} final_output_dir=${final_output_dir:-'/data/preprocessed'} if [ -d ${final_output_dir}/train ] \ && [ -d ${final_output_dir}/validation ] \ && [ -d ${final_output_dir}/test ] \ && [ -f ${final_output_dir}/feature_spec.yaml ]; then echo "Final conversion already done" else echo "Performing final conversion to a custom data format" python parquet_to_binary.py --parallel_jobs 40 --src_dir ${spark_output_path} \ --intermediate_dir ${conversion_intermediate_dir} \ --dst_dir ${final_output_dir} cp "${spark_output_path}/model_size.json" "${final_output_dir}/model_size.json" python split_dataset.py --dataset "${final_output_dir}" --output "${final_output_dir}/split" rm ${final_output_dir}/train_data.bin rm ${final_output_dir}/validation_data.bin rm ${final_output_dir}/test_data.bin rm ${final_output_dir}/model_size.json mv ${final_output_dir}/split/* ${final_output_dir} rm -rf ${final_output_dir}/split fi echo "Done preprocessing the Criteo Kaggle Dataset"
TensorFlow/LanguageModeling/BERT/scripts
scripts
run_pretraining_adam
#! /bin/bash # Copyright (c) 2019 NVIDIA CORPORATION. All rights reserved. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. echo "Container nvidia build = " $NVIDIA_BUILD_ID train_batch_size=${1:-16} eval_batch_size=${2:-8} learning_rate=${3:-"1e-4"} precision=${4:-"fp16"} use_xla=${5:-"true"} num_gpus=${6:-8} warmup_steps=${7:-"10000"} train_steps=${8:-1144000} save_checkpoints_steps=${9:-5000} bert_model=${10:-"large"} num_accumulation_steps=${11:-1} seq_len=${12:-512} max_pred_per_seq=${13:-80} DATA_DIR=data/tfrecord/lower_case_1_seq_len_${seq_len}_max_pred_${max_pred_per_seq}_masked_lm_prob_0.15_random_seed_12345_dupe_factor_5_shard_1472_test_split_10/books_wiki_en_corpus if [ "$bert_model" = "large" ] ; then export BERT_DIR=data/download/nvidia_pretrained/bert_tf_pretraining_large_lamb else export BERT_DIR=data/download/nvidia_pretrained/bert_tf_squad11_base_128 fi PREC="" if [ "$precision" = "fp16" ] ; then PREC="--amp" elif [ "$precision" = "fp32" ] ; then PREC="--noamp" elif [ "$precision" = "tf32" ] ; then PREC="--noamp" elif [ "$precision" = "manual_fp16" ] ; then PREC="--noamp --manual_fp16" else echo "Unknown <precision> argument" exit -2 fi if [ "$use_xla" = "true" ] ; then PREC="$PREC --use_xla" echo "XLA activated" else PREC="$PREC --nouse_xla" fi export GBS=$(expr $train_batch_size \* $num_gpus \* $num_accumulation_steps) printf -v TAG "tf_bert_pretraining_adam_%s_%s_gbs%d" "$bert_model" "$precision" $GBS DATESTAMP=`date +'%y%m%d%H%M%S'` #Edit to save logs & checkpoints in a different directory RESULTS_DIR=${RESULTS_DIR:-/results/${TAG}_${DATESTAMP}} LOGFILE=$RESULTS_DIR/$TAG.$DATESTAMP.log mkdir -m 777 -p $RESULTS_DIR printf "Saving checkpoints to %s\n" "$RESULTS_DIR" printf "Logs written to %s\n" "$LOGFILE" INPUT_FILES="$DATA_DIR/training" EVAL_FILES="$DATA_DIR/test" horovod_str="" mpi="" if [ $num_gpus -gt 1 ] ; then mpi="mpiexec --allow-run-as-root -np $num_gpus --bind-to socket" horovod_str="--horovod" fi CMD="$mpi python3 /workspace/bert/run_pretraining.py" CMD+=" --input_files_dir=$INPUT_FILES" CMD+=" --eval_files_dir=$EVAL_FILES" CMD+=" --output_dir=$RESULTS_DIR" CMD+=" --bert_config_file=$BERT_CONFIG" CMD+=" --do_train=True" CMD+=" --do_eval=True" CMD+=" --train_batch_size=$train_batch_size" CMD+=" --eval_batch_size=$eval_batch_size" CMD+=" --max_seq_length=$seq_len" CMD+=" --max_predictions_per_seq=$max_pred_per_seq" CMD+=" --num_train_steps=$train_steps" CMD+=" --num_warmup_steps=$warmup_steps" CMD+=" --num_accumulation_steps=$num_accumulation_steps" CMD+=" --save_checkpoints_steps=$save_checkpoints_steps" CMD+=" --learning_rate=$learning_rate" CMD+=" --optimizer_type=adam" CMD+=" $horovod_str $PREC" CMD+=" --allreduce_post_accumulation=True" #Check if all necessary files are available before training for DIR_or_file in $DATA_DIR $BERT_CONFIG $RESULTS_DIR; do if [ ! -d "$DIR_or_file" ] && [ ! -f "$DIR_or_file" ]; then echo "Error! $DIR_or_file directory missing. Please mount correctly" exit -1 fi done set -x if [ -z "$LOGFILE" ] ; then $CMD else ( $CMD ) |& tee $LOGFILE fi set +x
TensorFlow2/Segmentation/Contrib/UNet3P/data_preparation
data_preparation
verify_data
""" Verify for each image corresponding mask exist or not. Check against both train and val data """ import os import sys from omegaconf import DictConfig from tqdm import tqdm sys.path.append(os.path.abspath("./")) from utils.general_utils import join_paths from utils.images_utils import image_to_mask_name def check_image_and_mask(cfg, mode): """ Check and print names of those images whose mask are not found. """ images_path = join_paths( cfg.WORK_DIR, cfg.DATASET[mode].IMAGES_PATH ) mask_path = join_paths( cfg.WORK_DIR, cfg.DATASET[mode].MASK_PATH ) all_images = os.listdir(images_path) both_found = True for image in tqdm(all_images): mask_name = image_to_mask_name(image) if not ( os.path.exists( join_paths(images_path, image) ) and os.path.exists( join_paths(mask_path, mask_name) ) ): print(f"{mask_name} did not found against {image}") both_found = False return both_found def verify_data(cfg: DictConfig): """ For both train and val data, check for each image its corresponding mask exist or not. If not then stop the program. """ assert check_image_and_mask(cfg, "TRAIN"), \ "Train images and mask should be same in length" assert check_image_and_mask(cfg, "VAL"), \ "Validation images and mask should be same in length"
TensorFlow2/Classification/ConvNets/config/efficientnet_v1
efficientnet_v1
b4_cfg
import tensorflow as tf from config.defaults import Config # NOTE: this confile file can further be overridden by user-defined params provided at the command line config = dict( path_to_impl='model.efficientnet_model_v1', #data-related model params num_classes=1000, # must be the same as data.num_classes input_channels= 3, rescale_input= 1, # binary, mean_rgb=(0.485 * 255, 0.456 * 255, 0.406 * 255), # used when rescale_input=True std_rgb=(0.229 * 255, 0.224 * 255, 0.225 * 255), # used when rescale_input=True dtype= tf.float32, #used for input image normalization/casting, # tf.float32, tf.bfloat16, tf.float16, tf.float32, tf.bfloat16, # GUIDE # width depth resolution dropout # efficientnet_v1-b0 1.0 1.0 224 0.2 # 'efficientnet_v1-b1 1.0 1.1 240 0.2 # 'efficientnet_v1-b2 1.1 1.2 260 0.3 # 'efficientnet_v1-b3 1.2 1.4 300 0.3 # 'efficientnet_v1-b4 1.4 1.8 380 0.4 # 'efficientnet_v1-b5 1.6 2.2 456 0.4 # 'efficientnet_v1-b6 1.8 2.6 528 0.5 # 'efficientnet_v1-b7 2.0 3.1 600 0.5 # 'efficientnet_v1-b8 2.2 3.6 672 0.5 # 'efficientnet_v1-l2 4.3 5.3 800 0.5 width_coefficient= 1.4, depth_coefficient= 1.8, dropout_rate= 0.4, # image resolution must be set in tr/eval/predict configs below drop_connect_rate= 0.2, stem_base_filters= 32, top_base_filters= 1280, activation= 'swish', depth_divisor= 8, min_depth= None, use_se= 1, # binary batch_norm= 'syncbn', bn_momentum= 0.99, bn_epsilon= 1e-3, weight_init= 'fan_out', blocks= ( # (input_filters, output_filters, kernel_size, num_repeat,expand_ratio, strides, se_ratio) # pylint: disable=bad-whitespace dict(input_filters=32, output_filters=16, kernel_size=3, num_repeat=1, expand_ratio=1, strides=(1, 1), se_ratio=0.25,id_skip=True,fused_conv=False,conv_type='depthwise'), dict(input_filters=16, output_filters=24, kernel_size=3, num_repeat=2, expand_ratio=6, strides=(2, 2), se_ratio=0.25,id_skip=True,fused_conv=False,conv_type='depthwise'), dict(input_filters=24, output_filters=40, kernel_size=5, num_repeat=2, expand_ratio=6, strides=(2, 2), se_ratio=0.25,id_skip=True,fused_conv=False,conv_type='depthwise'), dict(input_filters=40, output_filters=80, kernel_size=3, num_repeat=3, expand_ratio=6, strides=(2, 2), se_ratio=0.25,id_skip=True,fused_conv=False,conv_type='depthwise'), dict(input_filters=80, output_filters=112, kernel_size=5, num_repeat=3, expand_ratio=6, strides=(1, 1), se_ratio=0.25,id_skip=True,fused_conv=False,conv_type='depthwise'), dict(input_filters=112, output_filters=192, kernel_size=5, num_repeat=4, expand_ratio=6, strides=(2, 2), se_ratio=0.25,id_skip=True,fused_conv=False,conv_type='depthwise'), dict(input_filters=192, output_filters=320, kernel_size=3, num_repeat=1, expand_ratio=6, strides=(1, 1), se_ratio=0.25,id_skip=True,fused_conv=False,conv_type='depthwise'), # pylint: enable=bad-whitespace ), ) # train_config = dict(lr_decay='cosine', # # max_epochs=500, # img_size=380, # batch_size=256, # save_checkpoint_freq=5, # lr_init=0.005, # weight_decay=5e-6, # epsilon=0.001, # resume_checkpoint=1, # enable_tensorboard=0 # ) # # eval_config = dict(img_size=380, # batch_size=256) # # data_config = dict( # data_dir='/data/', # augmenter_name='autoaugment', # mixup_alpha=0.0, # # # ) # runtime_config = dict(mode='train_and_eval', # model_dir='./output/', # use_amp=1, # use_xla=1, # log_steps=100 # ) # # config = dict(model=model_config, # train=train_config, # eval=eval_config, # data=data_config, # runtime=runtime_config, # )
PyTorch/Classification/GPUNet/triton/085ms/runner
runner
config_NVIDIA-DGX-1-(1x-V100-32GB)
batching: dynamic checkpoints: - name: 0.85ms url: https://api.ngc.nvidia.com/v2/models/nvidia/dle/gpunet_1_pyt_ckpt/versions/21.12.0_amp/zip configurations: - checkpoint: 0.85ms parameters: backend_accelerator: trt checkpoint: 0.85ms device_kind: gpu export_format: onnx export_precision: fp16 format: onnx max_batch_size: 64 number_of_model_instances: 2 precision: fp16 tensorrt_capture_cuda_graph: 0 torch_jit: none container_version: '21.12' datasets: - name: imagenet datasets_dir: datasets ensemble_model_name: null framework: PyTorch measurement_steps_offline: 8 measurement_steps_online: 32 model_name: GPUnet performance_tool: model_analyzer triton_container_image: nvcr.io/nvidia/tritonserver:21.12-py3 triton_custom_operations: null triton_dockerfile: null triton_load_model_method: explicit
TensorFlow/Detection/SSD/models/research/object_detection/builders
builders
input_reader_builder
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Input reader builder. Creates data sources for DetectionModels from an InputReader config. See input_reader.proto for options. Note: If users wishes to also use their own InputReaders with the Object Detection configuration framework, they should define their own builder function that wraps the build function. """ import tensorflow as tf from object_detection.data_decoders import tf_example_decoder from object_detection.protos import input_reader_pb2 parallel_reader = tf.contrib.slim.parallel_reader def build(input_reader_config): """Builds a tensor dictionary based on the InputReader config. Args: input_reader_config: A input_reader_pb2.InputReader object. Returns: A tensor dict based on the input_reader_config. Raises: ValueError: On invalid input reader proto. ValueError: If no input paths are specified. """ if not isinstance(input_reader_config, input_reader_pb2.InputReader): raise ValueError('input_reader_config not of type ' 'input_reader_pb2.InputReader.') if input_reader_config.WhichOneof('input_reader') == 'tf_record_input_reader': config = input_reader_config.tf_record_input_reader if not config.input_path: raise ValueError('At least one input path must be specified in ' '`input_reader_config`.') _, string_tensor = parallel_reader.parallel_read( config.input_path[:], # Convert `RepeatedScalarContainer` to list. reader_class=tf.TFRecordReader, num_epochs=(input_reader_config.num_epochs if input_reader_config.num_epochs else None), num_readers=input_reader_config.num_readers, shuffle=input_reader_config.shuffle, dtypes=[tf.string, tf.string], capacity=input_reader_config.queue_capacity, min_after_dequeue=input_reader_config.min_after_dequeue) label_map_proto_file = None if input_reader_config.HasField('label_map_path'): label_map_proto_file = input_reader_config.label_map_path decoder = tf_example_decoder.TfExampleDecoder( load_instance_masks=input_reader_config.load_instance_masks, instance_mask_type=input_reader_config.mask_type, label_map_proto_file=label_map_proto_file) return decoder.decode(string_tensor) raise ValueError('Unsupported input_reader_config.')
PyTorch/SpeechRecognition/QuartzNet/common/dali
dali
iterator
# Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import numpy as np import torch from nvidia.dali.plugin.base_iterator import LastBatchPolicy from nvidia.dali.plugin.pytorch import DALIGenericIterator from common.helpers import print_once from common.text import _clean_text, punctuation_map def normalize_string(s, symbols, punct_map): """ Normalizes string. Example: 'call me at 8:00 pm!' -> 'call me at eight zero pm' """ labels = set(symbols) try: text = _clean_text(s, ["english_cleaners"], punct_map).strip() return ''.join([tok for tok in text if all(t in labels for t in tok)]) except Exception as e: print_once(f"WARNING: Normalizing failed: {s} {e}") class DaliIterator(object): """Returns batches of data. Batches are in the form: (preprocessed_signal, preprocessed_signal_length, transcript, transcript_length) This iterator is not meant to be the entry point to a Dali pipeline. Use DataLoader instead. """ def __init__(self, dali_pipelines, transcripts, symbols, batch_size, reader_name, train_iterator: bool): self.transcripts = transcripts self.symbols = symbols self.batch_size = batch_size # in train pipeline shard_size is set to divisable by batch_size, # so PARTIAL policy is safe self.dali_it = DALIGenericIterator( dali_pipelines, ["audio", "label", "audio_shape"], reader_name=reader_name, dynamic_shape=True, auto_reset=True, last_batch_policy=LastBatchPolicy.DROP) @staticmethod def _str2list(s: str): """ Returns list of floats, that represents given string. '0.' denotes separator '1.' denotes 'a' '27.' denotes "'" Assumes, that the string is lower case. """ list = [] for c in s: if c == "'": list.append(27.) else: list.append(max(0., ord(c) - 96.)) return list @staticmethod def _pad_lists(lists: list, pad_val=0): """ Pads lists, so that all have the same size. Returns list with actual sizes of corresponding input lists """ max_length = 0 sizes = [] for li in lists: sizes.append(len(li)) max_length = max_length if len(li) < max_length else len(li) for li in lists: li += [pad_val] * (max_length - len(li)) return sizes def _gen_transcripts(self, labels, normalize_transcripts: bool = True): """ Generate transcripts in format expected by NN """ if normalize_transcripts: lists = [ self._str2list(normalize_string(self.transcripts[lab.item()], self.symbols, punctuation_map(self.symbols))) for lab in labels] else: lists = [self._str2list(self.transcripts[lab.item()]) for lab in labels] sizes = self._pad_lists(lists) return (torch.tensor(lists).cuda(), torch.tensor(sizes, dtype=torch.int32).cuda()) def __next__(self): data = self.dali_it.__next__() transcripts, transcripts_lengths = self._gen_transcripts( data[0]["label"]) return (data[0]["audio"], data[0]["audio_shape"][:, 1], transcripts, transcripts_lengths) def next(self): return self.__next__() def __iter__(self): return self # TODO: refactor class SyntheticDataIterator(object): def __init__(self, batch_size, nfeatures, feat_min=-5., feat_max=0., txt_min=0., txt_max=23., feat_lens_max=1760, txt_lens_max=231, regenerate=False): """ Args: batch_size nfeatures: number of features for melfbanks feat_min: minimum value in `feat` tensor, used for randomization feat_max: maximum value in `feat` tensor, used for randomization txt_min: minimum value in `txt` tensor, used for randomization txt_max: maximum value in `txt` tensor, used for randomization regenerate: If True, regenerate random tensors for every iterator step. If False, generate them only at start. """ self.batch_size = batch_size self.nfeatures = nfeatures self.feat_min = feat_min self.feat_max = feat_max self.feat_lens_max = feat_lens_max self.txt_min = txt_min self.txt_max = txt_max self.txt_lens_max = txt_lens_max self.regenerate = regenerate if not self.regenerate: (self.feat, self.feat_lens, self.txt, self.txt_lens ) = self._generate_sample() def _generate_sample(self): feat = ((self.feat_max - self.feat_min) * np.random.random_sample( (self.batch_size, self.nfeatures, self.feat_lens_max)) + self.feat_min) feat_lens = np.random.randint(0, int(self.feat_lens_max) - 1, size=self.batch_size) txt = (self.txt_max - self.txt_min) * np.random.random_sample( (self.batch_size, self.txt_lens_max)) + self.txt_min txt_lens = np.random.randint(0, int(self.txt_lens_max) - 1, size=self.batch_size) return (torch.Tensor(feat).cuda(), torch.Tensor(feat_lens).cuda(), torch.Tensor(txt).cuda(), torch.Tensor(txt_lens).cuda()) def __next__(self): if self.regenerate: return self._generate_sample() return self.feat, self.feat_lens, self.txt, self.txt_lens def next(self): return self.__next__() def __iter__(self): return self
PyTorch/SpeechSynthesis/Tacotron2/trtis_cpp/src/trt/util
util
dims1
/* * Copyright (c) 2019-2020, NVIDIA CORPORATION. All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * Neither the name of the NVIDIA CORPORATION nor the * names of its contributors may be used to endorse or promote products * derived from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL NVIDIA CORPORATION BE LIABLE FOR ANY * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #ifndef TT2I_DIMS1 #define TT2I_DIMS1 #include "NvInfer.h" namespace tts { class Dims1 : public nvinfer1::Dims { public: /** * @brief Create a new 1-dimension Dims struct. * * @param n The size of the single dimension. */ Dims1(const int n) : Dims() { nbDims = 1; d[0] = n; } }; } // namespace tts #endif
TensorFlow/LanguageModeling/BERT
BERT
fused_layer_norm
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import collections import copy import json import math import re import six import tensorflow as tf from tensorflow.python.framework import ops from tensorflow.contrib.layers.python.layers import utils from tensorflow.contrib.framework.python.ops import variables from tensorflow.python.ops import init_ops import numpy from tensorflow.python.ops import array_ops from tensorflow.python.framework import dtypes from tensorflow.python.ops import nn def fused_layer_norm(inputs, center=True, scale=True, activation_fn=None, reuse=None, variables_collections=None, outputs_collections=None, trainable=True, begin_norm_axis=1, begin_params_axis=-1, scope=None, use_fused_batch_norm=False): with tf.variable_scope( scope, 'LayerNorm', [inputs], reuse=reuse) as sc: inputs = ops.convert_to_tensor(inputs) inputs_shape = inputs.shape inputs_rank = inputs_shape.ndims if inputs_rank is None: raise ValueError('Inputs %s has undefined rank.' % inputs.name) dtype = inputs.dtype.base_dtype if begin_norm_axis < 0: begin_norm_axis = inputs_rank + begin_norm_axis if begin_params_axis >= inputs_rank or begin_norm_axis >= inputs_rank: raise ValueError('begin_params_axis (%d) and begin_norm_axis (%d) ' 'must be < rank(inputs) (%d)' % (begin_params_axis, begin_norm_axis, inputs_rank)) params_shape = inputs_shape[begin_params_axis:] if not params_shape.is_fully_defined(): raise ValueError( 'Inputs %s: shape(inputs)[%s:] is not fully defined: %s' % (inputs.name, begin_params_axis, inputs_shape)) # Allocate parameters for the beta and gamma of the normalization. beta, gamma = None, None if center: beta_collections = utils.get_variable_collections(variables_collections, 'beta') beta = variables.model_variable( 'beta', shape=params_shape, dtype=dtype, initializer=init_ops.zeros_initializer(), collections=beta_collections, trainable=trainable) if scale: gamma_collections = utils.get_variable_collections( variables_collections, 'gamma') gamma = variables.model_variable( 'gamma', shape=params_shape, dtype=dtype, initializer=init_ops.ones_initializer(), collections=gamma_collections, trainable=trainable) if use_fused_batch_norm: # get static TensorShape if fully defined, # otherwise retrieve shape tensor norm_shape = inputs.shape[begin_norm_axis:] if norm_shape.is_fully_defined(): bn_shape = [1, -1, 1, numpy.prod(norm_shape.as_list())] else: norm_shape = tf.shape(inputs)[begin_norm_axis:] bn_shape = [1, -1, 1, tf.reduce_prod(norm_shape)] if inputs.get_shape().is_fully_defined(): outputs_shape = inputs.get_shape() else: outputs_shape = tf.shape(inputs) inputs = array_ops.reshape(inputs, bn_shape) if inputs.get_shape().is_fully_defined(): # static inputs TensorShape fully defined after reshape. ones = array_ops.ones(inputs.get_shape()[1], dtype=dtypes.float32) zeros = array_ops.zeros(inputs.get_shape()[1], dtype=dtypes.float32) else: # static inputs TensorShape NOT fully defined after reshape. # must use dynamic shape, which means these input tensors # have to be created at runtime, which causes a slowdown. scale_shape = tf.shape(inputs)[1] ones = array_ops.ones(scale_shape, dtype=dtypes.float32) zeros = array_ops.zeros(scale_shape, dtype=dtypes.float32) outputs, mean, variance = nn.fused_batch_norm( inputs, ones, zeros, epsilon=1e-4, data_format="NCHW") outputs = array_ops.reshape(outputs, outputs_shape) if center and scale: outputs = outputs * gamma + beta elif center: outputs = outputs + beta elif scale: outputs = outputs * gamma else: # Calculate the moments on the last axis (layer activations). norm_axes = list(range(begin_norm_axis, inputs_rank)) mean, variance = nn.moments(inputs, norm_axes, keep_dims=True) # Compute layer normalization using the batch_normalization function. variance_epsilon = 1e-4 outputs = nn.batch_normalization( inputs, mean, variance, offset=beta, scale=gamma, variance_epsilon=variance_epsilon) outputs.set_shape(inputs_shape) if activation_fn is not None: outputs = activation_fn(outputs) return utils.collect_named_outputs(outputs_collections, sc.name, outputs)
TensorFlow/Segmentation/UNet_Industrial
UNet_Industrial
export_saved_model
# !/usr/bin/env python # -*- coding: utf-8 -*- # ============================================================================== # # Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # ============================================================================== """ Usage: python export_saved_model.py \ --activation_fn='relu' \ --batch_size=16 \ --data_format='NCHW' \ --input_dtype="fp32" \ --export_dir="exported_models" \ --model_checkpoint_path="path/to/checkpoint/model.ckpt-2500" \ --unet_variant='tinyUNet' \ --xla \ --amp """ from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import argparse import pprint os.environ["TF_CPP_MIN_LOG_LEVEL"] = "3" import tensorflow as tf from dllogger.logger import LOGGER from model.unet import UNet_v1 from model.blocks.activation_blck import authorized_activation_fn from utils.cmdline_helper import _add_bool_argument def get_export_flags(): parser = argparse.ArgumentParser(description="JoC-UNet_v1-TF-ExportFlags") parser.add_argument('--export_dir', default=None, required=True, type=str, help='The export directory.') parser.add_argument('--model_checkpoint_path', default=None, required=True, help='Checkpoint path.') parser.add_argument( '--data_format', choices=['NHWC', 'NCHW'], type=str, default="NCHW", required=False, help="""Which Tensor format is used for computation inside the mode""" ) parser.add_argument( '--input_dtype', choices=['fp32', 'fp16'], type=str, default="fp32", required=False, help="""Tensorflow dtype of the input tensor""" ) parser.add_argument( '--unet_variant', default="tinyUNet", choices=UNet_v1.authorized_models_variants, type=str, required=False, help="""Which model size is used. This parameter control directly the size and the number of parameters""" ) parser.add_argument( '--activation_fn', choices=authorized_activation_fn, type=str, default="relu", required=False, help="""Which activation function is used after the convolution layers""" ) _add_bool_argument( parser=parser, name="amp", default=False, required=False, help="Enable Automatic Mixed Precision Computation to maximise performance." ) _add_bool_argument( parser=parser, name="xla", default=False, required=False, help="Enable Tensorflow XLA to maximise performance." ) parser.add_argument('--batch_size', default=16, type=int, help='Evaluation batch size.') FLAGS, unknown_args = parser.parse_known_args() if len(unknown_args) > 0: for bad_arg in unknown_args: print("ERROR: Unknown command line arg: %s" % bad_arg) raise ValueError("Invalid command line arg(s)") return FLAGS def export_model(RUNNING_CONFIG): if RUNNING_CONFIG.amp: os.environ["TF_ENABLE_AUTO_MIXED_PRECISION_GRAPH_REWRITE"] = "1" model = UNet_v1( model_name="UNet_v1", input_format="NHWC", compute_format=RUNNING_CONFIG.data_format, n_output_channels=1, unet_variant=RUNNING_CONFIG.unet_variant, weight_init_method="he_normal", activation_fn=RUNNING_CONFIG.activation_fn ) config_proto = tf.ConfigProto() config_proto.allow_soft_placement = True config_proto.log_device_placement = False config_proto.gpu_options.allow_growth = True if RUNNING_CONFIG.xla: # Only working on single GPU LOGGER.log("XLA is activated - Experimental Feature") config_proto.graph_options.optimizer_options.global_jit_level = tf.OptimizerOptions.ON_1 config_proto.gpu_options.force_gpu_compatible = True # Force pinned memory run_config = tf.estimator.RunConfig( model_dir=None, tf_random_seed=None, save_summary_steps=1e9, # disabled save_checkpoints_steps=None, save_checkpoints_secs=None, session_config=config_proto, keep_checkpoint_max=None, keep_checkpoint_every_n_hours=1e9, # disabled log_step_count_steps=1e9, train_distribute=None, device_fn=None, protocol=None, eval_distribute=None, experimental_distribute=None ) estimator = tf.estimator.Estimator( model_fn=model, model_dir=RUNNING_CONFIG.model_checkpoint_path, config=run_config, params={'debug_verbosity': 0} ) LOGGER.log('[*] Exporting the model ...') input_type = tf.float32 if RUNNING_CONFIG.input_dtype else tf.float16 def get_serving_input_receiver_fn(): input_shape = [RUNNING_CONFIG.batch_size, 512, 512, 1] def serving_input_receiver_fn(): features = tf.placeholder(dtype=input_type, shape=input_shape, name='input_tensor') return tf.estimator.export.TensorServingInputReceiver(features=features, receiver_tensors=features) return serving_input_receiver_fn export_path = estimator.export_saved_model( export_dir_base=RUNNING_CONFIG.export_dir, serving_input_receiver_fn=get_serving_input_receiver_fn(), checkpoint_path=RUNNING_CONFIG.model_checkpoint_path ) LOGGER.log('[*] Done! path: `%s`' % export_path.decode()) if __name__ == '__main__': tf.logging.set_verbosity(tf.logging.ERROR) tf.disable_eager_execution() flags = get_export_flags() for endpattern in [".index", ".meta"]: file_to_check = flags.model_checkpoint_path + endpattern if not os.path.isfile(file_to_check): raise FileNotFoundError("The checkpoint file `%s` does not exist" % file_to_check) print(" ========================= Export Flags =========================\n") pprint.pprint(dict(flags._get_kwargs())) print("\n %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%") export_model(flags)
PyTorch/SpeechRecognition/wav2vec2/common
common
fairseq_fake_modules
# Copyright (c) 2023, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. '''Fake fairseq.* modules allowing to torch.load fairseq checkpoints.''' import sys class Dummy: pass class FakeModule: def __init__(self, classes=["AverageMeter", "TimeMeter", "StopwatchMeter"]): [setattr(self, cls, Dummy) for cls in classes] sys.modules["fairseq"] = Dummy() sys.modules["fairseq.data"] = Dummy() sys.modules["fairseq.data.dictionary"] = FakeModule(["Dictionary"]) sys.modules["fairseq.logging.meters"] = FakeModule() sys.modules["fairseq.meters"] = FakeModule()
Tools/DGLPyTorch/SyntheticGraphGeneration/syngen/generator/graph
graph
utils
# Copyright (c) 2023, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os import logging import math import multiprocessing from datetime import datetime from functools import partial from typing import Tuple, Union, Optional import cupy as cp import matplotlib.pyplot as plt import numpy as np from tqdm import tqdm from pylibraft.random import rmat from scipy import stats from syngen.utils import NDArray, infer_operator from syngen.utils.utils import infer_operator from syngen.utils.io_utils import dump_generated_graph from syngen.utils.memory_manager import MemoryManager from syngen.utils.types import NDArray logger = logging.getLogger(__name__) def move_ndarray_to_host(ndarray: NDArray): if isinstance(ndarray, np.ndarray): return ndarray elif isinstance(ndarray, cp.ndarray): return cp.asnumpy(ndarray) else: raise ValueError('supports only numpy and cupy ndarrays') def rearrange_graph( edge_list: NDArray, src_nodes: NDArray, dst_nodes: NDArray, assume_unique: bool = False, ) -> Tuple[NDArray, NDArray]: """ Transforms a bipartite graph from edge list format to lower_left and upper_right adjacency matrices. Returned matrices are in coordinate list format. """ operator = infer_operator(edge_list) if not isinstance(src_nodes, (np.ndarray, cp.ndarray)): raise ValueError('src_nodes: expected type NDArray, but %s was passed', type(src_nodes)) if not isinstance(dst_nodes, (np.ndarray, cp.ndarray)): raise ValueError('dst_nodes: expected type NDArray, but %s was passed', type(dst_nodes)) if not assume_unique: src_nodes = operator.unique(src_nodes) dst_nodes = operator.unique(dst_nodes) if operator.intersect1d(src_nodes, dst_nodes, assume_unique=True).size != 0: raise ValueError('node sets cannot intersect') edge_list = edge_list.flatten() node_set = operator.hstack([src_nodes, dst_nodes]) pos_to_new_id = operator.argsort(node_set) sorted_node_set = node_set[pos_to_new_id] pos_in_sorted_nodeset = operator.searchsorted(sorted_node_set, edge_list) # need to validate since errors could be ignored # https://docs.cupy.dev/en/stable/user_guide/difference.html#out-of-bounds-indices message = 'all ids in a graph should be in one of the node sets' if operator.any(pos_in_sorted_nodeset == len(sorted_node_set)): raise ValueError(message) if operator.any(sorted_node_set[pos_in_sorted_nodeset] != edge_list): raise ValueError(message) edge_list_mapped = pos_to_new_id[pos_in_sorted_nodeset].reshape(-1, 2) upper_right = edge_list_mapped[edge_list_mapped[:, 0] < len(src_nodes)] upper_right[:, 1] -= len(src_nodes) lower_left = edge_list_mapped[edge_list_mapped[:, 0] >= len(src_nodes)] lower_left[:, 0] -= len(src_nodes) return lower_left, upper_right def reindex_graph( edge_list: NDArray, return_counts: bool = False, ) -> Union[NDArray, Tuple[NDArray, int, int]]: """ Reindexes a graph by assigning node ids starting from 0. Returns the processed graph and, optionally, number of nodes and number of edges. """ operator = infer_operator(edge_list) nodes, inverse_flat = operator.unique(edge_list, return_inverse=True) edge_list_reindexed = inverse_flat.reshape(edge_list.shape) if return_counts: return edge_list_reindexed, len(nodes), len(edge_list) else: return edge_list_reindexed def get_reversed_part(part, gpu=False, operator=None): operator = operator or (cp if gpu else np) new_part = operator.empty_like(part) new_part[:, 0] = part[:, 1] new_part[:, 1] = part[:, 0] return new_part # Postprocessing def recreate_graph(lower: NDArray, upper: NDArray, offset: int, gpu=False): assert ( lower is not None and upper is not None ), "Upper and lower cannot be None" operator = cp if gpu else np lower[:, 0] = lower[:, 0] + offset upper[:, 1] = upper[:, 1] + offset new_graph = operator.concatenate((lower, upper), axis=0) return new_graph def recreate_bipartite_nondirected(graph, row_shape): upper = [(row, col + row_shape) for row, col in graph] lower = [(col, row) for row, col in upper] new_graph = upper + lower return new_graph def to_adj_matrix(graph, shape): matrix = np.zeros(shape=shape, dtype=np.bool) arr_indicies = np.array(graph) matrix[arr_indicies[:, 0], arr_indicies[:, 1]] = 1 return matrix def plot_graph_adj(graph, shape): graph_adj = to_adj_matrix(graph, shape=shape) return plt.imshow(graph_adj, cmap="binary", interpolation="nearest") def graph_to_snap_file(A, filename): np.savetxt(filename, A, fmt="%i", delimiter="\t") def effective_nonsquare_rmat_approximate( theta, E, A_shape, noise_scaling=1.0, batch_size=1000, dtype=np.int64, custom_samplers=None, generate_back_edges=False, verbose=False, ): """ This function generates list of edges using modified RMat approach Args: theta (np.array): seeding matrix, needs to be shape 2x2 E (int): number of edges to be generated A_shape (tuple): shape of resulting adjacency matrix. numbers has to be powers of 2 A_shape should be equal to (ceil(log2(X)),ceil(log2(Y))) X,Y are dimensions of original adjacency noise_scaling (float 0..1): noise scaling factor for good degree distribution batch_size (int): edges are generated in batches of batch_size size dtype (numpy dtype np.int32/np.int64): dtype of nodes id's custom_samplers (List[scipy.stats.rv_discrete]): samplers for each step of genration process generate_back_edges (bool): if True then generated edges will also have "back" edges. Not that setting to True for partite graphs makes no sense. Returns: A (np.array 2 x E): matrix containing in every row a signle edge. Edge is always directed 0'th column is FROM edge 1st is TO edge mtx_shape (tuple) - shape of adjecency matrix (A contains list of edges, this is Adjecency metrix shape) custom_samplers (List[scipy.stats.rv_discrete]) - list of samplers needed to generate edges from the same disctribution for multiple runs of the function Description: The generation will consist of theta^[n] (x) theta_p^[m] (x) theta_q^[l] ^[n] is kronecker power (x) is matrix kronecker product theta_p (2x1) and theta_q(1x2) are marginals of theta This way we can generate rectangular shape of adjecency matrix e.g. for bipatrite graphs """ def get_row_col_addres(thetas_n): thetas_r = [t.shape[0] for t in thetas_n] thetas_c = [t.shape[1] for t in thetas_n] row_n = np.prod(thetas_r) # theta_r**quadrant_sequence.shape[1] col_n = np.prod(thetas_c) # theta_c**quadrant_sequence.shape[1] row_adders = np.array( [ int(row_n / thetas_r[i] ** (i + 1)) % row_n for i in range(len(thetas_n)) ] ) # there has to be % as we can have thetas_r[i]==1 col_adders = np.array( [ int(col_n / thetas_c[i] ** (i + 1)) % col_n for i in range(len(thetas_n)) ] ) return row_adders, col_adders, thetas_r, thetas_c, row_n, col_n def parse_quadrants( quadrant_sequence, thetas_n, row_adders, col_addres, thetas_r, thetas_c, row_n, col_n, dtype=np.int64, ): N = len(thetas_n) new_edges = np.zeros( shape=(quadrant_sequence.shape[0], 2) ) # 2 because 0 col=rows_addresses, 1st col = columns row_addr = np.array(quadrant_sequence // thetas_c, dtype=dtype) col_addr = np.array(quadrant_sequence % thetas_c, dtype=dtype) row_adders = np.array( [int(row_n / thetas_r[i] ** (i + 1)) % row_n for i in range(N)] ) # there has to be % as we can have thetas_r[i]==1 col_adders = np.array( [int(col_n / thetas_c[i] ** (i + 1)) % col_n for i in range(N)] ) new_edges[:, 0] = np.sum(np.multiply(row_addr, row_adders), axis=1) new_edges[:, 1] = np.sum(np.multiply(col_addr, col_adders), axis=1) return new_edges if batch_size > E: # if bs>E batch_size = int(E // 2 * 2) if generate_back_edges: assert ( batch_size % 2 == 0 and batch_size >= 2 ), "batch size has to be odd and >1" assert ( np.abs((np.sum(theta) - 1.0)) < 1e-6 ), "Theta probabilities has to sum to 1.0" assert (theta.shape[0] == 2) and ( theta.shape[1] == 2 ), "Only 2x2 seeding matrixes are acceptable" assert len(A_shape) == 2, "A_shape needs to be of len 2" # get appropriate number of n,m,l always m=0 or l=0 (or both for rectangular adjecency) r = A_shape[0] c = A_shape[1] n = min(r, c) # theta^[n] (x) theta_p^[m] (x) theta_q^[l] m = max(0, r - c) # flake8: noqa l = max(0, c - r) # calc values of marginal theta matrixes theta_p = theta.sum(axis=1).reshape((2, -1)) # 2x1 theta_q = theta.sum(axis=0).reshape((1, -1)) # 1x2 # get all thetas thetas_n = [theta] * n + [theta_p] * m + [theta_q] * l # prepare samplers for each of n+m+l steps if custom_samplers is None: custom_samplers = [] for i in range(n + m + l): theta_n = thetas_n[ i ] # each of n+m+l steps have their own theta_n which can be theta/theta_p or theta_q + # noise noise = noise_scaling * np.random.uniform( -1, 1, size=theta_n.shape ) noise_to_add = np.multiply(theta_n, noise) theta_n = theta_n + noise_to_add theta_n = theta_n / np.sum(theta_n) cstm_n = "step_" + str(i) theta_r = theta_n.shape[0] theta_c = theta_n.shape[1] xk = tuple(range(theta_r * theta_c)) pk = theta_n.reshape(-1) cstm_s = stats.rv_discrete(name=cstm_n, values=(xk, pk)) custom_samplers.append(cstm_s) # Prepare all batch sizes needed for generation if batch_size == 0: batch_count = 0 # XXX: why does this happen anyways? else: batch_count = E // batch_size last_batch_size = E - batch_count * batch_size if last_batch_size % 2 > 0 and generate_back_edges: last_batch_size -= 1 A = np.zeros((E, 2), dtype=np.int64) num_sequences = batch_size last_num_sequences = last_batch_size if ( generate_back_edges ): # in case of generating back edges we need to sample just E/2 last_num_sequences = last_batch_size // 2 num_sequences = batch_size // 2 new_back_edges = np.zeros(shape=(num_sequences, 2)) quadrant_sequence = np.zeros(shape=(num_sequences, n + m + l), dtype=dtype) ( row_adders, col_addres, thetas_r, thetas_c, row_n, col_n, ) = get_row_col_addres(thetas_n) # generate sequences of quadrants from previously prepared samplers batch_itr = range(batch_count) if verbose: batch_itr = tqdm(batch_itr) for e in batch_itr: for i in range( n + m + l ): # each steps in generation has its own sampler smpl = custom_samplers[i].rvs(size=num_sequences) quadrant_sequence[:, i] = smpl # produce new edges new_edges = parse_quadrants( quadrant_sequence, thetas_n, row_adders, col_addres, thetas_r, thetas_c, row_n, col_n, dtype=dtype, ) if generate_back_edges: new_back_edges[:, [0, 1]] = new_edges[:, [1, 0]] # swap columns A[ e * batch_size: (e + 1) * batch_size: 2, : ] = new_edges # we need interleave so that back edges are "right after" normal edges A[ e * batch_size + 1: (e + 1) * batch_size: 2, : ] = new_back_edges else: A[e * batch_size: (e + 1) * batch_size, :] = new_edges # generate last batch if last_batch_size > 0: for i in range(n + m + l): smpl = custom_samplers[i].rvs(size=last_num_sequences) quadrant_sequence[:last_num_sequences, i] = smpl new_edges = parse_quadrants( quadrant_sequence[:last_num_sequences, :], thetas_n, row_adders, col_addres, thetas_r, thetas_c, row_n, col_n, dtype=dtype, ) if generate_back_edges: new_back_edges[:last_num_sequences, [0, 1]] = new_edges[ :last_num_sequences, [1, 0] ] # we need interleave so that back edges are "right after" normal edges A[ batch_count * batch_size: batch_count * batch_size + last_batch_size: 2, :, ] = new_edges # np.concatenate((new_edges,new_back_edges[:last_num_sequences,:]),axis=0) A[ batch_count * batch_size + 1: batch_count * batch_size + last_batch_size: 2, :, ] = new_back_edges[:last_num_sequences, :] else: A[ batch_count * batch_size: batch_count * batch_size + last_batch_size, :, ] = new_edges mtx_shape = ( np.prod([t.shape[0] for t in thetas_n]), np.prod([t.shape[1] for t in thetas_n]), ) # shape of resulting adjacency matrix return A, mtx_shape, custom_samplers def effective_nonsquare_rmat_exact( theta, E, A_shape, noise_scaling=1.0, batch_size=1000, dtype=np.int64, custom_samplers=None, remove_selfloops=False, generate_back_edges=False, return_node_ids=0, verbose=False, ): """ This function generates list of edges using modified RMat approach based on effective_nonsuqare_rmat_approximate Args: theta (np.array): seeding matrix, needs to be shape 2x2 E (int): number of edges to be generated A_shape (tuple): shape of resulting adjacency matrix. numbers has to be powers of 2 A_shape should be equal to (ceil(log2(X)),ceil(log2(Y))) X,Y are dimensions of original adjacency noise_scaling (float 0..1): noise scaling factor for good degree distribution batch_size (int): edges are generated in batches of batch_size size dtype (numpy dtype np.int32/np.int64): dtype of nodes id's remove_selfloops (bool): If true edges n->n will not be generated. Note that for partite graphs this makes no sense generate_back_edges (bool): if True then generated edges will also have "back" edges. Not that setting to True for partite graphs makes no sense. Returns: A (np.array 2 x E) - matrix containing in every row a signle edge. Edge is always directed 0'th column is FROM edge 1st is TO edge mtx_shape (tuple) - shape of adjecency matrix (A contains list of edges, this is Adjecency metrix shape) custom_samplers (List[scipy.stats.rv_discrete]) - list of samplers needed to generate edges from the same disctribution for multiple runs of the function Description: see effective_nonsuqare_rmat_approximate """ heuristics = 1.5 if verbose: print("Getting egdes") A, mtx_shape, cs = effective_nonsquare_rmat_approximate( theta, int(heuristics * E), A_shape, noise_scaling=noise_scaling, batch_size=batch_size, dtype=dtype, custom_samplers=custom_samplers, generate_back_edges=generate_back_edges, verbose=verbose, ) if generate_back_edges: A = A[ np.sort(np.unique(A, return_index=True, axis=0)[1]) ] # permutation is not needed here else: if verbose: print("Getting unique egdes") A = np.unique(A, axis=0) if verbose: print("Permuting edges") perm = np.random.permutation( A.shape[0] ) # we need to permute it as othervise unique returns edges in order A = A[perm] if remove_selfloops: if verbose: print("Removing selfloops") A = np.delete(A, np.where(A[:, 0] == A[:, 1]), axis=0) E_already_generated = A.shape[0] if E_already_generated >= E: if return_node_ids == 2: return A[:E, :], np.unique(A[:E, :][:, 0]), np.unique(A[:E, :][:, 1]), mtx_shape, cs if return_node_ids == 1: return A[:E, :], np.unique(A[:E, :]), mtx_shape, cs return A[:E, :], mtx_shape, cs else: while E_already_generated < E: if verbose: print("Generating some additional edges") E_to_generate = int(heuristics * (E - E_already_generated)) A_next, mtx_shape, cs = effective_nonsquare_rmat_approximate( theta, E_to_generate, A_shape, noise_scaling=noise_scaling, batch_size=batch_size, dtype=dtype, custom_samplers=cs, verbose=verbose, ) if remove_selfloops: A_next = np.delete( A_next, np.where(A_next[:, 0] == A_next[:, 1]), axis=0 ) A = np.concatenate((A, A_next), axis=0) if generate_back_edges: A = A[np.sort(np.unique(A, return_index=True, axis=0)[1])] else: A = np.unique(A, axis=0) perm = np.random.permutation(A.shape[0]) A = A[perm] E_already_generated = A.shape[0] if return_node_ids == 2: return A[:E, :], np.unique(A[:E, :][:, 0]), np.unique(A[:E, :][:, 1]), mtx_shape, cs if return_node_ids == 1: return A[:E, :], np.unique(A[:E, :]), mtx_shape, cs return A[:E, :], mtx_shape, cs def cupy_unique_axis0(array): # https://stackoverflow.com/questions/58662085/is-there-a-cupy-version-supporting-axis-option-in-cupy-unique-function-any sortarr = array[cp.lexsort(array.T[::-1])] mask = cp.empty(array.shape[0], dtype=cp.bool_) mask[0] = True mask[1:] = cp.any(sortarr[1:] != sortarr[:-1], axis=1) return sortarr[mask] def unique_axis0(ar: NDArray) -> NDArray: """ Uniform way of calling operator.unique(ar, axis=0). axis != None is not supported in cupy yet. This function provides a workaround for one of the cases. """ operator = infer_operator(ar) if operator == cp: return cupy_unique_axis0(ar) else: return np.unique(ar, axis=0) def generate_gpu_rmat( a, b, c, d, r_scale, c_scale, n_edges, noise=0.5, is_directed=False, has_self_loop=False, return_node_ids=0, ): if not is_directed and r_scale != c_scale: raise ValueError('undirected generation works only for square adj matrix') if not is_directed: n_edges = n_edges // 2 gen_graph = None HEURISTIC = 1.2 edges_to_generate = int(HEURISTIC * n_edges) theta_len = max(r_scale, c_scale) base_theta = [a, b, c, d] if noise > 0: full_theta = [] for i in range(theta_len): noise_uniform = noise * np.random.uniform( -1, 1, size=len(base_theta) ) noise_to_add = np.multiply(base_theta, noise_uniform) theta_n = base_theta + noise_to_add theta_n = theta_n / np.sum(theta_n) full_theta.append(theta_n) else: full_theta = base_theta * theta_len theta_cpu = np.array(full_theta, dtype=np.float32) theta = cp.asarray(theta_cpu) while gen_graph is None or gen_graph.shape[0] < n_edges: tmp = cp.empty((edges_to_generate, 2), dtype=cp.int32) seed = cp.random.randint(0, high=1_000_000, size=None, dtype=int) rmat(tmp, theta, r_scale, c_scale, seed=seed) # Remove self loops if not has_self_loop: tmp = tmp[tmp[:, 0] != tmp[:, 1]] # Keep only one sided edges if not is_directed: tmp = tmp[tmp[:, 0] <= tmp[:, 1]] if gen_graph is None: # Remove duplicates gen_graph = cupy_unique_axis0(tmp) else: gen_graph = cp.concatenate((gen_graph, tmp), axis=0) # Remove duplicates gen_graph = cupy_unique_axis0(gen_graph) gen_graph = gen_graph[:n_edges] if not is_directed: gen_graph_backward = cp.empty((n_edges, 2), dtype=cp.int32) gen_graph_backward[:, 0] = gen_graph[:, 1] gen_graph_backward[:, 1] = gen_graph[:, 0] gen_graph = cp.concatenate((gen_graph, gen_graph_backward), axis=0) gen_graph = cupy_unique_axis0( gen_graph ) if not has_self_loop: gen_graph = gen_graph[gen_graph[:, 0] != gen_graph[:, 1]] if return_node_ids == 2: return cp.asnumpy(gen_graph), cp.asnumpy(cp.unique(gen_graph[:, 0])), cp.asnumpy(cp.unique(gen_graph[:, 1])) if return_node_ids == 1: return cp.asnumpy(gen_graph), cp.asnumpy(cp.unique(gen_graph)) return cp.asnumpy(gen_graph) def generate_theta(base_theta, noise, theta_len, is_directed): if noise > 0: full_theta = [] for i in range(theta_len): noise_uniform = noise * np.random.uniform( -1, 1, size=len(base_theta) ) noise_to_add = np.multiply(base_theta, noise_uniform) theta_n = base_theta + noise_to_add if not is_directed: theta_n[2] = theta_n[1] theta_n = theta_n / np.sum(theta_n) full_theta.append(theta_n) else: full_theta = [base_theta] * theta_len return full_theta def prepare_chunks(full_theta, r_scale, c_scale, gpu_bytes_to_use, edges_to_generate): if r_scale > 32 or c_scale > 32: bytes_per_edge = 8 max_id = 9223372036854775807 # int64 max else: bytes_per_edge = 4 max_id = 2147483647 # int32 max bytes_to_generate = edges_to_generate * 2 * bytes_per_edge skip_theta = 0 # approximation while (bytes_to_generate >> 2 * skip_theta) > gpu_bytes_to_use \ or (bytes_to_generate >> 2 * skip_theta) > max_id: skip_theta += 1 if skip_theta == 0: return [], np.array([edges_to_generate]), full_theta, 0, r_scale, c_scale # chunk size is limited by the smaller side of the rectangular graph while abs(r_scale - c_scale) > skip_theta: skip_theta += 1 def repeat(a, scale): if scale == 1: return a return np.repeat(np.repeat(a, scale, axis=0), scale, axis=1) def tile(a, scale): if scale == 1: return a return np.tile(a, (scale, scale)) def prepare_prefixes(skip_theta): if skip_theta > 0: prefix_theta = full_theta[:skip_theta] gen_theta_len = max(r_scale, c_scale) - skip_theta prefix_edges = np.ones((1 << skip_theta, 1 << skip_theta), dtype=np.float64) prefixes = np.zeros((2, 1 << skip_theta, 1 << skip_theta), dtype=np.int32) for theta_idx, theta in enumerate(prefix_theta): pref_src = np.array([[0, 0], [1, 1]]) << theta_idx pref_dst = np.array([[0, 1], [0, 1]]) << theta_idx theta = np.array(theta, dtype=np.float64).reshape(2, 2) repeat_scale = 1 << (skip_theta - theta_idx - 1) tile_scale = 1 << theta_idx prefix_edges = prefix_edges * tile(repeat(theta, repeat_scale), tile_scale) prefixes[0] = prefixes[0] + tile(repeat(pref_src, repeat_scale), tile_scale) prefixes[1] = prefixes[1] + tile(repeat(pref_dst, repeat_scale), tile_scale) if r_scale != c_scale: # probabilities in the rectangular matrix should sum up to 1.0 r_len = 2 ** (r_scale - gen_theta_len) c_len = 2 ** (c_scale - gen_theta_len) prefix_edges[:r_len, :c_len] = prefix_edges[:r_len, :c_len] / prefix_edges[:r_len, :c_len].sum() prefixes[int(r_scale > c_scale), :r_len, :c_len] = \ prefixes[int(r_scale > c_scale), :r_len, :c_len] >> abs(r_scale - c_scale) prefix_edges = np.ceil(prefix_edges * edges_to_generate).astype(np.int32).reshape(-1) prefixes = prefixes.reshape(2, -1) else: prefixes = [] prefix_edges = np.array([edges_to_generate]) return prefixes, prefix_edges prefixes, prefix_edges = prepare_prefixes(skip_theta) while prefix_edges.max() * 2 * bytes_per_edge > gpu_bytes_to_use: skip_theta += 1 prefixes, prefix_edges = prepare_prefixes(skip_theta) generation_theta = full_theta[skip_theta:] return prefixes, prefix_edges, generation_theta, skip_theta, len(generation_theta), len(generation_theta) def _generate_gpu_chunk_rmat( chunk_info, prefixes, prefix_edges, has_self_loop, is_directed, generation_theta, r_log2_nodes, c_log2_nodes, r_pref_len, c_pref_len, row_len, gpus, dtype='int32', return_node_ids=0, memmap_kwargs: Optional = None, chunk_save_path_format: Optional[str] = None): chunk_id, chunk_end = chunk_info chunk_size = prefix_edges[chunk_id] if gpus > 1: gpu_id = int(multiprocessing.current_process()._identity[0]) % gpus else: gpu_id = 0 theta_cpu = np.array(generation_theta, dtype=np.float32) edge_list = None is_diagonal_chunk = ((chunk_id // row_len) == (chunk_id % row_len)) use_memmap = memmap_kwargs is not None if use_memmap: memmap_outfile = np.load(file=memmap_kwargs['filename'], mmap_mode='r+') with cp.cuda.Device(gpu_id): theta = cp.asarray(theta_cpu) while edge_list is None or edge_list.shape[0] < prefix_edges[chunk_id]: tmp = cp.empty((prefix_edges[chunk_id], 2), dtype=dtype) seed = cp.random.randint(0, high=1_000_000, size=None, dtype=int) rmat(tmp, theta, r_log2_nodes, c_log2_nodes, seed=seed) if not has_self_loop and is_diagonal_chunk: tmp = tmp[tmp[:, 0] != tmp[:, 1]] # undirected diagonal_case if not is_directed and is_diagonal_chunk: tmp = tmp[tmp[:, 0] <= tmp[:, 1]] tmp = cupy_unique_axis0(tmp) if edge_list is None: edge_list = tmp else: edge_list = cp.concatenate((edge_list, tmp), axis=0) del tmp edge_list = cupy_unique_axis0(edge_list) if len(prefix_edges) > 1: edge_list[:, 0] = (edge_list[:, 0] << r_pref_len) + prefixes[0][chunk_id] edge_list[:, 1] = (edge_list[:, 1] << c_pref_len) + prefixes[1][chunk_id] edge_list = edge_list[:prefix_edges[chunk_id]] if return_node_ids == 2: src_nodes_ids = cp.asnumpy(cp.unique(edge_list[:, 0])) dst_nodes_ids = cp.asnumpy(cp.unique(edge_list[:, 1])) if return_node_ids == 1: nodes_ids = cp.asnumpy(cp.unique(edge_list)) result = cp.asnumpy(edge_list) if use_memmap: memmap_outfile[chunk_end-chunk_size:chunk_end] = result del edge_list if chunk_save_path_format is not None: dump_generated_graph(chunk_save_path_format.format(chunk_id=chunk_id), result) result = len(result) if use_memmap: result = None if return_node_ids == 2: return result, src_nodes_ids, dst_nodes_ids if return_node_ids == 1: return result, nodes_ids return result def generate_gpu_chunked_rmat( a, b, c, d, r_scale, c_scale, n_edges, noise=0.5, is_directed=False, has_self_loop=False, gpus=None, return_node_ids=0, save_path: Optional[str] = None, verbose: bool = False, ): if not is_directed and r_scale != c_scale: raise ValueError('undirected generation works only for square adj matrix') base_theta = [a, b, c, d] theta_len = max(r_scale, c_scale) full_theta = generate_theta(base_theta, noise, theta_len, is_directed) if gpus is None: gpus = MemoryManager().get_available_gpus() gpu_bytes_to_use = MemoryManager().get_min_available_across_gpus_memory(gpus=gpus) gpu_bytes_to_use = math.floor(gpu_bytes_to_use * 0.10) prefixes, prefix_edges, generation_theta, prefix_len, r_log2_nodes, c_log2_nodes = \ prepare_chunks(full_theta, r_scale, c_scale, gpu_bytes_to_use, n_edges) chunk_ids = list(range(len(prefix_edges))) row_len = 1 << prefix_len r_pref_len = r_scale - len(generation_theta) c_pref_len = c_scale - len(generation_theta) if not is_directed: # generate a triangular adj matrix chunk_ids = [i for i in chunk_ids if (i // row_len) <= (i % row_len)] # reduce the diagonal chunks for i in range(prefix_len * 2): prefix_edges[i * row_len + i] = prefix_edges[i * row_len + i] // 2 if r_scale != c_scale: chunk_ids = [i for i in chunk_ids if (i // row_len) < 2 ** r_pref_len and (i % row_len) < 2 ** c_pref_len] is_single_chunk = len(chunk_ids) == 1 memmap_kwargs = None chunk_save_path_format = None use_memmap = False if save_path and os.path.isdir(save_path): chunk_save_path_format = os.path.join(save_path, 'chunk_{chunk_id}.npy') elif save_path and save_path.endswith('.npy'): use_memmap = True memmap_shape = (sum(prefix_edges[chunk_ids]), 2) memmap_dtype = np.uint64 if theta_len > 32 else np.uint32 memmap_kwargs = dict( filename=save_path, ) memmap_outfile = np.lib.format.open_memmap(save_path, dtype=memmap_dtype, shape=memmap_shape, mode='w+') dtype = cp.int64 if theta_len > 32 else cp.int32 _generate_gpu_chunk_rmat_p = partial( _generate_gpu_chunk_rmat, prefixes=prefixes, prefix_edges=prefix_edges, has_self_loop=has_self_loop, is_directed=is_directed, generation_theta=generation_theta, r_log2_nodes=r_log2_nodes, c_log2_nodes=c_log2_nodes, r_pref_len=r_pref_len, c_pref_len=c_pref_len, row_len=row_len, dtype=dtype, return_node_ids=return_node_ids, chunk_save_path_format=chunk_save_path_format, memmap_kwargs=memmap_kwargs, gpus=1 if is_single_chunk else gpus, ) if is_single_chunk: chunk_res = _generate_gpu_chunk_rmat_p((chunk_ids[0], prefix_edges[chunk_ids[0]])) if return_node_ids == 2: result, src_node_ids, dst_node_ids = chunk_res elif return_node_ids == 1: result, node_ids = chunk_res else: result = chunk_res if use_memmap: result = memmap_outfile else: multiprocessing.set_start_method('spawn', force=True) sub_res_lists = [] if return_node_ids == 2: src_node_ids_presence = np.full(2**r_scale, False) dst_node_ids_presence = np.full(2**c_scale, False) elif return_node_ids == 1: node_ids_presence = np.full(2**theta_len, False) with multiprocessing.Pool(processes=gpus) as pool: chunk_res = pool.imap_unordered(_generate_gpu_chunk_rmat_p, zip(chunk_ids, np.cumsum(prefix_edges[chunk_ids])), chunksize=(len(chunk_ids)+gpus-1) // gpus ) if verbose: chunk_res = tqdm(chunk_res, total=len(chunk_ids)) if return_node_ids == 2: for res, src_n_ids, dst_n_ids in chunk_res: sub_res_lists.append(res) src_node_ids_presence[src_n_ids] = True dst_node_ids_presence[dst_n_ids] = True elif return_node_ids == 1: for res, n_ids in chunk_res: sub_res_lists.append(res) node_ids_presence[n_ids] = True else: sub_res_lists = list(chunk_res) if use_memmap: result = memmap_outfile elif chunk_save_path_format is None: result = np.concatenate(sub_res_lists) else: result = int(np.sum(sub_res_lists)) if return_node_ids == 2: src_node_ids, = np.where(src_node_ids_presence) dst_node_ids, = np.where(dst_node_ids_presence) elif return_node_ids == 1: node_ids, = np.where(node_ids_presence) if return_node_ids == 2: return result, src_node_ids, dst_node_ids if return_node_ids == 1: return result, node_ids return result def get_degree_distribution(vertices, gpu=False, operator=None): operator = operator or (cp if gpu else np) _, degrees = operator.unique(vertices, return_counts=True) degree_values, degree_counts = operator.unique(degrees, return_counts=True) return degree_values, degree_counts class BaseLogger: """ Base logger class Args: logdir (str): path to the logging directory """ def __init__(self, logdir: str = "tmp"): self.logdir = logdir os.makedirs(self.logdir, exist_ok=True) currentDateAndTime = datetime.now() self.logname = ( f'{currentDateAndTime.strftime("%Y_%m_%d_%H_%M_%S")}.txt' ) self.logpath = os.path.join(self.logdir, self.logname) self.setup_logger() self.log("Initialized logger") def setup_logger(self): """ This function setups logger """ logging.basicConfig( filename=self.logpath, filemode="a", format="%(asctime)s| %(message)s", datefmt="%Y/%m/%d %H:%M:%S", level=logging.DEBUG, ) def log(self, msg: str): """ This function logs messages in debug mode Args: msg (str): message to be printed """ logging.debug(msg) def _reshuffle(X: NDArray, mask: NDArray, max_node_id: int) -> None: """ Shuffles dst nodes of edges specified by idx. Preserves degree distribution and keeps edge list sorted. """ operator = infer_operator(X) if not operator.any(mask): return target = X[mask, 1] operator.random.shuffle(target) X[mask, 1] = target src_node_mask = operator.zeros(max_node_id + 1, dtype=operator.bool_) src_node_mask[X[mask, 0]] = True to_sort_mask = operator.zeros(X.shape[0], dtype=operator.bool_) to_sort_mask = src_node_mask[X[:, 0]] to_sort = X[to_sort_mask] to_sort = to_sort[operator.lexsort(to_sort.T[::-1])] X[to_sort_mask] = to_sort def _find_correct_edges( X: NDArray, self_loops: bool = False, assume_sorted: bool = False, ) -> Tuple[NDArray, NDArray]: """ Finds duplicates and self loops in an edge list. """ operator = infer_operator(X) if not assume_sorted: X = X[operator.lexsort(X.T[::-1])] mask = operator.empty(X.shape[0], dtype=operator.bool_) mask[0] = True mask[1:] = operator.any(X[1:] != X[:-1], axis=1) if not self_loops: mask &= X[:, 0] != X[:, 1] return X, mask def postprocess_edge_list(X: NDArray, n_reshuffle: int = 0, self_loops: bool = False) -> NDArray: """ Removes multi-edges and (optionally) self-loops. If n_reshuffle > 0 is specified, edges are shuffled between nodes so that the degree distribution is preserved and less edges will be removed. Assumes node set is reindexed from min_id > 0 to max_id ~ N. """ max_node_id = X.max().item() X, mask = _find_correct_edges(X, self_loops=self_loops) for _ in range(n_reshuffle): _reshuffle(X, ~mask, max_node_id) X, mask = _find_correct_edges(X, self_loops=self_loops, assume_sorted=True) return X[mask]
PyTorch/SpeechSynthesis/Tacotron2/trtis_cpp/trtis_client
trtis_client
CMakeLists
cmake_minimum_required(VERSION 3.0 FATAL_ERROR) project(tacotron2_inference) if (DEFINED DEVEL AND NOT DEVEL EQUAL 0) if ("${CMAKE_CXX_COMPILER_ID}" MATCHES "GNU") # g++ warnings set(CPP_DEVEL_FLAGS "${CPP_DEVEL_FLAGS} -Wall") set(CPP_DEVEL_FLAGS "${CPP_DEVEL_FLAGS} -Werror") set(CPP_DEVEL_FLAGS "${CPP_DEVEL_FLAGS} -Wpedantic") set(CPP_DEVEL_FLAGS "${CPP_DEVEL_FLAGS} -Weffc++") set(CPP_DEVEL_FLAGS "${CPP_DEVEL_FLAGS} -Wextra") set(CPP_DEVEL_FLAGS "${CPP_DEVEL_FLAGS} -DDEVEL=1") # nvcc warnings set(CUDA_DEVEL_FLAGS "${CUDA_DEVEL_FLAGS} -Xcompiler -Wall") set(CUDA_DEVEL_FLAGS "${CUDA_DEVEL_FLAGS} -Xcompiler -Werror") set(CUDA_DEVEL_FLAGS "${CUDA_DEVEL_FLAGS} -Xcompiler -Weffc++") set(CUDA_DEVEL_FLAGS "${CUDA_DEVEL_FLAGS} -Xcompiler -Wextra") set(CUDA_DEVEL_FLAGS "${CUDA_DEVEL_FLAGS} -Xcompiler -DDEVEL=1") endif() endif() set(CMAKE_CXX_FLAGS_DEBUG "-g -O0") set(CMAKE_CXX_STANDARD 14) set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -fPIC") separate_arguments(CPP_DEVEL_FLAGS) separate_arguments(CUDA_DEVEL_FLAGS) add_subdirectory("src")
PyTorch/Recommendation/DLRM/preproc
preproc
parquet_to_binary
# Copyright (c) 2021 NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import numpy as np import pandas as pd import os from joblib import Parallel, delayed import glob import argparse import tqdm import subprocess def process_file(f, dst): label = '_c0' dense_columns = [f'_c{i}' for i in range(1, 14)] categorical_columns = [f'_c{i}' for i in range(14, 40)] all_columns_sorted = [f'_c{i}' for i in range(0, 40)] data = pd.read_parquet(f) data = data[all_columns_sorted] data[label] = data[label].astype(np.int32) data[dense_columns] = data[dense_columns].astype(np.float32) data[categorical_columns] = data[categorical_columns].astype(np.int32) data = data.to_records(index=False) data = data.tobytes() dst_file = dst + '/' + f.split('/')[-1] + '.bin' with open(dst_file, 'wb') as dst_fd: dst_fd.write(data) def main(): parser = argparse.ArgumentParser() parser.add_argument('--src_dir', type=str) parser.add_argument('--intermediate_dir', type=str) parser.add_argument('--dst_dir', type=str) parser.add_argument('--parallel_jobs', default=40, type=int) args = parser.parse_args() print('Processing train files...') train_src_files = glob.glob(args.src_dir + '/train/*.parquet') train_intermediate_dir = os.path.join(args.intermediate_dir, 'train') os.makedirs(train_intermediate_dir, exist_ok=True) Parallel(n_jobs=args.parallel_jobs)(delayed(process_file)(f, train_intermediate_dir) for f in tqdm.tqdm(train_src_files)) print('Train files conversion done') print('Processing test files...') test_src_files = glob.glob(args.src_dir + '/test/*.parquet') test_intermediate_dir = os.path.join(args.intermediate_dir, 'test') os.makedirs(test_intermediate_dir, exist_ok=True) Parallel(n_jobs=args.parallel_jobs)(delayed(process_file)(f, test_intermediate_dir) for f in tqdm.tqdm(test_src_files)) print('Test files conversion done') print('Processing validation files...') valid_src_files = glob.glob(args.src_dir + '/validation/*.parquet') valid_intermediate_dir = os.path.join(args.intermediate_dir, 'validation') os.makedirs(valid_intermediate_dir, exist_ok=True) Parallel(n_jobs=args.parallel_jobs)(delayed(process_file)(f, valid_intermediate_dir) for f in tqdm.tqdm(valid_src_files)) print('Validation files conversion done') os.makedirs(args.dst_dir, exist_ok=True) print('Concatenating train files') os.system(f'cat {train_intermediate_dir}/*.bin > {args.dst_dir}/train_data.bin') print('Concatenating test files') os.system(f'cat {test_intermediate_dir}/*.bin > {args.dst_dir}/test_data.bin') print('Concatenating validation files') os.system(f'cat {valid_intermediate_dir}/*.bin > {args.dst_dir}/validation_data.bin') print('Done') if __name__ == '__main__': main()
PyTorch/Segmentation/MaskRCNN/pytorch/configs/pascal_voc
pascal_voc
e2e_faster_rcnn_R_50_C4_1x_1_gpu_voc
MODEL: META_ARCHITECTURE: "GeneralizedRCNN" WEIGHT: "catalog://ImageNetPretrained/MSRA/R-50" RPN: PRE_NMS_TOP_N_TEST: 6000 POST_NMS_TOP_N_TEST: 300 ANCHOR_SIZES: (128, 256, 512) ROI_BOX_HEAD: NUM_CLASSES: 21 DATASETS: TRAIN: ("voc_2007_train", "voc_2007_val") TEST: ("voc_2007_test",) SOLVER: BASE_LR: 0.001 WEIGHT_DECAY: 0.0001 STEPS: (50000, ) MAX_ITER: 70000 IMS_PER_BATCH: 1 TEST: IMS_PER_BATCH: 1
TensorFlow2/Recommendation/WideAndDeep/triton/deployment_toolkit/triton_performance_runner/perf_analyzer
perf_analyzer
perf_config
# Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import Any from .exceptions import PerfAnalyzerException class PerfAnalyzerConfig: """ A config class to set arguments to the perf_analyzer. An argument set to None will use the perf_analyzer's default. """ perf_analyzer_args = [ "async", "sync", "measurement-interval", "measurement-mode", "measurement-request-count", "concurrency-range", "request-rate-range", "request-distribution", "request-intervals", "binary-search", "num-of-sequence", "latency-threshold", "max-threads", "stability-percentage", "max-trials", "percentile", "input-data", "shared-memory", "output-shared-memory-size", "sequence-length", "string-length", "string-data", ] perf_analyzer_multiple_args = [ "shape", ] input_to_options = [ "model-name", "model-version", "batch-size", "url", "protocol", "latency-report-file", "streaming", ] input_to_verbose = ["verbose", "extra-verbose"] def __init__(self): """ Construct a PerfAnalyzerConfig """ self._args = {k: None for k in self.perf_analyzer_args} self._multiple_args = {k: [] for k in self.perf_analyzer_multiple_args} self._options = { "-m": None, "-x": None, "-b": None, "-u": None, "-i": None, "-f": None, "-H": None, "-c": None, "-t": None, } self._verbose = {"-v": None, "-v -v": None} self._input_to_options = { "model-name": "-m", "model-version": "-x", "batch-size": "-b", "url": "-u", "protocol": "-i", "latency-report-file": "-f", "streaming": "-H", "concurrency": "-c", "threads": "-t", } self._input_to_verbose = {"verbose": "-v", "extra-verbose": "-v -v"} @classmethod def allowed_keys(cls): """ Returns ------- list of str The keys that are allowed to be passed into perf_analyzer """ return ( list(cls.perf_analyzer_args) + list(cls.perf_analyzer_multiple_args) + list(cls.input_to_options) + list(cls.input_to_verbose) ) def update_config(self, params=None): """ Allows setting values from a params dict Parameters ---------- params: dict keys are allowed args to perf_analyzer """ if params: for key in params: self[key] = params[key] def to_cli_string(self): """ Utility function to convert a config into a string of arguments to the perf_analyzer with CLI. Returns ------- str cli command string consisting of all arguments to the perf_analyzer set in the config, without the executable name. """ # single dashed options, then verbose flags, then main args args = [f"{k} {v}" for k, v in self._options.items() if v] args += [k for k, v in self._verbose.items() if v] args += [f"--{k}={v}" for k, v in self._args.items() if v] for k, v in self._multiple_args.items(): for item in v: args.append(f"--{k}={item}") return " ".join(args) def __getitem__(self, key: str): """ Gets an arguments value in config Parameters ---------- key : str The name of the argument to the perf_analyzer Returns ------- The value that the argument is set to in this config Raises ------ TritonModelAnalyzerException If argument not found in the config """ if key in self._args: return self._args[key] elif key in self._multiple_args: return self._multiple_args[key] elif key in self._input_to_options: return self._options[self._input_to_options[key]] elif key in self._input_to_verbose: return self._verbose[self._input_to_verbose[key]] else: raise PerfAnalyzerException(f"'{key}' Key not found in config") def __setitem__(self, key: str, value: Any): """ Sets an arguments value in config after checking if defined/supported. Parameters ---------- key : str The name of the argument to the perf_analyzer value : (any) The value to which the argument is being set Raises ------ TritonModelAnalyzerException If key is unsupported or undefined in the config class """ if key in self._args: self._args[key] = value elif key in self._multiple_args: self._multiple_args[key].append(value) elif key in self._input_to_options: self._options[self._input_to_options[key]] = value elif key in self._input_to_verbose: self._verbose[self._input_to_verbose[key]] = value else: raise PerfAnalyzerException( f"The argument '{key}' to the perf_analyzer " "is not supported by the model analyzer." )
TensorFlow2/Recommendation/DLRM_and_DCNv2/deployment/tf
tf
__init__
# Copyright (c) 2023, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # author: Tomasz Grel (tgrel@nvidia.com) #from constants import dense_model_name, hps_model_name from deployment.tf.deploy_dense import deploy_dense from deployment.tf.deploy_ensemble import deploy_ensemble from deployment.tf.deploy_sparse import deploy_sparse from deployment.tf.deploy_monolithic import deploy_monolithic
TensorFlow2/Recommendation/WideAndDeep/data/outbrain
outbrain
dataloader
# Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import cupy import horovod.tensorflow as hvd import tensorflow as tf from nvtabular.loader.tensorflow import KerasSequenceLoader from data.outbrain.defaults import LABEL_CHANNEL, MAP_FEATURE_CHANNEL, NUMERICAL_CHANNEL, ONEHOT_CHANNEL, \ MULTIHOT_CHANNEL cupy.random.seed(None) def seed_fn(): min_int, max_int = tf.int32.limits max_rand = max_int // hvd.size() # Generate a seed fragment on each worker seed_fragment = cupy.random.randint(0, max_rand).get() # Aggregate seed fragments from all Horovod workers seed_tensor = tf.constant(seed_fragment) reduced_seed = hvd.allreduce(seed_tensor, name="shuffle_seed", op=hvd.mpi_ops.Sum) return reduced_seed % max_rand def get_dataset(feature_spec, mapping, batch_size, buffer_size=0.1, parts_per_chunk=1, map_channel_enabled=False, shuffle=True): data_paths = feature_spec.get_paths_by_mapping(mapping) label_names = feature_spec.get_names_by_channel(LABEL_CHANNEL) cat_names = feature_spec.get_names_by_channel(ONEHOT_CHANNEL) + feature_spec.get_names_by_channel(MULTIHOT_CHANNEL) cont_names = feature_spec.get_names_by_channel(NUMERICAL_CHANNEL) if map_channel_enabled: cat_names += feature_spec.get_names_by_channel(MAP_FEATURE_CHANNEL) tf_dataset = KerasSequenceLoader( data_paths, batch_size=batch_size, label_names=label_names, cat_names=cat_names, cont_names=cont_names, engine="parquet", shuffle=shuffle, buffer_size=buffer_size, parts_per_chunk=parts_per_chunk, global_size=hvd.size(), global_rank=hvd.rank(), seed_fn=seed_fn, ) return tf_dataset def make_padding_function(multihot_hotness_dict): @tf.function(experimental_relax_shapes=True) def pad_batch(batch): batch = batch.copy() for feature, hotness in multihot_hotness_dict.items(): multihot_tuple = batch[feature] values = multihot_tuple[0][:, 0] row_lengths = multihot_tuple[1][:, 0] padded = tf.RaggedTensor.from_row_lengths( values, row_lengths, validate=False ).to_tensor(default_value=-1, shape=[None, hotness]) batch[feature] = padded return batch return pad_batch
TensorFlow2/Recommendation/WideAndDeep/triton/deployment_toolkit/triton_performance_runner
triton_performance_runner
runner
# Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # method from PEP-366 to support relative import in executed modules import logging import pathlib from typing import List, Optional if __package__ is None: __package__ = pathlib.Path(__file__).parent.name from ..core import EvaluationMode, MeasurementMode, OfflineMode, PerformanceTool from .model_analyzer import ModelAnalyzerRunner from .perf_analyzer import PerfAnalyzerRunner, PerfAnalyzerWarmupRunner LOGGER = logging.getLogger("triton_performance_runner") class TritonPerformanceRunner: def __init__( self, server_url: str, model_name: str, input_data: str, input_shapes: List[str], batch_sizes: List[int], concurrency: List[int], measurement_mode: MeasurementMode, measurement_interval: int, measurement_request_count: int, evaluation_mode: EvaluationMode, offline_mode: OfflineMode, output_shared_memory_size: int, performance_tool: PerformanceTool, model_repository: str, result_path: pathlib.Path, warmup: bool, timeout: Optional[int], verbose: bool, ): self._warmup_runner = None if warmup: LOGGER.info("Running warmup before the main test") self._warmup_runner = PerfAnalyzerWarmupRunner( server_url=server_url, model_name=model_name, input_data=input_data, input_shapes=input_shapes, batch_sizes=batch_sizes, concurrency=concurrency, measurement_mode=measurement_mode, measurement_interval=measurement_interval, measurement_request_count=measurement_request_count, evaluation_mode=evaluation_mode, offline_mode=offline_mode, output_shared_memory_size=output_shared_memory_size, timeout=timeout, ) if performance_tool == PerformanceTool.MODEL_ANALYZER: LOGGER.info("Using Model Analyzer for performance evaluation") self._runner = ModelAnalyzerRunner( server_url=server_url, model_name=model_name, input_data=input_data, input_shapes=input_shapes, batch_sizes=batch_sizes, concurrency=concurrency, measurement_mode=measurement_mode, measurement_interval=measurement_interval, measurement_request_count=measurement_request_count, evaluation_mode=evaluation_mode, offline_mode=offline_mode, output_shared_memory_size=output_shared_memory_size, model_repository=model_repository, result_path=result_path, timeout=timeout, verbose=verbose, ) elif performance_tool == PerformanceTool.PERF_ANALYZER: LOGGER.info("Using Perf Analyzer for performance evaluation") self._runner = PerfAnalyzerRunner( server_url=server_url, model_name=model_name, input_data=input_data, input_shapes=input_shapes, batch_sizes=batch_sizes, measurement_mode=measurement_mode, measurement_interval=measurement_interval, measurement_request_count=measurement_request_count, concurrency=concurrency, evaluation_mode=evaluation_mode, offline_mode=offline_mode, output_shared_memory_size=output_shared_memory_size, result_path=result_path, timeout=timeout, verbose=verbose, ) else: raise ValueError(f"Unsupported performance tool {performance_tool}") def run(self): if self._warmup_runner: self._warmup_runner.run() self._runner.run()
TensorFlow/Detection/SSD/models/research/slim/nets
nets
cyclegan_test
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Tests for tensorflow.contrib.slim.nets.cyclegan.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf from nets import cyclegan # TODO(joelshor): Add a test to check generator endpoints. class CycleganTest(tf.test.TestCase): def test_generator_inference(self): """Check one inference step.""" img_batch = tf.zeros([2, 32, 32, 3]) model_output, _ = cyclegan.cyclegan_generator_resnet(img_batch) with self.test_session() as sess: sess.run(tf.global_variables_initializer()) sess.run(model_output) def _test_generator_graph_helper(self, shape): """Check that generator can take small and non-square inputs.""" output_imgs, _ = cyclegan.cyclegan_generator_resnet(tf.ones(shape)) self.assertAllEqual(shape, output_imgs.shape.as_list()) def test_generator_graph_small(self): self._test_generator_graph_helper([4, 32, 32, 3]) def test_generator_graph_medium(self): self._test_generator_graph_helper([3, 128, 128, 3]) def test_generator_graph_nonsquare(self): self._test_generator_graph_helper([2, 80, 400, 3]) def test_generator_unknown_batch_dim(self): """Check that generator can take unknown batch dimension inputs.""" img = tf.placeholder(tf.float32, shape=[None, 32, None, 3]) output_imgs, _ = cyclegan.cyclegan_generator_resnet(img) self.assertAllEqual([None, 32, None, 3], output_imgs.shape.as_list()) def _input_and_output_same_shape_helper(self, kernel_size): img_batch = tf.placeholder(tf.float32, shape=[None, 32, 32, 3]) output_img_batch, _ = cyclegan.cyclegan_generator_resnet( img_batch, kernel_size=kernel_size) self.assertAllEqual(img_batch.shape.as_list(), output_img_batch.shape.as_list()) def input_and_output_same_shape_kernel3(self): self._input_and_output_same_shape_helper(3) def input_and_output_same_shape_kernel4(self): self._input_and_output_same_shape_helper(4) def input_and_output_same_shape_kernel5(self): self._input_and_output_same_shape_helper(5) def input_and_output_same_shape_kernel6(self): self._input_and_output_same_shape_helper(6) def _error_if_height_not_multiple_of_four_helper(self, height): self.assertRaisesRegexp( ValueError, 'The input height must be a multiple of 4.', cyclegan.cyclegan_generator_resnet, tf.placeholder(tf.float32, shape=[None, height, 32, 3])) def test_error_if_height_not_multiple_of_four_height29(self): self._error_if_height_not_multiple_of_four_helper(29) def test_error_if_height_not_multiple_of_four_height30(self): self._error_if_height_not_multiple_of_four_helper(30) def test_error_if_height_not_multiple_of_four_height31(self): self._error_if_height_not_multiple_of_four_helper(31) def _error_if_width_not_multiple_of_four_helper(self, width): self.assertRaisesRegexp( ValueError, 'The input width must be a multiple of 4.', cyclegan.cyclegan_generator_resnet, tf.placeholder(tf.float32, shape=[None, 32, width, 3])) def test_error_if_width_not_multiple_of_four_width29(self): self._error_if_width_not_multiple_of_four_helper(29) def test_error_if_width_not_multiple_of_four_width30(self): self._error_if_width_not_multiple_of_four_helper(30) def test_error_if_width_not_multiple_of_four_width31(self): self._error_if_width_not_multiple_of_four_helper(31) if __name__ == '__main__': tf.test.main()
PaddlePaddle/LanguageModeling/BERT/vocab
vocab
bert-large-uncased-vocab
[PAD] [unused0] [unused1] [unused2] [unused3] [unused4] [unused5] [unused6] [unused7] [unused8] [unused9] [unused10] [unused11] [unused12] [unused13] [unused14] [unused15] [unused16] [unused17] [unused18] [unused19] [unused20] [unused21] [unused22] [unused23] [unused24] [unused25] [unused26] [unused27] [unused28] [unused29] [unused30] [unused31] [unused32] [unused33] [unused34] [unused35] [unused36] [unused37] [unused38] [unused39] [unused40] [unused41] [unused42] [unused43] [unused44] [unused45] [unused46] [unused47] [unused48] [unused49] [unused50] [unused51] [unused52] [unused53] [unused54] [unused55] [unused56] [unused57] [unused58] [unused59] [unused60] [unused61] [unused62] [unused63] [unused64] [unused65] [unused66] [unused67] [unused68] [unused69] [unused70] [unused71] [unused72] [unused73] [unused74] [unused75] [unused76] [unused77] [unused78] [unused79] [unused80] [unused81] [unused82] [unused83] [unused84] [unused85] [unused86] [unused87] [unused88] [unused89] [unused90] [unused91] [unused92] [unused93] [unused94] [unused95] [unused96] [unused97] [unused98] [UNK] [CLS] [SEP] [MASK] [unused99] [unused100] [unused101] [unused102] [unused103] [unused104] [unused105] [unused106] [unused107] [unused108] [unused109] [unused110] [unused111] [unused112] [unused113] [unused114] [unused115] [unused116] [unused117] [unused118] [unused119] [unused120] [unused121] [unused122] [unused123] [unused124] [unused125] [unused126] [unused127] [unused128] [unused129] [unused130] [unused131] [unused132] [unused133] [unused134] [unused135] [unused136] [unused137] [unused138] [unused139] [unused140] [unused141] [unused142] [unused143] [unused144] [unused145] [unused146] [unused147] [unused148] [unused149] [unused150] [unused151] [unused152] [unused153] [unused154] [unused155] [unused156] [unused157] [unused158] [unused159] [unused160] [unused161] [unused162] [unused163] [unused164] [unused165] [unused166] [unused167] [unused168] [unused169] [unused170] [unused171] [unused172] [unused173] [unused174] [unused175] [unused176] [unused177] [unused178] [unused179] [unused180] [unused181] [unused182] [unused183] [unused184] [unused185] [unused186] [unused187] [unused188] [unused189] [unused190] [unused191] [unused192] [unused193] [unused194] [unused195] [unused196] [unused197] [unused198] [unused199] [unused200] [unused201] [unused202] [unused203] [unused204] [unused205] [unused206] [unused207] [unused208] [unused209] [unused210] [unused211] [unused212] [unused213] [unused214] [unused215] [unused216] [unused217] [unused218] [unused219] [unused220] [unused221] [unused222] [unused223] [unused224] [unused225] [unused226] [unused227] [unused228] [unused229] [unused230] [unused231] [unused232] [unused233] [unused234] [unused235] [unused236] [unused237] [unused238] [unused239] [unused240] [unused241] [unused242] [unused243] [unused244] [unused245] [unused246] [unused247] [unused248] [unused249] [unused250] [unused251] [unused252] [unused253] [unused254] [unused255] [unused256] [unused257] [unused258] [unused259] [unused260] [unused261] [unused262] [unused263] [unused264] [unused265] [unused266] [unused267] [unused268] [unused269] [unused270] [unused271] [unused272] [unused273] [unused274] [unused275] [unused276] [unused277] [unused278] [unused279] [unused280] [unused281] [unused282] [unused283] [unused284] [unused285] [unused286] [unused287] [unused288] [unused289] [unused290] [unused291] [unused292] [unused293] [unused294] [unused295] [unused296] [unused297] [unused298] [unused299] [unused300] [unused301] [unused302] [unused303] [unused304] [unused305] [unused306] [unused307] [unused308] [unused309] [unused310] [unused311] [unused312] [unused313] [unused314] [unused315] [unused316] [unused317] [unused318] [unused319] [unused320] [unused321] [unused322] [unused323] [unused324] [unused325] [unused326] [unused327] [unused328] [unused329] [unused330] [unused331] [unused332] [unused333] [unused334] [unused335] [unused336] [unused337] [unused338] [unused339] [unused340] [unused341] [unused342] [unused343] [unused344] [unused345] [unused346] [unused347] [unused348] [unused349] [unused350] [unused351] [unused352] [unused353] [unused354] [unused355] [unused356] [unused357] [unused358] [unused359] [unused360] [unused361] [unused362] [unused363] [unused364] [unused365] [unused366] [unused367] [unused368] [unused369] [unused370] [unused371] [unused372] [unused373] [unused374] [unused375] [unused376] [unused377] [unused378] [unused379] [unused380] [unused381] [unused382] [unused383] [unused384] [unused385] [unused386] [unused387] [unused388] [unused389] [unused390] [unused391] [unused392] [unused393] [unused394] [unused395] [unused396] [unused397] [unused398] [unused399] [unused400] [unused401] [unused402] [unused403] [unused404] [unused405] [unused406] [unused407] [unused408] [unused409] [unused410] [unused411] [unused412] [unused413] [unused414] [unused415] [unused416] [unused417] [unused418] [unused419] [unused420] [unused421] [unused422] [unused423] [unused424] [unused425] [unused426] [unused427] [unused428] [unused429] [unused430] [unused431] [unused432] [unused433] [unused434] [unused435] [unused436] [unused437] [unused438] [unused439] [unused440] [unused441] [unused442] [unused443] [unused444] [unused445] [unused446] [unused447] [unused448] [unused449] [unused450] [unused451] [unused452] [unused453] [unused454] [unused455] [unused456] [unused457] [unused458] [unused459] [unused460] [unused461] [unused462] [unused463] [unused464] [unused465] [unused466] [unused467] [unused468] [unused469] [unused470] [unused471] [unused472] [unused473] [unused474] [unused475] [unused476] [unused477] [unused478] [unused479] [unused480] [unused481] [unused482] [unused483] [unused484] [unused485] [unused486] [unused487] [unused488] [unused489] [unused490] [unused491] [unused492] [unused493] [unused494] [unused495] [unused496] [unused497] [unused498] [unused499] [unused500] [unused501] [unused502] [unused503] [unused504] [unused505] [unused506] [unused507] [unused508] [unused509] [unused510] [unused511] [unused512] [unused513] [unused514] [unused515] [unused516] [unused517] [unused518] [unused519] [unused520] [unused521] [unused522] [unused523] [unused524] [unused525] [unused526] [unused527] [unused528] [unused529] [unused530] [unused531] [unused532] [unused533] [unused534] [unused535] [unused536] [unused537] [unused538] [unused539] [unused540] [unused541] [unused542] [unused543] [unused544] [unused545] [unused546] [unused547] [unused548] [unused549] [unused550] [unused551] [unused552] [unused553] [unused554] [unused555] [unused556] [unused557] [unused558] [unused559] [unused560] [unused561] [unused562] [unused563] [unused564] [unused565] [unused566] [unused567] [unused568] [unused569] [unused570] [unused571] [unused572] [unused573] [unused574] [unused575] [unused576] [unused577] [unused578] [unused579] [unused580] [unused581] [unused582] [unused583] [unused584] [unused585] [unused586] [unused587] [unused588] [unused589] [unused590] [unused591] [unused592] [unused593] [unused594] [unused595] [unused596] [unused597] [unused598] [unused599] [unused600] [unused601] [unused602] [unused603] [unused604] [unused605] [unused606] [unused607] [unused608] [unused609] [unused610] [unused611] [unused612] [unused613] [unused614] [unused615] [unused616] [unused617] [unused618] [unused619] [unused620] [unused621] [unused622] [unused623] [unused624] [unused625] [unused626] [unused627] [unused628] [unused629] [unused630] [unused631] [unused632] [unused633] [unused634] [unused635] [unused636] [unused637] [unused638] [unused639] [unused640] [unused641] [unused642] [unused643] [unused644] [unused645] [unused646] [unused647] [unused648] [unused649] [unused650] [unused651] [unused652] [unused653] [unused654] [unused655] [unused656] [unused657] [unused658] [unused659] [unused660] [unused661] [unused662] [unused663] [unused664] [unused665] [unused666] [unused667] [unused668] [unused669] [unused670] [unused671] [unused672] [unused673] [unused674] [unused675] [unused676] [unused677] [unused678] [unused679] [unused680] [unused681] [unused682] [unused683] [unused684] [unused685] [unused686] [unused687] [unused688] [unused689] [unused690] [unused691] [unused692] [unused693] [unused694] [unused695] [unused696] [unused697] [unused698] [unused699] [unused700] [unused701] [unused702] [unused703] [unused704] [unused705] [unused706] [unused707] [unused708] [unused709] [unused710] [unused711] [unused712] [unused713] [unused714] [unused715] [unused716] [unused717] [unused718] [unused719] [unused720] [unused721] [unused722] [unused723] [unused724] [unused725] [unused726] [unused727] [unused728] [unused729] [unused730] [unused731] [unused732] [unused733] [unused734] [unused735] [unused736] [unused737] [unused738] [unused739] [unused740] [unused741] [unused742] [unused743] [unused744] [unused745] [unused746] [unused747] [unused748] [unused749] [unused750] [unused751] [unused752] [unused753] [unused754] [unused755] [unused756] [unused757] [unused758] [unused759] [unused760] [unused761] [unused762] [unused763] [unused764] [unused765] [unused766] [unused767] [unused768] [unused769] [unused770] [unused771] [unused772] [unused773] [unused774] [unused775] [unused776] [unused777] [unused778] [unused779] [unused780] [unused781] [unused782] [unused783] [unused784] [unused785] [unused786] [unused787] [unused788] [unused789] [unused790] [unused791] [unused792] [unused793] [unused794] [unused795] [unused796] [unused797] [unused798] [unused799] [unused800] [unused801] [unused802] [unused803] [unused804] [unused805] [unused806] [unused807] [unused808] [unused809] [unused810] [unused811] [unused812] [unused813] [unused814] [unused815] [unused816] [unused817] [unused818] [unused819] [unused820] [unused821] [unused822] [unused823] [unused824] [unused825] [unused826] [unused827] [unused828] [unused829] [unused830] [unused831] [unused832] [unused833] [unused834] [unused835] [unused836] [unused837] [unused838] [unused839] [unused840] [unused841] [unused842] [unused843] [unused844] [unused845] [unused846] [unused847] [unused848] [unused849] [unused850] [unused851] [unused852] [unused853] [unused854] [unused855] [unused856] [unused857] [unused858] [unused859] [unused860] [unused861] [unused862] [unused863] [unused864] [unused865] [unused866] [unused867] [unused868] [unused869] [unused870] [unused871] [unused872] [unused873] [unused874] [unused875] [unused876] [unused877] [unused878] [unused879] [unused880] [unused881] [unused882] [unused883] [unused884] [unused885] [unused886] [unused887] [unused888] [unused889] [unused890] [unused891] [unused892] [unused893] [unused894] [unused895] [unused896] [unused897] [unused898] [unused899] [unused900] [unused901] [unused902] [unused903] [unused904] [unused905] [unused906] [unused907] [unused908] [unused909] [unused910] [unused911] [unused912] [unused913] [unused914] [unused915] [unused916] [unused917] [unused918] [unused919] [unused920] [unused921] [unused922] [unused923] [unused924] [unused925] [unused926] [unused927] [unused928] [unused929] [unused930] [unused931] [unused932] [unused933] [unused934] [unused935] [unused936] [unused937] [unused938] [unused939] [unused940] [unused941] [unused942] [unused943] [unused944] [unused945] [unused946] [unused947] [unused948] [unused949] [unused950] [unused951] [unused952] [unused953] [unused954] [unused955] [unused956] [unused957] [unused958] [unused959] [unused960] [unused961] [unused962] [unused963] [unused964] [unused965] [unused966] [unused967] [unused968] [unused969] [unused970] [unused971] [unused972] [unused973] [unused974] [unused975] [unused976] [unused977] [unused978] [unused979] [unused980] [unused981] [unused982] [unused983] [unused984] [unused985] [unused986] [unused987] [unused988] [unused989] [unused990] [unused991] [unused992] [unused993] ! " # $ % & ' ( ) * + , - . / 0 1 2 3 4 5 6 7 8 9 : ; < = > ? @ [ \ ] ^ _ ` a b c d e f g h i j k l m n o p q r s t u v w x y z { | } ~ ¡ ¢ £ ¤ ¥ ¦ § ¨ © ª « ¬ ® ° ± ² ³ ´ µ ¶ · ¹ º » ¼ ½ ¾ ¿ × ß æ ð ÷ ø þ đ ħ ı ł ŋ œ ƒ ɐ ɑ ɒ ɔ ɕ ə ɛ ɡ ɣ ɨ ɪ ɫ ɬ ɯ ɲ ɴ ɹ ɾ ʀ ʁ ʂ ʃ ʉ ʊ ʋ ʌ ʎ ʐ ʑ ʒ ʔ ʰ ʲ ʳ ʷ ʸ ʻ ʼ ʾ ʿ ˈ ː ˡ ˢ ˣ ˤ α β γ δ ε ζ η θ ι κ λ μ ν ξ ο π ρ ς σ τ υ φ χ ψ ω а б в г д е ж з и к л м н о п р с т у ф х ц ч ш щ ъ ы ь э ю я ђ є і ј љ њ ћ ӏ ա բ գ դ ե թ ի լ կ հ մ յ ն ո պ ս վ տ ր ւ ք ־ א ב ג ד ה ו ז ח ט י ך כ ל ם מ ן נ ס ע ף פ ץ צ ק ר ש ת ، ء ا ب ة ت ث ج ح خ د ذ ر ز س ش ص ض ط ظ ع غ ـ ف ق ك ل م ن ه و ى ي ٹ پ چ ک گ ں ھ ہ ی ے अ आ उ ए क ख ग च ज ट ड ण त थ द ध न प ब भ म य र ल व श ष स ह ा ि ी ो । ॥ ং অ আ ই উ এ ও ক খ গ চ ছ জ ট ড ণ ত থ দ ধ ন প ব ভ ম য র ল শ ষ স হ া ি ী ে க ச ட த ந ன ப ம ய ர ல ள வ ா ி ு ே ை ನ ರ ಾ ක ය ර ල ව ා ก ง ต ท น พ ม ย ร ล ว ส อ า เ ་ ། ག ང ད ན པ བ མ འ ར ལ ས မ ა ბ გ დ ე ვ თ ი კ ლ მ ნ ო რ ს ტ უ ᄀ ᄂ ᄃ ᄅ ᄆ ᄇ ᄉ ᄊ ᄋ ᄌ ᄎ ᄏ ᄐ ᄑ ᄒ ᅡ ᅢ ᅥ ᅦ ᅧ ᅩ ᅪ ᅭ ᅮ ᅯ ᅲ ᅳ ᅴ ᅵ ᆨ ᆫ ᆯ ᆷ ᆸ ᆼ ᴬ ᴮ ᴰ ᴵ ᴺ ᵀ ᵃ ᵇ ᵈ ᵉ ᵍ ᵏ ᵐ ᵒ ᵖ ᵗ ᵘ ᵢ ᵣ ᵤ ᵥ ᶜ ᶠ ‐ ‑ ‒ – — ― ‖ ‘ ’ ‚ “ ” „ † ‡ • … ‰ ′ ″ › ‿ ⁄ ⁰ ⁱ ⁴ ⁵ ⁶ ⁷ ⁸ ⁹ ⁺ ⁻ ⁿ ₀ ₁ ₂ ₃ ₄ ₅ ₆ ₇ ₈ ₉ ₊ ₍ ₎ ₐ ₑ ₒ ₓ ₕ ₖ ₗ ₘ ₙ ₚ ₛ ₜ ₤ ₩ € ₱ ₹ ℓ № ℝ ™ ⅓ ⅔ ← ↑ → ↓ ↔ ↦ ⇄ ⇌ ⇒ ∂ ∅ ∆ ∇ ∈ − ∗ ∘ √ ∞ ∧ ∨ ∩ ∪ ≈ ≡ ≤ ≥ ⊂ ⊆ ⊕ ⊗ ⋅ ─ │ ■ ▪ ● ★ ☆ ☉ ♠ ♣ ♥ ♦ ♭ ♯ ⟨ ⟩ ⱼ ⺩ ⺼ ⽥ 、 。 〈 〉 《 》 「 」 『 』 〜 あ い う え お か き く け こ さ し す せ そ た ち っ つ て と な に ぬ ね の は ひ ふ へ ほ ま み む め も や ゆ よ ら り る れ ろ を ん ァ ア ィ イ ウ ェ エ オ カ キ ク ケ コ サ シ ス セ タ チ ッ ツ テ ト ナ ニ ノ ハ ヒ フ ヘ ホ マ ミ ム メ モ ャ ュ ョ ラ リ ル レ ロ ワ ン ・ ー 一 三 上 下 不 世 中 主 久 之 也 事 二 五 井 京 人 亻 仁 介 代 仮 伊 会 佐 侍 保 信 健 元 光 八 公 内 出 分 前 劉 力 加 勝 北 区 十 千 南 博 原 口 古 史 司 合 吉 同 名 和 囗 四 国 國 土 地 坂 城 堂 場 士 夏 外 大 天 太 夫 奈 女 子 学 宀 宇 安 宗 定 宣 宮 家 宿 寺 將 小 尚 山 岡 島 崎 川 州 巿 帝 平 年 幸 广 弘 張 彳 後 御 德 心 忄 志 忠 愛 成 我 戦 戸 手 扌 政 文 新 方 日 明 星 春 昭 智 曲 書 月 有 朝 木 本 李 村 東 松 林 森 楊 樹 橋 歌 止 正 武 比 氏 民 水 氵 氷 永 江 沢 河 治 法 海 清 漢 瀬 火 版 犬 王 生 田 男 疒 発 白 的 皇 目 相 省 真 石 示 社 神 福 禾 秀 秋 空 立 章 竹 糹 美 義 耳 良 艹 花 英 華 葉 藤 行 街 西 見 訁 語 谷 貝 貴 車 軍 辶 道 郎 郡 部 都 里 野 金 鈴 镇 長 門 間 阝 阿 陳 陽 雄 青 面 風 食 香 馬 高 龍 龸 fi fl ! ( ) , - . / : ? ~ the of and in to was he is as for on with that it his by at from her ##s she you had an were but be this are not my they one which or have him me first all also their has up who out been when after there into new two its ##a time would no what about said we over then other so more ##e can if like back them only some could ##i where just ##ing during before ##n do ##o made school through than now years most world may between down well three ##d year while will ##ed ##r ##y later ##t city under around did such being used state people part know against your many second university both national ##er these don known off way until re how even get head ... didn ##ly team american because de ##l born united film since still long work south us became any high again day family see right man eyes house season war states including took life north same each called name much place however go four group another found won area here going 10 away series left home music best make hand number company several never last john 000 very album take end good too following released game played little began district ##m old want those side held own early county ll league use west ##u face think ##es 2010 government ##h march came small general town june ##on line based something ##k september thought looked along international 2011 air july club went january october our august april york 12 few 2012 2008 east show member college 2009 father public ##us come men five set station church ##c next former november room party located december 2013 age got 2007 ##g system let love 2006 though every 2014 look song water century without body black night within great women single ve building large population river named band white started ##an once 15 20 should 18 2015 service top built british open death king moved local times children february book why 11 door need president order final road wasn although due major died village third knew 2016 asked turned st wanted say ##p together received main son served different ##en behind himself felt members power football law voice play ##in near park history 30 having 2005 16 ##man saw mother ##al army point front help english street art late hands games award ##ia young 14 put published country division across told 13 often ever french london center six red 2017 led days include light 25 find tell among species really according central half 2004 form original gave office making enough lost full opened must included live given german player run business woman community cup might million land 2000 court development 17 short round ii km seen class story always become sure research almost director council la ##2 career things using island ##z couldn car ##is 24 close force ##1 better free support control field students 2003 education married ##b nothing worked others record big inside level anything continued give james ##3 military established non returned feel does title written thing feet william far co association hard already 2002 ##ra championship human western 100 ##na department hall role various production 21 19 heart 2001 living fire version ##ers ##f television royal ##4 produced working act case society region present radio period looking least total keep england wife program per brother mind special 22 ##le am works soon ##6 political george services taken created ##7 further able reached david union joined upon done important social information either ##ic ##x appeared position ground lead rock dark election 23 board france hair course arms site police girl instead real sound ##v words moment ##te someone ##8 summer project announced san less wrote past followed ##5 blue founded al finally india taking records america ##ne 1999 design considered northern god stop battle toward european outside described track today playing language 28 call 26 heard professional low australia miles california win yet green ##ie trying blood ##ton southern science maybe everything match square 27 mouth video race recorded leave above ##9 daughter points space 1998 museum change middle common ##0 move tv post ##ta lake seven tried elected closed ten paul minister ##th months start chief return canada person sea release similar modern brought rest hit formed mr ##la 1997 floor event doing thomas 1996 robert care killed training star week needed turn finished railway rather news health sent example ran term michael coming currently yes forces despite gold areas 50 stage fact 29 dead says popular 2018 originally germany probably developed result pulled friend stood money running mi signed word songs child eventually met tour average teams minutes festival current deep kind 1995 decided usually eastern seemed ##ness episode bed added table indian private charles route available idea throughout centre addition appointed style 1994 books eight construction press mean wall friends remained schools study ##ch ##um institute oh chinese sometimes events possible 1992 australian type brown forward talk process food debut seat performance committee features character arts herself else lot strong russian range hours peter arm ##da morning dr sold ##ry quickly directed 1993 guitar china ##w 31 list ##ma performed media uk players smile ##rs myself 40 placed coach province towards wouldn leading whole boy official designed grand census ##el europe attack japanese henry 1991 ##re ##os cross getting alone action lower network wide washington japan 1990 hospital believe changed sister ##ar hold gone sir hadn ship ##ka studies academy shot rights below base bad involved kept largest ##ist bank future especially beginning mark movement section female magazine plan professor lord longer ##ian sat walked hill actually civil energy model families size thus aircraft completed includes data captain ##or fight vocals featured richard bridge fourth 1989 officer stone hear ##ism means medical groups management self lips competition entire lived technology leaving federal tournament bit passed hot independent awards kingdom mary spent fine doesn reported ##ling jack fall raised itself stay true studio 1988 sports replaced paris systems saint leader theatre whose market capital parents spanish canadian earth ##ity cut degree writing bay christian awarded natural higher bill ##as coast provided previous senior ft valley organization stopped onto countries parts conference queen security interest saying allowed master earlier phone matter smith winning try happened moving campaign los ##ley breath nearly mid 1987 certain girls date italian african standing fell artist ##ted shows deal mine industry 1986 ##ng everyone republic provide collection library student ##ville primary owned older via heavy 1st makes ##able attention anyone africa ##ri stated length ended fingers command staff skin foreign opening governor okay medal kill sun cover job 1985 introduced chest hell feeling ##ies success meet reason standard meeting novel 1984 trade source buildings ##land rose guy goal ##ur chapter native husband previously unit limited entered weeks producer operations mountain takes covered forced related roman complete successful key texas cold ##ya channel 1980 traditional films dance clear approximately 500 nine van prince question active tracks ireland regional silver author personal sense operation ##ine economic 1983 holding twenty isbn additional speed hour edition regular historic places whom shook movie km² secretary prior report chicago read foundation view engine scored 1982 units ask airport property ready immediately lady month listed contract ##de manager themselves lines ##ki navy writer meant ##ts runs ##ro practice championships singer glass commission required forest starting culture generally giving access attended test couple stand catholic martin caught executive ##less eye ##ey thinking chair quite shoulder 1979 hope decision plays defeated municipality whether structure offered slowly pain ice direction ##ion paper mission 1981 mostly 200 noted individual managed nature lives plant ##ha helped except studied computer figure relationship issue significant loss die smiled gun ago highest 1972 ##am male bring goals mexico problem distance commercial completely location annual famous drive 1976 neck 1978 surface caused italy understand greek highway wrong hotel comes appearance joseph double issues musical companies castle income review assembly bass initially parliament artists experience 1974 particular walk foot engineering talking window dropped ##ter miss baby boys break 1975 stars edge remember policy carried train stadium bar sex angeles evidence ##ge becoming assistant soviet 1977 upper step wing 1970 youth financial reach ##ll actor numerous ##se ##st nodded arrived ##ation minute ##nt believed sorry complex beautiful victory associated temple 1968 1973 chance perhaps metal ##son 1945 bishop ##et lee launched particularly tree le retired subject prize contains yeah theory empire ##ce suddenly waiting trust recording ##to happy terms camp champion 1971 religious pass zealand names 2nd port ancient tom corner represented watch legal anti justice cause watched brothers 45 material changes simply response louis fast ##ting answer 60 historical 1969 stories straight create feature increased rate administration virginia el activities cultural overall winner programs basketball legs guard beyond cast doctor mm flight results remains cost effect winter ##ble larger islands problems chairman grew commander isn 1967 pay failed selected hurt fort box regiment majority journal 35 edward plans ##ke ##ni shown pretty irish characters directly scene likely operated allow spring ##j junior matches looks mike houses fellow ##tion beach marriage ##ham ##ive rules oil 65 florida expected nearby congress sam peace recent iii wait subsequently cell ##do variety serving agreed please poor joe pacific attempt wood democratic piece prime ##ca rural mile touch appears township 1964 1966 soldiers ##men ##ized 1965 pennsylvania closer fighting claimed score jones physical editor ##ous filled genus specific sitting super mom ##va therefore supported status fear cases store meaning wales minor spain tower focus vice frank follow parish separate golden horse fifth remaining branch 32 presented stared ##id uses secret forms ##co baseball exactly ##ck choice note discovered travel composed truth russia ball color kiss dad wind continue ring referred numbers digital greater ##ns metres slightly direct increase 1960 responsible crew rule trees troops ##no broke goes individuals hundred weight creek sleep memory defense provides ordered code value jewish windows 1944 safe judge whatever corps realized growing pre ##ga cities alexander gaze lies spread scott letter showed situation mayor transport watching workers extended ##li expression normal ##ment chart multiple border ##ba host ##ner daily mrs walls piano ##ko heat cannot ##ate earned products drama era authority seasons join grade ##io sign difficult machine 1963 territory mainly ##wood stations squadron 1962 stepped iron 19th ##led serve appear sky speak broken charge knowledge kilometres removed ships article campus simple ##ty pushed britain ##ve leaves recently cd soft boston latter easy acquired poland ##sa quality officers presence planned nations mass broadcast jean share image influence wild offer emperor electric reading headed ability promoted yellow ministry 1942 throat smaller politician ##by latin spoke cars williams males lack pop 80 ##ier acting seeing consists ##ti estate 1961 pressure johnson newspaper jr chris olympics online conditions beat elements walking vote ##field needs carolina text featuring global block shirt levels francisco purpose females et dutch duke ahead gas twice safety serious turning highly lieutenant firm maria amount mixed daniel proposed perfect agreement affairs 3rd seconds contemporary paid 1943 prison save kitchen label administrative intended constructed academic nice teacher races 1956 formerly corporation ben nation issued shut 1958 drums housing victoria seems opera 1959 graduated function von mentioned picked build recognized shortly protection picture notable exchange elections 1980s loved percent racing fish elizabeth garden volume hockey 1941 beside settled ##ford 1940 competed replied drew 1948 actress marine scotland steel glanced farm steve 1957 risk tonight positive magic singles effects gray screen dog ##ja residents bus sides none secondary literature polish destroyed flying founder households 1939 lay reserve usa gallery ##ler 1946 industrial younger approach appearances urban ones 1950 finish avenue powerful fully growth page honor jersey projects advanced revealed basic 90 infantry pair equipment visit 33 evening search grant effort solo treatment buried republican primarily bottom owner 1970s israel gives jim dream bob remain spot 70 notes produce champions contact ed soul accepted ways del ##ally losing split price capacity basis trial questions ##ina 1955 20th guess officially memorial naval initial ##ization whispered median engineer ##ful sydney ##go columbia strength 300 1952 tears senate 00 card asian agent 1947 software 44 draw warm supposed com pro ##il transferred leaned ##at candidate escape mountains asia potential activity entertainment seem traffic jackson murder 36 slow product orchestra haven agency bbc taught website comedy unable storm planning albums rugby environment scientific grabbed protect ##hi boat typically 1954 1953 damage principal divided dedicated mount ohio ##berg pick fought driver ##der empty shoulders sort thank berlin prominent account freedom necessary efforts alex headquarters follows alongside des simon andrew suggested operating learning steps 1949 sweet technical begin easily 34 teeth speaking settlement scale ##sh renamed ray max enemy semi joint compared ##rd scottish leadership analysis offers georgia pieces captured animal deputy guest organized ##lin tony combined method challenge 1960s huge wants battalion sons rise crime types facilities telling path 1951 platform sit 1990s ##lo tells assigned rich pull ##ot commonly alive ##za letters concept conducted wearing happen bought becomes holy gets ocean defeat languages purchased coffee occurred titled ##q declared applied sciences concert sounds jazz brain ##me painting fleet tax nick ##ius michigan count animals leaders episodes ##line content ##den birth ##it clubs 64 palace critical refused fair leg laughed returning surrounding participated formation lifted pointed connected rome medicine laid taylor santa powers adam tall shared focused knowing yards entrance falls ##wa calling ##ad sources chosen beneath resources yard ##ite nominated silence zone defined ##que gained thirty 38 bodies moon ##ard adopted christmas widely register apart iran premier serves du unknown parties ##les generation ##ff continues quick fields brigade quiet teaching clothes impact weapons partner flat theater supreme 1938 37 relations ##tor plants suffered 1936 wilson kids begins ##age 1918 seats armed internet models worth laws 400 communities classes background knows thanks quarter reaching humans carry killing format kong hong setting 75 architecture disease railroad inc possibly wish arthur thoughts harry doors density ##di crowd illinois stomach tone unique reports anyway ##ir liberal der vehicle thick dry drug faced largely facility theme holds creation strange colonel ##mi revolution bell politics turns silent rail relief independence combat shape write determined sales learned 4th finger oxford providing 1937 heritage fiction situated designated allowing distribution hosted ##est sight interview estimated reduced ##ria toronto footballer keeping guys damn claim motion sport sixth stayed ##ze en rear receive handed twelve dress audience granted brazil ##well spirit ##ated noticed etc olympic representative eric tight trouble reviews drink vampire missing roles ranked newly household finals wave critics ##ee phase massachusetts pilot unlike philadelphia bright guns crown organizations roof 42 respectively clearly tongue marked circle fox korea bronze brian expanded sexual supply yourself inspired labour fc ##ah reference vision draft connection brand reasons 1935 classic driving trip jesus cells entry 1920 neither trail claims atlantic orders labor nose afraid identified intelligence calls cancer attacked passing stephen positions imperial grey jason 39 sunday 48 swedish avoid extra uncle message covers allows surprise materials fame hunter ##ji 1930 citizens figures davis environmental confirmed shit titles di performing difference acts attacks ##ov existing votes opportunity nor shop entirely trains opposite pakistan ##pa develop resulted representatives actions reality pressed ##ish barely wine conversation faculty northwest ends documentary nuclear stock grace sets eat alternative ##ps bag resulting creating surprised cemetery 1919 drop finding sarah cricket streets tradition ride 1933 exhibition target ear explained rain composer injury apartment municipal educational occupied netherlands clean billion constitution learn 1914 maximum classical francis lose opposition jose ontario bear core hills rolled ending drawn permanent fun ##tes ##lla lewis sites chamber ryan ##way scoring height 1934 ##house lyrics staring 55 officials 1917 snow oldest ##tic orange ##ger qualified interior apparently succeeded thousand dinner lights existence fans heavily 41 greatest conservative send bowl plus enter catch ##un economy duty 1929 speech authorities princess performances versions shall graduate pictures effective remembered poetry desk crossed starring starts passenger sharp ##ant acres ass weather falling rank fund supporting check adult publishing heads cm southeast lane ##burg application bc ##ura les condition transfer prevent display ex regions earl federation cool relatively answered besides 1928 obtained portion ##town mix ##ding reaction liked dean express peak 1932 ##tte counter religion chain rare miller convention aid lie vehicles mobile perform squad wonder lying crazy sword ##ping attempted centuries weren philosophy category ##ize anna interested 47 sweden wolf frequently abandoned kg literary alliance task entitled ##ay threw promotion factory tiny soccer visited matt fm achieved 52 defence internal persian 43 methods ##ging arrested otherwise cambridge programming villages elementary districts rooms criminal conflict worry trained 1931 attempts waited signal bird truck subsequent programme ##ol ad 49 communist details faith sector patrick carrying laugh ##ss controlled korean showing origin fuel evil 1927 ##ent brief identity darkness address pool missed publication web planet ian anne wings invited ##tt briefly standards kissed ##be ideas climate causing walter worse albert articles winners desire aged northeast dangerous gate doubt 1922 wooden multi ##ky poet rising funding 46 communications communication violence copies prepared ford investigation skills 1924 pulling electronic ##ak ##ial ##han containing ultimately offices singing understanding restaurant tomorrow fashion christ ward da pope stands 5th flow studios aired commissioned contained exist fresh americans ##per wrestling approved kid employed respect suit 1925 angel asking increasing frame angry selling 1950s thin finds ##nd temperature statement ali explain inhabitants towns extensive narrow 51 jane flowers images promise somewhere object fly closely ##ls 1912 bureau cape 1926 weekly presidential legislative 1921 ##ai ##au launch founding ##ny 978 ##ring artillery strike un institutions roll writers landing chose kevin anymore pp ##ut attorney fit dan billboard receiving agricultural breaking sought dave admitted lands mexican ##bury charlie specifically hole iv howard credit moscow roads accident 1923 proved wear struck hey guards stuff slid expansion 1915 cat anthony ##kin melbourne opposed sub southwest architect failure plane 1916 ##ron map camera tank listen regarding wet introduction metropolitan link ep fighter inch grown gene anger fixed buy dvd khan domestic worldwide chapel mill functions examples ##head developing 1910 turkey hits pocket antonio papers grow unless circuit 18th concerned attached journalist selection journey converted provincial painted hearing aren bands negative aside wondered knight lap survey ma ##ow noise billy ##ium shooting guide bedroom priest resistance motor homes sounded giant ##mer 150 scenes equal comic patients hidden solid actual bringing afternoon touched funds wedding consisted marie canal sr kim treaty turkish recognition residence cathedral broad knees incident shaped fired norwegian handle cheek contest represent ##pe representing beauty ##sen birds advantage emergency wrapped drawing notice pink broadcasting ##ong somehow bachelor seventh collected registered establishment alan assumed chemical personnel roger retirement jeff portuguese wore tied device threat progress advance ##ised banks hired manchester nfl teachers structures forever ##bo tennis helping saturday sale applications junction hip incorporated neighborhood dressed ceremony ##ds influenced hers visual stairs decades inner kansas hung hoped gain scheduled downtown engaged austria clock norway certainly pale protected 1913 victor employees plate putting surrounded ##ists finishing blues tropical ##ries minnesota consider philippines accept 54 retrieved 1900 concern anderson properties institution gordon successfully vietnam ##dy backing outstanding muslim crossing folk producing usual demand occurs observed lawyer educated ##ana kelly string pleasure budget items quietly colorado philip typical ##worth derived 600 survived asks mental ##ide 56 jake jews distinguished ltd 1911 sri extremely 53 athletic loud thousands worried shadow transportation horses weapon arena importance users tim objects contributed dragon douglas aware senator johnny jordan sisters engines flag investment samuel shock capable clark row wheel refers session familiar biggest wins hate maintained drove hamilton request expressed injured underground churches walker wars tunnel passes stupid agriculture softly cabinet regarded joining indiana ##ea ##ms push dates spend behavior woods protein gently chase morgan mention burning wake combination occur mirror leads jimmy indeed impossible singapore paintings covering ##nes soldier locations attendance sell historian wisconsin invasion argued painter diego changing egypt ##don experienced inches ##ku missouri vol grounds spoken switzerland ##gan reform rolling ha forget massive resigned burned allen tennessee locked values improved ##mo wounded universe sick dating facing pack purchase user ##pur moments ##ul merged anniversary 1908 coal brick understood causes dynasty queensland establish stores crisis promote hoping views cards referee extension ##si raise arizona improve colonial formal charged ##rt palm lucky hide rescue faces 95 feelings candidates juan ##ell goods 6th courses weekend 59 luke cash fallen ##om delivered affected installed carefully tries swiss hollywood costs lincoln responsibility ##he shore file proper normally maryland assistance jump constant offering friendly waters persons realize contain trophy 800 partnership factor 58 musicians cry bound oregon indicated hero houston medium ##ure consisting somewhat ##ara 57 cycle ##che beer moore frederick gotten eleven worst weak approached arranged chin loan universal bond fifteen pattern disappeared ##ney translated ##zed lip arab capture interests insurance ##chi shifted cave prix warning sections courts coat plot smell feed golf favorite maintain knife vs voted degrees finance quebec opinion translation manner ruled operate productions choose musician discovery confused tired separated stream techniques committed attend ranking kings throw passengers measure horror fan mining sand danger salt calm decade dam require runner ##ik rush associate greece ##ker rivers consecutive matthew ##ski sighed sq documents steam edited closing tie accused 1905 ##ini islamic distributed directors organisation bruce 7th breathing mad lit arrival concrete taste 08 composition shaking faster amateur adjacent stating 1906 twin flew ##ran tokyo publications ##tone obviously ridge storage 1907 carl pages concluded desert driven universities ages terminal sequence borough 250 constituency creative cousin economics dreams margaret notably reduce montreal mode 17th ears saved jan vocal ##ica 1909 andy ##jo riding roughly threatened ##ise meters meanwhile landed compete repeated grass czech regularly charges tea sudden appeal ##ung solution describes pierre classification glad parking ##ning belt physics 99 rachel add hungarian participate expedition damaged gift childhood 85 fifty ##red mathematics jumped letting defensive mph ##ux ##gh testing ##hip hundreds shoot owners matters smoke israeli kentucky dancing mounted grandfather emma designs profit argentina ##gs truly li lawrence cole begun detroit willing branches smiling decide miami enjoyed recordings ##dale poverty ethnic gay ##bi gary arabic 09 accompanied ##one ##ons fishing determine residential acid ##ary alice returns starred mail ##ang jonathan strategy ##ue net forty cook businesses equivalent commonwealth distinct ill ##cy seriously ##ors ##ped shift harris replace rio imagine formula ensure ##ber additionally scheme conservation occasionally purposes feels favor ##and ##ore 1930s contrast hanging hunt movies 1904 instruments victims danish christopher busy demon sugar earliest colony studying balance duties ##ks belgium slipped carter 05 visible stages iraq fifa ##im commune forming zero 07 continuing talked counties legend bathroom option tail clay daughters afterwards severe jaw visitors ##ded devices aviation russell kate ##vi entering subjects ##ino temporary swimming forth smooth ghost audio bush operates rocks movements signs eddie ##tz ann voices honorary 06 memories dallas pure measures racial promised 66 harvard ceo 16th parliamentary indicate benefit flesh dublin louisiana 1902 1901 patient sleeping 1903 membership coastal medieval wanting element scholars rice 62 limit survive makeup rating definitely collaboration obvious ##tan boss ms baron birthday linked soil diocese ##lan ncaa ##mann offensive shell shouldn waist ##tus plain ross organ resolution manufacturing adding relative kennedy 98 whilst moth marketing gardens crash 72 heading partners credited carlos moves cable ##zi marshall ##out depending bottle represents rejected responded existed 04 jobs denmark lock ##ating treated graham routes talent commissioner drugs secure tests reign restored photography ##gi contributions oklahoma designer disc grin seattle robin paused atlanta unusual ##gate praised las laughing satellite hungary visiting ##sky interesting factors deck poems norman ##water stuck speaker rifle domain premiered ##her dc comics actors 01 reputation eliminated 8th ceiling prisoners script ##nce leather austin mississippi rapidly admiral parallel charlotte guilty tools gender divisions fruit ##bs laboratory nelson fantasy marry rapid aunt tribe requirements aspects suicide amongst adams bone ukraine abc kick sees edinburgh clothing column rough gods hunting broadway gathered concerns ##ek spending ty 12th snapped requires solar bones cavalry ##tta iowa drinking waste index franklin charity thompson stewart tip flash landscape friday enjoy singh poem listening ##back eighth fred differences adapted bomb ukrainian surgery corporate masters anywhere ##more waves odd sean portugal orleans dick debate kent eating puerto cleared 96 expect cinema 97 guitarist blocks electrical agree involving depth dying panel struggle ##ged peninsula adults novels emerged vienna metro debuted shoes tamil songwriter meets prove beating instance heaven scared sending marks artistic passage superior 03 significantly shopping ##tive retained ##izing malaysia technique cheeks ##ola warren maintenance destroy extreme allied 120 appearing ##yn fill advice alabama qualifying policies cleveland hat battery smart authors 10th soundtrack acted dated lb glance equipped coalition funny outer ambassador roy possibility couples campbell dna loose ethan supplies 1898 gonna 88 monster ##res shake agents frequency springs dogs practices 61 gang plastic easier suggests gulf blade exposed colors industries markets pan nervous electoral charts legislation ownership ##idae mac appointment shield copy assault socialist abbey monument license throne employment jay 93 replacement charter cloud powered suffering accounts oak connecticut strongly wright colour crystal 13th context welsh networks voiced gabriel jerry ##cing forehead mp ##ens manage schedule totally remix ##ii forests occupation print nicholas brazilian strategic vampires engineers 76 roots seek correct instrumental und alfred backed hop ##des stanley robinson traveled wayne welcome austrian achieve 67 exit rates 1899 strip whereas ##cs sing deeply adventure bobby rick jamie careful components cap useful personality knee ##shi pushing hosts 02 protest ca ottoman symphony ##sis 63 boundary 1890 processes considering considerable tons ##work ##ft ##nia cooper trading dear conduct 91 illegal apple revolutionary holiday definition harder ##van jacob circumstances destruction ##lle popularity grip classified liverpool donald baltimore flows seeking honour approval 92 mechanical till happening statue critic increasingly immediate describe commerce stare ##ster indonesia meat rounds boats baker orthodox depression formally worn naked claire muttered sentence 11th emily document 77 criticism wished vessel spiritual bent virgin parker minimum murray lunch danny printed compilation keyboards false blow belonged 68 raising 78 cutting ##board pittsburgh ##up 9th shadows 81 hated indigenous jon 15th barry scholar ah ##zer oliver ##gy stick susan meetings attracted spell romantic ##ver ye 1895 photo demanded customers ##ac 1896 logan revival keys modified commanded jeans ##ious upset raw phil detective hiding resident vincent ##bly experiences diamond defeating coverage lucas external parks franchise helen bible successor percussion celebrated il lift profile clan romania ##ied mills ##su nobody achievement shrugged fault 1897 rhythm initiative breakfast carbon 700 69 lasted violent 74 wound ken killer gradually filmed °c dollars processing 94 remove criticized guests sang chemistry ##vin legislature disney ##bridge uniform escaped integrated proposal purple denied liquid karl influential morris nights stones intense experimental twisted 71 84 ##ld pace nazi mitchell ny blind reporter newspapers 14th centers burn basin forgotten surviving filed collections monastery losses manual couch description appropriate merely tag missions sebastian restoration replacing triple 73 elder julia warriors benjamin julian convinced stronger amazing declined versus merchant happens output finland bare barbara absence ignored dawn injuries ##port producers ##ram 82 luis ##ities kw admit expensive electricity nba exception symbol ##ving ladies shower sheriff characteristics ##je aimed button ratio effectively summit angle jury bears foster vessels pants executed evans dozen advertising kicked patrol 1889 competitions lifetime principles athletics ##logy birmingham sponsored 89 rob nomination 1893 acoustic ##sm creature longest ##tra credits harbor dust josh ##so territories milk infrastructure completion thailand indians leon archbishop ##sy assist pitch blake arrangement girlfriend serbian operational hence sad scent fur dj sessions hp refer rarely ##ora exists 1892 ##ten scientists dirty penalty burst portrait seed 79 pole limits rival 1894 stable alpha grave constitutional alcohol arrest flower mystery devil architectural relationships greatly habitat ##istic larry progressive remote cotton ##ics ##ok preserved reaches ##ming cited 86 vast scholarship decisions cbs joy teach 1885 editions knocked eve searching partly participation gap animated fate excellent ##ett na 87 alternate saints youngest ##ily climbed ##ita ##tors suggest ##ct discussion staying choir lakes jacket revenue nevertheless peaked instrument wondering annually managing neil 1891 signing terry ##ice apply clinical brooklyn aim catherine fuck farmers figured ninth pride hugh evolution ordinary involvement comfortable shouted tech encouraged taiwan representation sharing ##lia ##em panic exact cargo competing fat cried 83 1920s occasions pa cabin borders utah marcus ##isation badly muscles ##ance victorian transition warner bet permission ##rin slave terrible similarly shares seth uefa possession medals benefits colleges lowered perfectly mall transit ##ye ##kar publisher ##ened harrison deaths elevation ##ae asleep machines sigh ash hardly argument occasion parent leo decline 1888 contribution ##ua concentration 1000 opportunities hispanic guardian extent emotions hips mason volumes bloody controversy diameter steady mistake phoenix identify violin ##sk departure richmond spin funeral enemies 1864 gear literally connor random sergeant grab confusion 1865 transmission informed op leaning sacred suspended thinks gates portland luck agencies yours hull expert muscle layer practical sculpture jerusalem latest lloyd statistics deeper recommended warrior arkansas mess supports greg eagle 1880 recovered rated concerts rushed ##ano stops eggs files premiere keith ##vo delhi turner pit affair belief paint ##zing mate ##ach ##ev victim ##ology withdrew bonus styles fled ##ud glasgow technologies funded nbc adaptation ##ata portrayed cooperation supporters judges bernard justin hallway ralph ##ick graduating controversial distant continental spider bite ##ho recognize intention mixing ##ese egyptian bow tourism suppose claiming tiger dominated participants vi ##ru nurse partially tape ##rum psychology ##rn essential touring duo voting civilian emotional channels ##king apparent hebrew 1887 tommy carrier intersection beast hudson ##gar ##zo lab nova bench discuss costa ##ered detailed behalf drivers unfortunately obtain ##lis rocky ##dae siege friendship honey ##rian 1861 amy hang posted governments collins respond wildlife preferred operator ##po laura pregnant videos dennis suspected boots instantly weird automatic businessman alleged placing throwing ph mood 1862 perry venue jet remainder ##lli ##ci passion biological boyfriend 1863 dirt buffalo ron segment fa abuse ##era genre thrown stroke colored stress exercise displayed ##gen struggled ##tti abroad dramatic wonderful thereafter madrid component widespread ##sed tale citizen todd monday 1886 vancouver overseas forcing crying descent ##ris discussed substantial ranks regime 1870 provinces switch drum zane ted tribes proof lp cream researchers volunteer manor silk milan donated allies venture principle delivery enterprise ##ves ##ans bars traditionally witch reminded copper ##uk pete inter links colin grinned elsewhere competitive frequent ##oy scream ##hu tension texts submarine finnish defending defend pat detail 1884 affiliated stuart themes villa periods tool belgian ruling crimes answers folded licensed resort demolished hans lucy 1881 lion traded photographs writes craig ##fa trials generated beth noble debt percentage yorkshire erected ss viewed grades confidence ceased islam telephone retail ##ible chile m² roberts sixteen ##ich commented hampshire innocent dual pounds checked regulations afghanistan sung rico liberty assets bigger options angels relegated tribute wells attending leaf ##yan butler romanian forum monthly lisa patterns gmina ##tory madison hurricane rev ##ians bristol ##ula elite valuable disaster democracy awareness germans freyja ##ins loop absolutely paying populations maine sole prayer spencer releases doorway bull ##ani lover midnight conclusion ##sson thirteen lily mediterranean ##lt nhl proud sample ##hill drummer guinea ##ova murphy climb ##ston instant attributed horn ain railways steven ##ao autumn ferry opponent root traveling secured corridor stretched tales sheet trinity cattle helps indicates manhattan murdered fitted 1882 gentle grandmother mines shocked vegas produces ##light caribbean ##ou belong continuous desperate drunk historically trio waved raf dealing nathan bat murmured interrupted residing scientist pioneer harold aaron ##net delta attempting minority mini believes chorus tend lots eyed indoor load shots updated jail ##llo concerning connecting wealth ##ved slaves arrive rangers sufficient rebuilt ##wick cardinal flood muhammad whenever relation runners moral repair viewers arriving revenge punk assisted bath fairly breathe lists innings illustrated whisper nearest voters clinton ties ultimate screamed beijing lions andre fictional gathering comfort radar suitable dismissed hms ban pine wrist atmosphere voivodeship bid timber ##ned ##nan giants ##ane cameron recovery uss identical categories switched serbia laughter noah ensemble therapy peoples touching ##off locally pearl platforms everywhere ballet tables lanka herbert outdoor toured derek 1883 spaces contested swept 1878 exclusive slight connections ##dra winds prisoner collective bangladesh tube publicly wealthy thai ##ys isolated select ##ric insisted pen fortune ticket spotted reportedly animation enforcement tanks 110 decides wider lowest owen ##time nod hitting ##hn gregory furthermore magazines fighters solutions ##ery pointing requested peru reed chancellor knights mask worker eldest flames reduction 1860 volunteers ##tis reporting ##hl wire advisory endemic origins settlers pursue knock consumer 1876 eu compound creatures mansion sentenced ivan deployed guitars frowned involves mechanism kilometers perspective shops maps terminus duncan alien fist bridges ##pers heroes fed derby swallowed ##ros patent sara illness characterized adventures slide hawaii jurisdiction ##op organised ##side adelaide walks biology se ##ties rogers swing tightly boundaries ##rie prepare implementation stolen ##sha certified colombia edwards garage ##mm recalled ##ball rage harm nigeria breast ##ren furniture pupils settle ##lus cuba balls client alaska 21st linear thrust celebration latino genetic terror ##cia ##ening lightning fee witness lodge establishing skull ##ique earning hood ##ei rebellion wang sporting warned missile devoted activist porch worship fourteen package 1871 decorated ##shire housed ##ock chess sailed doctors oscar joan treat garcia harbour jeremy ##ire traditions dominant jacques ##gon ##wan relocated 1879 amendment sized companion simultaneously volleyball spun acre increases stopping loves belongs affect drafted tossed scout battles 1875 filming shoved munich tenure vertical romance pc ##cher argue ##ical craft ranging www opens honest tyler yesterday virtual ##let muslims reveal snake immigrants radical screaming speakers firing saving belonging ease lighting prefecture blame farmer hungry grows rubbed beam sur subsidiary ##cha armenian sao dropping conventional ##fer microsoft reply qualify spots 1867 sweat festivals ##ken immigration physician discover exposure sandy explanation isaac implemented ##fish hart initiated connect stakes presents heights householder pleased tourist regardless slip closest ##ction surely sultan brings riley preparation aboard slammed baptist experiment ongoing interstate organic playoffs ##ika 1877 130 ##tar hindu error tours tier plenty arrangements talks trapped excited sank ho athens 1872 denver welfare suburb athletes trick diverse belly exclusively yelled 1868 ##med conversion ##ette 1874 internationally computers conductor abilities sensitive hello dispute measured globe rocket prices amsterdam flights tigers inn municipalities emotion references 3d ##mus explains airlines manufactured pm archaeological 1873 interpretation devon comment ##ites settlements kissing absolute improvement suite impressed barcelona sullivan jefferson towers jesse julie ##tin ##lu grandson hi gauge regard rings interviews trace raymond thumb departments burns serial bulgarian scores demonstrated ##ix 1866 kyle alberta underneath romanized ##ward relieved acquisition phrase cliff reveals han cuts merger custom ##dar nee gilbert graduation ##nts assessment cafe difficulty demands swung democrat jennifer commons 1940s grove ##yo completing focuses sum substitute bearing stretch reception ##py reflected essentially destination pairs ##ched survival resource ##bach promoting doubles messages tear ##down ##fully parade florence harvey incumbent partial framework 900 pedro frozen procedure olivia controls ##mic shelter personally temperatures ##od brisbane tested sits marble comprehensive oxygen leonard ##kov inaugural iranian referring quarters attitude ##ivity mainstream lined mars dakota norfolk unsuccessful ##° explosion helicopter congressional ##sing inspector bitch seal departed divine ##ters coaching examination punishment manufacturer sink columns unincorporated signals nevada squeezed dylan dining photos martial manuel eighteen elevator brushed plates ministers ivy congregation ##len slept specialized taxes curve restricted negotiations likes statistical arnold inspiration execution bold intermediate significance margin ruler wheels gothic intellectual dependent listened eligible buses widow syria earn cincinnati collapsed recipient secrets accessible philippine maritime goddess clerk surrender breaks playoff database ##ified ##lon ideal beetle aspect soap regulation strings expand anglo shorter crosses retreat tough coins wallace directions pressing ##oon shipping locomotives comparison topics nephew ##mes distinction honors travelled sierra ibn ##over fortress sa recognised carved 1869 clients ##dan intent ##mar coaches describing bread ##ington beaten northwestern ##ona merit youtube collapse challenges em historians objective submitted virus attacking drake assume ##ere diseases marc stem leeds ##cus ##ab farming glasses ##lock visits nowhere fellowship relevant carries restaurants experiments 101 constantly bases targets shah tenth opponents verse territorial ##ira writings corruption ##hs instruction inherited reverse emphasis ##vic employee arch keeps rabbi watson payment uh ##ala nancy ##tre venice fastest sexy banned adrian properly ruth touchdown dollar boards metre circles edges favour comments ok travels liberation scattered firmly ##ular holland permitted diesel kenya den originated ##ral demons resumed dragged rider ##rus servant blinked extend torn ##ias ##sey input meal everybody cylinder kinds camps ##fe bullet logic ##wn croatian evolved healthy fool chocolate wise preserve pradesh ##ess respective 1850 ##ew chicken artificial gross corresponding convicted cage caroline dialogue ##dor narrative stranger mario br christianity failing trent commanding buddhist 1848 maurice focusing yale bike altitude ##ering mouse revised ##sley veteran ##ig pulls theology crashed campaigns legion ##ability drag excellence customer cancelled intensity excuse ##lar liga participating contributing printing ##burn variable ##rk curious bin legacy renaissance ##my symptoms binding vocalist dancer ##nie grammar gospel democrats ya enters sc diplomatic hitler ##ser clouds mathematical quit defended oriented ##heim fundamental hardware impressive equally convince confederate guilt chuck sliding ##ware magnetic narrowed petersburg bulgaria otto phd skill ##ama reader hopes pitcher reservoir hearts automatically expecting mysterious bennett extensively imagined seeds monitor fix ##ative journalism struggling signature ranch encounter photographer observation protests ##pin influences ##hr calendar ##all cruz croatia locomotive hughes naturally shakespeare basement hook uncredited faded theories approaches dare phillips filling fury obama ##ain efficient arc deliver min raid breeding inducted leagues efficiency axis montana eagles ##ked supplied instructions karen picking indicating trap anchor practically christians tomb vary occasional electronics lords readers newcastle faint innovation collect situations engagement 160 claude mixture ##feld peer tissue logo lean ##ration °f floors ##ven architects reducing ##our ##ments rope 1859 ottawa ##har samples banking declaration proteins resignation francois saudi advocate exhibited armor twins divorce ##ras abraham reviewed jo temporarily matrix physically pulse curled ##ena difficulties bengal usage ##ban annie riders certificate ##pi holes warsaw distinctive jessica ##mon mutual 1857 customs circular eugene removal loaded mere vulnerable depicted generations dame heir enormous lightly climbing pitched lessons pilots nepal ram google preparing brad louise renowned ##₂ liam ##ably plaza shaw sophie brilliant bills ##bar ##nik fucking mainland server pleasant seized veterans jerked fail beta brush radiation stored warmth southeastern nate sin raced berkeley joke athlete designation trunk ##low roland qualification archives heels artwork receives judicial reserves ##bed woke installation abu floating fake lesser excitement interface concentrated addressed characteristic amanda saxophone monk auto ##bus releasing egg dies interaction defender ce outbreak glory loving ##bert sequel consciousness http awake ski enrolled ##ress handling rookie brow somebody biography warfare amounts contracts presentation fabric dissolved challenged meter psychological lt elevated rally accurate ##tha hospitals undergraduate specialist venezuela exhibit shed nursing protestant fluid structural footage jared consistent prey ##ska succession reflect exile lebanon wiped suspect shanghai resting integration preservation marvel variant pirates sheep rounded capita sailing colonies manuscript deemed variations clarke functional emerging boxing relaxed curse azerbaijan heavyweight nickname editorial rang grid tightened earthquake flashed miguel rushing ##ches improvements boxes brooks 180 consumption molecular felix societies repeatedly variation aids civic graphics professionals realm autonomous receiver delayed workshop militia chairs trump canyon ##point harsh extending lovely happiness ##jan stake eyebrows embassy wellington hannah ##ella sony corners bishops swear cloth contents xi namely commenced 1854 stanford nashville courage graphic commitment garrison ##bin hamlet clearing rebels attraction literacy cooking ruins temples jenny humanity celebrate hasn freight sixty rebel bastard ##art newton ##ada deer ##ges ##ching smiles delaware singers ##ets approaching assists flame ##ph boulevard barrel planted ##ome pursuit ##sia consequences posts shallow invitation rode depot ernest kane rod concepts preston topic chambers striking blast arrives descendants montgomery ranges worlds ##lay ##ari span chaos praise ##ag fewer 1855 sanctuary mud fbi ##ions programmes maintaining unity harper bore handsome closure tournaments thunder nebraska linda facade puts satisfied argentine dale cork dome panama ##yl 1858 tasks experts ##ates feeding equation ##las ##ida ##tu engage bryan ##ax um quartet melody disbanded sheffield blocked gasped delay kisses maggie connects ##non sts poured creator publishers ##we guided ellis extinct hug gaining ##ord complicated ##bility poll clenched investigate ##use thereby quantum spine cdp humor kills administered semifinals ##du encountered ignore ##bu commentary ##maker bother roosevelt 140 plains halfway flowing cultures crack imprisoned neighboring airline ##ses ##view ##mate ##ec gather wolves marathon transformed ##ill cruise organisations carol punch exhibitions numbered alarm ratings daddy silently ##stein queens colours impression guidance liu tactical ##rat marshal della arrow ##ings rested feared tender owns bitter advisor escort ##ides spare farms grants ##ene dragons encourage colleagues cameras ##und sucked pile spirits prague statements suspension landmark fence torture recreation bags permanently survivors pond spy predecessor bombing coup ##og protecting transformation glow ##lands ##book dug priests andrea feat barn jumping ##chen ##ologist ##con casualties stern auckland pipe serie revealing ba ##bel trevor mercy spectrum yang consist governing collaborated possessed epic comprises blew shane ##ack lopez honored magical sacrifice judgment perceived hammer mtv baronet tune das missionary sheets 350 neutral oral threatening attractive shade aims seminary ##master estates 1856 michel wounds refugees manufacturers ##nic mercury syndrome porter ##iya ##din hamburg identification upstairs purse widened pause cared breathed affiliate santiago prevented celtic fisher 125 recruited byzantine reconstruction farther ##mp diet sake au spite sensation ##ert blank separation 105 ##hon vladimir armies anime ##lie accommodate orbit cult sofia archive ##ify ##box founders sustained disorder honours northeastern mia crops violet threats blanket fires canton followers southwestern prototype voyage assignment altered moderate protocol pistol ##eo questioned brass lifting 1852 math authored ##ual doug dimensional dynamic ##san 1851 pronounced grateful quest uncomfortable boom presidency stevens relating politicians chen barrier quinn diana mosque tribal cheese palmer portions sometime chester treasure wu bend download millions reforms registration ##osa consequently monitoring ate preliminary brandon invented ps eaten exterior intervention ports documented log displays lecture sally favourite ##itz vermont lo invisible isle breed ##ator journalists relay speaks backward explore midfielder actively stefan procedures cannon blond kenneth centered servants chains libraries malcolm essex henri slavery ##hal facts fairy coached cassie cats washed cop ##fi announcement item 2000s vinyl activated marco frontier growled curriculum ##das loyal accomplished leslie ritual kenny ##00 vii napoleon hollow hybrid jungle stationed friedrich counted ##ulated platinum theatrical seated col rubber glen 1840 diversity healing extends id provisions administrator columbus ##oe tributary te assured org ##uous prestigious examined lectures grammy ronald associations bailey allan essays flute believing consultant proceedings travelling 1853 kit kerala yugoslavia buddy methodist ##ith burial centres batman ##nda discontinued bo dock stockholm lungs severely ##nk citing manga ##ugh steal mumbai iraqi robot celebrity bride broadcasts abolished pot joel overhead franz packed reconnaissance johann acknowledged introduce handled doctorate developments drinks alley palestine ##nis ##aki proceeded recover bradley grain patch afford infection nationalist legendary ##ath interchange virtually gen gravity exploration amber vital wishes powell doctrine elbow screenplay ##bird contribute indonesian pet creates ##com enzyme kylie discipline drops manila hunger ##ien layers suffer fever bits monica keyboard manages ##hood searched appeals ##bad testament grande reid ##war beliefs congo ##ification ##dia si requiring ##via casey 1849 regret streak rape depends syrian sprint pound tourists upcoming pub ##xi tense ##els practiced echo nationwide guild motorcycle liz ##zar chiefs desired elena bye precious absorbed relatives booth pianist ##mal citizenship exhausted wilhelm ##ceae ##hed noting quarterback urge hectares ##gue ace holly ##tal blonde davies parked sustainable stepping twentieth airfield galaxy nest chip ##nell tan shaft paulo requirement ##zy paradise tobacco trans renewed vietnamese ##cker ##ju suggesting catching holmes enjoying md trips colt holder butterfly nerve reformed cherry bowling trailer carriage goodbye appreciate toy joshua interactive enabled involve ##kan collar determination bunch facebook recall shorts superintendent episcopal frustration giovanni nineteenth laser privately array circulation ##ovic armstrong deals painful permit discrimination ##wi aires retiring cottage ni ##sta horizon ellen jamaica ripped fernando chapters playstation patron lecturer navigation behaviour genes georgian export solomon rivals swift seventeen rodriguez princeton independently sox 1847 arguing entity casting hank criteria oakland geographic milwaukee reflection expanding conquest dubbed ##tv halt brave brunswick doi arched curtis divorced predominantly somerset streams ugly zoo horrible curved buenos fierce dictionary vector theological unions handful stability chan punjab segments ##lly altar ignoring gesture monsters pastor ##stone thighs unexpected operators abruptly coin compiled associates improving migration pin ##ose compact collegiate reserved ##urs quarterfinals roster restore assembled hurry oval ##cies 1846 flags martha ##del victories sharply ##rated argues deadly neo drawings symbols performer ##iel griffin restrictions editing andrews java journals arabia compositions dee pierce removing hindi casino runway civilians minds nasa hotels ##zation refuge rent retain potentially conferences suburban conducting ##tto ##tions ##tle descended massacre ##cal ammunition terrain fork souls counts chelsea durham drives cab ##bank perth realizing palestinian finn simpson ##dal betty ##ule moreover particles cardinals tent evaluation extraordinary ##oid inscription ##works wednesday chloe maintains panels ashley trucks ##nation cluster sunlight strikes zhang ##wing dialect canon ##ap tucked ##ws collecting ##mas ##can ##sville maker quoted evan franco aria buying cleaning eva closet provision apollo clinic rat ##ez necessarily ac ##gle ##ising venues flipped cent spreading trustees checking authorized ##sco disappointed ##ado notion duration trumpet hesitated topped brussels rolls theoretical hint define aggressive repeat wash peaceful optical width allegedly mcdonald strict copyright ##illa investors mar jam witnesses sounding miranda michelle privacy hugo harmony ##pp valid lynn glared nina 102 headquartered diving boarding gibson ##ncy albanian marsh routine dealt enhanced er intelligent substance targeted enlisted discovers spinning observations pissed smoking rebecca capitol visa varied costume seemingly indies compensation surgeon thursday arsenal westminster suburbs rid anglican ##ridge knots foods alumni lighter fraser whoever portal scandal ##ray gavin advised instructor flooding terrorist ##ale teenage interim senses duck teen thesis abby eager overcome ##ile newport glenn rises shame ##cc prompted priority forgot bomber nicolas protective 360 cartoon katherine breeze lonely trusted henderson richardson relax banner candy palms remarkable ##rio legends cricketer essay ordained edmund rifles trigger ##uri ##away sail alert 1830 audiences penn sussex siblings pursued indianapolis resist rosa consequence succeed avoided 1845 ##ulation inland ##tie ##nna counsel profession chronicle hurried ##una eyebrow eventual bleeding innovative cure ##dom committees accounting con scope hardy heather tenor gut herald codes tore scales wagon ##oo luxury tin prefer fountain triangle bonds darling convoy dried traced beings troy accidentally slam findings smelled joey lawyers outcome steep bosnia configuration shifting toll brook performers lobby philosophical construct shrine aggregate boot cox phenomenon savage insane solely reynolds lifestyle ##ima nationally holdings consideration enable edgar mo mama ##tein fights relegation chances atomic hub conjunction awkward reactions currency finale kumar underwent steering elaborate gifts comprising melissa veins reasonable sunshine chi solve trails inhabited elimination ethics huh ana molly consent apartments layout marines ##ces hunters bulk ##oma hometown ##wall ##mont cracked reads neighbouring withdrawn admission wingspan damned anthology lancashire brands batting forgive cuban awful ##lyn 104 dimensions imagination ##ade dante ##ship tracking desperately goalkeeper ##yne groaned workshops confident burton gerald milton circus uncertain slope copenhagen sophia fog philosopher portraits accent cycling varying gripped larvae garrett specified scotia mature luther kurt rap ##kes aerial 750 ferdinand heated es transported ##shan safely nonetheless ##orn ##gal motors demanding ##sburg startled ##brook ally generate caps ghana stained demo mentions beds ap afterward diary ##bling utility ##iro richards 1837 conspiracy conscious shining footsteps observer cyprus urged loyalty developer probability olive upgraded gym miracle insects graves 1844 ourselves hydrogen amazon katie tickets poets ##pm planes ##pan prevention witnessed dense jin randy tang warehouse monroe bang archived elderly investigations alec granite mineral conflicts controlling aboriginal carlo ##zu mechanics stan stark rhode skirt est ##berry bombs respected ##horn imposed limestone deny nominee memphis grabbing disabled ##als amusement aa frankfurt corn referendum varies slowed disk firms unconscious incredible clue sue ##zhou twist ##cio joins idaho chad developers computing destroyer 103 mortal tucker kingston choices yu carson 1800 os whitney geneva pretend dimension staged plateau maya ##une freestyle ##bc rovers hiv ##ids tristan classroom prospect ##hus honestly diploma lied thermal auxiliary feast unlikely iata ##tel morocco pounding treasury lithuania considerably 1841 dish 1812 geological matching stumbled destroying marched brien advances cake nicole belle settling measuring directing ##mie tuesday bassist capabilities stunned fraud torpedo ##list ##phone anton wisdom surveillance ruined ##ulate lawsuit healthcare theorem halls trend aka horizontal dozens acquire lasting swim hawk gorgeous fees vicinity decrease adoption tactics ##ography pakistani ##ole draws ##hall willie burke heath algorithm integral powder elliott brigadier jackie tate varieties darker ##cho lately cigarette specimens adds ##ree ##ensis ##inger exploded finalist cia murders wilderness arguments nicknamed acceptance onwards manufacture robertson jets tampa enterprises blog loudly composers nominations 1838 ai malta inquiry automobile hosting viii rays tilted grief museums strategies furious euro equality cohen poison surrey wireless governed ridiculous moses ##esh ##room vanished ##ito barnes attract morrison istanbul ##iness absent rotation petition janet ##logical satisfaction custody deliberately observatory comedian surfaces pinyin novelist strictly canterbury oslo monks embrace ibm jealous photograph continent dorothy marina doc excess holden allegations explaining stack avoiding lance storyline majesty poorly spike dos bradford raven travis classics proven voltage pillow fists butt 1842 interpreted ##car 1839 gage telegraph lens promising expelled casual collector zones ##min silly nintendo ##kh ##bra downstairs chef suspicious afl flies vacant uganda pregnancy condemned lutheran estimates cheap decree saxon proximity stripped idiot deposits contrary presenter magnus glacier im offense edwin ##ori upright ##long bolt ##ois toss geographical ##izes environments delicate marking abstract xavier nails windsor plantation occurring equity saskatchewan fears drifted sequences vegetation revolt ##stic 1843 sooner fusion opposing nato skating 1836 secretly ruin lease ##oc edit ##nne flora anxiety ruby ##ological ##mia tel bout taxi emmy frost rainbow compounds foundations rainfall assassination nightmare dominican ##win achievements deserve orlando intact armenia ##nte calgary valentine 106 marion proclaimed theodore bells courtyard thigh gonzalez console troop minimal monte everyday ##ence ##if supporter terrorism buck openly presbyterian activists carpet ##iers rubbing uprising ##yi cute conceived legally ##cht millennium cello velocity ji rescued cardiff 1835 rex concentrate senators beard rendered glowing battalions scouts competitors sculptor catalogue arctic ion raja bicycle wow glancing lawn ##woman gentleman lighthouse publish predicted calculated ##val variants ##gne strain ##ui winston deceased ##nus touchdowns brady caleb sinking echoed crush hon blessed protagonist hayes endangered magnitude editors ##tine estimate responsibilities ##mel backup laying consumed sealed zurich lovers frustrated ##eau ahmed kicking mit treasurer 1832 biblical refuse terrified pump agrees genuine imprisonment refuses plymouth ##hen lou ##nen tara trembling antarctic ton learns ##tas crap crucial faction atop ##borough wrap lancaster odds hopkins erik lyon ##eon bros ##ode snap locality tips empress crowned cal acclaimed chuckled ##ory clara sends mild towel ##fl ##day ##а wishing assuming interviewed ##bal ##die interactions eden cups helena ##lf indie beck ##fire batteries filipino wizard parted ##lam traces ##born rows idol albany delegates ##ees ##sar discussions ##ex notre instructed belgrade highways suggestion lauren possess orientation alexandria abdul beats salary reunion ludwig alright wagner intimate pockets slovenia hugged brighton merchants cruel stole trek slopes repairs enrollment politically underlying promotional counting boeing ##bb isabella naming ##и keen bacteria listing separately belfast ussr 450 lithuanian anybody ribs sphere martinez cock embarrassed proposals fragments nationals ##fs ##wski premises fin 1500 alpine matched freely bounded jace sleeve ##af gaming pier populated evident ##like frances flooded ##dle frightened pour trainer framed visitor challenging pig wickets ##fold infected email ##pes arose ##aw reward ecuador oblast vale ch shuttle ##usa bach rankings forbidden cornwall accordance salem consumers bruno fantastic toes machinery resolved julius remembering propaganda iceland bombardment tide contacts wives ##rah concerto macdonald albania implement daisy tapped sudan helmet angela mistress ##lic crop sunk finest ##craft hostile ##ute ##tsu boxer fr paths adjusted habit ballot supervision soprano ##zen bullets wicked sunset regiments disappear lamp performs app ##gia ##oa rabbit digging incidents entries ##cion dishes ##oi introducing ##ati ##fied freshman slot jill tackles baroque backs ##iest lone sponsor destiny altogether convert ##aro consensus shapes demonstration basically feminist auction artifacts ##bing strongest twitter halifax 2019 allmusic mighty smallest precise alexandra viola ##los ##ille manuscripts ##illo dancers ari managers monuments blades barracks springfield maiden consolidated electron ##end berry airing wheat nobel inclusion blair payments geography bee cc eleanor react ##hurst afc manitoba ##yu su lineup fitness recreational investments airborne disappointment ##dis edmonton viewing ##row renovation ##cast infant bankruptcy roses aftermath pavilion ##yer carpenter withdrawal ladder ##hy discussing popped reliable agreements rochester ##abad curves bombers 220 rao reverend decreased choosing 107 stiff consulting naples crawford tracy ka ribbon cops ##lee crushed deciding unified teenager accepting flagship explorer poles sanchez inspection revived skilled induced exchanged flee locals tragedy swallow loading hanna demonstrate ##ela salvador flown contestants civilization ##ines wanna rhodes fletcher hector knocking considers ##ough nash mechanisms sensed mentally walt unclear ##eus renovated madame ##cks crews governmental ##hin undertaken monkey ##ben ##ato fatal armored copa caves governance grasp perception certification froze damp tugged wyoming ##rg ##ero newman ##lor nerves curiosity graph 115 ##ami withdraw tunnels dull meredith moss exhibits neighbors communicate accuracy explored raiders republicans secular kat superman penny criticised ##tch freed update conviction wade ham likewise delegation gotta doll promises technological myth nationality resolve convent ##mark sharon dig sip coordinator entrepreneur fold ##dine capability councillor synonym blown swan cursed 1815 jonas haired sofa canvas keeper rivalry ##hart rapper speedway swords postal maxwell estonia potter recurring ##nn ##ave errors ##oni cognitive 1834 ##² claws nadu roberto bce wrestler ellie ##ations infinite ink ##tia presumably finite staircase 108 noel patricia nacional ##cation chill eternal tu preventing prussia fossil limbs ##logist ernst frog perez rene ##ace pizza prussian ##ios ##vy molecules regulatory answering opinions sworn lengths supposedly hypothesis upward habitats seating ancestors drank yield hd synthesis researcher modest ##var mothers peered voluntary homeland ##the acclaim ##igan static valve luxembourg alto carroll fe receptor norton ambulance ##tian johnston catholics depicting jointly elephant gloria mentor badge ahmad distinguish remarked councils precisely allison advancing detection crowded ##10 cooperative ankle mercedes dagger surrendered pollution commit subway jeffrey lesson sculptures provider ##fication membrane timothy rectangular fiscal heating teammate basket particle anonymous deployment ##ple missiles courthouse proportion shoe sec ##ller complaints forbes blacks abandon remind sizes overwhelming autobiography natalie ##awa risks contestant countryside babies scorer invaded enclosed proceed hurling disorders ##cu reflecting continuously cruiser graduates freeway investigated ore deserved maid blocking phillip jorge shakes dove mann variables lacked burden accompanying que consistently organizing provisional complained endless ##rm tubes juice georges krishna mick labels thriller ##uch laps arcade sage snail ##table shannon fi laurence seoul vacation presenting hire churchill surprisingly prohibited savannah technically ##oli 170 ##lessly testimony suited speeds toys romans mlb flowering measurement talented kay settings charleston expectations shattered achieving triumph ceremonies portsmouth lanes mandatory loser stretching cologne realizes seventy cornell careers webb ##ulating americas budapest ava suspicion ##ison yo conrad ##hai sterling jessie rector ##az 1831 transform organize loans christine volcanic warrant slender summers subfamily newer danced dynamics rhine proceeds heinrich gastropod commands sings facilitate easter ra positioned responses expense fruits yanked imported 25th velvet vic primitive tribune baldwin neighbourhood donna rip hay pr ##uro 1814 espn welcomed ##aria qualifier glare highland timing ##cted shells eased geometry louder exciting slovakia ##sion ##iz ##lot savings prairie ##ques marching rafael tonnes ##lled curtain preceding shy heal greene worthy ##pot detachment bury sherman ##eck reinforced seeks bottles contracted duchess outfit walsh ##sc mickey ##ase geoffrey archer squeeze dawson eliminate invention ##enberg neal ##eth stance dealer coral maple retire polo simplified ##ht 1833 hid watts backwards jules ##oke genesis mt frames rebounds burma woodland moist santos whispers drained subspecies ##aa streaming ulster burnt correspondence maternal gerard denis stealing ##load genius duchy ##oria inaugurated momentum suits placement sovereign clause thames ##hara confederation reservation sketch yankees lets rotten charm hal verses ultra commercially dot salon citation adopt winnipeg mist allocated cairo ##boy jenkins interference objectives ##wind 1820 portfolio armoured sectors ##eh initiatives ##world integrity exercises robe tap ab gazed ##tones distracted rulers 111 favorable jerome tended cart factories ##eri diplomat valued gravel charitable ##try calvin exploring chang shepherd terrace pdf pupil ##ural reflects ups ##rch governors shelf depths ##nberg trailed crest tackle ##nian ##ats hatred ##kai clare makers ethiopia longtime detected embedded lacking slapped rely thomson anticipation iso morton successive agnes screenwriter straightened philippe playwright haunted licence iris intentions sutton 112 logical correctly ##weight branded licked tipped silva ricky narrator requests ##ents greeted supernatural cow ##wald lung refusing employer strait gaelic liner ##piece zoe sabha ##mba driveway harvest prints bates reluctantly threshold algebra ira wherever coupled 240 assumption picks ##air designers raids gentlemen ##ean roller blowing leipzig locks screw dressing strand ##lings scar dwarf depicts ##nu nods ##mine differ boris ##eur yuan flip ##gie mob invested questioning applying ##ture shout ##sel gameplay blamed illustrations bothered weakness rehabilitation ##of ##zes envelope rumors miners leicester subtle kerry ##ico ferguson ##fu premiership ne ##cat bengali prof catches remnants dana ##rily shouting presidents baltic ought ghosts dances sailors shirley fancy dominic ##bie madonna ##rick bark buttons gymnasium ashes liver toby oath providence doyle evangelical nixon cement carnegie embarked hatch surroundings guarantee needing pirate essence ##bee filter crane hammond projected immune percy twelfth ##ult regent doctoral damon mikhail ##ichi lu critically elect realised abortion acute screening mythology steadily ##fc frown nottingham kirk wa minneapolis ##rra module algeria mc nautical encounters surprising statues availability shirts pie alma brows munster mack soup crater tornado sanskrit cedar explosive bordered dixon planets stamp exam happily ##bble carriers kidnapped ##vis accommodation emigrated ##met knockout correspondent violation profits peaks lang specimen agenda ancestry pottery spelling equations obtaining ki linking 1825 debris asylum ##20 buddhism teddy ##ants gazette ##nger ##sse dental eligibility utc fathers averaged zimbabwe francesco coloured hissed translator lynch mandate humanities mackenzie uniforms lin ##iana ##gio asset mhz fitting samantha genera wei rim beloved shark riot entities expressions indo carmen slipping owing abbot neighbor sidney ##av rats recommendations encouraging squadrons anticipated commanders conquered ##oto donations diagnosed ##mond divide ##iva guessed decoration vernon auditorium revelation conversations ##kers ##power herzegovina dash alike protested lateral herman accredited mg ##gent freeman mel fiji crow crimson ##rine livestock ##pped humanitarian bored oz whip ##lene ##ali legitimate alter grinning spelled anxious oriental wesley ##nin ##hole carnival controller detect ##ssa bowed educator kosovo macedonia ##sin occupy mastering stephanie janeiro para unaware nurses noon 135 cam hopefully ranger combine sociology polar rica ##eer neill ##sman holocaust ##ip doubled lust 1828 109 decent cooling unveiled ##card 1829 nsw homer chapman meyer ##gin dive mae reagan expertise ##gled darwin brooke sided prosecution investigating comprised petroleum genres reluctant differently trilogy johns vegetables corpse highlighted lounge pension unsuccessfully elegant aided ivory beatles amelia cain dubai sunny immigrant babe click ##nder underwater pepper combining mumbled atlas horns accessed ballad physicians homeless gestured rpm freak louisville corporations patriots prizes rational warn modes decorative overnight din troubled phantom ##ort monarch sheer ##dorf generals guidelines organs addresses ##zon enhance curling parishes cord ##kie linux caesar deutsche bavaria ##bia coleman cyclone ##eria bacon petty ##yama ##old hampton diagnosis 1824 throws complexity rita disputed ##₃ pablo ##sch marketed trafficking ##ulus examine plague formats ##oh vault faithful ##bourne webster ##ox highlights ##ient ##ann phones vacuum sandwich modeling ##gated bolivia clergy qualities isabel ##nas ##ars wears screams reunited annoyed bra ##ancy ##rate differential transmitter tattoo container poker ##och excessive resides cowboys ##tum augustus trash providers statute retreated balcony reversed void storey preceded masses leap laughs neighborhoods wards schemes falcon santo battlefield pad ronnie thread lesbian venus ##dian beg sandstone daylight punched gwen analog stroked wwe acceptable measurements dec toxic ##kel adequate surgical economist parameters varsity ##sberg quantity ella ##chy ##rton countess generating precision diamonds expressway ga ##ı 1821 uruguay talents galleries expenses scanned colleague outlets ryder lucien ##ila paramount ##bon syracuse dim fangs gown sweep ##sie toyota missionaries websites ##nsis sentences adviser val trademark spells ##plane patience starter slim ##borg toe incredibly shoots elliot nobility ##wyn cowboy endorsed gardner tendency persuaded organisms emissions kazakhstan amused boring chips themed ##hand llc constantinople chasing systematic guatemala borrowed erin carey ##hard highlands struggles 1810 ##ifying ##ced wong exceptions develops enlarged kindergarten castro ##ern ##rina leigh zombie juvenile ##most consul ##nar sailor hyde clarence intensive pinned nasty useless jung clayton stuffed exceptional ix apostolic 230 transactions ##dge exempt swinging cove religions ##ash shields dairy bypass 190 pursuing bug joyce bombay chassis southampton chat interact redesignated ##pen nascar pray salmon rigid regained malaysian grim publicity constituted capturing toilet delegate purely tray drift loosely striker weakened trinidad mitch itv defines transmitted ming scarlet nodding fitzgerald fu narrowly sp tooth standings virtue ##₁ ##wara ##cting chateau gloves lid ##nel hurting conservatory ##pel sinclair reopened sympathy nigerian strode advocated optional chronic discharge ##rc suck compatible laurel stella shi fails wage dodge 128 informal sorts levi buddha villagers ##aka chronicles heavier summoned gateway 3000 eleventh jewelry translations accordingly seas ##ency fiber pyramid cubic dragging ##ista caring ##ops android contacted lunar ##dt kai lisbon patted 1826 sacramento theft madagascar subtropical disputes ta holidays piper willow mare cane itunes newfoundland benny companions dong raj observe roar charming plaque tibetan fossils enacted manning bubble tina tanzania ##eda ##hir funk swamp deputies cloak ufc scenario par scratch metals anthem guru engaging specially ##boat dialects nineteen cecil duet disability messenger unofficial ##lies defunct eds moonlight drainage surname puzzle honda switching conservatives mammals knox broadcaster sidewalk cope ##ried benson princes peterson ##sal bedford sharks eli wreck alberto gasp archaeology lgbt teaches securities madness compromise waving coordination davidson visions leased possibilities eighty jun fernandez enthusiasm assassin sponsorship reviewer kingdoms estonian laboratories ##fy ##nal applies verb celebrations ##zzo rowing lightweight sadness submit mvp balanced dude ##vas explicitly metric magnificent mound brett mohammad mistakes irregular ##hing ##ass sanders betrayed shipped surge ##enburg reporters termed georg pity verbal bulls abbreviated enabling appealed ##are ##atic sicily sting heel sweetheart bart spacecraft brutal monarchy ##tter aberdeen cameo diane ##ub survivor clyde ##aries complaint ##makers clarinet delicious chilean karnataka coordinates 1818 panties ##rst pretending ar dramatically kiev bella tends distances 113 catalog launching instances telecommunications portable lindsay vatican ##eim angles aliens marker stint screens bolton ##rne judy wool benedict plasma europa spark imaging filmmaker swiftly ##een contributor ##nor opted stamps apologize financing butter gideon sophisticated alignment avery chemicals yearly speculation prominence professionally ##ils immortal institutional inception wrists identifying tribunal derives gains ##wo papal preference linguistic vince operative brewery ##ont unemployment boyd ##ured ##outs albeit prophet 1813 bi ##rr ##face ##rad quarterly asteroid cleaned radius temper ##llen telugu jerk viscount menu ##ote glimpse ##aya yacht hawaiian baden ##rl laptop readily ##gu monetary offshore scots watches ##yang ##arian upgrade needle xbox lea encyclopedia flank fingertips ##pus delight teachings confirm roth beaches midway winters ##iah teasing daytime beverly gambling bonnie ##backs regulated clement hermann tricks knot ##shing ##uring ##vre detached ecological owed specialty byron inventor bats stays screened unesco midland trim affection ##ander ##rry jess thoroughly feedback ##uma chennai strained heartbeat wrapping overtime pleaded ##sworth mon leisure oclc ##tate ##ele feathers angelo thirds nuts surveys clever gill commentator ##dos darren rides gibraltar ##nc ##mu dissolution dedication shin meals saddle elvis reds chaired taller appreciation functioning niece favored advocacy robbie criminals suffolk yugoslav passport constable congressman hastings vera ##rov consecrated sparks ecclesiastical confined ##ovich muller floyd nora 1822 paved 1827 cumberland ned saga spiral ##flow appreciated yi collaborative treating similarities feminine finishes ##ib jade import ##nse ##hot champagne mice securing celebrities helsinki attributes ##gos cousins phases ache lucia gandhi submission vicar spear shine tasmania biting detention constitute tighter seasonal ##gus terrestrial matthews ##oka effectiveness parody philharmonic ##onic 1816 strangers encoded consortium guaranteed regards shifts tortured collision supervisor inform broader insight theaters armour emeritus blink incorporates mapping ##50 ##ein handball flexible ##nta substantially generous thief ##own carr loses 1793 prose ucla romeo generic metallic realization damages mk commissioners zach default ##ther helicopters lengthy stems spa partnered spectators rogue indication penalties teresa 1801 sen ##tric dalton ##wich irving photographic ##vey dell deaf peters excluded unsure ##vable patterson crawled ##zio resided whipped latvia slower ecole pipes employers maharashtra comparable va textile pageant ##gel alphabet binary irrigation chartered choked antoine offs waking supplement ##wen quantities demolition regain locate urdu folks alt 114 ##mc scary andreas whites ##ava classrooms mw aesthetic publishes valleys guides cubs johannes bryant conventions affecting ##itt drain awesome isolation prosecutor ambitious apology captive downs atmospheric lorenzo aisle beef foul ##onia kidding composite disturbed illusion natives ##ffer emi rockets riverside wartime painters adolf melted ##ail uncertainty simulation hawks progressed meantime builder spray breach unhappy regina russians ##urg determining ##tation tram 1806 ##quin aging ##12 1823 garion rented mister diaz terminated clip 1817 depend nervously disco owe defenders shiva notorious disbelief shiny worcester ##gation ##yr trailing undertook islander belarus limitations watershed fuller overlooking utilized raphael 1819 synthetic breakdown klein ##nate moaned memoir lamb practicing ##erly cellular arrows exotic ##graphy witches 117 charted rey hut hierarchy subdivision freshwater giuseppe aloud reyes qatar marty sideways utterly sexually jude prayers mccarthy softball blend damien ##gging ##metric wholly erupted lebanese negro revenues tasted comparative teamed transaction labeled maori sovereignty parkway trauma gran malay 121 advancement descendant 2020 buzz salvation inventory symbolic ##making antarctica mps ##gas ##bro mohammed myanmar holt submarines tones ##lman locker patriarch bangkok emerson remarks predators kin afghan confession norwich rental emerge advantages ##zel rca ##hold shortened storms aidan ##matic autonomy compliance ##quet dudley atp ##osis 1803 motto documentation summary professors spectacular christina archdiocese flashing innocence remake ##dell psychic reef scare employ rs sticks meg gus leans ##ude accompany bergen tomas ##iko doom wages pools ##nch ##bes breasts scholarly alison outline brittany breakthrough willis realistic ##cut ##boro competitor ##stan pike picnic icon designing commercials washing villain skiing micro costumes auburn halted executives ##hat logistics cycles vowel applicable barrett exclaimed eurovision eternity ramon ##umi ##lls modifications sweeping disgust ##uck torch aviv ensuring rude dusty sonic donovan outskirts cu pathway ##band ##gun ##lines disciplines acids cadet paired ##40 sketches ##sive marriages ##⁺ folding peers slovak implies admired ##beck 1880s leopold instinct attained weston megan horace ##ination dorsal ingredients evolutionary ##its complications deity lethal brushing levy deserted institutes posthumously delivering telescope coronation motivated rapids luc flicked pays volcano tanner weighed ##nica crowds frankie gifted addressing granddaughter winding ##rna constantine gomez ##front landscapes rudolf anthropology slate werewolf ##lio astronomy circa rouge dreaming sack knelt drowned naomi prolific tracked freezing herb ##dium agony randall twisting wendy deposit touches vein wheeler ##bbled ##bor batted retaining tire presently compare specification daemon nigel ##grave merry recommendation czechoslovakia sandra ng roma ##sts lambert inheritance sheikh winchester cries examining ##yle comeback cuisine nave ##iv ko retrieve tomatoes barker polished defining irene lantern personalities begging tract swore 1809 175 ##gic omaha brotherhood ##rley haiti ##ots exeter ##ete ##zia steele dumb pearson 210 surveyed elisabeth trends ##ef fritz ##rf premium bugs fraction calmly viking ##birds tug inserted unusually ##ield confronted distress crashing brent turks resign ##olo cambodia gabe sauce ##kal evelyn 116 extant clusters quarry teenagers luna ##lers ##ister affiliation drill ##ashi panthers scenic libya anita strengthen inscriptions ##cated lace sued judith riots ##uted mint ##eta preparations midst dub challenger ##vich mock cf displaced wicket breaths enables schmidt analyst ##lum ag highlight automotive axe josef newark sufficiently resembles 50th ##pal flushed mum traits ##ante commodore incomplete warming titular ceremonial ethical 118 celebrating eighteenth cao lima medalist mobility strips snakes ##city miniature zagreb barton escapes umbrella automated doubted differs cooled georgetown dresden cooked fade wyatt rna jacobs carlton abundant stereo boost madras inning ##hia spur ip malayalam begged osaka groan escaping charging dose vista ##aj bud papa communists advocates edged tri ##cent resemble peaking necklace fried montenegro saxony goose glances stuttgart curator recruit grocery sympathetic ##tting ##fort 127 lotus randolph ancestor ##rand succeeding jupiter 1798 macedonian ##heads hiking 1808 handing fischer ##itive garbage node ##pies prone singular papua inclined attractions italia pouring motioned grandma garnered jacksonville corp ego ringing aluminum ##hausen ordering ##foot drawer traders synagogue ##play ##kawa resistant wandering fragile fiona teased var hardcore soaked jubilee decisive exposition mercer poster valencia hale kuwait 1811 ##ises ##wr ##eed tavern gamma 122 johan ##uer airways amino gil ##ury vocational domains torres ##sp generator folklore outcomes ##keeper canberra shooter fl beams confrontation ##lling ##gram feb aligned forestry pipeline jax motorway conception decay ##tos coffin ##cott stalin 1805 escorted minded ##nam sitcom purchasing twilight veronica additions passive tensions straw 123 frequencies 1804 refugee cultivation ##iate christie clary bulletin crept disposal ##rich ##zong processor crescent ##rol bmw emphasized whale nazis aurora ##eng dwelling hauled sponsors toledo mega ideology theatres tessa cerambycidae saves turtle cone suspects kara rusty yelling greeks mozart shades cocked participant ##tro shire spit freeze necessity ##cos inmates nielsen councillors loaned uncommon omar peasants botanical offspring daniels formations jokes 1794 pioneers sigma licensing ##sus wheelchair polite 1807 liquor pratt trustee ##uta forewings balloon ##zz kilometre camping explicit casually shawn foolish teammates nm hassan carrie judged satisfy vanessa knives selective cnn flowed ##lice eclipse stressed eliza mathematician cease cultivated ##roy commissions browns ##ania destroyers sheridan meadow ##rius minerals ##cial downstream clash gram memoirs ventures baha seymour archie midlands edith fare flynn invite canceled tiles stabbed boulder incorporate amended camden facial mollusk unreleased descriptions yoga grabs 550 raises ramp shiver ##rose coined pioneering tunes qing warwick tops 119 melanie giles ##rous wandered ##inal annexed nov 30th unnamed ##ished organizational airplane normandy stoke whistle blessing violations chased holders shotgun ##ctic outlet reactor ##vik tires tearing shores fortified mascot constituencies nc columnist productive tibet ##rta lineage hooked oct tapes judging cody ##gger hansen kashmir triggered ##eva solved cliffs ##tree resisted anatomy protesters transparent implied ##iga injection mattress excluding ##mbo defenses helpless devotion ##elli growl liberals weber phenomena atoms plug ##iff mortality apprentice howe convincing aaa swimmer barber leone promptly sodium def nowadays arise ##oning gloucester corrected dignity norm erie ##ders elders evacuated sylvia compression ##yar hartford pose backpack reasoning accepts 24th wipe millimetres marcel ##oda dodgers albion 1790 overwhelmed aerospace oaks 1795 showcase acknowledge recovering nolan ashe hurts geology fashioned disappearance farewell swollen shrug marquis wimbledon 124 rue 1792 commemorate reduces experiencing inevitable calcutta intel ##court murderer sticking fisheries imagery bloom 280 brake ##inus gustav hesitation memorable po viral beans accidents tunisia antenna spilled consort treatments aye perimeter ##gard donation hostage migrated banker addiction apex lil trout ##ously conscience ##nova rams sands genome passionate troubles ##lets ##set amid ##ibility ##ret higgins exceed vikings ##vie payne ##zan muscular ##ste defendant sucking ##wal ibrahim fuselage claudia vfl europeans snails interval ##garh preparatory statewide tasked lacrosse viktor ##lation angola ##hra flint implications employs teens patrons stall weekends barriers scrambled nucleus tehran jenna parsons lifelong robots displacement 5000 ##bles precipitation ##gt knuckles clutched 1802 marrying ecology marx accusations declare scars kolkata mat meadows bermuda skeleton finalists vintage crawl coordinate affects subjected orchestral mistaken ##tc mirrors dipped relied 260 arches candle ##nick incorporating wildly fond basilica owl fringe rituals whispering stirred feud tertiary slick goat honorable whereby skip ricardo stripes parachute adjoining submerged synthesizer ##gren intend positively ninety phi beaver partition fellows alexis prohibition carlisle bizarre fraternity ##bre doubts icy cbc aquatic sneak sonny combines airports crude supervised spatial merge alfonso ##bic corrupt scan undergo ##ams disabilities colombian comparing dolphins perkins ##lish reprinted unanimous bounced hairs underworld midwest semester bucket paperback miniseries coventry demise ##leigh demonstrations sensor rotating yan ##hler arrange soils ##idge hyderabad labs ##dr brakes grandchildren ##nde negotiated rover ferrari continuation directorate augusta stevenson counterpart gore ##rda nursery rican ave collectively broadly pastoral repertoire asserted discovering nordic styled fiba cunningham harley middlesex survives tumor tempo zack aiming lok urgent ##rade ##nto devils ##ement contractor turin ##wl ##ool bliss repaired simmons moan astronomical cr negotiate lyric 1890s lara bred clad angus pbs ##ience engineered posed ##lk hernandez possessions elbows psychiatric strokes confluence electorate lifts campuses lava alps ##ep ##ution ##date physicist woody ##page ##ographic ##itis juliet reformation sparhawk 320 complement suppressed jewel ##½ floated ##kas continuity sadly ##ische inability melting scanning paula flour judaism safer vague ##lm solving curb ##stown financially gable bees expired miserable cassidy dominion 1789 cupped 145 robbery facto amos warden resume tallest marvin ing pounded usd declaring gasoline ##aux darkened 270 650 sophomore ##mere erection gossip televised risen dial ##eu pillars ##link passages profound ##tina arabian ashton silicon nail ##ead ##lated ##wer ##hardt fleming firearms ducked circuits blows waterloo titans ##lina atom fireplace cheshire financed activation algorithms ##zzi constituent catcher cherokee partnerships sexuality platoon tragic vivian guarded whiskey meditation poetic ##late ##nga ##ake porto listeners dominance kendra mona chandler factions 22nd salisbury attitudes derivative ##ido ##haus intake paced javier illustrator barrels bias cockpit burnett dreamed ensuing ##anda receptors someday hawkins mattered ##lal slavic 1799 jesuit cameroon wasted tai wax lowering victorious freaking outright hancock librarian sensing bald calcium myers tablet announcing barack shipyard pharmaceutical ##uan greenwich flush medley patches wolfgang pt speeches acquiring exams nikolai ##gg hayden kannada ##type reilly ##pt waitress abdomen devastated capped pseudonym pharmacy fulfill paraguay 1796 clicked ##trom archipelago syndicated ##hman lumber orgasm rejection clifford lorraine advent mafia rodney brock ##ght ##used ##elia cassette chamberlain despair mongolia sensors developmental upstream ##eg ##alis spanning 165 trombone basque seeded interred renewable rhys leapt revision molecule ##ages chord vicious nord shivered 23rd arlington debts corpus sunrise bays blackburn centimetres ##uded shuddered gm strangely gripping cartoons isabelle orbital ##ppa seals proving ##lton refusal strengthened bust assisting baghdad batsman portrayal mara pushes spears og ##cock reside nathaniel brennan 1776 confirmation caucus ##worthy markings yemen nobles ku lazy viewer catalan encompasses sawyer ##fall sparked substances patents braves arranger evacuation sergio persuade dover tolerance penguin cum jockey insufficient townships occupying declining plural processed projection puppet flanders introduces liability ##yon gymnastics antwerp taipei hobart candles jeep wes observers 126 chaplain bundle glorious ##hine hazel flung sol excavations dumped stares sh bangalore triangular icelandic intervals expressing turbine ##vers songwriting crafts ##igo jasmine ditch rite ##ways entertaining comply sorrow wrestlers basel emirates marian rivera helpful ##some caution downward networking ##atory ##tered darted genocide emergence replies specializing spokesman convenient unlocked fading augustine concentrations resemblance elijah investigator andhra ##uda promotes bean ##rrell fleeing wan simone announcer ##ame ##bby lydia weaver 132 residency modification ##fest stretches ##ast alternatively nat lowe lacks ##ented pam tile concealed inferior abdullah residences tissues vengeance ##ided moisture peculiar groove zip bologna jennings ninja oversaw zombies pumping batch livingston emerald installations 1797 peel nitrogen rama ##fying ##star schooling strands responding werner ##ost lime casa accurately targeting ##rod underway ##uru hemisphere lester ##yard occupies 2d griffith angrily reorganized ##owing courtney deposited ##dd ##30 estadio ##ifies dunn exiled ##ying checks ##combe ##о ##fly successes unexpectedly blu assessed ##flower ##ه observing sacked spiders kn ##tail mu nodes prosperity audrey divisional 155 broncos tangled adjust feeds erosion paolo surf directory snatched humid admiralty screwed gt reddish ##nese modules trench lamps bind leah bucks competes ##nz ##form transcription ##uc isles violently clutching pga cyclist inflation flats ragged unnecessary ##hian stubborn coordinated harriet baba disqualified 330 insect wolfe ##fies reinforcements rocked duel winked embraced bricks ##raj hiatus defeats pending brightly jealousy ##xton ##hm ##uki lena gdp colorful ##dley stein kidney ##shu underwear wanderers ##haw ##icus guardians m³ roared habits ##wise permits gp uranium punished disguise bundesliga elise dundee erotic partisan pi collectors float individually rendering behavioral bucharest ser hare valerie corporal nutrition proportional ##isa immense ##kis pavement ##zie ##eld sutherland crouched 1775 ##lp suzuki trades endurance operas crosby prayed priory rory socially ##urn gujarat ##pu walton cube pasha privilege lennon floods thorne waterfall nipple scouting approve ##lov minorities voter dwight extensions assure ballroom slap dripping privileges rejoined confessed demonstrating patriotic yell investor ##uth pagan slumped squares ##cle ##kins confront bert embarrassment ##aid aston urging sweater starr yuri brains williamson commuter mortar structured selfish exports ##jon cds ##him unfinished ##rre mortgage destinations ##nagar canoe solitary buchanan delays magistrate fk ##pling motivation ##lier ##vier recruiting assess ##mouth malik antique 1791 pius rahman reich tub zhou smashed airs galway xii conditioning honduras discharged dexter ##pf lionel 129 debates lemon tiffany volunteered dom dioxide procession devi sic tremendous advertisements colts transferring verdict hanover decommissioned utter relate pac racism ##top beacon limp similarity terra occurrence ant ##how becky capt updates armament richie pal ##graph halloween mayo ##ssen ##bone cara serena fcc dolls obligations ##dling violated lafayette jakarta exploitation ##ime infamous iconic ##lah ##park kitty moody reginald dread spill crystals olivier modeled bluff equilibrium separating notices ordnance extinction onset cosmic attachment sammy expose privy anchored ##bil abbott admits bending baritone emmanuel policeman vaughan winged climax dresses denny polytechnic mohamed burmese authentic nikki genetics grandparents homestead gaza postponed metacritic una ##sby ##bat unstable dissertation ##rial ##cian curls obscure uncovered bronx praying disappearing ##hoe prehistoric coke turret mutations nonprofit pits monaco ##ي ##usion prominently dispatched podium ##mir uci ##uation 133 fortifications birthplace kendall ##lby ##oll preacher rack goodman ##rman persistent ##ott countless jaime recorder lexington persecution jumps renewal wagons ##11 crushing ##holder decorations ##lake abundance wrath laundry £1 garde ##rp jeanne beetles peasant ##sl splitting caste sergei ##rer ##ema scripts ##ively rub satellites ##vor inscribed verlag scrapped gale packages chick potato slogan kathleen arabs ##culture counterparts reminiscent choral ##tead rand retains bushes dane accomplish courtesy closes ##oth slaughter hague krakow lawson tailed elias ginger ##ttes canopy betrayal rebuilding turf ##hof frowning allegiance brigades kicks rebuild polls alias nationalism td rowan audition bowie fortunately recognizes harp dillon horrified ##oro renault ##tics ropes ##α presumed rewarded infrared wiping accelerated illustration ##rid presses practitioners badminton ##iard detained ##tera recognizing relates misery ##sies ##tly reproduction piercing potatoes thornton esther manners hbo ##aan ours bullshit ernie perennial sensitivity illuminated rupert ##jin ##iss ##ear rfc nassau ##dock staggered socialism ##haven appointments nonsense prestige sharma haul ##tical solidarity gps ##ook ##rata igor pedestrian ##uit baxter tenants wires medication unlimited guiding impacts diabetes ##rama sasha pas clive extraction 131 continually constraints ##bilities sonata hunted sixteenth chu planting quote mayer pretended abs spat ##hua ceramic ##cci curtains pigs pitching ##dad latvian sore dayton ##sted ##qi patrols slice playground ##nted shone stool apparatus inadequate mates treason ##ija desires ##liga ##croft somalia laurent mir leonardo oracle grape obliged chevrolet thirteenth stunning enthusiastic ##ede accounted concludes currents basil ##kovic drought ##rica mai ##aire shove posting ##shed pilgrimage humorous packing fry pencil wines smells 144 marilyn aching newest clung bon neighbours sanctioned ##pie mug ##stock drowning ##mma hydraulic ##vil hiring reminder lilly investigators ##ncies sour ##eous compulsory packet ##rion ##graphic ##elle cannes ##inate depressed ##rit heroic importantly theresa ##tled conway saturn marginal rae ##xia corresponds royce pact jasper explosives packaging aluminium ##ttered denotes rhythmic spans assignments hereditary outlined originating sundays lad reissued greeting beatrice ##dic pillar marcos plots handbook alcoholic judiciary avant slides extract masculine blur ##eum ##force homage trembled owens hymn trey omega signaling socks accumulated reacted attic theo lining angie distraction primera talbot ##key 1200 ti creativity billed ##hey deacon eduardo identifies proposition dizzy gunner hogan ##yam ##pping ##hol ja ##chan jensen reconstructed ##berger clearance darius ##nier abe harlem plea dei circled emotionally notation fascist neville exceeded upwards viable ducks ##fo workforce racer limiting shri ##lson possesses 1600 kerr moths devastating laden disturbing locking ##cture gal fearing accreditation flavor aide 1870s mountainous ##baum melt ##ures motel texture servers soda ##mb herd ##nium erect puzzled hum peggy examinations gould testified geoff ren devised sacks ##law denial posters grunted cesar tutor ec gerry offerings byrne falcons combinations ct incoming pardon rocking 26th avengers flared mankind seller uttar loch nadia stroking exposing ##hd fertile ancestral instituted ##has noises prophecy taxation eminent vivid pol ##bol dart indirect multimedia notebook upside displaying adrenaline referenced geometric ##iving progression ##ddy blunt announce ##far implementing ##lav aggression liaison cooler cares headache plantations gorge dots impulse thickness ashamed averaging kathy obligation precursor 137 fowler symmetry thee 225 hears ##rai undergoing ads butcher bowler ##lip cigarettes subscription goodness ##ically browne ##hos ##tech kyoto donor ##erty damaging friction drifting expeditions hardened prostitution 152 fauna blankets claw tossing snarled butterflies recruits investigative coated healed 138 communal hai xiii academics boone psychologist restless lahore stephens mba brendan foreigners printer ##pc ached explode 27th deed scratched dared ##pole cardiac 1780 okinawa proto commando compelled oddly electrons ##base replica thanksgiving ##rist sheila deliberate stafford tidal representations hercules ou ##path ##iated kidnapping lenses ##tling deficit samoa mouths consuming computational maze granting smirk razor fixture ideals inviting aiden nominal ##vs issuing julio pitt ramsey docks ##oss exhaust ##owed bavarian draped anterior mating ethiopian explores noticing ##nton discarded convenience hoffman endowment beasts cartridge mormon paternal probe sleeves interfere lump deadline ##rail jenks bulldogs scrap alternating justified reproductive nam seize descending secretariat kirby coupe grouped smash panther sedan tapping ##18 lola cheer germanic unfortunate ##eter unrelated ##fan subordinate ##sdale suzanne advertisement ##ility horsepower ##lda cautiously discourse luigi ##mans ##fields noun prevalent mao schneider everett surround governorate kira ##avia westward ##take misty rails sustainability 134 unused ##rating packs toast unwilling regulate thy suffrage nile awe assam definitions travelers affordable ##rb conferred sells undefeated beneficial torso basal repeating remixes ##pass bahrain cables fang ##itated excavated numbering statutory ##rey deluxe ##lian forested ramirez derbyshire zeus slamming transfers astronomer banana lottery berg histories bamboo ##uchi resurrection posterior bowls vaguely ##thi thou preserving tensed offence ##inas meyrick callum ridden watt langdon tying lowland snorted daring truman ##hale ##girl aura overly filing weighing goa infections philanthropist saunders eponymous ##owski latitude perspectives reviewing mets commandant radial ##kha flashlight reliability koch vowels amazed ada elaine supper ##rth ##encies predator debated soviets cola ##boards ##nah compartment crooked arbitrary fourteenth ##ctive havana majors steelers clips profitable ambush exited packers ##tile nude cracks fungi ##е limb trousers josie shelby tens frederic ##ος definite smoothly constellation insult baton discs lingering ##nco conclusions lent staging becker grandpa shaky ##tron einstein obstacles sk adverse elle economically ##moto mccartney thor dismissal motions readings nostrils treatise ##pace squeezing evidently prolonged 1783 venezuelan je marguerite beirut takeover shareholders ##vent denise digit airplay norse ##bbling imaginary pills hubert blaze vacated eliminating ##ello vine mansfield ##tty retrospective barrow borne clutch bail forensic weaving ##nett ##witz desktop citadel promotions worrying dorset ieee subdivided ##iating manned expeditionary pickup synod chuckle 185 barney ##rz ##ffin functionality karachi litigation meanings uc lick turbo anders ##ffed execute curl oppose ankles typhoon ##د ##ache ##asia linguistics compassion pressures grazing perfection ##iting immunity monopoly muddy backgrounds 136 namibia francesca monitors attracting stunt tuition ##ии vegetable ##mates ##quent mgm jen complexes forts ##ond cellar bites seventeenth royals flemish failures mast charities ##cular peruvian capitals macmillan ipswich outward frigate postgraduate folds employing ##ouse concurrently fiery ##tai contingent nightmares monumental nicaragua ##kowski lizard mal fielding gig reject ##pad harding ##ipe coastline ##cin ##nos beethoven humphrey innovations ##tam ##nge norris doris solicitor huang obey 141 ##lc niagara ##tton shelves aug bourbon curry nightclub specifications hilton ##ndo centennial dispersed worm neglected briggs sm font kuala uneasy plc ##nstein ##bound ##aking ##burgh awaiting pronunciation ##bbed ##quest eh optimal zhu raped greens presided brenda worries ##life venetian marxist turnout ##lius refined braced sins grasped sunderland nickel speculated lowell cyrillic communism fundraising resembling colonists mutant freddie usc ##mos gratitude ##run mural ##lous chemist wi reminds 28th steals tess pietro ##ingen promoter ri microphone honoured rai sant ##qui feather ##nson burlington kurdish terrorists deborah sickness ##wed ##eet hazard irritated desperation veil clarity ##rik jewels xv ##gged ##ows ##cup berkshire unfair mysteries orchid winced exhaustion renovations stranded obe infinity ##nies adapt redevelopment thanked registry olga domingo noir tudor ole ##atus commenting behaviors ##ais crisp pauline probable stirling wigan ##bian paralympics panting surpassed ##rew luca barred pony famed ##sters cassandra waiter carolyn exported ##orted andres destructive deeds jonah castles vacancy suv ##glass 1788 orchard yep famine belarusian sprang ##forth skinny ##mis administrators rotterdam zambia zhao boiler discoveries ##ride ##physics lucius disappointing outreach spoon ##frame qualifications unanimously enjoys regency ##iidae stade realism veterinary rodgers dump alain chestnut castile censorship rumble gibbs ##itor communion reggae inactivated logs loads ##houses homosexual ##iano ale informs ##cas phrases plaster linebacker ambrose kaiser fascinated 850 limerick recruitment forge mastered ##nding leinster rooted threaten ##strom borneo ##hes suggestions scholarships propeller documentaries patronage coats constructing invest neurons comet entirety shouts identities annoying unchanged wary ##antly ##ogy neat oversight ##kos phillies replay constance ##kka incarnation humble skies minus ##acy smithsonian ##chel guerrilla jar cadets ##plate surplus audit ##aru cracking joanna louisa pacing ##lights intentionally ##iri diner nwa imprint australians tong unprecedented bunker naive specialists ark nichols railing leaked pedal ##uka shrub longing roofs v8 captains neural tuned ##ntal ##jet emission medina frantic codex definitive sid abolition intensified stocks enrique sustain genoa oxide ##written clues cha ##gers tributaries fragment venom ##rity ##ente ##sca muffled vain sire laos ##ingly ##hana hastily snapping surfaced sentiment motive ##oft contests approximate mesa luckily dinosaur exchanges propelled accord bourne relieve tow masks offended ##ues cynthia ##mmer rains bartender zinc reviewers lois ##sai legged arrogant rafe rosie comprise handicap blockade inlet lagoon copied drilling shelley petals ##inian mandarin obsolete ##inated onward arguably productivity cindy praising seldom busch discusses raleigh shortage ranged stanton encouragement firstly conceded overs temporal ##uke cbe ##bos woo certainty pumps ##pton stalked ##uli lizzie periodic thieves weaker ##night gases shoving chooses wc ##chemical prompting weights ##kill robust flanked sticky hu tuberculosis ##eb ##eal christchurch resembled wallet reese inappropriate pictured distract fixing fiddle giggled burger heirs hairy mechanic torque apache obsessed chiefly cheng logging ##tag extracted meaningful numb ##vsky gloucestershire reminding ##bay unite ##lit breeds diminished clown glove 1860s ##ن ##ug archibald focal freelance sliced depiction ##yk organism switches sights stray crawling ##ril lever leningrad interpretations loops anytime reel alicia delighted ##ech inhaled xiv suitcase bernie vega licenses northampton exclusion induction monasteries racecourse homosexuality ##right ##sfield ##rky dimitri michele alternatives ions commentators genuinely objected pork hospitality fencing stephan warships peripheral wit drunken wrinkled quentin spends departing chung numerical spokesperson ##zone johannesburg caliber killers ##udge assumes neatly demographic abigail bloc ##vel mounting ##lain bentley slightest xu recipients ##jk merlin ##writer seniors prisons blinking hindwings flickered kappa ##hel 80s strengthening appealing brewing gypsy mali lashes hulk unpleasant harassment bio treaties predict instrumentation pulp troupe boiling mantle ##ffe ins ##vn dividing handles verbs ##onal coconut senegal 340 thorough gum momentarily ##sto cocaine panicked destined ##turing teatro denying weary captained mans ##hawks ##code wakefield bollywood thankfully ##16 cyril ##wu amendments ##bahn consultation stud reflections kindness 1787 internally ##ovo tex mosaic distribute paddy seeming 143 ##hic piers ##15 ##mura ##verse popularly winger kang sentinel mccoy ##anza covenant ##bag verge fireworks suppress thrilled dominate ##jar swansea ##60 142 reconciliation ##ndi stiffened cue dorian ##uf damascus amor ida foremost ##aga porsche unseen dir ##had ##azi stony lexi melodies ##nko angular integer podcast ants inherent jaws justify persona ##olved josephine ##nr ##ressed customary flashes gala cyrus glaring backyard ariel physiology greenland html stir avon atletico finch methodology ked ##lent mas catholicism townsend branding quincy fits containers 1777 ashore aragon ##19 forearm poisoning ##sd adopting conquer grinding amnesty keller finances evaluate forged lankan instincts ##uto guam bosnian photographed workplace desirable protector ##dog allocation intently encourages willy ##sten bodyguard electro brighter ##ν bihar ##chev lasts opener amphibious sal verde arte ##cope captivity vocabulary yields ##tted agreeing desmond pioneered ##chus strap campaigned railroads ##ович emblem ##dre stormed 501 ##ulous marijuana northumberland ##gn ##nath bowen landmarks beaumont ##qua danube ##bler attorneys th ge flyers critique villains cass mutation acc ##0s colombo mckay motif sampling concluding syndicate ##rell neon stables ds warnings clint mourning wilkinson ##tated merrill leopard evenings exhaled emil sonia ezra discrete stove farrell fifteenth prescribed superhero ##rier worms helm wren ##duction ##hc expo ##rator hq unfamiliar antony prevents acceleration fiercely mari painfully calculations cheaper ign clifton irvine davenport mozambique ##np pierced ##evich wonders ##wig ##cate ##iling crusade ware ##uel enzymes reasonably mls ##coe mater ambition bunny eliot kernel ##fin asphalt headmaster torah aden lush pins waived ##care ##yas joao substrate enforce ##grad ##ules alvarez selections epidemic tempted ##bit bremen translates ensured waterfront 29th forrest manny malone kramer reigning cookies simpler absorption 205 engraved ##ffy evaluated 1778 haze 146 comforting crossover ##abe thorn ##rift ##imo ##pop suppression fatigue cutter ##tr 201 wurttemberg ##orf enforced hovering proprietary gb samurai syllable ascent lacey tick lars tractor merchandise rep bouncing defendants ##yre huntington ##ground ##oko standardized ##hor ##hima assassinated nu predecessors rainy liar assurance lyrical ##uga secondly flattened ios parameter undercover ##mity bordeaux punish ridges markers exodus inactive hesitate debbie nyc pledge savoy nagar offset organist ##tium hesse marin converting ##iver diagram propulsion pu validity reverted supportive ##dc ministries clans responds proclamation ##inae ##ø ##rea ein pleading patriot sf birch islanders strauss hates ##dh brandenburg concession rd ##ob 1900s killings textbook antiquity cinematography wharf embarrassing setup creed farmland inequality centred signatures fallon 370 ##ingham ##uts ceylon gazing directive laurie ##tern globally ##uated ##dent allah excavation threads ##cross 148 frantically icc utilize determines respiratory thoughtful receptions ##dicate merging chandra seine 147 builders builds diagnostic dev visibility goddamn analyses dhaka cho proves chancel concurrent curiously canadians pumped restoring 1850s turtles jaguar sinister spinal traction declan vows 1784 glowed capitalism swirling install universidad ##lder ##oat soloist ##genic ##oor coincidence beginnings nissan dip resorts caucasus combustion infectious ##eno pigeon serpent ##itating conclude masked salad jew ##gr surreal toni ##wc harmonica 151 ##gins ##etic ##coat fishermen intending bravery ##wave klaus titan wembley taiwanese ransom 40th incorrect hussein eyelids jp cooke dramas utilities ##etta ##print eisenhower principally granada lana ##rak openings concord ##bl bethany connie morality sega ##mons ##nard earnings ##kara ##cine wii communes ##rel coma composing softened severed grapes ##17 nguyen analyzed warlord hubbard heavenly behave slovenian ##hit ##ony hailed filmmakers trance caldwell skye unrest coward likelihood ##aging bern sci taliban honolulu propose ##wang 1700 browser imagining cobra contributes dukes instinctively conan violinist ##ores accessories gradual ##amp quotes sioux ##dating undertake intercepted sparkling compressed 139 fungus tombs haley imposing rests degradation lincolnshire retailers wetlands tulsa distributor dungeon nun greenhouse convey atlantis aft exits oman dresser lyons ##sti joking eddy judgement omitted digits ##cts ##game juniors ##rae cents stricken une ##ngo wizards weir breton nan technician fibers liking royalty ##cca 154 persia terribly magician ##rable ##unt vance cafeteria booker camille warmer ##static consume cavern gaps compass contemporaries foyer soothing graveyard maj plunged blush ##wear cascade demonstrates ordinance ##nov boyle ##lana rockefeller shaken banjo izzy ##ense breathless vines ##32 ##eman alterations chromosome dwellings feudal mole 153 catalonia relics tenant mandated ##fm fridge hats honesty patented raul heap cruisers accusing enlightenment infants wherein chatham contractors zen affinity hc osborne piston 156 traps maturity ##rana lagos ##zal peering ##nay attendant dealers protocols subset prospects biographical ##cre artery ##zers insignia nuns endured ##eration recommend schwartz serbs berger cromwell crossroads ##ctor enduring clasped grounded ##bine marseille twitched abel choke https catalyst moldova italians ##tist disastrous wee ##oured ##nti wwf nope ##piration ##asa expresses thumbs 167 ##nza coca 1781 cheating ##ption skipped sensory heidelberg spies satan dangers semifinal 202 bohemia whitish confusing shipbuilding relies surgeons landings ravi baku moor suffix alejandro ##yana litre upheld ##unk rajasthan ##rek coaster insists posture scenarios etienne favoured appoint transgender elephants poked greenwood defences fulfilled militant somali 1758 chalk potent ##ucci migrants wink assistants nos restriction activism niger ##ario colon shaun ##sat daphne ##erated swam congregations reprise considerations magnet playable xvi ##р overthrow tobias knob chavez coding ##mers propped katrina orient newcomer ##suke temperate ##pool farmhouse interrogation ##vd committing ##vert forthcoming strawberry joaquin macau ponds shocking siberia ##cellular chant contributors ##nant ##ologists sped absorb hail 1782 spared ##hore barbados karate opus originates saul ##xie evergreen leaped ##rock correlation exaggerated weekday unification bump tracing brig afb pathways utilizing ##ners mod mb disturbance kneeling ##stad ##guchi 100th pune ##thy decreasing 168 manipulation miriam academia ecosystem occupational rbi ##lem rift ##14 rotary stacked incorporation awakening generators guerrero racist ##omy cyber derivatives culminated allie annals panzer sainte wikipedia pops zu austro ##vate algerian politely nicholson mornings educate tastes thrill dartmouth ##gating db ##jee regan differing concentrating choreography divinity ##media pledged alexandre routing gregor madeline ##idal apocalypse ##hora gunfire culminating elves fined liang lam programmed tar guessing transparency gabrielle ##gna cancellation flexibility ##lining accession shea stronghold nets specializes ##rgan abused hasan sgt ling exceeding ##₄ admiration supermarket ##ark photographers specialised tilt resonance hmm perfume 380 sami threatens garland botany guarding boiled greet puppy russo supplier wilmington vibrant vijay ##bius paralympic grumbled paige faa licking margins hurricanes ##gong fest grenade ripping ##uz counseling weigh ##sian needles wiltshire edison costly ##not fulton tramway redesigned staffordshire cache gasping watkins sleepy candidacy ##group monkeys timeline throbbing ##bid ##sos berth uzbekistan vanderbilt bothering overturned ballots gem ##iger sunglasses subscribers hooker compelling ang exceptionally saloon stab ##rdi carla terrifying rom ##vision coil ##oids satisfying vendors 31st mackay deities overlooked ambient bahamas felipe olympia whirled botanist advertised tugging ##dden disciples morales unionist rites foley morse motives creepy ##₀ soo ##sz bargain highness frightening turnpike tory reorganization ##cer depict biographer ##walk unopposed manifesto ##gles institut emile accidental kapoor ##dam kilkenny cortex lively ##13 romanesque jain shan cannons ##ood ##ske petrol echoing amalgamated disappears cautious proposes sanctions trenton ##ر flotilla aus contempt tor canary cote theirs ##hun conceptual deleted fascinating paso blazing elf honourable hutchinson ##eiro ##outh ##zin surveyor tee amidst wooded reissue intro ##ono cobb shelters newsletter hanson brace encoding confiscated dem caravan marino scroll melodic cows imam ##adi ##aneous northward searches biodiversity cora 310 roaring ##bers connell theologian halo compose pathetic unmarried dynamo ##oot az calculation toulouse deserves humour nr forgiveness tam undergone martyr pamela myths whore counselor hicks 290 heavens battleship electromagnetic ##bbs stellar establishments presley hopped ##chin temptation 90s wills nas ##yuan nhs ##nya seminars ##yev adaptations gong asher lex indicator sikh tobago cites goin ##yte satirical ##gies characterised correspond bubbles lure participates ##vid eruption skate therapeutic 1785 canals wholesale defaulted sac 460 petit ##zzled virgil leak ravens 256 portraying ##yx ghetto creators dams portray vicente ##rington fae namesake bounty ##arium joachim ##ota ##iser aforementioned axle snout depended dismantled reuben 480 ##ibly gallagher ##lau ##pd earnest ##ieu ##iary inflicted objections ##llar asa gritted ##athy jericho ##sea ##was flick underside ceramics undead substituted 195 eastward undoubtedly wheeled chimney ##iche guinness cb ##ager siding ##bell traitor baptiste disguised inauguration 149 tipperary choreographer perched warmed stationary eco ##ike ##ntes bacterial ##aurus flores phosphate ##core attacker invaders alvin intersects a1 indirectly immigrated businessmen cornelius valves narrated pill sober ul nationale monastic applicants scenery ##jack 161 motifs constitutes cpu ##osh jurisdictions sd tuning irritation woven ##uddin fertility gao ##erie antagonist impatient glacial hides boarded denominations interception ##jas cookie nicola ##tee algebraic marquess bahn parole buyers bait turbines paperwork bestowed natasha renee oceans purchases 157 vaccine 215 ##tock fixtures playhouse integrate jai oswald intellectuals ##cky booked nests mortimer ##isi obsession sept ##gler ##sum 440 scrutiny simultaneous squinted ##shin collects oven shankar penned remarkably ##я slips luggage spectral 1786 collaborations louie consolidation ##ailed ##ivating 420 hoover blackpool harness ignition vest tails belmont mongol skinner ##nae visually mage derry ##tism ##unce stevie transitional ##rdy redskins drying prep prospective ##21 annoyance oversee ##loaded fills ##books ##iki announces fda scowled respects prasad mystic tucson ##vale revue springer bankrupt 1772 aristotle salvatore habsburg ##geny dal natal nut pod chewing darts moroccan walkover rosario lenin punjabi ##ße grossed scattering wired invasive hui polynomial corridors wakes gina portrays ##cratic arid retreating erich irwin sniper ##dha linen lindsey maneuver butch shutting socio bounce commemorative postseason jeremiah pines 275 mystical beads bp abbas furnace bidding consulted assaulted empirical rubble enclosure sob weakly cancel polly yielded ##emann curly prediction battered 70s vhs jacqueline render sails barked detailing grayson riga sloane raging ##yah herbs bravo ##athlon alloy giggle imminent suffers assumptions waltz ##itate accomplishments ##ited bathing remixed deception prefix ##emia deepest ##tier ##eis balkan frogs ##rong slab ##pate philosophers peterborough grains imports dickinson rwanda ##atics 1774 dirk lan tablets ##rove clone ##rice caretaker hostilities mclean ##gre regimental treasures norms impose tsar tango diplomacy variously complain 192 recognise arrests 1779 celestial pulitzer ##dus bing libretto ##moor adele splash ##rite expectation lds confronts ##izer spontaneous harmful wedge entrepreneurs buyer ##ope bilingual translate rugged conner circulated uae eaton ##gra ##zzle lingered lockheed vishnu reelection alonso ##oom joints yankee headline cooperate heinz laureate invading ##sford echoes scandinavian ##dham hugging vitamin salute micah hind trader ##sper radioactive ##ndra militants poisoned ratified remark campeonato deprived wander prop ##dong outlook ##tani ##rix ##eye chiang darcy ##oping mandolin spice statesman babylon 182 walled forgetting afro ##cap 158 giorgio buffer ##polis planetary ##gis overlap terminals kinda centenary ##bir arising manipulate elm ke 1770 ak ##tad chrysler mapped moose pomeranian quad macarthur assemblies shoreline recalls stratford ##rted noticeable ##evic imp ##rita ##sque accustomed supplying tents disgusted vogue sipped filters khz reno selecting luftwaffe mcmahon tyne masterpiece carriages collided dunes exercised flare remembers muzzle ##mobile heck ##rson burgess lunged middleton boycott bilateral ##sity hazardous lumpur multiplayer spotlight jackets goldman liege porcelain rag waterford benz attracts hopeful battling ottomans kensington baked hymns cheyenne lattice levine borrow polymer clashes michaels monitored commitments denounced ##25 ##von cavity ##oney hobby akin ##holders futures intricate cornish patty ##oned illegally dolphin ##lag barlow yellowish maddie apologized luton plagued ##puram nana ##rds sway fanny łodz ##rino psi suspicions hanged ##eding initiate charlton ##por nak competent 235 analytical annex wardrobe reservations ##rma sect 162 fairfax hedge piled buckingham uneven bauer simplicity snyder interpret accountability donors moderately byrd continents ##cite ##max disciple hr jamaican ping nominees ##uss mongolian diver attackers eagerly ideological pillows miracles apartheid revolver sulfur clinics moran 163 ##enko ile katy rhetoric ##icated chronology recycling ##hrer elongated mughal pascal profiles vibration databases domination ##fare ##rant matthias digest rehearsal polling weiss initiation reeves clinging flourished impress ngo ##hoff ##ume buckley symposium rhythms weed emphasize transforming ##taking ##gence ##yman accountant analyze flicker foil priesthood voluntarily decreases ##80 ##hya slater sv charting mcgill ##lde moreno ##iu besieged zur robes ##phic admitting api deported turmoil peyton earthquakes ##ares nationalists beau clair brethren interrupt welch curated galerie requesting 164 ##ested impending steward viper ##vina complaining beautifully brandy foam nl 1660 ##cake alessandro punches laced explanations ##lim attribute clit reggie discomfort ##cards smoothed whales ##cene adler countered duffy disciplinary widening recipe reliance conducts goats gradient preaching ##shaw matilda quasi striped meridian cannabis cordoba certificates ##agh ##tering graffiti hangs pilgrims repeats ##ych revive urine etat ##hawk fueled belts fuzzy susceptible ##hang mauritius salle sincere beers hooks ##cki arbitration entrusted advise sniffed seminar junk donnell processors principality strapped celia mendoza everton fortunes prejudice starving reassigned steamer ##lund tuck evenly foreman ##ffen dans 375 envisioned slit ##xy baseman liberia rosemary ##weed electrified periodically potassium stride contexts sperm slade mariners influx bianca subcommittee ##rane spilling icao estuary ##nock delivers iphone ##ulata isa mira bohemian dessert ##sbury welcoming proudly slowing ##chs musee ascension russ ##vian waits ##psy africans exploit ##morphic gov eccentric crab peck ##ull entrances formidable marketplace groom bolted metabolism patton robbins courier payload endure ##ifier andes refrigerator ##pr ornate ##uca ruthless illegitimate masonry strasbourg bikes adobe ##³ apples quintet willingly niche bakery corpses energetic ##cliffe ##sser ##ards 177 centimeters centro fuscous cretaceous rancho ##yde andrei telecom tottenham oasis ordination vulnerability presiding corey cp penguins sims ##pis malawi piss ##48 correction ##cked ##ffle ##ryn countdown detectives psychiatrist psychedelic dinosaurs blouse ##get choi vowed ##oz randomly ##pol 49ers scrub blanche bruins dusseldorf ##using unwanted ##ums 212 dominique elevations headlights om laguna ##oga 1750 famously ignorance shrewsbury ##aine ajax breuning che confederacy greco overhaul ##screen paz skirts disagreement cruelty jagged phoebe shifter hovered viruses ##wes mandy ##lined ##gc landlord squirrel dashed ##ι ornamental gag wally grange literal spurs undisclosed proceeding yin ##text billie orphan spanned humidity indy weighted presentations explosions lucian ##tary vaughn hindus ##anga ##hell psycho 171 daytona protects efficiently rematch sly tandem ##oya rebranded impaired hee metropolis peach godfrey diaspora ethnicity prosperous gleaming dar grossing playback ##rden stripe pistols ##tain births labelled ##cating 172 rudy alba ##onne aquarium hostility ##gb ##tase shudder sumatra hardest lakers consonant creeping demos homicide capsule zeke liberties expulsion pueblo ##comb trait transporting ##ddin ##neck ##yna depart gregg mold ledge hangar oldham playboy termination analysts gmbh romero ##itic insist cradle filthy brightness slash shootout deposed bordering ##truct isis microwave tumbled sheltered cathy werewolves messy andersen convex clapped clinched satire wasting edo vc rufus ##jak mont ##etti poznan ##keeping restructuring transverse ##rland azerbaijani slovene gestures roommate choking shear ##quist vanguard oblivious ##hiro disagreed baptism ##lich coliseum ##aceae salvage societe cory locke relocation relying versailles ahl swelling ##elo cheerful ##word ##edes gin sarajevo obstacle diverted ##nac messed thoroughbred fluttered utrecht chewed acquaintance assassins dispatch mirza ##wart nike salzburg swell yen ##gee idle ligue samson ##nds ##igh playful spawned ##cise tease ##case burgundy ##bot stirring skeptical interceptions marathi ##dies bedrooms aroused pinch ##lik preferences tattoos buster digitally projecting rust ##ital kitten priorities addison pseudo ##guard dusk icons sermon ##psis ##iba bt ##lift ##xt ju truce rink ##dah ##wy defects psychiatry offences calculate glucose ##iful ##rized ##unda francaise ##hari richest warwickshire carly 1763 purity redemption lending ##cious muse bruises cerebral aero carving ##name preface terminology invade monty ##int anarchist blurred ##iled rossi treats guts shu foothills ballads undertaking premise cecilia affiliates blasted conditional wilder minors drone rudolph buffy swallowing horton attested ##hop rutherford howell primetime livery penal ##bis minimize hydro wrecked wrought palazzo ##gling cans vernacular friedman nobleman shale walnut danielle ##ection ##tley sears ##kumar chords lend flipping streamed por dracula gallons sacrifices gamble orphanage ##iman mckenzie ##gible boxers daly ##balls ##ان 208 ##ific ##rative ##iq exploited slated ##uity circling hillary pinched goldberg provost campaigning lim piles ironically jong mohan successors usaf ##tem ##ught autobiographical haute preserves ##ending acquitted comparisons 203 hydroelectric gangs cypriot torpedoes rushes chrome derive bumps instability fiat pets ##mbe silas dye reckless settler ##itation info heats ##writing 176 canonical maltese fins mushroom stacy aspen avid ##kur ##loading vickers gaston hillside statutes wilde gail kung sabine comfortably motorcycles ##rgo 169 pneumonia fetch ##sonic axel faintly parallels ##oop mclaren spouse compton interdisciplinary miner ##eni 181 clamped ##chal ##llah separates versa ##mler scarborough labrador ##lity ##osing rutgers hurdles como 166 burt divers ##100 wichita cade coincided ##erson bruised mla ##pper vineyard ##ili ##brush notch mentioning jase hearted kits doe ##acle pomerania ##ady ronan seizure pavel problematic ##zaki domenico ##ulin catering penelope dependence parental emilio ministerial atkinson ##bolic clarkson chargers colby grill peeked arises summon ##aged fools ##grapher faculties qaeda ##vial garner refurbished ##hwa geelong disasters nudged bs shareholder lori algae reinstated rot ##ades ##nous invites stainless 183 inclusive ##itude diocesan til ##icz denomination ##xa benton floral registers ##ider ##erman ##kell absurd brunei guangzhou hitter retaliation ##uled ##eve blanc nh consistency contamination ##eres ##rner dire palermo broadcasters diaries inspire vols brewer tightening ky mixtape hormone ##tok stokes ##color ##dly ##ssi pg ##ometer ##lington sanitation ##tility intercontinental apps ##adt ¹⁄₂ cylinders economies favourable unison croix gertrude odyssey vanity dangling ##logists upgrades dice middleweight practitioner ##ight 206 henrik parlor orion angered lac python blurted ##rri sensual intends swings angled ##phs husky attain peerage precinct textiles cheltenham shuffled dai confess tasting bhutan ##riation tyrone segregation abrupt ruiz ##rish smirked blackwell confidential browning amounted ##put vase scarce fabulous raided staple guyana unemployed glider shay ##tow carmine troll intervene squash superstar ##uce cylindrical len roadway researched handy ##rium ##jana meta lao declares ##rring ##tadt ##elin ##kova willem shrubs napoleonic realms skater qi volkswagen ##ł tad hara archaeologist awkwardly eerie ##kind wiley ##heimer ##24 titus organizers cfl crusaders lama usb vent enraged thankful occupants maximilian ##gaard possessing textbooks ##oran collaborator quaker ##ulo avalanche mono silky straits isaiah mustang surged resolutions potomac descend cl kilograms plato strains saturdays ##olin bernstein ##ype holstein ponytail ##watch belize conversely heroine perpetual ##ylus charcoal piedmont glee negotiating backdrop prologue ##jah ##mmy pasadena climbs ramos sunni ##holm ##tner ##tri anand deficiency hertfordshire stout ##avi aperture orioles ##irs doncaster intrigued bombed coating otis ##mat cocktail ##jit ##eto amir arousal sar ##proof ##act ##ories dixie pots ##bow whereabouts 159 ##fted drains bullying cottages scripture coherent fore poe appetite ##uration sampled ##ators ##dp derrick rotor jays peacock installment ##rro advisors ##coming rodeo scotch ##mot ##db ##fen ##vant ensued rodrigo dictatorship martyrs twenties ##н towed incidence marta rainforest sai scaled ##cles oceanic qualifiers symphonic mcbride dislike generalized aubrey colonization ##iation ##lion ##ssing disliked lublin salesman ##ulates spherical whatsoever sweating avalon contention punt severity alderman atari ##dina ##grant ##rop scarf seville vertices annexation fairfield fascination inspiring launches palatinate regretted ##rca feral ##iom elk nap olsen reddy yong ##leader ##iae garment transports feng gracie outrage viceroy insides ##esis breakup grady organizer softer grimaced 222 murals galicia arranging vectors ##rsten bas ##sb ##cens sloan ##eka bitten ara fender nausea bumped kris banquet comrades detector persisted ##llan adjustment endowed cinemas ##shot sellers ##uman peek epa kindly neglect simpsons talon mausoleum runaway hangul lookout ##cic rewards coughed acquainted chloride ##ald quicker accordion neolithic ##qa artemis coefficient lenny pandora tx ##xed ecstasy litter segunda chairperson gemma hiss rumor vow nasal antioch compensate patiently transformers ##eded judo morrow penis posthumous philips bandits husbands denote flaming ##any ##phones langley yorker 1760 walters ##uo ##kle gubernatorial fatty samsung leroy outlaw ##nine unpublished poole jakob ##ᵢ ##ₙ crete distorted superiority ##dhi intercept crust mig claus crashes positioning 188 stallion 301 frontal armistice ##estinal elton aj encompassing camel commemorated malaria woodward calf cigar penetrate ##oso willard ##rno ##uche illustrate amusing convergence noteworthy ##lma ##rva journeys realise manfred ##sable 410 ##vocation hearings fiance ##posed educators provoked adjusting ##cturing modular stockton paterson vlad rejects electors selena maureen ##tres uber ##rce swirled ##num proportions nanny pawn naturalist parma apostles awoke ethel wen ##bey monsoon overview ##inating mccain rendition risky adorned ##ih equestrian germain nj conspicuous confirming ##yoshi shivering ##imeter milestone rumours flinched bounds smacked token ##bei lectured automobiles ##shore impacted ##iable nouns nero ##leaf ismail prostitute trams ##lace bridget sud stimulus impressions reins revolves ##oud ##gned giro honeymoon ##swell criterion ##sms ##uil libyan prefers ##osition 211 preview sucks accusation bursts metaphor diffusion tolerate faye betting cinematographer liturgical specials bitterly humboldt ##ckle flux rattled ##itzer archaeologists odor authorised marshes discretion ##ов alarmed archaic inverse ##leton explorers ##pine drummond tsunami woodlands ##minate ##tland booklet insanity owning insert crafted calculus ##tore receivers ##bt stung ##eca ##nched prevailing travellers eyeing lila graphs ##borne 178 julien ##won morale adaptive therapist erica cw libertarian bowman pitches vita ##ional crook ##ads ##entation caledonia mutiny ##sible 1840s automation ##ß flock ##pia ironic pathology ##imus remarried ##22 joker withstand energies ##att shropshire hostages madeleine tentatively conflicting mateo recipes euros ol mercenaries nico ##ndon albuquerque augmented mythical bel freud ##child cough ##lica 365 freddy lillian genetically nuremberg calder 209 bonn outdoors paste suns urgency vin restraint tyson ##cera ##selle barrage bethlehem kahn ##par mounts nippon barony happier ryu makeshift sheldon blushed castillo barking listener taped bethel fluent headlines pornography rum disclosure sighing mace doubling gunther manly ##plex rt interventions physiological forwards emerges ##tooth ##gny compliment rib recession visibly barge faults connector exquisite prefect ##rlin patio ##cured elevators brandt italics pena 173 wasp satin ea botswana graceful respectable ##jima ##rter ##oic franciscan generates ##dl alfredo disgusting ##olate ##iously sherwood warns cod promo cheryl sino ##ة ##escu twitch ##zhi brownish thom ortiz ##dron densely ##beat carmel reinforce ##bana 187 anastasia downhill vertex contaminated remembrance harmonic homework ##sol fiancee gears olds angelica loft ramsay quiz colliery sevens ##cape autism ##hil walkway ##boats ruben abnormal ounce khmer ##bbe zachary bedside morphology punching ##olar sparrow convinces ##35 hewitt queer remastered rods mabel solemn notified lyricist symmetric ##xide 174 encore passports wildcats ##uni baja ##pac mildly ##ease bleed commodity mounds glossy orchestras ##omo damian prelude ambitions ##vet awhile remotely ##aud asserts imply ##iques distinctly modelling remedy ##dded windshield dani xiao ##endra audible powerplant 1300 invalid elemental acquisitions ##hala immaculate libby plata smuggling ventilation denoted minh ##morphism 430 differed dion kelley lore mocking sabbath spikes hygiene drown runoff stylized tally liberated aux interpreter righteous aba siren reaper pearce millie ##cier ##yra gaius ##iso captures ##ttering dorm claudio ##sic benches knighted blackness ##ored discount fumble oxidation routed ##ς novak perpendicular spoiled fracture splits ##urt pads topology ##cats axes fortunate offenders protestants esteem 221 broadband convened frankly hound prototypes isil facilitated keel ##sher sahara awaited bubba orb prosecutors 186 hem 520 ##xing relaxing remnant romney sorted slalom stefano ulrich ##active exemption folder pauses foliage hitchcock epithet 204 criticisms ##aca ballistic brody hinduism chaotic youths equals ##pala pts thicker analogous capitalist improvised overseeing sinatra ascended beverage ##tl straightforward ##kon curran ##west bois 325 induce surveying emperors sax unpopular ##kk cartoonist fused ##mble unto ##yuki localities ##cko ##ln darlington slain academie lobbying sediment puzzles ##grass defiance dickens manifest tongues alumnus arbor coincide 184 appalachian mustafa examiner cabaret traumatic yves bracelet draining heroin magnum baths odessa consonants mitsubishi ##gua kellan vaudeville ##fr joked null straps probation ##ław ceded interfaces ##pas ##zawa blinding viet 224 rothschild museo 640 huddersfield ##vr tactic ##storm brackets dazed incorrectly ##vu reg glazed fearful manifold benefited irony ##sun stumbling ##rte willingness balkans mei wraps ##aba injected ##lea gu syed harmless ##hammer bray takeoff poppy timor cardboard astronaut purdue weeping southbound cursing stalls diagonal ##neer lamar bryce comte weekdays harrington ##uba negatively ##see lays grouping ##cken ##henko affirmed halle modernist ##lai hodges smelling aristocratic baptized dismiss justification oilers ##now coupling qin snack healer ##qing gardener layla battled formulated stephenson gravitational ##gill ##jun 1768 granny coordinating suites ##cd ##ioned monarchs ##cote ##hips sep blended apr barrister deposition fia mina policemen paranoid ##pressed churchyard covert crumpled creep abandoning tr transmit conceal barr understands readiness spire ##cology ##enia ##erry 610 startling unlock vida bowled slots ##nat ##islav spaced trusting admire rig ##ink slack ##70 mv 207 casualty ##wei classmates ##odes ##rar ##rked amherst furnished evolve foundry menace mead ##lein flu wesleyan ##kled monterey webber ##vos wil ##mith ##на bartholomew justices restrained ##cke amenities 191 mediated sewage trenches ml mainz ##thus 1800s ##cula ##inski caine bonding 213 converts spheres superseded marianne crypt sweaty ensign historia ##br spruce ##post ##ask forks thoughtfully yukon pamphlet ames ##uter karma ##yya bryn negotiation sighs incapable ##mbre ##ntial actresses taft ##mill luce prevailed ##amine 1773 motionless envoy testify investing sculpted instructors provence kali cullen horseback ##while goodwin ##jos gaa norte ##ldon modify wavelength abd 214 skinned sprinter forecast scheduling marries squared tentative ##chman boer ##isch bolts swap fisherman assyrian impatiently guthrie martins murdoch 194 tanya nicely dolly lacy med ##45 syn decks fashionable millionaire ##ust surfing ##ml ##ision heaved tammy consulate attendees routinely 197 fuse saxophonist backseat malaya ##lord scowl tau ##ishly 193 sighted steaming ##rks 303 911 ##holes ##hong ching ##wife bless conserved jurassic stacey unix zion chunk rigorous blaine 198 peabody slayer dismay brewers nz ##jer det ##glia glover postwar int penetration sylvester imitation vertically airlift heiress knoxville viva ##uin 390 macon ##rim ##fighter ##gonal janice ##orescence ##wari marius belongings leicestershire 196 blanco inverted preseason sanity sobbing ##due ##elt ##dled collingwood regeneration flickering shortest ##mount ##osi feminism ##lat sherlock cabinets fumbled northbound precedent snaps ##mme researching ##akes guillaume insights manipulated vapor neighbour sap gangster frey f1 stalking scarcely callie barnett tendencies audi doomed assessing slung panchayat ambiguous bartlett ##etto distributing violating wolverhampton ##hetic swami histoire ##urus liable pounder groin hussain larsen popping surprises ##atter vie curt ##station mute relocate musicals authorization richter ##sef immortality tna bombings ##press deteriorated yiddish ##acious robbed colchester cs pmid ao verified balancing apostle swayed recognizable oxfordshire retention nottinghamshire contender judd invitational shrimp uhf ##icient cleaner longitudinal tanker ##mur acronym broker koppen sundance suppliers ##gil 4000 clipped fuels petite ##anne landslide helene diversion populous landowners auspices melville quantitative ##xes ferries nicky ##llus doo haunting roche carver downed unavailable ##pathy approximation hiroshima ##hue garfield valle comparatively keyboardist traveler ##eit congestion calculating subsidiaries ##bate serb modernization fairies deepened ville averages ##lore inflammatory tonga ##itch co₂ squads ##hea gigantic serum enjoyment retailer verona 35th cis ##phobic magna technicians ##vati arithmetic ##sport levin ##dation amtrak chow sienna ##eyer backstage entrepreneurship ##otic learnt tao ##udy worcestershire formulation baggage hesitant bali sabotage ##kari barren enhancing murmur pl freshly putnam syntax aces medicines resentment bandwidth ##sier grins chili guido ##sei framing implying gareth lissa genevieve pertaining admissions geo thorpe proliferation sato bela analyzing parting ##gor awakened ##isman huddled secrecy ##kling hush gentry 540 dungeons ##ego coasts ##utz sacrificed ##chule landowner mutually prevalence programmer adolescent disrupted seaside gee trusts vamp georgie ##nesian ##iol schedules sindh ##market etched hm sparse bey beaux scratching gliding unidentified 216 collaborating gems jesuits oro accumulation shaping mbe anal ##xin 231 enthusiasts newscast ##egan janata dewey parkinson 179 ankara biennial towering dd inconsistent 950 ##chet thriving terminate cabins furiously eats advocating donkey marley muster phyllis leiden ##user grassland glittering iucn loneliness 217 memorandum armenians ##ddle popularized rhodesia 60s lame ##illon sans bikini header orbits ##xx ##finger ##ulator sharif spines biotechnology strolled naughty yates ##wire fremantle milo ##mour abducted removes ##atin humming wonderland ##chrome ##ester hume pivotal ##rates armand grams believers elector rte apron bis scraped ##yria endorsement initials ##llation eps dotted hints buzzing emigration nearer ##tom indicators ##ulu coarse neutron protectorate ##uze directional exploits pains loire 1830s proponents guggenheim rabbits ritchie 305 hectare inputs hutton ##raz verify ##ako boilers longitude ##lev skeletal yer emilia citrus compromised ##gau pokemon prescription paragraph eduard cadillac attire categorized kenyan weddings charley ##bourg entertain monmouth ##lles nutrients davey mesh incentive practised ecosystems kemp subdued overheard ##rya bodily maxim ##nius apprenticeship ursula ##fight lodged rug silesian unconstitutional patel inspected coyote unbeaten ##hak 34th disruption convict parcel ##cl ##nham collier implicated mallory ##iac ##lab susannah winkler ##rber shia phelps sediments graphical robotic ##sner adulthood mart smoked ##isto kathryn clarified ##aran divides convictions oppression pausing burying ##mt federico mathias eileen ##tana kite hunched ##acies 189 ##atz disadvantage liza kinetic greedy paradox yokohama dowager trunks ventured ##gement gupta vilnius olaf ##thest crimean hopper ##ej progressively arturo mouthed arrondissement ##fusion rubin simulcast oceania ##orum ##stra ##rred busiest intensely navigator cary ##vine ##hini ##bies fife rowe rowland posing insurgents shafts lawsuits activate conor inward culturally garlic 265 ##eering eclectic ##hui ##kee ##nl furrowed vargas meteorological rendezvous ##aus culinary commencement ##dition quota ##notes mommy salaries overlapping mule ##iology ##mology sums wentworth ##isk ##zione mainline subgroup ##illy hack plaintiff verdi bulb differentiation engagements multinational supplemented bertrand caller regis ##naire ##sler ##arts ##imated blossom propagation kilometer viaduct vineyards ##uate beckett optimization golfer songwriters seminal semitic thud volatile evolving ridley ##wley trivial distributions scandinavia jiang ##ject wrestled insistence ##dio emphasizes napkin ##ods adjunct rhyme ##ricted ##eti hopeless surrounds tremble 32nd smoky ##ntly oils medicinal padded steer wilkes 219 255 concessions hue uniquely blinded landon yahoo ##lane hendrix commemorating dex specify chicks ##ggio intercity 1400 morley ##torm highlighting ##oting pang oblique stalled ##liner flirting newborn 1769 bishopric shaved 232 currie ##ush dharma spartan ##ooped favorites smug novella sirens abusive creations espana ##lage paradigm semiconductor sheen ##rdo ##yen ##zak nrl renew ##pose ##tur adjutant marches norma ##enity ineffective weimar grunt ##gat lordship plotting expenditure infringement lbs refrain av mimi mistakenly postmaster 1771 ##bara ras motorsports tito 199 subjective ##zza bully stew ##kaya prescott 1a ##raphic ##zam bids styling paranormal reeve sneaking exploding katz akbar migrant syllables indefinitely ##ogical destroys replaces applause ##phine pest ##fide 218 articulated bertie ##thing ##cars ##ptic courtroom crowley aesthetics cummings tehsil hormones titanic dangerously ##ibe stadion jaenelle auguste ciudad ##chu mysore partisans ##sio lucan philipp ##aly debating henley interiors ##rano ##tious homecoming beyonce usher henrietta prepares weeds ##oman ely plucked ##pire ##dable luxurious ##aq artifact password pasture juno maddy minsk ##dder ##ologies ##rone assessments martian royalist 1765 examines ##mani ##rge nino 223 parry scooped relativity ##eli ##uting ##cao congregational noisy traverse ##agawa strikeouts nickelodeon obituary transylvania binds depictions polk trolley ##yed ##lard breeders ##under dryly hokkaido 1762 strengths stacks bonaparte connectivity neared prostitutes stamped anaheim gutierrez sinai ##zzling bram fresno madhya ##86 proton ##lena ##llum ##phon reelected wanda ##anus ##lb ample distinguishing ##yler grasping sermons tomato bland stimulation avenues ##eux spreads scarlett fern pentagon assert baird chesapeake ir calmed distortion fatalities ##olis correctional pricing ##astic ##gina prom dammit ying collaborate ##chia welterweight 33rd pointer substitution bonded umpire communicating multitude paddle ##obe federally intimacy ##insky betray ssr ##lett ##lean ##lves ##therapy airbus ##tery functioned ud bearer biomedical netflix ##hire ##nca condom brink ik ##nical macy ##bet flap gma experimented jelly lavender ##icles ##ulia munro ##mian ##tial rye ##rle 60th gigs hottest rotated predictions fuji bu ##erence ##omi barangay ##fulness ##sas clocks ##rwood ##liness cereal roe wight decker uttered babu onion xml forcibly ##df petra sarcasm hartley peeled storytelling ##42 ##xley ##ysis ##ffa fibre kiel auditor fig harald greenville ##berries geographically nell quartz ##athic cemeteries ##lr crossings nah holloway reptiles chun sichuan snowy 660 corrections ##ivo zheng ambassadors blacksmith fielded fluids hardcover turnover medications melvin academies ##erton ro roach absorbing spaniards colton ##founded outsider espionage kelsey 245 edible ##ulf dora establishes ##sham ##tries contracting ##tania cinematic costello nesting ##uron connolly duff ##nology mma ##mata fergus sexes gi optics spectator woodstock banning ##hee ##fle differentiate outfielder refinery 226 312 gerhard horde lair drastically ##udi landfall ##cheng motorsport odi ##achi predominant quay skins ##ental edna harshly complementary murdering ##aves wreckage ##90 ono outstretched lennox munitions galen reconcile 470 scalp bicycles gillespie questionable rosenberg guillermo hostel jarvis kabul volvo opium yd ##twined abuses decca outpost ##cino sensible neutrality ##64 ponce anchorage atkins turrets inadvertently disagree libre vodka reassuring weighs ##yal glide jumper ceilings repertory outs stain ##bial envy ##ucible smashing heightened policing hyun mixes lai prima ##ples celeste ##bina lucrative intervened kc manually ##rned stature staffed bun bastards nairobi priced ##auer thatcher ##kia tripped comune ##ogan ##pled brasil incentives emanuel hereford musica ##kim benedictine biennale ##lani eureka gardiner rb knocks sha ##ael ##elled ##onate efficacy ventura masonic sanford maize leverage ##feit capacities santana ##aur novelty vanilla ##cter ##tour benin ##oir ##rain neptune drafting tallinn ##cable humiliation ##boarding schleswig fabian bernardo liturgy spectacle sweeney pont routledge ##tment cosmos ut hilt sleek universally ##eville ##gawa typed ##dry favors allegheny glaciers ##rly recalling aziz ##log parasite requiem auf ##berto ##llin illumination ##breaker ##issa festivities bows govern vibe vp 333 sprawled larson pilgrim bwf leaping ##rts ##ssel alexei greyhound hoarse ##dler ##oration seneca ##cule gaping ##ulously ##pura cinnamon ##gens ##rricular craven fantasies houghton engined reigned dictator supervising ##oris bogota commentaries unnatural fingernails spirituality tighten ##tm canadiens protesting intentional cheers sparta ##ytic ##iere ##zine widen belgarath controllers dodd iaaf navarre ##ication defect squire steiner whisky ##mins 560 inevitably tome ##gold chew ##uid ##lid elastic ##aby streaked alliances jailed regal ##ined ##phy czechoslovak narration absently ##uld bluegrass guangdong quran criticizing hose hari ##liest ##owa skier streaks deploy ##lom raft bose dialed huff ##eira haifa simplest bursting endings ib sultanate ##titled franks whitman ensures sven ##ggs collaborators forster organising ui banished napier injustice teller layered thump ##otti roc battleships evidenced fugitive sadie robotics ##roud equatorial geologist ##iza yielding ##bron ##sr internationale mecca ##diment sbs skyline toad uploaded reflective undrafted lal leafs bayern ##dai lakshmi shortlisted ##stick ##wicz camouflage donate af christi lau ##acio disclosed nemesis 1761 assemble straining northamptonshire tal ##asi bernardino premature heidi 42nd coefficients galactic reproduce buzzed sensations zionist monsieur myrtle ##eme archery strangled musically viewpoint antiquities bei trailers seahawks cured pee preferring tasmanian lange sul ##mail ##working colder overland lucivar massey gatherings haitian ##smith disapproval flaws ##cco ##enbach 1766 npr ##icular boroughs creole forums techno 1755 dent abdominal streetcar ##eson ##stream procurement gemini predictable ##tya acheron christoph feeder fronts vendor bernhard jammu tumors slang ##uber goaltender twists curving manson vuelta mer peanut confessions pouch unpredictable allowance theodor vascular ##factory bala authenticity metabolic coughing nanjing ##cea pembroke ##bard splendid 36th ff hourly ##ahu elmer handel ##ivate awarding thrusting dl experimentation ##hesion ##46 caressed entertained steak ##rangle biologist orphans baroness oyster stepfather ##dridge mirage reefs speeding ##31 barons 1764 227 inhabit preached repealed ##tral honoring boogie captives administer johanna ##imate gel suspiciously 1767 sobs ##dington backbone hayward garry ##folding ##nesia maxi ##oof ##ppe ellison galileo ##stand crimea frenzy amour bumper matrices natalia baking garth palestinians ##grove smack conveyed ensembles gardening ##manship ##rup ##stituting 1640 harvesting topography jing shifters dormitory ##carriage ##lston ist skulls ##stadt dolores jewellery sarawak ##wai ##zier fences christy confinement tumbling credibility fir stench ##bria ##plication ##nged ##sam virtues ##belt marjorie pba ##eem ##made celebrates schooner agitated barley fulfilling anthropologist ##pro restrict novi regulating ##nent padres ##rani ##hesive loyola tabitha milky olson proprietor crambidae guarantees intercollegiate ljubljana hilda ##sko ignorant hooded ##lts sardinia ##lidae ##vation frontman privileged witchcraft ##gp jammed laude poking ##than bracket amazement yunnan ##erus maharaja linnaeus 264 commissioning milano peacefully ##logies akira rani regulator ##36 grasses ##rance luzon crows compiler gretchen seaman edouard tab buccaneers ellington hamlets whig socialists ##anto directorial easton mythological ##kr ##vary rhineland semantic taut dune inventions succeeds ##iter replication branched ##pired jul prosecuted kangaroo penetrated ##avian middlesbrough doses bleak madam predatory relentless ##vili reluctance ##vir hailey crore silvery 1759 monstrous swimmers transmissions hawthorn informing ##eral toilets caracas crouch kb ##sett 295 cartel hadley ##aling alexia yvonne ##biology cinderella eton superb blizzard stabbing industrialist maximus ##gm ##orus groves maud clade oversized comedic ##bella rosen nomadic fulham montane beverages galaxies redundant swarm ##rot ##folia ##llis buckinghamshire fen bearings bahadur ##rom gilles phased dynamite faber benoit vip ##ount ##wd booking fractured tailored anya spices westwood cairns auditions inflammation steamed ##rocity ##acion ##urne skyla thereof watford torment archdeacon transforms lulu demeanor fucked serge ##sor mckenna minas entertainer ##icide caress originate residue ##sty 1740 ##ilised ##org beech ##wana subsidies ##ghton emptied gladstone ru firefighters voodoo ##rcle het nightingale tamara edmond ingredient weaknesses silhouette 285 compatibility withdrawing hampson ##mona anguish giggling ##mber bookstore ##jiang southernmost tilting ##vance bai economical rf briefcase dreadful hinted projections shattering totaling ##rogate analogue indicted periodical fullback ##dman haynes ##tenberg ##ffs ##ishment 1745 thirst stumble penang vigorous ##ddling ##kor ##lium octave ##ove ##enstein ##inen ##ones siberian ##uti cbn repeal swaying ##vington khalid tanaka unicorn otago plastered lobe riddle ##rella perch ##ishing croydon filtered graeme tripoli ##ossa crocodile ##chers sufi mined ##tung inferno lsu ##phi swelled utilizes £2 cale periodicals styx hike informally coop lund ##tidae ala hen qui transformations disposed sheath chickens ##cade fitzroy sas silesia unacceptable odisha 1650 sabrina pe spokane ratios athena massage shen dilemma ##drum ##riz ##hul corona doubtful niall ##pha ##bino fines cite acknowledging bangor ballard bathurst ##resh huron mustered alzheimer garments kinase tyre warship ##cp flashback pulmonary braun cheat kamal cyclists constructions grenades ndp traveller excuses stomped signalling trimmed futsal mosques relevance ##wine wta ##23 ##vah ##lter hoc ##riding optimistic ##´s deco sim interacting rejecting moniker waterways ##ieri ##oku mayors gdansk outnumbered pearls ##ended ##hampton fairs totals dominating 262 notions stairway compiling pursed commodities grease yeast ##jong carthage griffiths residual amc contraction laird sapphire ##marine ##ivated amalgamation dissolve inclination lyle packaged altitudes suez canons graded lurched narrowing boasts guise wed enrico ##ovsky rower scarred bree cub iberian protagonists bargaining proposing trainers voyages vans fishes ##aea ##ivist ##verance encryption artworks kazan sabre cleopatra hepburn rotting supremacy mecklenburg ##brate burrows hazards outgoing flair organizes ##ctions scorpion ##usions boo 234 chevalier dunedin slapping ##34 ineligible pensions ##38 ##omic manufactures emails bismarck 238 weakening blackish ding mcgee quo ##rling northernmost xx manpower greed sampson clicking ##ange ##horpe ##inations ##roving torre ##eptive ##moral symbolism 38th asshole meritorious outfits splashed biographies sprung astros ##tale 302 737 filly raoul nw tokugawa linden clubhouse ##apa tracts romano ##pio putin tags ##note chained dickson gunshot moe gunn rashid ##tails zipper ##bas ##nea contrasted ##ply ##udes plum pharaoh ##pile aw comedies ingrid sandwiches subdivisions 1100 mariana nokia kamen hz delaney veto herring ##words possessive outlines ##roup siemens stairwell rc gallantry messiah palais yells 233 zeppelin ##dm bolivar ##cede smackdown mckinley ##mora ##yt muted geologic finely unitary avatar hamas maynard rees bog contrasting ##rut liv chico disposition pixel ##erate becca dmitry yeshiva narratives ##lva ##ulton mercenary sharpe tempered navigate stealth amassed keynes ##lini untouched ##rrie havoc lithium ##fighting abyss graf southward wolverine balloons implements ngos transitions ##icum ambushed concacaf dormant economists ##dim costing csi rana universite boulders verity ##llon collin mellon misses cypress fluorescent lifeless spence ##ulla crewe shepard pak revelations ##م jolly gibbons paw ##dro ##quel freeing ##test shack fries palatine ##51 ##hiko accompaniment cruising recycled ##aver erwin sorting synthesizers dyke realities sg strides enslaved wetland ##ghan competence gunpowder grassy maroon reactors objection ##oms carlson gearbox macintosh radios shelton ##sho clergyman prakash 254 mongols trophies oricon 228 stimuli twenty20 cantonese cortes mirrored ##saurus bhp cristina melancholy ##lating enjoyable nuevo ##wny downfall schumacher ##ind banging lausanne rumbled paramilitary reflex ax amplitude migratory ##gall ##ups midi barnard lastly sherry ##hp ##nall keystone ##kra carleton slippery ##53 coloring foe socket otter ##rgos mats ##tose consultants bafta bison topping ##km 490 primal abandonment transplant atoll hideous mort pained reproduced tae howling ##turn unlawful billionaire hotter poised lansing ##chang dinamo retro messing nfc domesday ##mina blitz timed ##athing ##kley ascending gesturing ##izations signaled tis chinatown mermaid savanna jameson ##aint catalina ##pet ##hers cochrane cy chatting ##kus alerted computation mused noelle majestic mohawk campo octagonal ##sant ##hend 241 aspiring ##mart comprehend iona paralyzed shimmering swindon rhone ##eley reputed configurations pitchfork agitation francais gillian lipstick ##ilo outsiders pontifical resisting bitterness sewer rockies ##edd ##ucher misleading 1756 exiting galloway ##nging risked ##heart 246 commemoration schultz ##rka integrating ##rsa poses shrieked ##weiler guineas gladys jerking owls goldsmith nightly penetrating ##unced lia ##33 ignited betsy ##aring ##thorpe follower vigorously ##rave coded kiran knit zoology tbilisi ##28 ##bered repository govt deciduous dino growling ##bba enhancement unleashed chanting pussy biochemistry ##eric kettle repression toxicity nrhp ##arth ##kko ##bush ernesto commended outspoken 242 mca parchment sms kristen ##aton bisexual raked glamour navajo a2 conditioned showcased ##hma spacious youthful ##esa usl appliances junta brest layne conglomerate enchanted chao loosened picasso circulating inspect montevideo ##centric ##kti piazza spurred ##aith bari freedoms poultry stamford lieu ##ect indigo sarcastic bahia stump attach dvds frankenstein lille approx scriptures pollen ##script nmi overseen ##ivism tides proponent newmarket inherit milling ##erland centralized ##rou distributors credentials drawers abbreviation ##lco ##xon downing uncomfortably ripe ##oes erase franchises ##ever populace ##bery ##khar decomposition pleas ##tet daryl sabah ##stle ##wide fearless genie lesions annette ##ogist oboe appendix nair dripped petitioned maclean mosquito parrot rpg hampered 1648 operatic reservoirs ##tham irrelevant jolt summarized ##fp medallion ##taff ##− clawed harlow narrower goddard marcia bodied fremont suarez altering tempest mussolini porn ##isms sweetly oversees walkers solitude grimly shrines hk ich supervisors hostess dietrich legitimacy brushes expressive ##yp dissipated ##rse localized systemic ##nikov gettysburg ##js ##uaries dialogues muttering 251 housekeeper sicilian discouraged ##frey beamed kaladin halftime kidnap ##amo ##llet 1754 synonymous depleted instituto insulin reprised ##opsis clashed ##ctric interrupting radcliffe insisting medici 1715 ejected playfully turbulent ##47 starvation ##rini shipment rebellious petersen verification merits ##rified cakes ##charged 1757 milford shortages spying fidelity ##aker emitted storylines harvested seismic ##iform cheung kilda theoretically barbie lynx ##rgy ##tius goblin mata poisonous ##nburg reactive residues obedience ##евич conjecture ##rac 401 hating sixties kicker moaning motown ##bha emancipation neoclassical ##hering consoles ebert professorship ##tures sustaining assaults obeyed affluent incurred tornadoes ##eber ##zow emphasizing highlanders cheated helmets ##ctus internship terence bony executions legislators berries peninsular tinged ##aco 1689 amplifier corvette ribbons lavish pennant ##lander worthless ##chfield ##forms mariano pyrenees expenditures ##icides chesterfield mandir tailor 39th sergey nestled willed aristocracy devotees goodnight raaf rumored weaponry remy appropriations harcourt burr riaa ##lence limitation unnoticed guo soaking swamps ##tica collapsing tatiana descriptive brigham psalm ##chment maddox ##lization patti caliph ##aja akron injuring serra ##ganj basins ##sari astonished launcher ##church hilary wilkins sewing ##sf stinging ##fia ##ncia underwood startup ##ition compilations vibrations embankment jurist ##nity bard juventus groundwater kern palaces helium boca cramped marissa soto ##worm jae princely ##ggy faso bazaar warmly ##voking 229 pairing ##lite ##grate ##nets wien freaked ulysses rebirth ##alia ##rent mummy guzman jimenez stilled ##nitz trajectory tha woken archival professions ##pts ##pta hilly shadowy shrink ##bolt norwood glued migrate stereotypes devoid ##pheus 625 evacuate horrors infancy gotham knowles optic downloaded sachs kingsley parramatta darryl mor ##onale shady commence confesses kan ##meter ##placed marlborough roundabout regents frigates io ##imating gothenburg revoked carvings clockwise convertible intruder ##sche banged ##ogo vicky bourgeois ##mony dupont footing ##gum pd ##real buckle yun penthouse sane 720 serviced stakeholders neumann bb ##eers comb ##gam catchment pinning rallies typing ##elles forefront freiburg sweetie giacomo widowed goodwill worshipped aspirations midday ##vat fishery ##trick bournemouth turk 243 hearth ethanol guadalajara murmurs sl ##uge afforded scripted ##hta wah ##jn coroner translucent 252 memorials puck progresses clumsy ##race 315 candace recounted ##27 ##slin ##uve filtering ##mac howl strata heron leveled ##ays dubious ##oja ##т ##wheel citations exhibiting ##laya ##mics ##pods turkic ##lberg injunction ##ennial ##mit antibodies ##44 organise ##rigues cardiovascular cushion inverness ##zquez dia cocoa sibling ##tman ##roid expanse feasible tunisian algiers ##relli rus bloomberg dso westphalia bro tacoma 281 downloads ##ours konrad duran ##hdi continuum jett compares legislator secession ##nable ##gues ##zuka translating reacher ##gley ##ła aleppo ##agi tc orchards trapping linguist versatile drumming postage calhoun superiors ##mx barefoot leary ##cis ignacio alfa kaplan ##rogen bratislava mori ##vot disturb haas 313 cartridges gilmore radiated salford tunic hades ##ulsive archeological delilah magistrates auditioned brewster charters empowerment blogs cappella dynasties iroquois whipping ##krishna raceway truths myra weaken judah mcgregor ##horse mic refueling 37th burnley bosses markus premio query ##gga dunbar ##economic darkest lyndon sealing commendation reappeared ##mun addicted ezio slaughtered satisfactory shuffle ##eves ##thic ##uj fortification warrington ##otto resurrected fargo mane ##utable ##lei ##space foreword ox ##aris ##vern abrams hua ##mento sakura ##alo uv sentimental ##skaya midfield ##eses sturdy scrolls macleod ##kyu entropy ##lance mitochondrial cicero excelled thinner convoys perceive ##oslav ##urable systematically grind burkina 287 ##tagram ops ##aman guantanamo ##cloth ##tite forcefully wavy ##jou pointless ##linger ##tze layton portico superficial clerical outlaws ##hism burials muir ##inn creditors hauling rattle ##leg calais monde archers reclaimed dwell wexford hellenic falsely remorse ##tek dough furnishings ##uttered gabon neurological novice ##igraphy contemplated pulpit nightstand saratoga ##istan documenting pulsing taluk ##firmed busted marital ##rien disagreements wasps ##yes hodge mcdonnell mimic fran pendant dhabi musa ##nington congratulations argent darrell concussion losers regrets thessaloniki reversal donaldson hardwood thence achilles ritter ##eran demonic jurgen prophets goethe eki classmate buff ##cking yank irrational ##inging perished seductive qur sourced ##crat ##typic mustard ravine barre horizontally characterization phylogenetic boise ##dit ##runner ##tower brutally intercourse seduce ##bbing fay ferris ogden amar nik unarmed ##inator evaluating kyrgyzstan sweetness ##lford ##oki mccormick meiji notoriety stimulate disrupt figuring instructional mcgrath ##zoo groundbreaking ##lto flinch khorasan agrarian bengals mixer radiating ##sov ingram pitchers nad tariff ##cript tata ##codes ##emi ##ungen appellate lehigh ##bled ##giri brawl duct texans ##ciation ##ropolis skipper speculative vomit doctrines stresses 253 davy graders whitehead jozef timely cumulative haryana paints appropriately boon cactus ##ales ##pid dow legions ##pit perceptions 1730 picturesque ##yse periphery rune wr ##aha celtics sentencing whoa ##erin confirms variance 425 moines mathews spade rave m1 fronted fx blending alleging reared ##gl 237 ##paper grassroots eroded ##free ##physical directs ordeal ##sław accelerate hacker rooftop ##inia lev buys cebu devote ##lce specialising ##ulsion choreographed repetition warehouses ##ryl paisley tuscany analogy sorcerer hash huts shards descends exclude nix chaplin gaga ito vane ##drich causeway misconduct limo orchestrated glands jana ##kot u2 ##mple ##sons branching contrasts scoop longed ##virus chattanooga ##75 syrup cornerstone ##tized ##mind ##iaceae careless precedence frescoes ##uet chilled consult modelled snatch peat ##thermal caucasian humane relaxation spins temperance ##lbert occupations lambda hybrids moons mp3 ##oese 247 rolf societal yerevan ness ##ssler befriended mechanized nominate trough boasted cues seater ##hom bends ##tangle conductors emptiness ##lmer eurasian adriatic tian ##cie anxiously lark propellers chichester jock ev 2a ##holding credible recounts tori loyalist abduction ##hoot ##redo nepali ##mite ventral tempting ##ango ##crats steered ##wice javelin dipping laborers prentice looming titanium ##ː badges emir tensor ##ntation egyptians rash denies hawthorne lombard showers wehrmacht dietary trojan ##reus welles executing horseshoe lifeboat ##lak elsa infirmary nearing roberta boyer mutter trillion joanne ##fine ##oked sinks vortex uruguayan clasp sirius ##block accelerator prohibit sunken byu chronological diplomats ochreous 510 symmetrical 1644 maia ##tology salts reigns atrocities ##ия hess bared issn ##vyn cater saturated ##cycle ##isse sable voyager dyer yusuf ##inge fountains wolff ##39 ##nni engraving rollins atheist ominous ##ault herr chariot martina strung ##fell ##farlane horrific sahib gazes saetan erased ptolemy ##olic flushing lauderdale analytic ##ices 530 navarro beak gorilla herrera broom guadalupe raiding sykes 311 bsc deliveries 1720 invasions carmichael tajikistan thematic ecumenical sentiments onstage ##rians ##brand ##sume catastrophic flanks molten ##arns waller aimee terminating ##icing alternately ##oche nehru printers outraged ##eving empires template banners repetitive za ##oise vegetarian ##tell guiana opt cavendish lucknow synthesized ##hani ##mada finalized ##ctable fictitious mayoral unreliable ##enham embracing peppers rbis ##chio ##neo inhibition slashed togo orderly embroidered safari salty 236 barron benito totaled ##dak pubs simulated caden devin tolkien momma welding sesame ##ept gottingen hardness 630 shaman temeraire 620 adequately pediatric ##kit ck assertion radicals composure cadence seafood beaufort lazarus mani warily cunning kurdistan 249 cantata ##kir ares ##41 ##clusive nape townland geared insulted flutter boating violate draper dumping malmo ##hh ##romatic firearm alta bono obscured ##clave exceeds panorama unbelievable ##train preschool ##essed disconnected installing rescuing secretaries accessibility ##castle ##drive ##ifice ##film bouts slug waterway mindanao ##buro ##ratic halves ##ل calming liter maternity adorable bragg electrification mcc ##dote roxy schizophrenia ##body munoz kaye whaling 239 mil tingling tolerant ##ago unconventional volcanoes ##finder deportivo ##llie robson kaufman neuroscience wai deportation masovian scraping converse ##bh hacking bulge ##oun administratively yao 580 amp mammoth booster claremont hooper nomenclature pursuits mclaughlin melinda ##sul catfish barclay substrates taxa zee originals kimberly packets padma ##ality borrowing ostensibly solvent ##bri ##genesis ##mist lukas shreveport veracruz ##ь ##lou ##wives cheney tt anatolia hobbs ##zyn cyclic radiant alistair greenish siena dat independents ##bation conform pieter hyper applicant bradshaw spores telangana vinci inexpensive nuclei 322 jang nme soho spd ##ign cradled receptionist pow ##43 ##rika fascism ##ifer experimenting ##ading ##iec ##region 345 jocelyn maris stair nocturnal toro constabulary elgin ##kker msc ##giving ##schen ##rase doherty doping sarcastically batter maneuvers ##cano ##apple ##gai ##git intrinsic ##nst ##stor 1753 showtime cafes gasps lviv ushered ##thed fours restart astonishment transmitting flyer shrugs ##sau intriguing cones dictated mushrooms medial ##kovsky ##elman escorting gaped ##26 godfather ##door ##sell djs recaptured timetable vila 1710 3a aerodrome mortals scientology ##orne angelina mag convection unpaid insertion intermittent lego ##nated endeavor kota pereira ##lz 304 bwv glamorgan insults agatha fey ##cend fleetwood mahogany protruding steamship zeta ##arty mcguire suspense ##sphere advising urges ##wala hurriedly meteor gilded inline arroyo stalker ##oge excitedly revered ##cure earle introductory ##break ##ilde mutants puff pulses reinforcement ##haling curses lizards stalk correlated ##fixed fallout macquarie ##unas bearded denton heaving 802 ##ocation winery assign dortmund ##lkirk everest invariant charismatic susie ##elling bled lesley telegram sumner bk ##ogen ##к wilcox needy colbert duval ##iferous ##mbled allotted attends imperative ##hita replacements hawker ##inda insurgency ##zee ##eke casts ##yla 680 ives transitioned ##pack ##powering authoritative baylor flex cringed plaintiffs woodrow ##skie drastic ape aroma unfolded commotion nt preoccupied theta routines lasers privatization wand domino ek clenching nsa strategically showered bile handkerchief pere storing christophe insulting 316 nakamura romani asiatic magdalena palma cruises stripping 405 konstantin soaring ##berman colloquially forerunner havilland incarcerated parasites sincerity ##utus disks plank saigon ##ining corbin homo ornaments powerhouse ##tlement chong fastened feasibility idf morphological usable ##nish ##zuki aqueduct jaguars keepers ##flies aleksandr faust assigns ewing bacterium hurled tricky hungarians integers wallis 321 yamaha ##isha hushed oblivion aviator evangelist friars ##eller monograph ode ##nary airplanes labourers charms ##nee 1661 hagen tnt rudder fiesta transcript dorothea ska inhibitor maccabi retorted raining encompassed clauses menacing 1642 lineman ##gist vamps ##ape ##dick gloom ##rera dealings easing seekers ##nut ##pment helens unmanned ##anu ##isson basics ##amy ##ckman adjustments 1688 brutality horne ##zell sui ##55 ##mable aggregator ##thal rhino ##drick ##vira counters zoom ##01 ##rting mn montenegrin packard ##unciation ##♭ ##kki reclaim scholastic thugs pulsed ##icia syriac quan saddam banda kobe blaming buddies dissent ##lusion ##usia corbett jaya delle erratic lexie ##hesis 435 amiga hermes ##pressing ##leen chapels gospels jamal ##uating compute revolving warp ##sso ##thes armory ##eras ##gol antrim loki ##kow ##asian ##good ##zano braid handwriting subdistrict funky pantheon ##iculate concurrency estimation improper juliana ##his newcomers johnstone staten communicated ##oco ##alle sausage stormy ##stered ##tters superfamily ##grade acidic collateral tabloid ##oped ##rza bladder austen ##ellant mcgraw ##hay hannibal mein aquino lucifer wo badger boar cher christensen greenberg interruption ##kken jem 244 mocked bottoms cambridgeshire ##lide sprawling ##bbly eastwood ghent synth ##buck advisers ##bah nominally hapoel qu daggers estranged fabricated towels vinnie wcw misunderstanding anglia nothin unmistakable ##dust ##lova chilly marquette truss ##edge ##erine reece ##lty ##chemist ##connected 272 308 41st bash raion waterfalls ##ump ##main labyrinth queue theorist ##istle bharatiya flexed soundtracks rooney leftist patrolling wharton plainly alleviate eastman schuster topographic engages immensely unbearable fairchild 1620 dona lurking parisian oliveira ia indictment hahn bangladeshi ##aster vivo ##uming ##ential antonia expects indoors kildare harlan ##logue ##ogenic ##sities forgiven ##wat childish tavi ##mide ##orra plausible grimm successively scooted ##bola ##dget ##rith spartans emery flatly azure epilogue ##wark flourish ##iny ##tracted ##overs ##oshi bestseller distressed receipt spitting hermit topological ##cot drilled subunit francs ##layer eel ##fk ##itas octopus footprint petitions ufo ##say ##foil interfering leaking palo ##metry thistle valiant ##pic narayan mcpherson ##fast gonzales ##ym ##enne dustin novgorod solos ##zman doin ##raph ##patient ##meyer soluble ashland cuffs carole pendleton whistling vassal ##river deviation revisited constituents rallied rotate loomed ##eil ##nting amateurs augsburg auschwitz crowns skeletons ##cona bonnet 257 dummy globalization simeon sleeper mandal differentiated ##crow ##mare milne bundled exasperated talmud owes segregated ##feng ##uary dentist piracy props ##rang devlin ##torium malicious paws ##laid dependency ##ergy ##fers ##enna 258 pistons rourke jed grammatical tres maha wig 512 ghostly jayne ##achal ##creen ##ilis ##lins ##rence designate ##with arrogance cambodian clones showdown throttle twain ##ception lobes metz nagoya 335 braking ##furt 385 roaming ##minster amin crippled ##37 ##llary indifferent hoffmann idols intimidating 1751 261 influenza memo onions 1748 bandage consciously ##landa ##rage clandestine observes swiped tangle ##ener ##jected ##trum ##bill ##lta hugs congresses josiah spirited ##dek humanist managerial filmmaking inmate rhymes debuting grimsby ur ##laze duplicate vigor ##tf republished bolshevik refurbishment antibiotics martini methane newscasts royale horizons levant iain visas ##ischen paler ##around manifestation snuck alf chop futile pedestal rehab ##kat bmg kerman res fairbanks jarrett abstraction saharan ##zek 1746 procedural clearer kincaid sash luciano ##ffey crunch helmut ##vara revolutionaries ##tute creamy leach ##mmon 1747 permitting nes plight wendell ##lese contra ts clancy ipa mach staples autopsy disturbances nueva karin pontiac ##uding proxy venerable haunt leto bergman expands ##helm wal ##pipe canning celine cords obesity ##enary intrusion planner ##phate reasoned sequencing 307 harrow ##chon ##dora marred mcintyre repay tarzan darting 248 harrisburg margarita repulsed ##hur ##lding belinda hamburger novo compliant runways bingham registrar skyscraper ic cuthbert improvisation livelihood ##corp ##elial admiring ##dened sporadic believer casablanca popcorn ##29 asha shovel ##bek ##dice coiled tangible ##dez casper elsie resin tenderness rectory ##ivision avail sonar ##mori boutique ##dier guerre bathed upbringing vaulted sandals blessings ##naut ##utnant 1680 306 foxes pia corrosion hesitantly confederates crystalline footprints shapiro tirana valentin drones 45th microscope shipments texted inquisition wry guernsey unauthorized resigning 760 ripple schubert stu reassure felony ##ardo brittle koreans ##havan ##ives dun implicit tyres ##aldi ##lth magnolia ##ehan ##puri ##poulos aggressively fei gr familiarity ##poo indicative ##trust fundamentally jimmie overrun 395 anchors moans ##opus britannia armagh ##ggle purposely seizing ##vao bewildered mundane avoidance cosmopolitan geometridae quartermaster caf 415 chatter engulfed gleam purge ##icate juliette jurisprudence guerra revisions ##bn casimir brew ##jm 1749 clapton cloudy conde hermitage 278 simulations torches vincenzo matteo ##rill hidalgo booming westbound accomplishment tentacles unaffected ##sius annabelle flopped sloping ##litz dreamer interceptor vu ##loh consecration copying messaging breaker climates hospitalized 1752 torino afternoons winfield witnessing ##teacher breakers choirs sawmill coldly ##ege sipping haste uninhabited conical bibliography pamphlets severn edict ##oca deux illnesses grips ##pl rehearsals sis thinkers tame ##keepers 1690 acacia reformer ##osed ##rys shuffling ##iring ##shima eastbound ionic rhea flees littered ##oum rocker vomiting groaning champ overwhelmingly civilizations paces sloop adoptive ##tish skaters ##vres aiding mango ##joy nikola shriek ##ignon pharmaceuticals ##mg tuna calvert gustavo stocked yearbook ##urai ##mana computed subsp riff hanoi kelvin hamid moors pastures summons jihad nectar ##ctors bayou untitled pleasing vastly republics intellect ##η ##ulio ##tou crumbling stylistic sb ##ی consolation frequented h₂o walden widows ##iens 404 ##ignment chunks improves 288 grit recited ##dev snarl sociological ##arte ##gul inquired ##held bruise clube consultancy homogeneous hornets multiplication pasta prick savior ##grin ##kou ##phile yoon ##gara grimes vanishing cheering reacting bn distillery ##quisite ##vity coe dockyard massif ##jord escorts voss ##valent byte chopped hawke illusions workings floats ##koto ##vac kv annapolis madden ##onus alvaro noctuidae ##cum ##scopic avenge steamboat forte illustrates erika ##trip 570 dew nationalities bran manifested thirsty diversified muscled reborn ##standing arson ##lessness ##dran ##logram ##boys ##kushima ##vious willoughby ##phobia 286 alsace dashboard yuki ##chai granville myspace publicized tricked ##gang adjective ##ater relic reorganisation enthusiastically indications saxe ##lassified consolidate iec padua helplessly ramps renaming regulars pedestrians accents convicts inaccurate lowers mana ##pati barrie bjp outta someplace berwick flanking invoked marrow sparsely excerpts clothed rei ##ginal wept ##straße ##vish alexa excel ##ptive membranes aquitaine creeks cutler sheppard implementations ns ##dur fragrance budge concordia magnesium marcelo ##antes gladly vibrating ##rral ##ggles montrose ##omba lew seamus 1630 cocky ##ament ##uen bjorn ##rrick fielder fluttering ##lase methyl kimberley mcdowell reductions barbed ##jic ##tonic aeronautical condensed distracting ##promising huffed ##cala ##sle claudius invincible missy pious balthazar ci ##lang butte combo orson ##dication myriad 1707 silenced ##fed ##rh coco netball yourselves ##oza clarify heller peg durban etudes offender roast blackmail curvature ##woods vile 309 illicit suriname ##linson overture 1685 bubbling gymnast tucking ##mming ##ouin maldives ##bala gurney ##dda ##eased ##oides backside pinto jars racehorse tending ##rdial baronetcy wiener duly ##rke barbarian cupping flawed ##thesis bertha pleistocene puddle swearing ##nob ##tically fleeting prostate amulet educating ##mined ##iti ##tler 75th jens respondents analytics cavaliers papacy raju ##iente ##ulum ##tip funnel 271 disneyland ##lley sociologist ##iam 2500 faulkner louvre menon ##dson 276 ##ower afterlife mannheim peptide referees comedians meaningless ##anger ##laise fabrics hurley renal sleeps ##bour ##icle breakout kristin roadside animator clover disdain unsafe redesign ##urity firth barnsley portage reset narrows 268 commandos expansive speechless tubular ##lux essendon eyelashes smashwords ##yad ##bang ##claim craved sprinted chet somme astor wrocław orton 266 bane ##erving ##uing mischief ##amps ##sund scaling terre ##xious impairment offenses undermine moi soy contiguous arcadia inuit seam ##tops macbeth rebelled ##icative ##iot 590 elaborated frs uniformed ##dberg 259 powerless priscilla stimulated 980 qc arboretum frustrating trieste bullock ##nified enriched glistening intern ##adia locus nouvelle ollie ike lash starboard ee tapestry headlined hove rigged ##vite pollock ##yme thrive clustered cas roi gleamed olympiad ##lino pressured regimes ##hosis ##lick ripley ##ophone kickoff gallon rockwell ##arable crusader glue revolutions scrambling 1714 grover ##jure englishman aztec 263 contemplating coven ipad preach triumphant tufts ##esian rotational ##phus 328 falkland ##brates strewn clarissa rejoin environmentally glint banded drenched moat albanians johor rr maestro malley nouveau shaded taxonomy v6 adhere bunk airfields ##ritan 1741 encompass remington tran ##erative amelie mazda friar morals passions ##zai breadth vis ##hae argus burnham caressing insider rudd ##imov ##mini ##rso italianate murderous textual wainwright armada bam weave timer ##taken ##nh fra ##crest ardent salazar taps tunis ##ntino allegro gland philanthropic ##chester implication ##optera esq judas noticeably wynn ##dara inched indexed crises villiers bandit royalties patterned cupboard interspersed accessory isla kendrick entourage stitches ##esthesia headwaters ##ior interlude distraught draught 1727 ##basket biased sy transient triad subgenus adapting kidd shortstop ##umatic dimly spiked mcleod reprint nellie pretoria windmill ##cek singled ##mps 273 reunite ##orous 747 bankers outlying ##omp ##ports ##tream apologies cosmetics patsy ##deh ##ocks ##yson bender nantes serene ##nad lucha mmm 323 ##cius ##gli cmll coinage nestor juarez ##rook smeared sprayed twitching sterile irina embodied juveniles enveloped miscellaneous cancers dq gulped luisa crested swat donegal ref ##anov ##acker hearst mercantile ##lika doorbell ua vicki ##alla ##som bilbao psychologists stryker sw horsemen turkmenistan wits ##national anson mathew screenings ##umb rihanna ##agne ##nessy aisles ##iani ##osphere hines kenton saskatoon tasha truncated ##champ ##itan mildred advises fredrik interpreting inhibitors ##athi spectroscopy ##hab ##kong karim panda ##oia ##nail ##vc conqueror kgb leukemia ##dity arrivals cheered pisa phosphorus shielded ##riated mammal unitarian urgently chopin sanitary ##mission spicy drugged hinges ##tort tipping trier impoverished westchester ##caster 267 epoch nonstop ##gman ##khov aromatic centrally cerro ##tively ##vio billions modulation sedimentary 283 facilitating outrageous goldstein ##eak ##kt ld maitland penultimate pollard ##dance fleets spaceship vertebrae ##nig alcoholism als recital ##bham ##ference ##omics m2 ##bm trois ##tropical ##в commemorates ##meric marge ##raction 1643 670 cosmetic ravaged ##ige catastrophe eng ##shida albrecht arterial bellamy decor harmon ##rde bulbs synchronized vito easiest shetland shielding wnba ##glers ##ssar ##riam brianna cumbria ##aceous ##rard cores thayer ##nsk brood hilltop luminous carts keynote larkin logos ##cta ##ا ##mund ##quay lilith tinted 277 wrestle mobilization ##uses sequential siam bloomfield takahashi 274 ##ieving presenters ringo blazed witty ##oven ##ignant devastation haydn harmed newt therese ##peed gershwin molina rabbis sudanese 001 innate restarted ##sack ##fus slices wb ##shah enroll hypothetical hysterical 1743 fabio indefinite warped ##hg exchanging 525 unsuitable ##sboro gallo 1603 bret cobalt homemade ##hunter mx operatives ##dhar terraces durable latch pens whorls ##ctuated ##eaux billing ligament succumbed ##gly regulators spawn ##brick ##stead filmfare rochelle ##nzo 1725 circumstance saber supplements ##nsky ##tson crowe wellesley carrot ##9th ##movable primate drury sincerely topical ##mad ##rao callahan kyiv smarter tits undo ##yeh announcements anthologies barrio nebula ##islaus ##shaft ##tyn bodyguards 2021 assassinate barns emmett scully ##mah ##yd ##eland ##tino ##itarian demoted gorman lashed prized adventist writ ##gui alla invertebrates ##ausen 1641 amman 1742 align healy redistribution ##gf ##rize insulation ##drop adherents hezbollah vitro ferns yanking 269 php registering uppsala cheerleading confines mischievous tully ##ross 49th docked roam stipulated pumpkin ##bry prompt ##ezer blindly shuddering craftsmen frail scented katharine scramble shaggy sponge helix zaragoza 279 ##52 43rd backlash fontaine seizures posse cowan nonfiction telenovela wwii hammered undone ##gpur encircled irs ##ivation artefacts oneself searing smallpox ##belle ##osaurus shandong breached upland blushing rankin infinitely psyche tolerated docking evicted ##col unmarked ##lving gnome lettering litres musique ##oint benevolent ##jal blackened ##anna mccall racers tingle ##ocene ##orestation introductions radically 292 ##hiff ##باد 1610 1739 munchen plead ##nka condo scissors ##sight ##tens apprehension ##cey ##yin hallmark watering formulas sequels ##llas aggravated bae commencing ##building enfield prohibits marne vedic civilized euclidean jagger beforehand blasts dumont ##arney ##nem 740 conversions hierarchical rios simulator ##dya ##lellan hedges oleg thrusts shadowed darby maximize 1744 gregorian ##nded ##routed sham unspecified ##hog emory factual ##smo ##tp fooled ##rger ortega wellness marlon ##oton ##urance casket keating ley enclave ##ayan char influencing jia ##chenko 412 ammonia erebidae incompatible violins cornered ##arat grooves astronauts columbian rampant fabrication kyushu mahmud vanish ##dern mesopotamia ##lete ict ##rgen caspian kenji pitted ##vered 999 grimace roanoke tchaikovsky twinned ##analysis ##awan xinjiang arias clemson kazakh sizable 1662 ##khand ##vard plunge tatum vittorio ##nden cholera ##dana ##oper bracing indifference projectile superliga ##chee realises upgrading 299 porte retribution ##vies nk stil ##resses ama bureaucracy blackberry bosch testosterone collapses greer ##pathic ioc fifties malls ##erved bao baskets adolescents siegfried ##osity ##tosis mantra detecting existent fledgling ##cchi dissatisfied gan telecommunication mingled sobbed 6000 controversies outdated taxis ##raus fright slams ##lham ##fect ##tten detectors fetal tanned ##uw fray goth olympian skipping mandates scratches sheng unspoken hyundai tracey hotspur restrictive ##buch americana mundo ##bari burroughs diva vulcan ##6th distinctions thumping ##ngen mikey sheds fide rescues springsteen vested valuation ##ece ##ely pinnacle rake sylvie ##edo almond quivering ##irus alteration faltered ##wad 51st hydra ticked ##kato recommends ##dicated antigua arjun stagecoach wilfred trickle pronouns ##pon aryan nighttime ##anian gall pea stitch ##hei leung milos ##dini eritrea nexus starved snowfall kant parasitic cot discus hana strikers appleton kitchens ##erina ##partisan ##itha ##vius disclose metis ##channel 1701 tesla ##vera fitch 1735 blooded ##tila decimal ##tang ##bai cyclones eun bottled peas pensacola basha bolivian crabs boil lanterns partridge roofed 1645 necks ##phila opined patting ##kla ##lland chuckles volta whereupon ##nche devout euroleague suicidal ##dee inherently involuntary knitting nasser ##hide puppets colourful courageous southend stills miraculous hodgson richer rochdale ethernet greta uniting prism umm ##haya ##itical ##utation deterioration pointe prowess ##ropriation lids scranton billings subcontinent ##koff ##scope brute kellogg psalms degraded ##vez stanisław ##ructured ferreira pun astonishing gunnar ##yat arya prc gottfried ##tight excursion ##ographer dina ##quil ##nare huffington illustrious wilbur gundam verandah ##zard naacp ##odle constructive fjord kade ##naud generosity thrilling baseline cayman frankish plastics accommodations zoological ##fting cedric qb motorized ##dome ##otted squealed tackled canucks budgets situ asthma dail gabled grasslands whimpered writhing judgments ##65 minnie pv ##carbon bananas grille domes monique odin maguire markham tierney ##estra ##chua libel poke speedy atrium laval notwithstanding ##edly fai kala ##sur robb ##sma listings luz supplementary tianjin ##acing enzo jd ric scanner croats transcribed ##49 arden cv ##hair ##raphy ##lver ##uy 357 seventies staggering alam horticultural hs regression timbers blasting ##ounded montagu manipulating ##cit catalytic 1550 troopers ##meo condemnation fitzpatrick ##oire ##roved inexperienced 1670 castes ##lative outing 314 dubois flicking quarrel ste learners 1625 iq whistled ##class 282 classify tariffs temperament 355 folly liszt ##yles immersed jordanian ceasefire apparel extras maru fished ##bio harta stockport assortment craftsman paralysis transmitters ##cola blindness ##wk fatally proficiency solemnly ##orno repairing amore groceries ultraviolet ##chase schoolhouse ##tua resurgence nailed ##otype ##× ruse saliva diagrams ##tructing albans rann thirties 1b antennas hilarious cougars paddington stats ##eger breakaway ipod reza authorship prohibiting scoffed ##etz ##ttle conscription defected trondheim ##fires ivanov keenan ##adan ##ciful ##fb ##slow locating ##ials ##tford cadiz basalt blankly interned rags rattling ##tick carpathian reassured sync bum guildford iss staunch ##onga astronomers sera sofie emergencies susquehanna ##heard duc mastery vh1 williamsburg bayer buckled craving ##khan ##rdes bloomington ##write alton barbecue ##bians justine ##hri ##ndt delightful smartphone newtown photon retrieval peugeot hissing ##monium ##orough flavors lighted relaunched tainted ##games ##lysis anarchy microscopic hopping adept evade evie ##beau inhibit sinn adjustable hurst intuition wilton cisco 44th lawful lowlands stockings thierry ##dalen ##hila ##nai fates prank tb maison lobbied provocative 1724 4a utopia ##qual carbonate gujarati purcell ##rford curtiss ##mei overgrown arenas mediation swallows ##rnik respectful turnbull ##hedron ##hope alyssa ozone ##ʻi ami gestapo johansson snooker canteen cuff declines empathy stigma ##ags ##iner ##raine taxpayers gui volga ##wright ##copic lifespan overcame tattooed enactment giggles ##ador ##camp barrington bribe obligatory orbiting peng ##enas elusive sucker ##vating cong hardship empowered anticipating estrada cryptic greasy detainees planck sudbury plaid dod marriott kayla ##ears ##vb ##zd mortally ##hein cognition radha 319 liechtenstein meade richly argyle harpsichord liberalism trumpets lauded tyrant salsa tiled lear promoters reused slicing trident ##chuk ##gami ##lka cantor checkpoint ##points gaul leger mammalian ##tov ##aar ##schaft doha frenchman nirvana ##vino delgado headlining ##eron ##iography jug tko 1649 naga intersections ##jia benfica nawab ##suka ashford gulp ##deck ##vill ##rug brentford frazier pleasures dunne potsdam shenzhen dentistry ##tec flanagan ##dorff ##hear chorale dinah prem quezon ##rogated relinquished sutra terri ##pani flaps ##rissa poly ##rnet homme aback ##eki linger womb ##kson ##lewood doorstep orthodoxy threaded westfield ##rval dioceses fridays subsided ##gata loyalists ##biotic ##ettes letterman lunatic prelate tenderly invariably souza thug winslow ##otide furlongs gogh jeopardy ##runa pegasus ##umble humiliated standalone tagged ##roller freshmen klan ##bright attaining initiating transatlantic logged viz ##uance 1723 combatants intervening stephane chieftain despised grazed 317 cdc galveston godzilla macro simulate ##planes parades ##esses 960 ##ductive ##unes equator overdose ##cans ##hosh ##lifting joshi epstein sonora treacherous aquatics manchu responsive ##sation supervisory ##christ ##llins ##ibar ##balance ##uso kimball karlsruhe mab ##emy ignores phonetic reuters spaghetti 820 almighty danzig rumbling tombstone designations lured outset ##felt supermarkets ##wt grupo kei kraft susanna ##blood comprehension genealogy ##aghan ##verted redding ##ythe 1722 bowing ##pore ##roi lest sharpened fulbright valkyrie sikhs ##unds swans bouquet merritt ##tage ##venting commuted redhead clerks leasing cesare dea hazy ##vances fledged greenfield servicemen ##gical armando blackout dt sagged downloadable intra potion pods ##4th ##mism xp attendants gambia stale ##ntine plump asteroids rediscovered buds flea hive ##neas 1737 classifications debuts ##eles olympus scala ##eurs ##gno ##mute hummed sigismund visuals wiggled await pilasters clench sulfate ##ances bellevue enigma trainee snort ##sw clouded denim ##rank ##rder churning hartman lodges riches sima ##missible accountable socrates regulates mueller ##cr 1702 avoids solids himalayas nutrient pup ##jevic squat fades nec ##lates ##pina ##rona ##ου privateer tequila ##gative ##mpton apt hornet immortals ##dou asturias cleansing dario ##rries ##anta etymology servicing zhejiang ##venor ##nx horned erasmus rayon relocating £10 ##bags escalated promenade stubble 2010s artisans axial liquids mora sho yoo ##tsky bundles oldies ##nally notification bastion ##ths sparkle ##lved 1728 leash pathogen highs ##hmi immature 880 gonzaga ignatius mansions monterrey sweets bryson ##loe polled regatta brightest pei rosy squid hatfield payroll addict meath cornerback heaviest lodging ##mage capcom rippled ##sily barnet mayhem ymca snuggled rousseau ##cute blanchard 284 fragmented leighton chromosomes risking ##md ##strel ##utter corinne coyotes cynical hiroshi yeomanry ##ractive ebook grading mandela plume agustin magdalene ##rkin bea femme trafford ##coll ##lun ##tance 52nd fourier upton ##mental camilla gust iihf islamabad longevity ##kala feldman netting ##rization endeavour foraging mfa orr ##open greyish contradiction graz ##ruff handicapped marlene tweed oaxaca spp campos miocene pri configured cooks pluto cozy pornographic ##entes 70th fairness glided jonny lynne rounding sired ##emon ##nist remade uncover ##mack complied lei newsweek ##jured ##parts ##enting ##pg 293 finer guerrillas athenian deng disused stepmother accuse gingerly seduction 521 confronting ##walker ##going gora nostalgia sabres virginity wrenched ##minated syndication wielding eyre ##56 ##gnon ##igny behaved taxpayer sweeps ##growth childless gallant ##ywood amplified geraldine scrape ##ffi babylonian fresco ##rdan ##kney ##position 1718 restricting tack fukuoka osborn selector partnering ##dlow 318 gnu kia tak whitley gables ##54 ##mania mri softness immersion ##bots ##evsky 1713 chilling insignificant pcs ##uis elites lina purported supplemental teaming ##americana ##dding ##inton proficient rouen ##nage ##rret niccolo selects ##bread fluffy 1621 gruff knotted mukherjee polgara thrash nicholls secluded smoothing thru corsica loaf whitaker inquiries ##rrier ##kam indochina 289 marlins myles peking ##tea extracts pastry superhuman connacht vogel ##ditional ##het ##udged ##lash gloss quarries refit teaser ##alic ##gaon 20s materialized sling camped pickering tung tracker pursuant ##cide cranes soc ##cini ##typical ##viere anhalt overboard workout chores fares orphaned stains ##logie fenton surpassing joyah triggers ##itte grandmaster ##lass ##lists clapping fraudulent ledger nagasaki ##cor ##nosis ##tsa eucalyptus tun ##icio ##rney ##tara dax heroism ina wrexham onboard unsigned ##dates moshe galley winnie droplets exiles praises watered noodles ##aia fein adi leland multicultural stink bingo comets erskine modernized canned constraint domestically chemotherapy featherweight stifled ##mum darkly irresistible refreshing hasty isolate ##oys kitchener planners ##wehr cages yarn implant toulon elects childbirth yue ##lind ##lone cn rightful sportsman junctions remodeled specifies ##rgh 291 ##oons complimented ##urgent lister ot ##logic bequeathed cheekbones fontana gabby ##dial amadeus corrugated maverick resented triangles ##hered ##usly nazareth tyrol 1675 assent poorer sectional aegean ##cous 296 nylon ghanaian ##egorical ##weig cushions forbid fusiliers obstruction somerville ##scia dime earrings elliptical leyte oder polymers timmy atm midtown piloted settles continual externally mayfield ##uh enrichment henson keane persians 1733 benji braden pep 324 ##efe contenders pepsi valet ##isches 298 ##asse ##earing goofy stroll ##amen authoritarian occurrences adversary ahmedabad tangent toppled dorchester 1672 modernism marxism islamist charlemagne exponential racks unicode brunette mbc pic skirmish ##bund ##lad ##powered ##yst hoisted messina shatter ##ctum jedi vantage ##music ##neil clemens mahmoud corrupted authentication lowry nils ##washed omnibus wounding jillian ##itors ##opped serialized narcotics handheld ##arm ##plicity intersecting stimulating ##onis crate fellowships hemingway casinos climatic fordham copeland drip beatty leaflets robber brothel madeira ##hedral sphinx ultrasound ##vana valor forbade leonid villas ##aldo duane marquez ##cytes disadvantaged forearms kawasaki reacts consular lax uncles uphold ##hopper concepcion dorsey lass ##izan arching passageway 1708 researches tia internationals ##graphs ##opers distinguishes javanese divert ##uven plotted ##listic ##rwin ##erik ##tify affirmative signifies validation ##bson kari felicity georgina zulu ##eros ##rained ##rath overcoming ##dot argyll ##rbin 1734 chiba ratification windy earls parapet ##marks hunan pristine astrid punta ##gart brodie ##kota ##oder malaga minerva rouse ##phonic bellowed pagoda portals reclamation ##gur ##odies ##⁄₄ parentheses quoting allergic palette showcases benefactor heartland nonlinear ##tness bladed cheerfully scans ##ety ##hone 1666 girlfriends pedersen hiram sous ##liche ##nator 1683 ##nery ##orio ##umen bobo primaries smiley ##cb unearthed uniformly fis metadata 1635 ind ##oted recoil ##titles ##tura ##ια 406 hilbert jamestown mcmillan tulane seychelles ##frid antics coli fated stucco ##grants 1654 bulky accolades arrays caledonian carnage optimism puebla ##tative ##cave enforcing rotherham seo dunlop aeronautics chimed incline zoning archduke hellenistic ##oses ##sions candi thong ##ople magnate rustic ##rsk projective slant ##offs danes hollis vocalists ##ammed congenital contend gesellschaft ##ocating ##pressive douglass quieter ##cm ##kshi howled salim spontaneously townsville buena southport ##bold kato 1638 faerie stiffly ##vus ##rled 297 flawless realising taboo ##7th bytes straightening 356 jena ##hid ##rmin cartwright berber bertram soloists 411 noses 417 coping fission hardin inca ##cen 1717 mobilized vhf ##raf biscuits curate ##85 ##anial 331 gaunt neighbourhoods 1540 ##abas blanca bypassed sockets behold coincidentally ##bane nara shave splinter terrific ##arion ##erian commonplace juris redwood waistband boxed caitlin fingerprints jennie naturalized ##ired balfour craters jody bungalow hugely quilt glitter pigeons undertaker bulging constrained goo ##sil ##akh assimilation reworked ##person persuasion ##pants felicia ##cliff ##ulent 1732 explodes ##dun ##inium ##zic lyman vulture hog overlook begs northwards ow spoil ##urer fatima favorably accumulate sargent sorority corresponded dispersal kochi toned ##imi ##lita internacional newfound ##agger ##lynn ##rigue booths peanuts ##eborg medicare muriel nur ##uram crates millennia pajamas worsened ##breakers jimi vanuatu yawned ##udeau carousel ##hony hurdle ##ccus ##mounted ##pod rv ##eche airship ambiguity compulsion recapture ##claiming arthritis ##osomal 1667 asserting ngc sniffing dade discontent glendale ported ##amina defamation rammed ##scent fling livingstone ##fleet 875 ##ppy apocalyptic comrade lcd ##lowe cessna eine persecuted subsistence demi hoop reliefs 710 coptic progressing stemmed perpetrators 1665 priestess ##nio dobson ebony rooster itf tortricidae ##bbon ##jian cleanup ##jean ##øy 1721 eighties taxonomic holiness ##hearted ##spar antilles showcasing stabilized ##nb gia mascara michelangelo dawned ##uria ##vinsky extinguished fitz grotesque £100 ##fera ##loid ##mous barges neue throbbed cipher johnnie ##a1 ##mpt outburst ##swick spearheaded administrations c1 heartbreak pixels pleasantly ##enay lombardy plush ##nsed bobbie ##hly reapers tremor xiang minogue substantive hitch barak ##wyl kwan ##encia 910 obscene elegance indus surfer bribery conserve ##hyllum ##masters horatio ##fat apes rebound psychotic ##pour iteration ##mium ##vani botanic horribly antiques dispose paxton ##hli ##wg timeless 1704 disregard engraver hounds ##bau ##version looted uno facilitates groans masjid rutland antibody disqualification decatur footballers quake slacks 48th rein scribe stabilize commits exemplary tho ##hort ##chison pantry traversed ##hiti disrepair identifiable vibrated baccalaureate ##nnis csa interviewing ##iensis ##raße greaves wealthiest 343 classed jogged £5 ##58 ##atal illuminating knicks respecting ##uno scrubbed ##iji ##dles kruger moods growls raider silvia chefs kam vr cree percival ##terol gunter counterattack defiant henan ze ##rasia ##riety equivalence submissions ##fra ##thor bautista mechanically ##heater cornice herbal templar ##mering outputs ruining ligand renumbered extravagant mika blockbuster eta insurrection ##ilia darkening ferocious pianos strife kinship ##aer melee ##anor ##iste ##may ##oue decidedly weep ##jad ##missive ##ppel 354 puget unease ##gnant 1629 hammering kassel ob wessex ##lga bromwich egan paranoia utilization ##atable ##idad contradictory provoke ##ols ##ouring ##tangled knesset ##very ##lette plumbing ##sden ##¹ greensboro occult sniff 338 zev beaming gamer haggard mahal ##olt ##pins mendes utmost briefing gunnery ##gut ##pher ##zh ##rok 1679 khalifa sonya ##boot principals urbana wiring ##liffe ##minating ##rrado dahl nyu skepticism np townspeople ithaca lobster somethin ##fur ##arina ##−1 freighter zimmerman biceps contractual ##herton amend hurrying subconscious ##anal 336 meng clermont spawning ##eia ##lub dignitaries impetus snacks spotting twigs ##bilis ##cz ##ouk libertadores nic skylar ##aina ##firm gustave asean ##anum dieter legislatures flirt bromley trolls umar ##bbies ##tyle blah parc bridgeport crank negligence ##nction 46th constantin molded bandages seriousness 00pm siegel carpets compartments upbeat statehood ##dner ##edging marko 730 platt ##hane paving ##iy 1738 abbess impatience limousine nbl ##talk 441 lucille mojo nightfall robbers ##nais karel brisk calves replicate ascribed telescopes ##olf intimidated ##reen ballast specialization ##sit aerodynamic caliphate rainer visionary ##arded epsilon ##aday ##onte aggregation auditory boosted reunification kathmandu loco robyn 402 acknowledges appointing humanoid newell redeveloped restraints ##tained barbarians chopper 1609 italiana ##lez ##lho investigates wrestlemania ##anies ##bib 690 ##falls creaked dragoons gravely minions stupidity volley ##harat ##week musik ##eries ##uously fungal massimo semantics malvern ##ahl ##pee discourage embryo imperialism 1910s profoundly ##ddled jiangsu sparkled stat ##holz sweatshirt tobin ##iction sneered ##cheon ##oit brit causal smyth ##neuve diffuse perrin silvio ##ipes ##recht detonated iqbal selma ##nism ##zumi roasted ##riders tay ##ados ##mament ##mut ##rud 840 completes nipples cfa flavour hirsch ##laus calderon sneakers moravian ##ksha 1622 rq 294 ##imeters bodo ##isance ##pre ##ronia anatomical excerpt ##lke dh kunst ##tablished ##scoe biomass panted unharmed gael housemates montpellier ##59 coa rodents tonic hickory singleton ##taro 451 1719 aldo breaststroke dempsey och rocco ##cuit merton dissemination midsummer serials ##idi haji polynomials ##rdon gs enoch prematurely shutter taunton £3 ##grating ##inates archangel harassed ##asco 326 archway dazzling ##ecin 1736 sumo wat ##kovich 1086 honneur ##ently ##nostic ##ttal ##idon 1605 403 1716 blogger rents ##gnan hires ##ikh ##dant howie ##rons handler retracted shocks 1632 arun duluth kepler trumpeter ##lary peeking seasoned trooper ##mara laszlo ##iciencies ##rti heterosexual ##inatory ##ssion indira jogging ##inga ##lism beit dissatisfaction malice ##ately nedra peeling ##rgeon 47th stadiums 475 vertigo ##ains iced restroom ##plify ##tub illustrating pear ##chner ##sibility inorganic rappers receipts watery ##kura lucinda ##oulos reintroduced ##8th ##tched gracefully saxons nutritional wastewater rained favourites bedrock fisted hallways likeness upscale ##lateral 1580 blinds prequel ##pps ##tama deter humiliating restraining tn vents 1659 laundering recess rosary tractors coulter federer ##ifiers ##plin persistence ##quitable geschichte pendulum quakers ##beam bassett pictorial buffet koln ##sitor drills reciprocal shooters ##57 ##cton ##tees converge pip dmitri donnelly yamamoto aqua azores demographics hypnotic spitfire suspend wryly roderick ##rran sebastien ##asurable mavericks ##fles ##200 himalayan prodigy ##iance transvaal demonstrators handcuffs dodged mcnamara sublime 1726 crazed ##efined ##till ivo pondered reconciled shrill sava ##duk bal cad heresy jaipur goran ##nished 341 lux shelly whitehall ##hre israelis peacekeeping ##wled 1703 demetrius ousted ##arians ##zos beale anwar backstroke raged shrinking cremated ##yck benign towing wadi darmstadt landfill parana soothe colleen sidewalks mayfair tumble hepatitis ferrer superstructure ##gingly ##urse ##wee anthropological translators ##mies closeness hooves ##pw mondays ##roll ##vita landscaping ##urized purification sock thorns thwarted jalan tiberius ##taka saline ##rito confidently khyber sculptors ##ij brahms hammersmith inspectors battista fivb fragmentation hackney ##uls arresting exercising antoinette bedfordshire ##zily dyed ##hema 1656 racetrack variability ##tique 1655 austrians deteriorating madman theorists aix lehman weathered 1731 decreed eruptions 1729 flaw quinlan sorbonne flutes nunez 1711 adored downwards fable rasped 1712 moritz mouthful renegade shivers stunts dysfunction restrain translit 327 pancakes ##avio ##cision ##tray 351 vial ##lden bain ##maid ##oxide chihuahua malacca vimes ##rba ##rnier 1664 donnie plaques ##ually 337 bangs floppy huntsville loretta nikolay ##otte eater handgun ubiquitous ##hett eras zodiac 1634 ##omorphic 1820s ##zog cochran ##bula ##lithic warring ##rada dalai excused blazers mcconnell reeling bot este ##abi geese hoax taxon ##bla guitarists ##icon condemning hunts inversion moffat taekwondo ##lvis 1624 stammered ##rest ##rzy sousa fundraiser marylebone navigable uptown cabbage daniela salman shitty whimper ##kian ##utive programmers protections rm ##rmi ##rued forceful ##enes fuss ##tao ##wash brat oppressive reykjavik spartak ticking ##inkles ##kiewicz adolph horst maui protege straighten cpc landau concourse clements resultant ##ando imaginative joo reactivated ##rem ##ffled ##uising consultative ##guide flop kaitlyn mergers parenting somber ##vron supervise vidhan ##imum courtship exemplified harmonies medallist refining ##rrow ##ка amara ##hum 780 goalscorer sited overshadowed rohan displeasure secretive multiplied osman ##orth engravings padre ##kali ##veda miniatures mis ##yala clap pali rook ##cana 1692 57th antennae astro oskar 1628 bulldog crotch hackett yucatan ##sure amplifiers brno ferrara migrating ##gree thanking turing ##eza mccann ting andersson onslaught gaines ganga incense standardization ##mation sentai scuba stuffing turquoise waivers alloys ##vitt regaining vaults ##clops ##gizing digger furry memorabilia probing ##iad payton rec deutschland filippo opaque seamen zenith afrikaans ##filtration disciplined inspirational ##merie banco confuse grafton tod ##dgets championed simi anomaly biplane ##ceptive electrode ##para 1697 cleavage crossbow swirl informant ##lars ##osta afi bonfire spec ##oux lakeside slump ##culus ##lais ##qvist ##rrigan 1016 facades borg inwardly cervical xl pointedly 050 stabilization ##odon chests 1699 hacked ctv orthogonal suzy ##lastic gaulle jacobite rearview ##cam ##erted ashby ##drik ##igate ##mise ##zbek affectionately canine disperse latham ##istles ##ivar spielberg ##orin ##idium ezekiel cid ##sg durga middletown ##cina customized frontiers harden ##etano ##zzy 1604 bolsheviks ##66 coloration yoko ##bedo briefs slabs debra liquidation plumage ##oin blossoms dementia subsidy 1611 proctor relational jerseys parochial ter ##ici esa peshawar cavalier loren cpi idiots shamrock 1646 dutton malabar mustache ##endez ##ocytes referencing terminates marche yarmouth ##sop acton mated seton subtly baptised beige extremes jolted kristina telecast ##actic safeguard waldo ##baldi ##bular endeavors sloppy subterranean ##ensburg ##itung delicately pigment tq ##scu 1626 ##ound collisions coveted herds ##personal ##meister ##nberger chopra ##ricting abnormalities defective galician lucie ##dilly alligator likened ##genase burundi clears complexion derelict deafening diablo fingered champaign dogg enlist isotope labeling mrna ##erre brilliance marvelous ##ayo 1652 crawley ether footed dwellers deserts hamish rubs warlock skimmed ##lizer 870 buick embark heraldic irregularities ##ajan kiara ##kulam ##ieg antigen kowalski ##lge oakley visitation ##mbit vt ##suit 1570 murderers ##miento ##rites chimneys ##sling condemn custer exchequer havre ##ghi fluctuations ##rations dfb hendricks vaccines ##tarian nietzsche biking juicy ##duced brooding scrolling selangor ##ragan 352 annum boomed seminole sugarcane ##dna departmental dismissing innsbruck arteries ashok batavia daze kun overtook ##rga ##tlan beheaded gaddafi holm electronically faulty galilee fractures kobayashi ##lized gunmen magma aramaic mala eastenders inference messengers bf ##qu 407 bathrooms ##vere 1658 flashbacks ideally misunderstood ##jali ##weather mendez ##grounds 505 uncanny ##iii 1709 friendships ##nbc sacrament accommodated reiterated logistical pebbles thumped ##escence administering decrees drafts ##flight ##cased ##tula futuristic picket intimidation winthrop ##fahan interfered 339 afar francoise morally uta cochin croft dwarfs ##bruck ##dents ##nami biker ##hner ##meral nano ##isen ##ometric ##pres ##ан brightened meek parcels securely gunners ##jhl ##zko agile hysteria ##lten ##rcus bukit champs chevy cuckoo leith sadler theologians welded ##section 1663 jj plurality xander ##rooms ##formed shredded temps intimately pau tormented ##lok ##stellar 1618 charred ems essen ##mmel alarms spraying ascot blooms twinkle ##abia ##apes internment obsidian ##chaft snoop ##dav ##ooping malibu ##tension quiver ##itia hays mcintosh travers walsall ##ffie 1623 beverley schwarz plunging structurally m3 rosenthal vikram ##tsk 770 ghz ##onda ##tiv chalmers groningen pew reckon unicef ##rvis 55th ##gni 1651 sulawesi avila cai metaphysical screwing turbulence ##mberg augusto samba 56th baffled momentary toxin ##urian ##wani aachen condoms dali steppe ##3d ##app ##oed ##year adolescence dauphin electrically inaccessible microscopy nikita ##ega atv ##cel ##enter ##oles ##oteric ##ы accountants punishments wrongly bribes adventurous clinch flinders southland ##hem ##kata gough ##ciency lads soared ##ה undergoes deformation outlawed rubbish ##arus ##mussen ##nidae ##rzburg arcs ##ingdon ##tituted 1695 wheelbase wheeling bombardier campground zebra ##lices ##oj ##bain lullaby ##ecure donetsk wylie grenada ##arding ##ης squinting eireann opposes ##andra maximal runes ##broken ##cuting ##iface ##ror ##rosis additive britney adultery triggering ##drome detrimental aarhus containment jc swapped vichy ##ioms madly ##oric ##rag brant ##ckey ##trix 1560 1612 broughton rustling ##stems ##uder asbestos mentoring ##nivorous finley leaps ##isan apical pry slits substitutes ##dict intuitive fantasia insistent unreasonable ##igen ##vna domed hannover margot ponder ##zziness impromptu jian lc rampage stemming ##eft andrey gerais whichever amnesia appropriated anzac clicks modifying ultimatum cambrian maids verve yellowstone ##mbs conservatoire ##scribe adherence dinners spectra imperfect mysteriously sidekick tatar tuba ##aks ##ifolia distrust ##athan ##zle c2 ronin zac ##pse celaena instrumentalist scents skopje ##mbling comical compensated vidal condor intersect jingle wavelengths ##urrent mcqueen ##izzly carp weasel 422 kanye militias postdoctoral eugen gunslinger ##ɛ faux hospice ##for appalled derivation dwarves ##elis dilapidated ##folk astoria philology ##lwyn ##otho ##saka inducing philanthropy ##bf ##itative geek markedly sql ##yce bessie indices rn ##flict 495 frowns resolving weightlifting tugs cleric contentious 1653 mania rms ##miya ##reate ##ruck ##tucket bien eels marek ##ayton ##cence discreet unofficially ##ife leaks ##bber 1705 332 dung compressor hillsborough pandit shillings distal ##skin 381 ##tat ##you nosed ##nir mangrove undeveloped ##idia textures ##inho ##500 ##rise ae irritating nay amazingly bancroft apologetic compassionate kata symphonies ##lovic airspace ##lch 930 gifford precautions fulfillment sevilla vulgar martinique ##urities looting piccolo tidy ##dermott quadrant armchair incomes mathematicians stampede nilsson ##inking ##scan foo quarterfinal ##ostal shang shouldered squirrels ##owe 344 vinegar ##bner ##rchy ##systems delaying ##trics ars dwyer rhapsody sponsoring ##gration bipolar cinder starters ##olio ##urst 421 signage ##nty aground figurative mons acquaintances duets erroneously soyuz elliptic recreated ##cultural ##quette ##ssed ##tma ##zcz moderator scares ##itaire ##stones ##udence juniper sighting ##just ##nsen britten calabria ry bop cramer forsyth stillness ##л airmen gathers unfit ##umber ##upt taunting ##rip seeker streamlined ##bution holster schumann tread vox ##gano ##onzo strive dil reforming covent newbury predicting ##orro decorate tre ##puted andover ie asahi dept dunkirk gills ##tori buren huskies ##stis ##stov abstracts bets loosen ##opa 1682 yearning ##glio ##sir berman effortlessly enamel napoli persist ##peration ##uez attache elisa b1 invitations ##kic accelerating reindeer boardwalk clutches nelly polka starbucks ##kei adamant huey lough unbroken adventurer embroidery inspecting stanza ##ducted naia taluka ##pone ##roids chases deprivation florian ##jing ##ppet earthly ##lib ##ssee colossal foreigner vet freaks patrice rosewood triassic upstate ##pkins dominates ata chants ks vo ##400 ##bley ##raya ##rmed 555 agra infiltrate ##ailing ##ilation ##tzer ##uppe ##werk binoculars enthusiast fujian squeak ##avs abolitionist almeida boredom hampstead marsden rations ##ands inflated 334 bonuses rosalie patna ##rco 329 detachments penitentiary 54th flourishing woolf ##dion ##etched papyrus ##lster ##nsor ##toy bobbed dismounted endelle inhuman motorola tbs wince wreath ##ticus hideout inspections sanjay disgrace infused pudding stalks ##urbed arsenic leases ##hyl ##rrard collarbone ##waite ##wil dowry ##bant ##edance genealogical nitrate salamanca scandals thyroid necessitated ##! ##" ### ##$ ##% ##& ##' ##( ##) ##* ##+ ##, ##- ##. ##/ ##: ##; ##< ##= ##> ##? ##@ ##[ ##\ ##] ##^ ##_ ##` ##{ ##| ##} ##~ ##¡ ##¢ ##£ ##¤ ##¥ ##¦ ##§ ##¨ ##© ##ª ##« ##¬ ##® ##± ##´ ##µ ##¶ ##· ##º ##» ##¼ ##¾ ##¿ ##æ ##ð ##÷ ##þ ##đ ##ħ ##ŋ ##œ ##ƒ ##ɐ ##ɑ ##ɒ ##ɔ ##ɕ ##ə ##ɡ ##ɣ ##ɨ ##ɪ ##ɫ ##ɬ ##ɯ ##ɲ ##ɴ ##ɹ ##ɾ ##ʀ ##ʁ ##ʂ ##ʃ ##ʉ ##ʊ ##ʋ ##ʌ ##ʎ ##ʐ ##ʑ ##ʒ ##ʔ ##ʰ ##ʲ ##ʳ ##ʷ ##ʸ ##ʻ ##ʼ ##ʾ ##ʿ ##ˈ ##ˡ ##ˢ ##ˣ ##ˤ ##β ##γ ##δ ##ε ##ζ ##θ ##κ ##λ ##μ ##ξ ##ο ##π ##ρ ##σ ##τ ##υ ##φ ##χ ##ψ ##ω ##б ##г ##д ##ж ##з ##м ##п ##с ##у ##ф ##х ##ц ##ч ##ш ##щ ##ъ ##э ##ю ##ђ ##є ##і ##ј ##љ ##њ ##ћ ##ӏ ##ա ##բ ##գ ##դ ##ե ##թ ##ի ##լ ##կ ##հ ##մ ##յ ##ն ##ո ##պ ##ս ##վ ##տ ##ր ##ւ ##ք ##־ ##א ##ב ##ג ##ד ##ו ##ז ##ח ##ט ##י ##ך ##כ ##ל ##ם ##מ ##ן ##נ ##ס ##ע ##ף ##פ ##ץ ##צ ##ק ##ר ##ש ##ת ##، ##ء ##ب ##ت ##ث ##ج ##ح ##خ ##ذ ##ز ##س ##ش ##ص ##ض ##ط ##ظ ##ع ##غ ##ـ ##ف ##ق ##ك ##و ##ى ##ٹ ##پ ##چ ##ک ##گ ##ں ##ھ ##ہ ##ے ##अ ##आ ##उ ##ए ##क ##ख ##ग ##च ##ज ##ट ##ड ##ण ##त ##थ ##द ##ध ##न ##प ##ब ##भ ##म ##य ##र ##ल ##व ##श ##ष ##स ##ह ##ा ##ि ##ी ##ो ##। ##॥ ##ং ##অ ##আ ##ই ##উ ##এ ##ও ##ক ##খ ##গ ##চ ##ছ ##জ ##ট ##ড ##ণ ##ত ##থ ##দ ##ধ ##ন ##প ##ব ##ভ ##ম ##য ##র ##ল ##শ ##ষ ##স ##হ ##া ##ি ##ী ##ে ##க ##ச ##ட ##த ##ந ##ன ##ப ##ம ##ய ##ர ##ல ##ள ##வ ##ா ##ி ##ு ##ே ##ை ##ನ ##ರ ##ಾ ##ක ##ය ##ර ##ල ##ව ##ා ##ก ##ง ##ต ##ท ##น ##พ ##ม ##ย ##ร ##ล ##ว ##ส ##อ ##า ##เ ##་ ##། ##ག ##ང ##ད ##ན ##པ ##བ ##མ ##འ ##ར ##ལ ##ས ##မ ##ა ##ბ ##გ ##დ ##ე ##ვ ##თ ##ი ##კ ##ლ ##მ ##ნ ##ო ##რ ##ს ##ტ ##უ ##ᄀ ##ᄂ ##ᄃ ##ᄅ ##ᄆ ##ᄇ ##ᄉ ##ᄊ ##ᄋ ##ᄌ ##ᄎ ##ᄏ ##ᄐ ##ᄑ ##ᄒ ##ᅡ ##ᅢ ##ᅥ ##ᅦ ##ᅧ ##ᅩ ##ᅪ ##ᅭ ##ᅮ ##ᅯ ##ᅲ ##ᅳ ##ᅴ ##ᅵ ##ᆨ ##ᆫ ##ᆯ ##ᆷ ##ᆸ ##ᆼ ##ᴬ ##ᴮ ##ᴰ ##ᴵ ##ᴺ ##ᵀ ##ᵃ ##ᵇ ##ᵈ ##ᵉ ##ᵍ ##ᵏ ##ᵐ ##ᵒ ##ᵖ ##ᵗ ##ᵘ ##ᵣ ##ᵤ ##ᵥ ##ᶜ ##ᶠ ##‐ ##‑ ##‒ ##– ##— ##― ##‖ ##‘ ##’ ##‚ ##“ ##” ##„ ##† ##‡ ##• ##… ##‰ ##′ ##″ ##› ##‿ ##⁄ ##⁰ ##ⁱ ##⁴ ##⁵ ##⁶ ##⁷ ##⁸ ##⁹ ##⁻ ##ⁿ ##₅ ##₆ ##₇ ##₈ ##₉ ##₊ ##₍ ##₎ ##ₐ ##ₑ ##ₒ ##ₓ ##ₕ ##ₖ ##ₗ ##ₘ ##ₚ ##ₛ ##ₜ ##₤ ##₩ ##€ ##₱ ##₹ ##ℓ ##№ ##ℝ ##™ ##⅓ ##⅔ ##← ##↑ ##→ ##↓ ##↔ ##↦ ##⇄ ##⇌ ##⇒ ##∂ ##∅ ##∆ ##∇ ##∈ ##∗ ##∘ ##√ ##∞ ##∧ ##∨ ##∩ ##∪ ##≈ ##≡ ##≤ ##≥ ##⊂ ##⊆ ##⊕ ##⊗ ##⋅ ##─ ##│ ##■ ##▪ ##● ##★ ##☆ ##☉ ##♠ ##♣ ##♥ ##♦ ##♯ ##⟨ ##⟩ ##ⱼ ##⺩ ##⺼ ##⽥ ##、 ##。 ##〈 ##〉 ##《 ##》 ##「 ##」 ##『 ##』 ##〜 ##あ ##い ##う ##え ##お ##か ##き ##く ##け ##こ ##さ ##し ##す ##せ ##そ ##た ##ち ##っ ##つ ##て ##と ##な ##に ##ぬ ##ね ##の ##は ##ひ ##ふ ##へ ##ほ ##ま ##み ##む ##め ##も ##や ##ゆ ##よ ##ら ##り ##る ##れ ##ろ ##を ##ん ##ァ ##ア ##ィ ##イ ##ウ ##ェ ##エ ##オ ##カ ##キ ##ク ##ケ ##コ ##サ ##シ ##ス ##セ ##タ ##チ ##ッ ##ツ ##テ ##ト ##ナ ##ニ ##ノ ##ハ ##ヒ ##フ ##ヘ ##ホ ##マ ##ミ ##ム ##メ ##モ ##ャ ##ュ ##ョ ##ラ ##リ ##ル ##レ ##ロ ##ワ ##ン ##・ ##ー ##一 ##三 ##上 ##下 ##不 ##世 ##中 ##主 ##久 ##之 ##也 ##事 ##二 ##五 ##井 ##京 ##人 ##亻 ##仁 ##介 ##代 ##仮 ##伊 ##会 ##佐 ##侍 ##保 ##信 ##健 ##元 ##光 ##八 ##公 ##内 ##出 ##分 ##前 ##劉 ##力 ##加 ##勝 ##北 ##区 ##十 ##千 ##南 ##博 ##原 ##口 ##古 ##史 ##司 ##合 ##吉 ##同 ##名 ##和 ##囗 ##四 ##国 ##國 ##土 ##地 ##坂 ##城 ##堂 ##場 ##士 ##夏 ##外 ##大 ##天 ##太 ##夫 ##奈 ##女 ##子 ##学 ##宀 ##宇 ##安 ##宗 ##定 ##宣 ##宮 ##家 ##宿 ##寺 ##將 ##小 ##尚 ##山 ##岡 ##島 ##崎 ##川 ##州 ##巿 ##帝 ##平 ##年 ##幸 ##广 ##弘 ##張 ##彳 ##後 ##御 ##德 ##心 ##忄 ##志 ##忠 ##愛 ##成 ##我 ##戦 ##戸 ##手 ##扌 ##政 ##文 ##新 ##方 ##日 ##明 ##星 ##春 ##昭 ##智 ##曲 ##書 ##月 ##有 ##朝 ##木 ##本 ##李 ##村 ##東 ##松 ##林 ##森 ##楊 ##樹 ##橋 ##歌 ##止 ##正 ##武 ##比 ##氏 ##民 ##水 ##氵 ##氷 ##永 ##江 ##沢 ##河 ##治 ##法 ##海 ##清 ##漢 ##瀬 ##火 ##版 ##犬 ##王 ##生 ##田 ##男 ##疒 ##発 ##白 ##的 ##皇 ##目 ##相 ##省 ##真 ##石 ##示 ##社 ##神 ##福 ##禾 ##秀 ##秋 ##空 ##立 ##章 ##竹 ##糹 ##美 ##義 ##耳 ##良 ##艹 ##花 ##英 ##華 ##葉 ##藤 ##行 ##街 ##西 ##見 ##訁 ##語 ##谷 ##貝 ##貴 ##車 ##軍 ##辶 ##道 ##郎 ##郡 ##部 ##都 ##里 ##野 ##金 ##鈴 ##镇 ##長 ##門 ##間 ##阝 ##阿 ##陳 ##陽 ##雄 ##青 ##面 ##風 ##食 ##香 ##馬 ##高 ##龍 ##龸 ##fi ##fl ##! ##( ##) ##, ##- ##. ##/ ##: ##? ##~
Tools/DGLPyTorch/SyntheticGraphGeneration/syngen/cli
cli
__init__
# Copyright (c) 2023, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from syngen.cli.commands.synthesize import SynthesizeCommand from syngen.cli.commands.preprocess import PreprocessingCommand from syngen.cli.commands.mimic_dataset import MimicDatasetCommand from syngen.cli.commands.pretrain import PretrainCommand def get_parser(): parser = argparse.ArgumentParser( description="Synthetic Graph Generation Tool", formatter_class=argparse.ArgumentDefaultsHelpFormatter ) command = parser.add_subparsers(title="command") command.required = True SynthesizeCommand().init_parser(command) PreprocessingCommand().init_parser(command) MimicDatasetCommand().init_parser(command) PretrainCommand().init_parser(command) return parser
CUDA-Optimized/FastSpeech/fastspeech/trt/plugins/repeat
repeat
test_repeat_plugin
# Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # * Neither the name of the NVIDIA CORPORATION nor the # names of its contributors may be used to endorse or promote products # derived from this software without specific prior written permission. # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL NVIDIA CORPORATION BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. from __future__ import print_function import numpy as np import tensorrt as trt import pycuda.autoinit import pycuda.driver as cuda import ctypes import os import time import sys logger = trt.Logger(trt.Logger.INFO) PLUGIN_PATH = '/home/dahn/git/fastspeech/fastspeech/trt/plugins/repeat/RepeatPlugin.so' ctypes.cdll.LoadLibrary(PLUGIN_PATH) def get_plugin_creator(plugin_name): trt.init_libnvinfer_plugins(logger, '') plugin_creator_list = trt.get_plugin_registry().plugin_creator_list plugin_creator = None for c in plugin_creator_list: if c.name == plugin_name: plugin_creator = c return plugin_creator def build_engine(shape, shape2): plugin_creator = get_plugin_creator('RepeatPlugin') if plugin_creator == None: print('Plugin not found. Exiting') exit() builder = trt.Builder(logger) builder.max_batch_size = 1024 builder.max_workspace_size = 1 << 20 builder.fp16_mode = use_fp16 network = builder.create_network() tensor = network.add_input('input1', trt.DataType.FLOAT, shape) tensor2 = network.add_input('input2', trt.DataType.FLOAT, shape2) tensor = network.add_plugin_v2( [tensor, tensor2], plugin_creator.create_plugin('RepeatPlugin', trt.PluginFieldCollection([ trt.PluginField('maxOutputLength', np.array([MAX_OUTPUT_LENGTH], dtype=np.int32), trt.PluginFieldType.INT32) ])) ).get_output(0) network.mark_output(tensor) return builder.build_cuda_engine(network) def run_trt(input1, input2): batch_size = input1.shape[0] engine = build_engine(input1.shape[1:], input2.shape[1:]) context = engine.create_execution_context() d_input1 = cuda.mem_alloc(input1.nbytes) d_input2 = cuda.mem_alloc(input2.nbytes) output = np.zeros(shape=(batch_size, MAX_OUTPUT_LENGTH, input1.shape[2]), dtype=np.float32) d_output = cuda.mem_alloc(output.nbytes) cuda.memcpy_htod(d_input1, input1) cuda.memcpy_htod(d_input2, input2) bindings = [int(d_input1), int(d_input2), int(d_output)] start = time.time() context.execute(batch_size, bindings) end = time.time() time_elapsed = end - start print("time elapsed: {:06f}".format(time_elapsed)) cuda.memcpy_dtoh(output, d_output) return output use_fp16 = len(sys.argv) > 1 and sys.argv[1].isdigit() and int(sys.argv[1]) == 1 print('Use FP16:', use_fp16) ## # accuray test ## MAX_OUTPUT_LENGTH=8 inputs = np.array([ [[1, 2], [4, 5], [7, 8]], [[3, 4], [5, 6], [8, 9]] ], np.float32) masks = np.ones((2,3,1), np.float32) repeats = np.array([ [[0, 2, 10]], [[1, 2, 1]] ], np.float32) output = run_trt(inputs, repeats) print(output) print(output.shape) print(type(output)) output_mask = run_trt(masks, repeats) print(output_mask) print(output_mask.shape) print(type(output_mask)) ## # latency test ## # MAX_OUTPUT_LENGTH=1024 # inputs = np.full((16, 256, 384), 2, np.float32) # masks = np.ones((16, 256, 384), np.float32) # repeats = np.full((16, 256), 4, np.float32) # output = run_trt(inputs, repeats) # output_mask = run_trt(masks, repeats)
PyTorch/Classification/ConvNets/se-resnext101-32x4d/training/AMP
AMP
DGX1V_se-resnext101-32x4d_AMP_90E
python ./multiproc.py --nproc_per_node 8 ./launch.py --model se-resnext101-32x4d --precision AMP --mode convergence --platform DGX1V /imagenet --epochs 90 --mixup 0.0 --workspace ${1:-./} --raport-file raport.json
PyTorch/LanguageModeling/BERT/triton/deployment_toolkit/bermuda
bermuda
tensorrt
# Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import logging import sys from pathlib import Path from typing import Dict, NamedTuple, Optional, Union import numpy as np # pytype: disable=import-error try: import pycuda.autoinit import pycuda.driver as cuda except Exception as e: logging.getLogger(__name__).warning(f"Problems with importing pycuda package; {e}") # pytype: enable=import-error import tensorrt as trt # pytype: disable=import-error from ..core import BaseLoader, BaseRunner, BaseRunnerSession, Format, Model, TensorSpec from ..extensions import loaders, runners LOGGER = logging.getLogger(__name__) TRT_LOGGER = trt.Logger(trt.Logger.INFO) # documentation: # https://docs.nvidia.com/deeplearning/tensorrt/api/python_api/index.html # https://docs.nvidia.com/deeplearning/tensorrt/developer-guide/index.html#python_samples_section _NP_DTYPE2TRT_DTYPE = { np.dtype("float32"): trt.DataType.FLOAT, np.dtype("float16"): trt.DataType.HALF, np.dtype("int8"): trt.DataType.INT8, np.dtype("int32"): trt.DataType.INT32, np.dtype("bool"): trt.DataType.BOOL, } class TensorRTLoader(BaseLoader): def load(self, model_path: Union[str, Path], **_) -> Model: model_path = Path(model_path) LOGGER.debug(f"Loading TensorRT engine from {model_path}") engine = self._load_engine(model_path) if engine is None: LOGGER.debug("Unable to load engine without plugins. Loading plugins.") trt.init_libnvinfer_plugins(logger=TRT_LOGGER, namespace="") LOGGER.debug(f"Loading TensorRT engine with plugins from {model_path}") engine = self._load_engine(model_path) if engine is None: raise RuntimeError(f"Could not load ICudaEngine from {model_path}") inputs = {} outputs = {} for binding_idx in range(engine.num_bindings): name = engine.get_binding_name(binding_idx) is_input = engine.binding_is_input(binding_idx) dtype = np.dtype(trt.nptype(engine.get_binding_dtype(binding_idx))).name shape = engine.get_binding_shape(binding_idx) if is_input: inputs[name] = TensorSpec(name, dtype, shape) else: outputs[name] = TensorSpec(name, dtype, shape) return Model(engine, None, inputs, outputs) def _load_engine(self, model_path: Path): with model_path.open("rb") as fh, trt.Runtime(TRT_LOGGER) as runtime: engine = runtime.deserialize_cuda_engine(fh.read()) return engine class TRTBuffers(NamedTuple): x_host: Optional[Dict[str, object]] x_dev: Dict[str, object] y_pred_host: Dict[str, object] y_pred_dev: Dict[str, object] class TensorRTRunner(BaseRunner): def __init__(self): pass def init_inference(self, model: Model): return TensorRTRunnerSession(model=model) class TensorRTRunnerSession(BaseRunnerSession): def __init__(self, model: Model): super().__init__(model) assert isinstance(model.handle, trt.ICudaEngine) self._model = model self._has_dynamic_shapes = None self._context = None self._engine: trt.ICudaEngine = self._model.handle self._cuda_context = pycuda.autoinit.context self._input_names = None self._output_names = None self._buffers = None def __enter__(self): self._context = self._engine.create_execution_context() self._context.__enter__() self._input_names = [ self._engine[idx] for idx in range(self._engine.num_bindings) if self._engine.binding_is_input(idx) ] self._output_names = [ self._engine[idx] for idx in range(self._engine.num_bindings) if not self._engine.binding_is_input(idx) ] # all_binding_shapes_specified is True for models without dynamic shapes # so initially this variable is False for models with dynamic shapes self._has_dynamic_shapes = not self._context.all_binding_shapes_specified return self def __exit__(self, exc_type, exc_value, traceback): self._context.__exit__(exc_type, exc_value, traceback) self._input_names = None self._output_names = None # TODO: are cuda buffers dealloc automatically? self._buffers = None def __call__(self, x): buffers = self._prepare_buffers_if_needed(x) bindings = self._update_bindings(buffers) for name in self._input_names: cuda.memcpy_htod(buffers.x_dev[name], buffers.x_host[name]) self._cuda_context.push() self._context.execute_v2(bindings=bindings) self._cuda_context.pop() for name in self._output_names: cuda.memcpy_dtoh(buffers.y_pred_host[name], buffers.y_pred_dev[name]) return buffers.y_pred_host def _update_bindings(self, buffers: TRTBuffers): bindings = [None] * self._engine.num_bindings for name in buffers.y_pred_dev: binding_idx: int = self._engine[name] bindings[binding_idx] = buffers.y_pred_dev[name] for name in buffers.x_dev: binding_idx: int = self._engine[name] bindings[binding_idx] = buffers.x_dev[name] return bindings def _set_dynamic_input_shapes(self, x_host): def _is_shape_dynamic(input_shape): return any([dim is None or dim == -1 for dim in input_shape]) for name in self._input_names: bindings_idx = self._engine[name] data_shape = x_host[name].shape # pytype: disable=attribute-error if self._engine.is_shape_binding(bindings_idx): input_shape = self._context.get_shape(bindings_idx) if _is_shape_dynamic(input_shape): self._context.set_shape_input(bindings_idx, data_shape) else: input_shape = self._engine.get_binding_shape(bindings_idx) if _is_shape_dynamic(input_shape): self._context.set_binding_shape(bindings_idx, data_shape) assert self._context.all_binding_shapes_specified and self._context.all_shape_inputs_specified def _prepare_buffers_if_needed(self, x_host: Dict[str, object]): # pytype: disable=attribute-error new_batch_size = list(x_host.values())[0].shape[0] current_batch_size = list(self._buffers.y_pred_host.values())[0].shape[0] if self._buffers else 0 # pytype: enable=attribute-error if self._has_dynamic_shapes or new_batch_size != current_batch_size: # TODO: are CUDA buffers dealloc automatically? self._set_dynamic_input_shapes(x_host) y_pred_host = {} for name in self._output_names: shape = self._context.get_binding_shape(self._engine[name]) binding_idx: int = self._engine[name] dtype_from_trt_binding = np.dtype(trt.nptype(self._engine.get_binding_dtype(binding_idx))) dtype_from_model_spec = np.dtype(self._model.outputs[name].dtype) assert dtype_from_model_spec == dtype_from_trt_binding y_pred_host[name] = np.zeros(shape, dtype=dtype_from_model_spec) y_pred_dev = {name: cuda.mem_alloc(data.nbytes) for name, data in y_pred_host.items()} # cast host input into binding dtype def _cast_input(name, data): binding_idx: int = self._engine[name] np_dtype = trt.nptype(self._engine.get_binding_dtype(binding_idx)) return data.astype(np_dtype) x_host = {name: _cast_input(name, host_input) for name, host_input in x_host.items()} x_dev = { name: cuda.mem_alloc(host_input.nbytes) for name, host_input in x_host.items() if name in self._input_names # pytype: disable=attribute-error } self._buffers = TRTBuffers(None, x_dev, y_pred_host, y_pred_dev) return self._buffers._replace(x_host=x_host) if "pycuda.driver" in sys.modules: loaders.register_extension(Format.TRT.value, TensorRTLoader) runners.register_extension(Format.TRT.value, TensorRTRunner) else: LOGGER.warning("Do not register TensorRT extension due problems with importing pycuda.driver package.")
PyTorch/Forecasting/TFT
TFT
tft_torchhub
# Copyright (c) 2022, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os import sys import urllib.request from zipfile import ZipFile import torch from torch.utils.data import DataLoader NGC_CHECKPOINT_URLS = {} NGC_CHECKPOINT_URLS["electricity"] = "https://api.ngc.nvidia.com/v2/models/nvidia/dle/tft_base_pyt_ckpt_ds-electricity/versions/22.11.0_amp/zip" NGC_CHECKPOINT_URLS["traffic"] = "https://api.ngc.nvidia.com/v2/models/nvidia/dle/tft_base_pyt_ckpt_ds-traffic/versions/22.11.0_amp/zip" def _download_checkpoint(checkpoint, force_reload): model_dir = os.path.join(torch.hub._get_torch_home(), 'checkpoints') if not os.path.exists(model_dir): os.makedirs(model_dir) ckpt_file = os.path.join(model_dir, os.path.basename(checkpoint)) if not os.path.exists(ckpt_file) or force_reload: sys.stderr.write('Downloading checkpoint from {}\n'.format(checkpoint)) urllib.request.urlretrieve(checkpoint, ckpt_file) with ZipFile(ckpt_file, "r") as zf: zf.extractall(path=model_dir) return os.path.join(model_dir, "checkpoint.pt") def nvidia_tft(pretrained=True, **kwargs): from .modeling import TemporalFusionTransformer """Constructs a TFT model. For detailed information on model input and output, training recipies, inference and performance visit: github.com/NVIDIA/DeepLearningExamples and/or ngc.nvidia.com Args (type[, default value]): pretrained (bool, True): If True, returns a pretrained model. dataset (str, 'electricity'): loads selected model type electricity or traffic. Defaults to electricity """ ds_type = kwargs.get("dataset", "electricity") ckpt = _download_checkpoint(NGC_CHECKPOINT_URLS[ds_type], True) state_dict = torch.load(ckpt) config = state_dict['config'] model = TemporalFusionTransformer(config) if pretrained: model.load_state_dict(state_dict['model']) model.eval() return model def nvidia_tft_data_utils(**kwargs): from .data_utils import TFTDataset from .configuration import ElectricityConfig class Processing: @staticmethod def download_data(path): if not os.path.exists(os.path.join(path, "raw")): os.makedirs(os.path.join(path, "raw"), exist_ok=True) dataset_url = "https://archive.ics.uci.edu/ml/machine-learning-databases/00321/LD2011_2014.txt.zip" ckpt_file = os.path.join(path, "raw/electricity.zip") if not os.path.exists(ckpt_file): sys.stderr.write('Downloading checkpoint from {}\n'.format(dataset_url)) urllib.request.urlretrieve(dataset_url, ckpt_file) with ZipFile(ckpt_file, "r") as zf: zf.extractall(path=os.path.join(path, "raw/electricity/")) @staticmethod def preprocess(path): config = ElectricityConfig() if not os.path.exists(os.path.join(path, "processed")): os.makedirs(os.path.join(path, "processed"), exist_ok=True) from data_utils import standarize_electricity as standarize from data_utils import preprocess standarize(os.path.join(path, "raw/electricity")) preprocess(os.path.join(path, "raw/electricity/standarized.csv"), os.path.join(path, "processed/electricity_bin/"), config) @staticmethod def get_batch(path): config = ElectricityConfig() test_split = TFTDataset(os.path.join(path, "processed/electricity_bin/", "test.csv"), config) data_loader = DataLoader(test_split, batch_size=16, num_workers=0) for i, batch in enumerate(data_loader): if i == 40: break return batch return Processing()
PyTorch/SpeechRecognition/wav2vec2/scripts/docker
docker
build
#!/usr/bin/env bash # Copyright (c) 2023, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. docker build . --rm -t wav2vec2
TensorFlow2/Classification/ConvNets/efficientnet_v1
efficientnet_v1
README
# EfficientNet v1 For TensorFlow 2.6 This repository provides scripts and recipes to train EfficientNet v1-B0 and v1-B4 to achieve state-of-the-art accuracy. The content of the repository is maintained by NVIDIA and is tested against each NGC monthly released container to ensure consistent accuracy and performance over time. ## Table Of Contents - [Model overview](#model-overview) * [Model architecture](#model-architecture) * [Default configuration](#default-configuration) * [Feature support matrix](#feature-support-matrix) * [Features](#features) * [Mixed precision training](#mixed-precision-training) * [Enabling mixed precision](#enabling-mixed-precision) * [Enabling TF32](#enabling-tf32) - [Setup](#setup) * [Requirements](#requirements) - [Quick Start Guide](#quick-start-guide) - [Advanced](#advanced) * [Scripts and sample code](#scripts-and-sample-code) * [Parameters](#parameters) * [Command-line options](#command-line-options) * [Getting the data](#getting-the-data) * [Training process](#training-process) * [Multi-node](#multi-node) * [Inference process](#inference-process) - [Performance](#performance) * [Benchmarking](#benchmarking) * [Training performance benchmark](#training-performance-benchmark) * [Inference performance benchmark](#inference-performance-benchmark) * [Results](#results) * [Training accuracy results for EfficientNet v1-B0](#training-accuracy-results-for-efficientnet-v1-b0) * [Training accuracy: NVIDIA DGX A100 (8x A100 80GB)](#training-accuracy-nvidia-dgx-a100-8x-a100-80gb) * [Training accuracy: NVIDIA DGX-1 (8x V100 32GB)](#training-accuracy-nvidia-dgx-1-8x-v100-32gb) * [Training accuracy results for EfficientNet v1-B4](#training-accuracy-results-for-efficientnet-v1-b4) * [Training accuracy: NVIDIA DGX A100 (8x A100 80GB)](#training-accuracy-nvidia-dgx-a100-8x-a100-80gb) * [Training accuracy: NVIDIA DGX-1 (8x V100 32GB)](#training-accuracy-nvidia-dgx-1-8x-v100-32gb) * [Training performance results for EfficientNet v1-B0](#training-performance-results-for-efficientnet-v1-b0) * [Training performance: NVIDIA DGX A100 (8x A100 80GB)](#training-performance-nvidia-dgx-a100-8x-a100-80gb) * [Training performance: NVIDIA DGX-1 (8x V100 32GB)](#training-performance-nvidia-dgx-1-8x-v100-32gb) * [Training performance results for EfficientNet v1-B4](#training-performance-results-for-efficientnet-v1-b4) * [Training performance: NVIDIA DGX A100 (8x A100 80GB)](#training-performance-nvidia-dgx-a100-8x-a100-80gb) * [Training performance: NVIDIA DGX-1 (8x V100 32GB)](#training-performance-nvidia-dgx-1-8x-v100-32gb) * [Inference performance results for EfficientNet v1-B0](#inference-performance-results-for-efficientnet-v1-b0) * [Inference performance: NVIDIA DGX A100 (1x A100 80GB)](#inference-performance-nvidia-dgx-a100-1x-a100-80gb) * [Inference performance: NVIDIA DGX-1 (1x V100 32GB)](#inference-performance-nvidia-dgx-1-1x-v100-32gb) * [Inference performance results for EfficientNet v1-B4](#inference-performance-results-for-efficientnet-v1-b4) * [Inference performance: NVIDIA DGX A100 (1x A100 80GB)](#inference-performance-nvidia-dgx-a100-1x-a100-80gb) * [Inference performance: NVIDIA DGX-1 (1x V100 32GB)](#inference-performance-nvidia-dgx-1-1x-v100-32gb) - [Release notes](#release-notes) * [Changelog](#changelog) * [Known issues](#known-issues) ## Model overview ### Model architecture EfficientNet v1 is developed based on AutoML and Compound Scaling. In particular, a mobile-size baseline network called EfficientNet v1-B0 is developed from AutoML MNAS Mobile framework, the building block is mobile inverted bottleneck MBConv with squeeze-and-excitation optimization. Then, through a compound scaling method, this baseline is scaled up to obtain EfficientNet v1-B1 to B7. ![Efficientnet_structure](https://1.bp.blogspot.com/-Cdtb97FtgdA/XO3BHsB7oEI/AAAAAAAAEKE/bmtkonwgs8cmWyI5esVo8wJPnhPLQ5bGQCLcBGAs/s1600/image4.png) ### Default configuration Here is the Baseline EfficientNet v1-B0 structure. ![Efficientnet v1-B0](https://miro.medium.com/max/1106/1*5oQHqmvS_q9Pq_lZ_Rv51A.png) The following features are supported by this implementation: - General: - XLA support - Mixed precision support - Multi-GPU support using Horovod - Multi-node support using Horovod - Cosine LR Decay - Inference: - Support for inference on a single image is included - Support for inference on a batch of images is included ### Feature support matrix | Feature | EfficientNet |-----------------------|-------------------------- | |Horovod Multi-GPU training (NCCL) | Yes | |Multi-GPU training | Yes | |Multi-node training | Yes | |Automatic mixed precision (AMP) | Yes | |XLA | Yes | |Gradient Accumulation| Yes | |Stage-wise Training| Yes | #### Features **Multi-GPU training with Horovod** Our model uses Horovod to implement efficient multi-GPU training with NCCL. For details, refer to example sources in this repository or refer to the [TensorFlow tutorial](https://github.com/horovod/horovod/#usage). **Multi-node training with Horovod** Our model also uses Horovod to implement efficient multi-node training. **Automatic Mixed Precision (AMP)** Computation graphs can be modified by TensorFlow on runtime to support mixed precision training. A detailed explanation of mixed precision can be found in Appendix. **Gradient Accumulation** Gradient Accumulation is supported through a custom train_step function. This feature is enabled only when grad_accum_steps is greater than 1. ### Mixed precision training Mixed precision is the combined use of different numerical precisions in a computational method. [Mixed precision](https://arxiv.org/abs/1710.03740) training offers significant computational speedup by performing operations in half-precision format while storing minimal information in single-precision to retain as much information as possible in critical parts of the network. Since the introduction of [Tensor Cores](https://developer.nvidia.com/tensor-cores) in NVIDIA Volta, and following with both the NVIDIA Turing and NVIDIA Ampere architectures, significant training speedups are experienced by switching to mixed precision -- up to 3x overall speedup on the most arithmetically intense model architectures. Using [mixed precision training](https://docs.nvidia.com/deeplearning/performance/mixed-precision-training/index.html) previously required two steps: 1. Porting the model to use the FP16 data type where appropriate. 2. Adding loss scaling to preserve small gradient values. This can now be achieved using Automatic Mixed Precision (AMP) for TensorFlow to enable the full [mixed precision methodology](https://docs.nvidia.com/deeplearning/sdk/mixed-precision-training/index.html#tensorflow) in your existing TensorFlow model code. AMP enables mixed precision training on NVIDIA Volta, NVIDIA Turing, and NVIDIA Ampere GPU architectures automatically. The TensorFlow framework code makes all necessary model changes internally. In TF-AMP, the computational graph is optimized to use as few casts as necessary and maximize the use of FP16, and the loss scaling is automatically applied inside of supported optimizers. AMP can be configured to work with the existing tf.contrib loss scaling manager by disabling the AMP scaling with a single environment variable to perform only the automatic mixed-precision optimization. It accomplishes this by automatically rewriting all computation graphs with the necessary operations to enable mixed precision training and automatic loss scaling. For information about: - How to train using mixed precision, refer to the [Mixed Precision Training](https://arxiv.org/abs/1710.03740) paper and [Training With Mixed Precision](https://docs.nvidia.com/deeplearning/performance/mixed-precision-training/index.html) documentation. - Techniques used for mixed precision training, refer to the [Mixed-Precision Training of Deep Neural Networks](https://devblogs.nvidia.com/mixed-precision-training-deep-neural-networks/) blog. - How to access and enable AMP for TensorFlow, refer to [Using TF-AMP](https://docs.nvidia.com/deeplearning/dgx/tensorflow-user-guide/index.html#tfamp) from the TensorFlow User Guide. #### Enabling mixed precision Mixed precision is enabled in TensorFlow by using the Automatic Mixed Precision (TF-AMP) extension, which casts variables to half-precision upon retrieval while storing variables in single-precision format. Furthermore, to preserve small gradient magnitudes in backpropagation, a [loss scaling](https://docs.nvidia.com/deeplearning/sdk/mixed-precision-training/index.html#lossscaling) step must be included when applying gradients. In TensorFlow, loss scaling can be applied statically by using simple multiplication of loss by a constant value or automatically, by TF-AMP. Automatic mixed precision makes all the adjustments internally in TensorFlow, providing two benefits over manual operations. First, programmers need not modify network model code, reducing development and maintenance effort. Second, using AMP maintains forward and backward compatibility with all the APIs for defining and running TensorFlow models. To enable mixed precision, you can simply add the `--use_amp` to the command-line used to run the model. This will enable the following code: ``` if params.use_amp: policy = tf.keras.mixed_precision.experimental.Policy('mixed_float16', loss_scale='dynamic') tf.keras.mixed_precision.experimental.set_policy(policy) ``` #### Enabling TF32 TensorFloat-32 (TF32) is the new math mode in [NVIDIA A100](https://www.nvidia.com/en-us/data-center/a100/) GPUs for handling the matrix math, also called tensor operations. TF32 running on Tensor Cores in A100 GPUs can provide up to 10x speedups compared to single-precision floating-point math (FP32) on NVIDIA Volta GPUs. TF32 Tensor Cores can speed up networks using FP32, typically with no loss of accuracy. It is more robust than FP16 for models which require a high dynamic range for weights or activations. For more information, refer to the [TensorFloat-32 in the A100 GPU Accelerates AI Training, HPC up to 20x](https://blogs.nvidia.com/blog/2020/05/14/tensorfloat-32-precision-format/) blog post. TF32 is supported in the NVIDIA Ampere GPU architecture and is enabled by default. ## Setup The following section lists the requirements that you need to meet in order to start training the EfficientNet model. # Requirements This repository contains Dockerfile which extends the TensorFlow NGC container and encapsulates some dependencies. Aside from these dependencies, ensure you have the following components: - [NVIDIA Docker](https://github.com/NVIDIA/nvidia-docker) - [TensorFlow 21.09-py3] NGC container or later - Supported GPUs: - [NVIDIA Volta architecture](https://www.nvidia.com/en-us/data-center/volta-gpu-architecture/) - [NVIDIA Turing architecture](https://www.nvidia.com/en-us/geforce/turing/) - [NVIDIA Ampere architecture](https://www.nvidia.com/en-us/data-center/nvidia-ampere-gpu-architecture/) For more information about how to get started with NGC containers, refer to the following sections from the NVIDIA GPU Cloud Documentation and the Deep Learning Documentation: - [Getting Started Using NVIDIA GPU Cloud](https://docs.nvidia.com/ngc/ngc-getting-started-guide/index.html) - [Accessing And Pulling From The NGC Container Registry](https://docs.nvidia.com/deeplearning/frameworks/user-guide/index.html#accessing_registry) - [Running TensorFlow](https://docs.nvidia.com/deeplearning/frameworks/tensorflow-release-notes/running.html#running) As an alternative to the use of the Tensorflow2 NGC container, to set up the required environment or create your own container, refer to the versioned [NVIDIA Container Support Matrix](https://docs.nvidia.com/deeplearning/frameworks/support-matrix/index.html). For multi-node, the sample provided in this repository requires [Enroot](https://github.com/NVIDIA/enroot) and [Pyxis](https://github.com/NVIDIA/pyxis) set up on a [SLURM](https://slurm.schedmd.com) cluster. ## Quick Start Guide To train your model using mixed or TF32 precision with Tensor Cores or using FP32, perform the following steps using the default parameters of the EfficientNet model on the ImageNet dataset. For the specifics concerning training and inference, refer to the [Advanced](#advanced) section. 1. Clone the repository. ``` git clone https://github.com/NVIDIA/DeepLearningExamples.git cd DeepLearningExamples/TensorFlow2/Classification/ConvNets/efficientnet ``` 2. Download and prepare the dataset. `Runner.py` supports ImageNet with [TensorFlow Datasets (TFDS)](https://www.tensorflow.org/datasets/overview). Refer to the [TFDS ImageNet readme](https://github.com/tensorflow/datasets/blob/master/docs/catalog/imagenet2012.md) for manual download instructions. 3. Build EfficientNet on top of the NGC container. `bash ./scripts/docker/build.sh YOUR_DESIRED_CONTAINER_NAME` 4. Start an interactive session in the NGC container to run training/inference. **Ensure that `launch.sh` has the correct path to ImageNet on your machine and that this path is mounted onto the `/data` directory because this is where training and evaluation scripts search for data.** `bash ./scripts/docker/launch.sh YOUR_DESIRED_CONTAINER_NAME` 5. Start training. To run training for a standard configuration, **under the container default entry point `/workspace`**, run one of the scripts in the `./efficinetnet_v1/{B0,B4}/training/{AMP,TF32,FP32}/convergence_8x{A100-80G, V100-32G}.sh`. For example: `bash ./efficinetnet_v1/B0/training/AMP/convergence_8xA100-80G.sh` 6. Start validation/evaluation. To run validation/evaluation for a standard configuration, **under the container default entry point `/workspace`**, run one of the scripts in the `./efficinetnet_v1/{B0,B4}/evaluation/evaluation_{AMP,FP32,TF32}_8x{A100-80G, V100-32G}.sh`. The evaluation script is configured to use the checkpoint specified in `checkpoint` for evaluation. The specified checkpoint will be read from the location passed by `--model_dir'.For example: `bash ./efficinetnet_v1/B0/evaluation/evaluation_AMP_A100-80G.sh` 7. Start inference/predictions. To run inference for a standard configuration, **under the container default entry point `/workspace`**, run one of the scripts in the `./efficinetnet_v1/{B0,B4}/inference/inference_{AMP,FP32,TF32}.sh`. Ensure your JPEG images used to run inference on are mounted in the `/infer_data` directory with this folder structure : ``` infer_data | ├── images | | ├── image1.JPEG | | ├── image2.JPEG ``` For example: `bash ./efficinetnet_v1/B0/inference/inference_AMP.sh` Now that you have your model trained and evaluated, you can choose to compare your training results with our [Training accuracy results](#training-accuracy-results). You can also choose to benchmark your performance to [Training performance benchmark](#training-performance-results) or [Inference performance benchmark](#inference-performance-results). Following the steps in these sections will ensure that you achieve the same accuracy and performance results as stated in the [Results](#results) section. ## Advanced The following sections provide greater details of the dataset, running training and inference, and the training results. ### Scripts and sample code The repository is structured as follows: - `scripts/` - shell scripts to build and launch EfficientNet container on top of NGC container - `efficientnet_{v1,v2}` scripts to launch training, evaluation and inference - `model/` - building blocks and EfficientNet model definitions - `runtime/` - detailed procedure for each running mode - `utils/` - support util functions for learning rates, optimizers, etc. - `dataloader/` provides data pipeline utils - `config/` contains model definitions ### Parameters The hyper parameters can be grouped into model-specific hyperparameters (e.g., #layers ) and general hyperparameters (e.g., #training epochs). The model-specific hyperparameters are to be defined in a python module, which must be passed in the command line via --cfg ( `python main.py --cfg config/efficientnet_v1/b0_cfg.py`). To override model-specific hyperparameters, you can use a comma-separated list of k=v pairs (e.g., `python main.py --cfg config/efficientnet_v1/b0_cfg.py --mparams=bn_momentum=0.9,dropout=0.5`). The general hyperparameters and their default values can be found in `utils/cmdline_helper.py`. The user can override these hyperparameters in the command line (e.g., `python main.py --cfg config/efficientnet_v1/b0_cfg.py --data_dir xx --train_batch_size 128`). Here is a list of important hyperparameters: - `--mode` (`train_and_eval`,`train`,`eval`,`prediction`) - the default is `train_and_eval`. - `--use_amp` Set to True to enable AMP - `--use_xla` Set to True to enable XLA - `--model_dir` The folder where model checkpoints are saved (the default is `/workspace/output`) - `--data_dir` The folder where data resides (the default is `/data/`) - `--log_steps` The interval of steps between logging of batch level stats. - `--augmenter_name` Type of data augmentation - `--raug_num_layers` Number of layers used in the random data augmentation scheme - `--raug_magnitude` Strength of transformations applied in the random data augmentation scheme - `--cutmix_alpha` Cutmix parameter used in the last stage of training. - `--mixup_alpha` Mixup parameter used in the last stage of training. - `--defer_img_mixing` Move image mixing ops from the data loader to the model/GPU (faster training) - `--eval_img_size` Size of images used for evaluation - `--eval_batch_size` The evaluation batch size per GPU - `--n_stages` Number of stages used for stage-wise training - `--train_batch_size` The training batch size per GPU - `--train_img_size` Size of images used in the last stage of training - `--base_img_size` Size of images used in the first stage of training - `--max_epochs` The number of training epochs - `--warmup_epochs` The number of epochs of warmup - `--moving_average_decay` The decay weight used for EMA - `--lr_init` The learning rate for a batch size of 128, effective learning rate will be automatically scaled according to the global training batch size: `lr=lr_init * global_BS/128 where global_BS=train_batch_size*n_GPUs` - `--lr_decay` Learning rate decay policy - `--weight_decay` Weight decay coefficient - `--save_checkpoint_freq` Number of epochs to save checkpoints **NOTE**: Avoid changing the default values of the general hyperparameters provided in `utils/cmdline_helper.py`. The reason is that some other models supported by this repository may rely on such default values. If you want to change the values, override them via the command line. ### Command-line options To display the full list of available options and their descriptions, use the `-h` or `--help` command-line option, for example: `python main.py --help` ### Getting the data Refer to the [TFDS ImageNet readme](https://github.com/tensorflow/datasets/blob/master/docs/catalog/imagenet2012.md) for manual download instructions. To train on the ImageNet dataset, pass `$path_to_ImageNet_tfrecords` to `$data_dir` in the command-line. Name the TFRecords in the following scheme: - Training images - `/data/train-*` - Validation images - `/data/validation-*` ### Training process The training process can start from scratch or resume from a checkpoint. By default, bash script `scripts/{B0,B4}/training/{AMP,FP32,TF32}/convergence_8x{A100-80G,V100-32G}.sh` will start the training process with the following settings. - Use 8 GPUs by Horovod - Has XLA enabled - Saves checkpoints after every 10 epochs to `/workspace/output/` folder - AMP or FP32 or TF32 based on the folder `scripts/{B0,B4}/training/{AMP, FP32, TF32}` The training starts from scratch if `--model_dir` has no checkpoints in it. To resume from a checkpoint, place the checkpoint into `--model_dir` and make sure the `checkpoint` file points to it. #### Multi-node Multi-node runs can be launched on a Pyxis/enroot Slurm cluster (refer to [Requirements](#requirements)) with the `run_{B0,B4}_multinode.sub` script with the following command for a 4-node NVIDIA DGX A100 example: ``` PARTITION=<partition_name> sbatch N 4 --ntasks-per-node=8 run_B0_multinode.sub PARTITION=<partition_name> sbatch N 4 --ntasks-per-node=8 run_B4_multinode.sub ``` Checkpoints will be saved after `--save_checkpoint_freq` epochs at `checkpointdir`. The latest checkpoint will be automatically picked up to resume training in case it needs to be resumed. Cluster partition name has to be provided `<partition_name>`. Note that the `run_{B0,B4}_multinode.sub` script is a starting point that has to be adapted depending on the environment. In particular, pay attention to the variables such as `--container-image`, which handles the container image to train, and `--datadir`, which handles the location of the ImageNet data. Refer to the scripts to find the full list of variables to adjust for your system. ## Inference process Validation can be done either during training (when `--mode train_and_eval` is used) or in a post-training setting (`--mode eval`) on a checkpointed model. The evaluation script expects data in the tfrecord format. `bash ./scripts/{B0,B4}/evaluation/evaluation_{AMP,FP32,TF32}_{A100-80G,V100-32G}.sh` Metrics gathered through this process are as follows: ``` - eval_loss - eval_accuracy_top_1 - eval_accuracy_top_5 - avg_exp_per_second_eval - avg_exp_per_second_eval_per_GPU - avg_time_per_exp_eval : Average Latency - latency_90pct : 90% Latency - latency_95pct : 95% Latency - latency_99pct : 99% Latency ``` The scripts used for inference expect the inference data in the following directory structure: ``` infer_data | ├── images | | ├── image1.JPEG | | ├── image2.JPEG ``` Run: `bash ./scripts/{B0,B4}/inference/inference_{AMP,FP32,TF32}.sh` ## Performance The performance measurements in this document were conducted at the time of publication and may not reflect the performance achieved from NVIDIA’s latest software release. For the most up-to-date performance measurements, go to [NVIDIA Data Center Deep Learning Product Performance](https://developer.nvidia.com/deep-learning-performance-training-inference). ### Benchmarking The following section shows how to run benchmarks measuring the model performance in training and inference modes. #### Training performance benchmark Training benchmark for EfficientNet v1-B0 was run on NVIDIA DGX A100 80GB and NVIDIA DGX-1 V100 32GB. To benchmark training performance with other parameters, run: `bash ./scripts/B0/training/{AMP, FP32, TF32}/train_benchmark_8x{A100-80G, V100-32G}.sh` Training benchmark for EfficientNet v1-B4 was run on NVIDIA DGX A100 80GB and NVIDIA DGX-1 V100 32GB. `bash ./scripts/B4/training/{AMP, FP32, TF32}/train_benchmark_8x{A100-80G, V100-32G}.sh` #### Inference performance benchmark Inference benchmark for EfficientNet v1-B0 was run on NVIDIA DGX A100 80GB and NVIDIA DGX-1 V100 32GB. Inference benchmark for EfficientNet v1-B4 was run on NVIDIA DGX A100 80GB and NVIDIA DGX-1 V100 32GB. ### Results The following sections provide details on how we achieved our performance and accuracy in training and inference. #### Training accuracy results for EfficientNet v1-B0 ##### Training accuracy: NVIDIA DGX A100 (8x A100 80GB) Our results were obtained by running the training scripts in the tensorflow:21.09-tf2-py3 NGC container on NVIDIA DGX A100 (8x A100 80GB) GPUs. We evaluated the models using both the original and EMA weights and selected the higher accuracy to report. | GPUs | Accuracy - TF32 | Accuracy - mixed precision | Time to train - TF32 | Time to train - mixed precision | Time to train speedup (TF32 to mixed precision) | |----------|------------------|-----------------------------|-------------------------|----------------------------------|--------------------------------------------------------| | 8 | 77.60% | 77.59% | 19.5hrs | 8.5hrs | 2.29 | | 16 | 77.51% | 77.48% | 10hrs | 4.5hrs | 2.22 | ##### Training accuracy: NVIDIA DGX-1 (8x V100 32GB) Our results were obtained by running the training scripts in the tensorflow:21.09-tf2-py3 NGC container on NVIDIA DGX V100 (8x V100 32GB) GPUs. We evaluated the models using both the original and EMA weights and selected the higher accuracy to report. | GPUs | Accuracy - FP32 | Accuracy - mixed precision | Time to train - FP32 | Time to train - mixed precision | Time to train speedup (FP32 to mixed precision) | |----------|------------------|-----------------------------|-------------------------|----------------------------------|--------------------------------------------------------| | 8 | 77.67% | 77.69% | 49.0hrs | 38.0hrs | 1.29 | | 32 | 77.55% | 77.53% | 11.5hrs | 10hrs | 1.15 | #### Training accuracy results for EfficientNet v1-B4 ##### Training accuracy: NVIDIA DGX A100 (8x A100 80GB) Our results were obtained by running the training scripts in the tensorflow:21.09-tf2-py3 NGC container on multi-node NVIDIA DGX A100 (8x A100 80GB) GPUs. We evaluated the models using both the original and EMA weights and selected the higher accuracy to report. | GPUs | Accuracy - TF32 | Accuracy - mixed precision | Time to train - TF32 | Time to train - mixed precision | Time to train speedup (TF32 to mixed precision) | |----------|------------------|-----------------------------|-------------------------|----------------------------------|--------------------------------------------------------| | 32 | 82.98% | 83.13% | 38hrs | 14hrs | 2.00 | | 64 | 83.14% | 83.05% | 19hrs | 7hrs | 2.00 | ##### Training accuracy: NVIDIA DGX V100 (8x V100 32GB) Our results were obtained by running the training scripts in the tensorflow:21.09-tf2-py3 NGC container on NVIDIA DGX V100 (8x A100 32GB) GPUs. We evaluated the models using both the original and EMA weights and selected the higher accuracy to report. | GPUs | Accuracy - FP32 | Accuracy - mixed precision | Time to train - TF32 | Time to train - mixed precision | Time to train speedup (FP32 to mixed precision) | |----------|------------------|-----------------------------|-------------------------|----------------------------------|--------------------------------------------------------| | 32 | 82.64% | 82.88% | 97.0hrs | 41.0hrs | 2.37 | | 64 | 82.74% | 83.16% | 50.0hrs | 20.5hrs | 2.43 | #### Training performance results for EfficientNet v1-B0 ##### Training performance: NVIDIA DGX A100 (8x A100 80GB) | GPUs | Throughput - TF32 | Throughput - mixed precision | Throughput speedup (TF32 - mixed precision) | Weak scaling - TF32 | Weak scaling - mixed precision | |-----|------------------------|---------------------------------|-----------------------------------------------|------------------------|----------------------------------| | 1 | 1209 | 3454 | 2.85 | 1 | 1 | | 8 | 9119 | 20647 | 2.26 | 7.54 | 5.98 | | 16 | 17815 | 40644 | 2.28 | 14.74 | 11.77 | ##### Training performance: NVIDIA DGX-1 (8x V100 32GB) | GPUs | Throughput - FP32 | Throughput - mixed precision | Throughput speedup (FP32 - mixed precision) | Weak scaling - FP32 | Weak scaling - mixed precision | |-----|------------------------|---------------------------------|-----------------------------------------------|------------------------|----------------------------------| | 1 | 752 | 868 | 1.15 | 1 | 1 | | 8 | 4504 | 4880 | 1.08 | 5.99 | 5.62 | | 32 | 15309 | 18424 | 1.20 | 20.36 | 21.23 | #### Training performance results for EfficientNet v1-B4 ##### Training performance: NVIDIA DGX A100 (8x A100 80GB) | GPUs | Throughput - TF32 | Throughput - mixed precision | Throughput speedup (TF32 - mixed precision) | Weak scaling - TF32 | Weak scaling - mixed precision | |-----|------------------------|---------------------------------|-----------------------------------------------|------------------------|----------------------------------| | 1 | 165 | 470 | 2.85 | 1 | 1 | | 8 | 1308 | 3550 | 2.71 | 7.93 | 7.55 | | 32 | 4782 | 12908 | 2.70 | 28.98 | 27.46 | | 64 | 9473 | 25455 | 2.69 | 57.41 | 54.16 | ##### Training performance: NVIDIA DGX-1 (8x V100 32GB) | GPUs | Throughput - FP32 | Throughput - mixed precision | Throughput speedup (FP32 - mixed precision) | Weak scaling - TF32 | Weak scaling - mixed precision | |-----|------------------------|---------------------------------|-----------------------------------------------|------------------------|----------------------------------| | 1 | 79 | 211 | 2.67 | 1 | 1 | | 8 | 570 | 1258 | 2.21 | 7.22 | 5.96 | | 32 | 1855 | 4325 | 2.33 | 23.48 | 20.50 | | 64 | 3568 | 8643 | 2.42 | 45.16 | 40.96 | #### Inference performance results for EfficientNet v1-B0 ##### Inference performance: NVIDIA DGX A100 (1x A100 80GB) Our results were obtained by running the inferencing benchmarking script in the tensorflow:21.09-tf2-py3 NGC container on the NVIDIA DGX A100 (1x A100 80GB) GPU. FP16 Inference Latency | Batch size | Resolution | Throughput Avg | Latency Avg (ms) | Latency 90% (ms) |Latency 95% (ms) |Latency 99% (ms) | |-------------|------------------|--------|--------|--------|--------|--------| | 1 | 224x224 | 110.97 | 9.09 | 9.02 | 9.04 | 9.09 | #| 95.71 | 10.45 | 10.34 | 10.38 | 10.42 | | 8 | 224x224 | 874.91 | 9.12 | 9.04 | 9.08 | 9.12 | #| 616.02 | 12.99 | 12.81 | 12.88 | 12.98 | | 32 | 224x224 | 2188.84| 14.62 | 14.35 | 14.43 | 14.52 | | 1024 | 224x224 | 9729.85| 105.24 | 101.50 | 103.20 | 105.24 | TF32 Inference Latency | Batch size | Resolution | Throughput Avg | Latency Avg (ms) | Latency 90% (ms) | Latency 95% (ms) | Latency 99% (ms) | |------------|-----------------|---------|--------|--------|--------|--------| | 1 | 224x224 | 127.95 | 7.88 | 7.83 | 7.84 | 7.87 | # 119.26 | 8.38 | 8.30 | 8.34 | 8.37 | | 8 | 224x224 | 892.27 | 8.97 | 8.88 | 8.91 | 8.94 | # | 803.61 | 9.96 | 9.82 | 9.87 | 9.93 | 32 | 224x224 | 2185.02 | 14.65 | 14.33 | 14.43 | 14.54 | | 512 | 224x224 | 5253.19 | 97.46 | 96.57 | 97.03 | 97.46 | ##### Inference performance: NVIDIA DGX-1 (1x V100 32GB) FP16 Inference Latency | Batch size | Resolution | Throughput Avg | Latency Avg (ms) | Latency 90% (ms) |Latency 95% (ms) |Latency 99% (ms) | |-------------|------------------|--------|--------|--------|--------|--------| | 1 | 224x224 | 97.53 | 10.25 | 10.11 | 10.13 | 10.21 | | 8 | 224x224 | 752.72 | 10.63 | 10.49 | 10.54 | 10.59 | | 32 | 224x224 | 1768.05| 18.10 | 17.88 | 17.96 | 18.04 | | 512 | 224x224 | 5399.88| 94.82 | 92.85 | 93.89 | 94.82 | FP32 Inference Latency | Batch size | Resolution | Throughput Avg | Latency Avg (ms) | Latency 90% (ms) | Latency 95% (ms) | Latency 99% (ms) | |------------|-----------------|---------|--------|--------|--------|--------| | 1 | 224x224 | 97.01 | 10.31 | 10.17 | 10.22 | 10.28 | | 8 | 224x224 | 649.79 | 12.31 | 12.16 | 12.22 | 12.28 | | 32 | 224x224 | 1861.65 | 17.19 | 16.98 | 17.03 | 17.10 | | 256 | 224x224 | 2829.34 | 90.48 | 89.80 | 90.13 | 90.43 | #### Inference performance results for EfficientNet v1-B4 ##### Inference performance: NVIDIA DGX A100 (1x A100 80GB) Our results were obtained by running the inferencing benchmarking script in the tensorflow:21.09-tf2-py3 NGC container on the NVIDIA DGX A100 (1x A100 80GB) GPU. FP16 Inference Latency | Batch size | Resolution | Throughput Avg | Latency Avg (ms) | Latency 90% (ms) |Latency 95% (ms) |Latency 99% (ms) | |------------|------------------|--------|--------|--------|--------|--------| | 1 | 380x380 | 61.36 | 16.30 | 16.20 | 16.24 | 16.28 | #61.36 16.30 16.20 16.24 16.28| 45.40 | 22.03 | 21.82 | 21.90 | 21.99 | | 8 | 380x380 | 338.60 | 23.63 | 23.34 | 23.46 | 23.58 | | 32 | 380x380 | 971.68 | 32.93 | 32.46 | 32.61 | 32.76 | | 128 | 380x380 | 1497.21| 85.28 | 83.01 | 83.68 | 84.70 | TF32 Inference Latency | Batch size | Resolution | Throughput Avg | Latency Avg (ms) | Latency 90% (ms) | Latency 95% (ms) | Latency 99% (ms) | |------------|-----------------|---------|--------|--------|--------|--------| | 1 | 380x380 | 60.54 | 16.52 | 16.34 | 16.41 | 16.49 | | 8 | 380x380 | 366.82 | 21.81 | 21.48 | 21.61 | 21.75 | | 32 | 380x380 | 642.78 | 49.78 | 49.41 | 49.53 | 49.65 | | 64 | 380x380 | 714.55 | 89.54 | 89.00 | 89.17 | 89.34 | ##### Inference performance: NVIDIA DGX-1 (1x V100 32GB) FP16 Inference Latency | Batch size | Resolution | Throughput Avg | Latency Avg (ms) | Latency 90% (ms) |Latency 95% (ms) |Latency 99% (ms) | |------------|------------------|--------|--------|--------|--------|--------| | 1 | 380x380 | 55.71 | 17.95 | 17.68 | 17.93 | 17.86 | #61.36 16.30 16.20 16.24 16.28| 45.40 | 22.03 | 21.82 | 21.90 | 21.99 | | 8 | 380x380 | 256.72 | 31.16 | 30.92 | 31.02 | 31.12 | | 16 | 380x380 | 350.14 | 45.75 | 45.44 | 45.57 | 45.68 | | 64 | 380x380 | 805.21 | 79.46 | 78.74 | 78.86 | 79.01 | TF32 Inference Latency | Batch size | Resolution | Throughput Avg | Latency Avg (ms) | Latency 90% (ms) | Latency 95% (ms) | Latency 99% (ms) | |------------|-----------------|---------|--------|--------|--------|--------| | 1 | 380x380 | 49.03 | 20.40 | 20.03 | 20.18 | 20.34 | | 8 | 380x380 | 258.21 | 30.98 | 30.83 | 30.89 | 30.95 | | 16 | 380x380 | 310.84 | 51.47 | 51.26 | 51.34 | 51.42 | | 32 | 380x380 | 372.23 | 85.97 | 85.70 | 85.79 | 85.89 | ## Release notes ### Changelog February 2022 - Second release - Code Refactoring - Add support of graduate accumulation - Add support of exponential moving average evaluation - Update all accuracy and performance tables on V100 and A100 results March 2021 - Initial release ### Known issues - EfficientNet v1-B0 does not improve training speed by using AMP as compared to FP32, because of the CPU-bound Auto-augmentation.
TensorFlow/LanguageModeling/BERT/data
data
BooksDownloader
# Copyright (c) 2019 NVIDIA CORPORATION. All rights reserved. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import subprocess class BooksDownloader: def __init__(self, save_path): self.save_path = save_path pass def download(self): bookscorpus_download_command = 'python3 /workspace/bookcorpus/download_files.py --list /workspace/bookcorpus/url_list.jsonl --out' bookscorpus_download_command += ' ' + self.save_path + '/bookscorpus' bookscorpus_download_command += ' --trash-bad-count' bookscorpus_download_process = subprocess.run(bookscorpus_download_command, shell=True, check=True)
TensorFlow/Detection/SSD/models/research/object_detection/samples/configs
configs
faster_rcnn_resnet101_coco
# Faster R-CNN with Resnet-101 (v1), configuration for MSCOCO Dataset. # Users should configure the fine_tune_checkpoint field in the train config as # well as the label_map_path and input_path fields in the train_input_reader and # eval_input_reader. Search for "PATH_TO_BE_CONFIGURED" to find the fields that # should be configured. model { faster_rcnn { num_classes: 90 image_resizer { keep_aspect_ratio_resizer { min_dimension: 600 max_dimension: 1024 } } feature_extractor { type: 'faster_rcnn_resnet101' first_stage_features_stride: 16 } first_stage_anchor_generator { grid_anchor_generator { scales: [0.25, 0.5, 1.0, 2.0] aspect_ratios: [0.5, 1.0, 2.0] height_stride: 16 width_stride: 16 } } first_stage_box_predictor_conv_hyperparams { op: CONV regularizer { l2_regularizer { weight: 0.0 } } initializer { truncated_normal_initializer { stddev: 0.01 } } } first_stage_nms_score_threshold: 0.0 first_stage_nms_iou_threshold: 0.7 first_stage_max_proposals: 300 first_stage_localization_loss_weight: 2.0 first_stage_objectness_loss_weight: 1.0 initial_crop_size: 14 maxpool_kernel_size: 2 maxpool_stride: 2 second_stage_box_predictor { mask_rcnn_box_predictor { use_dropout: false dropout_keep_probability: 1.0 fc_hyperparams { op: FC regularizer { l2_regularizer { weight: 0.0 } } initializer { variance_scaling_initializer { factor: 1.0 uniform: true mode: FAN_AVG } } } } } second_stage_post_processing { batch_non_max_suppression { score_threshold: 0.0 iou_threshold: 0.6 max_detections_per_class: 100 max_total_detections: 300 } score_converter: SOFTMAX } second_stage_localization_loss_weight: 2.0 second_stage_classification_loss_weight: 1.0 } } train_config: { batch_size: 1 optimizer { momentum_optimizer: { learning_rate: { manual_step_learning_rate { initial_learning_rate: 0.0003 schedule { step: 900000 learning_rate: .00003 } schedule { step: 1200000 learning_rate: .000003 } } } momentum_optimizer_value: 0.9 } use_moving_average: false } gradient_clipping_by_norm: 10.0 fine_tune_checkpoint: "PATH_TO_BE_CONFIGURED/model.ckpt" from_detection_checkpoint: true data_augmentation_options { random_horizontal_flip { } } } train_input_reader: { tf_record_input_reader { input_path: "PATH_TO_BE_CONFIGURED/mscoco_train.record-?????-of-00100" } label_map_path: "PATH_TO_BE_CONFIGURED/mscoco_label_map.pbtxt" } eval_config: { num_examples: 8000 # Note: The below line limits the evaluation process to 10 evaluations. # Remove the below line to evaluate indefinitely. max_evals: 10 } eval_input_reader: { tf_record_input_reader { input_path: "PATH_TO_BE_CONFIGURED/mscoco_val.record-?????-of-00010" } label_map_path: "PATH_TO_BE_CONFIGURED/mscoco_label_map.pbtxt" shuffle: false num_readers: 1 }
PyTorch/Classification/GPUNet/triton/deployment_toolkit/library
library
utils
# Copyright (c) 2022, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from collections import Counter from typing import Callable, Dict, List, Optional import networkx as nx from ..core import ShapeSpec def infer_precision( nx_graph: nx.Graph, input_names: List[str], output_names: List[str], get_node_dtype_fn: Callable, ): node_dtypes = [nx_graph.nodes[node_name].get("dtype", None) for node_name in nx_graph.nodes] node_dtypes = [dt for dt in node_dtypes if dt is None or dt.kind not in ["i", "b"]] dtypes_counter = Counter(node_dtypes) return dtypes_counter.most_common()[0][0] def get_shapes_with_dynamic_axes(dataloader, batch_size_dim: Optional[int] = None): def _set_dynamic_shapes(t, shapes): for k, v in t.items(): shape = list(v.shape) for dim, s in enumerate(shape): if shapes[k][dim] != -1 and shapes[k][dim] != s: shapes[k][dim] = -1 def _mark_batch_axis(shape, batch_axis: int): shape = list(shape) shape[batch_axis] = -1 return tuple(shape) ## get all shapes from input and output tensors input_shapes = {} output_shapes = {} for batch in dataloader: _, x, y = batch for k, v in x.items(): input_shapes[k] = list(v.shape) for k, v in y.items(): output_shapes[k] = list(v.shape) break # based on max <max_num_iters> iterations, check which # dimensions differ to determine dynamic_axes max_num_iters = 100 for idx, batch in enumerate(dataloader): if idx >= max_num_iters: break _, x, y = batch _set_dynamic_shapes(x, input_shapes) _set_dynamic_shapes(y, output_shapes) if batch_size_dim is not None: input_shapes = {name: _mark_batch_axis(shape, batch_size_dim) for name, shape in input_shapes.items()} output_shapes = {name: _mark_batch_axis(shape, batch_size_dim) for name, shape in output_shapes.items()} return input_shapes, output_shapes def get_dynamic_axes(dataloader, batch_size_dim: Optional[int] = None): input_shapes, output_shapes = get_shapes_with_dynamic_axes(dataloader, batch_size_dim=batch_size_dim) all_shapes = {**input_shapes, **output_shapes} dynamic_axes = {} for k, shape in all_shapes.items(): for idx, s in enumerate(shape): if s == -1: dynamic_axes[k] = {idx: k + "_" + str(idx)} for k in all_shapes: if k in dynamic_axes: dynamic_axes[k].update({batch_size_dim: "batch_size_" + str(batch_size_dim)}) else: dynamic_axes[k] = {batch_size_dim: "batch_size_" + str(batch_size_dim)} return dynamic_axes def get_input_shapes(dataloader, max_batch_size=1) -> Dict[str, ShapeSpec]: def init_counters_and_shapes(x, counters, min_shapes, max_shapes): for k, v in x.items(): counters[k] = Counter() min_shapes[k] = [float("inf")] * v.ndim max_shapes[k] = [float("-inf")] * v.ndim counters = {} min_shapes: Dict[str, tuple] = {} max_shapes: Dict[str, tuple] = {} for idx, batch in enumerate(dataloader): ids, x, y = batch if idx == 0: init_counters_and_shapes(x, counters, min_shapes, max_shapes) for k, v in x.items(): shape = v.shape counters[k][shape] += 1 min_shapes[k] = tuple(min(a, b) for a, b in zip(min_shapes[k], shape)) max_shapes[k] = tuple(max(a, b) for a, b in zip(max_shapes[k], shape)) opt_shapes: Dict[str, tuple] = {} for k, v in counters.items(): opt_shapes[k] = v.most_common(1)[0][0] shapes = {} for k in opt_shapes.keys(): # same keys in min_shapes and max_shapes shapes[k] = ShapeSpec( min=(1,) + min_shapes[k][1:], max=(max_batch_size,) + max_shapes[k][1:], opt=(max_batch_size,) + opt_shapes[k][1:], ) return shapes
Tools/DGLPyTorch/SyntheticGraphGeneration/syngen/generator/graph
graph
base_graph_generator
# Copyright (c) 2023, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import abc import json import pickle from abc import ABC from typing import List, Optional, Set, Tuple import numpy as np from syngen.generator.graph.utils import BaseLogger from syngen.generator.graph.seeder import BaseSeeder class BaseGenerator(abc.ABC): """ BaseGenerator class """ JSON_ASSERTION = "Expected file to be json" @classmethod def get_generators(cls, include_parents=True): """ Recursively find subclasses Args: include_parents (bool): whether to include parents to other classes. (default: `True`) Returns: generators: dictionary with all the subclasses """ generators = dict() for child in cls.__subclasses__(): children = child.get_generators(include_parents) generators.update(children) if include_parents or not children: if abc.ABC not in child.__bases__ and BaseGenerator not in child.__bases__: generators[child.__name__] = child return generators def save(self, path): raise NotImplementedError() @classmethod def load(cls, path): raise NotImplementedError() class BaseGraphGenerator(BaseGenerator, ABC): """ Base class for all graph generators Args: *args: optional positional args **kwargs: optional key-word args """ def __init__( self, seed: Optional[int] = None, logdir: str = "./logs", gpu: bool = True, verbose: bool = False, *args, **kwargs, ): self._fit_results = None self.seeder = BaseSeeder(seed) self.seeder.reseed() self.logger = BaseLogger(logdir) self.logger.log(f"Using seed: {self.seeder.seed}") self.gpu = gpu self.verbose = verbose def fit( self, graph: List[Tuple[int, int]], is_directed: bool, *args, **kwargs ): """ Fits generator on the graph Args: graph (List[Tuple[int, int]]): graph to be fitted on is_directed (bool): flag indicating whether the graph is directed *args: optional positional args **kwargs: optional key-word args """ raise NotImplementedError() def generate( self, num_nodes: int, num_edges: int, is_directed: bool, *args, return_node_ids: bool = False, **kwargs, ) -> np.ndarray: """ Generates graph with approximately `num_nodes` and exactly `num_edges` from generator Args: num_nodes (int): approximate number of nodes to be generated num_edges (int): exact number of edges to be generated is_directed (bool): flag indicating whether the generated graph has to be directed return_node_ids (bool): flag indicating whether the generator has to return nodes_ids as the second output *args: optional positional args **kwargs: optional key-word args """ raise NotImplementedError() def set_fit_results(self, fit_results): self._fit_results = fit_results def get_fit_results(self): return self._fit_results def save_fit_results(self, save_path: str = "./fit_results.json"): """ Store fitted results into json file Args: save_path (str): path to the json file with the fitted result """ assert ( self._fit_results ), "There are no fit results to be saved, \ call fit method first or load the results from the file" assert save_path.endswith(".json"), self.JSON_ASSERTION with open(save_path, "w") as fjson: json.dump(self._fit_results, fjson) def load_fit_results(self, load_path: str = "./fit_results.json"): """load fitted results from json file Args: load_path (str): path to the json file with the fitted result """ assert load_path.endswith(".json"), self.JSON_ASSERTION with open(load_path, "r") as fjson: self._fit_results = json.load(fjson) def save(self, path): with open(path, 'wb') as file_handler: pickle.dump(self, file_handler, protocol=pickle.HIGHEST_PROTOCOL) @classmethod def load(cls, path): with open(path, 'rb') as file_handler: model = pickle.load(file_handler) return model @staticmethod def add_args(parser): return parser class BaseBipartiteGraphGenerator(BaseGenerator, ABC): """ Base class for all bipartite graph generators Args: *args: optional positional args **kwargs: optional key-word args """ def __init__( self, seed: Optional[int] = None, logdir: str = "./logs", gpu: bool = True, verbose: bool = False, *args, **kwargs, ): self._fit_src_dst_results = None self._fit_dst_src_results = None self.seeder = BaseSeeder(seed) self.seeder.reseed() self.logger = BaseLogger(logdir) self.logger.log(f"Using seed: {self.seeder.seed}") self.gpu = gpu self.verbose = verbose def fit( self, graph: List[Tuple[int, int]], src_set: Set[int], dst_set: Set[int], is_directed: bool, transform_graph: bool, *args, **kwargs, ): """ Fits generator on the graph Args: graph (List[Tuple[int, int]]): graph to be fitted on src_set (Set[int]): set of source nodes dst_set (Set[int]): set of destination nodes is_directed (bool): flag indicating whether the graph is directed *args: optional positional args **kwargs: optional key-word args """ raise NotImplementedError() def generate( self, num_nodes_src_set: int, num_nodes_dst_set: int, num_edges_src_dst: int, num_edges_dst_src: int, is_directed: bool, return_node_ids: bool = False, transform_graph: bool = True, *args, **kwargs, ): """ Generates graph with approximately `num_nodes_src_set`/`num_nodes_dst_set` nodes and exactly `num_edges_src_dst`/`num_edges_dst_src` edges from generator Args: num_nodes_src_set (int): approximate number of source nodes to be generated num_nodes_dst_set (int): approximate number of destination nodes to be generated num_edges_src_dst (int): exact number of source->destination edges to be generated num_edges_dst_src (int): exact number of destination->source to be generated is_directed (bool) flag indicating whether the generated graph has to be directed return_node_ids (bool): flag indicating whether the generator has to return nodes_ids as the second output *args: optional positional args **kwargs: optional key-word args """ raise NotImplementedError() def set_fit_results(self, fit_results): self._fit_src_dst_results, self._fit_dst_src_results = fit_results def get_fit_results(self): return self._fit_src_dst_results, self._fit_dst_src_results def save_fit_results(self, save_path: str = "./fit_results.json"): """ Stores fitted results into json file Args: save_path (str): path to the json file with the fitted result """ assert ( self._fit_src_dst_results or self._fit_dst_src_results ), "There are no fit results to be saved, \ call fit method first or load the results from the file" assert save_path.endswith(".json"), self.JSON_ASSERTION wrapped_results = { "fit_src_dst_results": self._fit_src_dst_results, "fit_dst_src_results": self._fit_dst_src_results, } with open(save_path, "w") as fjson: json.dump(wrapped_results, fjson) def load_fit_results(self, load_path: str = "./fit_results.json"): """ Loads fitted results from json file Args: load_path (str): path to the json file with the fitted result """ assert load_path.endswith(".json"), self.JSON_ASSERTION with open(load_path, "r") as fjson: wrapped_results = json.load(fjson) assert ( "fit_src_dst_results" in wrapped_results and "fit_dst_src_results" in wrapped_results ), "Required keys fit_src_dst_results and fit_dst_src_results keys in the json not found" self._fit_src_dst_results = wrapped_results["fit_src_dst_results"] self._fit_dst_src_results = wrapped_results["fit_dst_src_results"] def save(self, path): with open(path, 'wb') as file_handler: pickle.dump(self, file_handler, protocol=pickle.HIGHEST_PROTOCOL) @classmethod def load(cls, path): with open(path, 'rb') as file_handler: model = pickle.load(file_handler) return model @staticmethod def add_args(parser): return parser
CUDA-Optimized/FastSpeech/tacotron2
tacotron2
distributed
# BSD 3-Clause License # Copyright (c) 2018-2020, NVIDIA Corporation # All rights reserved. # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # * Redistributions of source code must retain the above copyright notice, this # list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright notice, # this list of conditions and the following disclaimer in the documentation # and/or other materials provided with the distribution. # * Neither the name of the copyright holder nor the names of its # contributors may be used to endorse or promote products derived from # this software without specific prior written permission. # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE # FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL # DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR # SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER # CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, # OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. """https://github.com/NVIDIA/tacotron2""" import torch import torch.distributed as dist from torch.nn.modules import Module from torch.autograd import Variable def _flatten_dense_tensors(tensors): """Flatten dense tensors into a contiguous 1D buffer. Assume tensors are of same dense type. Since inputs are dense, the resulting tensor will be a concatenated 1D buffer. Element-wise operation on this buffer will be equivalent to operating individually. Arguments: tensors (Iterable[Tensor]): dense tensors to flatten. Returns: A contiguous 1D buffer containing input tensors. """ if len(tensors) == 1: return tensors[0].contiguous().view(-1) flat = torch.cat([t.contiguous().view(-1) for t in tensors], dim=0) return flat def _unflatten_dense_tensors(flat, tensors): """View a flat buffer using the sizes of tensors. Assume that tensors are of same dense type, and that flat is given by _flatten_dense_tensors. Arguments: flat (Tensor): flattened dense tensors to unflatten. tensors (Iterable[Tensor]): dense tensors whose sizes will be used to unflatten flat. Returns: Unflattened dense tensors with sizes same as tensors and values from flat. """ outputs = [] offset = 0 for tensor in tensors: numel = tensor.numel() outputs.append(flat.narrow(0, offset, numel).view_as(tensor)) offset += numel return tuple(outputs) ''' This version of DistributedDataParallel is designed to be used in conjunction with the multiproc.py launcher included with this example. It assumes that your run is using multiprocess with 1 GPU/process, that the model is on the correct device, and that torch.set_device has been used to set the device. Parameters are broadcasted to the other processes on initialization of DistributedDataParallel, and will be allreduced at the finish of the backward pass. ''' class DistributedDataParallel(Module): def __init__(self, module): super(DistributedDataParallel, self).__init__() #fallback for PyTorch 0.3 if not hasattr(dist, '_backend'): self.warn_on_half = True else: self.warn_on_half = True if dist._backend == dist.dist_backend.GLOO else False self.module = module for p in self.module.state_dict().values(): if not torch.is_tensor(p): continue dist.broadcast(p, 0) def allreduce_params(): if(self.needs_reduction): self.needs_reduction = False buckets = {} for param in self.module.parameters(): if param.requires_grad and param.grad is not None: tp = type(param.data) if tp not in buckets: buckets[tp] = [] buckets[tp].append(param) if self.warn_on_half: if torch.cuda.HalfTensor in buckets: print("WARNING: gloo dist backend for half parameters may be extremely slow." + " It is recommended to use the NCCL backend in this case. This currently requires" + "PyTorch built from top of tree master.") self.warn_on_half = False for tp in buckets: bucket = buckets[tp] grads = [param.grad.data for param in bucket] coalesced = _flatten_dense_tensors(grads) dist.all_reduce(coalesced) coalesced /= dist.get_world_size() for buf, synced in zip(grads, _unflatten_dense_tensors(coalesced, grads)): buf.copy_(synced) for param in list(self.module.parameters()): def allreduce_hook(*unused): param._execution_engine.queue_callback(allreduce_params) if param.requires_grad: param.register_hook(allreduce_hook) def forward(self, *inputs, **kwargs): self.needs_reduction = True return self.module(*inputs, **kwargs) ''' def _sync_buffers(self): buffers = list(self.module._all_buffers()) if len(buffers) > 0: # cross-node buffer sync flat_buffers = _flatten_dense_tensors(buffers) dist.broadcast(flat_buffers, 0) for buf, synced in zip(buffers, _unflatten_dense_tensors(flat_buffers, buffers)): buf.copy_(synced) def train(self, mode=True): # Clear NCCL communicator and CUDA event cache of the default group ID, # These cache will be recreated at the later call. This is currently a # work-around for a potential NCCL deadlock. if dist._backend == dist.dist_backend.NCCL: dist._clear_group_cache() super(DistributedDataParallel, self).train(mode) self.module.train(mode) ''' ''' Modifies existing model to do gradient allreduce, but doesn't change class so you don't need "module" ''' def apply_gradient_allreduce(module): if not hasattr(dist, '_backend'): module.warn_on_half = True else: module.warn_on_half = True if dist._backend == dist.dist_backend.GLOO else False for p in module.state_dict().values(): if not torch.is_tensor(p): continue dist.broadcast(p, 0) def allreduce_params(): if(module.needs_reduction): module.needs_reduction = False buckets = {} for param in module.parameters(): if param.requires_grad and param.grad is not None: tp = param.data.dtype if tp not in buckets: buckets[tp] = [] buckets[tp].append(param) if module.warn_on_half: if torch.cuda.HalfTensor in buckets: print("WARNING: gloo dist backend for half parameters may be extremely slow." + " It is recommended to use the NCCL backend in this case. This currently requires" + "PyTorch built from top of tree master.") module.warn_on_half = False for tp in buckets: bucket = buckets[tp] grads = [param.grad.data for param in bucket] coalesced = _flatten_dense_tensors(grads) dist.all_reduce(coalesced) coalesced /= dist.get_world_size() for buf, synced in zip(grads, _unflatten_dense_tensors(coalesced, grads)): buf.copy_(synced) for param in list(module.parameters()): def allreduce_hook(*unused): Variable._execution_engine.queue_callback(allreduce_params) if param.requires_grad: param.register_hook(allreduce_hook) def set_needs_reduction(self, input, output): self.needs_reduction = True module.register_forward_hook(set_needs_reduction) return module
TensorFlow/Detection/SSD/models/research/slim/preprocessing
preprocessing
inception_preprocessing
# Copyright 2016 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Provides utilities to preprocess images for the Inception networks.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow as tf from tensorflow.python.ops import control_flow_ops def apply_with_random_selector(x, func, num_cases): """Computes func(x, sel), with sel sampled from [0...num_cases-1]. Args: x: input Tensor. func: Python function to apply. num_cases: Python int32, number of cases to sample sel from. Returns: The result of func(x, sel), where func receives the value of the selector as a python integer, but sel is sampled dynamically. """ sel = tf.random_uniform([], maxval=num_cases, dtype=tf.int32) # Pass the real x only to one of the func calls. return control_flow_ops.merge([ func(control_flow_ops.switch(x, tf.equal(sel, case))[1], case) for case in range(num_cases)])[0] def distort_color(image, color_ordering=0, fast_mode=True, scope=None): """Distort the color of a Tensor image. Each color distortion is non-commutative and thus ordering of the color ops matters. Ideally we would randomly permute the ordering of the color ops. Rather then adding that level of complication, we select a distinct ordering of color ops for each preprocessing thread. Args: image: 3-D Tensor containing single image in [0, 1]. color_ordering: Python int, a type of distortion (valid values: 0-3). fast_mode: Avoids slower ops (random_hue and random_contrast) scope: Optional scope for name_scope. Returns: 3-D Tensor color-distorted image on range [0, 1] Raises: ValueError: if color_ordering not in [0, 3] """ with tf.name_scope(scope, 'distort_color', [image]): if fast_mode: if color_ordering == 0: image = tf.image.random_brightness(image, max_delta=32. / 255.) image = tf.image.random_saturation(image, lower=0.5, upper=1.5) else: image = tf.image.random_saturation(image, lower=0.5, upper=1.5) image = tf.image.random_brightness(image, max_delta=32. / 255.) else: if color_ordering == 0: image = tf.image.random_brightness(image, max_delta=32. / 255.) image = tf.image.random_saturation(image, lower=0.5, upper=1.5) image = tf.image.random_hue(image, max_delta=0.2) image = tf.image.random_contrast(image, lower=0.5, upper=1.5) elif color_ordering == 1: image = tf.image.random_saturation(image, lower=0.5, upper=1.5) image = tf.image.random_brightness(image, max_delta=32. / 255.) image = tf.image.random_contrast(image, lower=0.5, upper=1.5) image = tf.image.random_hue(image, max_delta=0.2) elif color_ordering == 2: image = tf.image.random_contrast(image, lower=0.5, upper=1.5) image = tf.image.random_hue(image, max_delta=0.2) image = tf.image.random_brightness(image, max_delta=32. / 255.) image = tf.image.random_saturation(image, lower=0.5, upper=1.5) elif color_ordering == 3: image = tf.image.random_hue(image, max_delta=0.2) image = tf.image.random_saturation(image, lower=0.5, upper=1.5) image = tf.image.random_contrast(image, lower=0.5, upper=1.5) image = tf.image.random_brightness(image, max_delta=32. / 255.) else: raise ValueError('color_ordering must be in [0, 3]') # The random_* ops do not necessarily clamp. return tf.clip_by_value(image, 0.0, 1.0) def distorted_bounding_box_crop(image, bbox, min_object_covered=0.1, aspect_ratio_range=(0.75, 1.33), area_range=(0.05, 1.0), max_attempts=100, scope=None): """Generates cropped_image using a one of the bboxes randomly distorted. See `tf.image.sample_distorted_bounding_box` for more documentation. Args: image: 3-D Tensor of image (it will be converted to floats in [0, 1]). bbox: 3-D float Tensor of bounding boxes arranged [1, num_boxes, coords] where each coordinate is [0, 1) and the coordinates are arranged as [ymin, xmin, ymax, xmax]. If num_boxes is 0 then it would use the whole image. min_object_covered: An optional `float`. Defaults to `0.1`. The cropped area of the image must contain at least this fraction of any bounding box supplied. aspect_ratio_range: An optional list of `floats`. The cropped area of the image must have an aspect ratio = width / height within this range. area_range: An optional list of `floats`. The cropped area of the image must contain a fraction of the supplied image within in this range. max_attempts: An optional `int`. Number of attempts at generating a cropped region of the image of the specified constraints. After `max_attempts` failures, return the entire image. scope: Optional scope for name_scope. Returns: A tuple, a 3-D Tensor cropped_image and the distorted bbox """ with tf.name_scope(scope, 'distorted_bounding_box_crop', [image, bbox]): # Each bounding box has shape [1, num_boxes, box coords] and # the coordinates are ordered [ymin, xmin, ymax, xmax]. # A large fraction of image datasets contain a human-annotated bounding # box delineating the region of the image containing the object of interest. # We choose to create a new bounding box for the object which is a randomly # distorted version of the human-annotated bounding box that obeys an # allowed range of aspect ratios, sizes and overlap with the human-annotated # bounding box. If no box is supplied, then we assume the bounding box is # the entire image. sample_distorted_bounding_box = tf.image.sample_distorted_bounding_box( tf.shape(image), bounding_boxes=bbox, min_object_covered=min_object_covered, aspect_ratio_range=aspect_ratio_range, area_range=area_range, max_attempts=max_attempts, use_image_if_no_bounding_boxes=True) bbox_begin, bbox_size, distort_bbox = sample_distorted_bounding_box # Crop the image to the specified bounding box. cropped_image = tf.slice(image, bbox_begin, bbox_size) return cropped_image, distort_bbox def preprocess_for_train(image, height, width, bbox, fast_mode=True, scope=None, add_image_summaries=True): """Distort one image for training a network. Distorting images provides a useful technique for augmenting the data set during training in order to make the network invariant to aspects of the image that do not effect the label. Additionally it would create image_summaries to display the different transformations applied to the image. Args: image: 3-D Tensor of image. If dtype is tf.float32 then the range should be [0, 1], otherwise it would converted to tf.float32 assuming that the range is [0, MAX], where MAX is largest positive representable number for int(8/16/32) data type (see `tf.image.convert_image_dtype` for details). height: integer width: integer bbox: 3-D float Tensor of bounding boxes arranged [1, num_boxes, coords] where each coordinate is [0, 1) and the coordinates are arranged as [ymin, xmin, ymax, xmax]. fast_mode: Optional boolean, if True avoids slower transformations (i.e. bi-cubic resizing, random_hue or random_contrast). scope: Optional scope for name_scope. add_image_summaries: Enable image summaries. Returns: 3-D float Tensor of distorted image used for training with range [-1, 1]. """ with tf.name_scope(scope, 'distort_image', [image, height, width, bbox]): if bbox is None: bbox = tf.constant([0.0, 0.0, 1.0, 1.0], dtype=tf.float32, shape=[1, 1, 4]) if image.dtype != tf.float32: image = tf.image.convert_image_dtype(image, dtype=tf.float32) # Each bounding box has shape [1, num_boxes, box coords] and # the coordinates are ordered [ymin, xmin, ymax, xmax]. image_with_box = tf.image.draw_bounding_boxes(tf.expand_dims(image, 0), bbox) if add_image_summaries: tf.summary.image('image_with_bounding_boxes', image_with_box) distorted_image, distorted_bbox = distorted_bounding_box_crop(image, bbox) # Restore the shape since the dynamic slice based upon the bbox_size loses # the third dimension. distorted_image.set_shape([None, None, 3]) image_with_distorted_box = tf.image.draw_bounding_boxes( tf.expand_dims(image, 0), distorted_bbox) if add_image_summaries: tf.summary.image('images_with_distorted_bounding_box', image_with_distorted_box) # This resizing operation may distort the images because the aspect # ratio is not respected. We select a resize method in a round robin # fashion based on the thread number. # Note that ResizeMethod contains 4 enumerated resizing methods. # We select only 1 case for fast_mode bilinear. num_resize_cases = 1 if fast_mode else 4 distorted_image = apply_with_random_selector( distorted_image, lambda x, method: tf.image.resize_images(x, [height, width], method), num_cases=num_resize_cases) if add_image_summaries: tf.summary.image('cropped_resized_image', tf.expand_dims(distorted_image, 0)) # Randomly flip the image horizontally. distorted_image = tf.image.random_flip_left_right(distorted_image) # Randomly distort the colors. There are 1 or 4 ways to do it. num_distort_cases = 1 if fast_mode else 4 distorted_image = apply_with_random_selector( distorted_image, lambda x, ordering: distort_color(x, ordering, fast_mode), num_cases=num_distort_cases) if add_image_summaries: tf.summary.image('final_distorted_image', tf.expand_dims(distorted_image, 0)) distorted_image = tf.subtract(distorted_image, 0.5) distorted_image = tf.multiply(distorted_image, 2.0) return distorted_image def preprocess_for_eval(image, height, width, central_fraction=0.875, scope=None): """Prepare one image for evaluation. If height and width are specified it would output an image with that size by applying resize_bilinear. If central_fraction is specified it would crop the central fraction of the input image. Args: image: 3-D Tensor of image. If dtype is tf.float32 then the range should be [0, 1], otherwise it would converted to tf.float32 assuming that the range is [0, MAX], where MAX is largest positive representable number for int(8/16/32) data type (see `tf.image.convert_image_dtype` for details). height: integer width: integer central_fraction: Optional Float, fraction of the image to crop. scope: Optional scope for name_scope. Returns: 3-D float Tensor of prepared image. """ with tf.name_scope(scope, 'eval_image', [image, height, width]): if image.dtype != tf.float32: image = tf.image.convert_image_dtype(image, dtype=tf.float32) # Crop the central region of the image with an area containing 87.5% of # the original image. if central_fraction: image = tf.image.central_crop(image, central_fraction=central_fraction) if height and width: # Resize the image to the specified height and width. image = tf.expand_dims(image, 0) image = tf.image.resize_bilinear(image, [height, width], align_corners=False) image = tf.squeeze(image, [0]) image = tf.subtract(image, 0.5) image = tf.multiply(image, 2.0) return image def preprocess_image(image, height, width, is_training=False, bbox=None, fast_mode=True, add_image_summaries=True): """Pre-process one image for training or evaluation. Args: image: 3-D Tensor [height, width, channels] with the image. If dtype is tf.float32 then the range should be [0, 1], otherwise it would converted to tf.float32 assuming that the range is [0, MAX], where MAX is largest positive representable number for int(8/16/32) data type (see `tf.image.convert_image_dtype` for details). height: integer, image expected height. width: integer, image expected width. is_training: Boolean. If true it would transform an image for train, otherwise it would transform it for evaluation. bbox: 3-D float Tensor of bounding boxes arranged [1, num_boxes, coords] where each coordinate is [0, 1) and the coordinates are arranged as [ymin, xmin, ymax, xmax]. fast_mode: Optional boolean, if True avoids slower transformations. add_image_summaries: Enable image summaries. Returns: 3-D float Tensor containing an appropriately scaled image Raises: ValueError: if user does not provide bounding box """ if is_training: return preprocess_for_train(image, height, width, bbox, fast_mode, add_image_summaries=add_image_summaries) else: return preprocess_for_eval(image, height, width)
TensorFlow/Detection/SSD/models/research/object_detection/utils
utils
test_utils
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Contains functions which are convenient for unit testing.""" import numpy as np import tensorflow as tf from object_detection.core import anchor_generator from object_detection.core import box_coder from object_detection.core import box_list from object_detection.core import box_predictor from object_detection.core import matcher from object_detection.utils import shape_utils # Default size (both width and height) used for testing mask predictions. DEFAULT_MASK_SIZE = 5 class MockBoxCoder(box_coder.BoxCoder): """Simple `difference` BoxCoder.""" @property def code_size(self): return 4 def _encode(self, boxes, anchors): return boxes.get() - anchors.get() def _decode(self, rel_codes, anchors): return box_list.BoxList(rel_codes + anchors.get()) class MockMaskHead(object): """Simple maskhead that returns all zeros as mask predictions.""" def __init__(self, num_classes): self._num_classes = num_classes def predict(self, features): batch_size = tf.shape(features)[0] return tf.zeros((batch_size, 1, self._num_classes, DEFAULT_MASK_SIZE, DEFAULT_MASK_SIZE), dtype=tf.float32) class MockBoxPredictor(box_predictor.BoxPredictor): """Simple box predictor that ignores inputs and outputs all zeros.""" def __init__(self, is_training, num_classes, add_background_class=True): super(MockBoxPredictor, self).__init__(is_training, num_classes) self._add_background_class = add_background_class def _predict(self, image_features, num_predictions_per_location): image_feature = image_features[0] combined_feature_shape = shape_utils.combined_static_and_dynamic_shape( image_feature) batch_size = combined_feature_shape[0] num_anchors = (combined_feature_shape[1] * combined_feature_shape[2]) code_size = 4 zero = tf.reduce_sum(0 * image_feature) num_class_slots = self.num_classes if self._add_background_class: num_class_slots = num_class_slots + 1 box_encodings = zero + tf.zeros( (batch_size, num_anchors, 1, code_size), dtype=tf.float32) class_predictions_with_background = zero + tf.zeros( (batch_size, num_anchors, num_class_slots), dtype=tf.float32) predictions_dict = { box_predictor.BOX_ENCODINGS: box_encodings, box_predictor.CLASS_PREDICTIONS_WITH_BACKGROUND: class_predictions_with_background } return predictions_dict class MockKerasBoxPredictor(box_predictor.KerasBoxPredictor): """Simple box predictor that ignores inputs and outputs all zeros.""" def __init__(self, is_training, num_classes, add_background_class=True): super(MockKerasBoxPredictor, self).__init__( is_training, num_classes, False, False) self._add_background_class = add_background_class def _predict(self, image_features, **kwargs): image_feature = image_features[0] combined_feature_shape = shape_utils.combined_static_and_dynamic_shape( image_feature) batch_size = combined_feature_shape[0] num_anchors = (combined_feature_shape[1] * combined_feature_shape[2]) code_size = 4 zero = tf.reduce_sum(0 * image_feature) num_class_slots = self.num_classes if self._add_background_class: num_class_slots = num_class_slots + 1 box_encodings = zero + tf.zeros( (batch_size, num_anchors, 1, code_size), dtype=tf.float32) class_predictions_with_background = zero + tf.zeros( (batch_size, num_anchors, num_class_slots), dtype=tf.float32) predictions_dict = { box_predictor.BOX_ENCODINGS: box_encodings, box_predictor.CLASS_PREDICTIONS_WITH_BACKGROUND: class_predictions_with_background } return predictions_dict class MockAnchorGenerator(anchor_generator.AnchorGenerator): """Mock anchor generator.""" def name_scope(self): return 'MockAnchorGenerator' def num_anchors_per_location(self): return [1] def _generate(self, feature_map_shape_list): num_anchors = sum([shape[0] * shape[1] for shape in feature_map_shape_list]) return box_list.BoxList(tf.zeros((num_anchors, 4), dtype=tf.float32)) class MockMatcher(matcher.Matcher): """Simple matcher that matches first anchor to first groundtruth box.""" def _match(self, similarity_matrix, valid_rows): return tf.constant([0, -1, -1, -1], dtype=tf.int32) def create_diagonal_gradient_image(height, width, depth): """Creates pyramid image. Useful for testing. For example, pyramid_image(5, 6, 1) looks like: # [[[ 5. 4. 3. 2. 1. 0.] # [ 6. 5. 4. 3. 2. 1.] # [ 7. 6. 5. 4. 3. 2.] # [ 8. 7. 6. 5. 4. 3.] # [ 9. 8. 7. 6. 5. 4.]]] Args: height: height of image width: width of image depth: depth of image Returns: pyramid image """ row = np.arange(height) col = np.arange(width)[::-1] image_layer = np.expand_dims(row, 1) + col image_layer = np.expand_dims(image_layer, 2) image = image_layer for i in range(1, depth): image = np.concatenate((image, image_layer * pow(10, i)), 2) return image.astype(np.float32) def create_random_boxes(num_boxes, max_height, max_width): """Creates random bounding boxes of specific maximum height and width. Args: num_boxes: number of boxes. max_height: maximum height of boxes. max_width: maximum width of boxes. Returns: boxes: numpy array of shape [num_boxes, 4]. Each row is in form [y_min, x_min, y_max, x_max]. """ y_1 = np.random.uniform(size=(1, num_boxes)) * max_height y_2 = np.random.uniform(size=(1, num_boxes)) * max_height x_1 = np.random.uniform(size=(1, num_boxes)) * max_width x_2 = np.random.uniform(size=(1, num_boxes)) * max_width boxes = np.zeros(shape=(num_boxes, 4)) boxes[:, 0] = np.minimum(y_1, y_2) boxes[:, 1] = np.minimum(x_1, x_2) boxes[:, 2] = np.maximum(y_1, y_2) boxes[:, 3] = np.maximum(x_1, x_2) return boxes.astype(np.float32) def first_rows_close_as_set(a, b, k=None, rtol=1e-6, atol=1e-6): """Checks if first K entries of two lists are close, up to permutation. Inputs to this assert are lists of items which can be compared via numpy.allclose(...) and can be sorted. Args: a: list of items which can be compared via numpy.allclose(...) and are sortable. b: list of items which can be compared via numpy.allclose(...) and are sortable. k: a non-negative integer. If not provided, k is set to be len(a). rtol: relative tolerance. atol: absolute tolerance. Returns: boolean, True if input lists a and b have the same length and the first k entries of the inputs satisfy numpy.allclose() after sorting entries. """ if not isinstance(a, list) or not isinstance(b, list) or len(a) != len(b): return False if not k: k = len(a) k = min(k, len(a)) a_sorted = sorted(a[:k]) b_sorted = sorted(b[:k]) return all([ np.allclose(entry_a, entry_b, rtol, atol) for (entry_a, entry_b) in zip(a_sorted, b_sorted) ])
TensorFlow/LanguageModeling/BERT/biobert/scripts
scripts
run_pretraining_pubmed_base_phase_2
#! /bin/bash echo "Container nvidia build = " $NVIDIA_BUILD_ID init_checkpoint=${1} train_batch_size=${2:-16} learning_rate=${3:-"2.9e-4"} cased=${4:-false} precision=${5:-"fp16"} use_xla=${6:-true} num_gpu=${7:-16} warmup_steps=${8:-"434"} train_steps=${9:-4340} num_accumulation_steps=${10:-128} save_checkpoint_steps=${11:-5000} eval_batch_size=${12:-26} use_fp16="" if [ "$precision" = "fp16" ] ; then echo "fp16 activated!" use_fp16="--amp" else echo "fp32/tf32 activated!" use_fp16="--noamp" fi if [ "$use_xla" = "true" ] ; then use_xla_tag="--use_xla" echo "XLA activated" else use_xla_tag="--nouse_xla" fi if [ "$cased" = "true" ] ; then DO_LOWER_CASE=0 CASING_DIR_PREFIX="cased" else DO_LOWER_CASE=1 CASING_DIR_PREFIX="uncased" fi BERT_CONFIG=/workspace/bert/data/download/google_pretrained_weights/${CASING_DIR_PREFIX}_L-12_H-768_A-12/bert_config.json RESULTS_DIR=/results CHECKPOINTS_DIR=${RESULTS_DIR}/biobert_phase_2 mkdir -p ${CHECKPOINTS_DIR} INPUT_FILES_DIR="/workspace/bert/data/tfrecord/lower_case_${DO_LOWER_CASE}_seq_len_512_max_pred_80_masked_lm_prob_0.15_random_seed_12345_dupe_factor_5_shard_1472_test_split_10/pubmed_baseline/training" EVAL_FILES_DIR="/workspace/bert/data/tfrecord/lower_case_${DO_LOWER_CASE}_seq_len_512_max_pred_80_masked_lm_prob_0.15_random_seed_12345_dupe_factor_5_shard_1472_test_split_10/pubmed_baseline/test" if [ $num_gpu -gt 1 ] ; then mpi_command="mpirun -np $num_gpu -H localhost:$num_gpu \ --allow-run-as-root -bind-to none -map-by slot \ -x NCCL_DEBUG=INFO \ -x LD_LIBRARY_PATH \ -x PATH -mca pml ob1 -mca btl ^openib" use_hvd="--horovod" else mpi_command="" use_hvd="" fi export GBS=$(expr $train_batch_size \* $num_gpus \* num_accumulation_steps) printf -v TAG "tf_bert_bio_1n_phase2_cased_%s_%s_gbs%d" "$cased" "$precision" $GBS DATESTAMP=`date +'%y%m%d%H%M%S'` LOGFILE=$RESULTS_DIR/$TAG.$DATESTAMP.log printf "Logs written to %s\n" "$LOGFILE" $mpi python3 /workspace/bert/run_pretraining.py \ --input_files_dir=$INPUT_FILES_DIR \ --eval_files_dir=$EVAL_FILES_DIR \ --output_dir=$CHECKPOINTS_DIR \ --bert_config_file=$BERT_CONFIG \ --do_train=True \ --do_eval=True \ --train_batch_size=$train_batch_size \ --eval_batch_size=$eval_batch_size \ --max_seq_length=512 \ --max_predictions_per_seq=80 \ --num_train_steps=$train_steps \ --num_warmup_steps=$warmup_steps \ --save_checkpoints_steps=$save_checkpoint_steps \ --num_accumulation_steps=$num_accumulation_steps \ --learning_rate=$learning_rate \ --report_loss \ $use_hvd $use_xla_tag $use_fp16 \ --init_checkpoint=$INIT_CHECKPOINT |& tee $LOGFILE
PyTorch/Recommendation/DLRM/preproc/gpu
gpu
get_gpu_resources
#! /bin/bash ADDRS=`nvidia-smi --query-gpu=index --format=csv,noheader | sed -e ':a' -e 'N' -e'$!ba' -e 's/\n/","/g'` echo {\"name\": \"gpu\", \"addresses\":[\"$ADDRS\"]}
TensorFlow2/Recommendation/WideAndDeep/tests/feature_specs
feature_specs
less_numerical
channel_spec: label: - clicked map: [] multihot_categorical: - topic_id_list - entity_id_list - category_id_list numerical: - document_id_promo_ctr - publisher_id_promo_ctr - source_id_promo_ctr - document_id_promo_count - publish_time_days_since_published - ad_id_ctr - advertiser_id_ctr - campaign_id_ctr - ad_id_count - publish_time_promo_days_since_published onehot_categorical: - ad_id - document_id - platform - document_id_promo - campaign_id - advertiser_id - source_id - geo_location - geo_location_country - geo_location_state - publisher_id - source_id_promo - publisher_id_promo feature_spec: ad_id: cardinality: 250000 ad_id_count: {} ad_id_ctr: {} advertiser_id: cardinality: 2500 advertiser_id_ctr: {} campaign_id: cardinality: 5000 campaign_id_ctr: {} category_id_list: cardinality: 100 max_hotness: 3 clicked: {} document_id: cardinality: 300000 document_id_promo: cardinality: 100000 document_id_promo_count: {} document_id_promo_ctr: {} entity_id_list: cardinality: 10000 max_hotness: 3 geo_location: cardinality: 2500 geo_location_country: cardinality: 300 geo_location_state: cardinality: 2000 platform: cardinality: 4 publish_time_days_since_published: {} publish_time_promo_days_since_published: {} publisher_id: cardinality: 1000 publisher_id_promo: cardinality: 1000 publisher_id_promo_ctr: {} source_id: cardinality: 4000 source_id_promo: cardinality: 4000 source_id_promo_ctr: {} topic_id_list: cardinality: 350 max_hotness: 3 metadata: {} source_spec: test: - features: - clicked - ad_id - document_id - platform - document_id_promo - campaign_id - advertiser_id - source_id - geo_location - geo_location_country - geo_location_state - publisher_id - source_id_promo - publisher_id_promo - topic_id_list - entity_id_list - category_id_list - document_id_promo_ctr - publisher_id_promo_ctr - source_id_promo_ctr - document_id_promo_count - publish_time_days_since_published - ad_id_ctr - advertiser_id_ctr - campaign_id_ctr - ad_id_count - publish_time_promo_days_since_published files: - valid.csv type: csv train: - features: - clicked - ad_id - document_id - platform - document_id_promo - campaign_id - advertiser_id - source_id - geo_location - geo_location_country - geo_location_state - publisher_id - source_id_promo - publisher_id_promo - topic_id_list - entity_id_list - category_id_list - document_id_promo_ctr - publisher_id_promo_ctr - source_id_promo_ctr - document_id_promo_count - publish_time_days_since_published - ad_id_ctr - advertiser_id_ctr - campaign_id_ctr - ad_id_count - publish_time_promo_days_since_published files: - train.csv type: csv
Tools/PyTorch/TimeSeriesPredictionPlatform
TimeSeriesPredictionPlatform
launch_preproc
# Copyright (c) 2021-2022, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import warnings import hydra warnings.filterwarnings("ignore") @hydra.main(config_path="conf/", config_name="preproc_config") def main(cfg): print(cfg) preprocessor = hydra.utils.instantiate(cfg, _recursive_=False) train, valid, test = preprocessor.preprocess() preprocessor.fit_scalers(train) train = preprocessor.apply_scalers(train) valid = preprocessor.apply_scalers(valid) test = preprocessor.apply_scalers(test) train = preprocessor.impute(train) valid = preprocessor.impute(valid) test = preprocessor.impute(test) preprocessor.save_state() preprocessor.save_datasets(train, valid, test) if __name__ == "__main__": main()
PyTorch/LanguageModeling/BERT/triton/large/runner
runner
start_NVIDIA-DGX-1-(1x-V100-32GB)
# Copyright (c) 2021, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #!/bin/bash # Install Docker . /etc/os-release && \ curl -fsSL https://download.docker.com/linux/debian/gpg | apt-key add - && \ echo "deb [arch=amd64] https://download.docker.com/linux/debian buster stable" > /etc/apt/sources.list.d/docker.list && \ curl -s -L https://nvidia.github.io/nvidia-docker/gpgkey| apt-key add - && \ curl -s -L https://nvidia.github.io/nvidia-docker/$ID$VERSION_ID/nvidia-docker.list > /etc/apt/sources.list.d/nvidia-docker.list && \ apt-get update && \ apt-get install -y docker-ce docker-ce-cli containerd.io nvidia-docker2 # Install packages pip install -r triton/runner/requirements.txt # Evaluate Runner python3 -m "triton.large.runner.__main__" \ --config-path "triton/large/runner/config_NVIDIA-DGX-1-(1x-V100-32GB).yaml" \ --device 0